-Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.
-Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Object Detection Annotations is a dataset for object detection tasks - it contains Objects annotations for 302 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
The "26 Class Object Detection Dataset" comprises a comprehensive collection of images annotated with objects belonging to 26 distinct classes. Each class represents a common urban or outdoor element encountered in various scenarios. The dataset includes the following classes:
Bench Bicycle Branch Bus Bushes Car Crosswalk Door Elevator Fire Hydrant Green Light Gun Motorcycle Person Pothole Rat Red Light Scooter Stairs Stop Sign Traffic Cone Train Tree Truck Umbrella Yellow Light These classes encompass a wide range of objects commonly encountered in urban and outdoor environments, including transportation vehicles, traffic signs, pedestrian-related elements, and natural features. The dataset serves as a valuable resource for training and evaluating object detection models, particularly those focused on urban scene understanding and safety applications.
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Dataset Description: Car Object Detection in Road Traffic
Overview:
This dataset is designed for car object detection in road traffic scenes (Images with shape 1080x1920x3). The dataset is derived from publicly available video content on YouTube, specifically from the video with the Creative Commons Attribution license, available here.
https://youtu.be/MNn9qKG2UFI?si=uJz_WicTCl8zfrVl" alt="youtube video">
Source:
Annotation Details:
Use Cases:
Acknowledgments: We acknowledge and thank the creator of the original video for making it available under a Creative Commons Attribution license. Their contribution enables the development of datasets and research in the field of computer vision and object detection.
Disclaimer: This dataset is provided for educational and research purposes and should be used in compliance with YouTube's terms of service and the Creative Commons Attribution license.
Nexdata provides high-quality Annotated Imagery Data annotation for bounding box, polygon,segmentation,polyline, key points,image classification and image description. We have handled tons of data for autonomous driving, internet entertainment, retail, surveillance and security and etc.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is a curated collection of images featuring various cattle body parts aimed at facilitating object detection tasks. The dataset contains a total of 428 high-quality photos, meticulously annotated with three distinct classes: "Back," "Head," and "Leg."
The dataset can be downloaded using this link. The dataset is also available at Roboflow Universe.
A YOLOv7X model has been trained using the dataset and achieved a mAP of 99.6%. You can access the trained weights through this link.
Accurate and reliable identification of different cattle body parts is crucial for various agricultural and veterinary applications. This dataset aims to provide a valuable resource for researchers, developers, and enthusiasts working on object detection tasks involving cattle, ultimately contributing to advancements in livestock management, health monitoring, and related fields.
📦 Cattle_Body_Parts_OD.zip
┣ 📂 images
┃ ┣ 📜 image1.jpg
┃ ┣ 📜 image2.jpg
┃ ┗ ...
┗ 📂 annotations
┣ 📜 image1.json
┣ 📜 image2.json
â”— ...
Each annotation file corresponds to an image in the dataset and is formatted as per the LabelMe JSON standard. These annotations define the bounding box coordinates for each labeled body part, enabling straightforward integration into object detection pipelines.
This work is licensed under a Creative Commons Attribution 4.0 International License.
This dataset has been collected from publicly available sources. I do not claim ownership of the data and have no intention of infringing on any copyright. The material contained in this dataset is copyrighted to their respective owners. I have made every effort to ensure the data is accurate and complete, but I cannot guarantee its accuracy or completeness. If you believe any data in this dataset infringes on your copyright, please get in touch with me immediately so I can take appropriate action.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Fall Annotation is a dataset for object detection tasks - it contains Fall annotations for 1,497 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Text Annotation is a dataset for object detection tasks - it contains Kacamata Masker Tas Sepatu Topi CW6W annotations for 6,400 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Description:
👉 Download the dataset here
This dataset consists of a diverse collection of images featuring Paimon, a popular character from the game Genshin Impact. The images have been sourced from in-game gameplay footage and capture Paimon from various angles and in different sizes (scales), making the dataset suitable for training YOLO object detection models.
The dataset provides a comprehensive view of Paimon in different lighting conditions, game environments, and positions, ensuring the model can generalize well to similar characters or object detection tasks. While most annotations are accurately labeled, a small number of annotations may include minor inaccuracies due to manual labeling errors. This is ideal for researchers and developers working on character recognition, object detection in gaming environments, or other AI vision tasks.
Download Dataset
Dataset Features:
Image Format: .jpg files in 640×320 resolution.
Annotation Format: .txt files in YOLO format, containing bounding box data with:
class_id
x_center
y_center
width
height
Use Cases:
Character Detection in Games: Train YOLO models to detect and identify in-game characters or NPCs.
Gaming Analytics: Improve recognition of specific game elements for AI-powered game analytics tools.
Research: Contribute to academic research focused on object detection or computer vision in animated and gaming environments.
Data Structure:
Images: High-quality .jpg images captured from multiple perspectives, ensuring robust model training across various orientations and lighting scenarios.
Annotations: Each image has an associated .txt file that follows the YOLO format. The annotations are structured to include class identification, object location (center coordinates), and
bounding box dimensions.
Key Advantages:
Varied Angles and Scales: The dataset includes Paimon from multiple perspectives, aiding in creating more versatile and adaptable object detection models.
Real-World Scenario: Extracted from actual gameplay footage, the dataset simulates real-world detection challenges such as varying backgrounds, motion blur, and changing character scales.
Training Ready: Suitable for training YOLO models and other deep learning frameworks that require object detection capabilities.
This dataset is sourced from Kaggle.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The Hard Hat
dataset is an object detection dataset of workers in workplace settings that require a hard hat. Annotations also include examples of just "person" and "head," for when an individual may be present without a hard hart.
The original dataset has a 75/25 train-test split.
Example Image:
https://i.imgur.com/7spoIJT.png" alt="Example Image">
One could use this dataset to, for example, build a classifier of workers that are abiding safety code within a workplace versus those that may not be. It is also a good general dataset for practice.
Use the fork
or Download this Dataset
button to copy this dataset to your own Roboflow account and export it with new preprocessing settings (perhaps resized for your model's desired format or converted to grayscale), or additional augmentations to make your model generalize better. This particular dataset would be very well suited for Roboflow's new advanced Bounding Box Only Augmentations.
Image Preprocessing | Image Augmentation | Modify Classes
* v1
(resize-416x416-reflect): generated with the original 75/25 train-test split | No augmentations
* v2
(raw_75-25_trainTestSplit): generated with the original 75/25 train-test split | These are the raw, original images
* v3
(v3): generated with the original 75/25 train-test split | Modify Classes used to drop person
class | Preprocessing and Augmentation applied
* v5
(raw_HeadHelmetClasses): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person
class
* v8
(raw_HelmetClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head
and person
classes
* v9
(raw_PersonClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head
and helmet
classes
* v10
(raw_AllClasses): generated with a 70/20/10 train/valid/test split | These are the raw, original images
* v11
(augmented3x-AllClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied | 3x image generation | Trained with Roboflow's Fast Model
* v12
(augmented3x-HeadHelmetClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person
class | 3x image generation | Trained with Roboflow's Fast Model
* v13
(augmented3x-HeadHelmetClasses-AccurateModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person
class | 3x image generation | Trained with Roboflow's Accurate Model
* v14
(raw_HeadClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person
class, and remap/relabel helmet
class to head
Choosing Between Computer Vision Model Sizes | Roboflow Train
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.
Developers reduce 50% of their code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.
https://cdla.io/sharing-1-0/https://cdla.io/sharing-1-0/
Imagine a world where machines see like humans—where your model doesn’t just scan pixels, but truly understands faces.
Face detection is the gateway to computer vision. From unlocking smartphones to powering surveillance systems, it’s the first step toward machines that understand the world as we do.
This dataset is crafted for creators, engineers, and visionaries who want to build models that don't just see — they recognize.
Originally prepared and exported using Roboflow, this dataset includes a diverse collection of face images, carefully annotated for object detection tasks. It’s designed to help you train accurate, real-time face detection models using cutting-edge deep learning architectures.
The structure is simple and scalable — optimized for quick experimentation and production-level deployment.
Focused and minimal. One class. One purpose.
Because face detection is more than bounding boxes — it’s about interaction, identity, and trust. Whether you’re building an AI that understands presence, or a system that reacts to people in real-time, this dataset gives you the data to begin.
This dataset is released under Creative Commons Zero (CC0 1.0). Use it freely — in research, in production, or anywhere your ideas take you.
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Semantic PASCAL-Part dataset
The Semantic PASCAL-Part dataset is the RDF version of the famous PASCAL-Part dataset used for object detection in Computer Vision. Each image is annotated with bounding boxes containing a single object. Couples of bounding boxes are annotated with the part-whole relationship. For example, the bounding box of a car has the part-whole annotation with the bounding boxes of its wheels.
This original release joins Computer Vision with Semantic Web as the objects in the dataset are aligned with concepts from:
The provided Python 3 code (see the GitHub repo) is able to browse the dataset and convert it in RDF knowledge graph format. This new format easily allows the fostering of research in both Semantic Web and Machine Learning fields.
Structure of the semantic PASCAL-Part Dataset
This is the folder structure of the dataset:
semanticPascalPart
: it contains the refined images and annotations (e.g., small specific parts are merged into bigger parts) of the PASCAL-Part dataset in Pascal-voc style.
Annotations_set
: the test set annotations in .xml
format. For further information See the PASCAL VOC format here.Annotations_trainval
: the train and validation set annotations in .xml
format. For further information See the PASCAL VOC format here.JPEGImages_test
: the test set images in .jpg
format.JPEGImages_trainval
: the train and validation set images in .jpg
format.test.txt
: the 2416 image filenames in the test set.trainval.txt
: the 7687 image filenames in the train and validation set.The PASCAL-Part Ontology
The PASCAL-Part OWL ontology formalizes, through logical axioms, the part-of relationship between whole objects (22 classes) and their parts (39 classes). The ontology contains 85 logical axiomns in Description Logic in (for example) the following form:
Every potted_plant has exactly 1 plant AND
has exactly 1 pot
We provide two versions of the ontology: with and without cardinality constraints in order to allow users to experiment with or without them. The WordNet alignment is encoded in the ontology as annotations. We further provide the WordNet_Yago_alignment.csv
file with both WordNet and Yago alignments.
The ontology can be browsed with many Semantic Web tools such as:
Citing semantic PASCAL-Part
If you use semantic PASCAL-Part in your research, please use the following BibTeX entry
@article{DBLP:journals/ia/DonadelloS16,
author = {Ivan Donadello and
Luciano Serafini},
title = {Integration of numeric and symbolic information for semantic image
interpretation},
journal = {Intelligenza Artificiale},
volume = {10},
number = {1},
pages = {33--47},
year = {2016}
}
https://www.marketresearchforecast.com/privacy-policyhttps://www.marketresearchforecast.com/privacy-policy
The global image tagging and annotation services market is experiencing robust growth, driven by the increasing adoption of artificial intelligence (AI) and machine learning (ML) across diverse sectors. The market, estimated at $2.5 billion in 2025, is projected to expand at a Compound Annual Growth Rate (CAGR) of 18% from 2025 to 2033, reaching an estimated $10 billion by 2033. This significant expansion is fueled by several key factors. The automotive industry leverages image tagging and annotation for autonomous vehicle development, requiring vast amounts of labeled data for training AI algorithms. Similarly, the retail and e-commerce sectors utilize these services for image search, product recognition, and improved customer experiences. The healthcare industry benefits from advancements in medical image analysis, while the government and security sectors employ image annotation for surveillance and security applications. The rising availability of high-quality data, coupled with the decreasing cost of annotation services, further accelerates market growth. However, challenges remain. Data privacy concerns and the need for high-accuracy annotation can pose significant hurdles. The demand for specialized skills in data annotation also contributes to a potential bottleneck in the market's growth trajectory. Overcoming these challenges requires a collaborative approach, involving technological advancements in automation and the development of robust data governance frameworks. The market segmentation, encompassing various annotation types (image classification, object recognition/detection, boundary recognition, segmentation) and application areas (automotive, retail, BFSI, government, healthcare, IT, transportation, etc.), presents diverse opportunities for market players. The competitive landscape includes a mix of established players and emerging firms, each offering specialized services and targeting specific market segments. North America currently holds the largest market share due to early adoption of AI and ML technologies, while Asia-Pacific is anticipated to witness rapid growth in the coming years.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
## Overview
Annotation is a dataset for object detection tasks - it contains Objects annotations for 1,229 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
Overview This dataset is a collection of 2,000 Licensed and 8,000 HD damaged car images that are ready to use for optimizing the accuracy of computer vision models. All of the contents is sourced from PIXTA's stock library of 100M+ Asian-featured images and videos. PIXTA is the largest platform of visual materials in the Asia Pacific region offering fully-managed services, high quality contents and data, and powerful tools for businesses & organisations to enable their creative and machine learning projects.
Use cases for damaged car images (object detection data) The 2,000 Licensed and 8,000 HD Images of damaged car could be used for various AI & Computer Vision models: Damage Inspection, Insurance Value Evaluation, Residual Value Forecast,... Each data set is supported by both AI and human review process to ensure labelling consistency and accuracy. Contact us for more custom datasets.
Annotation Annotation is available for this dataset on demand, including:
Bounding box
Polygon
Segmentation ...
About PIXTA PIXTASTOCK is the largest Asian-featured stock platform providing data, contents, tools and services since 2005. PIXTA experiences 15 years of integrating advanced AI technology in managing, curating, processing over 100M visual materials and serving global leading brands for their creative and data demands. Visit us at https://www.pixta.ai/ or contact via our email contact@pixta.ai.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset consists of microscopic images of blood cells specifically designed for the detection of White Blood Cells (WBC). It is intended for object detection tasks where the goal is to accurately locate and identify WBCs within blood smear images. Researchers and developers can utilize this data to train machine learning models for medical applications such as automated blood cell analysis.
Images: The dataset contains high-resolution microscopic images of blood smears, where WBCs are scattered among Red Blood Cells (RBCs) and platelets. Each image is annotated with bounding boxes around the WBCs.
Annotations: The annotations are provided in YOLO format, where each bounding box is associated with a label for WBC.
images/: Contains the blood cell images in .jpg or .png format. labels/: Contains the annotation files in .txt format (YOLO format), with each file corresponding to an image. Image Size: Varies, but all images are in high resolution suitable for detection tasks.
Medical Image Analysis: This dataset can be used to build models for the automated detection of WBCs, which is a crucial step in diagnosing various blood-related disorders. Object Detection: Ideal for testing object detection algorithms like YOLO, Faster R-CNN, or SSD. Acknowledgments This dataset is created using publicly available microscopic blood cell images, annotated for educational and research purposes. It can be used for developing machine learning models for academic research, prototyping medical applications, or object detection benchmarking.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Data
The dataset consist of 5538 images of public spaces, annotated with steps, stairs, ramps and grab bars for stairs and ramps. The dataset has annotations 3564 of steps, 1492 of stairs, 143 of ramps and 922 of grab bars.
Each step annotation is attributed with an estimate of the height of the step, as falling into one of three categories: less than 3cm, 3cm to 7cm or more than 7cm. Additionally it is attributed with a 'type', with the possibilities 'doorstep', 'curb' or 'other'.
Stair annotations are attributed with the number of steps in the stair.
Ramps are attributed with an estimate of their width, also falling into three categories: less than 50cm, 50cm to 100cm and more than 100cm.
In order to preserve all additional attributes of the labels, the data is published in the CVAT XML format for images.
Annotating Process
The labelling has been done using bounding boxes around the objects. This format is compatible with many popular object detection models, e.g. the YOLO object model. A bounding box is placed so it contains exactly the visible part of the respective objects. This implies that only objects that are visible in the photo are annotated. This means in particular a photo of a stair or step from above, where the object cannot be seen, have not been annotated, even when a human viewer can possibly infer that there is a stair or a step from other features in the photo.
Steps
A step is annotated, when there is an vertical increment that functions as a passage between two surface areas intended human or vehicle traffic. This means that we have not included:
In particular, the bounding box of a step object contains exactly the incremental part of the step, but does not extend into the top or bottom horizontal surface any more than necessary to enclose entirely the incremental part. This has been chosen for consistency reasons, as including parts of the horizontal surfaces would imply a non-trivial choice of how much to include, which we deemed would most likely lead to more inconstistent annotations.
The height of the steps are estimated by the annotators, and are therefore not guarranteed to be accurate.
The type of the steps typically fall into the category 'doorstep' or 'curb'. Steps that are in a doorway, entrance or likewise are attributed as doorsteps. We also include in this category steps that are immediately leading to a doorway within a proximity of 1-2m. Steps between different types of pathways, e.g. between streets and sidewalks, are annotated as curbs. Any other type of step are annotated with 'other'. Many of the 'other' steps are for example steps to terraces.
Stairs
The stair label is used whenever two or more steps directly follow each other in a consistent pattern. All vertical increments are enclosed in the bounding box, as well as intermediate surfaces of the steps. However the top and bottom surface is not included more than necessary for the same reason as for steps, as described in the previous section.
The annotator counts the number of steps, and attribute this to the stair object label.
Ramps
Ramps have been annotated when a sloped passage way has been placed or built to connect two surface areas intended for human or vehicle traffic. This implies the same considerations as with steps. Alike also only the sloped part of a ramp is annotated, not including the bottom or top surface area.
For each ramp, the annotator makes an assessment of the width of the ramp in three categories: less than 50cm, 50cm to 100cm and more than 100cm. This parameter is visually hard to assess, and sometimes impossible due to the view of the ramp.
Grab Bars
Grab bars are annotated for hand rails and similar that are in direct connection to a stair or a ramp. While horizontal grab bars could also have been included, this was omitted due to the implied ambiguities of fences and similar objects. As the grab bar was originally intended as an attributal information to stairs and ramps, we chose to keep this focus. The bounding box encloses the part of the grab bar that functions as a hand rail for the stair or ramp.
Usage
As is often the case when annotating data, much information depends on the subjective assessment of the annotator. As each data point in this dataset has been annotated only by one person, caution should be taken if the data is applied.
Generally speaking, the mindset and usage guiding the annotations have been wheelchair accessibility. While we have strived to annotate at an object level, hopefully making the data more widely applicable than this, we state this explicitly as it may have swayed untrivial annotation choices.
The attributal data, such as step height or ramp width are highly subjective estimations. We still provide this data to give a post-hoc method to adjust which annotations to use. E.g. for some purposes, one may be interested in detecting only steps that are indeed more than 3cm. The attributal data makes it possible to sort away the steps less than 3cm, so a machine learning algorithm can be trained on this more appropriate dataset for that use case. We stress however, that one cannot expect to train accurate machine learning algorithms inferring the attributal data, as this is not accurate data in the first place.
We hope this dataset will be a useful building block in the endeavours for automating barrier detection and documentation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset was collected from the MyNursingHome dataset, available at https://data.mendeley.com/datasets/fpctx3svzd/1 , and curated to develop a synthetic indoor object detection dataset for autonomous mobile robots, or robots, for supporting researchers in detecting and classifying objects for computer vision and pattern recognition. From the original dataset containing 25 object categories, we selected six key categories—basket bin (499 images), sofa (499 images), human (499 images), table (500 images), chair (496 images), and door (500 images). Initially, we collected a total of 2,993 images from these categories; however, during the annotation process using Roboflow, we rejected 1 sofa, 10 tables, 9 chairs, and 12 door images due to quality concerns, such as poor image resolution or difficulty in identifying the object, resulting in a final dataset of 2,961 images. To ensure an effective training pipeline, we divided the dataset into 70% training (2,073 images), 20% validation (591 images), and 10% test (297 images). Preprocessing steps included auto-orientation and resizing all images to 640×640 pixels to maintain uniformity. To improve generalization for real-world applications, we applied data augmentation techniques, including horizontal and vertical flipping, 90-degree rotations (clockwise, counter-clockwise, and upside down), random rotations within -15° to +15°, shearing within ±10° horizontally and vertically, and brightness adjustments between -15% and +15%. This augmentation process expanded the dataset to 7,107 images, with 6,219 images for training (88%), 597 for validation (8%), and 297 for testing (4%). Moreover, this well-annotated, preprocessed, and augmented dataset significantly improves object detection performance in indoor settings.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
A YOLO Annotated Wind Turbine Surface Damage is a dataset of wind turbine surface damage composed of images from DTU - Drone inspection images of wind turbine dataset, split into 586x371 pixel images with YOLO format annotations for Dirt and Damage. The dataset consists of 13000 images, just under 3000 of which have instances of one of the two classes.
-Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.
-Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001