Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
## Overview
Janganum Gate Multiple Object Detection No Annotations is a dataset for object detection tasks - it contains Fortification Arch Elements annotations for 626 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
Facebook
Twitter-Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.
-Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
## Overview
Images For Annotation is a dataset for object detection tasks - it contains Objs annotations for 6,234 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [Public Domain license](https://creativecommons.org/licenses/Public Domain).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset consists of drone images that were obtained for agricultural field monitoring to detect weeds and crops through computer vision and machine learning approaches. The images were obtained through high-resolution UAVs and annotated using the LabelImg and Roboflow tool. Each image has a corresponding YOLO annotation file that contains bounding box information and class IDs for detected objects. The dataset includes:
Original images in .jpg format with a resolution of 585 × 438 pixels.
Annotation files (.txt) corresponding to each image, following the YOLO format: class_id x_center y_center width height.
A classes.txt file listing the object categories used in labeling (e.g., Weed, Crop).
The dataset is intended for use in machine learning model development, particularly for precision agriculture, weed detection, and plant health monitoring. It can be directly used for training YOLOv7 and other object detection models.
Facebook
TwitterIf you use this dataset, please cite this paper: Puertas, E.; De-Las-Heras, G.; Sánchez-Soriano, J.; Fernández-Andrés, J. Dataset: Variable Message Signal Annotated Images for Object Detection. Data 2022, 7, 41. https://doi.org/10.3390/data7040041 This dataset consists of Spanish road images taken from inside a vehicle, as well as annotations in XML files in PASCAL VOC format that indicate the location of Variable Message Signals within them. Also, a CSV file is attached with information regarding the geographic position, the folder where the image is located, and the text in Spanish. This can be used to train supervised learning computer vision algorithms, such as convolutional neural networks. Throughout this work, the process followed to obtain the dataset, image acquisition, and labeling, and its specifications are detailed. The dataset is constituted of 1216 instances, 888 positives, and 328 negatives, in 1152 jpg images with a resolution of 1280x720 pixels. These are divided into 576 real images and 576 images created from the data-augmentation technique. The purpose of this dataset is to help in road computer vision research since there is not one specifically for VMSs. The folder structure of the dataset is as follows: vms_dataset/ data.csv real_images/ imgs/ annotations/ data-augmentation/ imgs/ annotations/ In which: data.csv: Each row contains the following information separated by commas (,): image_name, x_min, y_min, x_max, y_max, class_name, lat, long, folder, text. real_images: Images extracted directly from the videos. data-augmentation: Images created using data-augmentation imgs: Image files in .jpg format. annotations: Annotation files in .xml format.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
A dataset containing annotations made on images of cultural heritage digital objects from the Rijksmuseum . The annotations resulted from the application of automatic object detection techniques. This dataset was used for the evaluation of the object detection model developed by the project Saint George on a Bike. The service achieved a precision of 79.4% and a recall of 65.7% in this dataset. The dataset contains:
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Old Annotation is a dataset for object detection tasks - it contains Objects annotations for 1,306 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterThe image data was created using the code in the following URL. Annotation is in COCO format, handmade by VOTT. https://www.kaggle.com/code/hidebu/make-image-files-for-annotation
The number of data is 217. One or more from each video_id and each event_attributes were randomly selected. The objects being annotated are the players and the chief referee.
Please Upvote if you would like.👋
Facebook
TwitterNexdata provides high-quality Annotated Imagery Data annotation for bounding box, polygon,segmentation,polyline, key points,image classification and image description. We have handled tons of data for autonomous driving, internet entertainment, retail, surveillance and security and etc.
Facebook
TwitterAttribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The dataset comprises of annotated video frames from positioned in a public space camera. The tracking of each individual in the camera's view has been achieved using the rectangle tool in the Computer Vision Annotation Tool (CVAT).
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2Fc5a8dc4f63fe85c64a5fead10fad3031%2Fpersons_gif.gif?generation=1690705558283123&alt=media" alt="">
images directory houses the original video frames, serving as the primary source of raw data. annotations.xml file provides the detailed annotation data for the images. boxes directory contains frames that visually represent the bounding box annotations, showing the locations of the tracked individuals within each frame. These images can be used to understand how the tracking has been implemented and to visualize the marked areas for each individual.The annotations are represented as rectangle bounding boxes that are placed around each individual. Each bounding box annotation contains the position ( xtl-ytl-xbr-ybr coordinates ) for the respective box within the frame.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2F4f274551e10db2754c4d8a16dff97b33%2Fcarbon%20(10).png?generation=1687776281548084&alt=media" alt="">
🚀 You can learn more about our high-quality unique datasets here
keywords: multiple people tracking, human detection dataset, object detection dataset, people tracking dataset, tracking human object interactions, human Identification tracking dataset, people detection annotations, detecting human in a crowd, human trafficking dataset, deep learning object tracking, multi-object tracking dataset, labeled web tracking dataset, large-scale object tracking dataset
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Facebook
Twitterhttps://dataintelo.com/privacy-and-policyhttps://dataintelo.com/privacy-and-policy
The global image annotation tool market size is projected to grow from approximately $700 million in 2023 to an estimated $2.5 billion by 2032, exhibiting a remarkable compound annual growth rate (CAGR) of 15.2% over the forecast period. The surging demand for machine learning and artificial intelligence applications is driving this robust market expansion. Image annotation tools are crucial for training AI models to recognize and interpret images, a necessity across diverse industries.
One of the key growth factors fueling the image annotation tool market is the rapid adoption of AI and machine learning technologies across various sectors. Organizations in healthcare, automotive, retail, and many other industries are increasingly leveraging AI to enhance operational efficiency, improve customer experiences, and drive innovation. Accurate image annotation is essential for developing sophisticated AI models, thereby boosting the demand for these tools. Additionally, the proliferation of big data analytics and the growing necessity to manage large volumes of unstructured data have amplified the need for efficient image annotation solutions.
Another significant driver is the increasing use of autonomous systems and applications. In the automotive industry, for instance, the development of autonomous vehicles relies heavily on annotated images to train algorithms for object detection, lane discipline, and navigation. Similarly, in the healthcare sector, annotated medical images are indispensable for developing diagnostic tools and treatment planning systems powered by AI. This widespread application of image annotation tools in the development of autonomous systems is a critical factor propelling market growth.
The rise of e-commerce and the digital retail landscape has also spurred demand for image annotation tools. Retailers are using these tools to optimize visual search features, personalize shopping experiences, and enhance inventory management through automated recognition of products and categories. Furthermore, advancements in computer vision technology have expanded the capabilities of image annotation tools, making them more accurate and efficient, which in turn encourages their adoption across various industries.
Data Annotation Software plays a pivotal role in the image annotation tool market by providing the necessary infrastructure for labeling and categorizing images efficiently. These software solutions are designed to handle various annotation tasks, from simple bounding boxes to complex semantic segmentation, enabling organizations to generate high-quality training datasets for AI models. The continuous advancements in data annotation software, including the integration of machine learning algorithms for automated labeling, have significantly enhanced the accuracy and speed of the annotation process. As the demand for AI-driven applications grows, the reliance on robust data annotation software becomes increasingly critical, supporting the development of sophisticated models across industries.
Regionally, North America holds the largest share of the image annotation tool market, driven by significant investments in AI and machine learning technologies and the presence of leading technology companies. Europe follows, with strong growth supported by government initiatives promoting AI research and development. The Asia Pacific region presents substantial growth opportunities due to the rapid digital transformation in emerging economies and increasing investments in technology infrastructure. Latin America and the Middle East & Africa are also expected to witness steady growth, albeit at a slower pace, due to the gradual adoption of advanced technologies.
The image annotation tool market by component is segmented into software and services. The software segment dominates the market, encompassing a variety of tools designed for different annotation tasks, from simple image labeling to complex polygonal, semantic, or instance segmentation. The continuous evolution of software platforms, integrating advanced features such as automated annotation and machine learning algorithms, has significantly enhanced the accuracy and efficiency of image annotations. Furthermore, the availability of open-source annotation tools has lowered the entry barrier, allowing more organizations to adopt these technologies.
Services associated with image ann
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Vehicle Detection Dataset
This dataset is designed for vehicle detection tasks, featuring a comprehensive collection of images annotated for object detection. This dataset, originally sourced from Roboflow (https://universe.roboflow.com/object-detection-sn8ac/ai-traffic-system), was exported on May 29, 2025, at 4:59 PM GMT and is now publicly available on Kaggle under the CC BY 4.0 license.
../train/images../valid/images../test/imagesThis dataset was created and exported via Roboflow, an end-to-end computer vision platform that facilitates collaboration, image collection, annotation, dataset creation, model training, and deployment. The dataset is part of the ai-traffic-system project (version 1) under the workspace object-detection-sn8ac. For more details, visit: https://universe.roboflow.com/object-detection-sn8ac/ai-traffic-system/dataset/1.
This dataset is ideal for researchers, data scientists, and developers working on vehicle detection and traffic monitoring systems. It can be used to: - Train and evaluate deep learning models for object detection, particularly using the YOLOv11 framework. - Develop AI-powered traffic management systems, autonomous driving applications, or urban mobility solutions. - Explore computer vision techniques for real-world traffic scenarios.
For advanced training notebooks compatible with this dataset, check out: https://github.com/roboflow/notebooks. To explore additional datasets and pre-trained models, visit: https://universe.roboflow.com.
The dataset is licensed under CC BY 4.0, allowing for flexible use, sharing, and adaptation, provided appropriate credit is given to the original source.
This dataset is a valuable resource for building robust vehicle detection models and advancing computer vision applications in traffic systems.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
A curated dataset for human face detection designed to enhance AI and computer vision models. Includes diverse images of human faces across demographics, lighting conditions, and angles — suitable for facial recognition, security systems, and emotion analysis applications.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
ODDS Smart Building Depth Dataset
The goal of this dataset is to facilitate research focusing on recognizing objects in smart buildings using the depth sensor mounted at the ceiling. This dataset contains annotations of depth images for eight frequently seen object classes. The classes are: person, backpack, laptop, gun, phone, umbrella, cup, and box.
We collected data from two settings. We had Kinect mounted at a 9.3 feet ceiling near to a 6 feet wide door. We also used a tripod with a horizontal extender holding the kinect at a similar height looking downwards. We asked about 20 volunteers to enter and exit a number of times each in different directions (3 times walking straight, 3 times walking towards left side, 3 times walking towards right side) holding objects in many different ways and poses underneath the Kinect. Each subject was using his/her own backpack, purse, laptop, etc. As a result, we considered varieties within the same object, e.g., for laptops, we considered Macbooks, HP laptops, Lenovo laptops of different years and models, and for backpacks, we considered backpacks, side bags, and purse of women. We asked the subjects to walk while holding it in many ways, e.g., for laptop, the laptop was fully open, partially closed, and fully closed while carried. Also, people hold laptops in front and side of their bodies, and underneath their elbow. The subjects carried their backpacks in their back, in their side at different levels from foot to shoulder. We wanted to collect data with real guns. However, bringing real guns to the office is prohibited. So, we obtained a few nerf guns and the subjects were carrying these guns pointing it to front, side, up, and down while walking.
The Annotated dataset is created following the structure of Pascal VOC devkit, so that the data preparation becomes simple and it can be used quickly with different with object detection libraries that are friendly to Pascal VOC style annotations (e.g. Faster-RCNN, YOLO, SSD). The annotated data consists of a set of images; each image has an annotation file giving a bounding box and object class label for each object in one of the eight classes present in the image. Multiple objects from multiple classes may be present in the same image. The dataset has 3 main directories:
1)DepthImages: Contains all the images of training set and validation set.
2)Annotations: Contains one xml file per image file, (e.g., 1.xml for image file 1.png). The xml file includes the bounding box annotations for all objects in the corresponding image.
3)ImagesSets: Contains two text files training_samples.txt and testing_samples.txt. The training_samples.txt file has the name of images used in training and the testing_samples.txt has the name of images used for testing. (We randomly choose 80%, 20% split)
The un-annotated data consists of several set of depth images. No ground-truth annotation is available for these images yet. These un-annotated sets contain several challenging scenarios and no data has been collected from this office during annotated dataset construction. Hence, it will provide a way to test generalization performance of the algorithm.
If you use ODDS Smart Building dataset in your work, please cite the following reference in any publications: @inproceedings{mithun2018odds, title={ODDS: Real-Time Object Detection using Depth Sensors on Embedded GPUs}, author={Niluthpol Chowdhury Mithun and Sirajum Munir and Karen Guo and Charles Shelton}, booktitle={ ACM/IEEE Conference on Information Processing in Sensor Networks (IPSN)}, year={2018}, }
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Thermal Image Annotation is a dataset for object detection tasks - it contains Objects annotations for 1,911 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterThis dataset was created by Nitish Katkade
Facebook
TwitterThis is a modified version of the "Vehicle Dataset with YOLOv5 Annotations" Dataset by İLKER GALIP ATA.
The original dataset had some annotations missing and mismatched files, which I have removed.
This dataset is particularly for YOLOv5.
This dataset is identical to the original dataset, there are two directories:
Train
Test
Each contains images and their labels in separate folders.
Note: The annotations and labels are in the same file for every image in text format.
Facebook
TwitterAttribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Doing manual annotations can sometimes be resource heavy, depending on the amount of data. This dataset was designed to created to use in conjunction with a semi-automatic annotation method based on linear interpolation. The dataset contains 799 images, where 679 lies in the trainingset, and the rest lies in the validationset. The images are taken from 6 different video streams, where a remote controlled wheel loader approaches a miniature dump truck at different angles. 4 of the videos are used in the trainingset. The labels can contain up to 5 classes which are:
0 - front wheel
1 - middle wheel
2 - back wheel
3 - tipping body
4 - cap
This dataset was used to train a YOLOv3 model, hence the labels will be written in the YOLO labeling format.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
## Overview
Janganum Gate Multiple Object Detection No Annotations is a dataset for object detection tasks - it contains Fortification Arch Elements annotations for 626 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).