Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Diverse Image Collection: Our dataset encompasses a wide range of general images covering various categories such as objects, scenes, people, and more. The images are carefully curated to offer a rich source of visual data.
Sindhi Language Titles: One of the distinctive features of our dataset is the inclusion of Sindhi language titles for each image.
Annotations in YOLO Format: To facilitate your object detection tasks, we have meticulously annotated the images in YOLO format, making it compatible with the YOLOv3 or YOLOv4 models. This ensures that you can jump right into training your model without the hassle of converting annotations.
Comprehensive Metadata: Each image in the dataset is accompanied by a YAML file providing additional metadata, including information about the image source, date of capture, and any relevant context that may be useful for your research.
By publishing this YOLO-style dataset with Sindhi language titles, we aim to contribute to the machine learning and computer vision community, fostering innovation and inclusivity in the field. We encourage you to explore, experiment, and create cutting-edge models using this dataset, and we look forward to seeing the incredible projects that emerge from it.
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
The dataset is structured for person object detection tasks, containing separate directories for training, validation, and testing. Each split has an images folder with corresponding images and a labels folder with annotation files.
Train Set: Contains images and annotations for model training.
Validation Set: Includes images and labels for model evaluation during training.
Test Set: Provides unseen images and labels for final model performance assessment.
Each annotation file (TXT format) corresponds to an image and likely contains bounding box coordinates and class labels. This structure follows standard object detection dataset formats, ensuring easy integration with detection models like yolo,RT-DETR.
๐ dataset/ โโโ ๐ train/ โ โโโ ๐ images/ โ โ โโโ ๐ผ image1.jpg (Training image) โ โ โโโ ๐ผ image2.jpg (Training image) โ โโโ ๐ labels/ โ โ โโโ ๐ image1.txt (Annotation for image1.jpg) โ โ โโโ ๐ image2.txt (Annotation for image2.jpg) โ โโโ ๐ val/ โ โโโ ๐ images/ โ โ โโโ ๐ผ image3.jpg (Validation image) โ โ โโโ ๐ผ image4.jpg (Validation image) โ โโโ ๐ labels/ โ โ โโโ ๐ image3.txt (Annotation for image3.jpg) โ โ โโโ ๐ image4.txt (Annotation for image4.jpg) โ โโโ ๐ test/ โ โโโ ๐ images/ โ โ โโโ ๐ผ image5.jpg (Test image) โ โ โโโ ๐ผ image6.jpg (Test image) โ โโโ ๐ labels/ โ โ โโโ ๐ image5.txt (Annotation for image5.jpg) โ โ โโโ ๐ image6.txt (Annotation for image6.jpg)
Facebook
Twitter-Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.
-Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
## Overview
YOLO Coco Data Format is a dataset for object detection tasks - it contains Objects annotations for 692 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [Public Domain license](https://creativecommons.org/licenses/Public Domain).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset consists of drone images that were obtained for agricultural field monitoring to detect weeds and crops through computer vision and machine learning approaches. The images were obtained through high-resolution UAVs and annotated using the LabelImg and Roboflow tool. Each image has a corresponding YOLO annotation file that contains bounding box information and class IDs for detected objects. The dataset includes:
Original images in .jpg format with a resolution of 585 ร 438 pixels.
Annotation files (.txt) corresponding to each image, following the YOLO format: class_id x_center y_center width height.
A classes.txt file listing the object categories used in labeling (e.g., Weed, Crop).
The dataset is intended for use in machine learning model development, particularly for precision agriculture, weed detection, and plant health monitoring. It can be directly used for training YOLOv7 and other object detection models.
Facebook
TwitterIf you use this dataset, please cite this paper: Puertas, E.; De-Las-Heras, G.; Sรกnchez-Soriano, J.; Fernรกndez-Andrรฉs, J. Dataset: Variable Message Signal Annotated Images for Object Detection. Data 2022, 7, 41. https://doi.org/10.3390/data7040041 This dataset consists of Spanish road images taken from inside a vehicle, as well as annotations in XML files in PASCAL VOC format that indicate the location of Variable Message Signals within them. Also, a CSV file is attached with information regarding the geographic position, the folder where the image is located, and the text in Spanish. This can be used to train supervised learning computer vision algorithms, such as convolutional neural networks. Throughout this work, the process followed to obtain the dataset, image acquisition, and labeling, and its specifications are detailed. The dataset is constituted of 1216 instances, 888 positives, and 328 negatives, in 1152 jpg images with a resolution of 1280x720 pixels. These are divided into 576 real images and 576 images created from the data-augmentation technique. The purpose of this dataset is to help in road computer vision research since there is not one specifically for VMSs. The folder structure of the dataset is as follows: vms_dataset/ data.csv real_images/ imgs/ annotations/ data-augmentation/ imgs/ annotations/ In which: data.csv: Each row contains the following information separated by commas (,): image_name, x_min, y_min, x_max, y_max, class_name, lat, long, folder, text. real_images: Images extracted directly from the videos. data-augmentation: Images created using data-augmentation imgs: Image files in .jpg format. annotations: Annotation files in .xml format.
Facebook
TwitterAttribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The dataset comprises of annotated video frames from positioned in a public space camera. The tracking of each individual in the camera's view has been achieved using the rectangle tool in the Computer Vision Annotation Tool (CVAT).
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2Fc5a8dc4f63fe85c64a5fead10fad3031%2Fpersons_gif.gif?generation=1690705558283123&alt=media" alt="">
images directory houses the original video frames, serving as the primary source of raw data. annotations.xml file provides the detailed annotation data for the images. boxes directory contains frames that visually represent the bounding box annotations, showing the locations of the tracked individuals within each frame. These images can be used to understand how the tracking has been implemented and to visualize the marked areas for each individual.The annotations are represented as rectangle bounding boxes that are placed around each individual. Each bounding box annotation contains the position ( xtl-ytl-xbr-ybr coordinates ) for the respective box within the frame.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2F4f274551e10db2754c4d8a16dff97b33%2Fcarbon%20(10).png?generation=1687776281548084&alt=media" alt="">
๐ You can learn more about our high-quality unique datasets here
keywords: multiple people tracking, human detection dataset, object detection dataset, people tracking dataset, tracking human object interactions, human Identification tracking dataset, people detection annotations, detecting human in a crowd, human trafficking dataset, deep learning object tracking, multi-object tracking dataset, labeled web tracking dataset, large-scale object tracking dataset
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is the full 2017 COCO object detection dataset (train and valid), which is a subset of the most recent 2020 COCO object detection dataset.
COCO is a large-scale object detection, segmentation, and captioning dataset of many object types easily recognizable by a 4-year-old. The data is initially collected and published by Microsoft. The original source of the data is here and the paper introducing the COCO dataset is here.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Welcome to the Vehicle Detection Image Dataset! This dataset is meticulously curated for object detection and tracking tasks, with a specific focus on vehicle detection. It serves as a valuable resource for researchers, developers, and enthusiasts seeking to advance the capabilities of computer vision systems.
The primary aim of this dataset is to facilitate precise object detection tasks, particularly in identifying and tracking vehicles within images. Whether you are engaged in academic research, developing commercial applications, or exploring the frontiers of computer vision, this dataset provides a solid foundation for your projects.
Both versions of the dataset undergo essential preprocessing steps, including resizing and orientation adjustments. Additionally, the Apply_Grayscale version undergoes augmentation to introduce grayscale variations, thereby enriching the dataset and improving model robustness.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F14850461%2F4f23bd8094c892d1b6986c767b42baf4%2Fv2.png?generation=1712264632232641&alt=media" alt="">
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F14850461%2Fbfb10eb2a4db31a62eb4615da824c387%2Fdetails_v1.png?generation=1712264660626280&alt=media" alt="">
To ensure compatibility with a wide range of object detection frameworks and tools, each version of the dataset is available in multiple formats:
These formats facilitate seamless integration into various machine learning frameworks and libraries, empowering users to leverage their preferred development environments.
In addition to image datasets, we also provide a video for real-time object detection evaluation. This video allows users to test the performance of their models in real-world scenarios, providing invaluable insights into the effectiveness of their detection algorithms.
To begin exploring the Vehicle Detection Image Dataset, simply download the version and format that best suits your project requirements. Whether you are an experienced practitioner or just embarking on your journey in computer vision, this dataset offers a valuable resource for advancing your understanding and capabilities in object detection and tracking tasks.
If you utilize this dataset in your work, we kindly request that you cite the following:
Parisa Karimi Darabi. (2024). Vehicle Detection Image Dataset: Suitable for Object Detection and tracking Tasks. Retrieved from https://www.kaggle.com/datasets/pkdarabi/vehicle-detection-image-dataset/
I welcome feedback and contributions from the Kaggle community to continually enhance the quality and usability of this dataset. Please feel free to reach out if you have suggestions, questions, or additional data and annotations to contribute. Together, we can drive innovation and progress in computer vision.
Facebook
TwitterThe Common Objects in Context (COCO) dataset is a widely recognized collection designed to spur object detection, segmentation, and captioning research. Created by Microsoft, COCO provides annotations, including object categories, keypoints, and more. The model it a valuable asset for machine learning practitioners and researchers. Today, many model architectures are benchmarked against COCO, which has enabled a standard system by which architectures can be compared.
While COCO is often touted to comprise over 300k images, it's pivotal to understand that this number includes diverse formats like keypoints, among others. Specifically, the labeled dataset for object detection stands at 123,272 images.
The full object detection labeled dataset is made available here, ensuring researchers have access to the most comprehensive data for their experiments. With that said, COCO has not released their test set annotations, meaning the test data doesn't come with labels. Thus, this data is not included in the dataset.
The Roboflow team has worked extensively with COCO. Here are a few links that may be helpful as you get started working with this dataset:
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Data Format is a dataset for object detection tasks - it contains Defects annotations for 1,502 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Product Types is a dataset for object detection tasks - it contains Square Circle annotations for 4,788 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
Twitter9,574 Images โ Handwriting OCR Data On Board of 8 Languages. The language distribution includes English, Spanish and Portuguese, etc. The data diversity includes multiple collecting scenes, multiple text carriers, multiple languages, multiple photographic angles. The collecting angeles are looking up angle, eye-level angle and looking down angle. In terms of annotation, row-level quadrilateral (polygon) annotation and content transcription are performed on the text. The dataset can be used for tasks such as handwriting OCR.
Data size
9,574 images, 243,240 bounding boxes
Language distribution
English, Spanish, Portuguese, French, German, Japanese, Italian and Dutch
Collecting environment
black boards, white boards, green boards
Device
cellphone
Photographic angle
eye-level angle, looking down angle, looking up angle
Data format
the image data format is .jpg and other common image formats, the annotation file data format is.json
Annotation content
line-level quadrilateral (polygon) bounding box annotation and transcription for the texts
Accuracy rate
the error bound of each vertex of quadrilateral bounding box is within 5 pixels, which is a qualified annotation, the accuracy of bounding boxes is not less than 95%; the texts transcription accuracy is not less than 95%
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Tfrecord Data Format is a dataset for object detection tasks - it contains Weapons And Animals annotations for 1,333 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterRace distribution : Asians, Caucasians, black people
Gender distribution : gender balance
Age distribution : ranging from teenager to the elderly, the middle-aged and young people are the majorities
Collecting environment : including indoor and outdoor scenes
Data diversity : different shooting heights, different ages, different light conditions, different collecting environment, clothes in different seasons, multiple human poses
Device : cameras
Data format : the data format is .jpg/mp4, the annotation file format is .json, the camera parameter file format is .json, the point cloud file format is .pcd
Accuracy : based on the accuracy of the poses, the accuracy exceeds 97%;the accuracy of labels of gender, race, age, collecting environment and clothes are more than 97%
Facebook
TwitterDataset of 3623 images from 1661 users (~2.18/user), mainly front/back ID documents, ideal for OCR training, document recognition, and automated identity verification tasks.
Facebook
TwitterMillion Level Document OCR and Structured Analysis Data, Including textbooks, e-books, teaching aids, patents, theses, etc. The annotated files include OCR annotations and structured analysis.
Data Size
10 Million Pages
Data Types
Textbooks, Chinese E-books, Teaching Reference Books, Paper, Chinese Journals, English Journals
Data Format
The original document file format is PDF, the document image file format is. png, the OCR annotation file format is JSON, and the structured parsing file format is markdown(Tables and formulas are in Latex format or screenshot links)
Data Accuracy
If the transcription of the text is basically accurate, the markings are basically aligned, and there are no obvious typos, it is considered correct annotation. Divided by punctuation marks, the number of correctly annotated sentences should not be less than 90%.
Facebook
TwitterBiometric Data
FileMarket provides a comprehensive Biometric Data set, ideal for enhancing AI applications in security, identity verification, and more. In addition to Biometric Data, we offer specialized datasets across Object Detection Data, Machine Learning (ML) Data, Large Language Model (LLM) Data, and Deep Learning (DL) Data. Each dataset is meticulously crafted to support the development of cutting-edge AI models.
Data Size: 20,000 IDs
Race Distribution: The dataset encompasses individuals from diverse racial backgrounds, including Black, Caucasian, Indian, and Asian groups.
Gender Distribution: The dataset equally represents all genders, ensuring a balanced and inclusive collection.
Age Distribution: The data spans a broad age range, including young, middle-aged, and senior individuals, providing comprehensive age coverage.
Collection Environment: Data has been gathered in both indoor and outdoor environments, ensuring variety and relevance for real-world applications.
Data Diversity: This dataset includes a rich variety of face poses, racial backgrounds, age groups, lighting conditions, and scenes, making it ideal for robust biometric model training.
Device: All data has been collected using mobile phones, reflecting common real-world usage scenarios.
Data Format: The data is provided in .jpg and .png formats, ensuring compatibility with various processing tools and systems.
Accuracy: The labels for face pose, race, gender, and age are highly accurate, exceeding 95%, making this dataset reliable for training high-performance biometric models.
Facebook
TwitterPopulation distribution : the race distribution is Asians, Caucasians and black people, the gender distribution is male and female, the age distribution is from children to the elderly
Collecting environment : including indoor and outdoor scenes (such as supermarket, mall and residential area, etc.)
Data diversity : different ages, different time periods, different cameras, different human body orientations and postures, different ages collecting environment
Device : surveillance cameras, the image resolution is not less than 1,9201,080
Data format : the image data format is .jpg, the annotation file format is .json
Annotation content : human body rectangular bounding boxes, 15 human body attributes
Quality Requirements : A rectangular bounding box of human body is qualified when the deviation is not more than 3 pixels, and the qualified rate of the bounding boxes shall not be lower than 97%;Annotation accuracy of attributes is over 97%
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset consists of five subsets with annotated images in COCO format, designed for object detection and tracking plant growth: 1. Cucumber_Train Dataset (for Faster R-CNN) - Includes training, validation, and test images of cucumbers from different angles. - Annotations: Bounding boxes in COCO format for object detection tasks.
Annotations: Bounding boxes in COCO format.
Pepper Dataset
Contains images of pepper plants for 24 hours at hourly intervals from a fixed angle.
Annotations: Bounding boxes in COCO format.
Cannabis Dataset
Contains images of cannabis plants for 24 hours at hourly intervals from a fixed angle.
Annotations: Bounding boxes in COCO format.
Cucumber Dataset
Contains images of cucumber plants for 24 hours at hourly intervals from a fixed angle.
Annotations: Bounding boxes in COCO format.
This dataset supports training and evaluation of object detection models across diverse crops.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Diverse Image Collection: Our dataset encompasses a wide range of general images covering various categories such as objects, scenes, people, and more. The images are carefully curated to offer a rich source of visual data.
Sindhi Language Titles: One of the distinctive features of our dataset is the inclusion of Sindhi language titles for each image.
Annotations in YOLO Format: To facilitate your object detection tasks, we have meticulously annotated the images in YOLO format, making it compatible with the YOLOv3 or YOLOv4 models. This ensures that you can jump right into training your model without the hassle of converting annotations.
Comprehensive Metadata: Each image in the dataset is accompanied by a YAML file providing additional metadata, including information about the image source, date of capture, and any relevant context that may be useful for your research.
By publishing this YOLO-style dataset with Sindhi language titles, we aim to contribute to the machine learning and computer vision community, fostering innovation and inclusivity in the field. We encourage you to explore, experiment, and create cutting-edge models using this dataset, and we look forward to seeing the incredible projects that emerge from it.