Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
CVAT Coco is a dataset for object detection tasks - it contains Defect Distance Event annotations for 9,899 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset contains 8,992 images of Uno cards and 26,976 labeled examples on various textured backgrounds.
This dataset was collected, processed, and released by Roboflow user Adam Crawshaw, released with a modified MIT license: https://firstdonoharm.dev/
https://i.imgur.com/P8jIKjb.jpg" alt="Image example">
Adam used this dataset to create an auto-scoring Uno application:
Fork or download this dataset and follow our How to train state of the art object detector YOLOv4 for more.
See here for how to use the CVAT annotation tool.
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless. :fa-spacer: Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Upload_cvat is a dataset for object detection tasks - it contains Weld annotations for 1,676 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Summerschool_cvat is a dataset for object detection tasks - it contains Flags annotations for 1,049 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Drone Image is a dataset for object detection tasks - it contains Car Motorcycle Van Bus Lorry annotations for 9,029 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Authors marked with an asterisk (*) have contributed equally to this publication.
We annotated a dataset for the detection of drainage outlets and ventilations on flat rooftops. The underlying high-resolution aerial images are orthophotos with a ground sampling distance of 7.5 cm, provided by the Office for Land Management and Geoinformation of the City of Bonn, Germany. The dataset was created through manual annotation using the Computer Vision Annotation Tool (CVAT) and comprises 740 image pairs. Each pair consists of a rooftop image and a corresponding annotated mask indicating the drainage outlets and ventilations. Since rooftops vary in size, we aimed to create image pairs that capture a single rooftop per image without overlaps or cutoffs. Consequently, the dimensions of each image pair differ. The dataset is split randomly into 80% for training, 10% for validation, and 10% for testing.
We provide the dataset in the Common Objects in Context (COCO) format for object detection tasks. In addition to the COCO-formatted dataset, we provide the dataset in its original, pairwise, format to support various machine learning tasks, such as semantic segmentation and panoptic segmentation, as well as to accommodate different data-loading requirements for diverse deep learning models.
If your object detection approach requires the 'category_id' to start from 0 instead of 1, please refer to the following guide: https://github.com/obss/sahi/discussions/336
For conversion to a completely different dataset format, such as YOLO, please see the repository: https://github.com/ultralytics/JSON2YOLO
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Data
The dataset consist of 5538 images of public spaces, annotated with steps, stairs, ramps and grab bars for stairs and ramps. The dataset has annotations 3564 of steps, 1492 of stairs, 143 of ramps and 922 of grab bars.
Each step annotation is attributed with an estimate of the height of the step, as falling into one of three categories: less than 3cm, 3cm to 7cm or more than 7cm. Additionally it is attributed with a 'type', with the possibilities 'doorstep', 'curb' or 'other'.
Stair annotations are attributed with the number of steps in the stair.
Ramps are attributed with an estimate of their width, also falling into three categories: less than 50cm, 50cm to 100cm and more than 100cm.
In order to preserve all additional attributes of the labels, the data is published in the CVAT XML format for images.
Annotating Process
The labelling has been done using bounding boxes around the objects. This format is compatible with many popular object detection models, e.g. the YOLO object model. A bounding box is placed so it contains exactly the visible part of the respective objects. This implies that only objects that are visible in the photo are annotated. This means in particular a photo of a stair or step from above, where the object cannot be seen, have not been annotated, even when a human viewer can possibly infer that there is a stair or a step from other features in the photo.
Steps
A step is annotated, when there is an vertical increment that functions as a passage between two surface areas intended human or vehicle traffic. This means that we have not included:
In particular, the bounding box of a step object contains exactly the incremental part of the step, but does not extend into the top or bottom horizontal surface any more than necessary to enclose entirely the incremental part. This has been chosen for consistency reasons, as including parts of the horizontal surfaces would imply a non-trivial choice of how much to include, which we deemed would most likely lead to more inconstistent annotations.
The height of the steps are estimated by the annotators, and are therefore not guarranteed to be accurate.
The type of the steps typically fall into the category 'doorstep' or 'curb'. Steps that are in a doorway, entrance or likewise are attributed as doorsteps. We also include in this category steps that are immediately leading to a doorway within a proximity of 1-2m. Steps between different types of pathways, e.g. between streets and sidewalks, are annotated as curbs. Any other type of step are annotated with 'other'. Many of the 'other' steps are for example steps to terraces.
Stairs
The stair label is used whenever two or more steps directly follow each other in a consistent pattern. All vertical increments are enclosed in the bounding box, as well as intermediate surfaces of the steps. However the top and bottom surface is not included more than necessary for the same reason as for steps, as described in the previous section.
The annotator counts the number of steps, and attribute this to the stair object label.
Ramps
Ramps have been annotated when a sloped passage way has been placed or built to connect two surface areas intended for human or vehicle traffic. This implies the same considerations as with steps. Alike also only the sloped part of a ramp is annotated, not including the bottom or top surface area.
For each ramp, the annotator makes an assessment of the width of the ramp in three categories: less than 50cm, 50cm to 100cm and more than 100cm. This parameter is visually hard to assess, and sometimes impossible due to the view of the ramp.
Grab Bars
Grab bars are annotated for hand rails and similar that are in direct connection to a stair or a ramp. While horizontal grab bars could also have been included, this was omitted due to the implied ambiguities of fences and similar objects. As the grab bar was originally intended as an attributal information to stairs and ramps, we chose to keep this focus. The bounding box encloses the part of the grab bar that functions as a hand rail for the stair or ramp.
Usage
As is often the case when annotating data, much information depends on the subjective assessment of the annotator. As each data point in this dataset has been annotated only by one person, caution should be taken if the data is applied.
Generally speaking, the mindset and usage guiding the annotations have been wheelchair accessibility. While we have strived to annotate at an object level, hopefully making the data more widely applicable than this, we state this explicitly as it may have swayed untrivial annotation choices.
The attributal data, such as step height or ramp width are highly subjective estimations. We still provide this data to give a post-hoc method to adjust which annotations to use. E.g. for some purposes, one may be interested in detecting only steps that are indeed more than 3cm. The attributal data makes it possible to sort away the steps less than 3cm, so a machine learning algorithm can be trained on this more appropriate dataset for that use case. We stress however, that one cannot expect to train accurate machine learning algorithms inferring the attributal data, as this is not accurate data in the first place.
We hope this dataset will be a useful building block in the endeavours for automating barrier detection and documentation.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset contains 74 images of aerial maritime photographs taken with via a Mavic Air 2 drone and 1,151 bounding boxes, consisting of docks, boats, lifts, jetskis, and cars. This is a multi class problem. This is an aerial object detection dataset. This is a maritime object detection dataset.
The drone was flown at 400 ft. No drones were harmed in the making of this dataset.
This dataset was collected and annotated by the Roboflow team, released with MIT license.
https://i.imgur.com/9ZYLQSO.jpg" alt="Image example">
This dataset is a great starter dataset for building an aerial object detection model with your drone.
Fork or download this dataset and follow our How to train state of the art object detector YOLOv4 for more. Stay tuned for particular tutorials on how to teach your UAV drone how to see and comprable airplane imagery and airplane footage.
See here for how to use the CVAT annotation tool that was used to create this dataset.
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless. :fa-spacer: Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
The BlueberryDCM dataset consists of 140 RGB images of blueberry canopies captured at varied spatial scales. All the images were acquired using smartphones in natural field light conditions in different orchards in the season of 2022, with 134 images in Mississippi and 6 images in Michigan. A total of 17,955 bounding box annotations were manually done in the VGG Image Annotator (VIA) (v2.0.12) for the blueberry instances of two fruit maturity classes, "Blue" and "Unblue", representing ripe and unripe fruit, respectively. In addition, for each maturity class, there are two sub-categories in the annotation, "visible", and "occluded", to indicate whether the fruit is fully visible in the canopy or partially occluded. The original annotation format exported from the VGG is VIA .json. The derived annotation files in two other formats, .xml (Pascal VOC format) and .txt (YOLO format with noralized xywh, with 0, 1, 2, and 3 denoting the four categories of "Unblue_visible", "Unblue_occluded", "Blue_visible", and "Blue_occluded" bluerries, respectively) are provided in the dataset for the compatibility of a wide range of object detectors. Hence, the dataset contains both the raw images (.jpg) and three corresponding annotations files (.json, .xml, and .txt) with the same file names, totaling about 107 MB in file size.
The dataset was used for in a study (see below) on the evaluation of YOLOv8 and YOLOv9 models for blueberry detection, counting, and maturity assessment. The detection accuracy of 93% mAP@50 was achieved by YOLOv8l, with an error of about 10 blueberries in fruit counting and an error of 3.6% in estimating the "Blue" fruit percentage. Software programs for the modeling work are made publicly available at: https://github.com/vicdxxx/BlueberryDetectionAndCounting. In addition, the blueberry dataset was also used as a preliminary database for developing an iOS-based mobile application, which is described in Deng, B., Lu, Y., WanderWeide, J., 2024. Development and preliminary evaluation of a deep learning-based fruit counting mobile application for highbush Blueberries. 2024 ASABE Annual International Meeting 2401022
Details about the dataset curation and statistics as well as modeling experiments are described in the journal article: Deng, B., Lu, Y., 2024. Detection, Counting, and Maturity Assessment of Blueberries in Canopy Images using YOLOv8 and YOLOv9. Smart Agricultural Technology. https://doi.org/10.1016/j.atech.2024.100620. If you use the dataset in published research, please consider citing the dataset or the journal article. Hopefully, you find the dataset useful.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
https://www.youtube.com/watch?v=4MA_6oZQz7s&ab_channel=tektronix475
Spotted caps, are the normal OK class (fully closed). Clean caps, are the bad or anomally target class (partially closed). One double prediction at 3:59. 100x100 classification accuracy, out of 200 samples. Inference over unseen test dataset. 150 epochs training. 700 samples training dataset, no data augmentation.
PREPROCESSING Auto-Orient: Applied Resize: Stretch to 416x416 Grayscale: Applied AUGMENTATIONS No augmentations were applied.
Anomaly detection with: Roboflow, tensorflow, google colab, Ultralytics, yolo v5, cvat,
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
CVAT Coco is a dataset for object detection tasks - it contains Defect Distance Event annotations for 9,899 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).