87 datasets found
  1. R

    Oriented Bounding Boxes Dataset

    • universe.roboflow.com
    zip
    Updated Apr 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Robot (2024). Oriented Bounding Boxes Dataset [Dataset]. https://universe.roboflow.com/robot-crknl/oriented-bounding-boxes-dataset/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 16, 2024
    Dataset authored and provided by
    Robot
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Robot O0Gq Bounding Boxes
    Description

    Oriented Bounding Boxes Dataset

    ## Overview
    
    Oriented Bounding Boxes Dataset is a dataset for object detection tasks - it contains Robot O0Gq annotations for 563 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  2. d

    Deep Learning Swimming Pool Oriented Bounding Boxes 2025 - Datasets -...

    • catalogue.data.wa.gov.au
    Updated Sep 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Deep Learning Swimming Pool Oriented Bounding Boxes 2025 - Datasets - data.wa.gov.au [Dataset]. https://catalogue.data.wa.gov.au/dataset/deep-learning-swimming-pool-oriented-bounding-boxes-2025
    Explore at:
    Dataset updated
    Sep 16, 2025
    Area covered
    Western Australia
    Description

    Vector dataset extracted using a deep learning oriented object detection model. Model is trained to identify and classify above and below swimming pools. Show full description

  3. R

    Boots Oriented Bounding Box Dataset

    • universe.roboflow.com
    zip
    Updated Aug 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    roboteam (2024). Boots Oriented Bounding Box Dataset [Dataset]. https://universe.roboflow.com/roboteam/boots-oriented-bounding-box/model/5
    Explore at:
    zipAvailable download formats
    Dataset updated
    Aug 9, 2024
    Dataset authored and provided by
    roboteam
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Box Bounding Boxes
    Description

    Boots Oriented Bounding Box

    ## Overview
    
    Boots Oriented Bounding Box is a dataset for object detection tasks - it contains Box annotations for 509 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  4. FAIR1M Satellite Imagery for Object Detection

    • kaggle.com
    zip
    Updated Dec 10, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Olly Powell (2023). FAIR1M Satellite Imagery for Object Detection [Dataset]. https://www.kaggle.com/datasets/ollypowell/fair1m-satellite-imagery-for-object-detection/data
    Explore at:
    zip(9190024314 bytes)Available download formats
    Dataset updated
    Dec 10, 2023
    Authors
    Olly Powell
    License

    Attribution-NonCommercial-ShareAlike 3.0 (CC BY-NC-SA 3.0)https://creativecommons.org/licenses/by-nc-sa/3.0/
    License information was derived automatically

    Description

    EDA notebook here.

    The paper for this dataset is found here, the dataset was used in the Gaofen Challenge hosted by the Aerospace Information Research Institute, Chinese Academy of Sciences.

    I have put this together because a few months ago I had a project that needed such a dataset for vehicle detection, and found there wasn't much out there with suitable resolution and quality. I ended up using the xView1 Dataset, which was pretty good, but noted at the time the FAIR1M had a lot of potential too.

    It's main points of difference of FAIR1M compared to many others in this space are: - Some geographical diversity: Asia, Europe, North America, Capetown, Sydney. Mostly Urban - Oriented bounding boxes - Most of the imagery is high resolution: 0.3m or 0.6m, which makes it just enough for small car detection.

    For comparison, xView-1 is larger and more geographically diverse, but has flat bounding boxes. If you want to try oriented bounding boxes, FAIR1M is worth a try.

    I could only find 240,852 spatially unique labels, the rest seem to be duplicates due to overlapping imagery. Though some of course would be in the hidden test set, which has not been made public. Anyway, that's still a lot of labels, so thanks to the organisers for making these available.

  5. f

    Data from: Multi-scale spatial fusion lightweight model for optical remote...

    • tandf.figshare.com
    png
    Updated Oct 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Qiyi He; Ao Xu; Zhiwei Ye; Shirui Sheng; Wen Zhou; Xudong Lai (2025). Multi-scale spatial fusion lightweight model for optical remote sensing image-based small object detection [Dataset]. http://doi.org/10.6084/m9.figshare.30328707.v1
    Explore at:
    pngAvailable download formats
    Dataset updated
    Oct 10, 2025
    Dataset provided by
    Taylor & Francis
    Authors
    Qiyi He; Ao Xu; Zhiwei Ye; Shirui Sheng; Wen Zhou; Xudong Lai
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Current remote sensing object detection frameworks often focus solely on the geometric relationship between true and predicted boxes, neglecting the intrinsic shapes of the boxes. In the field of remote sensing detection, there are numerous elongated bounding boxes. Variations in the shape and size of these boxes result in differences in their Intersection over Union (IoU) values, which is particularly noticeable when detecting small objects. Platforms with limited resources, such as satellites and unmanned drones, have strict requirements for detector storage space and computational complexity. This makes it challenging for existing methods to balance detection performance and computational demands. Therefore, this paper presents RS-YOLO, a lightweight framework that enhances You Only Look Once (YOLO) and is specifically designed for deployment on resource-limited platforms. RS-YOLO has developed a bounding box regression approach for remote sensing images, focusing on the shape and scale of the boundary boxes. Additionally, to improve the integration of multi-scale spatial features, RS-YOLO introduces a lightweight multi-scale hybrid attention module for cross-space fusion. The DOTA-v1.0 and HRSC2016 datasets were used to test our model, which was then compared to multiple state-of-the-art oriented object detection models. The results indicate that the detector introduced in this article achieves top performance while being lightweight and suitable for deployment on resource-limited platforms.

  6. Eagle Dataset (YOLOv11-OBB Format)

    • kaggle.com
    zip
    Updated Jul 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mridankan Mandal (2025). Eagle Dataset (YOLOv11-OBB Format) [Dataset]. https://www.kaggle.com/datasets/redzapdos123/eagle-dataset-yolov11-obb-format/code
    Explore at:
    zip(3192683527 bytes)Available download formats
    Dataset updated
    Jul 29, 2025
    Authors
    Mridankan Mandal
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Eagle Dataset - YOLOv11 OBB for Vehicle Detection

    High-resolution aerial imagery with 16,000+ oriented bounding boxes for vehicle detection, pre-formatted for Ultralytics YOLOv11.

    Context

    This dataset is a ready-to-use version of the original Eagle Dataset from the German Aerospace Center (DLR). The original dataset was created to benchmark object detection models on challenging aerial imagery, featuring vehicles at various orientations.

    This version has been converted to the YOLOv11-OBB (Oriented Bounding Box) format. The conversion makes the dataset directly compatible with modern deep learning frameworks like Ultralytics YOLO, allowing researchers and developers to train state-of-the-art object detectors with minimal setup.

    The dataset is ideal for tasks requiring precise localization of rotated objects, such as vehicle detection in parking lots, traffic monitoring, and urban planning from aerial viewpoints.

    Content

    The dataset is split into training, validation, and test sets, following a standard structure for computer vision tasks.

    Dataset Split & Counts:

    • Training Set: 159 images and labels
    • Validation Set: 53 images and labels
    • Test Set: 106 images and labels

    Directory Structure:

    EagleDatasetYOLO/
    ├── train/
    │  ├── images/   # 159 images
    │  └── labels/   # 159 .txt obb labels
    ├── val/
    │  ├── images/   # 53 images
    │  └── labels/   # 53 .txt obb labels
    ├── test/
    │  ├── images/   # 106 images
    │  └── labels/   # 106 .txt obb labels
    ├── data.yaml
    └── license.md
    

    Annotation Format (YOLOv11-OBB):

    Each .txt label file contains one object per line. The format for each object is: <class_id> <x_center> <y_center> <width> <height> <angle>

    • <class_id>: The class index (in this case, 0 for 'vehicle').
    • <x_center> <y_center>: The normalized center coordinates of the bounding box.
    • <width> <height>: The normalized width and height of the bounding box.
    • <angle>: The rotation angle of the box in radians, from -π/2 to π/2.

    data.yaml Configuration:

    A data.yaml file is included for easy integration with the Ultralytics framework.

    path: ../EagleDatasetYOLO
    train: train/images
    val: val/images
    test: test/images
    
    nc: 1
    names: ['vehicle']
    

    Acknowledgements and License

    This dataset is a conversion of the original work by the German Aerospace Center (DLR). The conversion to YOLOv11-OBB format was performed by Mridankan Mandal.

    The dataset is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International license (CC BY-NC-SA 4.0).

    If you use this dataset in your research, please cite the original creators and acknowledge the conversion work.

  7. Z

    Data from: DeepScoresV2

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jun 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tuggener, Lukas; Satyawan, Yvan Putra; Pacha, Alexander; Schmidhuber, Jürgen; Stadelmann, Thilo (2023). DeepScoresV2 [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_4012192
    Explore at:
    Dataset updated
    Jun 7, 2023
    Dataset provided by
    TU Wien
    ZHAW Datalab & USi
    ZHAW Datalab
    The Swiss AI Lab IDSIA (USI & SUPSI)
    Authors
    Tuggener, Lukas; Satyawan, Yvan Putra; Pacha, Alexander; Schmidhuber, Jürgen; Stadelmann, Thilo
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The DeepScoresV2 Dataset for Music Object Detection contains digitally rendered images of written sheet music, together with the corresponding ground truth to fit various types of machine learning models. A total of 151 Million different instances of music symbols, belonging to 135 different classes are annotated. The total Dataset contains 255,385 Images. For most researches, the dense version, containing 1714 of the most diverse and interesting images, should suffice.

    The dataset contains ground in the form of:

    Non-oriented bounding boxes

    Oriented bounding boxes

    Semantic segmentation

    Instance segmentation

    The accompaning paper The DeepScoresV2 Dataset and Benchmark for Music Object Detection published at ICPR2020 can be found here:

    https://digitalcollection.zhaw.ch/handle/11475/20647

    A toolkit for convenient loading and inspection of the data can be found here:

    https://github.com/yvan674/obb_anns

    Code to train baseline models can be found here:

    https://github.com/tuggeluk/mmdetection/tree/DSV2_Baseline_FasterRCNN

    https://github.com/tuggeluk/DeepWatershedDetection/tree/dwd_old

  8. MVTec Screws

    • kaggle.com
    zip
    Updated Feb 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Innat (2022). MVTec Screws [Dataset]. https://www.kaggle.com/datasets/ipythonx/mvtec-screws
    Explore at:
    zip(1363527362 bytes)Available download formats
    Dataset updated
    Feb 21, 2022
    Authors
    Innat
    Description

    MVTEC SCREWS DATASET

    Ref. https://www.mvtec.com/company/research/datasets/mvtec-screws

    The MVTec Screws dataset has been designed for oriented box detection. It contains 384 images of 13 different types of screws and nuts on a wooden background. All objects are labeled by oriented bounding boxes and their respective category. Overall, there are 4,426 of such annotations.

    PLEASE NOTE: LICENSE TERMS

    The data is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0).

    In particular, it is not allowed to use the dataset for commercial purposes. If you are unsure whether or not your application violates the non-commercial use clause of the license, please contact us via the form below.

  9. VSAI Dataset (YOLO11-OBB format)

    • kaggle.com
    zip
    Updated Aug 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mridankan Mandal (2025). VSAI Dataset (YOLO11-OBB format) [Dataset]. https://www.kaggle.com/datasets/redzapdos123/vsai-dataset-yolo11-obb-format/code
    Explore at:
    zip(8332516716 bytes)Available download formats
    Dataset updated
    Aug 29, 2025
    Authors
    Mridankan Mandal
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    VSAI Aerial Vehicle Detection Dataset (YOLO OBB Format):

    A cleaned, and reformatted version of the VSAI Dataset, specifically adapted for Oriented Bounding Box (OBB) vehicle detection using the YOLOv11 format.

    Overview:

    This dataset is designed for aerial/drone-based vehicle detection tasks. It is a modified version of the original VSAI Dataset v1 by the DroneVision Team. This version has been modified by Mridankan Mandal for the easy of training object detection models like the YOLO11-OBB models.

    The dataset is split into two classes: small-vehicle and large-vehicle. All annotations have been converted to the YOLOv11-OBB format, and the data is organized into training, validation, and testing sets.

    Key Features and Modifications:

    This dataset improves upon the original by incorporating several key modifications to make it more accessible and useful for modern computer vision tasks:

    • Format Conversion: The annotations have been converted to the YOLOv11-OBB format, which uses four corner points to define an oriented bounding box.
    • Data Cleaning: All image and annotation pairs where the label file was empty have been removed to ensure dataset quality.
    • Structured Splits: The dataset is pre-split into train (80%), validation (10%), and test (10%) sets, with the following image counts:
      • Train: 4,297 images
      • Validation: 537 images
      • Test: 538 images
      • Total: 5,372 images
    • Coordinate Normalization: All bounding box coordinates are normalized to a range of [0.0 - 1.0], making them ready for training without preprocessing.

    Directory Structure

    The dataset is organized in a standard YOLO format for easy integration with popular training frameworks.

    YOLOOBBVSAIDataset/
    ├── train/
    │  ├── images/   #Contains 4,297 image files.
    │  └── labels/   #Contains 4,297 .txt label files.
    ├── val/
    │  ├── images/   #Contains 537 image files.
    │  └── labels/   #Contains 537 .txt label files.
    ├── test/
    │  ├── images/   #Contains 538 image files.
    │  └── labels/   #Contains 538 .txt label files.
    ├── data.yaml    #Dataset configuration file.
    ├── license.md   #Full license details.
    └── ReadMe.md    #Dataset README file.
    

    Annotation Format:

    Each .txt label file contains one or more lines, with each line representing a single object in the YOLOv11-OBB format:

    class_id x1 y1 x2 y2 x3 y3 x4 y4

    • class_id: An integer representing the object class (0 for small-vehicle, 1 for large-vehicle).
    • (x1, y1)...(x4, y4): The four corner points of the oriented bounding box, with coordinates normalized between 0 and 1.

    data.yaml:

    To begin training a YOLO model with this dataset, you can use the provided data.yaml file. Simply update the path to the location of the dataset on your local machine.

    #The path to the root dataset directory.
    path: /path/to/YOLOOBBVSAIDataset/
    train: train/images
    val: val/images
    test: test/images
    
    #Number of classes.
    nc: 2
    
    #The Class names,
    names:
     0: small-vehicle
     1: large-vehicle
    

    License and Attribution:

    This dataset is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) license.

    • You are free to: Use, modify, and redistribute this dataset for non-commercial research and educational purposes.
    • You must: Provide proper attribution to both the original creators and the modifier, and release any derivative works under the same license.

    Proper Attribution:

    When using this dataset, attribute as follows:

    • Original VSAI Dataset v1 by DroneVision Team, licensed under CC BY-NC-SA 4.0.
    • Modified VSAI Dataset (YOLOv11-OBB Format) by Mridankan Mandal, licensed under CC BY-NC-SA 4.0.

    Citation:

    If you use this dataset in your research, use the following BibTeX entry to cite it:

    @dataset{vsai_yolo_obb_2025,
     title={VSAI Dataset (YOLOv11-OBB Format)},
     author={Mridankan Mandal},
     year={2025},
     note={Modified from original VSAI v1 dataset by DroneVision},
     license={CC BY-NC-SA 4.0}
    }
    
  10. MaxiDent-OBBox Maxillary Oriented Bounded Box

    • kaggle.com
    zip
    Updated Jun 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    maxxxw (2025). MaxiDent-OBBox Maxillary Oriented Bounded Box [Dataset]. https://www.kaggle.com/datasets/trickykestral/maxident-bbox-maxillary-tooth-bounded-box-dataset
    Explore at:
    zip(482283573 bytes)Available download formats
    Dataset updated
    Jun 9, 2025
    Authors
    maxxxw
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    This is a ready-to-use dataset consisting of X-ray images of the human jaw, with corresponding annotations for individual teeth. Each tooth is labeled using oriented bounding box (OBB) coordinates, making the dataset well-suited for tasks that require precise object localization and orientation awareness. There are a total of 17 classes representing teeth in upper jaw

    The annotations are formatted specifically for compatibility with YOLO-OBB (Oriented Bounding Box) models, enabling seamless integration into training pipelines for dental detection and analysis tasks.

  11. D

    DOTA Dataset

    • datasetninja.com
    • kaggle.com
    Updated Feb 25, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jian Ding; Nan Xue; Gui-Song Xia (2021). DOTA Dataset [Dataset]. https://datasetninja.com/dota
    Explore at:
    Dataset updated
    Feb 25, 2021
    Dataset provided by
    Dataset Ninja
    Authors
    Jian Ding; Nan Xue; Gui-Song Xia
    License

    https://captain-whu.github.io/DOTA/dataset.htmlhttps://captain-whu.github.io/DOTA/dataset.html

    Description

    In the past decade, significant progress in object detection has been made in natural images, but authors of the DOTA v2.0: Dataset of Object deTection in Aerial images note that this progress hasn't extended to aerial images. The main reason for this discrepancy is the substantial variations in object scale and orientation caused by the bird's-eye view of aerial images. One major obstacle to the development of object detection in aerial images (ODAI) is the lack of large-scale benchmark datasets. The DOTA dataset contains 1,793,658 object instances spanning 18 different categories, all annotated with oriented bounding box annotations (OBB). These annotations were collected from a total of 11,268 aerial images. Using this extensive and meticulously annotated dataset, the authors establish baselines covering ten state-of-the-art algorithms, each with over 70 different configurations. These configurations are evaluated for both speed and accuracy performance.

  12. Data from: Detection of Structural Components in Point Clouds of Existing RC...

    • zenodo.org
    bin
    Updated Jan 24, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ruodan LU; Ioannis Brilakis; Campbell R. Middleton; Ruodan LU; Ioannis Brilakis; Campbell R. Middleton (2020). Detection of Structural Components in Point Clouds of Existing RC Bridges [Dataset]. http://doi.org/10.5281/zenodo.1233845
    Explore at:
    binAvailable download formats
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Ruodan LU; Ioannis Brilakis; Campbell R. Middleton; Ruodan LU; Ioannis Brilakis; Campbell R. Middleton
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The cost and effort of modelling existing bridges from point clouds currently outweighs the perceived benefits of the resulting model. There is a pressing need to automate this process. Previous research has achieved the automatic generation of surface primitives combined with rule-based classification to create labelled cuboids and cylinders from point clouds. While these methods work well in synthetic datasets or idealized cases, they encounter huge challenges when dealing with real-world bridge point clouds, which are often unevenly distributed and suffer from occlusions. In addition, real bridge geometries are complicated. In this paper, we propose a novel top-down method to tackle these challenges for detecting slab, pier, pier cap, and girder components in reinforced concrete bridges. This method uses a slicing algorithm to separate the deck assembly from pier assemblies. It then detects and segment pier caps using their surface normal, and girders using oriented bounding boxes and density histograms. Finally, our method merges over-segments into individually labelled point clusters. The results of 10 real-world bridge point cloud experiments indicate that our method achieves an average detection precision of 98.8%. This is the first method of its kind to achieve robust detection performance for the four component types in reinforced concrete bridges and to directly produce labelled point clusters. Our work provides a solid foundation for future work in generating rich Industry Foundation Classes models from the labelled point clusters.

  13. R

    Building Facades Dataset

    • universe.roboflow.com
    zip
    Updated Oct 17, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    gsvvilavelhaplus (2024). Building Facades Dataset [Dataset]. https://universe.roboflow.com/gsvvilavelhaplus/building-facades/model/5
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 17, 2024
    Dataset authored and provided by
    gsvvilavelhaplus
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Building Facades Bounding Boxes
    Description

    This dataset is a project created to aid in land use classification of properties based on their facades on the streets. It is a bounding box object detection oriented dataset, but the objective is to try semi-supervised techniques to utilize the fewer annotated image examples as possible.

  14. Screws and Nuts Image

    • kaggle.com
    zip
    Updated Oct 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sujan Shirol (2023). Screws and Nuts Image [Dataset]. https://www.kaggle.com/datasets/sujan97/screws-and-nuts-image/data
    Explore at:
    zip(1363527362 bytes)Available download formats
    Dataset updated
    Oct 30, 2023
    Authors
    Sujan Shirol
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Original Author: MVTec Software GmbH, July 2020.

    Dataset contains 384 images of 13 different types of screws and nuts on a wooden background. The objects are labeled by oriented bounding boxes and their respective category. Overall, there are 4426 of such annotations. Initially, they have been selected randomly, such that approximately 70% of the instances of each category are within the training split, and 15% each in the validation and test splits.

    • folder images contains the screw images.
    • mvtec_screws.json contains the annotations for all images in COCO format.
    • mvtec_screws_train/val/test.json contain examplary splits as mentioned above, in COCO format.
    • mvtec_screws.hdict contains the DLDataset unsplitted.
    • mvtec_screws_split.hdict contains the DLDataset with splits.
  15. Data from: SemanticSugarBeets: A Multi-Task Framework and Dataset for...

    • zenodo.org
    zip
    Updated May 13, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gerardus Croonen; Gerardus Croonen; Andreas Trondl; Andreas Trondl; Julia Simon; Julia Simon; Daniel Steininger; Daniel Steininger (2025). SemanticSugarBeets: A Multi-Task Framework and Dataset for Inspecting Harvest and Storage Characteristics of Sugar Beets [Dataset]. http://doi.org/10.5281/zenodo.15393471
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 13, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Gerardus Croonen; Gerardus Croonen; Andreas Trondl; Andreas Trondl; Julia Simon; Julia Simon; Daniel Steininger; Daniel Steininger
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    SemanticSugarBeets is a comprehensive dataset and framework designed for analyzing post-harvest and post-storage sugar beets using monocular RGB images. It supports three key tasks: instance segmentation to identify and delineate individual sugar beets, semantic segmentation to classify specific regions of each beet (e.g., damage, soil adhesion, vegetation, and rot) and oriented object detection to estimate the size and mass of beets using reference objects. The dataset includes 952 annotated images with 2,920 sugar-beet instances, captured both before and after storage. Accompanying the dataset is a demo application and processing code, available on GitHub. For more details, refer to the paper presented at the Agriculture-Vision Workshop at CVPR 2025.

    Annotations and Learning Tasks

    The dataset supports three primary learning tasks, each designed to address specific aspects of sugar-beet analysis:

    1. Instance Segmentation
      Detect and delineate entire sugar-beet instances in an image. This task provides coarse-grained annotations for identifying individual beets, which is useful for counting and localization.

    2. Semantic Segmentation
      Perform fine-grained segmentation of each beet instance to classify its regions into specific categories relevant to quality assessment, such as:
      • Beet: healthy, undamaged beet surfaces
      • Cut: areas where the beet has been topped or trimmed
      • Leaf: residual vegetation attached to the beet
      • Soil: soil adhering to the beet's surface
      • Damage: visible damage on the beet
      • Rot: areas affected by rot

    3. Oriented Object Detection
      Detect and estimate the position and orientation of reference objects (folding-ruler elements and plastic signs) within the image. These objects can be used for scale estimation to calculate the absolute size and mass of sugar beets.

    Data Structure and Formats

    The dataset is organized into the following directories:

    • images: contains all RGB images in .jpg format with a resolution of 2120x1192 pixels, which correspond to the annotations in the instances and markers directories

    • instances: annotations and split files used in instance-segmentation experiments:
      • anno: instance contours for a single sugar-beet class in YOLO11 format
      • train/val/test.txt: lists of image IDs for training, validation and testing

    • markers: annotations and split files used in oriented-object-detection experiments:
      • anno: oriented-bounding-box annotations for two classes of markers in YOLO11 format:
        • 0: Ruler (folding-ruler element)
        • 1: Sign (numbered plastic sign)
      • train/val/test.txt: lists of image IDs for training, validation and testing

    • segmentation: annotations, image patches and split files used in semantic-segmentation experiments:
      • anno: single-channel segmentation masks for each individual beet, where pixel values correspond to the following classes:
        • 0: Background
        • 1: Beet
        • 2: Cut
        • 3: Leaf
        • 4: Soil
        • 5: Damage
        • 6: Rot
      • patches: image patches of individual sugar-beet instances cropped from the original images for convenience
      • train/val/test.txt: lists of beet IDs for training, validation, and testing

    File Naming Convention

    File names of images and annotations follow this format:

    ssb-

    • : a 5-digit number (e.g., 00001) identifying the group of recorded sugar beets
    • : either a or b, indicating the same group of beets captured before (a) or after flipping (b)
    • : a 3-digit number (e.g., 001) enumerating individual sugar beets within an image (used only for semantic segmentation)

    Example

    • ssb-00001a: group ID 00001, side a
    • ssb-00001a-001: group ID 00001, side a, beet instance 001

    Citing

    If you use the SemanticSugarBeets dataset or source code in your research, please cite the following paper to acknowledge the authors' contributions:

    Croonen, G., Trondl, A., Simon, J., Steininger, D., 2025. SemanticSugarBeets: A Multi-Task Framework and Dataset for Inspecting Harvest and Storage Characteristics of Sugar Beets. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.

  16. Chemistry Lab Image Dataset Covering 25 Apparatus Categories

    • figshare.com
    application/x-rar
    Updated Aug 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Md. Sakhawat Hossain; Md. Sadman Haque; Md. Mostafizur Rahman; Md. Mosaddik Mashrafi Mousum; Zobaer Ibn Razzaque; Robiul Awoul Robin (2025). Chemistry Lab Image Dataset Covering 25 Apparatus Categories [Dataset]. http://doi.org/10.6084/m9.figshare.29110433.v3
    Explore at:
    application/x-rarAvailable download formats
    Dataset updated
    Aug 3, 2025
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Md. Sakhawat Hossain; Md. Sadman Haque; Md. Mostafizur Rahman; Md. Mosaddik Mashrafi Mousum; Zobaer Ibn Razzaque; Robiul Awoul Robin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains 4,599 high-quality, annotated images of 25 commonly used chemistry lab apparatuses. The images, each containing structures in real-world settings, have been captured from different angles, backgrounds, and distances, while also undergoing variations in lighting to aid in the robustness of object detection models. Every image has been labeled using bounding box annotation in TXT (YOLO) format, alongside the class IDs and normalized bounding box coordinates, making object detection more precise. The annotations and bounding boxes have been built using the Roboflow platform.To achieve a better learning procedure, the dataset has been split into three sub-datasets: training, validation, and testing. The training dataset constitutes 70% of the entire dataset, with validation and testing at 20% and 10% respectively. In addition, all images undergo scaling to a standard of 640x640 pixels while being auto-oriented to rectify rotation discrepancies brought about by the EXIF metadata. The dataset is structured in three main folders - train, valid, and test, and each contains images/ and labels/ subfolders. Every image contains a label file containing class and bounding box data corresponding to each detected object.The whole dataset features 6,960 labeled instances per 25 apparatus categories including beakers, conical flasks, measuring cylinders, test tubes, among others. The dataset can be utilized for the development of automation systems, real-time monitoring and tracking systems, tools for safety monitoring, alongside AI educational tools.

  17. h

    aria-everyday-objects

    • huggingface.co
    Updated Sep 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Project Aria from Meta Reality Labs Research (2024). aria-everyday-objects [Dataset]. https://huggingface.co/datasets/projectaria/aria-everyday-objects
    Explore at:
    Dataset updated
    Sep 30, 2024
    Dataset authored and provided by
    Project Aria from Meta Reality Labs Research
    License

    https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/

    Description

    Aria Everyday Objects Dataset

    [Project Page] [Data Explorer] [Code] [Paper]

    Aria Everyday Objects (AEO) is a small, challenging 3D object detection dataset for egocentric data. AEO consists of approximately 45 minutes of egocentric data across 25 sequences captured by non-computer vision experts collected in a diverse set of locations throughout the US. Oriented 3D bounding boxes have been annotated for each sequence. Annotation is done in 3D, using the camera calibration, SLAM… See the full description on the dataset page: https://huggingface.co/datasets/projectaria/aria-everyday-objects.

  18. Aerial Vehicle OBB Dataset

    • kaggle.com
    zip
    Updated Aug 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mridankan Mandal (2025). Aerial Vehicle OBB Dataset [Dataset]. https://www.kaggle.com/datasets/redzapdos123/aerial-vehicle-obb-dataset
    Explore at:
    zip(11517085012 bytes)Available download formats
    Dataset updated
    Aug 29, 2025
    Authors
    Mridankan Mandal
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Aerial Vehicles OBB Dataset (YOLOv11-OBB Format):

    A large scale, merged dataset for oriented vehicle detection in aerial imagery, preformatted for YOLOv11-OBB models.

    Overview:

    This dataset combines three distinct aerial imagery collections—**VSAI**, DroneVehicles, and DIOR-R, into a unified resource for training and benchmarking oriented object detection models. It has been specifically preprocessed and formatted for use with Ultralytics' YOLOv11-OBB models.

    The primary goal is to provide a detailed dataset for tasks like aerial surveillance, traffic monitoring, and vehicle detection from a drone's perspective. All annotations have been converted to the YOLO OBB format, and the classes have been simplified for focused vehicle detection tasks.

    Key Features:

    • Merged & Simplified: Combines three popular aerial vehicle datasets.
    • Two Class System: Simplifies detection by categorizing all objects into small-vehicle and large-vehicle.
    • YOLOv11-OBB Ready: Preformatted with normalized OBB annotations and a data.yaml configuration file for immediate use in YOLO training pipelines.
    • Cleaned & Split: Empty annotations have been removed, and the data is organized into standard train, validation, and test sets.

    Data Description:

    Source Datasets:

    1. VSAI Dataset: Contains aerial imagery for traffic analysis by DroneVision.
    2. DroneVehicles Dataset: A collection of vehicle images from a drone's perspective, originally provided in YOLO OBB format.
    3. DIOR-R Dataset: A large scale benchmark for object detection in optical remote sensing images. Only the 'vehicle' class was extracted for this merged dataset.

    Preprocessing and Modifications:

    • Class Merging: All vehicle types from the source datasets were mapped to two parent classes: small-vehicle and large-vehicle. The vehicle class from the DIOR-R dataset was mapped to large-vehicle.
    • Data Cleaning: Image and label pairs with empty annotation files were removed to ensure dataset integrity.
    • Formatting: All annotations were converted to the YOLOv11-OBB format, with coordinates normalized between 0 and 1.

    Classes:

    Class IDClass NameSource Dataset(s)
    0small-vehicleVSAI, DroneVehicles
    1large-vehicleVSAI, DroneVehicles, DIOR-R

    Dataset Statistics:

    • Total Labeled Images: 29,125
      • Training Set: 18,274 images
      • Validation Set: 5,420 images
      • Test Set: 5,431 images

    Annotation Format:

    Each image has a corresponding .txt label file. Each line in the file represents one object in the YOLOv11-OBB format: class_id x1 y1 x2 y2 x3 y3 x4 y4

    • class_id: The class index (0 for small-vehicle, 1 for large-vehicle).
    • (x1, y1)...(x4, y4): The four corner points of the oriented bounding box, with all coordinates normalized to a range of [0, 1].

    File and Folder Structure:

    The dataset is organized into a standard YOLO directory structure for easy integration with training programs.

    RoadVehiclesYOLOOBBDataset/
    ├── train/
    │  ├── images/ #18,274 images
    │  └── labels/ #18,274 labels
    ├── val/
    │  ├── images/ #5,420 images
    │  └── labels/ #5,420 labels
    ├── test/
    │  ├── images/ #5,431 images
    │  └── labels/ #5,431 labels
    ├── data.yaml  #YOLO dataset configuration file.
    └── ReadMe.md  #Documentation
    

    Usage:

    To use this dataset with YOLOv11 or other compatible frameworks, simply point your training script to the included data.yaml file.

    Example data.yaml:

    #Dataset configuration.
    path: RoadVehiclesYOLOOBBDataset/
    train: train/images
    val: val/images
    test: test/images
    
    #Number of classes.
    nc: 2
    
    #Class names.
    names:
     0: small-vehicle
     1: large-vehicle
    

    License:

    This merged dataset is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0), which is the most restrictive license among its sources.

    • You are free to:
      • Share and adapt the material for any non-commercial purpose.
    • Under the following terms:
      • Attribution: You must give appropriate credit to the original authors and the creator of this merged dataset.
      • NonCommercial: You may not use the material for commercial purposes.
      • ShareAlike: If you remix, transform, or build upon the material, you must distribute your contributions under the same license.

    Citation and Attribution:

    When using this dataset, please provide attribution to all original sources as follows:

    - VSAI_Dataset: by DroneVision, licensed under CC BY-NC-SA 4.0.
    - DroneVehicles Dataset: by Yiming Sun, Bing Cao, Pengfei Zhu, and Qin G. Hu and modified by Mridankan Mandal, licensed under CC BY-NC-SA 4.0.
    - DIOR-R dataset: by the DIOR...
    
  19. Urban Surveillance Dataset – 8,085 Groups, 10 Categories with Bounding Box...

    • nexdata.ai
    Updated Dec 5, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2023). Urban Surveillance Dataset – 8,085 Groups, 10 Categories with Bounding Box Annotations [Dataset]. https://www.nexdata.ai/datasets/computervision/1092
    Explore at:
    Dataset updated
    Dec 5, 2023
    Dataset authored and provided by
    Nexdata
    Variables measured
    Device, Data size, Data format, Accuracy rata, Data diversity, Collecting time, Collecting angle, Annotation content, Collecting environment
    Description

    This dataset contains 8,085 groups of images across 10 categories The collection scenes include street, snack street, shop entrance, corridor, community entrance, construction site, etc. The data diversity includes multiple scenes, different time periods(day, night), different photographic angles. Each image is annotated with rectangular bounding boxes for urban objects. This data can be used for tasks such as urban object detection, smart city management, public safety monitoring, and AI-driven city infrastructure analysis.

  20. h

    cctv-knife-detection-dataset

    • huggingface.co
    Updated Nov 18, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Fredrik H (2025). cctv-knife-detection-dataset [Dataset]. https://huggingface.co/datasets/Simuletic/cctv-knife-detection-dataset
    Explore at:
    Dataset updated
    Nov 18, 2025
    Authors
    Fredrik H
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is an open-source synthetic dataset for computer vision object detection, focused on people holding knives in public and semi-public environments, viewed from CCTV and surveillance camera perspectives. It is designed to help train and evaluate YOLO, YOLOv8, YOLOWorld, Detectron, and other object detection models for threat recognition, security analytics, and abnormal behavior detection. Key Features Classes: person, knife Annotations: YOLO format (bounding boxes, normalized) Image Type:… See the full description on the dataset page: https://huggingface.co/datasets/Simuletic/cctv-knife-detection-dataset.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Robot (2024). Oriented Bounding Boxes Dataset [Dataset]. https://universe.roboflow.com/robot-crknl/oriented-bounding-boxes-dataset/dataset/2

Oriented Bounding Boxes Dataset

oriented-bounding-boxes-dataset

Explore at:
zipAvailable download formats
Dataset updated
Apr 16, 2024
Dataset authored and provided by
Robot
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Variables measured
Robot O0Gq Bounding Boxes
Description

Oriented Bounding Boxes Dataset

## Overview

Oriented Bounding Boxes Dataset is a dataset for object detection tasks - it contains Robot O0Gq annotations for 563 images.

## Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

  ## License

  This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Search
Clear search
Close search
Google apps
Main menu