100+ datasets found
  1. Packages Object Detection Dataset - augmented-v1

    • public.roboflow.com
    zip
    Updated Jan 14, 2021
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow Community (2021). Packages Object Detection Dataset - augmented-v1 [Dataset]. https://public.roboflow.com/object-detection/packages-dataset/5
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 14, 2021
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Roboflow Community
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Bounding Boxes of packages
    Description

    About This Dataset

    The Roboflow Packages dataset is a collection of packages located at the doors of various apartments and homes. Packages are flat envelopes, small boxes, and large boxes. Some images contain multiple annotated packages.

    Usage

    This dataset may be used as a good starter dataset to track and identify when a package has been delivered to a home. Perhaps you want to know when a package arrives to claim it quickly or prevent package theft.

    If you plan to use this dataset and adapt it to your own front door, it is recommended that you capture and add images from the context of your specific camera position. You can easily add images to this dataset via the web UI or via the Roboflow Upload API.

    About Roboflow

    Roboflow enables teams to build better computer vision models faster. We provide tools for image collection, organization, labeling, preprocessing, augmentation, training and deployment. :fa-spacer: Developers reduce boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:

    Roboflow Wordmark

  2. Image augmentation dataset

    • kaggle.com
    zip
    Updated Sep 10, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lyndia Lu (2022). Image augmentation dataset [Dataset]. https://www.kaggle.com/datasets/lyndialu/augment-test-set
    Explore at:
    zip(387658303 bytes)Available download formats
    Dataset updated
    Sep 10, 2022
    Authors
    Lyndia Lu
    Description

    Dataset

    This dataset was created by Lyndia Lu

    Contents

  3. R

    Image Augmentation Dataset

    • universe.roboflow.com
    zip
    Updated Apr 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Data Augmentation (2024). Image Augmentation Dataset [Dataset]. https://universe.roboflow.com/data-augmentation-d7svr/image-augmentation-4ax9o
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 2, 2024
    Dataset authored and provided by
    Data Augmentation
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Fractured Bounding Boxes
    Description

    Image Augmentation

    ## Overview
    
    Image Augmentation is a dataset for object detection tasks - it contains Fractured annotations for 702 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  4. Data from: Variable Message Signal annotated images for object detection

    • zenodo.org
    • portalcientifico.universidadeuropea.com
    zip
    Updated Oct 2, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas (2022). Variable Message Signal annotated images for object detection [Dataset]. http://doi.org/10.5281/zenodo.5904211
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 2, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    If you use this dataset, please cite this paper: Puertas, E.; De-Las-Heras, G.; Sánchez-Soriano, J.; Fernández-Andrés, J. Dataset: Variable Message Signal Annotated Images for Object Detection. Data 2022, 7, 41. https://doi.org/10.3390/data7040041

    This dataset consists of Spanish road images taken from inside a vehicle, as well as annotations in XML files in PASCAL VOC format that indicate the location of Variable Message Signals within them. Also, a CSV file is attached with information regarding the geographic position, the folder where the image is located, and the text in Spanish. This can be used to train supervised learning computer vision algorithms, such as convolutional neural networks. Throughout this work, the process followed to obtain the dataset, image acquisition, and labeling, and its specifications are detailed. The dataset is constituted of 1216 instances, 888 positives, and 328 negatives, in 1152 jpg images with a resolution of 1280x720 pixels. These are divided into 576 real images and 576 images created from the data-augmentation technique. The purpose of this dataset is to help in road computer vision research since there is not one specifically for VMSs.

    The folder structure of the dataset is as follows:

    • vms_dataset/
      • data.csv
      • real_images/
        • imgs/
        • annotations/
      • data-augmentation/
        • imgs/
        • annotations/

    In which:

    • data.csv: Each row contains the following information separated by commas (,): image_name, x_min, y_min, x_max, y_max, class_name, lat, long, folder, text.
    • real_images: Images extracted directly from the videos.
    • data-augmentation: Images created using data-augmentation
    • imgs: Image files in .jpg format.
    • annotations: Annotation files in .xml format.
  5. Augmented Under Water Object Detection

    • kaggle.com
    zip
    Updated Oct 29, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    kendor74 (2024). Augmented Under Water Object Detection [Dataset]. https://www.kaggle.com/datasets/kendor74/augmented-under-water-object-detection
    Explore at:
    zip(2046995671 bytes)Available download formats
    Dataset updated
    Oct 29, 2024
    Authors
    kendor74
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Underwater Object Detection Dataset (COCO Format)

    Dataset Overview

    This dataset is structured for underwater object detection tasks, following the COCO annotation format. It contains both real and augmented images of various underwater objects (e.g., fish, coral, ROVs). Images are grouped into classes, and all annotations are stored in a single JSON file for ease of access and compatibility with most object detection frameworks.

    Dataset Structure

    The dataset folder structure is as follows:

    Underwater_Object_Detection_Dataset/
    ├── combined_images/
    │  ├── animal_fish/
    │  │  ├── real_and_augmented_image1.jpg
    │  │  ├── real_and_augmented_image2.jpg
    │  │  └── ...
    │  ├── plant/
    │  │  ├── real_and_augmented_image1.jpg
    │  │  └── ...
    │  ├── rov/
    │  │  ├── real_and_augmented_image1.jpg
    │  │  └── ...
    │  ├── test/
    │  │  ├── test_image1.jpg
    │  │  ├── test_image2.jpg
    │  │  └── ...
    │  ├── mixed_categories/
    │  │  ├── mixed_image1.jpg
    │  │  ├── mixed_image2.jpg
    │  │  └── ...
    │  └── ...
    ├── combined_annotations.json
    

    Folder Details

    • combined_images/: Contains subfolders for each class, with each folder containing both real and augmented images for that class.
    • test/: Contains images specifically for testing the model, kept separate from the main classes.
    • mixed_categories/: Contains images with multiple object classes in a single image, allowing for multi-object detection tasks.
    • combined_annotations.json: A single JSON file with all image and annotation information, formatted in COCO-style for seamless integration with object detection models.

    Annotations (combined_annotations.json)

    The combined_annotations.json file follows the COCO format, structured into three main sections: images, annotations, and categories.

    Example JSON Structure

    {
      "images": [
        {
          "id": 1,
          "file_name": "vid_000159_frame0000008.jpg",
          "width": 480,
          "height": 270
        },
        {
          "id": 2,
          "file_name": "vid_000339_frame0000012.jpg",
          "width": 480,
          "height": 270
        }
        // Additional images
      ],
      
      "annotations": [
        {
          "segmentation": [],
          "area": 343.875,
          "iscrowd": 0,
          "image_id": 1,
          "bbox": [238.0, 165.0, 18.0, 23.0],
          "category_id": 1,
          "id": 221
        },
        {
          "segmentation": [],
          "area": 500.25,
          "iscrowd": 0,
          "image_id": 2,
          "bbox": [120.0, 140.0, 25.0, 20.0],
          "category_id": 2,
          "id": 222
        }
        // Additional annotations
      ],
      
      "categories": [
        {
          "supercategory": "marine_life",
          "id": 1,
          "name": "fish"
        },
        {
          "supercategory": "marine_life",
          "id": 2,
          "name": "coral"
        },
        {
          "supercategory": "vehicle",
          "id": 3,
          "name": "rov"
        }
        // Additional categories
      ]
    }
    

    JSON Key Explanations

    • images: Contains metadata about each image:

      • "id": Unique identifier for the image.
      • "file_name": File name within its respective class folder.
      • "width" and "height": Dimensions of the image in pixels.
    • annotations: Lists each object annotation with the following details:

      • "segmentation": For polygonal segmentation (empty here as we use bounding boxes only).
      • "area": Area of the bounding box.
      • "iscrowd": Set to 0 for individual objects, 1 if dense clustering.
      • "image_id": Corresponds to the id in images, linking the annotation to its image.
      • "bbox": Bounding box in [x_min, y_min, width, height] format.
      • "category_id": Refers to the object’s class in categories.
      • "id": Unique ID for each annotation.
    • categories: Lists unique object classes in the dataset:

      • "supercategory": High-level grouping for the class.
      • "id": Unique ID for each class.
      • "name": Name of the object class.

    Usage Recommendations

    This dataset is suitable for: - Training and validation for underwater object detection models. - Benchmarking and testing on object detection algorithms. - Exploring domain adaptation using real and augmented underwater images.

    Additional Notes

    • The test/ folder is intended exclusively for testing the model, helping to evaluate its performance on unseen data.
    • The mixed_categories/ folder contains images with multiple object types, making it suitable for multi-object detection challenges, where models need to detect several classes in the same image.
  6. Doors Image Dataset | Indoor Object Detection

    • kaggle.com
    zip
    Updated Feb 18, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    DataCluster Labs (2023). Doors Image Dataset | Indoor Object Detection [Dataset]. https://www.kaggle.com/datasets/dataclusterlabs/doors-doors
    Explore at:
    zip(556294883 bytes)Available download formats
    Dataset updated
    Feb 18, 2023
    Authors
    DataCluster Labs
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    This dataset is collected by DataCluster Labs. To download full dataset or to submit a request for your new data collection needs, please drop a mail to: sales@datacluster.ai

    This dataset is an extremely challenging set of over 3,000+ images of excavator vehicles from multiple construction site. These images captured and crowdsourced from over 2000+ different locations, where each image is manually reviewed and verified by computer vision professionals at Datacluster Labs. It contains a wide variety of indoor door images. This dataset can be used scene classification and domestic object detection.

    Optimized for Generative AI, Visual Question Answering, Image Classification, and LMM development, this dataset provides a strong basis for achieving robust model performance.

    Dataset Features

    • Dataset size : 3000+ images
    • Captured by : Over 2000+ crowdsource contributors
    • Resolution : HD and above (1920x1080 and above)
    • Location : Captured with 2000+ locations
    • Diversity : Various lighting conditions like day, night, varied distances, view points etc.
    • Device used : Captured using mobile phones in 2020-2022
    • Usage : Image classification, domestic object detection, objects relationship understanding etc.

    Available Annotation formats

    COCO, YOLO, PASCAL-VOC, Tf-Record

    The images in this dataset are exclusively owned by Data Cluster Labs and were not downloaded from the internet. To access a larger portion of the training dataset for research and commercial purposes, a license can be purchased. Contact us at sales@datacluster.ai Visit www.datacluster.ai to know more.

  7. R

    Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges

    • public.roboflow.com
    zip
    Updated Sep 30, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Northeastern University - China (2022). Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges [Dataset]. https://public.roboflow.com/object-detection/hard-hat-workers/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 30, 2022
    Dataset authored and provided by
    Northeastern University - China
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Bounding Boxes of Workers
    Description

    Overview

    The Hard Hat dataset is an object detection dataset of workers in workplace settings that require a hard hat. Annotations also include examples of just "person" and "head," for when an individual may be present without a hard hart.

    The original dataset has a 75/25 train-test split.

    Example Image: https://i.imgur.com/7spoIJT.png" alt="Example Image">

    Use Cases

    One could use this dataset to, for example, build a classifier of workers that are abiding safety code within a workplace versus those that may not be. It is also a good general dataset for practice.

    Using this Dataset

    Use the fork or Download this Dataset button to copy this dataset to your own Roboflow account and export it with new preprocessing settings (perhaps resized for your model's desired format or converted to grayscale), or additional augmentations to make your model generalize better. This particular dataset would be very well suited for Roboflow's new advanced Bounding Box Only Augmentations.

    Dataset Versions:

    Image Preprocessing | Image Augmentation | Modify Classes * v1 (resize-416x416-reflect): generated with the original 75/25 train-test split | No augmentations * v2 (raw_75-25_trainTestSplit): generated with the original 75/25 train-test split | These are the raw, original images * v3 (v3): generated with the original 75/25 train-test split | Modify Classes used to drop person class | Preprocessing and Augmentation applied * v5 (raw_HeadHelmetClasses): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class * v8 (raw_HelmetClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and person classes * v9 (raw_PersonClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and helmet classes * v10 (raw_AllClasses): generated with a 70/20/10 train/valid/test split | These are the raw, original images * v11 (augmented3x-AllClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied | 3x image generation | Trained with Roboflow's Fast Model * v12 (augmented3x-HeadHelmetClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Fast Model * v13 (augmented3x-HeadHelmetClasses-AccurateModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Accurate Model * v14 (raw_HeadClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class, and remap/relabel helmet class to head

    Choosing Between Computer Vision Model Sizes | Roboflow Train

    About Roboflow

    Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

    Developers reduce 50% of their code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

    Roboflow Workmark

  8. Experimental results of YOLOv8+WIOU.

    • plos.figshare.com
    xls
    Updated Mar 21, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Meiling Shi; Dongling Zheng; Tianhao Wu; Wenjing Zhang; Ruijie Fu; Kailiang Huang (2024). Experimental results of YOLOv8+WIOU. [Dataset]. http://doi.org/10.1371/journal.pone.0299902.t006
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Mar 21, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Meiling Shi; Dongling Zheng; Tianhao Wu; Wenjing Zhang; Ruijie Fu; Kailiang Huang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Accurate identification of small tea buds is a key technology for tea harvesting robots, which directly affects tea quality and yield. However, due to the complexity of the tea plantation environment and the diversity of tea buds, accurate identification remains an enormous challenge. Current methods based on traditional image processing and machine learning fail to effectively extract subtle features and morphology of small tea buds, resulting in low accuracy and robustness. To achieve accurate identification, this paper proposes a small object detection algorithm called STF-YOLO (Small Target Detection with Swin Transformer and Focused YOLO), which integrates the Swin Transformer module and the YOLOv8 network to improve the detection ability of small objects. The Swin Transformer module extracts visual features based on a self-attention mechanism, which captures global and local context information of small objects to enhance feature representation. The YOLOv8 network is an object detector based on deep convolutional neural networks, offering high speed and precision. Based on the YOLOv8 network, modules including Focus and Depthwise Convolution are introduced to reduce computation and parameters, increase receptive field and feature channels, and improve feature fusion and transmission. Additionally, the Wise Intersection over Union loss is utilized to optimize the network. Experiments conducted on a self-created dataset of tea buds demonstrate that the STF-YOLO model achieves outstanding results, with an accuracy of 91.5% and a mean Average Precision of 89.4%. These results are significantly better than other detectors. Results show that, compared to mainstream algorithms (YOLOv8, YOLOv7, YOLOv5, and YOLOx), the model improves accuracy and F1 score by 5-20.22 percentage points and 0.03-0.13, respectively, proving its effectiveness in enhancing small object detection performance. This research provides technical means for the accurate identification of small tea buds in complex environments and offers insights into small object detection. Future research can further optimize model structures and parameters for more scenarios and tasks, as well as explore data augmentation and model fusion methods to improve generalization ability and robustness.

  9. R

    Image Augmentation And Annotation Dataset

    • universe.roboflow.com
    zip
    Updated Jun 24, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Vico (2022). Image Augmentation And Annotation Dataset [Dataset]. https://universe.roboflow.com/vico/image-augmentation-and-annotation/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 24, 2022
    Dataset authored and provided by
    Vico
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Objects Bounding Boxes
    Description

    Image Augmentation And Annotation

    ## Overview
    
    Image Augmentation And Annotation is a dataset for object detection tasks - it contains Objects annotations for 431 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  10. PotHole Detector Dataset Augmented

    • kaggle.com
    zip
    Updated Nov 4, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    VincentMaes (2024). PotHole Detector Dataset Augmented [Dataset]. https://www.kaggle.com/datasets/vincenttgre/pothole-detector-dataset-augmented
    Explore at:
    zip(1507748590 bytes)Available download formats
    Dataset updated
    Nov 4, 2024
    Authors
    VincentMaes
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Road Damage and Pothole Detection Dataset

    Overview

    This dataset is specifically curated for object detection tasks aimed at identifying and classifying road damage and potholes. The original dataset on which this augmented dataset is based, included images labeled with four distinct classes: - Pothole - Alligator Crack - Long Crack - Lat Crack However, for training the model for detecting road damages, it has been placed into 1 class, namely the "Pothole" class, which now also includes the alligator, longitudinal, and lateral cracks.

    Data Augmentation

    To enhance the robustness and generalization capability of models trained on this dataset, extensive data augmentation techniques have been applied. The augmentation pipeline includes:

    • Horizontal Flip (50% probability)
    • Vertical Flip (10% probability)
    • Random Rotation by 90 degrees (50% probability)
    • Rotation (±10 degrees, 50% probability)
    • Random Brightness and Contrast adjustments (50% probability)
    • Gaussian Blur (30% probability)
    • Color Jitter (30% probability)
    • Random Scaling (±10% scale, 50% probability)
    • Perspective Transformations (scale range 0.05 to 0.1, 30% probability)

    These augmentations ensure that models can learn to recognize road damages under various conditions and viewpoints, improving their detection performance.

    Bounding Box Parameters

    Bounding boxes are provided in the YOLO format, ensuring easy integration with popular object detection frameworks. The bounding boxes are adjusted to correspond with the augmented images to maintain annotation accuracy.

    Classes

    The dataset includes the following class:

    Class ID Class Name 0 Pothole

    Data Split

    The dataset is divided into training, validation, and testing sets with the following proportions:

    • Training: 85%
    • Validation: 7%
    • Testing: 8%

    This split ensures a sufficient amount of data for training the model while maintaining enough data for validation and testing to assess model performance accurately.

    Conclusion

    This dataset aims to aid researchers and developers in building and fine-tuning models for road damage detection, contributing to safer and more efficient road maintenance systems.

  11. Performance comparison of the OFIDA and several SOTA data augmentation...

    • plos.figshare.com
    xls
    Updated May 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan (2024). Performance comparison of the OFIDA and several SOTA data augmentation methods for image classification. [Dataset]. http://doi.org/10.1371/journal.pone.0302124.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    May 2, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Performance comparison of the OFIDA and several SOTA data augmentation methods for image classification.

  12. A dataset for window and blind states detection

    • figshare.com
    bin
    Updated Aug 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Seunghyeon Wang (2024). A dataset for window and blind states detection [Dataset]. http://doi.org/10.6084/m9.figshare.26403004.v1
    Explore at:
    binAvailable download formats
    Dataset updated
    Aug 5, 2024
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Seunghyeon Wang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The data was constructed for detecting window and blind states. All images were annotated in XML format using LabelImg for object detection tasks. The results of applying the Faster R-CNN based model include detected images and loss graphs for both training and validation in this dataset. Additionally, the raw data with other annotations can be used for applications such as semantic segmentation and image captioning.

  13. Vehicle Detection Dataset image

    • kaggle.com
    zip
    Updated May 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daud shah (2025). Vehicle Detection Dataset image [Dataset]. https://www.kaggle.com/datasets/daudshah/vehicle-detection-dataset
    Explore at:
    zip(545957939 bytes)Available download formats
    Dataset updated
    May 29, 2025
    Authors
    Daud shah
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Vehicle Detection Dataset

    This dataset is designed for vehicle detection tasks, featuring a comprehensive collection of images annotated for object detection. This dataset, originally sourced from Roboflow (https://universe.roboflow.com/object-detection-sn8ac/ai-traffic-system), was exported on May 29, 2025, at 4:59 PM GMT and is now publicly available on Kaggle under the CC BY 4.0 license.

    Overview

    • Purpose: The dataset supports the development of computer vision models for detecting various types of vehicles in traffic scenarios.
    • Classes: The dataset includes annotations for 7 vehicle types:
      • Bicycle
      • Bus
      • Car
      • Motorbike
      • Rickshaw
      • Truck
      • Van
    • Number of Images: The dataset contains 9,440 images, split into training, validation, and test sets:
      • Training: Images located in ../train/images
      • Validation: Images located in ../valid/images
      • Test: Images located in ../test/images
    • Annotation Format: Images are annotated in YOLOv11 format, suitable for training state-of-the-art object detection models.
    • Pre-processing: Each image has been resized to 640x640 pixels (stretched). No additional image augmentation techniques were applied.

    Source and Creation

    This dataset was created and exported via Roboflow, an end-to-end computer vision platform that facilitates collaboration, image collection, annotation, dataset creation, model training, and deployment. The dataset is part of the ai-traffic-system project (version 1) under the workspace object-detection-sn8ac. For more details, visit: https://universe.roboflow.com/object-detection-sn8ac/ai-traffic-system/dataset/1.

    Usage

    This dataset is ideal for researchers, data scientists, and developers working on vehicle detection and traffic monitoring systems. It can be used to: - Train and evaluate deep learning models for object detection, particularly using the YOLOv11 framework. - Develop AI-powered traffic management systems, autonomous driving applications, or urban mobility solutions. - Explore computer vision techniques for real-world traffic scenarios.

    For advanced training notebooks compatible with this dataset, check out: https://github.com/roboflow/notebooks. To explore additional datasets and pre-trained models, visit: https://universe.roboflow.com.

    License

    The dataset is licensed under CC BY 4.0, allowing for flexible use, sharing, and adaptation, provided appropriate credit is given to the original source.

    This dataset is a valuable resource for building robust vehicle detection models and advancing computer vision applications in traffic systems.

  14. Parameters setting.

    • plos.figshare.com
    xls
    Updated May 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan (2024). Parameters setting. [Dataset]. http://doi.org/10.1371/journal.pone.0302124.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    May 2, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Image data augmentation plays a crucial role in data augmentation (DA) by increasing the quantity and diversity of labeled training data. However, existing methods have limitations. Notably, techniques like image manipulation, erasing, and mixing can distort images, compromising data quality. Accurate representation of objects without confusion is a challenge in methods like auto augment and feature augmentation. Preserving fine details and spatial relationships also proves difficult in certain techniques, as seen in deep generative models. To address these limitations, we propose OFIDA, an object-focused image data augmentation algorithm. OFIDA implements one-to-many enhancements that not only preserve essential target regions but also elevate the authenticity of simulating real-world settings and data distributions. Specifically, OFIDA utilizes a graph-based structure and object detection to streamline augmentation. Specifically, by leveraging graph properties like connectivity and hierarchy, it captures object essence and context for improved comprehension in real-world scenarios. Then, we introduce DynamicFocusNet, a novel object detection algorithm built on the graph framework. DynamicFocusNet merges dynamic graph convolutions and attention mechanisms to flexibly adjust receptive fields. Finally, the detected target images are extracted to facilitate one-to-many data augmentation. Experimental results validate the superiority of our OFIDA method over state-of-the-art methods across six benchmark datasets.

  15. Yolo tiger and lion labelled detection

    • kaggle.com
    zip
    Updated Sep 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Junkie75 (2024). Yolo tiger and lion labelled detection [Dataset]. https://www.kaggle.com/datasets/junkie75/yolo-tiger-and-lion-labelled-detection/discussion
    Explore at:
    zip(64999035 bytes)Available download formats
    Dataset updated
    Sep 10, 2024
    Authors
    Junkie75
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    This dataset contains images of lions and tigers sourced from the Open Images Dataset V6 and labeled specifically for object detection using the YOLO format. The dataset focuses on two classes: lion and tiger, with annotations provided for each image in a YOLO-compatible .txt file format. This dataset is ideal for training machine learning models for wildlife detection and classification tasks, particularly in distinguishing between these two majestic big cats. Key Features:

    Classes: Lion and Tiger
    Annotations: YOLO format, with bounding box coordinates and class labels provided in separate .txt files for each image.
    Source: Images sourced from Open Images Dataset V6, which is published under the Creative Commons Attribution 4.0 International License (CC BY 4.0).
    Application: Suitable for object detection models like YOLO, SSD, or Faster R-CNN.
    

    Usage:

    The dataset can be used for training, validating, or testing object detection models. Each image is accompanied by a corresponding YOLO annotation file, making it easy to integrate into any YOLO-based pipeline. Attribution:

    This dataset is derived from the Open Images Dataset V6, and proper attribution must be given. Please credit the Open Images Dataset when using or sharing this dataset in any format.

  16. R

    Data Augmentation Data Adjust 5k Dataset

    • universe.roboflow.com
    zip
    Updated May 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    grapevine data augmentation (2025). Data Augmentation Data Adjust 5k Dataset [Dataset]. https://universe.roboflow.com/grapevine-data-augmentation/data-augmentation-data-adjust-5k
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 29, 2025
    Dataset authored and provided by
    grapevine data augmentation
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Leaves Bounding Boxes
    Description

    Data Augmentation Data Adjust 5k

    ## Overview
    
    Data Augmentation Data Adjust 5k is a dataset for object detection tasks - it contains Leaves annotations for 4,994 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  17. u

    Building facade images for window states

    • rdr.ucl.ac.uk
    • resodate.org
    zip
    Updated Mar 15, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Seunghyeon Wang; Ivan Korolija; Dimitrios Rovas (2022). Building facade images for window states [Dataset]. http://doi.org/10.5522/04/14993589.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 15, 2022
    Dataset provided by
    University College London
    Authors
    Seunghyeon Wang; Ivan Korolija; Dimitrios Rovas
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The data was used for "Impact of Traditional Augmentation Methods on Window States Detection", which is a conference paper of CLIMA2022. The main purpose of this data is for reproductivity of proposed methods. All images are annotated with XML format using LabelImg. Additionally, this dataset may be used for other object detection and segmentation tasks as a possible application.

  18. m

    BRAGAN: a GAN-augmented dataset of Brazilian roadkill animals for object...

    • data.mendeley.com
    Updated Aug 20, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Henrique Souza de Abreu Martins (2025). BRAGAN: a GAN-augmented dataset of Brazilian roadkill animals for object detection [Dataset]. http://doi.org/10.17632/ck88dwffgd.2
    Explore at:
    Dataset updated
    Aug 20, 2025
    Authors
    Henrique Souza de Abreu Martins
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    BRAGAN is a new dataset of Brazilian wildlife developed for object detection tasks, combining real images with synthetic samples generated by Generative Adversarial Networks (GANs). It focuses on five medium and large-sized mammal species frequently involved in roadkill incidents on Brazilian highways: lowland tapir (Tapirus terrestris), jaguarundi (Herpailurus yagouaroundi), maned wolf (Chrysocyon brachyurus), puma (Puma concolor), and giant anteater (Myrmecophaga tridactyla). Its primary goal is to provide a standardized and expanded resource for biodiversity conservation research, wildlife monitoring technologies, and computer vision applications, with an emphasis on automated wildlife detection.

    The dataset builds upon the original BRA-Dataset by Ferrante et al. (2022), which was constructed from structured internet searches and manually curated with bounding box annotations. However, while the BRA-Dataset faced limitations in size and variability, BRAGAN introduces a new stage of dataset expansion through GAN-based synthetic image generation, substantially improving both the quantity and diversity of samples. In its final version, BRAGAN comprises approximately 9,238 images, divided into three main groups:

    Real images — original photographs from the BRA-Dataset. Total: 1,823.

    Classically augmented images — transformations applied to real samples, including rotations (RT), horizontal flips (HF), vertical flips (VF), and horizontal (HS) and vertical shifts (VS). Total: 7,300.

    GAN-generated images — synthetic samples created using WGAN-GP models trained separately for each species on preprocessed subsets of the original data. All generated images underwent visual inspection to ensure morphological fidelity and proper framing before inclusion. Total: 115.

    The dataset follows an organized directory structure with images/ and labels/ folders, each divided into train/ and val/ subsets, following an 80–20 split. Images are provided in .jpg format, while annotations follow the YOLO standard in .txt files (class_id x_center y_center width height, with normalized coordinates). The file naming convention explicitly encodes the species and the augmentation type for reproducibility.

    Designed to be compatible with multiple object detection architectures, BRAGAN has been evaluated on YOLOv5, YOLOv8, and YOLOv11 (variants n, s, and m), enabling the assessment of dataset expansion across different computational settings and performance requirements.

    By combining real data, classical augmentations, and high-quality synthetic samples, the BRAGAN provides a valuable resource for wildlife detection, environmental monitoring, and conservation research, especially in contexts where image availability for rare or threatened species is limited.

  19. CADOT Dataset

    • kaggle.com
    zip
    Updated May 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tú Hoàng (2025). CADOT Dataset [Dataset]. https://www.kaggle.com/datasets/grizmo/cadot-dataset
    Explore at:
    zip(206172834 bytes)Available download formats
    Dataset updated
    May 28, 2025
    Authors
    Tú Hoàng
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    The CADOT dataset is introduced as part of the Grand Challenge at IEEE ICIP 2025, aiming to push forward the development of advanced object detection techniques in remote sensing imagery, particularly focused on dense urban environments. The competition is organized by LabCom IRISER, in collaboration with IGN (Institut national de l'information géographique et forestière), and encourages the use of AI-based data augmentation to enhance model robustness.

    Competition Context

    The challenge calls for the detection of small objects in high-resolution optical satellite imagery, which is inherently complex due to occlusions, diverse object types, and varied urban layouts. Participants are expected to develop detection pipelines that are not only accurate but also robust under real-world remote sensing constraints.

    Dataset Description

    The CADOT dataset comprises high-resolution aerial images captured over a dense urban area in the Île-de-France region, France. Each image is carefully annotated with 14 object categories including buildings, roads, vehicles, trees, and various other urban components. The imagery comes from IGN and reflects a realistic and challenging setting for object detection models due to factors like shadows, perspective distortion, and dense object arrangements.

    • Data is derived from multi-view aerial imagery
    • Images are orthorectified to remove perspective distortion
    • Labeling is provided in geospatially aligned formats
    • Annotations are polygonal or bounding-box-based (depending on release phase)

    Data Reformatting

    To facilitate easier use of the dataset in machine learning workflows, I have reformatted the original data into the following versions:

    • Images in .jpg and .png format (cropped and full-frame)
    • Annotations converted to COCO JSON and YOLO format
    • Train/val/test splits based on geographic segmentation
    • A preview subset for quick experimentation

    License and Original Source

    For full licensing terms and official documentation, please refer to the official challenge page: 🔗 https://cadot.onrender.com/

  20. Performance evaluation of semantic segmentation on the CITYSCAPES validation...

    • figshare.com
    xls
    Updated May 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan (2024). Performance evaluation of semantic segmentation on the CITYSCAPES validation set using mIoU. [Dataset]. http://doi.org/10.1371/journal.pone.0302124.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    May 2, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Meng Zhang; Yina Guo; Haidong Wang; Hong Shangguan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Performance evaluation of semantic segmentation on the CITYSCAPES validation set using mIoU.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Roboflow Community (2021). Packages Object Detection Dataset - augmented-v1 [Dataset]. https://public.roboflow.com/object-detection/packages-dataset/5
Organization logo

Packages Object Detection Dataset - augmented-v1

Explore at:
zipAvailable download formats
Dataset updated
Jan 14, 2021
Dataset provided by
Roboflowhttps://roboflow.com/
Authors
Roboflow Community
License

CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically

Variables measured
Bounding Boxes of packages
Description

About This Dataset

The Roboflow Packages dataset is a collection of packages located at the doors of various apartments and homes. Packages are flat envelopes, small boxes, and large boxes. Some images contain multiple annotated packages.

Usage

This dataset may be used as a good starter dataset to track and identify when a package has been delivered to a home. Perhaps you want to know when a package arrives to claim it quickly or prevent package theft.

If you plan to use this dataset and adapt it to your own front door, it is recommended that you capture and add images from the context of your specific camera position. You can easily add images to this dataset via the web UI or via the Roboflow Upload API.

About Roboflow

Roboflow enables teams to build better computer vision models faster. We provide tools for image collection, organization, labeling, preprocessing, augmentation, training and deployment. :fa-spacer: Developers reduce boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:

Roboflow Wordmark

Search
Clear search
Close search
Google apps
Main menu