100+ datasets found
  1. n

    FishEye8K: A Benchmark and Dataset for Fisheye Camera Object Detection -...

    • scidm.nchc.org.tw
    Updated Jul 12, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). FishEye8K: A Benchmark and Dataset for Fisheye Camera Object Detection - Dataset - 國網中心Dataset平台 [Dataset]. https://scidm.nchc.org.tw/dataset/fisheye8k
    Explore at:
    Dataset updated
    Jul 12, 2024
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    With the advance of AI, road object detection has been a prominent topic in computer vision, mostly using perspective cameras. Fisheye lens provides omnidirectional wide coverage for using fewer cameras to monitor road intersections, however with view distortions. To our knowledge, there is no existing open dataset prepared for traffic surveillance on fisheye cameras. This paper introduces an open FishEye8K benchmark dataset for road object detection tasks, which comprises 157K bounding boxes across five classes (Pedestrian, Bike, Car, Bus, and Truck). In addition, we present benchmark results of State-of-The-Art (SoTA) models, including variations of YOLOv5, YOLOR, YOLO7, and YOLOv8. The dataset comprises 8,000 images recorded in 22 videos using 18 fisheye cameras for traffic monitoring in Hsinchu, Taiwan, at resolutions of 1080x1080 and 1280x1280. The data annotation and validation process were arduous and time-consuming, due to the ultra-wide panoramic and hemispherical fisheye camera images with large distortion and numerous road participants, particularly people riding scooters. To avoid bias, frames from a particular camera were assigned to either the training or test sets, maintaining a ratio of about 70:30 for both the number of images and bounding boxes in each class. Experimental results show that YOLOv8 and YOLOR outperform on input sizes 640x640 and 1280x1280, respectively. copyright

  2. PASCAL VOC 2012

    • kaggle.com
    Updated Jul 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Banuprasad B (2025). PASCAL VOC 2012 [Dataset]. https://www.kaggle.com/datasets/banuprasadb/pascal-voc-2012
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 5, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Banuprasad B
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    The PASCAL Visual Object Classes (VOC) 2012 dataset is a benchmark in object recognition, widely used for training and evaluating models in computer vision tasks.

    The dataset has been modified to include only the image data and labels in YOLO format. The original annotation files have been removed, and object labels were converted using provided scripts(from Ultralytics) to be compatible with YOLO-based object detection models.

  3. h

    object-detection-bench

    • huggingface.co
    Updated May 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    JigsawStack (2025). object-detection-bench [Dataset]. https://huggingface.co/datasets/JigsawStack/object-detection-bench
    Explore at:
    Dataset updated
    May 28, 2025
    Dataset authored and provided by
    JigsawStack
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Object Detection Bench

    This dataset is a customized version of the RealworldQA dataset, specifically tailored for object detection and segmentation benchmarking tasks.

      Dataset Description
    

    This benchmark dataset contains real-world images with questions, answers, and custom prompts designed for evaluating object detection and segmentation models. Each sample includes:

    Image: Real-world photographs Question: Original question about the image content Answer: Ground truth… See the full description on the dataset page: https://huggingface.co/datasets/JigsawStack/object-detection-bench.

  4. S

    A Benchmark Dataset for Fine-Grained Object Detection and Recognition Based...

    • scidb.cn
    Updated Feb 20, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    吴有明; 刁文辉; 索玉玺; 孙显 (2025). A Benchmark Dataset for Fine-Grained Object Detection and Recognition Based on Single-Look Complex SAR Images (FAIR-CSAR-V1.0) [Dataset]. http://doi.org/10.57760/sciencedb.radars.00019
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 20, 2025
    Dataset provided by
    Science Data Bank
    Authors
    吴有明; 刁文辉; 索玉玺; 孙显
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    FAIR-CSAR-V1.0 dataset, constructed on single-look complex (SLC) images of Gaofen-3 satellite, is the largest and most finely annotated SAR image dataset for fine-grained target to date. FAIR-CSAR-V1.0 aims to advance related technologies in SAR image object detection, recognition, and target characteristic understanding. The dataset is developed by Key Laboratory of Target Cognition and Application Technology (TCAT) at the Aerospace Information Research Institute, Chinese Academy of Sciences.FAIR-CSAR-V1.0 comprises 175 scenes of Gaofen-3 Level-1 SLC products, covering 32 global regions including airports, oil refineries, ports, and rivers. With a total data volume of 250 GB and over 340,000 instances, FAIR-CSAR-V1.0 covers 5 main categories and 22 subcategories, providing detailed annotations for imaging parameters (e.g., radar center frequency, pulse repetition frequency) and target characteristics (e.g., satellite-ground relative azimuthal angle, key scattering point distribution).FAIR-CSAR-V1.0 consists of two sub-datasets: the SL dataset and the FSI dataset. The SL dataset, acquired in spotlight mode with a nominal resolution of 1 meter, contains 170,000 instances across 22 target classes. The FSI dataset, acquired in fine stripmap mode with a nominal resolution of 5 meters, includes 170,000 instances across 3 target classes. Figure 1 presents an overview of the dataset.Data paper and citation format:[1] Youming Wu, Wenhui Diao, Yuxi Suo, Xian Sun. A Benchmark Dataset for Fine-Grained Object Detection and Recognition Based on Single-Look Complex SAR Images (FAIR-CSAR-V1.0) [OL]. Journal of Radars, 2025. https://radars.ac.cn/web/data/getData?dataType=FAIR_CSAR_en&pageType=en.[2] Y. Wu, Y. Suo, Q. Meng, W. Dai, T. Miao, W. Zhao, Z. Yan, W. Diao, G. Xie, Q. Ke, Y. Zhao, K. Fu and X. Sun, FAIR-CSAR: A Benchmark Dataset for Fine-Grained Object Detection and Recognition Based on Single-Look Complex SAR Images[J]. IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1-22, 2025, doi: 10.1109/TGRS.2024.3519891.

  5. R

    Pintel Firedet Benchmark Dataset

    • universe.roboflow.com
    zip
    Updated May 25, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    PINTEL (2023). Pintel Firedet Benchmark Dataset [Dataset]. https://universe.roboflow.com/pintel-f4hbd/pintel-firedet-benchmark-dataset
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 25, 2023
    Dataset authored and provided by
    PINTEL
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Fire Bounding Boxes
    Description

    PINTEL FireDet Benchmark Dataset

    ## Overview
    
    PINTEL FireDet Benchmark Dataset is a dataset for object detection tasks - it contains Fire annotations for 390 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  6. Z

    Data from: Tiny Robotics Dataset and Benchmark for Continual Object...

    • nde-dev.biothings.io
    • data.niaid.nih.gov
    • +1more
    Updated Mar 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Francesco, Pasti (2025). Tiny Robotics Dataset and Benchmark for Continual Object Detection [Dataset]. https://nde-dev.biothings.io/resources?id=zenodo_13834549
    Explore at:
    Dataset updated
    Mar 11, 2025
    Dataset authored and provided by
    Francesco, Pasti
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset for TiROD: Tiny Robotics Dataset and Benchmark for Continual Object DetectionOfficial Website -> https://pastifra.github.io/TiROD/

    Code -> https://github.com/pastifra/TiROD_code

    Video -> https://www.youtube.com/watch?v=e76m3ol1i4I

    Paper -> https://arxiv.org/abs/2409.16215

  7. Cloud_satelite_maxar_dataset

    • kaggle.com
    zip
    Updated Feb 22, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    kishor datta gupta (2025). Cloud_satelite_maxar_dataset [Dataset]. https://www.kaggle.com/datasets/kishor1123/cloud-satelite-maxar-dataset
    Explore at:
    zip(98887314 bytes)Available download formats
    Dataset updated
    Feb 22, 2025
    Authors
    kishor datta gupta
    License

    http://opendatacommons.org/licenses/dbcl/1.0/http://opendatacommons.org/licenses/dbcl/1.0/

    Description

    CDD & CloudNet: A Benchmark Dataset & Model for Object Detection Performance

    Overview

    Welcome to the CloudNet repository. This project provides a cloud detection dataset and a pre-trained model designed to enhance object detection accuracy in remote sensing aerial images, particularly in challenging cloud-covered scenarios. The dataset comprises two classes: cloud and non-cloud images, sourced from the publicly available Maxar "Hurricane Ian" repository.

    CloudNet Dataset

    Description

    The CloudNet dataset consists of cloud and non-cloud images, facilitating research in cloud detection for object detection in remote sensing imagery.

    Dataset Structure

    • cloud_images: Contains images with clouds.
    • non_cloud_images: Includes images without clouds.

    CloudNet Model

    Description

    The CloudNet model is a pre-trained model specifically designed for cloud detection in remote sensing imagery. It is trained on the CloudNet dataset and serves as a valuable tool for enhancing object detection accuracy in the presence of clouds.

    Model Weights

    You can download the pre-trained CloudNet model weights from the following link: CloudNet Model Weights

    Citation

    If you find the CloudNet dataset or model useful in your research, please cite our work using the following BibTeX entry:

    @INPROCEEDINGS{10747011,
     author={Haque, Mohd Ariful and Rifat, Rakib Hossain and Kamal, Marufa and George, Roy and Gupta, Kishor Datta and Shujaee, Khalil},
     booktitle={2024 Fifth International Conference on Intelligent Data Science Technologies and Applications (IDSTA)}, 
     title={CDD & CloudNet: A Benchmark Dataset & Model for Object Detection Performance}, 
     year={2024},
     volume={},
     number={},
     pages={118-122},
     abstract={Aerial imagery obtained through remote sensing is extensively utilized across diverse industries, particularly for object detection applications where it has demonstrated considerable efficacy. However, clouds in these images can obstruct evaluation and detection tasks. This study therefore involved the compilation of a cloud dataset, which categorized images into two classes: those containing clouds and those without. These images were sourced from the publicly available Maxar ‘Hurricane Ian’ repository, which contains images from various natural events. We demonstrated the impact of cloud removal during pre-processing on object detection using this dataset and employed six CNN models, including a custom model, for cloud detection benchmarking. These models were used to detect objects in aerial images from two other events in the Maxar dataset. Our results show significant improvements in precision, recall, and F1-score for CNN models, along with optimized training times for object detection in the CloudNet+YOLO combination. The findings demonstrate the effectiveness of our approach in improving object detection accuracy and efficiency in remote sensing imagery, particularly in challenging cloud-covered scenarios.},
     keywords={Training;Industries;Accuracy;Object detection;Benchmark testing;Data science;Data models;Remote sensing;Cloud Detection;Dataset;Deep Learning;CNN;ResNet;Vgg16;DenseNet169;EfficientNet;MobileNet},
     doi={10.1109/IDSTA62194.2024.10747011},
     ISSN={},
     month={Sep.},}
    
    

    License

    The CloudNet dataset and model are released under the License.

  8. FG-OVD: Fine-grained Open-Vocabulary Object Detection Benchmark Suite

    • data.europa.eu
    unknown
    Updated Apr 23, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zenodo (2024). FG-OVD: Fine-grained Open-Vocabulary Object Detection Benchmark Suite [Dataset]. https://data.europa.eu/data/datasets/oai-zenodo-org-11060560?locale=en
    Explore at:
    unknown(164986011)Available download formats
    Dataset updated
    Apr 23, 2024
    Dataset authored and provided by
    Zenodohttp://zenodo.org/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    A collection of annotations for PACO images containing free-form fine-grained textual captions of objects, their parts, and their attributes. It also comprises several sets of negative captions that can be used to test and evaluate the fine-grained recognition ability of open-vocabulary models.

  9. Underwater Objects Dataset

    • universe.roboflow.com
    zip
    Updated May 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow 100 (2023). Underwater Objects Dataset [Dataset]. https://universe.roboflow.com/roboflow-100/underwater-objects-5v7p8/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 7, 2023
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Roboflow 100
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Underwater Objects Bounding Boxes
    Description

    This dataset was originally created by Yimin Chen. To see the current project, which may have been updated since this version, please go here: https://universe.roboflow.com/workspace-txxpz/underwater-detection.

    This dataset is part of RF100, an Intel-sponsored initiative to create a new object detection benchmark for model generalizability.

    Access the RF100 Github repo: https://github.com/roboflow-ai/roboflow-100-benchmark

  10. Cyclist Dataset for Object Detection

    • kaggle.com
    zip
    Updated Mar 15, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SemiEmptyGlass (2022). Cyclist Dataset for Object Detection [Dataset]. https://www.kaggle.com/datasets/semiemptyglass/cyclist-dataset
    Explore at:
    zip(2319730694 bytes)Available download formats
    Dataset updated
    Mar 15, 2022
    Authors
    SemiEmptyGlass
    License

    Attribution-NonCommercial-ShareAlike 3.0 (CC BY-NC-SA 3.0)https://creativecommons.org/licenses/by-nc-sa/3.0/
    License information was derived automatically

    Description

    Cyclist Dataset

    Tsinghua-Daimler Cyclist Detection Benchmark Dataset in yolo format for Object Detection

    Context

    I'm not owner the of this dataset, all the credit goes to X. Li, F. Flohr, Y. Yang, H. Xiong, M. Braun, S. Pan, K. Li and D. M. Gavrila, the creators of this dataset.

    Content

    • img size - 2048x1024
    • 13.7k labeled images (1000 images have no cyclists)
    • labels in yolo format: id center_x center_y width height (relative to image width and height)

    Example yolo bounding box:

    0 0.41015625 0.44140625 0.0341796875 0.11328125
    

    Acknowledgments

    License Terms

    This dataset is made freely available non-commercial purposes such as academic research, teaching, scientific publications, or personal experimentation. Permission is granted to use, copy, and distribute the data given that you agree:

    • That the dataset comes "AS IS", without express or implied warranty. Although every effort has been made to ensure accuracy, Daimler (or the website host) does not accept any responsibility for errors or omissions.
    • That you include a reference to the above publication in any published work that makes use of the dataset.
    • That if you have altered the content of the dataset or created derivative work, prominent notices are made so that any recipients know that they are not receiving the original data.
    • That you may not use or distribute the dataset or any derivative work for commercial purposes as, for example, licensing or selling the data, or using the data with a purpose to procure a commercial gain.
    • That this original license notice is retained with all copies or derivatives of the dataset.
    • That all rights not expressly granted to you are reserved by Daimler.

    Cite

    X. Li, F. Flohr, Y. Yang, H. Xiong, M. Braun, S. Pan, K. Li and D. M. Gavrila. A New Benchmark for Vision-Based Cyclist Detection. In Proc. of the IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, pp.1028-1033, 2016.
    
  11. DUTS Salient Object Detection Dataset

    • kaggle.com
    zip
    Updated Nov 6, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Balraj Ashwath (2020). DUTS Salient Object Detection Dataset [Dataset]. https://www.kaggle.com/datasets/balraj98/duts-saliency-detection-dataset/discussion
    Explore at:
    zip(410744652 bytes)Available download formats
    Dataset updated
    Nov 6, 2020
    Authors
    Balraj Ashwath
    Description

    Context

    DUTS is a large-scale saliency detection dataset, containing 10,553 training images and 5,019 test images. All training images are collected from the ImageNet DET training/val sets, while test images are collected from the ImageNet DET test set and the SUN data set. Both the training and test set contain very challenging scenarios for saliency detection. Accurate pixel-level ground truths were manually annotated by 50 subjects.

    Acknowledgements

    This dataset is obtained from the official DUTS dataset homepage. Any work based on the dataset checkpoints should cite:

    @inproceedings{wang2017,
    title={Learning to Detect Salient Objects with Image-level Supervision},
    author={Wang, Lijun and Lu, Huchuan and Wang, Yifan and Feng, Mengyang 
    and Wang, Dong, and Yin, Baocai and Ruan, Xiang}, 
    booktitle={CVPR},
    year={2017}
    }
    

    Copyright claim of the ground-truth annotation

    All rights reserved by the original authors of DUTS Image Dataset.

  12. r

    VIMER-UFO Benchmark

    • resodate.org
    • service.tib.eu
    Updated Dec 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dongning Ma; Pengfei Zhao; Xun Jiao (2024). VIMER-UFO Benchmark [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvdmltZXItdWZvLWJlbmNobWFyaw==
    Explore at:
    Dataset updated
    Dec 16, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Dongning Ma; Pengfei Zhao; Xun Jiao
    Description

    The VIMER-UFO benchmark dataset consists of 8 computer vision tasks: CPLFW, Market1501, DukeMTMC, MSMT-17, Veri-776, VehicleId, VeriWild, and SOP.

  13. VisDrone Dataset

    • kaggle.com
    zip
    Updated Jun 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Banuprasad B (2025). VisDrone Dataset [Dataset]. https://www.kaggle.com/datasets/banuprasadb/visdrone-dataset/data
    Explore at:
    zip(2251268022 bytes)Available download formats
    Dataset updated
    Jun 28, 2025
    Authors
    Banuprasad B
    License

    http://www.gnu.org/licenses/agpl-3.0.htmlhttp://www.gnu.org/licenses/agpl-3.0.html

    Description

    The VisDrone dataset is a large-scale visual object detection and tracking benchmark captured by drones. Developed by the AISKYEYE team at Tianjin University, it aims to facilitate research in computer vision tasks such as object detection, object tracking, and crowd analysis in aerial imagery.

    The dataset consists of high-resolution images and videos collected using drones flying over urban and suburban environments across various cities in China. These scenes include pedestrians, vehicles, bicycles, and other common objects, captured under different lighting conditions, angles, and motion patterns.

    The dataset has been modified to include only the image data and labels in YOLO format. The original annotation files have been removed, and object labels were converted using provided scripts(from Ultralytics) to be compatible with YOLO-based object detection models.

  14. R

    Testing Benchmark Dataset

    • universe.roboflow.com
    zip
    Updated Nov 28, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mark (2024). Testing Benchmark Dataset [Dataset]. https://universe.roboflow.com/mark-fpqrq/testing-benchmark/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 28, 2024
    Dataset authored and provided by
    Mark
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Label Bounding Boxes
    Description

    Testing Benchmark

    ## Overview
    
    Testing Benchmark is a dataset for object detection tasks - it contains Label annotations for 504 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  15. h

    TiROD

    • huggingface.co
    Updated Sep 24, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Francesco Pasti (2024). TiROD [Dataset]. https://huggingface.co/datasets/pastifra/TiROD
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 24, 2024
    Authors
    Francesco Pasti
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset Website Dataset video Code Paper

  16. h

    m4-sar

    • huggingface.co
    Updated May 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    wchao0601 (2025). m4-sar [Dataset]. https://huggingface.co/datasets/wchao0601/m4-sar
    Explore at:
    Dataset updated
    May 14, 2025
    Authors
    wchao0601
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    M4-SAR: A Multi-Resolution, Multi-Polarization, Multi-Scene, Multi-Source Dataset and Benchmark for Optical-SAR Fusion Object Detection 2025

      Dataset description
    

    M4-SAR: A Multi-Resolution, Multi-Polarization, Multi-Scene, Multi-Source Dataset and Benchmark for Optical-SAR Fusion Object Detection

      Contact
    

    If you have any questions, please feel free to contact me via email at wchao0601@163.com

      Citation
    

    If our work is helpful, you can cite our paper:… See the full description on the dataset page: https://huggingface.co/datasets/wchao0601/m4-sar.

  17. O

    DIOR

    • opendatalab.com
    zip
    Updated Mar 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Technical University of Munich (2023). DIOR [Dataset]. https://opendatalab.com/OpenDataLab/DIOR
    Explore at:
    zip(7846302896 bytes)Available download formats
    Dataset updated
    Mar 3, 2023
    Dataset provided by
    Technical University of Munich
    Zhengzhou Institute of Surveying and Mapping
    Northwestern Polytechnical University
    Description

    "DIOR" is a large-scale benchmark dataset for object detection in optical remote sensing images, which consists of 23,463 images and 192,518 object instances annotated with horizontal bounding boxes.

  18. NCSTP: A Benchmark Dataset for Non-Cooperative Space Target Perception

    • figshare.com
    zip
    Updated Jun 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yuxuan Liu (2025). NCSTP: A Benchmark Dataset for Non-Cooperative Space Target Perception [Dataset]. http://doi.org/10.6084/m9.figshare.28606754.v4
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 16, 2025
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Yuxuan Liu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The automatic, accurate perception of targets in space is a crucial prerequisite for many aerospace missions, such as on-orbit maintenance and target monitoring. Therefore, research on perception technologies within images from spaceborne cameras, is of great significance. The recent, rapid development of deep learning has revealed its potential for application to space target perception. However, implementing deep learning models typically requires large-scale labeled datasets. In this study, we build a multitask benchmark space target dataset, NCSTP, to address the limitations of current datasets. First, we collect and modify various space target models for satellites, space debris, and space rocks. By importing them into a realistic space environment simulated by Blender, 200,000 images are generated with different target sizes, poses, lighting conditions, and backgrounds. Then, the data are annotated to ensure the dataset supports simultaneous space target detection, recognition and component segmentation. NCSTP has 10 fine-grained classes of satellites, 6 classes of space debris, and 4 classes of space rocks. All the data can be used for training space target detection and recognition models. We further annotate the body, solar panels, antennas, and observation payloads of each satellite for component segmentation. Finally, we test a series of state-of-the-art object detection and semantic segmentation models on the dataset to establish a benchmark.2025.6.16: A smaller version NCSTP-10000 is available now

  19. t

    PASCAL Visual Object Classes Challenge - Dataset - LDM

    • service.tib.eu
    Updated Dec 3, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). PASCAL Visual Object Classes Challenge - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/pascal-visual-object-classes-challenge
    Explore at:
    Dataset updated
    Dec 3, 2024
    Description

    The PASCAL Visual Object Classes Challenge (VOC) is a benchmark dataset for object detection and semantic segmentation.

  20. t

    Pigdetect: a diverse and challenging benchmark dataset for the detection of...

    • service.tib.eu
    Updated May 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Pigdetect: a diverse and challenging benchmark dataset for the detection of pigs in images - Vdataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/goe-doi-10-25625-i6uye9
    Explore at:
    Dataset updated
    May 16, 2025
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Note: To better find the files to download, select "Change View: Tree". The dataset contains: 2931 images from conventional pig farming with object detection annotations in yolo and coco format with predefined training, validation and test splits Trained model weights for pig detection A thorough explanation of all files contained in this data repository can be found in ReadMe.txt.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2024). FishEye8K: A Benchmark and Dataset for Fisheye Camera Object Detection - Dataset - 國網中心Dataset平台 [Dataset]. https://scidm.nchc.org.tw/dataset/fisheye8k

FishEye8K: A Benchmark and Dataset for Fisheye Camera Object Detection - Dataset - 國網中心Dataset平台

Explore at:
Dataset updated
Jul 12, 2024
License

Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically

Description

With the advance of AI, road object detection has been a prominent topic in computer vision, mostly using perspective cameras. Fisheye lens provides omnidirectional wide coverage for using fewer cameras to monitor road intersections, however with view distortions. To our knowledge, there is no existing open dataset prepared for traffic surveillance on fisheye cameras. This paper introduces an open FishEye8K benchmark dataset for road object detection tasks, which comprises 157K bounding boxes across five classes (Pedestrian, Bike, Car, Bus, and Truck). In addition, we present benchmark results of State-of-The-Art (SoTA) models, including variations of YOLOv5, YOLOR, YOLO7, and YOLOv8. The dataset comprises 8,000 images recorded in 22 videos using 18 fisheye cameras for traffic monitoring in Hsinchu, Taiwan, at resolutions of 1080x1080 and 1280x1280. The data annotation and validation process were arduous and time-consuming, due to the ultra-wide panoramic and hemispherical fisheye camera images with large distortion and numerous road participants, particularly people riding scooters. To avoid bias, frames from a particular camera were assigned to either the training or test sets, maintaining a ratio of about 70:30 for both the number of images and bounding boxes in each class. Experimental results show that YOLOv8 and YOLOR outperform on input sizes 640x640 and 1280x1280, respectively. copyright

Search
Clear search
Close search
Google apps
Main menu