21 datasets found
  1. Sartorius Segmentation - Detectron2 [Training] -v3

    • kaggle.com
    zip
    Updated Dec 8, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Y-Haneji (2021). Sartorius Segmentation - Detectron2 [Training] -v3 [Dataset]. https://www.kaggle.com/datasets/hanejiyuto/sartorius-segmentation-detectron2-training-v3
    Explore at:
    zip(2949182647 bytes)Available download formats
    Dataset updated
    Dec 8, 2021
    Authors
    Y-Haneji
    Description

    Dataset

    This dataset was created by Y-Haneji

    Contents

  2. Pre-processed (in Detectron2 and YOLO format) planetary images and boulder...

    • data.europa.eu
    • data-staging.niaid.nih.gov
    • +1more
    unknown
    Updated Jul 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zenodo (2025). Pre-processed (in Detectron2 and YOLO format) planetary images and boulder labels collected during the BOULDERING Marie Skłodowska-Curie Global fellowship [Dataset]. https://data.europa.eu/data/datasets/oai-zenodo-org-14250874?locale=no
    Explore at:
    unknown(601409488)Available download formats
    Dataset updated
    Jul 3, 2025
    Dataset authored and provided by
    Zenodohttp://zenodo.org/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This database contains 4976 planetary images of boulder fields located on Earth, Mars and Moon. The data was collected during the BOULDERING Marie Skłodowska-Curie Global fellowship between October 2021 and 2024. The data was already splitted into train, validation and test datasets, but feel free to re-organize the labels at your convenience. For each image, all of the boulder outlines within the image were carefully mapped in QGIS. More information about the labelling procedure can be found in the following manuscript (https://agupubs.onlinelibrary.wiley.com/doi/full/10.1029/2023JE008013). This dataset differs from the previous dataset included along with the manuscript https://zenodo.org/records/8171052, as it contains more mapped images, especially of boulder populations around young impact structures on the Moon (cold spots). In addition, the boulder outlines were also pre-processed so that it can be ingested directly in YOLOv8. A description of what is what is given in the README.txt file (in addition in how to load the custom datasets in Detectron2 and YOLO). Most of the other files are mostly self-explanatory. Please see previous dataset or manuscript for more information. If you want to have more information about specific lunar and martian planetary images, the IDs of the images are still available in the name of the file. Use this ID to find more information (e.g., M121118602_00875_image.png, ID M121118602 ca be used on https://pilot.wr.usgs.gov/). I will also upload the raw data from which this pre-processed dataset was generated (see https://zenodo.org/records/14250970). Thanks to this database, you can easily train a Detectron2 Mask R-CNN or YOLO instance segmentation models to automatically detect boulders. How to cite: Please refer to the "how to cite" section of the readme file of https://github.com/astroNils/YOLOv8-BeyondEarth. Structure: . └── boulder2024/ ├── jupyter-notebooks/ │ └── REGISTERING_BOULDER_DATASET_IN_DETECTRON2.ipynb ├── test/ │ └── images/ │ ├──

  3. Microcontroller Segmentation

    • kaggle.com
    zip
    Updated Jul 26, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gilbert Tanner (2020). Microcontroller Segmentation [Dataset]. https://www.kaggle.com/tannergi/microcontroller-segmentation
    Explore at:
    zip(17530203 bytes)Available download formats
    Dataset updated
    Jul 26, 2020
    Authors
    Gilbert Tanner
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description
  4. H

    Replication Data for: Training Deep Convolutional Object Detectors for...

    • dataverse.harvard.edu
    • search.dataone.org
    Updated Apr 16, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tomasz Gandor (2022). Replication Data for: Training Deep Convolutional Object Detectors for Images Affected by Lossy Compression [Dataset]. http://doi.org/10.7910/DVN/UHEP3C
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 16, 2022
    Dataset provided by
    Harvard Dataverse
    Authors
    Tomasz Gandor
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    This collection contains the trained models and object detection results of 2 architectures found in the Detectron2 library, on the MS COCO val2017 dataset, under different JPEG compresion level Q = {5, 12, 19, 26, 33, 40, 47, 54, 61, 68, 75, 82, 89, 96} (14 levels per trained model). Architectures: F50 – Faster R-CNN on ResNet-50 with FPN R50 – RetinaNet on ResNet-50 with FPN Training type: D2 – Detectron2 Model ZOO pre-trained 1x model (90.000 iterations, batch 16) STD – standard 1x training (90.000 iterations) on original train2017 dataset Q20 – 1x training (90.000 iterations) on train2017 dataset degraded to Q=20 Q40 – 1x training (90.000 iterations) on train2017 dataset degraded to Q=40 T20 – extra 1x training on top of D2 on train2017 dataset degraded to Q=20 T40 – extra 1x training on top of D2 on train2017 dataset degraded to Q=40 Model and metrics files models_FasterRCNN.tar.gz (F50-STD, F50-Q20, …) models_RetinaNet.tar.gz (R50-STD, R50-Q20, …) For every model there are 3 files: config.yaml – the Detectron2 config of the model. model_final.pth – the weights (training snapshot) in PyTorch format. metrics.json – training metrics (like time, total loss, etc.) every 20 iterations. The D2 models were not included, because they are available from the Detectron2 Model ZOO, as faster_rcnn_R_50_FPN_1x (F50-D2) and retinanet_R_50_FPN_1x (R50-D2). Result files F50-results.tar.gz – results for Faster R-CNN models (inluding D2). R50-results.tar.gz – results for RetinaNet models (inluding D2). For every model there are 14 subdirectories, e.g. evaluator_dump_R50x1_005 through evaluator_dump_R50x1_096, for each of the JPEG Q values. Each such folder contains: coco_instances_results.json – all detected objects (image id, bounding box, class index and confidence). results.json – AP metrics as computed by COCO API. Source code for processing the data The data can be processed using our code, published at: https://github.com/tgandor/urban_oculus. Additional dependencies for the source code: COCO API Detectron2

  5. Microcontroller Detection

    • kaggle.com
    zip
    Updated Nov 25, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gilbert Tanner (2019). Microcontroller Detection [Dataset]. https://www.kaggle.com/tannergi/microcontroller-detection
    Explore at:
    zip(8745225 bytes)Available download formats
    Dataset updated
    Nov 25, 2019
    Authors
    Gilbert Tanner
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Context

    As a electronics and computer science student I often work with microcontroller and microcomputers. That's why when I looked for objects to build my own object detection dataset they instantly came to mind.

    If you want to get started using the data-set feel free to check out my blog posts showing you how to train a model on the data-set with the Tensorflow Object Detection API or Detectron2.

  6. American Sign Language Poly Dataset

    • universe.roboflow.com
    zip
    Updated Apr 15, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Team Roboflow (2022). American Sign Language Poly Dataset [Dataset]. https://universe.roboflow.com/team-roboflow/american-sign-language-poly/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 15, 2022
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Team Roboflow
    Area covered
    United States
    Variables measured
    Signs Bounding Boxes
    Description

    This dataset includes all letters A through Z in American Sign Language labeled with polygon labels. See this blog post for how to train with Detectron2: https://blog.roboflow.com/p/4482cb2b-f378-48f6-bd58-df2b784670cf/

  7. R

    Data from: Leaf Disease Detection System Dataset

    • universe.roboflow.com
    zip
    Updated Apr 19, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Plant Disease Detection (2025). Leaf Disease Detection System Dataset [Dataset]. https://universe.roboflow.com/plant-disease-detection-m91t5/leaf-disease-detection-system
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 19, 2025
    Dataset authored and provided by
    Plant Disease Detection
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Leaf Disease Bounding Boxes
    Description

    This project aims to train Custam Plant disease dataset on Faster RCNN using Detectron2

  8. Sarnet Search And Rescue Dataset

    • universe.roboflow.com
    zip
    Updated Jun 16, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow Public (2022). Sarnet Search And Rescue Dataset [Dataset]. https://universe.roboflow.com/roboflow-public/sarnet-search-and-rescue/dataset/5
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 16, 2022
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Roboflow Public
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    SaR Bounding Boxes
    Description

    Description from the SaRNet: A Dataset for Deep Learning Assisted Search and Rescue with Satellite Imagery GitHub Repository * The "Note" was added by the Roboflow team.

    Satellite Imagery for Search And Rescue Dataset - ArXiv

    This is a single class dataset consisting of tiles of satellite imagery labeled with potential 'targets'. Labelers were instructed to draw boxes around anything they suspect may a paraglider wing, missing in a remote area of Nevada. Volunteers were shown examples of similar objects already in the environment for comparison. The missing wing, as it was found after 3 weeks, is shown below.

    https://michaeltpublic.s3.amazonaws.com/images/anomaly_small.jpg" alt="anomaly">

    The dataset contains the following:

    SetImagesAnnotations
    Train18083048
    Validate490747
    Test254411
    Total25524206

    The data is in the COCO format, and is directly compatible with faster r-cnn as implemented in Facebook's Detectron2.

    Getting hold of the Data

    Download the data here: sarnet.zip

    Or follow these steps

    # download the dataset
    wget https://michaeltpublic.s3.amazonaws.com/sarnet.zip
    
    # extract the files
    unzip sarnet.zip
    

    ***Note* with Roboflow, you can download the data here** (original, raw images, with annotations): https://universe.roboflow.com/roboflow-public/sarnet-search-and-rescue/ (download v1, original_raw-images) * Download the dataset in COCO JSON format, or another format of choice, and import them to Roboflow after unzipping the folder to get started on your project.

    Getting started

    Get started with a Faster R-CNN model pretrained on SaRNet: SaRNet_Demo.ipynb

    Source Code for Paper

    Source code for the paper is located here: SaRNet_train_test.ipynb

    Cite this dataset

    @misc{thoreau2021sarnet,
       title={SaRNet: A Dataset for Deep Learning Assisted Search and Rescue with Satellite Imagery}, 
       author={Michael Thoreau and Frazer Wilson},
       year={2021},
       eprint={2107.12469},
       archivePrefix={arXiv},
       primaryClass={eess.IV}
    }
    

    Acknowledgment

    The source data was generously provided by Planet Labs, Airbus Defence and Space, and Maxar Technologies.

  9. Modularized_own_code_Det2

    • kaggle.com
    zip
    Updated Jul 29, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Akarsh Rastogi (2021). Modularized_own_code_Det2 [Dataset]. https://www.kaggle.com/akarshrastogi/modularized-own-code-det2
    Explore at:
    zip(1615357 bytes)Available download formats
    Dataset updated
    Jul 29, 2021
    Authors
    Akarsh Rastogi
    Description

    A mostly up to date mirror of 10.75.129.40/DataInsights/GE-medicalimaging-train.git (Can be ahead since using it for testing)

    Only for Genpact DS Team.

    Detectron2 modularized codebase for training+prediction+submission+visualization on Kaggle's

    vinbigdata Chest X-ray competition

  10. Input and output data (images + boulder labels, model setup, model weights...

    • data.europa.eu
    unknown
    Updated Jul 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zenodo (2023). Input and output data (images + boulder labels, model setup, model weights and more) for the manuscript "Automatic characterization of boulders on planetary surfaces from high-resolution satellite images" [Dataset]. https://data.europa.eu/data/datasets/oai-zenodo-org-8171052?locale=no
    Explore at:
    unknown(265067013)Available download formats
    Dataset updated
    Jul 22, 2023
    Dataset authored and provided by
    Zenodohttp://zenodo.org/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    File 1: raw_data_BOULDERING.zip Size: 8.8 GB Summary: It contains all of the rasters (planetary images) and labeled boulders (raw data): a boulder-mapping file, which is the manually digitized outline of boulders. a ROM file (stands for Region of Mapping), which depicts the image patches on which the boulder mapping has been conducted. a global-tiles file, which shows all of the image patches within a raster. There are multiple locations/images per planetary body. Structure: . └── raw_data/ ├── earth/ │ └── image_name/ │ ├── shp/ │ │ ├──

  11. PeopleSansPeople (PeopleSansPeople: A Synthetic Data Generator for...

    • opendatalab.com
    zip
    Updated Dec 20, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unity Technologies (2021). PeopleSansPeople (PeopleSansPeople: A Synthetic Data Generator for Human-Centric Computer Vision) [Dataset]. https://opendatalab.com/OpenDataLab/PeopleSansPeople
    Explore at:
    zip(1547423033 bytes)Available download formats
    Dataset updated
    Dec 20, 2021
    Dataset provided by
    Unity Technologieshttps://unity.com/
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    We release a human-centric synthetic data generator PeopleSansPeople which contains simulation-ready 3D human assets, a parameterized lighting and camera system, and generates 2D and 3D bounding box, instance and semantic segmentation, and COCO pose labels. Using PeopleSansPeople, we performed benchmark synthetic data training using a Detectron2 Keypoint R-CNN variant [1]. We found that pre-training a network using synthetic data and fine-tuning on target real-world data (few-shot transfer to limited subsets of COCO-person train [2]) resulted in a keypoint AP of 60.37±0.48 (COCO test-dev2017) outperforming models trained with the same real data alone (keypoint AP of 55.80) and pre-trained with ImageNet (keypoint AP of 57.50). This freely-available data generator should enable a wide range of research into the emerging field of simulation to real transfer learning in the critical area of human-centric computer vision.

  12. HuBMap COCO Dataset 512x512 Tiled

    • kaggle.com
    zip
    Updated Nov 20, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sreevishnu Damodaran (2020). HuBMap COCO Dataset 512x512 Tiled [Dataset]. https://www.kaggle.com/datasets/sreevishnudamodaran/hubmap-coco-dataset-512x512-tiled
    Explore at:
    zip(739767398 bytes)Available download formats
    Dataset updated
    Nov 20, 2020
    Authors
    Sreevishnu Damodaran
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    This Dataset contains HuBMap Dataset in COCO format to use in any Object Detection and Instance Segmentation Task.

    COCO format easily supports Segmentation Frameworks such as AdelaiDet, Detectron2, TensorFlow etc.

    The dataset is structured with images split into directories and no downscaling was done.

    The following notebook explains how to convert custom annotations to COCO format:

    https://www.kaggle.com/sreevishnudamodaran/build-custom-coco-annotations-512x512-tiled

    Thanks to the Kaggle community and staff for all the support!

    Please don't miss to upvote and comment if you like my work :)

    Hope I everyone finds this useful!

    Directory Structure:

       - coco_train
         - images(contains images in jpg format)
           - original_tiff_image_name
             - tile_column_number
               - image
               .
               .
               .
              .
              .
              .
            .
            .
            .
         - train.json (contains all the segmentation annotations in coco 
         -       format with proper relative path of the images)
    
  13. EfficientDet Pytorch

    • kaggle.com
    zip
    Updated Apr 15, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mathurin Aché (2020). EfficientDet Pytorch [Dataset]. https://www.kaggle.com/mathurinache/efficientdet
    Explore at:
    zip(683867967 bytes)Available download formats
    Dataset updated
    Apr 15, 2020
    Authors
    Mathurin Aché
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    EfficientDet (PyTorch) This is a work in progress PyTorch implementation of EfficientDet.

    It is based on the

    official Tensorflow implementation by Mingxing Tan and the Google Brain team paper by Mingxing Tan, Ruoming Pang, Quoc V. Le EfficientDet: Scalable and Efficient Object Detection I am aware there are other PyTorch implementations. Their approach didn't fit well with my aim to replicate the Tensorflow models closely enough to allow weight ports while still maintaining a PyTorch feel and a high degree of flexibility for future additions. So, this is built from scratch and leverages my previous EfficientNet work.

    Updates / Tasks 2020-4-15 Taking a pause on training, some high priority things came up. There are signs of life on the training branch, was working the basic augs before priority switch, loss fn appeared to be doing something sane with distributed training working, no proper eval yet, init not correct yet. I will get to it, with SOTA training config and good performance as the end goal (as with my EfficientNet work).

    2020-04-11 Cleanup post-processing. Less code and a five-fold throughput increase on the smaller models. D0 running > 130 img/s on a single 2080Ti, D1 > 130 img/s on dual 2080Ti up to D7 @ 8.5 img/s.

    2020-04-10 Replace generate_detections with PyTorch impl using torchvision batched_nms. Significant performance increase with minor (+/-.001 mAP) score differences. Quite a bit faster than original TF impl on a GPU now.

    2020-04-09 Initial code with working validation posted. Yes, it's a little slow, but I think faster than the official impl on a GPU if you leave AMP enabled. Post processing needs some love.

    Core Tasks Feature extraction from my EfficientNet implementations (https://github.com/rwightman/gen-efficientnet-pytorch or https://github.com/rwightman/pytorch-image-models) Low level blocks / helpers (SeparableConv, create_pool2d (same padding), etc) PyTorch implementation of BiFPN, BoxNet, ClassNet modules and related submodules Port Tensorflow checkpoints to PyTorch -- initial D1 checkpoint converted, state_dict loaded, on to validation.... Basic MS COCO validation script Temporary (hacky) COCO dataset and transform Port reference TF anchor and object detection code Verify model output sanity Integrate MSCOCO eval metric calcs Some cleanup, testing Submit to test-dev server, all good Add torch hub support and pretrained URL based weight download Change module dependencies from 'timm' to minimal 'geffnet' for backbone, bring some of the layers here leaving as timm for now, as the training code will use many timm functions that I leverage to reproduce SOTA EfficientNet training in PyTorch Remove redundant bias layers that exist in the official impl and weights Add visualization support Performance improvements, numpy TF detection code -> optimized PyTorch Verify/fix Torchscript and ONNX export compatibility Possible Future Tasks Training (object detection) reimplementation w/ Rand/AutoAugment, etc Training (semantic segmentation) experiments Integration with Detectron2 / MMDetection codebases Addition and cleanup of EfficientNet based U-Net and DeepLab segmentation models that I've used in past projects Addition and cleanup of OpenImages dataset/training support from a past project Exploration of instance segmentation possibilities... If you are an organization is interested in sponsoring and any of this work, or prioritization of the possible future directions interests you, feel free to contact me (issue, LinkedIn, Twitter, hello at rwightman dot com). I will setup a github sponser if there is any interest.

    Models Variant Download mAP (val2017) mAP (test-dev2017) mAP (Tensorflow official test-dev2017) D0 tf_efficientdet_d0.pth 32.8 TBD 33.8 D1 tf_efficientdet_d1.pth 38.5 TBD 39.6 D2 tf_efficientdet_d2.pth 42.0 42.5 43 D3 tf_efficientdet_d3.pth 45.3 TBD 45.8 D4 tf_efficientdet_d4.pth 48.3 TBD 49.4 D5 tf_efficientdet_d5.pth 49.6 TBD 50.7 D6 tf_efficientdet_d6.pth 50.6 TBD 51.7 D7 tf_efficientdet_d7.pth 50.9 51.2 52.2 Usage Environment Setup Tested in a Python 3.7 or 3.8 conda environment in Linux with:

    PyTorch 1.4 PyTorch Image Models (timm) 0.1.20, pip install timm or local install from (https://github.com/rwightman/pytorch-image-models) Apex AMP master (as of 2020-04) NOTE - There is a conflict/bug with Numpy 1.18+ and pycocotools, force install numpy <= 1.17.5 or the coco eval will fail, the validation script will still save the output JSON and that can be run through eval again later.

    Dataset Setup MSCOCO 2017 validation data:

    wget http://images.cocodataset.org/zips/val2017.zip wget http://images.cocodataset.org/annotations/annotations_trainval2017.zip unzip val2017.zip unzip annotations_trainval2017.zip MSCOCO 2017 test-dev data:

    wget http://images.cocodataset.org/zips/test2017.zip unzip -q test2017.zip wget http://images.cocodat...

  14. R

    Uftir Particles Detection Dataset

    • universe.roboflow.com
    zip
    Updated Jul 19, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    uFTIR Particles (2023). Uftir Particles Detection Dataset [Dataset]. https://universe.roboflow.com/uftir-particles/uftir-particles-detection/model/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 19, 2023
    Dataset authored and provided by
    uFTIR Particles
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Particle Polygons
    Description

    In the context of this project, the samples for µ-FTIR analysis contained up to a few thousands particles. The integrated particle detection tool (Particle Wizard - OMNIC Picta) gave poor performances and an AI segmentation tool was needed. Using this dataset, we trained a Detectron2 neural network that was used within GEPARD, an open source software used to improve Raman and FTIR target acquisition and data analysis. With Roboflow, it is possible to export this dataset to various format and use these data to train different architecture of segmentation neural networks.

  15. Z

    COCO dataset and neural network weights for micro-FTIR particle detection on...

    • data.niaid.nih.gov
    Updated Aug 13, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Schowing, Thibault (2024). COCO dataset and neural network weights for micro-FTIR particle detection on filters. [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10839526
    Explore at:
    Dataset updated
    Aug 13, 2024
    Dataset provided by
    HES-SO Vaud
    Authors
    Schowing, Thibault
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The IMPTOX project has received funding from the EU's H2020 framework programme for research and innovation under grant agreement n. 965173. Imptox is part of the European MNP cluster on human health.

    More information about the project here.

    Description: This repository includes the trained weights and a custom COCO-formatted dataset used for developing and testing a Faster R-CNN R_50_FPN_3x object detector, specifically designed to identify particles in micro-FTIR filter images.

    Contents:

    Weights File (neuralNetWeights_V3.pth):

    Format: .pth

    Description: This file contains the trained weights for a Faster R-CNN model with a ResNet-50 backbone and a Feature Pyramid Network (FPN), trained for 3x schedule. These weights are specifically tuned for detecting particles in micro-FTIR filter images.

    Custom COCO Dataset (uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip):

    Format: .zip

    Description: This zip archive contains a custom COCO-formatted dataset, including JPEG images and their corresponding annotation file. The dataset consists of images of micro-FTIR filters with annotated particles.

    Contents:

    Images: JPEG format images of micro-FTIR filters.

    Annotations: A JSON file in COCO format providing detailed annotations of the particles in the images.

    Management: The dataset can be managed and manipulated using the Pycocotools library, facilitating easy integration with existing COCO tools and workflows.

    Applications: The provided weights and dataset are intended for researchers and practitioners in the field of microscopy and particle detection. The dataset and model can be used for further training, validation, and fine-tuning of object detection models in similar domains.

    Usage Notes:

    The neuralNetWeights_V3.pth file should be loaded into a PyTorch model compatible with the Faster R-CNN architecture, such as Detectron2.

    The contents of uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip should be extracted and can be used with any COCO-compatible object detection framework for training and evaluation purposes.

    Code can be found on the related Github repository.

  16. Paper Grading Computer Vision Models

    • kaggle.com
    Updated Apr 10, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kerem NAYMAN (2022). Paper Grading Computer Vision Models [Dataset]. https://www.kaggle.com/datasets/keremnayman/paper-grading-computer-vision-models
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 10, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Kerem NAYMAN
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This folder has AI models that are required to run the Paper Grading project.

    In total, more than 100.000 visual data was used to train these models. Most of the data were created by me. r50.pth is a Detectron2 Faster R-CNN model, others are ResNet models trained using Fast.Ai API.

    Model explanations: - abcde.pkl: ResNet Image Classification model with 18 layers. It can predict which letter was marked by the student. It has five different output classes: A-B-C-D-E on 224x224 input images. - config.yaml: Configuration file for Faster R-CNN model. For more details please check Detectron2 documentation. - dogru_yanlis.pkl: ResNet Image Classification model with 18 layers, which can classify if there is a tick (✓) or cross (X) on 48x48 input images. - eslestirme.pkl: ResNet Image Classification model with 34 layers, which can predict the letter (A-H) written by the students on 48x48 images. - gecerli_gecersiz.pkl: ResNet Image Classification model with 18 layers which can predict if the multiple-choice question marked by the student is valid or not (empty, marked more than one letter, etc.) - r50.pth: Faster R-CNN model with 101 layers. Its aim is to detect questions on photos or scans of real exam papers. It can detect: 1. True-False 2. Matching 3. Gap Filling 4. Multiple Choice question types as well as Name-Surname box. Please check the image below: https://tasarimciogretmen-storage.s3.eu-south-1.amazonaws.com/Resim1.png" alt="">

  17. IP102 COCO Format Annotations for Object Detection

    • kaggle.com
    zip
    Updated Oct 26, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    eljazouly (2025). IP102 COCO Format Annotations for Object Detection [Dataset]. https://www.kaggle.com/datasets/eljazouly/ip102-coco-annotations/discussion
    Explore at:
    zip(1897481 bytes)Available download formats
    Dataset updated
    Oct 26, 2025
    Authors
    eljazouly
    Description

    IP102 COCO Format Annotations

    This dataset contains preprocessed annotations for the IP102 Insect Pest Recognition Dataset converted to COCO format, making it ready for object detection models like DETR, Faster R-CNN, YOLO, and other modern detectors.

    About IP102 Dataset

    IP102 is a large-scale benchmark dataset for insect pest recognition containing: - 75,222 images of insect pests - 102 categories of agricultural pests - Images collected from real agricultural scenarios

    What's Included

    This dataset provides: - train_annotations.json - Training set annotations in COCO format - val_annotations.json - Validation set annotations in COCO format
    - test_annotations.json (optional) - Test set annotations

    Format Specification

    Annotations follow the standard COCO Object Detection format: json { "images": [ { "id": 1, "file_name": "image_001.jpg", "width": 640, "height": 480 } ], "annotations": [ { "id": 1, "image_id": 1, "category_id": 5, "bbox": [x, y, width, height], "area": 12345, "iscrowd": 0 } ], "categories": [ { "id": 1, "name": "rice_leaf_roller", "supercategory": "insect" } ] }

    Usage Example

    import json
    from pycocotools.coco import COCO
    
    # Load annotations
    with open('/kaggle/input/ip102-coco-annotations/train_annotations.json') as f:
      coco_data = json.load(f)
    
    # Or use COCO API
    coco = COCO('/kaggle/input/ip102-coco-annotations/train_annotations.json')
    
    print(f"Number of images: {len(coco_data['images'])}")
    print(f"Number of annotations: {len(coco_data['annotations'])}")
    print(f"Number of categories: {len(coco_data['categories'])}")
    

    🔗 Compatible With

    • DETR (Detection Transformer)
    • Faster R-CNN
    • Mask R-CNN
    • RetinaNet
    • YOLOv5/v8 (with conversion)
    • Detectron2
    • ✅ Any framework supporting COCO format

    Dataset Statistics

    • Total Images: ~75,000
    • Classes: 102 insect pest categories
    • Format: COCO JSON
    • Task: Object Detection / Instance Segmentation

    Citation

    If you use this dataset, please cite the original IP102 paper: @article{wu2019ip102, title={IP102: A Large-Scale Benchmark Dataset for Insect Pest Recognition}, author={Wu, Xiaoping and Zhan, Chi and Lai, Yu-Kun and Cheng, Ming-Ming and Yang, Jufeng}, journal={CVPR}, year={2019} }

    📝 Notes

    • Annotations created from original IP102 bounding box labels
    • Validated for training modern object detection models
    • Compatible with PyTorch, TensorFlow, and other frameworks
    • Preprocessed to save computation time on Kaggle

    Updates

    • v1.0 (2025-01-XX): Initial release with train/val splits

    Acknowledgments

    Original dataset by Wu et al. (CVPR 2019). This is a format conversion for easier integration with modern detection frameworks.

    Ready to train your insect detection model! 🐛🔍 ```

    Tags (choisissez 5-10) :

    object detection
    computer vision
    agriculture
    coco format
    insect recognition
    pest detection
    deep learning
    detr
    dataset
    annotation
    

    License:

    CC BY-NC-SA 4.0 (same as original IP102)
    

    ou ``` Database: Open Database, Contents: © Original Authors

  18. potholes, cracks and openmanholes (Road Hazards)

    • kaggle.com
    zip
    Updated Feb 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sabid Rahman (2025). potholes, cracks and openmanholes (Road Hazards) [Dataset]. https://www.kaggle.com/datasets/sabidrahman/pothole-cracks-and-openmanhole
    Explore at:
    zip(1078256309 bytes)Available download formats
    Dataset updated
    Feb 23, 2025
    Authors
    Sabid Rahman
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F23345571%2F4471e4ade50676d782d4787f77aa08ad%2F1000_F_256252609_6WIHRGbpzSaVQwioubxwgXdSJTNONNcK.jpg?generation=1739209341333909&alt=media" alt="">

    This dataset contains 2,700 augmented images, organized into training and validation folders, and focuses on detecting potholes, cracks, and open manholes on roads. To improve the robustness and generalization capability of detection models, the dataset has been augmented using various techniques that enhance data diversity. Annotations are available for all three categories, making the dataset fully compatible with both YOLO and Faster R-CNN architectures. Specifically, it includes YOLO format (.txt) files for use with YOLOv5, YOLOv7, and YOLOv8, as well as COCO JSON annotations suitable for Faster R-CNN, Detectron2, and MMDetection frameworks. Additionally, the dataset directory contains separate subfolders for each class—potholes, cracks, and open manholes—along with their respective annotation files, which facilitates easier access and class-wise analysis. Overall, this dataset is ready for direct use in modern object detection pipelines.

    • Potholes: class 0
    • Cracks: class 1
    • Open Manholes: class 2

    Included in the Dataset: - Bounding Box Annotations in YOLO Format (.txt files) - Format: YOLOv8 & YOLO11 compatible - Purpose: Ready for training YOLO-based object detection models

    • Folder Structure Organized into:

      • train/ folder
      • valid/ folder
      • Class-specific folders
      • An all_classes/ folder for combined access Benefit: Easy access for training, validation, and augmentation tasks
    • Dual Format Support

      • COCO JSON Annotations Included -Compatible with models like Faster R-CNN Enables flexibility across different object detection frameworks
    • Use Cases Targeted

      • Model training
      • Model testing
      • Custom data augmentation
      • Specific focus: Road safety and infrastructure detection

    Here's a clear breakdown of the folder structure:

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F23345571%2F023b40c98bf858c58394d6ed2393bfc3%2FScreenshot%202025-05-01%20202438.png?generation=1746109541780835&alt=media" alt="">

  19. FloorPlan

    • kaggle.com
    zip
    Updated Jun 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mazhar Rehan (2025). FloorPlan [Dataset]. https://www.kaggle.com/datasets/mazharrehan/floorplan
    Explore at:
    zip(6274969 bytes)Available download formats
    Dataset updated
    Jun 16, 2025
    Authors
    Mazhar Rehan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    FloorPlan Dataset

    This dataset contains color-coded architectural floorplans for residential plots of 5 Marla, 10 Marla, and 20 Marla, with a focus on Ground Floor (GF) plans. Each image is annotated using a consistent color scheme to indicate different room types and structural elements — making it ideal for applications in AI training, architectural analysis, and layout understanding.

    📁 Dataset Structure

    FloorPlan_Metadata/ ├── data_JSON/ # Individual JSON metadata files per image └── floor_plan_metadata_consolidated.csv # Combined metadata for all images

    Metadata_Generation/ # Scripts + logs used to generate metadata ├── 0_generate_metadata.ipynb ├── 1.0_generate_metadata.py ├── 1.1_generate_metadata.py └── Workflow.txt

    dataset/ # 308 floorplan images in .png format ├── 10Marla_GF_FP_001_V01.png ├── 10Marla_GF_FP_002_V01.png └── ...

    🖼️ Floorplan Images

    • Format: .png
    • Total: 308
    • Sizes Covered: 5 Marla, 10 Marla, 20 Marla
    • Level: Ground Floor only
    • Naming Convention: PlotSize_Floor_FP_Index_Version.png
      • Example: 10Marla_GF_FP_002_V01.png

    Each image is manually verified to maintain architectural accuracy and uniform labeling standards.

    📊 Metadata Details

    1. floor_plan_metadata_consolidated.csv

    A CSV file containing all metadata entries for the floorplans in structured format.

    Key fields: - file_name - plot_size (5, 10, or 20 Marla) - floor_level (GF) - num_bedrooms, num_bathrooms, num_kitchens, etc. - orientation, total_area, construction_year (if available)

    2. data_JSON/

    Each .json file contains metadata for a specific image — ideal for structured processing and integration into pipelines.

    ⚙️ Metadata Generation Scripts

    Scripts used to generate the metadata from floorplan images are included in the Metadata_Generation/ folder:

    • 0_generate_metadata.ipynb: Jupyter notebook version for visual walkthrough
    • 1.0_generate_metadata.py, 1.1_generate_metadata.py: Python scripts for automated generation
    • Workflow.txt: Execution guide for reproducibility

    These tools allow you to modify or extend metadata extraction logic for your own projects.

    🚀 Use Cases

    • AI models for architectural layout parsing
    • Floorplan segmentation and annotation
    • Generative design systems
    • Dataset for training scene understanding models (e.g., Detectron2, Mask R-CNN)
    • Real estate visual analytics

    📜 License

    License: CC BY 4.0 (Attribution)

    You are free to use, share, and adapt the data with proper credit to the dataset creator.

    🙌 Acknowledgment

    This dataset was compiled and shared to aid the research and development of intelligent systems in architecture, construction, and real estate technology. If you find it useful, consider citing or linking this Kaggle page.

    📫 Contact

    For suggestions, feedback, or collaboration opportunities, feel free to reach out:

  20. Sorghum Crop Line Detection Dataset

    • kaggle.com
    zip
    Updated Feb 28, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gabriel Fernandes Carvalho (2024). Sorghum Crop Line Detection Dataset [Dataset]. https://www.kaggle.com/datasets/gabrielfcarvalho/sorghum-crop-line-detection-dataset/code
    Explore at:
    zip(244194618 bytes)Available download formats
    Dataset updated
    Feb 28, 2024
    Authors
    Gabriel Fernandes Carvalho
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    UAV-Captured Sorghum Crop Line Detection Dataset

    Description

    This dataset contains UAV-captured images of sorghum fields, annotated for crop line detection. It has been curated to facilitate machine learning research, particularly for developing and evaluating algorithms for agricultural monitoring and analysis.

    The dataset has been divided into six separate folders, each formatted for compatibility with different object detection architectures:

    • 416x416_augmented: Prepared for use with Detectron2 architectures, such as RetinaNet and Faster R-CNN, with images augmented and resized to 416x416 pixels.
    • sorghumfield.v3-416x416_augmented.mt-yolov6: Contains images augmented and tailored for use with the YOLOv6 Meituan architecture.
    • sorghumfield.v3-416x416_augmented.yolov5pytorch: Formatted specifically for the YOLOv5 architecture implemented in PyTorch.
    • sorghumfield.v3-416x416_augmented.yolov8: Adapted for the latest YOLOv8 architecture, with the same augmentation and resizing.
    • sorghumfield.v3i.darknet: Designed for use with YOLOv3, YOLOv4 and YOLOv7 architectures within the Darknet framework.
    • sorghumfield.v9i.yolov8_synthetic: An updated set that incorporates synthetic images generated to augment the YOLOv8 dataset.

    Each folder contains images that have been manually annotated with bounding boxes to identify crop lines. Annotations were performed using LabelBox, and the data has been segregated into training, validation, and testing sets.

    Data Augmentation and Synthetic Data

    Data augmentation techniques such as rotations, translations, scaling, and flipping have been applied to increase the diversity and robustness of the dataset. Additionally, synthetic data has been generated and included to enhance the dataset further, providing additional variability and complexity for more effective training of object detection models.

    Intended Use

    This dataset is intended for use by researchers and practitioners in the fields of computer vision and agriculture technology. It is particularly useful for those developing object detection models for agricultural applications.

    Acknowledgments

    When utilizing this dataset, please reference the original source of the sorghum images made available by Purdue University and the manual annotations provided in this work.

    Citation

    If you use this dataset in your research, please cite the following: - Fernandes, G., & Pedro, J. (2023). "Aplicabilidade de Técnicas de Inteligência Artificial na Análise Automática de Imagens Agrícolas Aéreas". Undergraduate Thesis, UnB. - J. Ribera, F. He, Y. Chen, A. F. Habib, and E. J. Delp, "Estimating Phenotypic Traits From UAV Based RGB Imagery", ACM SIGKDD Conference on Knowledge Discovery and Data Mining Workshop - August 2016, San Francisco, CA - J. Ribera, D. Güera, E. J. Delp, "Locating Objects Without Bounding Boxes", Computer Vision and Pattern Recognition (CVPR), June 2019, Long Beach, CA. arXiv:1806.07564.

    License

    The dataset is available for non-commercial research and educational purposes. For any other use, please contact the authors for permission.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Y-Haneji (2021). Sartorius Segmentation - Detectron2 [Training] -v3 [Dataset]. https://www.kaggle.com/datasets/hanejiyuto/sartorius-segmentation-detectron2-training-v3
Organization logo

Sartorius Segmentation - Detectron2 [Training] -v3

Explore at:
zip(2949182647 bytes)Available download formats
Dataset updated
Dec 8, 2021
Authors
Y-Haneji
Description

Dataset

This dataset was created by Y-Haneji

Contents

Search
Clear search
Close search
Google apps
Main menu