9 datasets found
  1. h

    COCO8-pose

    • huggingface.co
    Updated Dec 18, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ultralytics (2024). COCO8-pose [Dataset]. https://huggingface.co/datasets/Ultralytics/COCO8-pose
    Explore at:
    Dataset updated
    Dec 18, 2024
    Dataset authored and provided by
    Ultralytics
    License

    https://choosealicense.com/licenses/agpl-3.0/https://choosealicense.com/licenses/agpl-3.0/

    Description

    Ultralytics COCO8-pose Dataset

    Ultralytics COCO8-pose is a small, but versatile pose detection dataset composed of the first 8 images of the COCO train 2017 set, 4 for training and 4 for validation. This dataset is ideal for testing and debugging object detection models, or for experimenting with new detection approaches. With 8 images, it is small enough to be easily manageable, yet diverse enough to test training pipelines for errors and act as a sanity check before training… See the full description on the dataset page: https://huggingface.co/datasets/Ultralytics/COCO8-pose.

  2. Rooftop Drainage Outlets and Ventilations Dataset

    • zenodo.org
    zip
    Updated Apr 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lukas Arzoumanidis*; Lukas Arzoumanidis*; Julius Knechtel*; Julius Knechtel*; Gizem Sen*; Weilian Li; Weilian Li; Youness Dehbi; Youness Dehbi; Gizem Sen* (2025). Rooftop Drainage Outlets and Ventilations Dataset [Dataset]. http://doi.org/10.5281/zenodo.14040571
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 1, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Lukas Arzoumanidis*; Lukas Arzoumanidis*; Julius Knechtel*; Julius Knechtel*; Gizem Sen*; Weilian Li; Weilian Li; Youness Dehbi; Youness Dehbi; Gizem Sen*
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Authors marked with an asterisk (*) have contributed equally to this publication.

    We annotated a dataset for the detection of drainage outlets and ventilations on flat rooftops. The underlying high-resolution aerial images are orthophotos with a ground sampling distance of 7.5 cm, provided by the Office for Land Management and Geoinformation of the City of Bonn, Germany. The dataset was created through manual annotation using the Computer Vision Annotation Tool (CVAT) and comprises 740 image pairs. Each pair consists of a rooftop image and a corresponding annotated mask indicating the drainage outlets and ventilations. Since rooftops vary in size, we aimed to create image pairs that capture a single rooftop per image without overlaps or cutoffs. Consequently, the dimensions of each image pair differ. The dataset is split randomly into 80% for training, 10% for validation, and 10% for testing.

    We provide the dataset in the Common Objects in Context (COCO) format for object detection tasks. In addition to the COCO-formatted dataset, we provide the dataset in its original, pairwise, format to support various machine learning tasks, such as semantic segmentation and panoptic segmentation, as well as to accommodate different data-loading requirements for diverse deep learning models.

    If your object detection approach requires the 'category_id' to start from 0 instead of 1, please refer to the following guide: https://github.com/obss/sahi/discussions/336
    For conversion to a completely different dataset format, such as YOLO, please see the repository: https://github.com/ultralytics/JSON2YOLO

  3. Underwater images from OBSEA fish detection training dataset (YOLO)

    • zenodo.org
    zip
    Updated Feb 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pol Baños Castelló; Pol Baños Castelló; Oriol Prat I Bayarri; Oriol Prat I Bayarri; Enoc Martínez Padró; Enoc Martínez Padró; Marco Francescangeli; Marco Francescangeli; Joaquin del Rio; Joaquin del Rio (2025). Underwater images from OBSEA fish detection training dataset (YOLO) [Dataset]. http://doi.org/10.5281/zenodo.14888440
    Explore at:
    zipAvailable download formats
    Dataset updated
    Feb 24, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Pol Baños Castelló; Pol Baños Castelló; Oriol Prat I Bayarri; Oriol Prat I Bayarri; Enoc Martínez Padró; Enoc Martínez Padró; Marco Francescangeli; Marco Francescangeli; Joaquin del Rio; Joaquin del Rio
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains labelled underwater pictures taken at the OBSEA underwater observatory. The pictures have already been splited into train, validation and test folders for training a YOLO AI model. This is a substet of this dataset, with the data arranged to simplify the training process.

    Train a YOLO model with this dataset:

    1. Install dependencies with the command "pip3 install ultralytics"
    2. Download and unzip the file
    3. Edit the data.yaml file to match the "train", "valid" and "test" paths in your filesystem
    4. Train the model with the command: "yolo task=detect mode=train model=yolov8n.pt data=data.yaml"

    Done! In order to optimize the results it is encouraged to tune the hyperparameters to fit your application.

    Training dataset

    This dataset contains labeled images for fish detection acquired at OBSEA Underwater Observatory (NW Mediterranean sea).

    Data preprocessing

    Several data augmentation techniques have been used to improve the training. The configuration can be found in args.yaml file.

    Data splitting

    Data has been randomly splitted in 70% training, 20% validation and 10% test. The splits are already included in the training dataset.

    Classes, labels and annotations

    The following classes are included in the dataset:

    • Chromis chromis: WoRMS id 127000
    • Coris julis: WoRMS id 126963
    • Dactylopterus volitans: WoRMS id None None
    • Dentex dentex: WoRMS id 273962
    • Diplodus cervinus: WoRMS id 127051
    • Diplodus puntazzo: WoRMS id 127052
    • Diplodus sargus: WoRMS id 127053
    • Diplodus vulgaris: WoRMS id 127054
    • Diver: scuba diver, used mainly to prevent divers to be detected as fish
    • Epinephelus costae: WoRMS id 127034
    • Epinephelus marginatus: WoRMS id 127036
    • Mullus surmuletus: WoRMS id 126986
    • Muraena helena: WoRMS id 126303
    • Myliobatidae: WoRMS id None None
    • Oblada melanura: WoRMS id None None
    • Octopus vulgaris: WoRMS id None None
    • Parablennius gattorugine: WoRMS id 126770
    • Sarpa salpa: WoRMS id 127064
    • Sciaena umbra: WoRMS id None None
    • Seriola dumerili: WoRMS id 126816
    • Serranus cabrilla: WoRMS id 127041
    • Sparus aurata: WoRMS id 151523
    • Symphodus mediterraneus: WoRMS id 273569

    Data sources

    Pictures where acquired by several underwater cameras, deployed at OBSEA, model Linovision IPC608.

    Data quality

    Images have been manually selected to include as much variety as possible in terms of light and water turbidity.

    Spatial coverage

    All pictures where taken at OBSEA underwater observatory, off-the-coast of Vilanova i la Geltrú, Spain. GPS coordinates

    Longitude Latitudedepth
    1.7525741.1821220 m

    Contact information

    For further technical inquiries or additional information about the annotated dataset, please contact enoc.martinez@upc.edu

  4. Bangla/Bengali License Plate with Number Detection

    • kaggle.com
    Updated Jul 21, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ashikur Rahman Shad (2025). Bangla/Bengali License Plate with Number Detection [Dataset]. https://www.kaggle.com/datasets/arstrxshad/banglabengali-license-plate-with-number-detection/code
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 21, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Ashikur Rahman Shad
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Bangla LPDB - A (YOLO Ready)

    This dataset is a reformatted and enhanced version of the Bangla LPDB - A dataset, originally published by Ataher Sams and Homaira Huda Shomee. It has been meticulously prepared to be plug-and-play for YOLO (You Only Look Once) object detection models, making it incredibly easy for researchers and developers to use for license plate detection tasks in Bangladeshi vehicles.

    Original Dataset Acknowledgment

    This dataset is built upon

    Ataher Sams, & Homaira Huda Shomee. (2021). Bangla LPDB - A (Version v1) [Data set]. International Conference on Digital Image Computing: Techniques and Applications (IEEE DICTA), Gold Coast, Queensland Australia. Zenodo. https://doi.org/10.5281/zenodo.4718238

    We extend our sincerest gratitude to them for creating such a comprehensive and vital resource for the research community.

    Key Enhancements

    While the original Bangla LPDB - A dataset is an excellent collection, this version provides significant improvements for immediate use with YOLO models:

    • YOLO Dataset Structure: The dataset has been meticulously restructured to follow the standard YOLO format, including all necessary annotation files.
    • dataset.yaml Included: A dataset.yaml file is provided for seamless integration with popular deep learning frameworks like Ultralytics YOLO.
    • Class Names: Separate files for both Bengali and English class names are included, offering flexibility for various applications.
    • Pre-split Data: The dataset is pre-split into optimal ratios for training, validation, and testing:
      • Training Set: 70%
      • Validation Set: 20%
      • Testing Set: 10%

    Preview

    Vehicle to License Plate https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F24265111%2F3976ff0d5a259dbd70dc017964bf7d47%2Fvehicle-to-license-plate.png?generation=1753064814234359&alt=media" alt="">

    License Plate to Text https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F24265111%2F8f5461031162387a7aba8ed7f31c6eda%2Flicense-plate-to-text.png?generation=1753064874309479&alt=media" alt="">

    The Original Paper

    H. H. Shomee and A. Sams, "License Plate Detection and Recognition System for All Types of Bangladeshi Vehicles Using Multi-step Deep Learning Model," 2021 Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 2021, pp. 01-07, https://doi.org/10.1109/DICTA52665.2021.9647284.

    Dataset Citation

    Users of this dataset are required to cite the original research paper, which introduces the Bangla LPDB - A dataset and its applications. Please use the following citation:

    @INPROCEEDINGS{9647284,
     author={Shomee, H. H. and Sams, A.},
     booktitle={2021 Digital Image Computing: Techniques and Applications (DICTA)},
     title={License Plate Detection and Recognition System for All Types of Bangladeshi Vehicles Using Multi-step Deep Learning Model},
     year={2021},
     pages={01-07},
     doi={10.1109/DICTA52665.2021.9647284}
    }
    

    Modified by

    Ashikur Rahman Shad

    ashikur-rahman-shad (Ashikur Rahman Shad) · GitHub

  5. Eagle Dataset (YOLOv11-OBB Format)

    • kaggle.com
    Updated Jul 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mridankan Mandal (2025). Eagle Dataset (YOLOv11-OBB Format) [Dataset]. https://www.kaggle.com/datasets/redzapdos123/eagle-dataset-yolov11-obb-format/discussion
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 29, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Mridankan Mandal
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Eagle Dataset - YOLOv11 OBB for Vehicle Detection

    High-resolution aerial imagery with 16,000+ oriented bounding boxes for vehicle detection, pre-formatted for Ultralytics YOLOv11.

    Context

    This dataset is a ready-to-use version of the original Eagle Dataset from the German Aerospace Center (DLR). The original dataset was created to benchmark object detection models on challenging aerial imagery, featuring vehicles at various orientations.

    This version has been converted to the YOLOv11-OBB (Oriented Bounding Box) format. The conversion makes the dataset directly compatible with modern deep learning frameworks like Ultralytics YOLO, allowing researchers and developers to train state-of-the-art object detectors with minimal setup.

    The dataset is ideal for tasks requiring precise localization of rotated objects, such as vehicle detection in parking lots, traffic monitoring, and urban planning from aerial viewpoints.

    Content

    The dataset is split into training, validation, and test sets, following a standard structure for computer vision tasks.

    Dataset Split & Counts:

    • Training Set: 159 images and labels
    • Validation Set: 53 images and labels
    • Test Set: 106 images and labels

    Directory Structure:

    EagleDatasetYOLO/
    ├── train/
    │  ├── images/   # 159 images
    │  └── labels/   # 159 .txt obb labels
    ├── val/
    │  ├── images/   # 53 images
    │  └── labels/   # 53 .txt obb labels
    ├── test/
    │  ├── images/   # 106 images
    │  └── labels/   # 106 .txt obb labels
    ├── data.yaml
    └── license.md
    

    Annotation Format (YOLOv11-OBB):

    Each .txt label file contains one object per line. The format for each object is: <class_id> <x_center> <y_center> <width> <height> <angle>

    • <class_id>: The class index (in this case, 0 for 'vehicle').
    • <x_center> <y_center>: The normalized center coordinates of the bounding box.
    • <width> <height>: The normalized width and height of the bounding box.
    • <angle>: The rotation angle of the box in radians, from -π/2 to π/2.

    data.yaml Configuration:

    A data.yaml file is included for easy integration with the Ultralytics framework.

    path: ../EagleDatasetYOLO
    train: train/images
    val: val/images
    test: test/images
    
    nc: 1
    names: ['vehicle']
    

    Acknowledgements and License

    This dataset is a conversion of the original work by the German Aerospace Center (DLR). The conversion to YOLOv11-OBB format was performed by Mridankan Mandal.

    The dataset is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International license (CC BY-NC-SA 4.0).

    If you use this dataset in your research, please cite the original creators and acknowledge the conversion work.

  6. t

    Crossroad Camera Dataset - Mobility Aid Users

    • repository.tugraz.at
    zip
    Updated May 13, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ludwig Mohr; Nadezda Kirillova; Horst Possegger; Horst Bischof; Ludwig Mohr; Nadezda Kirillova; Horst Possegger; Horst Bischof (2025). Crossroad Camera Dataset - Mobility Aid Users [Dataset]. http://doi.org/10.3217/2gat1-pev27
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 13, 2025
    Dataset provided by
    Graz University of Technology
    Authors
    Ludwig Mohr; Nadezda Kirillova; Horst Possegger; Horst Bischof; Ludwig Mohr; Nadezda Kirillova; Horst Possegger; Horst Bischof
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Time period covered
    Oct 2022
    Description

    The most vulnerable group of traffic participants are pedestrians using mobility aids. While there has been significant progress in the robustness and reliability of camera based general pedestrian detection systems, pedestrians reliant on mobility aids are highly underrepresented in common datasets for object detection and classification.

    To bridge this gap and enable research towards robust and reliable detection systems which may be employed in traffic monitoring, scheduling, and planning, we present this dataset of a pedestrian crossing scenario taken from an elevated traffic monitoring perspective together with ground truth annotations (Yolo format [1]). Classes present in the dataset are pedestrian (without mobility aids), as well as pedestrians using wheelchairs, rollators/wheeled walkers, crutches, and walking canes. The dataset comes with official training, validation, and test splits.

    An in-depth description of the dataset can be found in [2]. If you make use of this dataset in your work, research or publication, please cite this work as:

    @inproceedings{mohr2023mau,
    author = {Mohr, Ludwig and Kirillova, Nadezda and Possegger, Horst and Bischof, Horst},
    title = {{A Comprehensive Crossroad Camera Dataset of Mobility Aid Users}},
    booktitle = {Proceedings of the 34th British Machine Vision Conference ({BMVC}2023)},
    year = {2023}
    }

    Archive mobility.zip contains the full detection dataset in Yolo format with images, ground truth labels and meta data, archive mobility_class_hierarchy.zip contains labels and meta files (Yolo format) for training with class hierarchy using e.g. the modified version of Yolo v5/v8 available under [3].
    To use this dataset with Yolo, you will need to download and extract the zip archive and change the path entry in dataset.yaml to the directory where you extracted the archive to.

    [1] https://github.com/ultralytics/ultralytics
    [2] coming soon
    [3] coming soon

  7. Z

    Pre-processed (in Detectron2 and YOLO format) planetary images and boulder...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Nov 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lapotre, Mathieu (2024). Pre-processed (in Detectron2 and YOLO format) planetary images and boulder labels collected during the BOULDERING Marie Skłodowska-Curie Global fellowship [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_14250873
    Explore at:
    Dataset updated
    Nov 30, 2024
    Dataset provided by
    Prieur, Nils
    Amaro, Brian
    Gonzalez, Emiliano
    Lapotre, Mathieu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This database contains 4976 planetary images of boulder fields located on Earth, Mars and Moon. The data was collected during the BOULDERING Marie Skłodowska-Curie Global fellowship between October 2021 and 2024. The data was already splitted into train, validation and test datasets, but feel free to re-organize the labels at your convenience.

    For each image, all of the boulder outlines within the image were carefully mapped in QGIS. More information about the labelling procedure can be found in the following manuscript (https://agupubs.onlinelibrary.wiley.com/doi/full/10.1029/2023JE008013). This dataset differs from the previous dataset included along with the manuscript https://zenodo.org/records/8171052, as it contains more mapped images, especially of boulder populations around young impact structures on the Moon (cold spots). In addition, the boulder outlines were also pre-processed so that it can be ingested directly in YOLOv8.

    A description of what is what is given in the README.txt file (in addition in how to load the custom datasets in Detectron2 and YOLO). Most of the other files are mostly self-explanatory. Please see previous dataset or manuscript for more information. If you want to have more information about specific lunar and martian planetary images, the IDs of the images are still available in the name of the file. Use this ID to find more information (e.g., M121118602_00875_image.png, ID M121118602 ca be used on https://pilot.wr.usgs.gov/). I will also upload the raw data from which this pre-processed dataset was generated (see https://zenodo.org/records/14250970).

    Thanks to this database, you can easily train a Detectron2 Mask R-CNN or YOLO instance segmentation models to automatically detect boulders.

    How to cite:

    Please refer to the "how to cite" section of the readme file of https://github.com/astroNils/YOLOv8-BeyondEarth.

    Structure:

    . └── boulder2024/ ├── jupyter-notebooks/ │ └── REGISTERING_BOULDER_DATASET_IN_DETECTRON2.ipynb ├── test/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── train/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── validation/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── detectron2_inst_seg_boulder_dataset.json ├── README.txt ├── yolo_inst_seg_boulder_dataset.yaml

    detectron2_inst_seg_boulder_dataset.json

    is a json file containing the masks as expected by Detectron2 (see https://detectron2.readthedocs.io/en/latest/tutorials/datasets.html for more information on the format). In order to use this custom dataset, you need to register the dataset before using it in the training. There is an example how to do that in the jupyter-notebooks folder. You need to have detectron2, and all of its depedencies installed.

    yolo_inst_seg_boulder_dataset.yaml

    can be used as it is, however you need to update the paths in the .yaml file, to the test, train and validation folders. More information about the YOLO format can be found here (https://docs.ultralytics.com/datasets/segment/).

  8. iSAID Dataset (YOLO11-seg format)

    • kaggle.com
    Updated Jun 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mridankan Mandal (2025). iSAID Dataset (YOLO11-seg format) [Dataset]. https://www.kaggle.com/datasets/redzapdos123/isaid-dataset-yolo11-seg-format
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 25, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Mridankan Mandal
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    iSAID-YOLO11-Seg: A Cleaned YOLO11 Segmentation Conversion of the iSAID Aerial Instance Segmentation Dataset

    Overview A large-scale aerial imagery dataset reformatted from the original iSAID instance segmentation benchmark into Ultralytics’ YOLO11 segmentation format. Polygon masks have been normalized and cleaned for seamless training, validation, and inference with YOLO11-seg models .

    Dataset Composition

    • Total Images:

      • Train: 28,029
      • Validation: 9,512
      • Test (unlabeled): 19,377
    • Total Annotated Instances: 655,451 across 15 object categories (plus one “unlabeled” class for background).

    • Image Resolution: Original 6000×6000 tiles split into 800×800 crops.

    • Annotation Files: YOLO11 .txt segmentation labels (normalized coordinates).

    Directory Structure

    iSAID-YOLO11-Seg/
    ├── images/
    │  ├── train/    # 28,029 images
    │  ├── val/     # 9,512 images
    │  └── test/    # 19,377 images
    ├── labels/
    │  ├── train/    # 28,029 .txt polygon labels
    │  └── val/     # 9,512 .txt polygon labels
    ├── data.yaml    # Dataset configuration
    ├── ReadMe.md    # Dataset description and usage
    └── license.md    # License terms :contentReference[oaicite:2]{index=2}
    

    Annotation Format

    Each label file contains one polygon instance per line:

  9. Synthetic Gloomhaven Monsters

    • kaggle.com
    zip
    Updated Aug 30, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Eric de Potter (2020). Synthetic Gloomhaven Monsters [Dataset]. https://www.kaggle.com/ericdepotter/synthetic-gloomhaven-monsters
    Explore at:
    zip(0 bytes)Available download formats
    Dataset updated
    Aug 30, 2020
    Authors
    Eric de Potter
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Context

    One of my passions is playing board games with my friends. However one of them lives abroad and so we like to stream the game when playing with him. However instead of just having a normal stream, I wanted to show some additional information about the monsters that are on the game board. This originated in a fun project to train CNNs in order to detect these monsters.

    Content

    To have enough training data, I made a little project in UE4 to generate these training images. For each image there is a mask for every monster that appears in it. The dataset also includes annotations for the train images in the COCO format (annotations.json) and labes for the bounding box in Darknet format in the folder labels.

    There is a training and validation subset for the images, labels and masks folders. The structure is as follows: for the first training image containing an earth_demon and harrower_infester:

    • The image is stored at images/train/image_1.png
    • The label-file is stored at labels/train/label_1.png. This file contains two lines. One line for each monster. A line is constructed as follows: class_id center_x center_y width height. Note that the position and dimensions are relative to the image width and height.
    • There are two mask images located at masks/train. One is named image_1_mask_0_harrower_infester.png and the other image_1_mask_1_earth_demon.png.

    The code for generating this dataset and training a MaskRCNN and YoloV5 model can be found at https://github.com/ericdepotter/Gloomhaven-Monster-Recognizer.

    Acknowledgements

    I took pictures for the images of the monsters myself. The images of the game tiles I obtained from this collection of Gloomhaven assets.

    Inspiration

    This is a classic object detection or object segmentation problem.

  10. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Ultralytics (2024). COCO8-pose [Dataset]. https://huggingface.co/datasets/Ultralytics/COCO8-pose

COCO8-pose

coco8-pose

Ultralytics/COCO8-pose

Explore at:
Dataset updated
Dec 18, 2024
Dataset authored and provided by
Ultralytics
License

https://choosealicense.com/licenses/agpl-3.0/https://choosealicense.com/licenses/agpl-3.0/

Description

Ultralytics COCO8-pose Dataset

Ultralytics COCO8-pose is a small, but versatile pose detection dataset composed of the first 8 images of the COCO train 2017 set, 4 for training and 4 for validation. This dataset is ideal for testing and debugging object detection models, or for experimenting with new detection approaches. With 8 images, it is small enough to be easily manageable, yet diverse enough to test training pipelines for errors and act as a sanity check before training… See the full description on the dataset page: https://huggingface.co/datasets/Ultralytics/COCO8-pose.

Search
Clear search
Close search
Google apps
Main menu