100+ datasets found
  1. R

    Wine Label Segmentation Dataset

    • universe.roboflow.com
    zip
    Updated Sep 29, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SommeliAI (2025). Wine Label Segmentation Dataset [Dataset]. https://universe.roboflow.com/sommeliai/wine-label-segmentation-uootv/model/22
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 29, 2025
    Dataset authored and provided by
    SommeliAI
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Wine Labels Polygons
    Description

    Wine Label Segmentation

    ## Overview
    
    Wine Label Segmentation is a dataset for instance segmentation tasks - it contains Wine Labels annotations for 4,010 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  2. R

    Coffee Segmentation Labels Dataset

    • universe.roboflow.com
    zip
    Updated Mar 18, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    IITK (2023). Coffee Segmentation Labels Dataset [Dataset]. https://universe.roboflow.com/iitk-71h2z/coffee-segmentation-labels
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 18, 2023
    Dataset authored and provided by
    IITK
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Coffee Labels Polygons
    Description

    Coffee Segmentation Labels

    ## Overview
    
    Coffee Segmentation Labels is a dataset for instance segmentation tasks - it contains Coffee Labels annotations for 1,159 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  3. Grape Bunch Detect and Segment

    • kaggle.com
    Updated Jul 11, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nicolaas Regnier (2023). Grape Bunch Detect and Segment [Dataset]. https://www.kaggle.com/datasets/nicolaasregnier/grape-grapes
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 11, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Nicolaas Regnier
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    Introduction This dataset aims to explore the realm of object detection and segmentation with a specific focus on its applications in agriculture. The primary objective is to employ YOLOv8 and SAM techniques to develop robust models for detecting grape bunches.

    Dataset Description The dataset comprises four trained models utilizing YOLOv8 architecture. It includes two single-class models, one utilizing object detection and the other employing instance segmentation for grape detection. Additionally, there are two multi-class models capable of predicting and detecting different grape varietals. All models were trained using the large model from the Ultralytics repository (https://github.com/ultralytics/ultralytics).

    The dataset encompasses four grape varietals: - Pinot Noir: 102 images and labels - Chardonnay: 39 images and labels from me 47 from thsant - Sauvignon Blanc: 42 images and labels - Pinot Gris: 111 images and labels

    Total used for training: 341

    Note that the training of the segmentation models used a total of 20 images from each for a total of 100.

    Datasets Used for Training To see the dataset (e.g train/test/val folders) used for training the multi class object detection model please see the following zip file and note book:

    https://www.kaggle.com/code/nicolaasregnier/buildtraindataset/output?select=train-dataset-multi-class.zip

    To build a custom train-dataset please follow the instructions in the notebook: https://www.kaggle.com/code/nicolaasregnier/buildtraindataset/

    The labels used for training the multi-class instance segmentation model are under the folder SAMPreds

    Data Sources The dataset incorporates two primary data sources. The first source is a collection of images captured using an iPad Air 2 RGB camera. These images possess a resolution of 3226x2449 pixels and an 8-megapixel quality. The second source is contributed by GitHub user thsant, who has created an impressive project available at https://github.com/thsant/wgisd/tree/master.

    To label the data, a base model from a previous dataset was utilized, and the annotation process was carried out using LabelImg (https://github.com/heartexlabs/labelImg). It is important to note that some annotations from thsant's dataset required modifications for completeness.

    Implementation Steps The data preparation involved the utilization of classes and functions from the "my_SAM" (https://github.com/regs08/my_SAM) and "KaggleUtils" (https://github.com/regs08/KaggleUtils) repositories, facilitating the creation of training sets and the application of SAM techniques.

    For model training, the YOLOv8 architecture with default hyperparameters was employed. The object detection models underwent 50 epochs of training, while the instance segmentation models were trained for 75 epochs.

    Segment Anything (SAM) from https://segment-anything.com/ was applied to the bbox-labeled data to generate images and corresponding masks for the instance segmentation models. No further editing of the images occurred after applying SAM.

    Evaluation and Inference The evaluation metrics utilized were Mean Average Precision (mAP). The following mAP values were obtained:

    Single-class object detection: - mAP50: 0.85449 - mAP50-95: 0.56177

    Multi-class object detection: - mAP50: 0.85336 - mAP50-95: 0.56316

    Single-class instance segmentation: - mAP50: (value not provided) - mAP50-95: (value not provided)

    Multi-class instance segmentation: - mAP50: 0.89436 - mAP50-95: 0.62785

    For more comprehensive metrics, please refer to the results folder corresponding to the model of interest.

  4. R

    Annotation Label Dataset

    • universe.roboflow.com
    zip
    Updated Jan 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shrijith Nambiar (2024). Annotation Label Dataset [Dataset]. https://universe.roboflow.com/shrijith-nambiar-5v3no/annotation-label
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 2, 2024
    Dataset authored and provided by
    Shrijith Nambiar
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Gun Polygons
    Description

    Annotation Label

    ## Overview
    
    Annotation Label is a dataset for instance segmentation tasks - it contains Gun annotations for 968 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  5. Z

    Data from: TimberVision: A Multi-Task Dataset and Framework for...

    • data-staging.niaid.nih.gov
    Updated Feb 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Steininger, Daniel; Simon, Julia; Trondl, Andreas; Murschitz, Markus (2025). TimberVision: A Multi-Task Dataset and Framework for Log-Component Segmentation and Tracking in Autonomous Forestry Operations [Dataset]. https://data-staging.niaid.nih.gov/resources?id=zenodo_14825845
    Explore at:
    Dataset updated
    Feb 6, 2025
    Dataset provided by
    Austrian Institute of Technology
    Authors
    Steininger, Daniel; Simon, Julia; Trondl, Andreas; Murschitz, Markus
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    TimberVision is a dataset and framework for tree-trunk detection and tracking based on RGB images. It combines the advantages of oriented object detection and instance segmentation for optimizing robustness and efficiency, as described in the corresponding paper presented at WACV 2025. This repository contains images and annotations of the dataset as well as associated files. Source code, models, configuration files and further documentation can be found on our GitHub page.

    Data Structure

    The repository provides the following subdirectories:

    images: all images included in the TimberVision dataset

    labels: annotations corresponding to each image in YOLOv8 instance-segmentation format

    labels_eval: additional annotations

    mot: ground-truth annotations for multi-object-tracking evaluation in custom format

    timberseg: custom annotations for selected images from the TimberSeg dataset

    videos: complete video files used for evaluating multi-object-tracking (annotated keyframes sampled from each file are included in the images and labels directories)

    scene_parameters.csv: annotations of four scene parameters for each image describing trunk properties and context (see the paper for details)

    train/val/test.txt: original split files used for training, validation and testing of oriented-object-detection and instance-segmentation models with YOLOv8

    sources.md: references and licenses for images used in the open-source subset

    Subsets

    TimberVision consists of multiple subsets for different application scenarios. To identify them, file names of images and annotations include the following prefixes:

    tvc: core dataset recorded in forests and other outdoor locations

    tvh: images depicting harvesting scenarios in forests with visible machinery

    tvl: images depicting loading scenarios in more structured environments with visible machinery

    tvo: a small set of third-party open-source images for evaluating generalization

    tvt: keyframes extracted from videos at 2 fps for tracking evaluation

    Citing

    If you use the TimberVision dataset for your research, please cite the original paper: Steininger, D., Simon, J., Trondl, A., Murschitz, M., 2025. TimberVision: A Multi-Task Dataset and Framework for Log-Component Segmentation and Tracking in Autonomous Forestry Operations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV).

  6. m

    AutoNaVIT-C : Vision-Based Path and Obstacle Segmentation Dataset for...

    • data.mendeley.com
    Updated Apr 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jeevan S (2025). AutoNaVIT-C : Vision-Based Path and Obstacle Segmentation Dataset for Autonomous Driving - XML Compatible [Dataset]. http://doi.org/10.17632/8zhhjhyt35.1
    Explore at:
    Dataset updated
    Apr 14, 2025
    Authors
    Jeevan S
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    AutoNaVIT is a meticulously developed dataset designed to accelerate research in autonomous navigation, semantic scene understanding, and object segmentation through deep learning. This release includes only the annotation labels in XML format, aligned with high-resolution frames extracted from a controlled driving sequence at Vellore Institute of Technology – Chennai Campus (VIT-C). The corresponding images will be included in Version 2 of the dataset.

    Class Annotations The dataset features carefully annotated bounding boxes for the following three essential classes relevant to real-time navigation and path planning in autonomous vehicles:

    Kerb – 1,377 instances

    Obstacle – 258 instances

    Path – 532 instances

    All annotations were produced using Roboflow with human-verified precision, ensuring consistent, high-quality data that supports robust model development for urban and semi-urban scenarios.

    Data Capture Specifications The source video was captured using a Sony IMX890 sensor, under stable daylight lighting. Below are the capture parameters:

    Sensor Size: 1/1.56", 50 MP

    Lens: 6P optical configuration

    Aperture: ƒ/1.8

    Focal Length: 24mm equivalent

    Pixel Size: 1.0 µm

    Features: Optical Image Stabilization (OIS), PDAF autofocus

    Video Duration: 4 minutes 11 seconds

    Frame Rate: 2 FPS

    Total Annotated Frames: 504

    Format Compatibility and Model Support AutoNaVIT annotations are provided in Pascal VOC-compatible XML format, making them directly usable with models that support the Pascal VOC standard. The dataset is immediately compatible with:

    Pascal VOC

    As XML is a structured, extensible format, these annotations can be easily adapted for use with additional object detection frameworks that support XML-based label schemas.

    Benchmark Results To assess dataset utility, a YOLOv8 segmentation model was trained on the full dataset (including images). The model achieved the following results:

    Mean Average Precision (mAP): 96.5%

    Precision: 92.2%

    Recall: 94.4%

    These metrics demonstrate the dataset’s effectiveness in training models for autonomous vehicle perception and obstacle detection.

    Disclaimer and Attribution Requirement By downloading or using this dataset, users agree to the terms outlined in the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND 4.0):

    This dataset is available solely for academic and non-commercial research purposes.

    Proper attribution must be provided as follows: “Dataset courtesy of Vellore Institute of Technology – Chennai Campus.” This citation must appear in all research papers, presentations, or any work derived from this dataset.

    Redistribution, public hosting, commercial use, or modification is prohibited without prior written permission from VIT-C.

    Use of this dataset implies acceptance of these terms. All rights not explicitly granted are retained by VIT-C.

  7. Food Images and Labels Dataset for YoloV5

    • kaggle.com
    zip
    Updated Mar 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    CALEB STEPHEN URK20AI1009 (2023). Food Images and Labels Dataset for YoloV5 [Dataset]. https://www.kaggle.com/calebstephen/food-images-and-labels-dataset-for-yolov5
    Explore at:
    zip(41436337 bytes)Available download formats
    Dataset updated
    Mar 22, 2023
    Authors
    CALEB STEPHEN URK20AI1009
    Description

    This dataset contains 810 images of 12 different classes of food types. The dataset contains food that is generically found across the globe like Pizzas, Burgers, Fries, etc., and some food items that are geographically specific to India. Those include Idli, Vada, Chapathi, etc. In order for the Yolo model to recognize extremely generic items like fruits and common ingredients, the dataset was trained on Apples, Bananas, Rice, Tomatoes, etc. This dataset was created using roboflow's dataset creator present on the roboflow website. The data was augmented using roboflow's dataset augmentation methods like Flip 90 degrees and different ranges of saturation. The dataset can be used with YoloV5 and YoloV8 as well.

  8. R

    Auto Label Dataset

    • universe.roboflow.com
    zip
    Updated Mar 13, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Open Data Science (2024). Auto Label Dataset [Dataset]. https://universe.roboflow.com/open-data-science/auto-label-unuoz/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 13, 2024
    Dataset authored and provided by
    Open Data Science
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Buildings 4jY2 Polygons
    Description

    Auto Label

    ## Overview
    
    Auto Label is a dataset for instance segmentation tasks - it contains Buildings 4jY2 annotations for 7,839 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
    
  9. m

    AutoNaVIT : Vision-Based Path and Obstacle Segmentation Dataset for...

    • data.mendeley.com
    Updated Apr 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jeevan S (2025). AutoNaVIT : Vision-Based Path and Obstacle Segmentation Dataset for Autonomous Driving - TXT Compatible [Dataset]. http://doi.org/10.17632/nh645b8ds8.1
    Explore at:
    Dataset updated
    Apr 14, 2025
    Authors
    Jeevan S
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    AutoNaVIT is a meticulously curated dataset developed to assist research in autonomous navigation, scene understanding, and deep learning-based object segmentation. This release contains only the annotation labels in TXT format corresponding to high-resolution frames extracted from a recorded driving sequence at Vellore Institute of Technology – Chennai Campus (VIT-C). The corresponding images will be made available in Version 2 of the dataset soon.

    The dataset features manually annotated bounding boxes and labels for three essential classes critical for autonomous vehicle navigation:

    Kerb – 1,377 instances

    Obstacle – 258 instances

    Path – 532 instances

    All annotations were created using Roboflow, ensuring high fidelity and consistency, which is vital for real-world autonomous driving applications in both urban and semi-urban environments.

    Data Capture Specifications Source imagery was recorded using a Sony IMX890 sensor with the following specifications:

    Sensor Size: 1/1.56", 50 MP

    Lens: 6P, ƒ/1.8, 24mm equivalent, 1.0 µm pixels

    Features: OIS (Optical Image Stabilization), PDAF autofocus

    Video Duration: 4 min 11 sec

    Frame Rate: 2 FPS

    Total Annotated Frames: 504

    Format Compatibility and Model Support AutoNaVIT annotations are provided in standard TXT format, enabling direct compatibility with the following 13 models:

    yolokeras

    yolov4pytorch

    darknet

    yolov5-obb

    yolov8-obb

    imt-yolov6

    yolov4scaled

    yolov5pytorch

    yolov7pytorch

    yolov8

    yolov9

    yolov11

    yolov12

    As the dataset adheres to standard YOLO TXT annotations, it can easily be adapted for other models or frameworks that support TXT-based annotations.

    Benchmark Results To evaluate the dataset’s performance, a YOLOv8-based segmentation model was trained on the complete dataset (images + annotations). The model achieved:

    Mean Average Precision (mAP): 96.5%

    Precision: 92.2%

    Recall: 94.4%

    These results confirm the dataset's high utility and reliability in training segmentation models for autonomous vehicle perception systems.

    Disclaimer and Attribution Requirement By accessing or using this dataset, users agree to the terms outlined under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND 4.0):

    Usage is permitted solely for non-commercial academic and research purposes.

    Proper attribution must be given, stating: “Dataset courtesy of Vellore Institute of Technology – Chennai Campus.” This acknowledgment must be included in all forms of publication, presentation, or dissemination of work utilizing this dataset.

    Redistribution, commercial use, modification, or public hosting of the dataset is prohibited without explicit written permission from VIT-C.

    Use of this dataset implies acceptance of these terms. All rights not explicitly granted are reserved by VIT-C.

  10. Cell Counting (Roboflow) – Custom Segmentation

    • kaggle.com
    Updated Sep 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nhut Nguyen (2025). Cell Counting (Roboflow) – Custom Segmentation [Dataset]. https://www.kaggle.com/datasets/tensura3607/cell-counting-roboflow-segmentation-masks/versions/7
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 10, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Nhut Nguyen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Overview

    This dataset is derived from the [Cell Counting v5 dataset on Roboflow] (https://universe.roboflow.com/cell-counting-hapu2/cell-counting-so7h7 ).
    The original dataset was provided in YOLOv8 object detection format.
    We created binary masks suitable for UNet-based semantic segmentation tasks.

    Additionally, we generated augmented images to increase dataset variability.

    Dataset Composition

    • Train/Valid/Test Splits
      Each split contains:

      • images/: Source images
      • labels/: YOLO annotation files (kept for reference)
      • masks_binary/: Binary masks for semantic segmentation
    • Augmented Images

      • Directory: aug_inference_only/images/
      • Contains 105 augmented images generated from the original 35 images
      • No masks or labels are provided for these augmentations
      • Intended for inference/visualization only (not for training or evaluation)

    Data Augmentation

    Each of the 35 original images was augmented with 3 additional variations, resulting in 105 augmented images.

    Augmentation methods include:
    - Random rotation (−90° to 90°)
    - Flipping (horizontal, vertical, both)
    - Shifting and scaling
    - Brightness/contrast adjustment
    - Gaussian noise injection

    Source

    License

    CC BY 4.0 – This dataset can be shared and adapted with appropriate attribution.

  11. Precious Gemstone Identification

    • kaggle.com
    zip
    Updated Mar 28, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GauravKamath02 (2024). Precious Gemstone Identification [Dataset]. https://www.kaggle.com/datasets/gauravkamath02/precious-gemstone-identification
    Explore at:
    zip(7743109183 bytes)Available download formats
    Dataset updated
    Mar 28, 2024
    Authors
    GauravKamath02
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Precious Gemstone Identification

    Description: This comprehensive dataset comprises annotated images of a diverse range of precious gemstones meticulously curated for gemstone identification tasks. With 87 classes of gemstones for classification unique varieties including Chalcedony Blue, Amber, Aventurine Yellow, Dumortierite, Pearl, Aventurine Green, and many others, this dataset serves as a valuable resource for training and evaluating machine learning models in gemstone recognition.

    Gemstone Variety: The dataset encompasses a wide spectrum of precious gemstones, ranging from well-known varieties like Emerald, Ruby, Sapphire, and Diamond to lesser-known gems such as Benitoite, Larimar, and Sphene.

    Dataset Split: Train Set: 92% (46404 images) Validation Set: 4% (1932 images) Test Set: 4% (1932 images)

    Preprocessing: Images in the dataset have been preprocessed to ensure consistency and quality:

    • Auto-Orient: Applied to correct orientation inconsistencies.
    • Resize: Images are uniformly resized to 640x640 pixels.
    • Tiling: Organized into a grid of 3 rows x 2 columns for efficient processing.

    Augmentations: To enhance model robustness and generalization, each training example has been augmented with various transformations:

    • Flip: Horizontal and Vertical flips are applied.
    • Rotation: Random rotation between -15° and +15°.
    • Shear: Horizontal and Vertical shearing with a range of ±10°.
    • Saturation: Adjusted randomly between -15% and +15%.
    • Brightness: Random brightness adjustment between -10% and +10%.

    File Formats Available:

    • COCO Segmentation: COCO (Common Objects in Context) Segmentation format is commonly used for semantic segmentation tasks. It provides annotations for object segmentation, where each object instance is labeled with a mask indicating its outline.
    • COCO: COCO format is a widely used standard for object detection and instance segmentation tasks. It includes annotations for bounding boxes around objects, along with corresponding class labels and segmentation masks if applicable.
    • TensorFlow : TensorFlow format typically refers to a data format compatible with TensorFlow, a popular deep learning framework. It often includes annotations in a format suitable for training object detection and segmentation models using TensorFlow.
    • VOC: VOC (Visual Object Classes) format is a standard format for object detection and classification tasks. It includes annotations for bounding boxes around objects, along with class labels and metadata, following the PASCAL VOC dataset format.
    • YOLOv8-obb: YOLOv8-obb format is specific to the YOLO (You Only Look Once) object detection model architecture. It typically includes annotations for object bounding boxes in YOLO format, where each bounding box is defined by its center coordinates, width, height, and class label.
    • YOLOv9 Segmentation: YOLOv9 Segmentation format is tailored for semantic segmentation tasks using the YOLOv9 architecture. It provides annotations for pixel-wise segmentation masks corresponding to object instances, enabling accurate segmentation of objects in images.
    • Server Benchmark: The Server Benchmark format is used for annotated images with bounding boxes for object detection tasks. Each annotation entry in the JSON-like structure contains details about a specific object instance within an image.

    Disclaimer:

    The images included in this dataset were sourced from various online platforms, primarily from minerals.net and www.rasavgems.com websites, as well as other online datasets. We have curated and annotated these datasets for the purpose of gemstone identification and made them available in different formats. We do not claim ownership of the original images, and we do not claim to own these images. Any trademarks, logos, or copyrighted materials belong to their respective owners.

    Researchers, enthusiasts and developers interested in gemstone identification, machine learning, and computer vision applications will find this dataset invaluable for training and benchmarking gemstone recognition algorithms.

  12. f

    Data from: Hierarchical Deep Learning Framework for Automated Marine...

    • figshare.com
    bin
    Updated Dec 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bjørn Christian Weinbach (2024). Hierarchical Deep Learning Framework for Automated Marine Vegetation and Fauna Analysis Using ROV Video Data [Dataset]. http://doi.org/10.6084/m9.figshare.25688718.v4
    Explore at:
    binAvailable download formats
    Dataset updated
    Dec 9, 2024
    Dataset provided by
    figshare
    Authors
    Bjørn Christian Weinbach
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Experimental data for the paper "Hierarchical Deep Learning Framework for Automated Marine Vegetation and Fauna Analysis Using ROV Video Data."This dataset supports the study "Hierarchical Deep Learning Framework for Automated Marine Vegetation and Fauna Analysis Using ROV Video Data" by providing resources essential for reproducing and validating the research findings.Dataset Contents and Structure:Hierarchical Model Weights: - .pth files containing trained weights for all alpha regularization values used in hierarchical classification models.MaskRCNN-Segmented Objects: - .jpg files representing segmented objects detected by the MaskRCNN model. - Accompanied by maskrcnn-segmented-objects-dataset.parquet, which includes metadata and classifications: - Columns:masked_image: Path to the segmented image file.confidence: Confidence score for the prediction.predicted_species: Predicted species label.species: True species label.MaskRCNN Weights: - Trained MaskRCNN model weights, including hierarchical CNN models integrated with MaskRCNN in the processing pipeline.Pre-Trained Models:.pt files for all object detectors trained on the Esefjorden Marine Vegetation Segmentation Dataset (EMVSD) in YOLO txt format.Segmented Object Outputs: - Segmentation outputs and datasets for the following models: - RT-DETR: - Segmented objects: rtdetr-segmented-objects/ - Dataset: rtdetr-segmented-objects-dataset.parquet - YOLO-SAG: - Segmented objects: yolosag-segmented-objects/ - Dataset: yolosag-segmented-objects-dataset.parquet - YOLOv11: - Segmented objects: yolov11-segmented-objects/ - Dataset: yolov11-segmented-objects-dataset.parquet - YOLOv8: - Segmented objects: yolov8-segmented-objects/ - Dataset: yolov8-segmented-objects-dataset.parquet - YOLOv9: - Segmented objects: yolov9-segmented-objects/ - Dataset: yolov9-segmented-objects-dataset.parquetUsage Instructions:1. Download and extract the dataset.2. Utilize the Python scripts provided in the associated GitHub repository for evaluation and inference: https://github.com/Ci2Lab/FjordVisionReproducibility:The dataset includes pre-trained weights, segmentation outputs, and experimental results to facilitate reproducibility. The .parquet files and segmented object directories follow a standardized format to ensure consistency.Licensing:This dataset is released under the CC-BY 4.0 license, permitting reuse with proper attribution.Related Materials:- GitHub Repository: https://github.com/Ci2Lab/FjordVision

  13. m

    AutoNaVIT : Vision-Based Path and Obstacle Segmentation Dataset for...

    • data.mendeley.com
    Updated Apr 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jeevan S (2025). AutoNaVIT : Vision-Based Path and Obstacle Segmentation Dataset for Autonomous Driving - CSV Compatible [Dataset]. http://doi.org/10.17632/kb9sgg7x2p.1
    Explore at:
    Dataset updated
    Apr 14, 2025
    Authors
    Jeevan S
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    AutoNaVIT is a carefully designed dataset intended to advance research in autonomous navigation, semantic scene understanding, and deep learning-based object segmentation. This release includes only the annotation labels in CSV format, corresponding to high-resolution frames extracted from a driving sequence recorded at Vellore Institute of Technology – Chennai Campus (VIT-C). The corresponding images will be provided in Version 2 of the dataset.

    The dataset comprises manually annotated bounding boxes for three key classes that are critical for path planning and perception in autonomous vehicle systems:

    Kerb – 1,377 instances

    Obstacle – 258 instances

    Path – 532 instances

    All annotations were generated using Roboflow, with precise, human-verified labeling for consistent, high-quality data—essential for training robust models that generalize well to real-world urban and semi-urban driving scenarios.

    Data Capture Specifications The video footage used for annotation was recorded using a Sony IMX890 camera sensor under stable daylight conditions, with the following details:

    Sensor Size: 1/1.56", 50 MP

    Lens: 6P optical configuration

    Aperture: ƒ/1.8

    Focal Length: 24mm equivalent

    Pixel Size: 1.0 µm

    Features: Optical Image Stabilization (OIS), PDAF autofocus

    Video Duration: 4 minutes 11 seconds

    Frame Rate: 2 FPS

    Total Annotated Frames: 504

    Format Compatibility and Model Support AutoNaVIT’s annotations are made available in standard CSV format, enabling direct compatibility with the following three models:

    Multiclass

    TensorFlow CSV

    RetinaNet

    Since CSV is a highly adaptable format, the annotations can be easily modified or reformatted to suit other deep learning models or pipelines that support CSV-based label structures.

    Benchmark Results To validate the dataset's effectiveness, a segmentation model using YOLOv8 was trained with the full dataset (images + annotations). The resulting performance metrics were:

    Mean Average Precision (mAP): 96.5%

    Precision: 92.2%

    Recall: 94.4%

    These metrics confirm the dataset’s value in developing perception systems for autonomous vehicles, particularly for object detection and path segmentation tasks.

    Disclaimer and Attribution Requirement By accessing or using this dataset, users agree to the following terms under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND 4.0):

    The dataset is available for non-commercial academic and research purposes only.

    Proper attribution must be included as: “Dataset courtesy of Vellore Institute of Technology – Chennai Campus.” This citation must appear in all forms of publication, presentation, or dissemination using this dataset.

    Redistribution, commercial usage, public hosting, or modification of the dataset is not permitted without explicit written consent from VIT-C.

    Use of the dataset indicates acceptance of these conditions. All rights not explicitly granted are reserved by VIT-C.

  14. R

    Rust Labels Dataset

    • universe.roboflow.com
    zip
    Updated Oct 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Computer Vision (2025). Rust Labels Dataset [Dataset]. https://universe.roboflow.com/computer-vision-nl3zu/rust-labels-otqcp
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 23, 2025
    Dataset authored and provided by
    Computer Vision
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Objects Polygons
    Description

    Rust Labels

    ## Overview
    
    Rust Labels is a dataset for instance segmentation tasks - it contains Objects annotations for 2,858 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  15. Blueberry segmentation with Segment Anything Model

    • kaggle.com
    Updated Sep 17, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhengkun_Li3969 (2024). Blueberry segmentation with Segment Anything Model [Dataset]. https://www.kaggle.com/datasets/zhengkunli3969/blueberry-segmentation-with-segment-anything-model
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 17, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Zhengkun_Li3969
    Description

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F19272950%2F3136c2a234771726dd17c29a758ba365%2Fb0.png?generation=1709156580593893&alt=media" alt="">

    Fig. 1: Diagram of the proposed blueberry fruit phenotyping workflow, involving four stages: data collection, dataset generation, model training, and phenotyping traits extraction. Our mobile platform equipped with a multi-view imaging system (top, left and right) was used to scan the blueberry plants through navigating over crop rows. On the basis of fruit/cluster detection dataset, we leverage a maturity classifier and a segmentation foundation model, SAM, to generate a semantic instance dataset for immature, semi-mature, and mature fruits segmentation. We proposed a lightweight improved YOLOv8 model for fruit cluster detection and blueberry segmentation for plant-scale and cluster-scale phenotyping traits extraction, including yield, maturity, cluster number and compactness.

    Dataset generation: https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F19272950%2F7a06785e03056ac75a41f0ba881c7ca2%2Fb1.png?generation=1709156618386382&alt=media" alt=""> Fig 2: Illumination of the proposed automated pixel-wise labels generation for immature, semi-mature, and mature blueberry fruits (genotype: keecrisp). From left to right: (a) bounding box labels of blueberries from our previous manual detection dataset [27]; (b) three-classes boxes labels (immature-yellow, semi-mature-red, mature-blue) re-classified with a maturity classifier; (c) pixel-wise mask labels of blueberry fruits with Segment Anything Model.

    References

    If you find this work or code useful, please cite:

    @article{li2025-robotic blueberry phenotyping,
     title={In-field blueberry fruit phenotyping with a MARS-PhenoBot and customized BerryNet},
     author={Li, Zhengkun and Xu, Rui and Li, Changying and Munoz, Patricio and Takeda, Fumiomi and Leme, Bruno},
     journal={Computers and Electronics in Agriculture},
     volume={232},
     pages={110057},
     year={2025},
     publisher={Elsevier}
    }
    
  16. R

    Page Number Segmentation Dataset

    • universe.roboflow.com
    zip
    Updated Aug 12, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Projects (2025). Page Number Segmentation Dataset [Dataset]. https://universe.roboflow.com/projects-c8tfq/page-number-segmentation/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Aug 12, 2025
    Dataset authored and provided by
    Projects
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Page Polygons
    Description

    Page Number Segmentation

    ## Overview
    
    Page Number Segmentation is a dataset for instance segmentation tasks - it contains Page annotations for 711 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  17. R

    Labeling Test Dataset

    • universe.roboflow.com
    zip
    Updated Jan 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Visages (2025). Labeling Test Dataset [Dataset]. https://universe.roboflow.com/visages/labeling-test-gsjhl
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 23, 2025
    Dataset authored and provided by
    Visages
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Ear Polygons
    Description

    Labeling Test

    ## Overview
    
    Labeling Test is a dataset for instance segmentation tasks - it contains Ear annotations for 402 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  18. Number of images in each dataset.

    • plos.figshare.com
    xls
    Updated Jan 31, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Qingyang Wang; Yazhi Luo; Ying Zhao; Shuhao Wang; Yiru Niu; Jinxi Di; Jia Guo; Guorong Lan; Lei Yang; Yu Shan Mao; Yuan Tu; Dingrong Zhong; Pei Zhang (2025). Number of images in each dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0317996.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jan 31, 2025
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Qingyang Wang; Yazhi Luo; Ying Zhao; Shuhao Wang; Yiru Niu; Jinxi Di; Jia Guo; Guorong Lan; Lei Yang; Yu Shan Mao; Yuan Tu; Dingrong Zhong; Pei Zhang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Compared with histological examination of lung cancer, cytology is less invasive and provides better preservation of complete morphology and detail. However, traditional cytological diagnosis requires an experienced pathologist to evaluate all sections individually under a microscope, which is a time-consuming process with low interobserver consistency. With the development of deep neural networks, the You Only Look Once (YOLO) object-detection model has been recognized for its impressive speed and accuracy. Thus, in this study, we developed a model for intraoperative cytological segmentation of pulmonary lesions based on the YOLOv8 algorithm, which labels each instance by segmenting the image at the pixel level. The model achieved a mean pixel accuracy and mean intersection over union of 0.80 and 0.70, respectively, on the test set. At the image level, the accuracy and area under the receiver operating characteristic curve values for malignant and benign (or normal) lesions were 91.0% and 0.90, respectively. In addition, the model was deemed suitable for diagnosing pleural fluid cytology and bronchoalveolar lavage fluid cytology images. The model predictions were strongly correlated with pathologist diagnoses and the gold standard, indicating the model’s ability to make clinical-level decisions during initial diagnosis. Thus, the proposed method is useful for rapidly localizing lung cancer cells based on microscopic images and outputting image interpretation results.

  19. R

    Label Real Data Dataset

    • universe.roboflow.com
    zip
    Updated Oct 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    label (2024). Label Real Data Dataset [Dataset]. https://universe.roboflow.com/label-xhpov/label-real-data/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 1, 2024
    Dataset authored and provided by
    label
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Bags Polygons
    Description

    Label Real Data

    ## Overview
    
    Label Real Data is a dataset for instance segmentation tasks - it contains Bags annotations for 318 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [Public Domain license](https://creativecommons.org/licenses/Public Domain).
    
  20. R

    Label Data Dataset

    • universe.roboflow.com
    zip
    Updated Apr 11, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    KHKT 20252 (2025). Label Data Dataset [Dataset]. https://universe.roboflow.com/khkt-20252/label-data-khcp8/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 11, 2025
    Dataset authored and provided by
    KHKT 20252
    Variables measured
    Objects Polygons
    Description

    Label Data

    ## Overview
    
    Label Data is a dataset for instance segmentation tasks - it contains Objects annotations for 1,955 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
SommeliAI (2025). Wine Label Segmentation Dataset [Dataset]. https://universe.roboflow.com/sommeliai/wine-label-segmentation-uootv/model/22

Wine Label Segmentation Dataset

wine-label-segmentation-uootv

wine-label-segmentation-dataset

Explore at:
2 scholarly articles cite this dataset (View in Google Scholar)
zipAvailable download formats
Dataset updated
Sep 29, 2025
Dataset authored and provided by
SommeliAI
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Variables measured
Wine Labels Polygons
Description

Wine Label Segmentation

## Overview

Wine Label Segmentation is a dataset for instance segmentation tasks - it contains Wine Labels annotations for 4,010 images.

## Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

  ## License

  This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Search
Clear search
Close search
Google apps
Main menu