100+ datasets found
  1. Image Annotation Services | Image Labeling for AI & ML |Computer Vision...

    • datarade.ai
    Updated Dec 29, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2023). Image Annotation Services | Image Labeling for AI & ML |Computer Vision Data| Annotated Imagery Data [Dataset]. https://datarade.ai/data-products/nexdata-image-annotation-services-ai-assisted-labeling-nexdata
    Explore at:
    .bin, .json, .xml, .csv, .xls, .sql, .txtAvailable download formats
    Dataset updated
    Dec 29, 2023
    Dataset authored and provided by
    Nexdata
    Area covered
    Uzbekistan, Morocco, Qatar, Taiwan, United States of America, Philippines, Jamaica, Korea (Republic of), Ireland, Montenegro
    Description
    1. Overview We provide various types of Annotated Imagery Data annotation services, including:
    2. Bounding box
    3. Polygon
    4. Segmentation
    5. Polyline
    6. Key points
    7. Image classification
    8. Image description ...
    9. Our Capacity
    10. Platform: Our platform supports human-machine interaction and semi-automatic labeling, increasing labeling efficiency by more than 30% per annotator.It has successfully been applied to nearly 5,000 projects.
    • Annotation Tools: Nexdata's platform integrates 30 sets of annotation templates, covering audio, image, video, point cloud and text.

    -Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.

    -Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001

    1. About Nexdata Nexdata has global data processing centers and more than 20,000 professional annotators, supporting on-demand data annotation services, such as speech, image, video, point cloud and Natural Language Processing (NLP) Data, etc. Please visit us at https://www.nexdata.ai/computerVisionTraining?source=Datarade
  2. R

    Object Detection Annotations Dataset

    • universe.roboflow.com
    zip
    Updated Jun 20, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    farah-mohsen-samy (2025). Object Detection Annotations Dataset [Dataset]. https://universe.roboflow.com/farah-mohsen-samy/object-detection-annotations/dataset/4
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 20, 2025
    Dataset authored and provided by
    farah-mohsen-samy
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Objects Bounding Boxes
    Description

    Object Detection Annotations

    ## Overview
    
    Object Detection Annotations is a dataset for object detection tasks - it contains Objects annotations for 302 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  3. 26 Class Object detection dataset

    • kaggle.com
    • gts.ai
    Updated Feb 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mohamed Gobara (2024). 26 Class Object detection dataset [Dataset]. https://www.kaggle.com/datasets/mohamedgobara/26-class-object-detection-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 6, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Mohamed Gobara
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    The "26 Class Object Detection Dataset" comprises a comprehensive collection of images annotated with objects belonging to 26 distinct classes. Each class represents a common urban or outdoor element encountered in various scenarios. The dataset includes the following classes:

    Bench Bicycle Branch Bus Bushes Car Crosswalk Door Elevator Fire Hydrant Green Light Gun Motorcycle Person Pothole Rat Red Light Scooter Stairs Stop Sign Traffic Cone Train Tree Truck Umbrella Yellow Light These classes encompass a wide range of objects commonly encountered in urban and outdoor environments, including transportation vehicles, traffic signs, pedestrian-related elements, and natural features. The dataset serves as a valuable resource for training and evaluating object detection models, particularly those focused on urban scene understanding and safety applications.

  4. Traffic Road Object Detection Dataset using YOLO.

    • kaggle.com
    Updated Nov 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ilyesBoukraa (2023). Traffic Road Object Detection Dataset using YOLO. [Dataset]. https://www.kaggle.com/datasets/boukraailyesali/traffic-road-object-detection-dataset-using-yolo
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 8, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    ilyesBoukraa
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Dataset Description: Car Object Detection in Road Traffic

    Overview:

    This dataset is designed for car object detection in road traffic scenes (Images with shape 1080x1920x3). The dataset is derived from publicly available video content on YouTube, specifically from the video with the Creative Commons Attribution license, available here. https://youtu.be/MNn9qKG2UFI?si=uJz_WicTCl8zfrVl" alt="youtube video">

    Source:

    • Video Source: YouTube Video.
    • License: Creative Commons Attribution (reuse allowed) more details here.
    • Dataset Contents: The dataset consists of a collection of image frames extracted from the video. Each image frame captures various scenes from road traffic. Car objects within these frames are annotated with bounding boxes.

    Annotation Details:

    • Bounding Boxes: Each image frame contains annotated bounding boxes around car objects, marking their locations in the scene.
    • Classes: The dataset is focused on car object detection, and car objects are labeled as the target class (aka one class only).
    • Data Format: Images are provided in JPEG format.
    • Annotation files are provided in YOLO text format.
    • We used labelImg GUI to label this dataset in YOLO format, more details are in this GitHub repo.

    Use Cases:

    • Object Detection: This dataset can be used to train and evaluate object detection models, with an emphasis on detecting cars in road traffic scenarios.

    Acknowledgments: We acknowledge and thank the creator of the original video for making it available under a Creative Commons Attribution license. Their contribution enables the development of datasets and research in the field of computer vision and object detection.

    Disclaimer: This dataset is provided for educational and research purposes and should be used in compliance with YouTube's terms of service and the Creative Commons Attribution license.

  5. Image Annotation Services | Image Labeling for AI & ML |Computer Vision...

    • data.nexdata.ai
    Updated Aug 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2024). Image Annotation Services | Image Labeling for AI & ML |Computer Vision Data| Annotated Imagery Data [Dataset]. https://data.nexdata.ai/products/nexdata-image-annotation-services-ai-assisted-labeling-nexdata
    Explore at:
    Dataset updated
    Aug 3, 2024
    Dataset authored and provided by
    Nexdata
    Area covered
    Greece, Nicaragua, Singapore, Thailand, Colombia, Kyrgyzstan, Japan, Puerto Rico, Croatia, Belgium
    Description

    Nexdata provides high-quality Annotated Imagery Data annotation for bounding box, polygon,segmentation,polyline, key points,image classification and image description. We have handled tons of data for autonomous driving, internet entertainment, retail, surveillance and security and etc.

  6. R

    Cattle Body Parts For Object Detection Dataset

    • universe.roboflow.com
    zip
    Updated Apr 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ali KHalili (2025). Cattle Body Parts For Object Detection Dataset [Dataset]. https://universe.roboflow.com/ali-khalili/cattle-body-parts-dataset-for-object-detection
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 29, 2025
    Dataset authored and provided by
    Ali KHalili
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Temp3 Bounding Boxes
    Description

    Cattle Body Parts Image Dataset for Object Detection

    This dataset is a curated collection of images featuring various cattle body parts aimed at facilitating object detection tasks. The dataset contains a total of 428 high-quality photos, meticulously annotated with three distinct classes: "Back," "Head," and "Leg."

    The dataset can be downloaded using this link. The dataset is also available at Roboflow Universe.

    A YOLOv7X model has been trained using the dataset and achieved a mAP of 99.6%. You can access the trained weights through this link.

    Motivation

    Accurate and reliable identification of different cattle body parts is crucial for various agricultural and veterinary applications. This dataset aims to provide a valuable resource for researchers, developers, and enthusiasts working on object detection tasks involving cattle, ultimately contributing to advancements in livestock management, health monitoring, and related fields.

    Data

    Overview

    • Total Images: 428
    • Classes: Back, Head, Leg
    • Annotations: Bounding boxes for each class

    Contents

    📦 Cattle_Body_Parts_OD.zip
     ┣ 📂 images
     ┃ ┣ 📜 image1.jpg
     ┃ ┣ 📜 image2.jpg
     ┃ ┗ ...
     ┗ 📂 annotations
      ┣ 📜 image1.json
      ┣ 📜 image2.json
      â”— ...
    

    Annotation Format

    Each annotation file corresponds to an image in the dataset and is formatted as per the LabelMe JSON standard. These annotations define the bounding box coordinates for each labeled body part, enabling straightforward integration into object detection pipelines.

    License

    This work is licensed under a Creative Commons Attribution 4.0 International License.

    Disclaimer

    This dataset has been collected from publicly available sources. I do not claim ownership of the data and have no intention of infringing on any copyright. The material contained in this dataset is copyrighted to their respective owners. I have made every effort to ensure the data is accurate and complete, but I cannot guarantee its accuracy or completeness. If you believe any data in this dataset infringes on your copyright, please get in touch with me immediately so I can take appropriate action.

  7. R

    Fall Annotation Dataset

    • universe.roboflow.com
    zip
    Updated Apr 5, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    capstone10 (2025). Fall Annotation Dataset [Dataset]. https://universe.roboflow.com/capstone10/fall-annotation-sdrzw/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 5, 2025
    Dataset authored and provided by
    capstone10
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Fall Bounding Boxes
    Description

    Fall Annotation

    ## Overview
    
    Fall Annotation is a dataset for object detection tasks - it contains Fall annotations for 1,497 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  8. Data from: Text Annotation Dataset

    • universe.roboflow.com
    zip
    Updated Jun 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tugas Akhir Real Time Object Detection (2024). Text Annotation Dataset [Dataset]. https://universe.roboflow.com/tugas-akhir-real-time-object-detection/text-annotation-81da0
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 7, 2024
    Dataset provided by
    Object detection
    Authors
    Tugas Akhir Real Time Object Detection
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Kacamata Masker Tas Sepatu Topi CW6W Bounding Boxes
    Description

    Text Annotation

    ## Overview
    
    Text Annotation is a dataset for object detection tasks - it contains Kacamata Masker Tas Sepatu Topi CW6W annotations for 6,400 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  9. P

    Paimon Dataset YOLO Detection Dataset

    • paperswithcode.com
    • gts.ai
    Updated Dec 3, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). Paimon Dataset YOLO Detection Dataset [Dataset]. https://paperswithcode.com/dataset/paimon-dataset-yolo-detection
    Explore at:
    Dataset updated
    Dec 3, 2024
    Description

    Description:

    👉 Download the dataset here

    This dataset consists of a diverse collection of images featuring Paimon, a popular character from the game Genshin Impact. The images have been sourced from in-game gameplay footage and capture Paimon from various angles and in different sizes (scales), making the dataset suitable for training YOLO object detection models.

    The dataset provides a comprehensive view of Paimon in different lighting conditions, game environments, and positions, ensuring the model can generalize well to similar characters or object detection tasks. While most annotations are accurately labeled, a small number of annotations may include minor inaccuracies due to manual labeling errors. This is ideal for researchers and developers working on character recognition, object detection in gaming environments, or other AI vision tasks.

    Download Dataset

    Dataset Features:

    Image Format: .jpg files in 640×320 resolution.

    Annotation Format: .txt files in YOLO format, containing bounding box data with:

    class_id

    x_center

    y_center

    width

    height

    Use Cases:

    Character Detection in Games: Train YOLO models to detect and identify in-game characters or NPCs.

    Gaming Analytics: Improve recognition of specific game elements for AI-powered game analytics tools.

    Research: Contribute to academic research focused on object detection or computer vision in animated and gaming environments.

    Data Structure:

    Images: High-quality .jpg images captured from multiple perspectives, ensuring robust model training across various orientations and lighting scenarios.

    Annotations: Each image has an associated .txt file that follows the YOLO format. The annotations are structured to include class identification, object location (center coordinates), and

    bounding box dimensions.

    Key Advantages:

    Varied Angles and Scales: The dataset includes Paimon from multiple perspectives, aiding in creating more versatile and adaptable object detection models.

    Real-World Scenario: Extracted from actual gameplay footage, the dataset simulates real-world detection challenges such as varying backgrounds, motion blur, and changing character scales.

    Training Ready: Suitable for training YOLO models and other deep learning frameworks that require object detection capabilities.

    This dataset is sourced from Kaggle.

  10. R

    Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges

    • public.roboflow.com
    zip
    Updated Sep 30, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Northeastern University - China (2022). Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges [Dataset]. https://public.roboflow.com/object-detection/hard-hat-workers/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 30, 2022
    Dataset authored and provided by
    Northeastern University - China
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Bounding Boxes of Workers
    Description

    Overview

    The Hard Hat dataset is an object detection dataset of workers in workplace settings that require a hard hat. Annotations also include examples of just "person" and "head," for when an individual may be present without a hard hart.

    The original dataset has a 75/25 train-test split.

    Example Image: https://i.imgur.com/7spoIJT.png" alt="Example Image">

    Use Cases

    One could use this dataset to, for example, build a classifier of workers that are abiding safety code within a workplace versus those that may not be. It is also a good general dataset for practice.

    Using this Dataset

    Use the fork or Download this Dataset button to copy this dataset to your own Roboflow account and export it with new preprocessing settings (perhaps resized for your model's desired format or converted to grayscale), or additional augmentations to make your model generalize better. This particular dataset would be very well suited for Roboflow's new advanced Bounding Box Only Augmentations.

    Dataset Versions:

    Image Preprocessing | Image Augmentation | Modify Classes * v1 (resize-416x416-reflect): generated with the original 75/25 train-test split | No augmentations * v2 (raw_75-25_trainTestSplit): generated with the original 75/25 train-test split | These are the raw, original images * v3 (v3): generated with the original 75/25 train-test split | Modify Classes used to drop person class | Preprocessing and Augmentation applied * v5 (raw_HeadHelmetClasses): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class * v8 (raw_HelmetClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and person classes * v9 (raw_PersonClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and helmet classes * v10 (raw_AllClasses): generated with a 70/20/10 train/valid/test split | These are the raw, original images * v11 (augmented3x-AllClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied | 3x image generation | Trained with Roboflow's Fast Model * v12 (augmented3x-HeadHelmetClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Fast Model * v13 (augmented3x-HeadHelmetClasses-AccurateModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Accurate Model * v14 (raw_HeadClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class, and remap/relabel helmet class to head

    Choosing Between Computer Vision Model Sizes | Roboflow Train

    About Roboflow

    Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

    Developers reduce 50% of their code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

    Roboflow Workmark

  11. Face Detection Dataset

    • kaggle.com
    Updated Jun 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Adil Shamim (2025). Face Detection Dataset [Dataset]. https://www.kaggle.com/datasets/adilshamim8/face-detection
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 7, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Adil Shamim
    License

    https://cdla.io/sharing-1-0/https://cdla.io/sharing-1-0/

    Description

    Imagine a world where machines see like humans—where your model doesn’t just scan pixels, but truly understands faces.

    Face detection is the gateway to computer vision. From unlocking smartphones to powering surveillance systems, it’s the first step toward machines that understand the world as we do.

    This dataset is crafted for creators, engineers, and visionaries who want to build models that don't just see — they recognize.

    Originally prepared and exported using Roboflow, this dataset includes a diverse collection of face images, carefully annotated for object detection tasks. It’s designed to help you train accurate, real-time face detection models using cutting-edge deep learning architectures.

    What's Inside

    • train/: A collection of labeled images for model training
    • valid/: A separate set of images for validation
    • data.yaml: A configuration file to streamline your training pipeline

    The structure is simple and scalable — optimized for quick experimentation and production-level deployment.

    Class Label

    • 0: Face

    Focused and minimal. One class. One purpose.

    Format

    • YOLO annotation format
    • Bounding boxes normalized for compatibility
    • Ready for YOLOv5, YOLOv8, and other object detection frameworks

    Why This Dataset?

    Because face detection is more than bounding boxes — it’s about interaction, identity, and trust. Whether you’re building an AI that understands presence, or a system that reacts to people in real-time, this dataset gives you the data to begin.

    Ideal For

    • Training face detection models
    • Deploying vision systems in real-world environments
    • Academic research, experiments, and prototypes
    • Edge AI and mobile applications

    License

    This dataset is released under Creative Commons Zero (CC0 1.0). Use it freely — in research, in production, or anywhere your ideas take you.

  12. thief detection dataset

    • kaggle.com
    Updated Mar 30, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    JanstyLewis7 (2025). thief detection dataset [Dataset]. https://www.kaggle.com/datasets/janstylewis7/thief-detection-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 30, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    JanstyLewis7
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description
    • Content: 2364 images with YOLOv8 annotations for classes ['car', 'face hiding', 'gun', 'human', 'human in hurry', 'human_brakingdoor', 'human_lockunlocking', 'weapon'].
    • Annotations: Bounding box coordinates in YOLO format [
  13. The Semantic PASCAL-Part Dataset

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Apr 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ivan Donadello; Ivan Donadello; Luciano Serafini; Luciano Serafini (2025). The Semantic PASCAL-Part Dataset [Dataset]. http://doi.org/10.5281/zenodo.5878773
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 24, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Ivan Donadello; Ivan Donadello; Luciano Serafini; Luciano Serafini
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The Semantic PASCAL-Part dataset

    The Semantic PASCAL-Part dataset is the RDF version of the famous PASCAL-Part dataset used for object detection in Computer Vision. Each image is annotated with bounding boxes containing a single object. Couples of bounding boxes are annotated with the part-whole relationship. For example, the bounding box of a car has the part-whole annotation with the bounding boxes of its wheels.

    This original release joins Computer Vision with Semantic Web as the objects in the dataset are aligned with concepts from:

    • the provided supporting ontology;
    • the WordNet database through its synstes;
    • the Yago ontology.

    The provided Python 3 code (see the GitHub repo) is able to browse the dataset and convert it in RDF knowledge graph format. This new format easily allows the fostering of research in both Semantic Web and Machine Learning fields.

    Structure of the semantic PASCAL-Part Dataset

    This is the folder structure of the dataset:

    • semanticPascalPart: it contains the refined images and annotations (e.g., small specific parts are merged into bigger parts) of the PASCAL-Part dataset in Pascal-voc style.
      • Annotations_set: the test set annotations in .xml format. For further information See the PASCAL VOC format here.
      • Annotations_trainval: the train and validation set annotations in .xml format. For further information See the PASCAL VOC format here.
      • JPEGImages_test: the test set images in .jpg format.
      • JPEGImages_trainval: the train and validation set images in .jpg format.
      • test.txt: the 2416 image filenames in the test set.
      • trainval.txt: the 7687 image filenames in the train and validation set.

    The PASCAL-Part Ontology

    The PASCAL-Part OWL ontology formalizes, through logical axioms, the part-of relationship between whole objects (22 classes) and their parts (39 classes). The ontology contains 85 logical axiomns in Description Logic in (for example) the following form:

    Every potted_plant has exactly 1 plant AND
              has exactly 1 pot
    

    We provide two versions of the ontology: with and without cardinality constraints in order to allow users to experiment with or without them. The WordNet alignment is encoded in the ontology as annotations. We further provide the WordNet_Yago_alignment.csv file with both WordNet and Yago alignments.

    The ontology can be browsed with many Semantic Web tools such as:

    • Protégé: a graphical tool for ongology modelling;
    • OWLAPI: Java API for manipulating OWL ontologies;
    • rdflib: Python API for working with the RDF format.
    • RDF stores: databases for storing and semantically retrieve RDF triples. See here for some examples.

    Citing semantic PASCAL-Part

    If you use semantic PASCAL-Part in your research, please use the following BibTeX entry

    @article{DBLP:journals/ia/DonadelloS16,
     author  = {Ivan Donadello and
            Luciano Serafini},
     title   = {Integration of numeric and symbolic information for semantic image
            interpretation},
     journal  = {Intelligenza Artificiale},
     volume  = {10},
     number  = {1},
     pages   = {33--47},
     year   = {2016}
    }
    
  14. I

    Image Tagging and Annotation Services Report

    • marketresearchforecast.com
    doc, pdf, ppt
    Updated Mar 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Market Research Forecast (2025). Image Tagging and Annotation Services Report [Dataset]. https://www.marketresearchforecast.com/reports/image-tagging-and-annotation-services-33888
    Explore at:
    ppt, pdf, docAvailable download formats
    Dataset updated
    Mar 14, 2025
    Dataset authored and provided by
    Market Research Forecast
    License

    https://www.marketresearchforecast.com/privacy-policyhttps://www.marketresearchforecast.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The global image tagging and annotation services market is experiencing robust growth, driven by the increasing adoption of artificial intelligence (AI) and machine learning (ML) across diverse sectors. The market, estimated at $2.5 billion in 2025, is projected to expand at a Compound Annual Growth Rate (CAGR) of 18% from 2025 to 2033, reaching an estimated $10 billion by 2033. This significant expansion is fueled by several key factors. The automotive industry leverages image tagging and annotation for autonomous vehicle development, requiring vast amounts of labeled data for training AI algorithms. Similarly, the retail and e-commerce sectors utilize these services for image search, product recognition, and improved customer experiences. The healthcare industry benefits from advancements in medical image analysis, while the government and security sectors employ image annotation for surveillance and security applications. The rising availability of high-quality data, coupled with the decreasing cost of annotation services, further accelerates market growth. However, challenges remain. Data privacy concerns and the need for high-accuracy annotation can pose significant hurdles. The demand for specialized skills in data annotation also contributes to a potential bottleneck in the market's growth trajectory. Overcoming these challenges requires a collaborative approach, involving technological advancements in automation and the development of robust data governance frameworks. The market segmentation, encompassing various annotation types (image classification, object recognition/detection, boundary recognition, segmentation) and application areas (automotive, retail, BFSI, government, healthcare, IT, transportation, etc.), presents diverse opportunities for market players. The competitive landscape includes a mix of established players and emerging firms, each offering specialized services and targeting specific market segments. North America currently holds the largest market share due to early adoption of AI and ML technologies, while Asia-Pacific is anticipated to witness rapid growth in the coming years.

  15. R

    Annotation Dataset

    • universe.roboflow.com
    zip
    Updated Jun 19, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Detection objet (2025). Annotation Dataset [Dataset]. https://universe.roboflow.com/detection-objet-lkg54/annotation-c6dmx/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 19, 2025
    Dataset authored and provided by
    Detection objet
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Objects Bounding Boxes
    Description

    Annotation

    ## Overview
    
    Annotation is a dataset for object detection tasks - it contains Objects annotations for 1,229 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
    
  16. d

    Object Detection Data| Annotated Imagery Data| Damaged Car Images | AI...

    • datarade.ai
    Updated Aug 31, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pixta AI (2022). Object Detection Data| Annotated Imagery Data| Damaged Car Images | AI Training Data | 2,000 Licensed & 8,000 HD Images [Dataset]. https://datarade.ai/data-products/2-annotated-imagery-data-global-damaged-car-images-2-000-pixta-ai
    Explore at:
    .json, .xml, .csv, .txtAvailable download formats
    Dataset updated
    Aug 31, 2022
    Dataset authored and provided by
    Pixta AI
    Area covered
    Thailand, Australia, Malaysia, Netherlands, Austria, Germany, New Zealand, Norway, Canada, Philippines
    Description
    1. Overview This dataset is a collection of 2,000 Licensed and 8,000 HD damaged car images that are ready to use for optimizing the accuracy of computer vision models. All of the contents is sourced from PIXTA's stock library of 100M+ Asian-featured images and videos. PIXTA is the largest platform of visual materials in the Asia Pacific region offering fully-managed services, high quality contents and data, and powerful tools for businesses & organisations to enable their creative and machine learning projects.

    2. Use cases for damaged car images (object detection data) The 2,000 Licensed and 8,000 HD Images of damaged car could be used for various AI & Computer Vision models: Damage Inspection, Insurance Value Evaluation, Residual Value Forecast,... Each data set is supported by both AI and human review process to ensure labelling consistency and accuracy. Contact us for more custom datasets.

    3. Annotation Annotation is available for this dataset on demand, including:

    4. Bounding box

    5. Polygon

    6. Segmentation ...

    7. About PIXTA PIXTASTOCK is the largest Asian-featured stock platform providing data, contents, tools and services since 2005. PIXTA experiences 15 years of integrating advanced AI technology in managing, curating, processing over 100M visual materials and serving global leading brands for their creative and data demands. Visit us at https://www.pixta.ai/ or contact via our email contact@pixta.ai.

  17. WBC object detection dataset YOLOv8

    • kaggle.com
    Updated Sep 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Syed M Faizan Ahmed (2024). WBC object detection dataset YOLOv8 [Dataset]. https://www.kaggle.com/datasets/smfaizanahmed/wbc-object-detection-dataset-yolov8
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 26, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Syed M Faizan Ahmed
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    White Blood Cell (WBC) Detection in Microscopic Blood Cell Images

    Overview

    This dataset consists of microscopic images of blood cells specifically designed for the detection of White Blood Cells (WBC). It is intended for object detection tasks where the goal is to accurately locate and identify WBCs within blood smear images. Researchers and developers can utilize this data to train machine learning models for medical applications such as automated blood cell analysis.

    Dataset Content

    Images: The dataset contains high-resolution microscopic images of blood smears, where WBCs are scattered among Red Blood Cells (RBCs) and platelets. Each image is annotated with bounding boxes around the WBCs.

    Annotations: The annotations are provided in YOLO format, where each bounding box is associated with a label for WBC.

    File Structure:

    images/: Contains the blood cell images in .jpg or .png format. labels/: Contains the annotation files in .txt format (YOLO format), with each file corresponding to an image. Image Size: Varies, but all images are in high resolution suitable for detection tasks.

    Applications

    Medical Image Analysis: This dataset can be used to build models for the automated detection of WBCs, which is a crucial step in diagnosing various blood-related disorders. Object Detection: Ideal for testing object detection algorithms like YOLO, Faster R-CNN, or SSD. Acknowledgments This dataset is created using publicly available microscopic blood cell images, annotated for educational and research purposes. It can be used for developing machine learning models for academic research, prototyping medical applications, or object detection benchmarking.

  18. Image Dataset of Accessibility Barriers

    • zenodo.org
    • explore.openaire.eu
    zip
    Updated Mar 25, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jakob Stolberg; Jakob Stolberg (2022). Image Dataset of Accessibility Barriers [Dataset]. http://doi.org/10.5281/zenodo.6382090
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 25, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jakob Stolberg; Jakob Stolberg
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The Data
    The dataset consist of 5538 images of public spaces, annotated with steps, stairs, ramps and grab bars for stairs and ramps. The dataset has annotations 3564 of steps, 1492 of stairs, 143 of ramps and 922 of grab bars.

    Each step annotation is attributed with an estimate of the height of the step, as falling into one of three categories: less than 3cm, 3cm to 7cm or more than 7cm. Additionally it is attributed with a 'type', with the possibilities 'doorstep', 'curb' or 'other'.

    Stair annotations are attributed with the number of steps in the stair.

    Ramps are attributed with an estimate of their width, also falling into three categories: less than 50cm, 50cm to 100cm and more than 100cm.

    In order to preserve all additional attributes of the labels, the data is published in the CVAT XML format for images.

    Annotating Process
    The labelling has been done using bounding boxes around the objects. This format is compatible with many popular object detection models, e.g. the YOLO object model. A bounding box is placed so it contains exactly the visible part of the respective objects. This implies that only objects that are visible in the photo are annotated. This means in particular a photo of a stair or step from above, where the object cannot be seen, have not been annotated, even when a human viewer can possibly infer that there is a stair or a step from other features in the photo.

    Steps
    A step is annotated, when there is an vertical increment that functions as a passage between two surface areas intended human or vehicle traffic. This means that we have not included:

    • Increments that are to high to reasonably be considered at passage.
    • Increments that does not lead to a surface intended for human or vehicle traffic, e.g. a 'step' in front of a wall or a curb in front of a bush.

    In particular, the bounding box of a step object contains exactly the incremental part of the step, but does not extend into the top or bottom horizontal surface any more than necessary to enclose entirely the incremental part. This has been chosen for consistency reasons, as including parts of the horizontal surfaces would imply a non-trivial choice of how much to include, which we deemed would most likely lead to more inconstistent annotations.

    The height of the steps are estimated by the annotators, and are therefore not guarranteed to be accurate.

    The type of the steps typically fall into the category 'doorstep' or 'curb'. Steps that are in a doorway, entrance or likewise are attributed as doorsteps. We also include in this category steps that are immediately leading to a doorway within a proximity of 1-2m. Steps between different types of pathways, e.g. between streets and sidewalks, are annotated as curbs. Any other type of step are annotated with 'other'. Many of the 'other' steps are for example steps to terraces.

    Stairs
    The stair label is used whenever two or more steps directly follow each other in a consistent pattern. All vertical increments are enclosed in the bounding box, as well as intermediate surfaces of the steps. However the top and bottom surface is not included more than necessary for the same reason as for steps, as described in the previous section.

    The annotator counts the number of steps, and attribute this to the stair object label.

    Ramps
    Ramps have been annotated when a sloped passage way has been placed or built to connect two surface areas intended for human or vehicle traffic. This implies the same considerations as with steps. Alike also only the sloped part of a ramp is annotated, not including the bottom or top surface area.

    For each ramp, the annotator makes an assessment of the width of the ramp in three categories: less than 50cm, 50cm to 100cm and more than 100cm. This parameter is visually hard to assess, and sometimes impossible due to the view of the ramp.

    Grab Bars
    Grab bars are annotated for hand rails and similar that are in direct connection to a stair or a ramp. While horizontal grab bars could also have been included, this was omitted due to the implied ambiguities of fences and similar objects. As the grab bar was originally intended as an attributal information to stairs and ramps, we chose to keep this focus. The bounding box encloses the part of the grab bar that functions as a hand rail for the stair or ramp.

    Usage
    As is often the case when annotating data, much information depends on the subjective assessment of the annotator. As each data point in this dataset has been annotated only by one person, caution should be taken if the data is applied.

    Generally speaking, the mindset and usage guiding the annotations have been wheelchair accessibility. While we have strived to annotate at an object level, hopefully making the data more widely applicable than this, we state this explicitly as it may have swayed untrivial annotation choices.

    The attributal data, such as step height or ramp width are highly subjective estimations. We still provide this data to give a post-hoc method to adjust which annotations to use. E.g. for some purposes, one may be interested in detecting only steps that are indeed more than 3cm. The attributal data makes it possible to sort away the steps less than 3cm, so a machine learning algorithm can be trained on this more appropriate dataset for that use case. We stress however, that one cannot expect to train accurate machine learning algorithms inferring the attributal data, as this is not accurate data in the first place.

    We hope this dataset will be a useful building block in the endeavours for automating barrier detection and documentation.

  19. m

    SyntheticIndoorObjectDetectionDataset

    • data.mendeley.com
    Updated Mar 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nafiz Fahad (2025). SyntheticIndoorObjectDetectionDataset [Dataset]. http://doi.org/10.17632/nnph98d3kc.2
    Explore at:
    Dataset updated
    Mar 25, 2025
    Authors
    Nafiz Fahad
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset was collected from the MyNursingHome dataset, available at https://data.mendeley.com/datasets/fpctx3svzd/1 , and curated to develop a synthetic indoor object detection dataset for autonomous mobile robots, or robots, for supporting researchers in detecting and classifying objects for computer vision and pattern recognition. From the original dataset containing 25 object categories, we selected six key categories—basket bin (499 images), sofa (499 images), human (499 images), table (500 images), chair (496 images), and door (500 images). Initially, we collected a total of 2,993 images from these categories; however, during the annotation process using Roboflow, we rejected 1 sofa, 10 tables, 9 chairs, and 12 door images due to quality concerns, such as poor image resolution or difficulty in identifying the object, resulting in a final dataset of 2,961 images. To ensure an effective training pipeline, we divided the dataset into 70% training (2,073 images), 20% validation (591 images), and 10% test (297 images). Preprocessing steps included auto-orientation and resizing all images to 640×640 pixels to maintain uniformity. To improve generalization for real-world applications, we applied data augmentation techniques, including horizontal and vertical flipping, 90-degree rotations (clockwise, counter-clockwise, and upside down), random rotations within -15° to +15°, shearing within ±10° horizontally and vertically, and brightness adjustments between -15% and +15%. This augmentation process expanded the dataset to 7,107 images, with 6,219 images for training (88%), 597 for validation (8%), and 297 for testing (4%). Moreover, this well-annotated, preprocessed, and augmented dataset significantly improves object detection performance in indoor settings.

  20. D

    YOLO Annotated Wind Turbine Surface Damage Dataset

    • datasetninja.com
    Updated Oct 22, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Foster, Ashley; Best, Oscar; Gianni, Mario (2023). YOLO Annotated Wind Turbine Surface Damage Dataset [Dataset]. https://datasetninja.com/yolo-annotated-wind-turbine-surface-damage
    Explore at:
    Dataset updated
    Oct 22, 2023
    Dataset provided by
    Dataset Ninja
    Authors
    Foster, Ashley; Best, Oscar; Gianni, Mario
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    A YOLO Annotated Wind Turbine Surface Damage is a dataset of wind turbine surface damage composed of images from DTU - Drone inspection images of wind turbine dataset, split into 586x371 pixel images with YOLO format annotations for Dirt and Damage. The dataset consists of 13000 images, just under 3000 of which have instances of one of the two classes.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Nexdata (2023). Image Annotation Services | Image Labeling for AI & ML |Computer Vision Data| Annotated Imagery Data [Dataset]. https://datarade.ai/data-products/nexdata-image-annotation-services-ai-assisted-labeling-nexdata
Organization logo

Image Annotation Services | Image Labeling for AI & ML |Computer Vision Data| Annotated Imagery Data

Explore at:
.bin, .json, .xml, .csv, .xls, .sql, .txtAvailable download formats
Dataset updated
Dec 29, 2023
Dataset authored and provided by
Nexdata
Area covered
Uzbekistan, Morocco, Qatar, Taiwan, United States of America, Philippines, Jamaica, Korea (Republic of), Ireland, Montenegro
Description
  1. Overview We provide various types of Annotated Imagery Data annotation services, including:
  2. Bounding box
  3. Polygon
  4. Segmentation
  5. Polyline
  6. Key points
  7. Image classification
  8. Image description ...
  9. Our Capacity
  10. Platform: Our platform supports human-machine interaction and semi-automatic labeling, increasing labeling efficiency by more than 30% per annotator.It has successfully been applied to nearly 5,000 projects.
  • Annotation Tools: Nexdata's platform integrates 30 sets of annotation templates, covering audio, image, video, point cloud and text.

-Secure Implementation: NDA is signed to gurantee secure implementation and Annotated Imagery Data is destroyed upon delivery.

-Quality: Multiple rounds of quality inspections ensures high quality data output, certified with ISO9001

  1. About Nexdata Nexdata has global data processing centers and more than 20,000 professional annotators, supporting on-demand data annotation services, such as speech, image, video, point cloud and Natural Language Processing (NLP) Data, etc. Please visit us at https://www.nexdata.ai/computerVisionTraining?source=Datarade
Search
Clear search
Close search
Google apps
Main menu