100+ datasets found
  1. R

    Phone Data Annotate Dataset

    • universe.roboflow.com
    zip
    Updated Aug 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    phone dataset annotate (2024). Phone Data Annotate Dataset [Dataset]. https://universe.roboflow.com/phone-dataset-annotate/phone-data-annotate
    Explore at:
    zipAvailable download formats
    Dataset updated
    Aug 6, 2024
    Dataset authored and provided by
    phone dataset annotate
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Phone Hand Bounding Boxes
    Description

    Phone Data Annotate

    ## Overview
    
    Phone Data Annotate is a dataset for object detection tasks - it contains Phone Hand annotations for 522 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  2. d

    Data from: X-ray CT data with semantic annotations for the paper "A workflow...

    • catalog.data.gov
    • agdatacommons.nal.usda.gov
    Updated Jun 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Agricultural Research Service (2025). X-ray CT data with semantic annotations for the paper "A workflow for segmenting soil and plant X-ray CT images with deep learning in Google’s Colaboratory" [Dataset]. https://catalog.data.gov/dataset/x-ray-ct-data-with-semantic-annotations-for-the-paper-a-workflow-for-segmenting-soil-and-p-d195a
    Explore at:
    Dataset updated
    Jun 5, 2025
    Dataset provided by
    Agricultural Research Service
    Description

    Leaves from genetically unique Juglans regia plants were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA). Soil samples were collected in Fall of 2017 from the riparian oak forest located at the Russell Ranch Sustainable Agricultural Institute at the University of California Davis. The soil was sieved through a 2 mm mesh and was air dried before imaging. A single soil aggregate was scanned at 23 keV using the 10x objective lens with a pixel resolution of 650 nanometers on beamline 8.3.2 at the ALS. Additionally, a drought stressed almond flower bud (Prunus dulcis) from a plant housed at the University of California, Davis, was scanned using a 4x lens with a pixel resolution of 1.72 µm on beamline 8.3.2 at the ALS Raw tomographic image data was reconstructed using TomoPy. Reconstructions were converted to 8-bit tif or png format using ImageJ or the PIL package in Python before further processing. Images were annotated using Intel’s Computer Vision Annotation Tool (CVAT) and ImageJ. Both CVAT and ImageJ are free to use and open source. Leaf images were annotated in following Théroux-Rancourt et al. (2020). Specifically, Hand labeling was done directly in ImageJ by drawing around each tissue; with 5 images annotated per leaf. Care was taken to cover a range of anatomical variation to help improve the generalizability of the models to other leaves. All slices were labeled by Dr. Mina Momayyezi and Fiona Duong.To annotate the flower bud and soil aggregate, images were imported into CVAT. The exterior border of the bud (i.e. bud scales) and flower were annotated in CVAT and exported as masks. Similarly, the exterior of the soil aggregate and particulate organic matter identified by eye were annotated in CVAT and exported as masks. To annotate air spaces in both the bud and soil aggregate, images were imported into ImageJ. A gaussian blur was applied to the image to decrease noise and then the air space was segmented using thresholding. After applying the threshold, the selected air space region was converted to a binary image with white representing the air space and black representing everything else. This binary image was overlaid upon the original image and the air space within the flower bud and aggregate was selected using the “free hand” tool. Air space outside of the region of interest for both image sets was eliminated. The quality of the air space annotation was then visually inspected for accuracy against the underlying original image; incomplete annotations were corrected using the brush or pencil tool to paint missing air space white and incorrectly identified air space black. Once the annotation was satisfactorily corrected, the binary image of the air space was saved. Finally, the annotations of the bud and flower or aggregate and organic matter were opened in ImageJ and the associated air space mask was overlaid on top of them forming a three-layer mask suitable for training the fully convolutional network. All labeling of the soil aggregate and soil aggregate images was done by Dr. Devin Rippner. These images and annotations are for training deep learning models to identify different constituents in leaves, almond buds, and soil aggregates Limitations: For the walnut leaves, some tissues (stomata, etc.) are not labeled and only represent a small portion of a full leaf. Similarly, both the almond bud and the aggregate represent just one single sample of each. The bud tissues are only divided up into buds scales, flower, and air space. Many other tissues remain unlabeled. For the soil aggregate annotated labels are done by eye with no actual chemical information. Therefore particulate organic matter identification may be incorrect. Resources in this dataset:Resource Title: Annotated X-ray CT images and masks of a Forest Soil Aggregate. File Name: forest_soil_images_masks_for_testing_training.zipResource Description: This aggregate was collected from the riparian oak forest at the Russell Ranch Sustainable Agricultural Facility. The aggreagate was scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 0,0,0; pores spaces have a value of 250,250, 250; mineral solids have a value= 128,0,0; and particulate organic matter has a value of = 000,128,000. These files were used for training a model to segment the forest soil aggregate and for testing the accuracy, precision, recall, and f1 score of the model.Resource Title: Annotated X-ray CT images and masks of an Almond bud (P. Dulcis). File Name: Almond_bud_tube_D_P6_training_testing_images_and_masks.zipResource Description: Drought stressed almond flower bud (Prunis dulcis) from a plant housed at the University of California, Davis, was scanned by X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 4x lens with a pixel resolution of 1.72 µm using. For masks, the background has a value of 0,0,0; air spaces have a value of 255,255, 255; bud scales have a value= 128,0,0; and flower tissues have a value of = 000,128,000. These files were used for training a model to segment the almond bud and for testing the accuracy, precision, recall, and f1 score of the model.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads Resource Title: Annotated X-ray CT images and masks of Walnut leaves (J. Regia) . File Name: 6_leaf_training_testing_images_and_masks_for_paper.zipResource Description: Stems were collected from genetically unique J. regia accessions at the 117 USDA-ARS-NCGR in Wolfskill Experimental Orchard, Winters, California USA to use as scion, and were grafted by Sierra Gold Nursery onto a commonly used commercial rootstock, RX1 (J. microcarpa × J. regia). We used a common rootstock to eliminate any own-root effects and to simulate conditions for a commercial walnut orchard setting, where rootstocks are commonly used. The grafted saplings were repotted and transferred to the Armstrong lathe house facility at the University of California, Davis in June 2019, and kept under natural light and temperature. Leaves from each accession and treatment were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 170,170,170; Epidermis value= 85,85,85; Mesophyll value= 0,0,0; Bundle Sheath Extension value= 152,152,152; Vein value= 220,220,220; Air value = 255,255,255.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads

  3. R

    Data from: Annotate Dataset

    • universe.roboflow.com
    zip
    Updated Dec 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    test (2024). Annotate Dataset [Dataset]. https://universe.roboflow.com/test-gslwr/annotate-7rpzn/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Dec 7, 2024
    Dataset authored and provided by
    test
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Pastry Bounding Boxes
    Description

    Annotate

    ## Overview
    
    Annotate is a dataset for object detection tasks - it contains Pastry annotations for 887 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  4. h

    fgan-annotate-dataset

    • huggingface.co
    Updated Feb 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aaron Emmanuel Enejo (2024). fgan-annotate-dataset [Dataset]. https://huggingface.co/datasets/aaronemmanuel/fgan-annotate-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 7, 2024
    Authors
    Aaron Emmanuel Enejo
    Description

    Dataset Card for fgan-annotate-dataset

    This dataset has been created with Argilla. As shown in the sections below, this dataset can be loaded into Argilla as explained in Load with Argilla, or used directly with the datasets library in Load with datasets.

      Dataset Summary
    

    This dataset contains:

    A dataset configuration file conforming to the Argilla dataset format named argilla.yaml. This configuration file will be used to configure the dataset when using the… See the full description on the dataset page: https://huggingface.co/datasets/aaronemmanuel/fgan-annotate-dataset.

  5. R

    Crack Annotate Dataset

    • universe.roboflow.com
    zip
    Updated Jun 9, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    crackpatternproject (2022). Crack Annotate Dataset [Dataset]. https://universe.roboflow.com/crackpatternproject/crack-annotate
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 9, 2022
    Dataset authored and provided by
    crackpatternproject
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Crack Pattern Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Civil Engineering Monitoring: The "crack-annotate" model can be leveraged to analyze buildings, bridges, and other infrastructural facilities for any cracks, helping in the prediction of structural failures and planning timely repairs.

    2. Art Conservation: This model can be beneficial in assessing the preservation status of historic monuments, artifacts or art pieces. By identifying the type of crack patterns, experts can decide on appropriate restoration techniques.

    3. Earthquake Damage Assessment: Post-earthquake, the model can help identify and classify the damage on various structures, assisting in the damage assessment phase and prioritization of repair works.

    4. Construction Quality Control: During construction, the model can inspect structures for premature cracks, indicative of issues in materials used or workmanship, allowing early corrective actions.

    5. Aerospace Engineering: "crack-annotate" could be useful for inspecting the exterior of spacecraft or aircraft, helping identify potential damages that may need maintenance.

  6. e

    Image Annotation Tool - Dataset - B2FIND

    • b2find.eudat.eu
    Updated May 9, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). Image Annotation Tool - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/6411692f-8edd-5ce6-a7e6-3bb08304d7c3
    Explore at:
    Dataset updated
    May 9, 2023
    Description

    Image annotation tool is a web application that allows users to mark zones of interest in an image. These zones are then converted to TEI P5 code snippet that can be used in your document to connect the image and the text. This tool was developed to help students and teachers at the Faculty of Arts, Charles University to mark and annotate images of manuscripts.

  7. R

    Polygon Annotate Dataset

    • universe.roboflow.com
    zip
    Updated Apr 26, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    AI (2022). Polygon Annotate Dataset [Dataset]. https://universe.roboflow.com/ai-lsyuc/polygon-annotate
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 26, 2022
    Dataset authored and provided by
    AI
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Peralatan Ayam Bounding Boxes
    Description

    Polygon Annotate

    ## Overview
    
    Polygon Annotate is a dataset for object detection tasks - it contains Peralatan Ayam annotations for 1,408 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  8. Z

    Mouse Grooming Detection Annotated Dataset

    • data.niaid.nih.gov
    • zenodo.org
    Updated Apr 6, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Vivek Kumar (2021). Mouse Grooming Detection Annotated Dataset [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_4646087
    Explore at:
    Dataset updated
    Apr 6, 2021
    Dataset provided by
    Hao He
    Asaf Peer
    Brian Q Geuther
    Gautam Sabnis
    Vivek Kumar
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    This dataset contains 2 records. The first record is the annotated dataset. The second record contains a built singularity image containing the code and trained model for predicting on new videos.

    We generated 1,253 video clips which total 2,637,363 frames. Each video had variable duration, depending upon the grooming prediction length. Annotators were required to provide a "Grooming" or "Not Grooming" annotation for each frame. The annotated dataset is stored in the h5 record and is described as follows:

    First level grouping is Train/Validation split

    Second level grouping is by Video Clip

    Each video contains 5 datasets

    nframe

    Number of frames in this video

    Shape: 1

    video

    Raw Video

    Shape: nframes x 112 x 112

    label

    Labels for each frame

    0 = not grooming, 1 = grooming

    Shape: nframes

    mask

    Information for whether or not annotators agreed

    0 = disagree, 1 = agree

    When annotators disagree, label contains the values from the first person to annotate the frame

    Shape: nframes

    nlabelers

    Number of annotators that have labeled the video clip

    Shape: 1

  9. e

    Analyzing Dataset Annotation Quality Management in the Wild - Dataset -...

    • b2find.eudat.eu
    Updated Jul 18, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). Analyzing Dataset Annotation Quality Management in the Wild - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/475ea996-70d3-5fc1-97d5-386e6830c6b5
    Explore at:
    Dataset updated
    Jul 18, 2023
    Description

    This is the accompanying data for the paper "Analyzing Dataset Annotation Quality Management in the Wild". Data quality is crucial for training accurate, unbiased, and trustworthy machine learning models and their correct evaluation. Recent works, however, have shown that even popular datasets used to train and evaluate state-of-the-art models contain a non-negligible amount of erroneous annotations, bias or annotation artifacts. There exist best practices and guidelines regarding annotation projects. But to the best of our knowledge, no large-scale analysis has been performed as of yet on how quality management is actually conducted when creating natural language datasets and whether these recommendations are followed. Therefore, we first survey and summarize recommended quality management practices for dataset creation as described in the literature and provide suggestions on how to apply them. Then, we compile a corpus of 591 scientific publications introducing text datasets and annotate it for quality-related aspects, such as annotator management, agreement, adjudication or data validation. Using these annotations, we then analyze how quality management is conducted in practice. We find that a majority of the annotated publications apply good or very good quality management. However, we deem the effort of 30% of the works as only subpar. Our analysis also shows common errors, especially with using inter-annotator agreement and computing annotation error rates.

  10. Suicidal Ideation - Reddit Dataset

    • kaggle.com
    Updated Dec 13, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Varun (2023). Suicidal Ideation - Reddit Dataset [Dataset]. https://www.kaggle.com/datasets/rvarun11/suicidal-ideation-reddit-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 13, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Varun
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    The posts were manually annotated all the posts as Suicidal or Non-Suicidal based on the following rules: 1. Suicidal Text - Posts that conveyed definite signs of suicidal ideation or even showed signs of suffering extremely from mental health illnesses like depression etc. were marked in this category due to their relation with suicidal intent. - Posts that included detailed planning of suicide or asked questions related to committing suicide, for eg. “Hello, hypothetically what would be a good way to go without loved ones knowing?”. - Posts like "I weather today is so awful that it makes me want to kill myself hahaha" were carefully removed. - These posts were marked as “1”. 2. Non Suicidal Text - Posts that did not have anything related to suicide or self-harm were marked in this category. - Posts that used words related to suicide or self-harm in the context of news or information. - Posts that talked about suicide of some other person at some other time.
    - These posts were marked as “0”. This was the default category.

    Our annotators included one university professor and three university students who were very carefully instructed on how to annotate each post. The instructions are given below: 1. Select only one of the two categories mentioned above. 2. To select the default category in case of any doubt. 3. To remove any ambiguous posts which seemed very confusing after discussing with other annotators. 3. Maximum 100-200 posts were to be annotated in one session to avoid any mental fatigue. 4. Since the majority of posts in the dataset were extremely long (with words > 1000), a maximum of two annotation sessions were allowed in a day.

    Once the annotators completed their tasks, they were divided into pairs of two where they verified the annotations of the other annotator. Any disagreement was carefully resolved and the final annotation was mutually agreed upon by the pair. This helped in validating each annotation.

  11. R

    Label&annotate Dataset

    • universe.roboflow.com
    zip
    Updated Jun 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Garbage Labeling (2025). Label&annotate Dataset [Dataset]. https://universe.roboflow.com/garbage-labeling/label-annotate/model/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 5, 2025
    Dataset authored and provided by
    Garbage Labeling
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Cardboard Bounding Boxes
    Description

    Label&Annotate

    ## Overview
    
    Label&Annotate is a dataset for object detection tasks - it contains Cardboard annotations for 1,207 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  12. a

    Cat Annotation Dataset Merged

    • academictorrents.com
    bittorrent
    Updated Jul 2, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weiwei Zhang and Jian Sun and Xiaoou Tang (2014). Cat Annotation Dataset Merged [Dataset]. https://academictorrents.com/details/c501571c29d16d7f41d159d699d0e7fb37092cbd
    Explore at:
    bittorrent(1980831996)Available download formats
    Dataset updated
    Jul 2, 2014
    Dataset authored and provided by
    Weiwei Zhang and Jian Sun and Xiaoou Tang
    License

    https://academictorrents.com/nolicensespecifiedhttps://academictorrents.com/nolicensespecified

    Description

    Cat Annotation Dataset The CAT dataset includes 10,000 cat images. For each image, we annotate the head of cat with nine points, two for eyes, one for mouth, and six for ears. The detail configuration of the annotation was shown in Figure 6 of the original paper: Weiwei Zhang, Jian Sun, and Xiaoou Tang, "Cat Head Detection - How to Effectively Exploit Shape and Texture Features", Proc. of European Conf. Computer Vision, vol. 4, pp.802-816, 2008. ### Format The annotation data are stored in a file with the name of the corresponding cat image plus ".cat", one annotation file for each cat image. For each annotation file, the annotation data are stored in the following sequence: 1. Number of points (always 9) 2. Left Eye 3. Right Eye 4. Mouth 5. Left Ear-1 6. Left Ear-2 7. Left Ear-3 8. Right Ear-1 9. Right Ear-2 10. Right Ear-3 ### Training, Validation, and Testing We randomly divide the data into three sets: 5,000 images for training, 2,000 images for valid

  13. R

    Bottle Annotation Dataset

    • universe.roboflow.com
    zip
    Updated Sep 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Project (2024). Bottle Annotation Dataset [Dataset]. https://universe.roboflow.com/project-qnryu/bottle-annotation-eklbh
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 15, 2024
    Dataset authored and provided by
    Project
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Bottle Color Quadrant Bounding Boxes
    Description

    Bottle Annotation

    ## Overview
    
    Bottle Annotation is a dataset for object detection tasks - it contains Bottle Color Quadrant annotations for 303 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  14. e

    Annotation Curricula to Implicitly Train Non-Expert Annotators - Dataset -...

    • b2find.eudat.eu
    Updated Oct 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). Annotation Curricula to Implicitly Train Non-Expert Annotators - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/a5f6640f-4c4c-59be-b3e9-a53b79b57c97
    Explore at:
    Dataset updated
    Oct 31, 2023
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Annotation studies often require annotators to familiarize themselves with the task, its annotation scheme, and the data domain. This can be overwhelming in the beginning, mentally taxing, and induce errors into the resulting annotations; especially in citizen science or crowd sourcing scenarios where domain expertise is not required and only annotation guidelines are provided. To alleviate these issues, we propose annotation curricula, a novel approach to implicitly train annotators. We gradually introduce annotators into the task by ordering instances that are annotated according to a learning curriculum. To do so, we first formalize annotation curricula for sentence- and paragraph-level annotation tasks, define an ordering strategy, and identify well-performing heuristics and interactively trained models on three existing English datasets. We then conduct a user study with 40 voluntary participants who are asked to identify the most fitting misconception for English tweets about the Covid-19 pandemic. Our results show that using a simple heuristic to order instances can already significantly reduce the total annotation time while preserving a high annotation quality. Annotation curricula thus can provide a novel way to improve data collection. To facilitate future research, we further share our code and data consisting of 2,400 annotations.

  15. h

    Annotated-MIDI-Dataset

    • huggingface.co
    Updated Mar 27, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alex (2025). Annotated-MIDI-Dataset [Dataset]. https://huggingface.co/datasets/asigalov61/Annotated-MIDI-Dataset
    Explore at:
    Dataset updated
    Mar 27, 2025
    Authors
    Alex
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Annotated MIDI Dataset

      Comprehensive annotated MIDI dataset with original lyrics, lyrics summaries, lyrics sentiments, music descriptions, illustrations, pre-trained MIDI classification model and helper Python code
    
    
    
    
    
    
    
    
    
      Annotated MIDI Dataset LIVE demos
    

    Music Sentence Transformer Advanced MIDI Classifer Descriptive Music Transformer

      Installation
    

    from huggingface_hub import snapshot_download

    repo_id = "asigalov61/Annotated-MIDI-Dataset" repo_type =… See the full description on the dataset page: https://huggingface.co/datasets/asigalov61/Annotated-MIDI-Dataset.

  16. Data from: Region-based Annotation Data of Fire Images for Intelligent...

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Jan 23, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata; Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata (2022). Region-based Annotation Data of Fire Images for Intelligent Surveillance System [Dataset]. http://doi.org/10.5281/zenodo.5893854
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 23, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata; Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata
    Description

    This data presents fire segmentation annotation data on 12 commonly used and publicly available “VisiFire Dataset” videos from http://signal.ee.bilkent.edu.tr/VisiFire/. This annotations dataset was obtained by per-frame, manual hand annotation over the fire region with 2,684 total annotated frames. Since this annotation provides per-frame segmentation data, it offers a new and unique fire motion feature to the existing video, unlike other fire segmentation data that are collected from different still images. The annotations dataset also provides ground truth for segmentation task on videos. With segmentation task, it offers better insight on how well a machine learning model understood, not only detecting whether a fire is present, but also its exact location by calculating metrics such as Intersection over Union (IoU) with this annotations data. This annotations data is a tremendously useful addition to train, develop, and create a much better smart surveillance system for early detection in high-risk fire hotspots area.

  17. R

    Words Annotate Dataset

    • universe.roboflow.com
    zip
    Updated Jan 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Thanhpro (2025). Words Annotate Dataset [Dataset]. https://universe.roboflow.com/thanhpro/words-annotate/dataset/11
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 16, 2025
    Dataset authored and provided by
    Thanhpro
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Words Bounding Boxes
    Description

    Words Annotate

    ## Overview
    
    Words Annotate is a dataset for object detection tasks - it contains Words annotations for 208 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  18. e

    Annotate - Dataset - B2FIND

    • b2find.eudat.eu
    Updated Apr 28, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). Annotate - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/217de1c1-ec46-59bb-bafd-3a47cab2fdc4
    Explore at:
    Dataset updated
    Apr 28, 2023
    Description

    Annotate is a web and desktop application that should simplify the process of transforming photos of manuscripts to a browsable collection. It also allows users to annotate parts of the displayed images.

  19. C

    Annotations for ConfLab: A Data Collection Concept, Dataset, and Benchmark...

    • data.4tu.nl
    Updated Jun 7, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chirag Raman; Jose Vargas Quiros; Stephanie Tan; Ashraful Islam; Ekin Gedik; Hayley Hung (2022). Annotations for ConfLab: A Data Collection Concept, Dataset, and Benchmark for Machine Analysis of Free-Standing Social Interactions in the Wild [Dataset]. http://doi.org/10.4121/20017664.v3
    Explore at:
    Dataset updated
    Jun 7, 2022
    Dataset provided by
    4TU.ResearchData
    Authors
    Chirag Raman; Jose Vargas Quiros; Stephanie Tan; Ashraful Islam; Ekin Gedik; Hayley Hung
    License

    https://data.4tu.nl/info/fileadmin/user_upload/Documenten/4TU.ResearchData_Restricted_Data_2022.pdfhttps://data.4tu.nl/info/fileadmin/user_upload/Documenten/4TU.ResearchData_Restricted_Data_2022.pdf

    Description

    This file contains the annotations for the ConfLab dataset, including actions (speaking status), pose, and F-formations.

    ------------------

    ./actions/speaking_status:

    ./processed: the processed speaking status files, aggregated into a single data frame per segment. Skipped rows in the raw data (see https://josedvq.github.io/covfee/docs/output for details) have been imputed using the code at: https://github.com/TUDelft-SPC-Lab/conflab/tree/master/preprocessing/speaking_status

    The processed annotations consist of:

    ./speaking: The first row contains person IDs matching the sensor IDs,

    The rest of the row contains binary speaking status annotations at 60fps for the corresponding 2 min video segment (7200 frames).

    ./confidence: Same as above. These annotations reflect the continuous-valued rating of confidence of the annotators in their speaking annotation.

    To load these files with pandas: pd.read_csv(p, index_col=False)


    ./raw-covfee.zip: the raw outputs from speaking status annotation for each of the eight annotated 2-min video segments. These were were output by the covfee annotation tool (https://github.com/josedvq/covfee)

    Annotations were done at 60 fps.

    --------------------

    ./pose:

    ./coco: the processed pose files in coco JSON format, aggregated into a single data frame per video segment. These files have been generated from the raw files using the code at: https://github.com/TUDelft-SPC-Lab/conflab-keypoints

    To load in Python: f = json.load(open('/path/to/cam2_vid3_seg1_coco.json'))

    The skeleton structure (limbs) is contained within each file in:

    f['categories'][0]['skeleton']

    and keypoint names at:

    f['categories'][0]['keypoints']

    ./raw-covfee.zip: the raw outputs from continuous pose annotation. These were were output by the covfee annotation tool (https://github.com/josedvq/covfee)

    Annotations were done at 60 fps.

    ---------------------

    ./f_formations:

    seg 2: 14:00 onwards, for videos of the form x2xxx.MP4 in /video/raw/ for the relevant cameras (2,4,6,8,10).

    seg 3: for videos of the form x3xxx.MP4 in /video/raw/ for the relevant cameras (2,4,6,8,10).

    Note that camera 10 doesn't include meaningful subject information/body parts that are not already covered in camera 8.

    First column: time stamp

    Second column: "()" delineates groups, "<>" delineates subjects, cam X indicates the best camera view for which a particular group exists.


    phone.csv: time stamp (pertaining to seg3), corresponding group, ID of person using the phone

  20. Tomato Diseases for Object Detection

    • kaggle.com
    Updated Jun 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Iqbal Pahlevi A (2024). Tomato Diseases for Object Detection [Dataset]. https://www.kaggle.com/datasets/iqbalpahlevi/tomato-diseases-for-object-detection
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 1, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Iqbal Pahlevi A
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    This is the dataset for Tomato Diseases Object Detection. I take the original dataset from here and then annotate it manually by using label-studio library in Python.

    This dataset consists of several folders, each folder represent one tomato disease. And for each folder, there are two folders, images and annotations.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
phone dataset annotate (2024). Phone Data Annotate Dataset [Dataset]. https://universe.roboflow.com/phone-dataset-annotate/phone-data-annotate

Phone Data Annotate Dataset

phone-data-annotate

phone-data-annotate-dataset

Explore at:
zipAvailable download formats
Dataset updated
Aug 6, 2024
Dataset authored and provided by
phone dataset annotate
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Variables measured
Phone Hand Bounding Boxes
Description

Phone Data Annotate

## Overview

Phone Data Annotate is a dataset for object detection tasks - it contains Phone Hand annotations for 522 images.

## Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

  ## License

  This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Search
Clear search
Close search
Google apps
Main menu