100+ datasets found
  1. d

    TagX Data Annotation | Automated Annotation | AI-assisted labeling with...

    • datarade.ai
    Updated Aug 14, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    TagX (2022). TagX Data Annotation | Automated Annotation | AI-assisted labeling with human verification | Customized annotation | Data for AI & LLMs [Dataset]. https://datarade.ai/data-products/data-annotation-services-for-artificial-intelligence-and-data-tagx
    Explore at:
    .json, .xml, .csv, .xls, .txtAvailable download formats
    Dataset updated
    Aug 14, 2022
    Dataset authored and provided by
    TagX
    Area covered
    Saint Barthélemy, Sint Eustatius and Saba, Egypt, Lesotho, Central African Republic, Comoros, Georgia, Guatemala, Estonia, Cabo Verde
    Description

    TagX data annotation services are a set of tools and processes used to accurately label and classify large amounts of data for use in machine learning and artificial intelligence applications. The services are designed to be highly accurate, efficient, and customizable, allowing for a wide range of data types and use cases.

    The process typically begins with a team of trained annotators reviewing and categorizing the data, using a variety of annotation tools and techniques, such as text classification, image annotation, and video annotation. The annotators may also use natural language processing and other advanced techniques to extract relevant information and context from the data.

    Once the data has been annotated, it is then validated and checked for accuracy by a team of quality assurance specialists. Any errors or inconsistencies are corrected, and the data is then prepared for use in machine learning and AI models.

    TagX annotation services can be applied to a wide range of data types, including text, images, videos, and audio. The services can be customized to meet the specific needs of each client, including the type of data, the level of annotation required, and the desired level of accuracy.

    TagX data annotation services provide a powerful and efficient way to prepare large amounts of data for use in machine learning and AI applications, allowing organizations to extract valuable insights and improve their decision-making processes.

  2. f

    MAIA—A machine learning assisted image annotation method for environmental...

    • plos.figshare.com
    pdf
    Updated May 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Zurowietz; Daniel Langenkämper; Brett Hosking; Henry A. Ruhl; Tim W. Nattkemper (2023). MAIA—A machine learning assisted image annotation method for environmental monitoring and exploration [Dataset]. http://doi.org/10.1371/journal.pone.0207498
    Explore at:
    pdfAvailable download formats
    Dataset updated
    May 30, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Martin Zurowietz; Daniel Langenkämper; Brett Hosking; Henry A. Ruhl; Tim W. Nattkemper
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Digital imaging has become one of the most important techniques in environmental monitoring and exploration. In the case of the marine environment, mobile platforms such as autonomous underwater vehicles (AUVs) are now equipped with high-resolution cameras to capture huge collections of images from the seabed. However, the timely evaluation of all these images presents a bottleneck problem as tens of thousands or more images can be collected during a single dive. This makes computational support for marine image analysis essential. Computer-aided analysis of environmental images (and marine images in particular) with machine learning algorithms is promising, but challenging and different to other imaging domains because training data and class labels cannot be collected as efficiently and comprehensively as in other areas. In this paper, we present Machine learning Assisted Image Annotation (MAIA), a new image annotation method for environmental monitoring and exploration that overcomes the obstacle of missing training data. The method uses a combination of autoencoder networks and Mask Region-based Convolutional Neural Network (Mask R-CNN), which allows human observers to annotate large image collections much faster than before. We evaluated the method with three marine image datasets featuring different types of background, imaging equipment and object classes. Using MAIA, we were able to annotate objects of interest with an average recall of 84.1% more than twice as fast as compared to “traditional” annotation methods, which are purely based on software-supported direct visual inspection and manual annotation. The speed gain increases proportionally with the size of a dataset. The MAIA approach represents a substantial improvement on the path to greater efficiency in the annotation of large benthic image collections.

  3. P

    Data from: ImageNet Dataset

    • paperswithcode.com
    Updated Apr 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jia Deng; Wei Dong; Richard Socher; Li-Jia Li; Kai Li; Fei-Fei Li (2024). ImageNet Dataset [Dataset]. https://paperswithcode.com/dataset/imagenet
    Explore at:
    Dataset updated
    Apr 15, 2024
    Authors
    Jia Deng; Wei Dong; Richard Socher; Li-Jia Li; Kai Li; Fei-Fei Li
    Description

    The ImageNet dataset contains 14,197,122 annotated images according to the WordNet hierarchy. Since 2010 the dataset is used in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC), a benchmark in image classification and object detection. The publicly released dataset contains a set of manually annotated training images. A set of test images is also released, with the manual annotations withheld. ILSVRC annotations fall into one of two categories: (1) image-level annotation of a binary label for the presence or absence of an object class in the image, e.g., “there are cars in this image” but “there are no tigers,” and (2) object-level annotation of a tight bounding box and class label around an object instance in the image, e.g., “there is a screwdriver centered at position (20,25) with width of 50 pixels and height of 30 pixels”. The ImageNet project does not own the copyright of the images, therefore only thumbnails and URLs of images are provided.

    Total number of non-empty WordNet synsets: 21841 Total number of images: 14197122 Number of images with bounding box annotations: 1,034,908 Number of synsets with SIFT features: 1000 Number of images with SIFT features: 1.2 million

  4. Data for the evaluation of the MAIA method for image annotation

    • zenodo.org
    • eprints.soton.ac.uk
    csv
    Updated Jan 24, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Zurowietz; Martin Zurowietz; Daniel Langenkämper; Daniel Langenkämper; Brett Hosking; Brett Hosking; Henry A Ruhl; Tim W Nattkemper; Henry A Ruhl; Tim W Nattkemper (2020). Data for the evaluation of the MAIA method for image annotation [Dataset]. http://doi.org/10.5281/zenodo.1453836
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Martin Zurowietz; Martin Zurowietz; Daniel Langenkämper; Daniel Langenkämper; Brett Hosking; Brett Hosking; Henry A Ruhl; Tim W Nattkemper; Henry A Ruhl; Tim W Nattkemper
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains all annotations and annotation candidates that were used for the evaluation of the MAIA method for image annotation. Each row in the CSVs represents one annotation candidate or final annotation. Annotation candidates have the label "OOI candidate" (label_id 9974). All other entries represent final reviewed annotations. Each CSV contains the information for one of the three image datasets that were used in the evaluation.

    Visual exploration of the data is possible in the BIIGLE 2.0 image annotation system at https://biigle.de/projects/139 using the login maia@example.com and the password MAIApaper.

  5. n

    Data from: Efficient Deep Learning Methods for Medical Image Analysis

    • curate.nd.edu
    pdf
    Updated Nov 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yaopeng Peng (2024). Efficient Deep Learning Methods for Medical Image Analysis [Dataset]. http://doi.org/10.7274/27147567.v1
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Nov 11, 2024
    Dataset provided by
    University of Notre Dame
    Authors
    Yaopeng Peng
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    Medical image analysis plays a critical role in a range of medical applications, including diagnosis, treatment planning, and monitoring disease progression. However, it presents significant challenges due to the inherent complexity of the human body, as well as variability in image acquisition techniques, noise, and artifacts.

    Although deep learning methods have demonstrated considerable promise in medical image analysis, they frequently necessitate large volumes of annotated data for effective model training. Acquiring such annotated data can be particularly challenging in medical imaging due to factors such as the complexity of medical images and the imperative to uphold patient privacy. Furthermore, the annotation process is both time-consuming and costly, requiring the specialized expertise of medical professionals. Consequently, the limited availability of annotated data for training deep learning models often results in overfitting and suboptimal generalization to new data.

    Advances in medical image analysis have benefited from progress in foundational models originally developed for natural image domains. Innovations such as the integration of topological features into image representations and the application of Vision Transformers (ViTs) to capture global dependencies have proven valuable. However, these models often face significant challenges, including high computational costs and inference latency. Thus, there is an urgent need to develop approaches that are both data-efficient and computationally efficient to overcome these limitations. This dissertation presents six methods designed to improve segmentation and classification performance across both medical and natural scene domains. These methods include selecting the most informative slices for annotation, utilizing unlabeled slices, extracting additional topological information from existing datasets, and developing efficient Vision Transformer models to enhance performance while reducing computational costs.

    First, we employ an unsupervised method to identify the most effective and representative 2D slices from 3D calf muscle images for annotation. Subsequently, we generate pseudo-labels for all unlabeled slices and train a 3D segmentation model using both the labeled and pseudo-labeled slices. Second, we enhance the model by refining the pseudo-labels with a bi-directional hierarchical Earth Mover's Distance (bi-HEMD) algorithm and fine-tuning the segmentation results using the Primal-Dual Interior Point Method (IPM). Third, we develop a method that integrates both topological features and features extracted by a convolutional neural network (CNN) to improve performance. Fourth, we introduce a Group Vision Transformer mechanism to reduce computational complexity and model parameters, while enhancing feature diversity and reducing feature redundancy. Finally, we develop two Vision Transformer models to improve segmentation performance for detecting thin-cap fibroatheroma (TCFA) in intravascular optical coherence tomography (IVOCT) images and for skin lesion and polyp segmentation.

    The performance of image recognition in both medical and natural domains can be further enhanced by developing more advanced models. Accordingly, we propose four promising future directions. First, we aim to utilize the Wavelet Transform to mitigate information loss during the down-sampling process, thereby improving detection of small objects. Second, we plan to develop a Multi-Branch Vision Transformer to capture features across various scales while reducing computational costs and inference latency. Third, we intend to create a hierarchical Hilbert Mamba framework for image recognition, which will introduce greater spatial locality and facilitate smoother transitions among image tokens. Finally, we propose to develop a semi-supervised model for medical image segmentation, based on the Segment Anything Model, to address challenges associated with sparse annotations.

  6. n

    Data from: New Deep Learning Methods for Medical Image Analysis and...

    • curate.nd.edu
    pdf
    Updated Nov 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pengfei Gu (2024). New Deep Learning Methods for Medical Image Analysis and Scientific Data Generation and Compression [Dataset]. http://doi.org/10.7274/26156719.v1
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Nov 11, 2024
    Dataset provided by
    University of Notre Dame
    Authors
    Pengfei Gu
    License

    https://www.law.cornell.edu/uscode/text/17/106https://www.law.cornell.edu/uscode/text/17/106

    Description

    Medical image analysis is critical to biological studies, health research, computer- aided diagnoses, and clinical applications. Recently, deep learning (DL) techniques have achieved remarkable successes in medical image analysis applications. However, these techniques typically require large amounts of annotations to achieve satisfactory performance. Therefore, in this dissertation, we seek to address this critical problem: How can we develop efficient and effective DL algorithms for medical image analysis while reducing annotation efforts? To address this problem, we have outlined two specific aims: (A1) Utilize existing annotations effectively from advanced models; (A2) extract generic knowledge directly from unannotated images.

    To achieve the aim (A1): First, we introduce a new data representation called TopoImages, which encodes the local topology of all the image pixels. TopoImages can be complemented with the original images to improve medical image analysis tasks. Second, we propose a new augmentation method, SAMAug-C, that lever- ages the Segment Anything Model (SAM) to augment raw image input and enhance medical image classification. Third, we propose two advanced DL architectures, kCBAC-Net and ConvFormer, to enhance the performance of 2D and 3D medical image segmentation. We also present a gate-regularized network training (GrNT) approach to improve multi-scale fusion in medical image segmentation. To achieve the aim (A2), we propose a novel extension of known Masked Autoencoders (MAEs) for self pre-training, i.e., models pre-trained on the same target dataset, specifically for 3D medical image segmentation.

    Scientific visualization is a powerful approach for understanding and analyzing various physical or natural phenomena, such as climate change or chemical reactions. However, the cost of scientific simulations is high when factors like time, ensemble, and multivariate analyses are involved. Additionally, scientists can only afford to sparsely store the simulation outputs (e.g., scalar field data) or visual representations (e.g., streamlines) or visualization images due to limited I/O bandwidths and storage space. Therefore, in this dissertation, we seek to address this critical problem: How can we develop efficient and effective DL algorithms for scientific data generation and compression while reducing simulation and storage costs?

    To tackle this problem: First, we propose a DL framework that generates un- steady vector fields data from a set of streamlines. Based on this method, domain scientists only need to store representative streamlines at simulation time and recon- struct vector fields during post-processing. Second, we design a novel DL method that translates scalar fields to vector fields. Using this approach, domain scientists only need to store scalar field data at simulation time and generate vector fields from their scalar field counterparts afterward. Third, we present a new DL approach that compresses a large collection of visualization images generated from time-varying data for communicating volume visualization results.

  7. R

    Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges

    • public.roboflow.com
    zip
    Updated Sep 30, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Northeastern University - China (2022). Hard Hat Workers Object Detection Dataset - resize-416x416-reflectEdges [Dataset]. https://public.roboflow.com/object-detection/hard-hat-workers/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 30, 2022
    Dataset authored and provided by
    Northeastern University - China
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Bounding Boxes of Workers
    Description

    Overview

    The Hard Hat dataset is an object detection dataset of workers in workplace settings that require a hard hat. Annotations also include examples of just "person" and "head," for when an individual may be present without a hard hart.

    The original dataset has a 75/25 train-test split.

    Example Image: https://i.imgur.com/7spoIJT.png" alt="Example Image">

    Use Cases

    One could use this dataset to, for example, build a classifier of workers that are abiding safety code within a workplace versus those that may not be. It is also a good general dataset for practice.

    Using this Dataset

    Use the fork or Download this Dataset button to copy this dataset to your own Roboflow account and export it with new preprocessing settings (perhaps resized for your model's desired format or converted to grayscale), or additional augmentations to make your model generalize better. This particular dataset would be very well suited for Roboflow's new advanced Bounding Box Only Augmentations.

    Dataset Versions:

    Image Preprocessing | Image Augmentation | Modify Classes * v1 (resize-416x416-reflect): generated with the original 75/25 train-test split | No augmentations * v2 (raw_75-25_trainTestSplit): generated with the original 75/25 train-test split | These are the raw, original images * v3 (v3): generated with the original 75/25 train-test split | Modify Classes used to drop person class | Preprocessing and Augmentation applied * v5 (raw_HeadHelmetClasses): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class * v8 (raw_HelmetClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and person classes * v9 (raw_PersonClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop head and helmet classes * v10 (raw_AllClasses): generated with a 70/20/10 train/valid/test split | These are the raw, original images * v11 (augmented3x-AllClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied | 3x image generation | Trained with Roboflow's Fast Model * v12 (augmented3x-HeadHelmetClasses-FastModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Fast Model * v13 (augmented3x-HeadHelmetClasses-AccurateModel): generated with a 70/20/10 train/valid/test split | Preprocessing and Augmentation applied, Modify Classes used to drop person class | 3x image generation | Trained with Roboflow's Accurate Model * v14 (raw_HeadClassOnly): generated with a 70/20/10 train/valid/test split | Modify Classes used to drop person class, and remap/relabel helmet class to head

    Choosing Between Computer Vision Model Sizes | Roboflow Train

    About Roboflow

    Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

    Developers reduce 50% of their code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

    Roboflow Workmark

  8. Data from: Improving automated annotation of benthic survey images using...

    • zenodo.org
    • data.niaid.nih.gov
    • +1more
    bin, txt
    Updated May 31, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Oscar Beijbom; Tali Treibitz; David I. Kline; Gal Eyal; Adi Khen; Benjamin Neal; Yossi Loya; B. Greg Mitchell; David Kriegman; Oscar Beijbom; Tali Treibitz; David I. Kline; Gal Eyal; Adi Khen; Benjamin Neal; Yossi Loya; B. Greg Mitchell; David Kriegman (2022). Data from: Improving automated annotation of benthic survey images using wide-band fluorescence [Dataset]. http://doi.org/10.5061/dryad.t4362
    Explore at:
    bin, txtAvailable download formats
    Dataset updated
    May 31, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Oscar Beijbom; Tali Treibitz; David I. Kline; Gal Eyal; Adi Khen; Benjamin Neal; Yossi Loya; B. Greg Mitchell; David Kriegman; Oscar Beijbom; Tali Treibitz; David I. Kline; Gal Eyal; Adi Khen; Benjamin Neal; Yossi Loya; B. Greg Mitchell; David Kriegman
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Large-scale imaging techniques are used increasingly for ecological surveys. However, manual analysis can be prohibitively expensive, creating a bottleneck between collected images and desired data-products. This bottleneck is particularly severe for benthic surveys, where millions of images are obtained each year. Recent automated annotation methods may provide a solution, but reflectance images do not always contain sufficient information for adequate classification accuracy. In this work, the FluorIS, a low-cost modified consumer camera, was used to capture wide-band wide-field-of-view fluorescence images during a field deployment in Eilat, Israel. The fluorescence images were registered with standard reflectance images, and an automated annotation method based on convolutional neural networks was developed. Our results demonstrate a 22% reduction of classification error-rate when using both images types compared to only using reflectance images. The improvements were large, in particular, for coral reef genera Platygyra, Acropora and Millepora, where classification recall improved by 38%, 33%, and 41%, respectively. We conclude that convolutional neural networks can be used to combine reflectance and fluorescence imagery in order to significantly improve automated annotation accuracy and reduce the manual annotation bottleneck.

  9. Image Dataset of Accessibility Barriers

    • zenodo.org
    zip
    Updated Mar 25, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jakob Stolberg; Jakob Stolberg (2022). Image Dataset of Accessibility Barriers [Dataset]. http://doi.org/10.5281/zenodo.6382090
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 25, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jakob Stolberg; Jakob Stolberg
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The Data
    The dataset consist of 5538 images of public spaces, annotated with steps, stairs, ramps and grab bars for stairs and ramps. The dataset has annotations 3564 of steps, 1492 of stairs, 143 of ramps and 922 of grab bars.

    Each step annotation is attributed with an estimate of the height of the step, as falling into one of three categories: less than 3cm, 3cm to 7cm or more than 7cm. Additionally it is attributed with a 'type', with the possibilities 'doorstep', 'curb' or 'other'.

    Stair annotations are attributed with the number of steps in the stair.

    Ramps are attributed with an estimate of their width, also falling into three categories: less than 50cm, 50cm to 100cm and more than 100cm.

    In order to preserve all additional attributes of the labels, the data is published in the CVAT XML format for images.

    Annotating Process
    The labelling has been done using bounding boxes around the objects. This format is compatible with many popular object detection models, e.g. the YOLO object model. A bounding box is placed so it contains exactly the visible part of the respective objects. This implies that only objects that are visible in the photo are annotated. This means in particular a photo of a stair or step from above, where the object cannot be seen, have not been annotated, even when a human viewer can possibly infer that there is a stair or a step from other features in the photo.

    Steps
    A step is annotated, when there is an vertical increment that functions as a passage between two surface areas intended human or vehicle traffic. This means that we have not included:

    • Increments that are to high to reasonably be considered at passage.
    • Increments that does not lead to a surface intended for human or vehicle traffic, e.g. a 'step' in front of a wall or a curb in front of a bush.

    In particular, the bounding box of a step object contains exactly the incremental part of the step, but does not extend into the top or bottom horizontal surface any more than necessary to enclose entirely the incremental part. This has been chosen for consistency reasons, as including parts of the horizontal surfaces would imply a non-trivial choice of how much to include, which we deemed would most likely lead to more inconstistent annotations.

    The height of the steps are estimated by the annotators, and are therefore not guarranteed to be accurate.

    The type of the steps typically fall into the category 'doorstep' or 'curb'. Steps that are in a doorway, entrance or likewise are attributed as doorsteps. We also include in this category steps that are immediately leading to a doorway within a proximity of 1-2m. Steps between different types of pathways, e.g. between streets and sidewalks, are annotated as curbs. Any other type of step are annotated with 'other'. Many of the 'other' steps are for example steps to terraces.

    Stairs
    The stair label is used whenever two or more steps directly follow each other in a consistent pattern. All vertical increments are enclosed in the bounding box, as well as intermediate surfaces of the steps. However the top and bottom surface is not included more than necessary for the same reason as for steps, as described in the previous section.

    The annotator counts the number of steps, and attribute this to the stair object label.

    Ramps
    Ramps have been annotated when a sloped passage way has been placed or built to connect two surface areas intended for human or vehicle traffic. This implies the same considerations as with steps. Alike also only the sloped part of a ramp is annotated, not including the bottom or top surface area.

    For each ramp, the annotator makes an assessment of the width of the ramp in three categories: less than 50cm, 50cm to 100cm and more than 100cm. This parameter is visually hard to assess, and sometimes impossible due to the view of the ramp.

    Grab Bars
    Grab bars are annotated for hand rails and similar that are in direct connection to a stair or a ramp. While horizontal grab bars could also have been included, this was omitted due to the implied ambiguities of fences and similar objects. As the grab bar was originally intended as an attributal information to stairs and ramps, we chose to keep this focus. The bounding box encloses the part of the grab bar that functions as a hand rail for the stair or ramp.

    Usage
    As is often the case when annotating data, much information depends on the subjective assessment of the annotator. As each data point in this dataset has been annotated only by one person, caution should be taken if the data is applied.

    Generally speaking, the mindset and usage guiding the annotations have been wheelchair accessibility. While we have strived to annotate at an object level, hopefully making the data more widely applicable than this, we state this explicitly as it may have swayed untrivial annotation choices.

    The attributal data, such as step height or ramp width are highly subjective estimations. We still provide this data to give a post-hoc method to adjust which annotations to use. E.g. for some purposes, one may be interested in detecting only steps that are indeed more than 3cm. The attributal data makes it possible to sort away the steps less than 3cm, so a machine learning algorithm can be trained on this more appropriate dataset for that use case. We stress however, that one cannot expect to train accurate machine learning algorithms inferring the attributal data, as this is not accurate data in the first place.

    We hope this dataset will be a useful building block in the endeavours for automating barrier detection and documentation.

  10. Detection performance of the trained Mask R-CNN model on each validation...

    • plos.figshare.com
    • figshare.com
    xls
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Zurowietz; Daniel Langenkämper; Brett Hosking; Henry A. Ruhl; Tim W. Nattkemper (2023). Detection performance of the trained Mask R-CNN model on each validation subset VΓ. [Dataset]. http://doi.org/10.1371/journal.pone.0207498.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Martin Zurowietz; Daniel Langenkämper; Brett Hosking; Henry A. Ruhl; Tim W. Nattkemper
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Detection performance of the trained Mask R-CNN model on each validation subset VΓ.

  11. T

    open_images_v4

    • tensorflow.org
    • paperswithcode.com
    • +1more
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    open_images_v4 [Dataset]. https://www.tensorflow.org/datasets/catalog/open_images_v4
    Explore at:
    Description

    Open Images is a dataset of ~9M images that have been annotated with image-level labels and object bounding boxes.

    The training set of V4 contains 14.6M bounding boxes for 600 object classes on 1.74M images, making it the largest existing dataset with object location annotations. The boxes have been largely manually drawn by professional annotators to ensure accuracy and consistency. The images are very diverse and often contain complex scenes with several objects (8.4 per image on average). Moreover, the dataset is annotated with image-level labels spanning thousands of classes.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('open_images_v4', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/open_images_v4-original-2.0.0.png" alt="Visualization" width="500px">

  12. MCR LTER: Coral Reef: Computer Vision: Multi-annotator Comparison of Coral...

    • search.dataone.org
    • portal.edirepository.org
    Updated Apr 30, 2015
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Moorea Coral Reef LTER; Peter Edmunds (2015). MCR LTER: Coral Reef: Computer Vision: Multi-annotator Comparison of Coral Photo Quadrat Analysis [Dataset]. https://search.dataone.org/view/https%3A%2F%2Fpasta.lternet.edu%2Fpackage%2Fmetadata%2Feml%2Fknb-lter-mcr%2F5013%2F3
    Explore at:
    Dataset updated
    Apr 30, 2015
    Dataset provided by
    Long Term Ecological Research Networkhttp://www.lternet.edu/
    Authors
    Moorea Coral Reef LTER; Peter Edmunds
    Time period covered
    Jan 1, 2008
    Area covered
    Description

    This repository contains the Moorea portion of a larger data package published in conjuncture with: "Towards automated annotation of benthic survey images: variability of human experts and operational modes of automation", Beijbom et al. PLOS One, 2015.

      The rest of the data package is hosted at the Dryad data repository (doi:10.5061/dryad.m5pr3).
    
      The larger data package is an aggregate dataset from four Pacific coral reef monitoring projects in: Moorea (French Polynesia), 
      the northern Line Islands, Nanwan Bay (Taiwan) and Heron Reef (Australia). It contains 5090 coral reef survey images, 
      and 251,988 random-point annotations by coral ecology experts. The point-annotations indicate the dominant benthic 
      substrate at 10 to 200 random point locations per image, using a label-set of 20 categories. In addition, 200 images from each 
      location have been cross-annotated by 6 experts, for a total of 7 sets of annotations for each image. This set of cross-annotations 
      can be used to contextualize the performance of automated annotation methods for coral reef ecology. The full data package can 
      also be used by computer-vision and machine learning researchers to develop object classification, image segmentation, and 
      domain transfer learning methods. These data contain a subset of the raw data from which dataset knb-lter-mcr.4 is derived.
    
  13. Data from: Region-based Annotation Data of Fire Images for Intelligent...

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Jan 23, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata; Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata (2022). Region-based Annotation Data of Fire Images for Intelligent Surveillance System [Dataset]. http://doi.org/10.5281/zenodo.5574537
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 23, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata; Wahyono; Andi Dharmawan; Agus Harjoko; Chrystian; Faisal Dharma Adhinata
    Description

    This data presents fire segmentation annotation data on 12 commonly used and publicly available “VisiFire Dataset” videos from http://signal.ee.bilkent.edu.tr/VisiFire/. This annotations dataset was obtained by per-frame, manual hand annotation over the fire region with 2,684 total annotated frames. Since this annotation provides per-frame segmentation data, it offers a new and unique fire motion feature to the existing video, unlike other fire segmentation data that are collected from different still images. The annotations dataset also provides ground truth for segmentation task on videos. With segmentation task, it offers better insight on how well a machine learning model understood, not only detecting whether a fire is present, but also its exact location by calculating metrics such as Intersection over Union (IoU) with this annotations data. This annotations data is a tremendously useful addition to train, develop, and create a much better smart surveillance system for early detection in high-risk fire hotspots area.

  14. Training datasets and final models from paper ''RootPainter: Deep Learning...

    • zenodo.org
    • explore.openaire.eu
    • +1more
    zip
    Updated Apr 16, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abraham Goerge Smith; Eusun Han; Jens Petersen; Niels Alvin Faircloth Olsen; Christian Giese; Miriam Athmann; Dorte Bodin Dresbøll; Kristian Thorup-Kristensen; Abraham Goerge Smith; Eusun Han; Jens Petersen; Niels Alvin Faircloth Olsen; Christian Giese; Miriam Athmann; Dorte Bodin Dresbøll; Kristian Thorup-Kristensen (2020). Training datasets and final models from paper ''RootPainter: Deep Learning Segmentation of Biological Images with Corrective Annotation' [Dataset]. http://doi.org/10.5281/zenodo.3754046
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 16, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Abraham Goerge Smith; Eusun Han; Jens Petersen; Niels Alvin Faircloth Olsen; Christian Giese; Miriam Athmann; Dorte Bodin Dresbøll; Kristian Thorup-Kristensen; Abraham Goerge Smith; Eusun Han; Jens Petersen; Niels Alvin Faircloth Olsen; Christian Giese; Miriam Athmann; Dorte Bodin Dresbøll; Kristian Thorup-Kristensen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    See paper ''RootPainter: Deep Learning Segmentation of Biological Images with Corrective Annotation' for an explanation of how the models and datasets were created.

    The images are extracted from the following larger datasets using the RootPainter software:

    Nodules: http://doi.org/10.5281/zenodo.3753603

    Biopores: http://doi.org/10.5281/zenodo.3753969

    Roots: http://doi.org/10.5281/zenodo.3527713

  15. P

    MS COCO Dataset

    • paperswithcode.com
    Updated Apr 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tsung-Yi Lin; Michael Maire; Serge Belongie; Lubomir Bourdev; Ross Girshick; James Hays; Pietro Perona; Deva Ramanan; C. Lawrence Zitnick; Piotr Dollár, MS COCO Dataset [Dataset]. https://paperswithcode.com/dataset/coco
    Explore at:
    Dataset updated
    Apr 15, 2024
    Authors
    Tsung-Yi Lin; Michael Maire; Serge Belongie; Lubomir Bourdev; Ross Girshick; James Hays; Pietro Perona; Deva Ramanan; C. Lawrence Zitnick; Piotr Dollár
    Description

    The MS COCO (Microsoft Common Objects in Context) dataset is a large-scale object detection, segmentation, key-point detection, and captioning dataset. The dataset consists of 328K images.

    Splits: The first version of MS COCO dataset was released in 2014. It contains 164K images split into training (83K), validation (41K) and test (41K) sets. In 2015 additional test set of 81K images was released, including all the previous test images and 40K new images.

    Based on community feedback, in 2017 the training/validation split was changed from 83K/41K to 118K/5K. The new split uses the same images and annotations. The 2017 test set is a subset of 41K images of the 2015 test set. Additionally, the 2017 release contains a new unannotated dataset of 123K images.

    Annotations: The dataset has annotations for

    object detection: bounding boxes and per-instance segmentation masks with 80 object categories, captioning: natural language descriptions of the images (see MS COCO Captions), keypoints detection: containing more than 200,000 images and 250,000 person instances labeled with keypoints (17 possible keypoints, such as left eye, nose, right hip, right ankle), stuff image segmentation – per-pixel segmentation masks with 91 stuff categories, such as grass, wall, sky (see MS COCO Stuff), panoptic: full scene segmentation, with 80 thing categories (such as person, bicycle, elephant) and a subset of 91 stuff categories (grass, sky, road), dense pose: more than 39,000 images and 56,000 person instances labeled with DensePose annotations – each labeled person is annotated with an instance id and a mapping between image pixels that belong to that person body and a template 3D model. The annotations are publicly available only for training and validation images.

  16. c

    Annotations for ACRIN-HNSCC-FDG-PET-CT Collection

    • cancerimagingarchive.net
    • stage.cancerimagingarchive.net
    csv, dicom, n/a
    Updated Dec 23, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Cancer Imaging Archive (2022). Annotations for ACRIN-HNSCC-FDG-PET-CT Collection [Dataset]. http://doi.org/10.7937/JVGC-AQ36
    Explore at:
    n/a, csv, dicomAvailable download formats
    Dataset updated
    Dec 23, 2022
    Dataset authored and provided by
    The Cancer Imaging Archive
    License

    https://www.cancerimagingarchive.net/data-usage-policies-and-restrictions/https://www.cancerimagingarchive.net/data-usage-policies-and-restrictions/

    Time period covered
    Nov 13, 2023
    Dataset funded by
    National Cancer Institutehttp://www.cancer.gov/
    Description

    This dataset contains image annotations derived from the NCI Clinical Trial "ACRIN-HNSCC-FDG-PET-CT (ACRIN 6685)”. This dataset was generated as part of an NCI project to augment TCIA datasets with annotations that will improve their value for cancer researchers and AI developers.

    Annotation Protocol

    For each patient, all scans were reviewed to identify and annotate the clinically relevant time points and sequences/series. Scans were initially annotated by an international team of radiologists holding MBBS degrees or higher, which were then reviewed by US-based board-certified radiologists to ensure accuracy. In a typical patient all available time points were annotated. The following annotation rules were followed:
    1. PERCIST criteria was followed for PET imaging. Specifically, the lesions estimated to have the most elevated SUVmax were annotated.
    2. RECIST 1.1 was otherwise generally followed for MR and CT imaging. A maximum of 5 lesions were annotated per patient scan (timepoint); no more than 2 per organ. The same 5 lesions were annotated at each time point. Lymph nodes were annotated if >1 cm in short axis. Other lesions were annotated if >1 cm. If the primary lesion is < 1 cm, it was still annotated.
    3. Three-dimensional segmentations of lesions were created in the axial plane. If no axial plane was available, lesions were annotated in the coronal plane.
    4. MRIs were annotated using the T1-weighted axial post contrast sequence, fat saturated if available.
    5. CTs were annotated using the axial post contrast series. If not available, the non contrast series were annotated.
    6. PET/CTs were annotated on the CT and attenuation corrected PET images.
    7. If the post contrast CT was performed the same day as the PET/CT, the non contrast CT portion of the PET/CT was not annotated.
    8. Lesions were labeled separately.
    9. The volume of each annotated lesion was calculated and reported in cubic centimeters [cc] in the Annotation Metadata CSV.
    10. Seed points were automatically generated, but reviewed by a radiologist.
    11. A “negative” annotation was created for any exam without findings.
    At each time point:
    1. A seed point (kernel) was created for each segmented structure. The seed points for each segmentation are provided in a separate DICOM RTSTRUCT file.
    2. SNOMED-CT “Anatomic Region Sequence” and “Segmented Property Category Code Sequence” and codes were inserted for all segmented structures.
    3. “Tracking ID” and “Tracking UID” tags were inserted for each segmented structure to enable longitudinal lesion tracking.
    4. Imaging time point codes were inserted to help identify each annotation in the context of the clinical trial assessment protocol.
      1. “Clinical Trial Time Point ID” was used to encode time point type using one of the following strings as applicable: “pre-dose” or “post-chemotherapy”.
      2. Content Item in “Acquisition Context Sequence” was added containing "Time Point Type" using Concept Code Sequence (0040,A168) selected from:
        1. (255235001, SCT, “Pre-dose”) (in this trial, both the CT/MRI and PET/CT, while being different timepoints, are pre-treatment)

    Important supplementary information and sample code

    1. A spreadsheet containing key details about the annotations is available in the Data Access section below.
    2. A Jupyter notebook demonstrating how to use the NBIA Data Retriever Command-Line Interface application and the REST API to access these data can be found in the Additional Resources section below.

  17. d

    Open Images Dataset V4

    • dataportal.asia
    csv, txt, zip
    Updated Sep 16, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    scidm.nchc.org.tw (2021). Open Images Dataset V4 [Dataset]. https://dataportal.asia/id/dataset/212582112_open-images
    Explore at:
    csv(1194033454), zip, csv(475854), csv(45227339), csv(3048336194), txt(71460), csv(1229488697), csv(638407721), csv(216137230), csv(17074036), csv(26460276), csv(52174204), csv(19801709), csv(59773452), csv(5369002), csv(69570530), csv(376764810), csv(2068622216), csv(23008122), csv(8792832), csv(11255), csv(15428650)Available download formats
    Dataset updated
    Sep 16, 2021
    Dataset provided by
    scidm.nchc.org.tw
    Description

    Open Images V4

    https://storage.googleapis.com/openimages/web/index.html

    Open Images is a dataset of ~9 million images that have been annotated with image-level labels and object bounding boxes. The training set of V4 contains 14.6M bounding boxes for 600 object classes on 1.74M images, making it the largest existing dataset with object location annotations. The boxes have been largely manually drawn by professional annotators to ensure accuracy and consistency. The images are very diverse and often contain complex scenes with several objects (8.4 per image on average). Moreover, the dataset is annotated with image-level labels spanning thousands of classes.

    Licenses

    The annotations are licensed by Google Inc. under CC BY 4.0 license. The images are listed as having a CC BY 2.0 license. Note: while we tried to identify images that are licensed under a Creative Commons Attribution license, we make no representations or warranties regarding the license status of each image and you should verify the license for each image yourself.

  18. D

    Mapillary Vistas Dataset

    • datasetninja.com
    Updated Dec 8, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gerhard Neuhold; Tobias Ollmann; Samuel Rota Bulo (2020). Mapillary Vistas Dataset [Dataset]. https://datasetninja.com/mapillary-vistas-dataset
    Explore at:
    Dataset updated
    Dec 8, 2020
    Dataset provided by
    Dataset Ninja
    Authors
    Gerhard Neuhold; Tobias Ollmann; Samuel Rota Bulo
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    The Mapillary Vistas Dataset is a substantial, street-level image dataset containing 25,000 high-resolution images annotated across 124 classes (70 instance-specific, 46 stuff, 8 void or crowd). Annotation adopts a dense, fine-grained style using polygons to delineate individual objects. The dataset, authored by the dataset creators, is notably larger, by a factor of 5, than the combined fine annotations in Cityscapes. It encompasses images captured worldwide, encompassing diverse weather, seasonal, and daytime conditions. The images are sourced from various devices such as mobile phones, tablets, action cameras, professional capturing rigs, and different experienced photographers, thus embracing diversity, detail richness, and global coverage.

  19. S

    Data from: Towards automated annotation of benthic survey images:...

    • data.subak.org
    • data.niaid.nih.gov
    • +2more
    csv
    Updated Feb 16, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    University of California, San Diego (2023). Data from: Towards automated annotation of benthic survey images: variability of human experts and operational modes of automation [Dataset]. https://data.subak.org/dataset/data-from-towards-automated-annotation-of-benthic-survey-images-variability-of-human-experts-an
    Explore at:
    csvAvailable download formats
    Dataset updated
    Feb 16, 2023
    Dataset provided by
    University of California, San Diego
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Global climate change and other anthropogenic stressors have heightened the need to rapidly characterize ecological changes in marine benthic communities across large scales. Digital photography enables rapid collection of survey images to meet this need, but the subsequent image annotation is typically a time consuming, manual task. We investigated the feasibility of using automated point-annotation to expedite cover estimation of the 17 dominant benthic categories from survey-images captured at four Pacific coral reefs. Inter- and intra- annotator variability among six human experts was quantified and compared to semi- and fully- automated annotation methods, which are made available at coralnet.ucsd.edu. Our results indicate high expert agreement for identification of coral genera, but lower agreement for algal functional groups, in particular between turf algae and crustose coralline algae. This indicates the need for unequivocal definitions of algal groups, careful training of multiple annotators, and enhanced imaging technology. Semi-automated annotation, where 50% of the annotation decisions were performed automatically, yielded cover estimate errors comparable to those of the human experts. Furthermore, fully-automated annotation yielded rapid, unbiased cover estimates but with increased variance. These results show that automated annotation can increase spatial coverage and decrease time and financial outlay for image-based reef surveys.

  20. D

    Data Annotation Tool Market Report

    • marketresearchforecast.com
    doc, pdf, ppt
    Updated Dec 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Market Research Forecast (2024). Data Annotation Tool Market Report [Dataset]. https://www.marketresearchforecast.com/reports/data-annotation-tool-market-10075
    Explore at:
    doc, ppt, pdfAvailable download formats
    Dataset updated
    Dec 9, 2024
    Dataset authored and provided by
    Market Research Forecast
    License

    https://www.marketresearchforecast.com/privacy-policyhttps://www.marketresearchforecast.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The size of the Data Annotation Tool Market market was valued at USD 3.9 USD billion in 2023 and is projected to reach USD 6.64 USD billion by 2032, with an expected CAGR of 7.9% during the forecast period. A Data Annotation Tool is defined as the software that can be employed to make annotations to data hence helping a learning computer model learn patterns. These tools provide a way of segregating the data types to include images, texts, and audio, as well as videos. Some of the subcategories of annotation include images such as bounding boxes, segmentation, text such as entity recognition, sentiment analysis, audio such as transcription, sound labeling, and video such as object tracking. Other common features depend on the case but they commonly consist of interfaces, cooperation with others, suggestion of labels, and quality assurance. It can be used in the automotive industry (object detection for self-driving cars), text processing (classification of text), healthcare (medical imaging), and retail (recommendation). These tools get applied in training good quality, accurately labeled data sets for the engineering of efficient AI systems. Key drivers for this market are: Increasing Adoption of Cloud-based Managed Services to Drive Market Growth. Potential restraints include: Adverse Health Effect May Hamper Market Growth. Notable trends are: Growing Implementation of Touch-based and Voice-based Infotainment Systems to Increase Adoption of Intelligent Cars.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
TagX (2022). TagX Data Annotation | Automated Annotation | AI-assisted labeling with human verification | Customized annotation | Data for AI & LLMs [Dataset]. https://datarade.ai/data-products/data-annotation-services-for-artificial-intelligence-and-data-tagx

TagX Data Annotation | Automated Annotation | AI-assisted labeling with human verification | Customized annotation | Data for AI & LLMs

Explore at:
.json, .xml, .csv, .xls, .txtAvailable download formats
Dataset updated
Aug 14, 2022
Dataset authored and provided by
TagX
Area covered
Saint Barthélemy, Sint Eustatius and Saba, Egypt, Lesotho, Central African Republic, Comoros, Georgia, Guatemala, Estonia, Cabo Verde
Description

TagX data annotation services are a set of tools and processes used to accurately label and classify large amounts of data for use in machine learning and artificial intelligence applications. The services are designed to be highly accurate, efficient, and customizable, allowing for a wide range of data types and use cases.

The process typically begins with a team of trained annotators reviewing and categorizing the data, using a variety of annotation tools and techniques, such as text classification, image annotation, and video annotation. The annotators may also use natural language processing and other advanced techniques to extract relevant information and context from the data.

Once the data has been annotated, it is then validated and checked for accuracy by a team of quality assurance specialists. Any errors or inconsistencies are corrected, and the data is then prepared for use in machine learning and AI models.

TagX annotation services can be applied to a wide range of data types, including text, images, videos, and audio. The services can be customized to meet the specific needs of each client, including the type of data, the level of annotation required, and the desired level of accuracy.

TagX data annotation services provide a powerful and efficient way to prepare large amounts of data for use in machine learning and AI applications, allowing organizations to extract valuable insights and improve their decision-making processes.

Search
Clear search
Close search
Google apps
Main menu