100+ datasets found
  1. i

    Human Annotated GAN Images

    • ieee-dataport.org
    Updated Apr 25, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Juanyong Duan (2019). Human Annotated GAN Images [Dataset]. https://ieee-dataport.org/documents/human-annotated-gan-images
    Explore at:
    Dataset updated
    Apr 25, 2019
    Authors
    Juanyong Duan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    we have collected an annotated dataset that contains 600 fake images and 400 real images. All the fake images are generated by a generative adversarial net and all the real images are downsampled images from the ImageNet dataset. These images are evaluated by 10 workers from the Amazon Mechanical Turk (AMT) based on eight carefully defined attributes.

  2. h

    vTikz-human-annotated

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    zu6yn4 xgma0i, vTikz-human-annotated [Dataset]. https://huggingface.co/datasets/zu6yn4xgma0i/vTikz-human-annotated
    Explore at:
    Authors
    zu6yn4 xgma0i
    Description

    zu6yn4xgma0i/vTikz-human-annotated dataset hosted on Hugging Face and contributed by the HF Datasets community

  3. Data from: AckSent: Human Annotated Dataset of Support and Sentiments in...

    • zenodo.org
    Updated Nov 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Author; Author (2024). AckSent: Human Annotated Dataset of Support and Sentiments in Dissertation Acknowledgments [Dataset]. http://doi.org/10.5281/zenodo.13283331
    Explore at:
    Dataset updated
    Nov 5, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Author; Author
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Time period covered
    Aug 10, 2024
    Description

    This data is supplementary to the paper "AckSent: Human Annotated Dataset of Support and Sentiments in Dissertation Acknowledgments" .

  4. P

    HANNA Dataset

    • library.toponeai.link
    Updated Aug 25, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Cyril Chhun; Pierre Colombo; Chloé Clavel; Fabian M. Suchanek (2022). HANNA Dataset [Dataset]. https://library.toponeai.link/dataset/hanna
    Explore at:
    Dataset updated
    Aug 25, 2022
    Authors
    Cyril Chhun; Pierre Colombo; Chloé Clavel; Fabian M. Suchanek
    Description

    HANNA, a large annotated dataset of Human-ANnotated NArratives for Automatic Story Generation (ASG) evaluation, has been designed for the benchmarking of automatic metrics for ASG. HANNA contains 1,056 stories generated from 96 prompts from the WritingPrompts dataset. Each prompt is linked to a human story and to 10 stories generated by different ASG systems. Each story was annotated on six human criteria (Relevance, Coherence, Empathy, Surprise, Engagement and Complexity) by three raters. HANNA also contains the scores produced by 72 automatic metrics on each story.

  5. e

    A Human-Annotated Dataset for Language Modeling and Named Entity Recognition...

    • b2find.eudat.eu
    Updated Oct 10, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). A Human-Annotated Dataset for Language Modeling and Named Entity Recognition in Medieval Documents - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/11273bbd-d9bb-5dcf-8fb8-e4e732017a1a
    Explore at:
    Dataset updated
    Oct 10, 2024
    Description

    This is an open dataset of sentences from 19th and 20th century letterpress reprints of documents from the Hussite era. The dataset contains a corpus for language modeling and human annotations for named entity recognition (NER).

  6. h

    HumAID-event-type

    • huggingface.co
    Updated Nov 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Qatar Computing Research Institute (2024). HumAID-event-type [Dataset]. https://huggingface.co/datasets/QCRI/HumAID-event-type
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 6, 2024
    Dataset authored and provided by
    Qatar Computing Research Institute
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    HumAID: Human-Annotated Disaster Incidents Data from Twitter

      Dataset Summary
    

    The HumAID Twitter dataset consists of several thousands of manually annotated tweets that has been collected during 19 major natural disaster events including earthquakes, hurricanes, wildfires, and floods, which happened from 2016 to 2019 across different parts of the World. The annotations in the provided datasets consists of following humanitarian categories. The dataset consists only english… See the full description on the dataset page: https://huggingface.co/datasets/QCRI/HumAID-event-type.

  7. e

    A Human-Annotated Dataset of Scanned Images and OCR Texts from Medieval...

    • b2find.eudat.eu
    Updated Apr 5, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). A Human-Annotated Dataset of Scanned Images and OCR Texts from Medieval Documents: Supplementary Materials - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/be292644-684e-59f4-8abd-032eed84cc39
    Explore at:
    Dataset updated
    Apr 5, 2024
    Description

    These are supplementary materials for an open dataset of scanned images and OCR texts from 19th and 20th century letterpress reprints of documents from the Hussite era. The dataset contains human annotations for layout analysis, OCR evaluation, and language identification and is available at http://hdl.handle.net/11234/1-4615. These supplementary materials contain OCR texts from different OCR engines for book pages for which we have both high-resolution scanned images and annotations for OCR evaluation.

  8. e

    A Human-Annotated Dataset of Scanned Images and OCR Texts from Medieval...

    • b2find.eudat.eu
    Updated Nov 2, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). A Human-Annotated Dataset of Scanned Images and OCR Texts from Medieval Documents - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/05858562-edc8-5284-9dea-bf6d1666b5c9
    Explore at:
    Dataset updated
    Nov 2, 2023
    Description

    This is an open dataset of scanned images and OCR texts from 19th and 20th century letterpress reprints of documents from the Hussite era. The dataset contains human annotations for layout analysis, OCR evaluation, and language identification.

  9. r

    A Structured Human-Annotated Dataset for Food Extrusion Literature

    • researchdata.edu.au
    • data.csiro.au
    datadownload
    Updated Feb 27, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jordan Pennells; Mr Jordan Pennells; Mr Jordan Pennells (2025). A Structured Human-Annotated Dataset for Food Extrusion Literature [Dataset]. http://doi.org/10.25919/R4Y6-R260
    Explore at:
    datadownloadAvailable download formats
    Dataset updated
    Feb 27, 2025
    Dataset provided by
    Commonwealth Scientific and Industrial Research Organisation
    Authors
    Jordan Pennells; Mr Jordan Pennells; Mr Jordan Pennells
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Time period covered
    Oct 17, 2022 - Feb 26, 2025
    Description

    This dataset is a manually curated collection of structured data extracted from peer-reviewed food extrusion research articles. The dataset captures key parameters relevant to food extrusion processes, including equipment configurations, processing conditions, formulation details, and characterization methods. It is intended to support literature synthesis, meta-analyses, and knowledge representation in food extrusion research. This dataset provides a searchable, structured repository for researchers to efficiently access and analyse trends in food extrusion studies beyond what is typically available in standard academic databases. Lineage: This dataset was manually curated from 335 peer-reviewed food extrusion research articles sourced from the Web of Science database. The literature search used the following search syntax: "extru*" (Topic) AND "food" (Topic) NOT “packaging” (Topic). WoS Category filter: Food Science Technology, Nutrition & Dietetics, and Agriculture Dairy Animal Science. Key parameters—including equipment configurations, processing conditions, formulation details, and characterisation methods—were extracted, structured, and categorised by a domain expert in food engineering following a predefined schema. Citation screening was performed to ensure dataset quality.

  10. Z

    RUSSE'2018: Human-Annotated Sense-Disambiguated Word Contexts for Russian

    • data.niaid.nih.gov
    • explore.openaire.eu
    Updated Jan 21, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ustalov, Dmitry (2020). RUSSE'2018: Human-Annotated Sense-Disambiguated Word Contexts for Russian [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_1117227
    Explore at:
    Dataset updated
    Jan 21, 2020
    Dataset authored and provided by
    Ustalov, Dmitry
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    This dataset contains human-annotated sense identifiers for 2562 contexts of 20 words used in the RUSSE'2018 shared task on Word Sense Induction and Disambiguation for the Russian language; part of the bts-rnc evaluation dataset. These sense identifiers are disambiguated as according to the sense inventory of the Large Explanatory Dictionary of Russian.

    The annotation is done on December 1, 2017, on the Yandex.Toloka crowdsourcing platform. In particular, 80 pre-annotated contexts are used for training the human annotators, 2562 contexts are annotated by humans such that each context was annotated by 9 different annotators. The annotation reliability is indicated by a high value of Krippendorff's α = 0.83. After the annotation, every context was additionally inspected (“curated”) by the organizers of the shared task.

    The following words are represented: акция (action / stock), байка (yarn / tale), гвоздика (carnation / nail), гипербола (hyperbole), град (avalanche), гусеница (grub), домино (domino), кабачок (marrow / pub), капот (hood), карьер (mine / career), кок (cook), крона (top / crown), круп (croup), мандарин (mandarine), рок (fate / rock), слог (syllable), стопка (glass, stack), таз (bowl), такса (rate / badger-dog), шах (shah / check).

    The following files are included in this dataset:

    Toloka assignments (training: tasks-train.tsv, annotation: tasks-test.tsv)

    Toloka output (non-aggregated: assignments_01-12-2017.tsv.xz, aggregated: aggregated_results_pool_1036853_2017_12_01.tsv)

    annotator agreement report (agreement.txt)

    curated report (report-curated.tsv.xz and a supplementary file tasks-eval.tsv.xz)

    the final aggregated dataset (bts-rnc-crowd.tsv)

    The bts-rnc-crowd.tsv file has the following format: id, lemma, sense_id, left hand side context, word form, right hand side context, list of senses. The encoding is UTF-8 and the line breaks are LF (UNIX).

  11. Human Dataset

    • kaggle.com
    Updated Sep 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Fares Elmenshawii (2023). Human Dataset [Dataset]. https://www.kaggle.com/datasets/fareselmenshawii/human-dataset/data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 2, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Fares Elmenshawii
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    This dataset consists of 17,300 images and includes two annotation files, each in a distinct format. The first file, labeled "Label," contains annotations at their original scale, while the second file, named "yolo_format_labels," provides annotations in YOLO format. The dataset was created by utilizing the OIDv4 toolkit, a specialized tool for gathering data from Google Open Images. It's important to emphasize that this dataset is exclusively dedicated to human detection.

    This dataset is a valuable resource for training deep learning models tailored for human detection tasks. The images in the dataset are of high quality and have been meticulously annotated with bounding boxes encompassing the regions where humans are present. Annotations are available in two formats: the original scale, which represents pixel coordinates of the bounding boxes, and the YOLO format, which represents bounding box coordinates in a normalized form.

    The dataset was carefully curated by scraping relevant images from Google Open Images using the OIDv4 toolkit. Only images relevant to human detection tasks were included. Consequently, it is an ideal choice for training deep learning models specifically designed for human detection tasks.

  12. h

    Data from: PlainFact

    • huggingface.co
    Updated Mar 12, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhiwen You (2025). PlainFact [Dataset]. https://huggingface.co/datasets/uzw/PlainFact
    Explore at:
    Dataset updated
    Mar 12, 2025
    Authors
    Zhiwen You
    License

    Attribution-ShareAlike 3.0 (CC BY-SA 3.0)https://creativecommons.org/licenses/by-sa/3.0/
    License information was derived automatically

    Description

    PlainFact is a high-quality human-annotated dataset with fine-grained explanation (i.e., added information) annotations designed for Plain Language Summarization tasks, along with PlainQAFact factuality evaluation framework. It is collected from the Cochrane database sampled from CELLS dataset (Guo et al., 2024). PlainFact is a sentence-level benchmark that splits the summaries into sentences with fine-grained explanation annotations. In total, we have 200 plain language summary-abstract… See the full description on the dataset page: https://huggingface.co/datasets/uzw/PlainFact.

  13. Annotated dataset for the semantic segmentation of radishes

    • zenodo.org
    zip
    Updated Jan 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David Colliaux; David Colliaux (2023). Annotated dataset for the semantic segmentation of radishes [Dataset]. http://doi.org/10.5281/zenodo.7555485
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 21, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    David Colliaux; David Colliaux
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This folder contains pictures of radishes collected on the PMF experimental field during Spring 2017. There are two kinds of labeled images in the following folders:


    - human annotations: human annotators draw polygons around each plant and those were then refined using an active contours algorithm.
    - machine annotations: An SVM trained on the human annotations was used to produce labeled images. Images with bad segmentation were manually discarded.

    Each of these folder contains an images folder containing original pictures and a labels folder containing binary segmentation masks.

  14. R

    Human Object Dectection Dataset

    • universe.roboflow.com
    zip
    Updated Oct 9, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Human Detection Using yolov8 (2023). Human Object Dectection Dataset [Dataset]. https://universe.roboflow.com/human-detection-using-yolov8/human-object-dectection/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 9, 2023
    Dataset authored and provided by
    Human Detection Using yolov8
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Humans Bounding Boxes
    Description

    Human Object Dectection

    ## Overview
    
    Human Object Dectection is a dataset for object detection tasks - it contains Humans annotations for 1,980 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  15. d

    Annotated Imagery Data | 50K+ Human Full Body Images (Japan) | Computer...

    • datarade.ai
    Updated Aug 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pixta AI (2023). Annotated Imagery Data | 50K+ Human Full Body Images (Japan) | Computer Vision Data [Dataset]. https://datarade.ai/data-products/5-000-human-full-body-with-multiple-attributes-for-ai-ml-m-pixta-ai
    Explore at:
    .json, .xml, .csv, .txtAvailable download formats
    Dataset updated
    Aug 31, 2023
    Dataset authored and provided by
    Pixta AI
    Area covered
    Japan
    Description
    1. Overview This dataset is a collection of 50,000+ images of Human full body with multiple attributes that are ready to use for optimizing the accuracy of computer vision models. All of the contents is sourced from PIXTA's stock library of 100M+ Asian-featured images and videos. PIXTA is the largest platform of visual materials in the Asia Pacific region offering fully-managed services, high quality contents and data, and powerful tools for businesses & organisations to enable their creative and machine learning projects.

    2. Annotated Imagery Data of human in full body images This dataset contains 50,000+ images of human in full body. The dataset has been annotated in face bounding box face, body bounding box and Attribute of mask, wheelchair, stroller, umbrella, suitcase, bag, backpack, laptop, cellphone,...

    3. About PIXTA PIXTASTOCK is the largest Asian-featured stock platform providing data, contents, tools and services since 2005. PIXTA experiences 15 years of integrating advanced AI technology in managing, curating, processing over 100M visual materials and serving global leading brands for their creative and data demands.

  16. Z

    Crowdsourced LibriTTS Speech Prominence Annotations

    • data.niaid.nih.gov
    • zenodo.org
    Updated Dec 18, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Morrison, Max (2023). Crowdsourced LibriTTS Speech Prominence Annotations [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10402792
    Explore at:
    Dataset updated
    Dec 18, 2023
    Dataset authored and provided by
    Morrison, Max
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset corresponding to the ICASSP 2024 paper "Crowdsourced and Automatic Speech Prominence Estimation" [link]

    This dataset is useful for training machine learning models to perform automatic emphasis annotaiton, as well as downstream tasks such as emphasis-controlled TTS, emotion recognition, and text summarization. The dataset is described in Section 3 (Emphasis Annotation Dataset). The contents of this section are copied below for convenience.

    We used our crowdsourced annotation system to perform human annotation on one eighth of the train-clean-100 partition of the LibriTTS [1] dataset. Specifically, participants annotated 3,626 utterances with a total length of 6.42 hours and 69,809 words from 18 speakers (9 male and 9 female). We collected at least one annotation of all 3,626 utterances, at least two annotations of 2,259 of those utterances, at least four annotations of 974 utterances, and at least eight annotations of 453 utterances. We did this in order to explore (in Section 6) whether it is more cost-effective to train a system on multiple annotations of fewer utterances or fewer annotations of more utterances. We paid 298 annotators to annotate batches of 20 utterances, where each batch takes approximately 15 minutes. We paid $3.34 for each completed batch (estimated $13.35 per hour). Annotators each annotated between one and six batches. We recruited on MTurk US residents with an approval rating of at least 99 and at least 1000 approved tasks. Today, microlabor platforms like MTurk are plagued by automated task-completion software agents (bots) that randomly fill out surveys. We filtered out bots by excluding annotations from an additional 107 annotators that marked more than 2/3 of words as emphasized in eight or more utterances of the 20 utterances in a batch. Annotators who fail the bot filter are blocked from performing further annotation. We also recorded participants' native country and language, but note these may be unreliable as many MTurk workers use VPNs to subvert IP region filters on MTurk [2].

    The average Cohen Kappa score for annotators with at least one overlapping utterance is 0.226 (i.e., ``Fair'' agreement)---but not all annotators annotate the same utterances, and this overemphasizes pairs of annotators with low overlap. Therefore, we use a one-parameter logistic model (i.e., a Rasch model) computed via py-irt [3], which predicts heldout annotations from scores of overlapping annotators with 77.7% accuracy (50% is random).

    The structure of this dataset is a single JSON file of word-aligned emphasis annotations. The JSON references file stems of the LibriTTS dataset, which can be found here. All code used in the creation of the dataset can be found here. The format of the JSON file is as follows.

    {

    { "annotations": [ { "score": [ , , ... ], "stem": , "words": [ [ , ,

    ], [ , ,

    ], ... ] }, ... ], "country": , "language": }, ... }

    [1] Zen et al., “LibriTTS: A corpus derived from LibriSpeech for text-to-speech,” in Interspeech, 2019.[2] Moss et al., “Bots or inattentive humans? Identifying sources of low-quality data in online platforms,” PsyArXiv preprint PsyArXiv:wr8ds, 2021.[3] John Patrick Lalor and Pedro Rodriguez, “py-irt: A scalable item response theory library for Python,” INFORMS Journal on Computing, 2023.

  17. p

    Annotated Imagery Data | 50K+ Human Full Body Images (Japan) | Computer...

    • data.pixta.ai
    Updated Aug 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pixta AI (2024). Annotated Imagery Data | 50K+ Human Full Body Images (Japan) | Computer Vision Data [Dataset]. https://data.pixta.ai/products/5-000-human-full-body-with-multiple-attributes-for-ai-ml-m-pixta-ai
    Explore at:
    Dataset updated
    Aug 19, 2024
    Dataset authored and provided by
    Pixta AI
    Area covered
    Japan
    Description

    50,000+ human full body Annotated Imagery images with multiple attributes ready for AI & ML models

  18. Human Face (Ear Detection with Annotation)

    • kaggle.com
    Updated Jun 28, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Harsh Ghadiya (2022). Human Face (Ear Detection with Annotation) [Dataset]. https://www.kaggle.com/datasets/harshghadiya/human-face-ear-detection-with-annotation/suggestions
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 28, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Harsh Ghadiya
    Description

    This dataset is for detecting the ears of humans. In this dataset, an annotation file of ears is also available for training and validation images in JSON format.

    This dataset will helpful for detecting face and ears as well, using maskRCNN and YOLO.

    • train and val folders are of images.
    • train_json.json is annotation file of train images.
    • val_json.json is annotation file of val images

    This dataset contains images of human faces with annotated ear regions. The dataset consists of a total of 440 images, with each image containing one or more human faces. The dataset is useful for developing computer vision models for ear detection in human faces, which can be used in various applications such as biometric authentication and surveillance. The images were collected from various sources and have varying resolutions and lighting conditions. The dataset includes a CSV file with annotations for the ear regions in each image.

  19. v

    Data from: A draft annotation and overview of the human genome

    • res1catalogd-o-tdatad-o-tgov.vcapture.xyz
    • healthdata.gov
    • +2more
    Updated Jul 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    National Institutes of Health (2025). A draft annotation and overview of the human genome [Dataset]. https://res1catalogd-o-tdatad-o-tgov.vcapture.xyz/dataset/a-draft-annotation-and-overview-of-the-human-genome
    Explore at:
    Dataset updated
    Jul 24, 2025
    Dataset provided by
    National Institutes of Health
    Description

    Background The recent draft assembly of the human genome provides a unified basis for describing genomic structure and function. The draft is sufficiently accurate to provide useful annotation, enabling direct observations of previously inferred biological phenomena. Results We report here a functionally annotated human gene index placed directly on the genome. The index is based on the integration of public transcript, protein, and mapping information, supplemented with computational prediction. We describe numerous global features of the genome and examine the relationship of various genetic maps with the assembly. In addition, initial sequence analysis reveals highly ordered chromosomal landscapes associated with paralogous gene clusters and distinct functional compartments. Finally, these annotation data were synthesized to produce observations of gene density and number that accord well with historical estimates. Such a global approach had previously been described only for chromosomes 21 and 22, which together account for 2.2% of the genome. Conclusions We estimate that the genome contains 65,000-75,000 transcriptional units, with exon sequences comprising 4%. The creation of a comprehensive gene index requires the synthesis of all available computational and experimental evidence.

  20. h

    MPII_Human_Pose_Dataset

    • huggingface.co
    Updated May 7, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Voxel51 (2024). MPII_Human_Pose_Dataset [Dataset]. https://huggingface.co/datasets/Voxel51/MPII_Human_Pose_Dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 7, 2024
    Dataset authored and provided by
    Voxel51
    License

    https://choosealicense.com/licenses/bsd-2-clause/https://choosealicense.com/licenses/bsd-2-clause/

    Description

    Dataset Card for MPII Human Pose

    MPII Human Pose dataset is a state of the art benchmark for evaluation of articulated human pose estimation. The dataset includes around 25K images containing over 40K people with annotated body joints. The images were systematically collected using an established taxonomy of every day human activities. Overall the dataset covers 410 human activities and each image is provided with an activity label. Each image was extracted from a YouTube video… See the full description on the dataset page: https://huggingface.co/datasets/Voxel51/MPII_Human_Pose_Dataset.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Juanyong Duan (2019). Human Annotated GAN Images [Dataset]. https://ieee-dataport.org/documents/human-annotated-gan-images

Human Annotated GAN Images

Explore at:
Dataset updated
Apr 25, 2019
Authors
Juanyong Duan
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

we have collected an annotated dataset that contains 600 fake images and 400 real images. All the fake images are generated by a generative adversarial net and all the real images are downsampled images from the ImageNet dataset. These images are evaluated by 10 workers from the Amazon Mechanical Turk (AMT) based on eight carefully defined attributes.

Search
Clear search
Close search
Google apps
Main menu