7 datasets found
  1. h

    VizWiz-VQA

    • huggingface.co
    Updated Oct 4, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    LMMs-Lab (2024). VizWiz-VQA [Dataset]. https://huggingface.co/datasets/lmms-lab/VizWiz-VQA
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 4, 2024
    Dataset authored and provided by
    LMMs-Lab
    Description

    Dataset Card for "VizWiz-VQA"

      Large-scale Multi-modality Models Evaluation Suite
    

    Accelerating the development of large-scale multi-modality models (LMMs) with lmms-eval

    🏠 Homepage | 📚 Documentation | 🤗 Huggingface Datasets

      This Dataset
    

    This is a formatted version of VizWiz-VQA. It is used in our lmms-eval pipeline to allow for one-click evaluations of large multi-modality models. @inproceedings{gurari2018vizwiz, title={Vizwiz grand challenge:… See the full description on the dataset page: https://huggingface.co/datasets/lmms-lab/VizWiz-VQA.

  2. h

    VizWiz

    • huggingface.co
    Updated Jul 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    CV as NLP (2025). VizWiz [Dataset]. https://huggingface.co/datasets/Multimodal-Fatima/VizWiz
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 6, 2025
    Dataset authored and provided by
    CV as NLP
    Description

    Multimodal-Fatima/VizWiz dataset hosted on Hugging Face and contributed by the HF Datasets community

  3. h

    VizWiz

    • huggingface.co
    Updated Apr 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    HuggingFaceM4 (2024). VizWiz [Dataset]. https://huggingface.co/datasets/HuggingFaceM4/VizWiz
    Explore at:
    Dataset updated
    Apr 6, 2024
    Dataset authored and provided by
    HuggingFaceM4
    Description

    The VizWiz-VQA dataset originates from a natural visual question answering setting where blind people each took an image and recorded a spoken question about it, together with 10 crowdsourced answers per visual question. The proposed challenge addresses the following two tasks for this dataset: predict the answer to a visual question and (2) predict whether a visual question cannot be answered.

  4. h

    vizwiz-train-val_ef_id

    • huggingface.co
    Updated Sep 15, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    سامیار دقیقی راد (2025). vizwiz-train-val_ef_id [Dataset]. https://huggingface.co/datasets/samydgr2/vizwiz-train-val_ef_id
    Explore at:
    Dataset updated
    Sep 15, 2025
    Authors
    سامیار دقیقی راد
    Description

    samydgr2/vizwiz-train-val_ef_id dataset hosted on Hugging Face and contributed by the HF Datasets community

  5. h

    MMEB-eval-VizWiz-beir

    • huggingface.co
    Updated Jul 15, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zilin Xiao (2025). MMEB-eval-VizWiz-beir [Dataset]. https://huggingface.co/datasets/MrZilinXiao/MMEB-eval-VizWiz-beir
    Explore at:
    Dataset updated
    Jul 15, 2025
    Authors
    Zilin Xiao
    Description

    MrZilinXiao/MMEB-eval-VizWiz-beir dataset hosted on Hugging Face and contributed by the HF Datasets community

  6. h

    CogniORPOData-VizWiz_Captions

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhang, CogniORPOData-VizWiz_Captions [Dataset]. https://huggingface.co/datasets/KniKna/CogniORPOData-VizWiz_Captions
    Explore at:
    Authors
    Zhang
    Description

    CogniORPOData-VizWiz_Captions Dataset

      VizWiz-Captions_merged_with_complexity.parquet
    

    Background of the original dataset: The images in this dataset are taken by visually impaired people using mobile phones, involving a variety of daily activities, most of which require reading some text.

    Extracted quantity: 7750 image-caption pairs are extracted from this dataset.

    Included content: Original image vectors, metadata information, 5 generated new text descriptions from… See the full description on the dataset page: https://huggingface.co/datasets/KniKna/CogniORPOData-VizWiz_Captions.

  7. KVQA

    • huggingface.co
    Updated Dec 11, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SK Telecom (2019). KVQA [Dataset]. https://huggingface.co/datasets/skt/KVQA
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 11, 2019
    Dataset provided by
    SK텔레콤https://sktelecom.com/
    Authors
    SK Telecom
    License

    https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/

    Description

    We also provide KVQA blog pages in both Korean and English. SK텔레콤은 사회적 가치 추구를 위한 다양한 사업을 진행하고 있습니다. 기업이 먼저 앞장서서 사회 속에 혼재된 사회적 이슈를 발굴하고, 이를 해결하기 위한 사회적 책임을 지는 것이 지속가능한 경영의 출발이라고 생각합니다. 2019년 4월부터 이 기술의 현지화를 위해 사회적 기업인 테스트웍스와 협업하여 자발적으로 지원한 우리나라의 시각장애인들로부터 데이터를 수집하였고, 영문으로 공개된 VizWiz 데이터셋 중 현지화가 가능한 일부를 한국어로 번역하여 시각적 질의응답 기술을 한국어로 학습시킬 수 있는 데이터셋을 만들었습니다.

      논문
    
    
    
    
    
    
    
      AI for Social Good workshop at NeurIPS (Kim & Lim et al., 2019)
    

    PDF

      시각적 질의응답
    

    시각적 질의응답은 이미지가 주어지고 그… See the full description on the dataset page: https://huggingface.co/datasets/skt/KVQA.

  8. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
LMMs-Lab (2024). VizWiz-VQA [Dataset]. https://huggingface.co/datasets/lmms-lab/VizWiz-VQA

VizWiz-VQA

lmms-lab/VizWiz-VQA

Explore at:
126 scholarly articles cite this dataset (View in Google Scholar)
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Oct 4, 2024
Dataset authored and provided by
LMMs-Lab
Description

Dataset Card for "VizWiz-VQA"

  Large-scale Multi-modality Models Evaluation Suite

Accelerating the development of large-scale multi-modality models (LMMs) with lmms-eval

🏠 Homepage | 📚 Documentation | 🤗 Huggingface Datasets

  This Dataset

This is a formatted version of VizWiz-VQA. It is used in our lmms-eval pipeline to allow for one-click evaluations of large multi-modality models. @inproceedings{gurari2018vizwiz, title={Vizwiz grand challenge:… See the full description on the dataset page: https://huggingface.co/datasets/lmms-lab/VizWiz-VQA.

Search
Clear search
Close search
Google apps
Main menu