Dataset Card for "VizWiz-VQA"
Large-scale Multi-modality Models Evaluation Suite
Accelerating the development of large-scale multi-modality models (LMMs) with lmms-eval
🏠 Homepage | 📚 Documentation | 🤗 Huggingface Datasets
This Dataset
This is a formatted version of VizWiz-VQA. It is used in our lmms-eval pipeline to allow for one-click evaluations of large multi-modality models. @inproceedings{gurari2018vizwiz, title={Vizwiz grand challenge:… See the full description on the dataset page: https://huggingface.co/datasets/lmms-lab/VizWiz-VQA.
Multimodal-Fatima/VizWiz dataset hosted on Hugging Face and contributed by the HF Datasets community
The VizWiz-VQA dataset originates from a natural visual question answering setting where blind people each took an image and recorded a spoken question about it, together with 10 crowdsourced answers per visual question. The proposed challenge addresses the following two tasks for this dataset: predict the answer to a visual question and (2) predict whether a visual question cannot be answered.
samydgr2/vizwiz-train-val_ef_id dataset hosted on Hugging Face and contributed by the HF Datasets community
MrZilinXiao/MMEB-eval-VizWiz-beir dataset hosted on Hugging Face and contributed by the HF Datasets community
CogniORPOData-VizWiz_Captions Dataset
VizWiz-Captions_merged_with_complexity.parquet
Background of the original dataset: The images in this dataset are taken by visually impaired people using mobile phones, involving a variety of daily activities, most of which require reading some text.
Extracted quantity: 7750 image-caption pairs are extracted from this dataset.
Included content: Original image vectors, metadata information, 5 generated new text descriptions from… See the full description on the dataset page: https://huggingface.co/datasets/KniKna/CogniORPOData-VizWiz_Captions.
https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/
We also provide KVQA blog pages in both Korean and English. SK텔레콤은 사회적 가치 추구를 위한 다양한 사업을 진행하고 있습니다. 기업이 먼저 앞장서서 사회 속에 혼재된 사회적 이슈를 발굴하고, 이를 해결하기 위한 사회적 책임을 지는 것이 지속가능한 경영의 출발이라고 생각합니다. 2019년 4월부터 이 기술의 현지화를 위해 사회적 기업인 테스트웍스와 협업하여 자발적으로 지원한 우리나라의 시각장애인들로부터 데이터를 수집하였고, 영문으로 공개된 VizWiz 데이터셋 중 현지화가 가능한 일부를 한국어로 번역하여 시각적 질의응답 기술을 한국어로 학습시킬 수 있는 데이터셋을 만들었습니다.
논문
AI for Social Good workshop at NeurIPS (Kim & Lim et al., 2019)
시각적 질의응답
시각적 질의응답은 이미지가 주어지고 그… See the full description on the dataset page: https://huggingface.co/datasets/skt/KVQA.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Dataset Card for "VizWiz-VQA"
Large-scale Multi-modality Models Evaluation Suite
Accelerating the development of large-scale multi-modality models (LMMs) with lmms-eval
🏠 Homepage | 📚 Documentation | 🤗 Huggingface Datasets
This Dataset
This is a formatted version of VizWiz-VQA. It is used in our lmms-eval pipeline to allow for one-click evaluations of large multi-modality models. @inproceedings{gurari2018vizwiz, title={Vizwiz grand challenge:… See the full description on the dataset page: https://huggingface.co/datasets/lmms-lab/VizWiz-VQA.