13 datasets found
  1. h

    commonsense_qa

    • huggingface.co
    • paperswithcode.com
    • +1more
    Updated May 18, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tel Aviv University (2022). commonsense_qa [Dataset]. https://huggingface.co/datasets/tau/commonsense_qa
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 18, 2022
    Dataset authored and provided by
    Tel Aviv University
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset Card for "commonsense_qa"

      Dataset Summary
    

    CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers. The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation split, and "Question token split", see paper for details.… See the full description on the dataset page: https://huggingface.co/datasets/tau/commonsense_qa.

  2. o

    Cosmos QA (Commonsense QA)

    • opendatabay.com
    .undefined
    Updated Jun 28, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Datasimple (2025). Cosmos QA (Commonsense QA) [Dataset]. https://www.opendatabay.com/data/ai-ml/8c24035b-6a93-4935-bc78-00707e2ba71b
    Explore at:
    .undefinedAvailable download formats
    Dataset updated
    Jun 28, 2025
    Dataset authored and provided by
    Datasimple
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Area covered
    Data Science and Analytics
    Description

    The Cosmos QA dataset is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. The dataset focuses on reading between the lines over a diverse collection of people's everyday narratives, asking questions concerning on the likely causes or effects of events that require reasoning beyond the exact text spans in the context.

    This allows for much more sophisticated models to be built and evaluated, and could lead to better performance on real-world tasks

    How to use the dataset In order to use the Cosmos QA dataset, you will need to first download the data files from the Kaggle website. Once you have downloaded the files, you will need to unzip them and then place them in a directory on your computer.

    Once you have the data files placed on your computer, you can begin using the dataset for commonsense-based reading comprehension tasks. The first step is to load the context file into a text editor such as Microsoft Word or Adobe Acrobat Reader. Once the context file is open, you will need to locate the section of text that contains the question that you want to answer.

    Once you have located the section of text containing the question, you will need to read through thecontext in order to determine what type of answer would be most appropriate. After carefully reading throughthe context, you should then look at each of the answer choices and selectthe one that best fits with what you have read

    Research Ideas This dataset can be used to develop and evaluate commonsense-based reading comprehension models. This dataset can be used to improve and customize question answering systems for educational or customer service applications. This dataset can be used to study how human beings process and understand narratives, in order to better design artificial intelligence systems that can do the same

    Columns File: validation.csv

    Column name Description context The context of the question. (String) answer0 The first answer option. (String) answer1 The second answer option. (String) answer2 The third answer option. (String) answer3 The fourth answer option. (String) label The correct answer to the question. (String) File: train.csv

    Column name Description context The context of the question. (String) answer0 The first answer option. (String) answer1 The second answer option. (String) answer2 The third answer option. (String) answer3 The fourth answer option. (String) label The correct answer to the question. (String) File: test.csv

    Column name Description context The context of the question. (String) answer0 The first answer option. (String) answer1 The second answer option. (String) answer2 The third answer option. (String) answer3 The fourth answer option. (String) label The correct answer to the question. (String)

    License

    CC0

    Original Data Source: Cosmos QA (Commonsense QA)

  3. h

    fedrag-commonsense-qa

    • huggingface.co
    Updated Apr 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andrei Fajardo (2025). fedrag-commonsense-qa [Dataset]. https://huggingface.co/datasets/nerdai/fedrag-commonsense-qa
    Explore at:
    Dataset updated
    Apr 10, 2025
    Authors
    Andrei Fajardo
    Description

    nerdai/fedrag-commonsense-qa dataset hosted on Hugging Face and contributed by the HF Datasets community

  4. u

    RESPONSE: Dataset for Commonsense Reasoning about Disaster Management

    • rdr.ucl.ac.uk
    txt
    Updated Jun 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aissatou Diallo (2024). RESPONSE: Dataset for Commonsense Reasoning about Disaster Management [Dataset]. http://doi.org/10.5522/04/26010064.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Jun 25, 2024
    Dataset provided by
    University College London
    Authors
    Aissatou Diallo
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    This dataset contains 1789 data instances with problem identification, missing resource, time-dependent questions and answers pairs for disaster management.

  5. h

    commonsense_cot_partial_raw

    • huggingface.co
    Updated Jan 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter Chung (2024). commonsense_cot_partial_raw [Dataset]. https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_raw
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 26, 2024
    Authors
    Peter Chung
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Commonsense QA CoT (Partial, Raw, No Human Annotation)

      Dataset Summary
    

    Seeded by the CommonsenseQA dataset (tau/commonsense_qa) this preliminary set randomly samples 1,000 question-answer entries and uses Mixtral (mistralai/Mixtral-8x7B-Instruct-v0.1) to generate 3 unique CoT (Chain-of-Thought) rationales. This was created as the preliminary step towards fine-tuning a LM (language model) to specialize on commonsense reasoning. The working hypothesis, inspired by the… See the full description on the dataset page: https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_raw.

  6. h

    commonsense_cot_partial_annotated_v0.1

    • huggingface.co
    Updated Feb 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter Chung (2024). commonsense_cot_partial_annotated_v0.1 [Dataset]. https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_annotated_v0.1
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 6, 2024
    Authors
    Peter Chung
    Description

    Commonsense QA CoT (Partial, Annotated) v0.1

      Dataset Summary
    

    This dataset is a human-annotated subset of randomly sampled question-answer entries from the CommonsenseQA dataset (tau/commonsense_qa). The 'rationales' for each QA pair were created using a two-part method. First, Mixtral (mistralai/Mixtral-8x7B-Instruct-v0.1) was used to generate 3 unique CoT (Chain-of-Thought) explanations. Next, human evaluation was applied to distill the random sampling down to a cohesive… See the full description on the dataset page: https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_annotated_v0.1.

  7. social_i_qa

    • huggingface.co
    • opendatalab.com
    Updated May 23, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    social_i_qa [Dataset]. https://huggingface.co/datasets/allenai/social_i_qa
    Explore at:
    Dataset updated
    May 23, 2024
    Dataset provided by
    Allen Institute for AIhttp://allenai.org/
    Authors
    Ai2
    Description

    We introduce Social IQa: Social Interaction QA, a new question-answering benchmark for testing social commonsense intelligence. Contrary to many prior benchmarks that focus on physical or taxonomic knowledge, Social IQa focuses on reasoning about people’s actions and their social implications. For example, given an action like "Jesse saw a concert" and a question like "Why did Jesse do this?", humans can easily infer that Jesse wanted "to see their favorite performer" or "to enjoy the music", and not "to see what's happening inside" or "to see if it works". The actions in Social IQa span a wide variety of social situations, and answer candidates contain both human-curated answers and adversarially-filtered machine-generated candidates. Social IQa contains over 37,000 QA pairs for evaluating models’ abilities to reason about the social implications of everyday events and situations. (Less)

  8. T

    piqa

    • tensorflow.org
    Updated Dec 16, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). piqa [Dataset]. https://www.tensorflow.org/datasets/catalog/piqa
    Explore at:
    Dataset updated
    Dec 16, 2022
    Description

    Physical IQa: Physical Interaction QA, a new commonsense QA benchmark for naive physics reasoning focusing on how we interact with everyday objects in everyday situations. This dataset focuses on affordances of objects, i.e., what actions each physical object affords (e.g., it is possible to use a shoe as a doorstop), and what physical interactions a group of objects afford (e.g., it is possible to place an apple on top of a book, but not the other way around). The dataset requires reasoning about both the prototypical use of objects (e.g., shoes are used for walking) and non-prototypical but practically plausible use of objects (e.g., shoes can be used as a doorstop). The dataset includes 20,000 QA pairs that are either multiple-choice or true/false questions.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('piqa', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

  9. P

    PIQA Dataset

    • paperswithcode.com
    • library.toponeai.link
    Updated Jan 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yonatan Bisk; Rowan Zellers; Ronan Le Bras; Jianfeng Gao; Yejin Choi (2025). PIQA Dataset [Dataset]. https://paperswithcode.com/dataset/piqa
    Explore at:
    Dataset updated
    Jan 5, 2025
    Authors
    Yonatan Bisk; Rowan Zellers; Ronan Le Bras; Jianfeng Gao; Yejin Choi
    Description

    PIQA is a dataset for commonsense reasoning, and was created to investigate the physical knowledge of existing models in NLP.

  10. P

    Rainbow Dataset

    • paperswithcode.com
    • opendatalab.com
    • +1more
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nicholas Lourie; Ronan Le Bras; Chandra Bhagavatula; Yejin Choi, Rainbow Dataset [Dataset]. https://paperswithcode.com/dataset/rainbow
    Explore at:
    Authors
    Nicholas Lourie; Ronan Le Bras; Chandra Bhagavatula; Yejin Choi
    Description

    Rainbow is multi-task benchmark for common-sense reasoning that uses different existing QA datasets: aNLI, Cosmos QA, HellaSWAG. Physical IQa, Social IQa, WinoGrande.

  11. P

    RuCoS Dataset

    • paperswithcode.com
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tatiana Shavrina; Alena Fenogenova; Anton Emelyanov; Denis Shevelev; Ekaterina Artemova; Valentin Malykh; Vladislav Mikhailov; Maria Tikhonova; Andrey Chertok; Andrey Evlampiev, RuCoS Dataset [Dataset]. https://paperswithcode.com/dataset/rucos
    Explore at:
    Authors
    Tatiana Shavrina; Alena Fenogenova; Anton Emelyanov; Denis Shevelev; Ekaterina Artemova; Valentin Malykh; Vladislav Mikhailov; Maria Tikhonova; Andrey Chertok; Andrey Evlampiev
    Description

    Russian reading comprehension with Commonsense reasoning (RuCoS) is a large-scale reading comprehension dataset that requires commonsense reasoning. RuCoS consists of queries automatically generated from CNN/Daily Mail news articles; the answer to each query is a text span from a summarizing passage of the corresponding news. The goal of RuCoS is to evaluate a machine`s ability of commonsense reasoning in reading comprehension.

    Example {'source': 'Lenta', 'passage': { 'text': 'Мать двух мальчиков, брошенных отцом в московском аэропорту Шереметьево, забрала их. Об этом сообщили ТАСС в пресс-службе министерства образования и науки Хабаровского края. Сейчас младший ребенок посещает детский сад, а старший ходит в школу. В учебных заведениях с ними по необходимости работают штатные психологи. Также министерство социальной защиты населения рассматривает вопрос о бесплатном оздоровлении детей в летнее время. Через несколько дней после того, как Виктор Гаврилов бросил своих детей в аэропорту, он явился с повинной к следователям в городе Батайске Ростовской области. @context Бросившего детей в Шереметьево отца задержали за насилие над женой @context Россиянина заподозрили в истязании брошенных в Шереметьево детей @context Оставивший двоих детей в Шереметьево россиянин сам пришел к следователям', 'entities': [ {'start': 60, 'end': 71, 'text': 'Шереметьево'}, {'start': 102, 'end': 106, 'text': 'ТАСС'}, {'start': 155, 'end': 172, 'text': 'Хабаровского края'}, {'start': 470, 'end': 485, 'text': 'Виктор Гаврилов'}, {'start': 563, 'end': 571, 'text': 'Батайске'}, {'start': 572, 'end': 590, 'text': 'Ростовской области'}, {'start': 620, 'end': 631, 'text': 'Шереметьево'}, {'start': 725, 'end': 736, 'text': 'Шереметьево'}, {'start': 778, 'end': 789, 'text': 'Шереметьево'} ] }, 'qas': [ { 'query': '26 января @placeholder бросил сыновей в возрасте пяти и семи лет в Шереметьево.', 'answers': [ {'start': 470, 'end': 485, 'text': 'Виктор Гаврилов'} ], 'idx': 0 } ], 'idx': 0 }

    How did we collect data? All text examples were collected from open news sources, then automatically filtered with QA systems to prevent obvious questions to infiltrate the dataset. The texts were then filtered by IPM frequency of the contained words and, finally, manually reviewed.

  12. h

    chatgpt4-commonsense-qa

    • huggingface.co
    Updated Jun 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mesolitica (2025). chatgpt4-commonsense-qa [Dataset]. https://huggingface.co/datasets/mesolitica/chatgpt4-commonsense-qa
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 24, 2025
    Dataset authored and provided by
    Mesolitica
    Description

    Synthetic CommonSense

    Generated using ChatGPT4, originally from https://huggingface.co/datasets/commonsense_qa Notebook at https://github.com/mesolitica/malaysian-dataset/tree/master/question-answer/chatgpt4-commonsense

    synthetic-commonsense.jsonl, 36332 rows, 7.34 MB.

      Example data
    

    {'question': '1. Seseorang yang bersara mungkin perlu kembali bekerja jika mereka apa? A. mempunyai hutang B. mencari pendapatan C. meninggalkan pekerjaan D. memerlukan… See the full description on the dataset page: https://huggingface.co/datasets/mesolitica/chatgpt4-commonsense-qa.

  13. P

    Visual Question Answering v2.0 Dataset

    • paperswithcode.com
    Updated Mar 15, 2017
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yash Goyal; Tejas Khot; Douglas Summers-Stay; Dhruv Batra; Devi Parikh (2017). Visual Question Answering v2.0 Dataset [Dataset]. https://paperswithcode.com/dataset/visual-question-answering-v2-0
    Explore at:
    Dataset updated
    Mar 15, 2017
    Authors
    Yash Goyal; Tejas Khot; Douglas Summers-Stay; Dhruv Batra; Devi Parikh
    Description

    Visual Question Answering (VQA) v2.0 is a dataset containing open-ended questions about images. These questions require an understanding of vision, language and commonsense knowledge to answer. It is the second version of the VQA dataset.

    265,016 images (COCO and abstract scenes) At least 3 questions (5.4 questions on average) per image 10 ground truth answers per question 3 plausible (but likely incorrect) answers per question Automatic evaluation metric

    The first version of the dataset was released in October 2015.

  14. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Tel Aviv University (2022). commonsense_qa [Dataset]. https://huggingface.co/datasets/tau/commonsense_qa

commonsense_qa

CommonsenseQA

tau/commonsense_qa

Explore at:
67 scholarly articles cite this dataset (View in Google Scholar)
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
May 18, 2022
Dataset authored and provided by
Tel Aviv University
License

MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically

Description

Dataset Card for "commonsense_qa"

  Dataset Summary

CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers. The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation split, and "Question token split", see paper for details.… See the full description on the dataset page: https://huggingface.co/datasets/tau/commonsense_qa.

Search
Clear search
Close search
Google apps
Main menu