19 datasets found
  1. h

    commonsense_qa

    • huggingface.co
    • opendatalab.com
    Updated Nov 2, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tel Aviv University (2018). commonsense_qa [Dataset]. https://huggingface.co/datasets/tau/commonsense_qa
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 2, 2018
    Dataset authored and provided by
    Tel Aviv University
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset Card for "commonsense_qa"

      Dataset Summary
    

    CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers. The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation split, and "Question token split", see paper for details.… See the full description on the dataset page: https://huggingface.co/datasets/tau/commonsense_qa.

  2. Commonsense QA

    • kaggle.com
    zip
    Updated Nov 25, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Darien Schettler (2023). Commonsense QA [Dataset]. https://www.kaggle.com/datasets/dschettler8845/commonsense-qa
    Explore at:
    zip(2070878 bytes)Available download formats
    Dataset updated
    Nov 25, 2023
    Authors
    Darien Schettler
    Description

    Dataset

    This dataset was created by Darien Schettler

    Contents

  3. h

    commonsense_cot_partial_raw

    • huggingface.co
    Updated Jan 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter Chung (2024). commonsense_cot_partial_raw [Dataset]. https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_raw
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 26, 2024
    Authors
    Peter Chung
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Commonsense QA CoT (Partial, Raw, No Human Annotation)

      Dataset Summary
    

    Seeded by the CommonsenseQA dataset (tau/commonsense_qa) this preliminary set randomly samples 1,000 question-answer entries and uses Mixtral (mistralai/Mixtral-8x7B-Instruct-v0.1) to generate 3 unique CoT (Chain-of-Thought) rationales. This was created as the preliminary step towards fine-tuning a LM (language model) to specialize on commonsense reasoning. The working hypothesis, inspired by the… See the full description on the dataset page: https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_raw.

  4. Cosmos QA (Commonsense QA)

    • kaggle.com
    zip
    Updated Nov 20, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2022). Cosmos QA (Commonsense QA) [Dataset]. https://www.kaggle.com/datasets/thedevastator/cosmos-qa-a-large-scale-commonsense-based-readin/code
    Explore at:
    zip(7812205 bytes)Available download formats
    Dataset updated
    Nov 20, 2022
    Authors
    The Devastator
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Cosmos QA (Commonsense QA)

    Pushing Commonsense Reasoning to the Next Level

    Source

    Huggingface Hub: link

    About this dataset

    The Cosmos QA dataset is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. The dataset focuses on reading between the lines over a diverse collection of people's everyday narratives, asking questions concerning on the likely causes or effects of events that require reasoning beyond the exact text spans in the context.

    This allows for much more sophisticated models to be built and evaluated, and could lead to better performance on real-world tasks

    How to use the dataset

    In order to use the Cosmos QA dataset, you will need to first download the data files from the Kaggle website. Once you have downloaded the files, you will need to unzip them and then place them in a directory on your computer.

    Once you have the data files placed on your computer, you can begin using the dataset for commonsense-based reading comprehension tasks. The first step is to load the context file into a text editor such as Microsoft Word or Adobe Acrobat Reader. Once the context file is open, you will need to locate the section of text that contains the question that you want to answer.

    Once you have located the section of text containing the question, you will need to read through thecontext in order to determine what type of answer would be most appropriate. After carefully reading throughthe context, you should then look at each of the answer choices and selectthe one that best fits with what you have read

    Research Ideas

    • This dataset can be used to develop and evaluate commonsense-based reading comprehension models.
    • This dataset can be used to improve and customize question answering systems for educational or customer service applications.
    • This dataset can be used to study how human beings process and understand narratives, in order to better design artificial intelligence systems that can do the same

    Acknowledgements

    License

    License: CC0 1.0 Universal (CC0 1.0) - Public Domain Dedication No Copyright - You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission. See Other Information.

    Columns

    File: validation.csv | Column name | Description | |:--------------|:---------------------------------------------| | context | The context of the question. (String) | | answer0 | The first answer option. (String) | | answer1 | The second answer option. (String) | | answer2 | The third answer option. (String) | | answer3 | The fourth answer option. (String) | | label | The correct answer to the question. (String) |

    File: train.csv | Column name | Description | |:--------------|:---------------------------------------------| | context | The context of the question. (String) | | answer0 | The first answer option. (String) | | answer1 | The second answer option. (String) | | answer2 | The third answer option. (String) | | answer3 | The fourth answer option. (String) | | label | The correct answer to the question. (String) |

    File: test.csv | Column name | Description | |:--------------|:---------------------------------------------| | context | The context of the question. (String) | | answer0 | The first answer option. (String) | | answer1 | The second answer option. (String) | | answer2 | The third answer option. (String) | | answer3 | The fourth answer option. (String) | | label | The correct answer to the question. (String) |

  5. n

    chatgpt4-commonsense-qa

    • nerq.ai
    json
    Updated Mar 24, 2026
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    mesolitica (2026). chatgpt4-commonsense-qa [Dataset]. https://nerq.ai/dataset/chatgpt4-commonsense-qa
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Mar 24, 2026
    Dataset authored and provided by
    mesolitica
    Variables measured
    Downloads, Trust Score
    Description

    An AI assistant for common sense QA.

  6. h

    fedrag-commonsense-qa

    • huggingface.co
    Updated Jun 20, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andrei Fajardo (2023). fedrag-commonsense-qa [Dataset]. https://huggingface.co/datasets/nerdai/fedrag-commonsense-qa
    Explore at:
    Dataset updated
    Jun 20, 2023
    Authors
    Andrei Fajardo
    Description

    nerdai/fedrag-commonsense-qa dataset hosted on Hugging Face and contributed by the HF Datasets community

  7. T

    cosmos_qa

    • tensorflow.org
    • opendatalab.com
    • +1more
    Updated Dec 6, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). cosmos_qa [Dataset]. https://www.tensorflow.org/datasets/catalog/cosmos_qa
    Explore at:
    Dataset updated
    Dec 6, 2022
    Description

    Cosmos QA is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. It focuses on reading between the lines over a diverse collection of people's everyday narratives, asking questions concerning on the likely causes or effects of events that require reasoning beyond the exact text spans in the context.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('cosmos_qa', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

  8. u

    RESPONSE: Dataset for Commonsense Reasoning about Disaster Management

    • rdr.ucl.ac.uk
    txt
    Updated Jun 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aissatou Diallo (2024). RESPONSE: Dataset for Commonsense Reasoning about Disaster Management [Dataset]. http://doi.org/10.5522/04/26010064.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Jun 25, 2024
    Dataset provided by
    University College London
    Authors
    Aissatou Diallo
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    This dataset contains 1789 data instances with problem identification, missing resource, time-dependent questions and answers pairs for disaster management.

  9. Social IQa (Social Interaction Q&A)

    • kaggle.com
    zip
    Updated Nov 20, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2022). Social IQa (Social Interaction Q&A) [Dataset]. https://www.kaggle.com/datasets/thedevastator/social-i-qa-a-dataset-for-social-inquiry-questio/code
    Explore at:
    zip(2024126 bytes)Available download formats
    Dataset updated
    Nov 20, 2022
    Authors
    The Devastator
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Social IQa (Social Interaction Q&A)

    Question-answering benchmark for testing commonsense social intelligence

    Source

    Huggingface Hub: link

    About this dataset

    We introduce Social IQa: Social Interaction QA, a new question-answering benchmark for testing social commonsense intelligence. Contrary to many prior benchmarks that focus on physical or taxonomic knowledge, Social IQa focuses on reasoning about people’s actions and their social implications. For example, given an action like "Jesse saw a concert" and a question like "Why did Jesse do this?", humans can easily infer that Jesse wanted "to see their favorite performer" or "to enjoy the music", and not "to see what's happening inside" or "to see if it works". The actions in Social IQa span a wide variety of social situations, and answer candidates contain both human-curated answers and adversarially-filtered machine-generated candidates. Social IQa contains over 37,000 QA pairs for evaluating models’ abilities to reason about the social implications of everyday events and situations. (Less)

    How to use the dataset

    This dataset can be used to train and test models for social inquiry question answering. The questions and answers in the dataset have been annotations by experts, and the dataset has been verified for accuracy.

    Research Ideas

    • The dataset can be used to train a model to answer questions about social topics.
    • The dataset can be used to improve question-answering systems for social inquiry.
    • The dataset can be used to generate new questions about social topics

    Acknowledgements

    Huggingface Hub: link

    License

    License: CC0 1.0 Universal (CC0 1.0) - Public Domain Dedication No Copyright - You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission. See Other Information.

    Columns

    File: validation.csv | Column name | Description | |:--------------|:------------------------------------------------------| | context | The context of the question. (String) | | answerA | One of the possible answers to the question. (String) | | answerB | One of the possible answers to the question. (String) | | answerC | One of the possible answers to the question. (String) | | label | The correct answer to the question. (String) |

    File: train.csv | Column name | Description | |:--------------|:------------------------------------------------------| | context | The context of the question. (String) | | answerA | One of the possible answers to the question. (String) | | answerB | One of the possible answers to the question. (String) | | answerC | One of the possible answers to the question. (String) | | label | The correct answer to the question. (String) |

  10. h

    commonsense_cot_partial_annotated_prelim

    • huggingface.co
    Updated Feb 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter Chung (2024). commonsense_cot_partial_annotated_prelim [Dataset]. https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_annotated_prelim
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 6, 2024
    Authors
    Peter Chung
    Description

    Commonsense QA CoT (Partial, Annotated) - PRELIMINARY

      Dataset Summary
    

    This dataset is a human-annotated subset of randomly sampled question-answer entries from the CommonsenseQA dataset (tau/commonsense_qa). The 'rationales' for each QA pair were created using a two-part method. First, Mixtral (mistralai/Mixtral-8x7B-Instruct-v0.1) was used to generate 3 unique CoT (Chain-of-Thought) explanations. Next, human evaluation was applied to distill the random sampling down to a… See the full description on the dataset page: https://huggingface.co/datasets/peterkchung/commonsense_cot_partial_annotated_prelim.

  11. CommonsenseQA (Multiple-Choice Q&A)

    • kaggle.com
    zip
    Updated Nov 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2022). CommonsenseQA (Multiple-Choice Q&A) [Dataset]. https://www.kaggle.com/datasets/thedevastator/new-commonsenseqa-dataset-for-multiple-choice-qu
    Explore at:
    zip(712030 bytes)Available download formats
    Dataset updated
    Nov 21, 2022
    Authors
    The Devastator
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    CommonsenseQA (Multiple-Choice Q&A)

    12,102 questions with one correct answer and four distractor answers

    Source

    Huggingface Hub: link

    About this dataset

    CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers. The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation split, and "Question token split", see paper for details.

    How to use the dataset

    Research Ideas

    • This dataset can be used to train a model to predict the correct answers to multiple-choice questions.
    • This dataset can be used to evaluate the performance of different models on the CommonsenseQA dataset.
    • This dataset can be used to discover new types of commonsense knowledge required to predict the correct answers to questions in the CommonsenseQA dataset

    Acknowledgements

    License

    License: CC0 1.0 Universal (CC0 1.0) - Public Domain Dedication No Copyright - You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission. See Other Information.

    Columns

    File: validation.csv | Column name | Description | |:--------------|:---------------------------------------------------------------| | answerKey | The correct answer to the question. (String) | | choices | The four possible answers for each question. (List of strings) |

    File: train.csv | Column name | Description | |:--------------|:---------------------------------------------------------------| | answerKey | The correct answer to the question. (String) | | choices | The four possible answers for each question. (List of strings) |

    File: test.csv | Column name | Description | |:--------------|:---------------------------------------------------------------| | answerKey | The correct answer to the question. (String) | | choices | The four possible answers for each question. (List of strings) |

  12. T

    piqa

    • tensorflow.org
    Updated Dec 16, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). piqa [Dataset]. https://www.tensorflow.org/datasets/catalog/piqa
    Explore at:
    Dataset updated
    Dec 16, 2022
    Description

    Physical IQa: Physical Interaction QA, a new commonsense QA benchmark for naive physics reasoning focusing on how we interact with everyday objects in everyday situations. This dataset focuses on affordances of objects, i.e., what actions each physical object affords (e.g., it is possible to use a shoe as a doorstop), and what physical interactions a group of objects afford (e.g., it is possible to place an apple on top of a book, but not the other way around). The dataset requires reasoning about both the prototypical use of objects (e.g., shoes are used for walking) and non-prototypical but practically plausible use of objects (e.g., shoes can be used as a doorstop). The dataset includes 20,000 QA pairs that are either multiple-choice or true/false questions.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('piqa', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

  13. n

    reasoning_llama8b_preds_train_commonsense_qa

    • nerq.ai
    json
    Updated Mar 25, 2026
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    DaniilOr (2026). reasoning_llama8b_preds_train_commonsense_qa [Dataset]. https://nerq.ai/dataset/reasoning-llama8b-preds-train-commonsense-qa
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Mar 25, 2026
    Dataset authored and provided by
    DaniilOr
    Variables measured
    Downloads, Trust Score
    Description

    An AI tool for reasoning and common sense question-answering tasks.

  14. OpenBookQA (Multi-step Reasoning)

    • kaggle.com
    zip
    Updated Nov 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2022). OpenBookQA (Multi-step Reasoning) [Dataset]. https://www.kaggle.com/datasets/thedevastator/openbookqa-a-new-dataset-for-advanced-question-a
    Explore at:
    zip(826782 bytes)Available download formats
    Dataset updated
    Nov 21, 2022
    Authors
    The Devastator
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    OpenBookQA: A New Dataset for Advanced Question-Answering

    Multi-step Reasoning, Commonsense Knowledge, and Rich Text Comprehension

    Source

    Huggingface Hub: link

    About this dataset

    OpenBookQA aims to promote research in advanced question-answering, probing a deeper understanding of both the topic (with salient facts summarized as an open book, also provided with the dataset) and the language it is expressed in. In particular, it contains questions that require multi-step reasoning, use of additional common and commonsense knowledge, and rich text comprehension. OpenBookQA is a new kind of question-answering dataset modeled after open book exams for assessing human understanding of a subject.

    With OpenBookQA, we hope to push the boundaries of what current QA models can do and advance the state-of-the-art in this field. In addition to providing a challenging benchmark for existing models, we hope that this dataset will encourage new model architectures that can better handle complex questions and reasoning

    How to use the dataset

    Research Ideas

    • Questions that require multi-step reasoning,
    • Use of additional common and commonsense knowledge,
    • Rich text comprehension

    Acknowledgements

    License

    License: CC0 1.0 Universal (CC0 1.0) - Public Domain Dedication No Copyright - You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission. See Other Information.

    Columns

    File: main_test.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------------------------| | question_stem | The column 'question_stem' contains the stem of the question. (String) | | choices | The column 'choices' contains a list of answers to choose from. (List) | | answerKey | The column 'answerKey' contains the index of the correct answer in the choices list. (Integer) |

    File: main_train.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------------------------| | question_stem | The column 'question_stem' contains the stem of the question. (String) | | choices | The column 'choices' contains a list of answers to choose from. (List) | | answerKey | The column 'answerKey' contains the index of the correct answer in the choices list. (Integer) |

    File: additional_train.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------------------------| | question_stem | The column 'question_stem' contains the stem of the question. (String) | | choices | The column 'choices' contains a list of answers to choose from. (List) | | answerKey | The column 'answerKey' contains the index of the correct answer in the choices list. (Integer) |

    File: additional_test.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------------------------| | question_stem | The column 'question_stem' contains the stem of the question. (String) | | choices | The column 'choices' contains a list of answers to choose from. (List) | | answerKey | The column 'answerKey' contains the index of the correct answer in the choices list. (Integer) |

    File: additional_validation.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------------------------| | question_stem | The column 'question_stem' contains the stem of the question. (String) | | choices | The column 'choices' contains a list of answers to choose from. (List) | | answerKey | The column 'answerKey' contains the index of the correct answer in the choices list. (Integer) |

    **File: ...

  15. social_i_qa

    • huggingface.co
    • opendatalab.com
    Updated Aug 29, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ai2 (2023). social_i_qa [Dataset]. https://huggingface.co/datasets/allenai/social_i_qa
    Explore at:
    Dataset updated
    Aug 29, 2023
    Dataset provided by
    Allen Institute for AIhttp://allenai.org/
    Authors
    Ai2
    Description

    We introduce Social IQa: Social Interaction QA, a new question-answering benchmark for testing social commonsense intelligence. Contrary to many prior benchmarks that focus on physical or taxonomic knowledge, Social IQa focuses on reasoning about people’s actions and their social implications. For example, given an action like "Jesse saw a concert" and a question like "Why did Jesse do this?", humans can easily infer that Jesse wanted "to see their favorite performer" or "to enjoy the music", and not "to see what's happening inside" or "to see if it works". The actions in Social IQa span a wide variety of social situations, and answer candidates contain both human-curated answers and adversarially-filtered machine-generated candidates. Social IQa contains over 37,000 QA pairs for evaluating models’ abilities to reason about the social implications of everyday events and situations. (Less)

  16. n

    KA_tau_commonsense_qa_openai.gpt-3.5-turbo-0125_5_5_5

    • nerq.ai
    json
    Updated Mar 22, 2026
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    MorishT (2026). KA_tau_commonsense_qa_openai.gpt-3.5-turbo-0125_5_5_5 [Dataset]. https://nerq.ai/dataset/tau-commonsense-qa
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Mar 22, 2026
    Dataset authored and provided by
    MorishT
    Variables measured
    Downloads, Trust Score
    Description

    HuggingFace Dataset by MorishT

  17. JCommonsenseQA

    • huggingface.co
    Updated Apr 27, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SB Intuitions (2025). JCommonsenseQA [Dataset]. https://huggingface.co/datasets/sbintuitions/JCommonsenseQA
    Explore at:
    Dataset updated
    Apr 27, 2025
    Dataset provided by
    Authors
    SB Intuitions
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    評価スコアの再現性確保と SB Intuitions 修正版の公開用クローン ソース: yahoojapan/JGLUE on GitHub

    datasets/jcommonsenseqa-v1.1

      JCommonsenseQA
    

    JCommonsenseQA is a Japanese version of CommonsenseQA (Talmor+, 2019), which is a multiple-choice question answering dataset that requires commonsense reasoning ability. It is built using crowdsourcing with seeds extracted from the knowledge base ConceptNet.

      Licensing Information
    

    Creative Commons Attribution Share Alike 4.0 International

      Citation… See the full description on the dataset page: https://huggingface.co/datasets/sbintuitions/JCommonsenseQA.
    
  18. h

    chatgpt4-commonsense-qa

    • huggingface.co
    Updated Mar 2, 2026
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mesolitica (2026). chatgpt4-commonsense-qa [Dataset]. https://huggingface.co/datasets/mesolitica/chatgpt4-commonsense-qa
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 2, 2026
    Dataset authored and provided by
    Mesolitica
    Description

    Synthetic CommonSense

    Generated using ChatGPT4, originally from https://huggingface.co/datasets/commonsense_qa Notebook at https://github.com/mesolitica/malaysian-dataset/tree/master/question-answer/chatgpt4-commonsense

    synthetic-commonsense.jsonl, 36332 rows, 7.34 MB.

      Example data
    

    {'question': '1. Seseorang yang bersara mungkin perlu kembali bekerja jika mereka apa? A. mempunyai hutang B. mencari pendapatan C. meninggalkan pekerjaan D. memerlukan… See the full description on the dataset page: https://huggingface.co/datasets/mesolitica/chatgpt4-commonsense-qa.

  19. h

    StrategyQA

    • huggingface.co
    Updated Mar 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chi (2025). StrategyQA [Dataset]. https://huggingface.co/datasets/ChilleD/StrategyQA
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 11, 2025
    Authors
    Chi
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    ChilleD/StrategyQA dataset hosted on Hugging Face and contributed by the HF Datasets community

  20. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Tel Aviv University (2018). commonsense_qa [Dataset]. https://huggingface.co/datasets/tau/commonsense_qa

commonsense_qa

CommonsenseQA

tau/commonsense_qa

Explore at:
99 scholarly articles cite this dataset (View in Google Scholar)
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Nov 2, 2018
Dataset authored and provided by
Tel Aviv University
License

MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically

Description

Dataset Card for "commonsense_qa"

  Dataset Summary

CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers. The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation split, and "Question token split", see paper for details.… See the full description on the dataset page: https://huggingface.co/datasets/tau/commonsense_qa.

Search
Clear search
Close search
Google apps
Main menu