100+ datasets found
  1. Global ranking of LLM tools in 2023

    • statista.com
    Updated Jun 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Global ranking of LLM tools in 2023 [Dataset]. https://www.statista.com/statistics/1458138/leading-llm-tools/
    Explore at:
    Dataset updated
    Jun 25, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2023
    Area covered
    Worldwide
    Description

    In 2023, Claude 3 Opus was the large language model (LLM) tool that had the largest average worldwide, with an average total of ***** percent. Close behind, in second place, was Gemini 1.5 Pro with an average of about ** percent.

  2. h

    NeurIPS-LLM-data

    • huggingface.co
    Updated Mar 4, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Upaya (2024). NeurIPS-LLM-data [Dataset]. https://huggingface.co/datasets/upaya07/NeurIPS-LLM-data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 4, 2024
    Dataset authored and provided by
    Upaya
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    🤖 We curated this dataset for NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day. 🚀 Our Birbal-7B-V1 fine-tuned on this dataset achieved 🏆 first rank 🏆 in the competition.

    Here is high-level diagram of our data preparation strategy:

      Natural Instructions Dataset Preparation
    

    Natural Instructionsdataset is a community effort to create a large collection of tasks and their natural language definitions/instructions. As show in above diagram, we sample from… See the full description on the dataset page: https://huggingface.co/datasets/upaya07/NeurIPS-LLM-data.

  3. LLM market size in Japan FY 2024-2028

    • statista.com
    • ai-chatbox.pro
    Updated Jun 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). LLM market size in Japan FY 2024-2028 [Dataset]. https://www.statista.com/statistics/1550077/japan-large-language-model-market-size/
    Explore at:
    Dataset updated
    Jun 6, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Area covered
    Japan
    Description

    The value of the large language model (LLM) market in Japan was projected to reach ** billion Japanese yen in fiscal year 2024. Partly based on the assumption that the market would diversify with the release of specialized and cheaper LLMs from fiscal year 2025 onward, the market size was forecast to more than quadruple by fiscal year 2028.

  4. h

    long-llm-data

    • huggingface.co
    Updated Aug 25, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NaN (2024). long-llm-data [Dataset]. https://huggingface.co/datasets/namespace-Pt/long-llm-data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Aug 25, 2024
    Authors
    NaN
    Description

    namespace-Pt/long-llm-data dataset hosted on Hugging Face and contributed by the HF Datasets community

  5. d

    Image and Video Description Data | 1 PB | Multimodal Data | GenAI | LLM Data...

    • datarade.ai
    Updated Jan 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2025). Image and Video Description Data | 1 PB | Multimodal Data | GenAI | LLM Data | Large Language Model(LLM) Data| AI Datasets [Dataset]. https://datarade.ai/data-products/nexdata-image-and-video-description-data-1-pb-multimoda-nexdata
    Explore at:
    .bin, .json, .xml, .csv, .xls, .sql, .txtAvailable download formats
    Dataset updated
    Jan 3, 2025
    Dataset authored and provided by
    Nexdata
    Area covered
    Malta, Israel, Ecuador, Canada, United Arab Emirates, Mexico, Belgium, Czech Republic, Netherlands, Finland
    Description
    1. Image Description Data Data Size: 500 million pairs Image Type: generic scene(portrait, landscapes, animals,etc), human action, picture book, magazine, PPT&chart, App screenshot, and etc. Resolution: 4K+ Description Language: English, Spanish, Portuguese, French, Korean, German, Chinese, Japanese Description Length: text length is no less than 250 words Format: the image format is .jpg, the annotation format is .json, and the description format is .txt

    2. Video Description Data Data Size: 10 million pairs Image Type: generic scene(portrait, landscapes, animals,etc), ads, TV sports, documentaries Resolution: 1080p+ Description Language: English, Spanish, Portuguese, French, Korean, German, Chinese, Japanese Description Length: text length is no less than 250 words Format: .mp4,.mov,.avi and other common formats;.xlsx (annotation file format)

    3. About Nexdata Nexdata owns off-the-shelf PB-level Large Language Model(LLM) Data, 1 million hours of Audio Data and 800TB of Annotated Imagery Data. These ready-to-go data supports instant delivery, quickly improve the accuracy of AI models. For more details, please visit us at https://www.nexdata.ai/datasets/llm?source=Datarade

  6. Large Language Model Llm Market Report | Global Forecast From 2025 To 2033

    • dataintelo.com
    csv, pdf, pptx
    Updated Oct 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dataintelo (2024). Large Language Model Llm Market Report | Global Forecast From 2025 To 2033 [Dataset]. https://dataintelo.com/report/large-language-model-llm-market
    Explore at:
    pdf, pptx, csvAvailable download formats
    Dataset updated
    Oct 5, 2024
    Dataset authored and provided by
    Dataintelo
    License

    https://dataintelo.com/privacy-and-policyhttps://dataintelo.com/privacy-and-policy

    Time period covered
    2024 - 2032
    Area covered
    Global
    Description

    Large Language Model (LLM) Market Outlook



    The global market size for Large Language Models (LLMs) was valued at approximately USD 2.3 billion in 2023 and is projected to reach an astounding USD 15.8 billion by 2032, growing at a robust Compound Annual Growth Rate (CAGR) of 23.5%. The exponential growth of this market can be attributed to the increasing demand for AI-driven solutions across various sectors including healthcare, finance, and retail, among others. The rising adoption of natural language processing (NLP) technologies and advancements in machine learning algorithms are key factors driving this market.



    One of the major growth factors for the LLM market is the rapid development and adoption of artificial intelligence (AI) and machine learning technologies. The expanding capabilities of LLMs in understanding and generating human-like text have opened up new avenues for their application. This has led to increased investments in AI research and development, further propelling the advancements in LLM technologies. Moreover, the integration of LLMs with other advanced technologies such as cloud computing, big data, and IoT is enhancing their functionality and expanding their applicability across different sectors.



    Another crucial growth driver is the growing demand for automated customer service solutions. Businesses are increasingly deploying LLMs to improve customer engagement and satisfaction by providing instant, accurate, and personalized responses to customer queries. The ability of LLMs to understand and process natural language inputs makes them ideal for applications in chatbots, virtual assistants, and other automated customer service tools. This not only enhances customer experience but also significantly reduces operational costs for businesses by minimizing the need for human intervention.



    The healthcare sector is also witnessing a significant impact from the adoption of LLMs. These models are being utilized for various applications such as patient data management, diagnostics, and personalized medicine. The ability of LLMs to analyze large volumes of unstructured data and extract meaningful insights is revolutionizing the way healthcare providers deliver services. This is leading to improved patient outcomes, reduced medical errors, and more efficient healthcare delivery systems. Additionally, the ongoing advancements in AI technologies are expected to further enhance the capabilities of LLMs, driving their adoption in the healthcare sector.



    Regionally, North America is anticipated to dominate the LLM market, owing to the presence of major AI and technology companies, along with significant investments in AI research and development. The region's well-established IT infrastructure and high adoption rate of advanced technologies are further contributing to this growth. However, Asia Pacific is expected to witness the highest growth rate during the forecast period, driven by increasing digital transformation initiatives, rising investments in AI technology, and growing awareness about the benefits of LLMs in various applications.



    Component Analysis



    The LLM market can be segmented by component into software, hardware, and services. The software segment holds the largest share in the market, driven by the increasing demand for advanced AI software solutions that can leverage LLM capabilities. With the continuous advancements in machine learning algorithms and NLP technologies, the software segment is expected to maintain its dominance. Software solutions that incorporate LLMs are being used across various applications, from content generation to real-time language translation, making them indispensable tools for businesses and consumers alike.



    The hardware segment is also experiencing significant growth, as the deployment of LLMs requires substantial computational power. High-performance computing hardware, including Graphics Processing Units (GPUs) and Tensor Processing Units (TPUs), are essential for training and deploying LLMs. The increasing demand for powerful hardware solutions to support the computational requirements of LLMs is driving investments in this segment. Moreover, technological advancements in hardware components are enhancing the efficiency and performance of LLMs, further fueling their adoption.



    The services segment encompasses a wide range of offerings, including consulting, implementation, and maintenance services. As businesses increasingly adopt LLMs, the demand for specialized services to support the deployment and integration of these models is growing. Consulting services are

  7. Total funding of LLM developers worldwide in 2023

    • ai-chatbox.pro
    • statista.com
    Updated Mar 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2024). Total funding of LLM developers worldwide in 2023 [Dataset]. https://www.ai-chatbox.pro/?_=%2Fstatistics%2F1446568%2Fllm-developer-funding-2023%2F%23XgboD02vawLbpWJjSPEePEUG%2FVFd%2Bik%3D
    Explore at:
    Dataset updated
    Mar 19, 2024
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2023
    Area covered
    Worldwide
    Description

    OpenAI remains the most heavily funded large language model (LLM) developer worldwide in 2023. Its most well known product, ChatGPT, launched something of a generative AI revolution in late 2022. Its backing by Microsoft has made OpenAI a leading champion in the LLM competition, though competitors like Anthropic, funded by Google, are closing the gap.

  8. Ranking of LLM tools in solving math problems 2024

    • statista.com
    Updated Oct 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2024). Ranking of LLM tools in solving math problems 2024 [Dataset]. https://www.statista.com/statistics/1458141/leading-math-llm-tools/
    Explore at:
    Dataset updated
    Oct 25, 2024
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Mar 2024
    Area covered
    Worldwide
    Description

    As of March 2024, OpenAI o1 was the large language model (LLM) tool that had the best benchmark score in solving math problems, with a score of 94.8 percent. Close behind, in second place, was OpenAI o1-mini, followed by GPT-4o.

  9. m

    Large Language Model (LLM) Data | Machine Learning (ML) Data | AI Training...

    • data.mealme.ai
    Updated Jan 23, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    MealMe (2025). Large Language Model (LLM) Data | Machine Learning (ML) Data | AI Training Data (RAG) for 1M+ Global Grocery, Restaurant, and Retail Stores [Dataset]. https://data.mealme.ai/products/ai-training-data-rag-for-grocery-restaurant-and-retail-ra-mealme
    Explore at:
    Dataset updated
    Jan 23, 2025
    Dataset authored and provided by
    MealMe
    Area covered
    Uzbekistan, Venezuela, South Sudan, Madagascar, Somalia, Wallis and Futuna, Sao Tome and Principe, Austria, Bosnia and Herzegovina, Greenland
    Description

    Comprehensive training data on 1M+ stores across the US & Canada. Includes detailed menus, inventory, pricing, and availability. Ideal for AI/ML models, powering retrieval-augmented generation, search, and personalization systems.

  10. Energy consumption when training LLMs in 2022 (in MWh)

    • statista.com
    Updated Jun 30, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2024). Energy consumption when training LLMs in 2022 (in MWh) [Dataset]. https://www.statista.com/statistics/1384401/energy-use-when-training-llm-models/
    Explore at:
    Dataset updated
    Jun 30, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2022
    Area covered
    Worldwide
    Description

    Energy consumption of artificial intelligence (AI) models in training is considerable, with both GPT-3, the original release of the current iteration of OpenAI's popular ChatGPT, and Gopher consuming well over **********-megawatt hours of energy simply for training. As this is only for the training model it is likely that the energy consumption for the entire usage and lifetime of GPT-3 and other large language models (LLMs) is significantly higher. The largest consumer of energy, GPT-3, consumed roughly the equivalent of *** Germans in 2022. While not a staggering amount, it is a considerable use of energy. Energy savings through AI While it is undoubtedly true that training LLMs takes a considerable amount of energy, the energy savings are also likely to be substantial. Any AI model that improves processes by minute numbers might save hours on shipment, liters of fuel, or dozens of computations. Each one of these uses energy as well and the sum of energy saved through a LLM might vastly outperform its energy cost. A good example is mobile phone operators, of which a ***** expect that AI might reduce power consumption by *** to ******* percent. Considering that much of the world uses mobile phones this would be a considerable energy saver. Emissions are considerable The amount of CO2 emissions from training LLMs is also considerable, with GPT-3 producing nearly *** tonnes of CO2. This again could be radically changed based on the types of energy production creating the emissions. Most data center operators for instance would prefer to have nuclear energy play a key role, a significantly low-emission energy producer.

  11. Data and code from "Testing the reliability of a large language model to...

    • figshare.com
    xlsx
    Updated Jun 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andrew Gougherty (2024). Data and code from "Testing the reliability of a large language model to extract ecological information from the scientific literature" [Dataset]. http://doi.org/10.6084/m9.figshare.24646302.v2
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Jun 11, 2024
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Andrew Gougherty
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Note the LLM used in these scripts will be discontinued in July 2024. For updated versions of the scripts wrapped into an R package see https://github.com/agougher/geminus/datForReview - contains abstracts analyzed by LLMpalmExtract-forSubmission - R code for interacting with LLM

  12. B

    Data from: TruthEval: A Dataset to Evaluate LLM Truthfulness and Reliability...

    • borealisdata.ca
    Updated Jul 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aisha Khatun; Dan Brown (2024). TruthEval: A Dataset to Evaluate LLM Truthfulness and Reliability [Dataset]. http://doi.org/10.5683/SP3/5MZWBV
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 30, 2024
    Dataset provided by
    Borealis
    Authors
    Aisha Khatun; Dan Brown
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Large Language Model (LLM) evaluation is currently one of the most important areas of research, with existing benchmarks proving to be insufficient and not completely representative of LLMs' various capabilities. We present a curated collection of challenging statements on sensitive topics for LLM benchmarking called TruthEval. These statements were curated by hand and contain known truth values. The categories were chosen to distinguish LLMs' abilities from their stochastic nature. Details of collection method and use cases can be found in this paper: TruthEval: A Dataset to Evaluate LLM Truthfulness and Reliability

  13. deepseek-llm-7b-base

    • kaggle.com
    Updated Jan 30, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Younus_Mohamed (2025). deepseek-llm-7b-base [Dataset]. https://www.kaggle.com/datasets/younusmohamed/deepseek-llm-7b-base/data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 30, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Younus_Mohamed
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    DeepSeek Model Dataset

    Overview

    This dataset contains the DeepSeek model, a [brief description of the model, e.g., "state-of-the-art language model for natural language processing tasks"]. The model is designed for [specific use cases, e.g., "text generation, sentiment analysis, etc."].

    Contents

    • model_weights/: Directory containing the model weights.
    • config.json: Configuration file for the model.
    • inference_example.ipynb: Jupyter Notebook demonstrating how to load and use the model.
    • requirements.txt: List of Python dependencies.

    Usage

    1. Download the dataset from Kaggle.
    2. Install the required dependencies using pip install -r requirements.txt.
    3. Open the inference_example.ipynb notebook to see how to load the model and perform inference.

    License

    This dataset is licensed under [license name, e.g., "MIT License"]. See the LICENSE file for more details.

    Acknowledgments

    • Deepseek

    Contact

    For questions or issues, please contact

  14. h

    EverythingLM-data

    • huggingface.co
    Updated Aug 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kai Howard (2023). EverythingLM-data [Dataset]. https://huggingface.co/datasets/totally-not-an-llm/EverythingLM-data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Aug 4, 2023
    Authors
    Kai Howard
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    EverythingLM Dataset

    EverythingLM is a diverse instruct dataset consisting of ~1k sets of system prompts, instructions, and corresponding responses. These sets were generated using principles from both evol-instruct and Orca. The dataset encompasses a wide array of topics and interactions.

      Categories:
    

    Reasoning Creative Writing General Knowledge Brainstorming Search Query Coding Basic Instruct

    We also leverage various system prompts for evol-instruct and for responding… See the full description on the dataset page: https://huggingface.co/datasets/totally-not-an-llm/EverythingLM-data.

  15. Uses of LLMs in healthcare organizations in the United States 2024

    • statista.com
    Updated Jun 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Uses of LLMs in healthcare organizations in the United States 2024 [Dataset]. https://www.statista.com/statistics/1469378/uses-for-llm-use-in-healthcare-in-the-us/
    Explore at:
    Dataset updated
    Jun 24, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Feb 12, 2024 - Mar 15, 2024
    Area covered
    United States
    Description

    As of 2024, at least one fifth of respondents working in healthcare organizations reported that they used large language models for answering patient questions and medical chatbots. Furthermore, ** percent of healthcare organizations used LLMs for biomedical research.

  16. Firms planned LLM model usage in commercial deployments worldwide 2024

    • statista.com
    Updated Jun 26, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Firms planned LLM model usage in commercial deployments worldwide 2024 [Dataset]. https://www.statista.com/statistics/1485176/choice-of-llm-models-for-commercial-deployment-global/
    Explore at:
    Dataset updated
    Jun 26, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2024
    Area covered
    Worldwide
    Description

    As of 2024, over **** the global firms planned to use LLMs (LLama and LLama-like models), while ** percent chose to use embedding models (BERT and family) in their commercial deployment. Additionally, only ***** percent planned to utilize multi-modal models.

  17. D

    Data Annotation Tools Market Report

    • archivemarketresearch.com
    doc, pdf, ppt
    Updated Feb 18, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Archive Market Research (2025). Data Annotation Tools Market Report [Dataset]. https://www.archivemarketresearch.com/reports/data-annotation-tools-market-4890
    Explore at:
    pdf, doc, pptAvailable download formats
    Dataset updated
    Feb 18, 2025
    Dataset authored and provided by
    Archive Market Research
    License

    https://www.archivemarketresearch.com/privacy-policyhttps://www.archivemarketresearch.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    global
    Variables measured
    Market Size
    Description

    The Data Annotation Tools Market size was valued at USD 1.31 billion in 2023 and is projected to reach USD 6.72 billion by 2032, exhibiting a CAGR of 26.3 % during the forecasts period. The data annotation tools market is concerned with software applications that are used to tag as well as sort data for machine learning and artificial intelligence industries. They assist in development of training sets by tagging images, text, voice and video by relevant data and information. Some of the API’s that use reinforcement learning include training models for computer vision, natural language processing and speech recognition. Currently, tendencies in the market refer to the use of automated and semiautomated techniques for the process of annotation. Also, there is a rise in the demand for annotation tool with support for various form of data and support for AI marketing platforms. The application of AI and machine learning solutions in several industries is boosting the demand hence continues to propel the growth and competition in the market. Recent developments include: In November 2023, Appen Limited, a high-quality data provider for the AI lifecycle, chose Amazon Web Services (AWS) as its primary cloud for AI solutions and innovation. As Appen utilizes additional enterprise solutions for AI data source, annotation, and model validation, the firms are expanding their collaboration with a multi-year deal. Appen is strengthening its AI data platform, which serves as the bridge between people and AI, by integrating cutting-edge AWS services. , In September 2023, Labelbox launched Large Language Model (LLM) solution to assist organizations in innovating with generative AI and deepen the partnership with Google Cloud. With the introduction of large language models (LLMs), enterprises now have a plethora of chances to generate new competitive advantages and commercial value. LLM systems have the ability to revolutionize a wide range of intelligent applications; nevertheless, in many cases, organizations will need to adjust or finetune LLMs in order to align with human preferences. Labelbox, as part of an expanded cooperation, is leveraging Google Cloud's generative AI capabilities to assist organizations in developing LLM solutions with Vertex AI. Labelbox's AI platform will be integrated with Google Cloud's leading AI and Data Cloud tools, including Vertex AI and Google Cloud's Model Garden repository, allowing ML teams to access cutting-edge machine learning (ML) models for vision and natural language processing (NLP) and automate key workflows. , In March 2023, has released the most recent version of Enlitic Curie, a platform aimed at improving radiology department workflow. This platform includes Curie|ENDEX, which uses natural language processing and computer vision to analyze and process medical images, and Curie|ENCOG, which uses artificial intelligence to detect and protect medical images in Health Information Security. , In November 2022, Appen Limited, a global leader in data for the AI Lifecycle, announced its partnership with CLEAR Global, a nonprofit organization dedicated to ensuring access to essential information and amplifying voices across languages. This collaboration aims to develop a speech-based healthcare FAQ bot tailored for Sheng, a Nairobi slang language. .

  18. HiST-LLM

    • zenodo.org
    bin, json
    Updated Jan 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jakob Elias Hauser; Jakob Elias Hauser (2025). HiST-LLM [Dataset]. http://doi.org/10.5281/zenodo.14671248
    Explore at:
    bin, jsonAvailable download formats
    Dataset updated
    Jan 16, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jakob Elias Hauser; Jakob Elias Hauser
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Large Language Models' Expert-level Global History Knowledge Benchmark (HiST-LLM)

    Large Language Models (LLMs) have the potential to transform humanities and social science research, yet their history knowledge and comprehension at a graduate level remains untested. Benchmarking LLMs in history is particularly challenging, given that human knowledge of history is inherently unbalanced, with more information available on Western history and recent periods. We introduce the History Seshat Test for LLMs (Hist-LLM), based on a subset of the Seshat Global History Databank, which provides a structured representation of human historical knowledge, containing 36,000 data points across 600 historical societies and over 2,700 scholarly references. This dataset covers every major world region from the Neolithic period to the Industrial Revolution and includes information reviewed and assembled by history experts and graduate research assistants. Using this dataset, we benchmark a total of seven models from the Gemini, OpenAI, and Llama families. We find that, in a four-choice format, LLMs have a balanced accuracy ranging from 33.6% (Llama-3.1-8B) to 46% (GPT-4-Turbo), outperforming random guessing (25%) but falling short of expert comprehension. LLMs perform better on earlier historical periods. Regionally, performance is more even but still better for the Americas and lowest in Oceania and Sub-Saharan Africa for the more advanced models. Our benchmark shows that while LLMs possess some expert-level historical knowledge, there is considerable room for improvement.

    Dataset links

    Dataset Repository (Github)

    Croissant Metadata (Github)

    Usage

    This dataset can be used to benchmark LLMs on their expert level history knowledge.

    Loading the dataset

    using Python and Pandas:

    import pandas as pd
    main = pd.read_parquet("Neurips_HiST-LLM.parquet")
    ref = pd.read_parquet("references.parquet") 

    Dataset metadata

    Dataset metadata documented in the croissant.json file.

    Model Fingerprints

    When model fingerprint are available we created extra columns for each model fingerprint. These columns are named via the following pattern .

    Column Descriptions

    additional_review

    Boolean This column describes whether datapoints underwent additional expert review. See section 3.2 of the Paper.

    Q

    The multiple choice question.

    A

    The expected completion of the prompt.

    polity old id

    ID for polity according to Seshat ids.

    start year str

    String for when polity started existing (in BCE/CE format).

    end year str

    String for when polity stopped existing (in BCE/CE format).

    start year int

    Int for when polity started existing (in BCE/CE format).

    end year int

    Int for when polity stopped existing (in BCE/CE format).

    name

    Polity name.

    nga

    Natural Geographic Area for Polity.

    world_region

    The world region of a NGA (based on the UN regions with some modifications)

    category

    Immediate parent category of fact from Seshat codebook.

    root cat

    Major category of fact.

    value

    Value of data point.

    variable

    Variable of data point.

    id

    Request id for openai batch requests.

    description

    Description provided by RAs for fact.

  19. s

    Large Language Model (LLM) Data | 10 Million POI Average Noise Levels | 35 B...

    • storefront.silencio.network
    Updated Apr 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Silencio Network (2025). Large Language Model (LLM) Data | 10 Million POI Average Noise Levels | 35 B + Data Points | 100% Traceable Consent [Dataset]. https://storefront.silencio.network/products/ai-training-data-global-hyper-local-average-noise-levels-silencio-network
    Explore at:
    Dataset updated
    Apr 9, 2025
    Dataset provided by
    Quickkonnect UG
    Authors
    Silencio Network
    Area covered
    Anguilla, Central African Republic, Hungary, Svalbard and Jan Mayen, Mauritania, Timor-Leste, Faroe Islands, Uzbekistan, French Guiana, Chile
    Description

    Silencio provides the world’s largest real-world street and venue noise-level dataset, combining over 35 billion datapoints with AI-powered interpolation. Fully anonymized, user-consented, and ready for AI training, urban analysis, and mobility insights. Available in raw format.

  20. d

    TagX Data collection for AI/ ML training | LLM data | Data collection for AI...

    • datarade.ai
    .json, .csv, .xls
    Updated Jun 18, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    TagX (2021). TagX Data collection for AI/ ML training | LLM data | Data collection for AI development & model finetuning | Text, image, audio, and document data [Dataset]. https://datarade.ai/data-products/data-collection-and-capture-services-tagx
    Explore at:
    .json, .csv, .xlsAvailable download formats
    Dataset updated
    Jun 18, 2021
    Dataset authored and provided by
    TagX
    Area covered
    Equatorial Guinea, Belize, Saudi Arabia, Qatar, Iceland, Benin, Djibouti, Russian Federation, Colombia, Antigua and Barbuda
    Description

    We offer comprehensive data collection services that cater to a wide range of industries and applications. Whether you require image, audio, or text data, we have the expertise and resources to collect and deliver high-quality data that meets your specific requirements. Our data collection methods include manual collection, web scraping, and other automated techniques that ensure accuracy and completeness of data.

    Our team of experienced data collectors and quality assurance professionals ensure that the data is collected and processed according to the highest standards of quality. We also take great care to ensure that the data we collect is relevant and applicable to your use case. This means that you can rely on us to provide you with clean and useful data that can be used to train machine learning models, improve business processes, or conduct research.

    We are committed to delivering data in the format that you require. Whether you need raw data or a processed dataset, we can deliver the data in your preferred format, including CSV, JSON, or XML. We understand that every project is unique, and we work closely with our clients to ensure that we deliver the data that meets their specific needs. So if you need reliable data collection services for your next project, look no further than us.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Statista (2025). Global ranking of LLM tools in 2023 [Dataset]. https://www.statista.com/statistics/1458138/leading-llm-tools/
Organization logo

Global ranking of LLM tools in 2023

Explore at:
Dataset updated
Jun 25, 2025
Dataset authored and provided by
Statistahttp://statista.com/
Time period covered
2023
Area covered
Worldwide
Description

In 2023, Claude 3 Opus was the large language model (LLM) tool that had the largest average worldwide, with an average total of ***** percent. Close behind, in second place, was Gemini 1.5 Pro with an average of about ** percent.

Search
Clear search
Close search
Google apps
Main menu