65 datasets found
  1. ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact,...

    • figshare.com
    csv
    Updated May 8, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Davood Khodadad (2025). ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact, and Collaboration [Dataset]. http://doi.org/10.6084/m9.figshare.28536422.v1
    Explore at:
    csvAvailable download formats
    Dataset updated
    May 8, 2025
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Davood Khodadad
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Abstract:This dataset presents survey responses from first-year engineering students on their use of ChatGPT and other AI tools in a project-based learning environment. Collected as part of a study on AI’s role in engineering education, the data captures key insights into how students utilize ChatGPT for coding assistance, conceptual understanding, and collaborative work. The dataset includes responses on frequency of AI usage, perceived benefits and challenges, ethical concerns, and the impact of AI on learning outcomes and problem-solving skills.With AI increasingly integrated into education, this dataset provides valuable empirical evidence for researchers, educators, and policymakers interested in AI-assisted learning, STEM education, and academic integrity. It enables further analysis of student perceptions, responsible AI use, and the evolving role of generative AI in higher education.By making this dataset publicly available, we aim to support future research on AI literacy, pedagogy, and best practices for integrating AI into engineering and science curricula..................................................................................................................................................................Related PublicationThis dataset supports the findings presented in the following peer-reviewed article:ChatGPT in Engineering Education: A Breakthrough or a Challenge?Davood KhodadadPublished: 7 May 2025 | Physics Education, Volume 60, Number 4© 2025 The Author(s). Published by IOP Publishing LtdCitation: Davood Khodadad 2025 Phys. Educ. 60 045006DOI: 10.1088/1361-6552/add073If you use or reference this dataset, please consider citing the above publication......................................................................................................................................................................Description of the data and file structureTitle: ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact, and CollaborationDescription of Data Collection:This dataset was collected through a survey distributed via the Canvas learning platform following the completion of group projects in an introductory engineering course. The survey aimed to investigate how students engaged with ChatGPT and other AI tools in a project-based learning environment, particularly in relation to coding, report writing, idea generation, and collaboration.The survey consisted of 15 questions:12 multiple-choice questions to capture quantitative insights on AI usage patterns, frequency, and perceived benefits.3 open-ended questions to collect qualitative perspectives on challenges, ethical concerns, and students' reflections on AI-assisted learning.Key areas assessed in the survey include:Students’ prior familiarity with AI tools before the course.Frequency and purpose of ChatGPT usage (e.g., coding assistance, conceptual learning, collaboration).Perceived benefits and limitations of using AI tools in an engineering learning environment.Ethical considerations, including concerns about over-reliance and academic integrity.The dataset provides valuable empirical insights into the evolving role of AI in STEM education and can support further research on AI-assisted learning, responsible AI usage, and best practices for integrating AI tools in engineering education.

  2. Z

    A dataset to investigate ChatGPT for enhancing Students' Learning Experience...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jun 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Taibi, Davide (2024). A dataset to investigate ChatGPT for enhancing Students' Learning Experience via Concept Maps [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_12076680
    Explore at:
    Dataset updated
    Jun 19, 2024
    Dataset provided by
    Schicchi, Daniele
    Taibi, Davide
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset was compiled to examine the use of ChatGPT 3.5 in educational settings, particularly for creating and personalizing concept maps. The data has been organized into three folders: Maps, Texts, and Questionnaires. The Maps folder contains the graphical representation of the concept maps and the PlanUML code for drawing them in Italian and English. The Texts folder contains the source text used as input for the map's creation The Questionnaires folder includes the students' responses to the three administered questionnaires.

  3. S

    Test dataset of ChatGPT in medical field

    • scidb.cn
    Updated Mar 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    robin shen (2023). Test dataset of ChatGPT in medical field [Dataset]. http://doi.org/10.57760/sciencedb.o00130.00001
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 3, 2023
    Dataset provided by
    Science Data Bank
    Authors
    robin shen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The researcher tests the QA capability of ChatGPT in the medical field from the following aspects:1. Test their reserve capacity for medical knowledge2. Check their ability to read literature and understand medical literature3. Test their ability of auxiliary diagnosis after reading case data4. Test its error correction ability for case data5. Test its ability to standardize medical terms6. Test their evaluation ability to experts7. Check their ability to evaluate medical institutionsThe conclusion is:ChatGPT has great potential in the application of medical and health care, and may directly replace human beings or even professionals at a certain level in some fields;The researcher preliminarily believe that ChatGPT has basic medical knowledge and the ability of multiple rounds of dialogue, and its ability to understand Chinese is not weak;ChatGPT has the ability to read, understand and correct cases;ChatGPT has the ability of information extraction and terminology standardization, and is quite excellent;ChatGPT has the reasoning ability of medical knowledge;ChatGPT has the ability of continuous learning. After continuous training, its level has improved significantly;ChatGPT does not have the academic evaluation ability of Chinese medical talents, and the results are not ideal;ChatGPT does not have the academic evaluation ability of Chinese medical institutions, and the results are not ideal;ChatGPT is an epoch-making product, which can become a useful assistant for medical diagnosis and treatment, knowledge service, literature reading, review and paper writing.

  4. f

    Data_Sheet_1_Advanced large language models and visualization tools for data...

    • frontiersin.figshare.com
    txt
    Updated Aug 8, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jorge Valverde-Rebaza; Aram González; Octavio Navarro-Hinojosa; Julieta Noguez (2024). Data_Sheet_1_Advanced large language models and visualization tools for data analytics learning.csv [Dataset]. http://doi.org/10.3389/feduc.2024.1418006.s001
    Explore at:
    txtAvailable download formats
    Dataset updated
    Aug 8, 2024
    Dataset provided by
    Frontiers
    Authors
    Jorge Valverde-Rebaza; Aram González; Octavio Navarro-Hinojosa; Julieta Noguez
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    IntroductionIn recent years, numerous AI tools have been employed to equip learners with diverse technical skills such as coding, data analysis, and other competencies related to computational sciences. However, the desired outcomes have not been consistently achieved. This study aims to analyze the perspectives of students and professionals from non-computational fields on the use of generative AI tools, augmented with visualization support, to tackle data analytics projects. The focus is on promoting the development of coding skills and fostering a deep understanding of the solutions generated. Consequently, our research seeks to introduce innovative approaches for incorporating visualization and generative AI tools into educational practices.MethodsThis article examines how learners perform and their perspectives when using traditional tools vs. LLM-based tools to acquire data analytics skills. To explore this, we conducted a case study with a cohort of 59 participants among students and professionals without computational thinking skills. These participants developed a data analytics project in the context of a Data Analytics short session. Our case study focused on examining the participants' performance using traditional programming tools, ChatGPT, and LIDA with GPT as an advanced generative AI tool.ResultsThe results shown the transformative potential of approaches based on integrating advanced generative AI tools like GPT with specialized frameworks such as LIDA. The higher levels of participant preference indicate the superiority of these approaches over traditional development methods. Additionally, our findings suggest that the learning curves for the different approaches vary significantly. Since learners encountered technical difficulties in developing the project and interpreting the results. Our findings suggest that the integration of LIDA with GPT can significantly enhance the learning of advanced skills, especially those related to data analytics. We aim to establish this study as a foundation for the methodical adoption of generative AI tools in educational settings, paving the way for more effective and comprehensive training in these critical areas.DiscussionIt is important to highlight that when using general-purpose generative AI tools such as ChatGPT, users must be aware of the data analytics process and take responsibility for filtering out potential errors or incompleteness in the requirements of a data analytics project. These deficiencies can be mitigated by using more advanced tools specialized in supporting data analytics tasks, such as LIDA with GPT. However, users still need advanced programming knowledge to properly configure this connection via API. There is a significant opportunity for generative AI tools to improve their performance, providing accurate, complete, and convincing results for data analytics projects, thereby increasing user confidence in adopting these technologies. We hope this work underscores the opportunities and needs for integrating advanced LLMs into educational practices, particularly in developing computational thinking skills.

  5. Estimated water consumption for training GPT-3 2023

    • statista.com
    • ai-chatbox.pro
    Updated Nov 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2024). Estimated water consumption for training GPT-3 2023 [Dataset]. https://www.statista.com/statistics/1536925/gpt-3-estimated-water-consumption-training/
    Explore at:
    Dataset updated
    Nov 19, 2024
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Jul 2023
    Area covered
    Worldwide
    Description

    GPT-3's water consumption for the training phase was estimated at roughly 4.8 billion liters of water, when assuming the model was trained on Microsoft's Iowa data center (OpeanAI has disclosed that the data center was used for training parts of the GPT-4 model). If the model were to have been fully trained in the Washington data center, water consumption could have been as high as 15 billion liters. That would've amounted to more than Microsoft's total water withdrawals in 2023.

  6. i

    "ChatGPT vs. Student: A Dataset for Source Classification of Computer...

    • ieee-dataport.org
    Updated Jul 19, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ALI ABDULLAH S ALQAHTANI (2023). "ChatGPT vs. Student: A Dataset for Source Classification of Computer Science Answers [Dataset]. https://ieee-dataport.org/documents/chatgpt-vs-student-dataset-source-classification-computer-science-answers
    Explore at:
    Dataset updated
    Jul 19, 2023
    Authors
    ALI ABDULLAH S ALQAHTANI
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    along with the corresponding answers from students and ChatGPT.

  7. n

    A comparative evaluation of ChatGPT 3.5 and ChatGPT 4 in responses to...

    • data.niaid.nih.gov
    • zenodo.org
    • +1more
    zip
    Updated Jun 4, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Scott McGrath (2024). A comparative evaluation of ChatGPT 3.5 and ChatGPT 4 in responses to selected genetics questions - Full study data [Dataset]. http://doi.org/10.5061/dryad.s4mw6m9cv
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 4, 2024
    Dataset provided by
    University of California, Berkeley
    Authors
    Scott McGrath
    License

    https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html

    Description

    Objective: Our objective is to evaluate the efficacy of ChatGPT 4 in accurately and effectively delivering genetic information, building on previous findings with ChatGPT 3.5. We focus on assessing the utility, limitations, and ethical implications of using ChatGPT in medical settings. Materials and Methods: A structured questionnaire, including the Brief User Survey (BUS-15) and custom questions, was developed to assess ChatGPT 4's clinical value. An expert panel of genetic counselors and clinical geneticists independently evaluated ChatGPT 4's responses to these questions. We also involved comparative analysis with ChatGPT 3.5, utilizing descriptive statistics and using R for data analysis. Results: ChatGPT 4 demonstrated improvements over 3.5 in context recognition, relevance, and informativeness. However, performance variability and concerns about the naturalness of the output were noted. No significant difference in accuracy was found between ChatGPT 3.5 and 4.0. Notably, the efficacy of ChatGPT 4 varied significantly across different genetic conditions, with specific differences identified between responses related to BRCA1 and HFE. Discussion and Conclusion: This study highlights ChatGPT 4's potential in genomics, noting significant advancements over its predecessor. Despite these improvements, challenges remain, including the risk of outdated information and the necessity of ongoing refinement. The variability in performance across different genetic conditions underscores the need for expert oversight and continuous AI training. ChatGPT 4, while showing promise, emphasizes the importance of balancing technological innovation with ethical responsibility in healthcare information delivery. Methods Study Design This study was conducted to evaluate the performance of ChatGPT 4 (March 23rd, 2023) Model) in the context of genetic counseling and education. The evaluation involved a structured questionnaire, which included questions selected from the Brief User Survey (BUS-15) and additional custom questions designed to assess the clinical value of ChatGPT 4's responses. Questionnaire Development The questionnaire was built on Qualtrics, which comprised twelve questions: seven selected from the BUS-15 preceded by two additional questions that we designed. The initial questions focused on quality and answer relevancy: 1. The overall quality of the Chatbot’s response is: (5-point Likert: Very poor to Very Good) 2. The Chatbot delivered an answer that provided the relevant information you would include if asked the question. (5-point Likert: Strongly disagree to Strongly agree) The BUS-15 questions (7-point Likert: Strongly disagree to Strongly agree) focused on: 1. Recognition and facilitation of users’ goal and intent: Chatbot seems able to recognize the user’s intent and guide the user to its goals. 2. Relevance of information: The chatbot provides relevant and appropriate information/answer to people at each stage to make them closer to their goal. 3. Maxim of quantity: The chatbot responds in an informative way without adding too much information. 4. Resilience to failure: Chatbot seems able to find ways to respond appropriately even when it encounters situations or arguments it is not equipped to handle. 5. Understandability and politeness: The chatbot seems able to understand input and convey correct statements and answers without ambiguity and with acceptable manners. 6. Perceived conversational credibility: The chatbot responds in a credible and informative way without adding too much information. 7. Meet the neurodiverse needs: Chatbot seems able to meet needs and be used by users independently form their health conditions, well-being, age, etc. Expert Panel and Data Collection A panel of experts (two genetic counselors and two clinical geneticists) was provided with a link to the survey containing the questions. They independently evaluated the responses from ChatGPT 4 without discussing the questions or answers among themselves until after the survey submission. This approach ensured unbiased evaluation.

  8. m

    The Impact of AI and ChatGPT on Bangladeshi University Students

    • data.mendeley.com
    Updated Jan 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Md Jhirul Islam (2025). The Impact of AI and ChatGPT on Bangladeshi University Students [Dataset]. http://doi.org/10.17632/zykphpvbr7.2
    Explore at:
    Dataset updated
    Jan 6, 2025
    Authors
    Md Jhirul Islam
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Bangladesh
    Description

    The data set records the perceptions of Bangladeshi university students on the influence that AI tools, especially ChatGPT, have on their academic practices, learning experiences, and problem-solving abilities. The varying role of AI in education, which covers common usage statistics, what AI does to our creative abilities, its impact on our learning, and whether it could invade our privacy. This dataset reveals perspective on how AI tools are changing education in the country and offering valuable information for researchers, educators, policymakers, to understand trends, challenges, and opportunities in the adoption of AI in the academic contex.

    Methodology Data Collection Method: Online survey using google from Participants: A total of 3,512 students from various Bangladeshi universities participated. Survey Questions:The survey included questions on demographic information, frequency of AI tool usage, perceived benefits, concerns regarding privacy, and impacts on creativity and learning.

    Sampling Technique: Random sampling of university students Data Collection Period: June 2024 to December 2024

    Privacy Compliance This dataset has been anonymized to remove any personally identifiable information (PII). It adheres to relevant privacy regulations to ensure the confidentiality of participants.

    For further inquiries, please contact: Name: Md Jhirul Islam, Daffodil International University Email: jhirul15-4063@diu.edu.bd Phone: 01316317573

  9. Large Language Model content safety considerations text data

    • m.nexdata.ai
    • nexdata.ai
    Updated Jan 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2024). Large Language Model content safety considerations text data [Dataset]. https://m.nexdata.ai/datasets/llm/1349?source=Github
    Explore at:
    Dataset updated
    Jan 26, 2024
    Dataset authored and provided by
    Nexdata
    Variables measured
    Language, Data size, Data content, Storage format, Collecting type, Collecting method
    Description

    Large Language Model content safety considerations text data, about 570,000 in total, this dataset can be used for tasks such as LLM training, chatgpt

  10. h

    GPT-4-Prompts

    • huggingface.co
    Updated Dec 22, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Erfan zare chavoshi (2024). GPT-4-Prompts [Dataset]. https://huggingface.co/datasets/erfanzar/GPT-4-Prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 22, 2024
    Authors
    Erfan zare chavoshi
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Multi-Turn Conversational Prompts from ChatGPT-4 (10K+ Tokens) Abstract: This dataset offers a valuable collection of multi-turn conversational prompts generated by ChatGPT-4, carefully curated for diverse prompt styles (chatml, gemma, llama). Each prompt exceeds 10,000 tokens, providing ample context and inspiration for training and evaluating large language models. Ideal for researchers and developers interested in exploring advanced conversational AI capabilities. Table of Contents:… See the full description on the dataset page: https://huggingface.co/datasets/erfanzar/GPT-4-Prompts.

  11. m

    Data from: Artificial Intelligence Adoption Prediction Model: Would...

    • data.mendeley.com
    Updated Feb 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Christa van Staden (2024). Artificial Intelligence Adoption Prediction Model: Would ChatGPT-3.5 be adopted in English poetry classrooms? [Dataset]. http://doi.org/10.17632/289jtphg33.2
    Explore at:
    Dataset updated
    Feb 26, 2024
    Authors
    Christa van Staden
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is version 2 of the dataset created and used to explore ChatGPT-3.5's ability to write, justify and analyse English poems. This version was created after the reviewers decision that this paper may be published, if some changes are made.

    The purpose of the research was to determine if ChatGPT-3.5 would be adopted in English poetry classrooms. As none of the theoretical models were applicable, the Artificial Intelligence Adoption Prediction Model (AIAPM) was designed. Based on this model, an Artificial Intelligence Adoption Prediction tool (AIAPT) was designed to calculate an Adoption Prediction Score (APS). Then, ChatGPT-3.5's ability to write, justify and analyse poems were explored.

    It was found that ChatGPT-3.5 could write, justify, and analyse poems, but it could also make errors and hallucinate convincingly. Thus, the AIAPT was used to calculate the Adoption Prediction Score. The APS was 9, thus all factors of the AIAPM could drive the adoption decision. Thus, it could be predicted that ChatGPT-3.5 would be adopted in English poetry classrooms, both for ethical and unethical purposes. Based on the results, a few pro-active strategies were suggested.

    This dataset contains all data created and used during the research, including the poems which were integrated in the paper: "An Artificial Intelligence Adoption Prediction Model to determine if ChatGPT-3.5 would be adopted in English poetry classrooms" which was submitted toe Heliyon for publication.

  12. d

    Replication Data for: ChatGPT outperforms crowd-workers for text-annotation...

    • search.dataone.org
    Updated Nov 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gilardi, Fabrizio; Alizadeh, Meysam; Kubli, Maël (2023). Replication Data for: ChatGPT outperforms crowd-workers for text-annotation tasks [Dataset]. http://doi.org/10.7910/DVN/PQYF6M
    Explore at:
    Dataset updated
    Nov 8, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Gilardi, Fabrizio; Alizadeh, Meysam; Kubli, Maël
    Description

    Many NLP applications require manual text annotations for a variety of tasks, notably to train classifiers or evaluate the performance of unsupervised models. Depending on the size and degree of complexity, the tasks may be conducted by crowd-workers on platforms such as MTurk as well as trained annotators, such as research assistants. Using four samples of tweets and news articles (n = 6,183), we show that ChatGPT outperforms crowd-workers for several annotation tasks, including relevance, stance, topics, and frame detection. Across the four datasets, the zero-shot accuracy of ChatGPT exceeds that of crowd-workers by about 25 percentage points on average, while ChatGPT's intercoder agreement exceeds that of both crowd-workers and trained annotators for all tasks. Moreover, the per-annotation cost of ChatGPT is less than $0.003---about thirty times cheaper than MTurk. These results demonstrate the potential of large language models to drastically increase the efficiency of text classification.

  13. m

    Date Set: ChatGPT as an education and learning tool for engineering,...

    • data.mendeley.com
    Updated Jun 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    RAVINDRA BHARDWAJ (2025). Date Set: ChatGPT as an education and learning tool for engineering, technology and general studies: performance analysis of ChatGPT 3.0 on CSE, GATE and JEE examinations of India [Dataset]. http://doi.org/10.17632/995zwcz5yt.2
    Explore at:
    Dataset updated
    Jun 25, 2025
    Authors
    RAVINDRA BHARDWAJ
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    India
    Description

    This is the raw data that is used in the publication: ChatGPT as an education and learning tool for engineering, technology and general studies: performance analysis of ChatGPT 3.0 on CSE, GATE and JEE examinations of India.

  14. Z

    Data from: ChatGPT's performance in dentistry and allergy-immunology...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Sep 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Eggmann, Florin (2023). ChatGPT's performance in dentistry and allergy-immunology assessments: a comparative study [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_8331146
    Explore at:
    Dataset updated
    Sep 30, 2023
    Dataset provided by
    Eggmann, Florin
    Weiger, Roland
    Trachsel, Tina
    Fuchs, Alexander
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Data on ChatGPT 3's and ChatGPT 4's performance on self-assessment questions for dentistry (SFLEDM) and allergy and clinical immunology (EEAACI), sourced from the University of Bern’s Institute for Medical Education platform.

  15. IT data center systems total spending worldwide 2012-2025

    • statista.com
    • ai-chatbox.pro
    Updated Jun 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). IT data center systems total spending worldwide 2012-2025 [Dataset]. https://www.statista.com/statistics/314596/total-data-center-systems-worldwide-spending-forecast/
    Explore at:
    Dataset updated
    Jun 25, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Area covered
    Worldwide
    Description

    Worldwide spending on data center systems is projected to reach over, *** billion U.S. dollars in 2025, marking a significant ** percent increase from 2024. This growth reflects the ongoing digital transformation across industries and the increasing demand for advanced computing capabilities. The surge in data center investments is closely tied to the rapid expansion of artificial intelligence technologies, particularly with the wake of generative AI. AI chips fuel market growth The rise in data center spending aligns with the booming AI chip market, which is expected to reach ** billion U.S. dollars by 2025. Nvidia has emerged as a leader in this space, with its data center revenue skyrocketing due to the crucial role its GPUs play in training and running large language models like ChatGPT. The global GPU market, valued at ** billion U.S. dollars in 2024, is a key driver of this growth, powering advancements in machine learning and deep learning applications. Semiconductor industry adapts to AI demands The broader semiconductor industry is also evolving to meet the demands of AI technologies. With global semiconductor revenues surpassing *** billion U.S. dollars in 2023, the market is expected to approach *** billion U.S. dollars in 2024. AI chips are becoming increasingly prevalent in servers, data centers and storage infrastructures. This trend is reflected in the data centers and storage semiconductor market, which is projected to grow from ** billion U.S. dollars in 2023 to *** billion U.S. dollars by 2025, driven by the development of image sensors and edge AI processors.

  16. ChatGPT Evaluation Dataset v.2.0

    • zenodo.org
    • data.niaid.nih.gov
    Updated Oct 31, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jan Kocoń; Jan Kocoń; Przemysław Kazienko; Przemysław Kazienko (2024). ChatGPT Evaluation Dataset v.2.0 [Dataset]. http://doi.org/10.5281/zenodo.14019715
    Explore at:
    Dataset updated
    Oct 31, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jan Kocoń; Jan Kocoń; Przemysław Kazienko; Przemysław Kazienko
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Oct 2023
    Description

    We tested ChatGPT on 25 tasks focusing on solving common NLP problems and requiring analytical reasoning. These tasks include (1) a relatively simple binary classification of texts like spam, humor, sarcasm, aggression detection, or grammatical correctness of the text; (2) a more complex multiclass and multi-label classification of texts such as sentiment analysis, emotion recognition; (3) reasoning with the personal context, i.e., personalized versions of the problems that make use of additional information about text perception of a given user (user’s examples provided to ChatGPT); (4) semantic annotation and acceptance of the text going towards natural language understanding (NLU) like word sense disambiguation (WSD), and (5) answering questions based on the input text. More information in the paper: https://www.sciencedirect.com/science/article/pii/S156625352300177X

  17. Energy consumption when training LLMs in 2022 (in MWh)

    • statista.com
    Updated Jun 30, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Energy consumption when training LLMs in 2022 (in MWh) [Dataset]. https://www.statista.com/statistics/1384401/energy-use-when-training-llm-models/
    Explore at:
    Dataset updated
    Jun 30, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2022
    Area covered
    Worldwide
    Description

    Energy consumption of artificial intelligence (AI) models in training is considerable, with both GPT-3, the original release of the current iteration of OpenAI's popular ChatGPT, and Gopher consuming well over **********-megawatt hours of energy simply for training. As this is only for the training model it is likely that the energy consumption for the entire usage and lifetime of GPT-3 and other large language models (LLMs) is significantly higher. The largest consumer of energy, GPT-3, consumed roughly the equivalent of *** Germans in 2022. While not a staggering amount, it is a considerable use of energy. Energy savings through AI While it is undoubtedly true that training LLMs takes a considerable amount of energy, the energy savings are also likely to be substantial. Any AI model that improves processes by minute numbers might save hours on shipment, liters of fuel, or dozens of computations. Each one of these uses energy as well and the sum of energy saved through a LLM might vastly outperform its energy cost. A good example is mobile phone operators, of which a ***** expect that AI might reduce power consumption by *** to ******* percent. Considering that much of the world uses mobile phones this would be a considerable energy saver. Emissions are considerable The amount of CO2 emissions from training LLMs is also considerable, with GPT-3 producing nearly *** tonnes of CO2. This again could be radically changed based on the types of energy production creating the emissions. Most data center operators for instance would prefer to have nuclear energy play a key role, a significantly low-emission energy producer.

  18. f

    Data from: Enhancing self-directed learning with custom GPT AI facilitation...

    • tandf.figshare.com
    docx
    Updated Jun 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Wang Shalong; Zuo Yi; Zou Bin; Liu Ganglei; Zhou Jinyu; Zheng Yanwen; Zhang Zequn; Yuan Lianwen; Ren Feng (2025). Enhancing self-directed learning with custom GPT AI facilitation among medical students: A randomized controlled trial [Dataset]. http://doi.org/10.6084/m9.figshare.29039163.v1
    Explore at:
    docxAvailable download formats
    Dataset updated
    Jun 17, 2025
    Dataset provided by
    Taylor & Francis
    Authors
    Wang Shalong; Zuo Yi; Zou Bin; Liu Ganglei; Zhou Jinyu; Zheng Yanwen; Zhang Zequn; Yuan Lianwen; Ren Feng
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This study aims to assess the impact of LearnGuide, a specialized ChatGPT tool designed to support self-directed learning among medical students. In this 14-week randomized controlled trial (ClinicalTrials.gov NCT06276049), 103 medical students were assigned to either an intervention group, which received 12 weeks of problem-based training with LearnGuide support, or a control group, which received identical training without AI assistance. Primary and secondary outcomes, including Self-Directed Learning Scale scores at 6 and 12 weeks, Cornell Critical Thinking Test Level Z scores, and Global Flow Scores, were evaluated with a 14-week follow-up. Mann-Whitney U tests were used for statistical comparisons between the groups. At 6 weeks, the intervention group showed a marginally higher median Self-Directed Learning Scale score, which further improved by 12 weeks (4.15 [95% CI, 0.82 to 7.48]; p = 0.01) and was sustained at the 14-week follow-up. Additionally, this group demonstrated notable improvements in the Cornell Critical Thinking Test Score at 12 weeks (7.11 [95% CI, 4.50 to 9.72]; p 

  19. f

    Data Sheet 1_Large language models generating synthetic clinical datasets: a...

    • frontiersin.figshare.com
    xlsx
    Updated Feb 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Austin A. Barr; Joshua Quan; Eddie Guo; Emre Sezgin (2025). Data Sheet 1_Large language models generating synthetic clinical datasets: a feasibility and comparative analysis with real-world perioperative data.xlsx [Dataset]. http://doi.org/10.3389/frai.2025.1533508.s001
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Feb 5, 2025
    Dataset provided by
    Frontiers
    Authors
    Austin A. Barr; Joshua Quan; Eddie Guo; Emre Sezgin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    BackgroundClinical data is instrumental to medical research, machine learning (ML) model development, and advancing surgical care, but access is often constrained by privacy regulations and missing data. Synthetic data offers a promising solution to preserve privacy while enabling broader data access. Recent advances in large language models (LLMs) provide an opportunity to generate synthetic data with reduced reliance on domain expertise, computational resources, and pre-training.ObjectiveThis study aims to assess the feasibility of generating realistic tabular clinical data with OpenAI’s GPT-4o using zero-shot prompting, and evaluate the fidelity of LLM-generated data by comparing its statistical properties to the Vital Signs DataBase (VitalDB), a real-world open-source perioperative dataset.MethodsIn Phase 1, GPT-4o was prompted to generate a dataset with qualitative descriptions of 13 clinical parameters. The resultant data was assessed for general errors, plausibility of outputs, and cross-verification of related parameters. In Phase 2, GPT-4o was prompted to generate a dataset using descriptive statistics of the VitalDB dataset. Fidelity was assessed using two-sample t-tests, two-sample proportion tests, and 95% confidence interval (CI) overlap.ResultsIn Phase 1, GPT-4o generated a complete and structured dataset comprising 6,166 case files. The dataset was plausible in range and correctly calculated body mass index for all case files based on respective heights and weights. Statistical comparison between the LLM-generated datasets and VitalDB revealed that Phase 2 data achieved significant fidelity. Phase 2 data demonstrated statistical similarity in 12/13 (92.31%) parameters, whereby no statistically significant differences were observed in 6/6 (100.0%) categorical/binary and 6/7 (85.71%) continuous parameters. Overlap of 95% CIs were observed in 6/7 (85.71%) continuous parameters.ConclusionZero-shot prompting with GPT-4o can generate realistic tabular synthetic datasets, which can replicate key statistical properties of real-world perioperative data. This study highlights the potential of LLMs as a novel and accessible modality for synthetic data generation, which may address critical barriers in clinical data access and eliminate the need for technical expertise, extensive computational resources, and pre-training. Further research is warranted to enhance fidelity and investigate the use of LLMs to amplify and augment datasets, preserve multivariate relationships, and train robust ML models.

  20. o

    Data from: Comparing ChatGPT’s ability to rate the degree of stereotypes and...

    • ourarchive.otago.ac.nz
    Updated Jul 12, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chao-Cheng Lin; Zaine Akuhata-Huntington; Che-Wei Hsu (2023). Comparing ChatGPT’s ability to rate the degree of stereotypes and the consistency of stereotype attribution with those of medical students in New Zealand in developing a similarity rating test: a methodological study [Dataset]. https://ourarchive.otago.ac.nz/esploro/outputs/dataset/Comparing-ChatGPTs-ability-to-rate-the/9926722371001891
    Explore at:
    Dataset updated
    Jul 12, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Chao-Cheng Lin; Zaine Akuhata-Huntington; Che-Wei Hsu
    Time period covered
    Jul 12, 2023
    Area covered
    New Zealand
    Description

    Dataset for Lin, C.-C., Akuhata-Huntington, Z., & Hsu, C.-W. (2023). Comparing ChatGPT’s ability to rate the degree of stereotypes and the consistency of stereotype attribution with those of medical students in New Zealand in developing a similarity rating test: a methodological study. Journal of Educational Evaluation for Health Professions, 20:17. https://doi.org/10.3352/jeehp.2023.20.17. Learning about one’s implicit bias is crucial for improving one’s cultural competency and thereby reducing health inequity. To evaluate bias among medical students following a previously developed cultural training program targeting New Zealand Māori, we developed a text-based, self-evaluation tool called the Similarity Rating Test (SRT).

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Davood Khodadad (2025). ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact, and Collaboration [Dataset]. http://doi.org/10.6084/m9.figshare.28536422.v1
Organization logo

ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact, and Collaboration

Explore at:
csvAvailable download formats
Dataset updated
May 8, 2025
Dataset provided by
Figsharehttp://figshare.com/
Authors
Davood Khodadad
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

Abstract:This dataset presents survey responses from first-year engineering students on their use of ChatGPT and other AI tools in a project-based learning environment. Collected as part of a study on AI’s role in engineering education, the data captures key insights into how students utilize ChatGPT for coding assistance, conceptual understanding, and collaborative work. The dataset includes responses on frequency of AI usage, perceived benefits and challenges, ethical concerns, and the impact of AI on learning outcomes and problem-solving skills.With AI increasingly integrated into education, this dataset provides valuable empirical evidence for researchers, educators, and policymakers interested in AI-assisted learning, STEM education, and academic integrity. It enables further analysis of student perceptions, responsible AI use, and the evolving role of generative AI in higher education.By making this dataset publicly available, we aim to support future research on AI literacy, pedagogy, and best practices for integrating AI into engineering and science curricula..................................................................................................................................................................Related PublicationThis dataset supports the findings presented in the following peer-reviewed article:ChatGPT in Engineering Education: A Breakthrough or a Challenge?Davood KhodadadPublished: 7 May 2025 | Physics Education, Volume 60, Number 4© 2025 The Author(s). Published by IOP Publishing LtdCitation: Davood Khodadad 2025 Phys. Educ. 60 045006DOI: 10.1088/1361-6552/add073If you use or reference this dataset, please consider citing the above publication......................................................................................................................................................................Description of the data and file structureTitle: ChatGPT in Engineering Education: Survey Data on AI Usage, Learning Impact, and CollaborationDescription of Data Collection:This dataset was collected through a survey distributed via the Canvas learning platform following the completion of group projects in an introductory engineering course. The survey aimed to investigate how students engaged with ChatGPT and other AI tools in a project-based learning environment, particularly in relation to coding, report writing, idea generation, and collaboration.The survey consisted of 15 questions:12 multiple-choice questions to capture quantitative insights on AI usage patterns, frequency, and perceived benefits.3 open-ended questions to collect qualitative perspectives on challenges, ethical concerns, and students' reflections on AI-assisted learning.Key areas assessed in the survey include:Students’ prior familiarity with AI tools before the course.Frequency and purpose of ChatGPT usage (e.g., coding assistance, conceptual learning, collaboration).Perceived benefits and limitations of using AI tools in an engineering learning environment.Ethical considerations, including concerns about over-reliance and academic integrity.The dataset provides valuable empirical insights into the evolving role of AI in STEM education and can support further research on AI-assisted learning, responsible AI usage, and best practices for integrating AI tools in engineering education.

Search
Clear search
Close search
Google apps
Main menu