100+ datasets found
  1. h

    llm-system-prompts-benchmark

    • huggingface.co
    Updated 10 thg 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Naomi Bashkansky (2024). llm-system-prompts-benchmark [Dataset]. https://huggingface.co/datasets/Naomibas/llm-system-prompts-benchmark
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    10 thg 1, 2024
    Authors
    Naomi Bashkansky
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Dataset Card for Dataset Name

    This datset is a collection of 100 system prompts for large language models.

      Dataset Details
    
    
    
    
    
      Dataset Description
    

    These 100 system prompts test a model's ability to follow grammatical patterns; answer basic multiple choice questions; act according to a particular persona; memorize information; and speak in French. Files:

    hundred_system_prompts.py: refer to this to see the (prompt, probe, function) triplets, as well as the… See the full description on the dataset page: https://huggingface.co/datasets/Naomibas/llm-system-prompts-benchmark.

  2. h

    malicious-llm-prompts

    • huggingface.co
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sagar Patel, malicious-llm-prompts [Dataset]. https://huggingface.co/datasets/codesagar/malicious-llm-prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Authors
    Sagar Patel
    Description

    Dataset Card for "malicious-llm-prompts"

    More Information needed

  3. h

    paper-llm-prompts

    • huggingface.co
    Updated 3 thg 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    B F (2023). paper-llm-prompts [Dataset]. https://huggingface.co/datasets/beephids/paper-llm-prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    3 thg 7, 2023
    Authors
    B F
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    beephids/paper-llm-prompts dataset hosted on Hugging Face and contributed by the HF Datasets community

  4. h

    synthetic_multilingual_llm_prompts

    • huggingface.co
    Updated 11 thg 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gretel.ai (2024). synthetic_multilingual_llm_prompts [Dataset]. https://huggingface.co/datasets/gretelai/synthetic_multilingual_llm_prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    11 thg 6, 2024
    Dataset provided by
    Gretel.ai
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Image generated by DALL-E. See prompt for more details

      📝🌐 Synthetic Multilingual LLM Prompts
    

    Welcome to the "Synthetic Multilingual LLM Prompts" dataset! This comprehensive collection features 1,250 synthetic LLM prompts generated using Gretel Navigator, available in seven different languages. To ensure accuracy and diversity in prompts, and translation quality and consistency across the different languages, we employed Gretel Navigator both as a generation tool and as an… See the full description on the dataset page: https://huggingface.co/datasets/gretelai/synthetic_multilingual_llm_prompts.

  5. h

    Official_LLM_System_Prompts

    • huggingface.co
    Updated 4 thg 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nymbo (2025). Official_LLM_System_Prompts [Dataset]. https://huggingface.co/datasets/Nymbo/Official_LLM_System_Prompts
    Explore at:
    Dataset updated
    4 thg 11, 2025
    Authors
    Nymbo
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Official LLM System Prompts

    This short dataset contains a few system prompts leaked from proprietary models. Contains date-stamped prompts from OpenAI, Anthropic, MS Copilot, GitHub Copilot, Grok, and Perplexity.

  6. h

    LLM_Math

    • huggingface.co
    Updated 18 thg 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    LangChain Hub Prompts (2023). LLM_Math [Dataset]. https://huggingface.co/datasets/LangChainHub-Prompts/LLM_Math
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    18 thg 1, 2023
    Dataset authored and provided by
    LangChain Hub Prompts
    Description

    Description of LLM Math

    Prompt designed to optionally output iPython syntax to be run in order to better answer math questions.

      Inputs
    

    This is a description of the inputs that the prompt expects. question: User question to be answered.

      Usage
    

    Below is a code snippet for how to use the prompt. from langchain.prompts import load_prompt from langchain.chains import LLMMathChain

    llm = ... prompt = load_prompt('lc://prompts/llm_math/

  7. h

    awesome-chatgpt-prompts

    • huggingface.co
    Updated 15 thg 12, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Fatih Kadir Akın (2023). awesome-chatgpt-prompts [Dataset]. https://huggingface.co/datasets/fka/awesome-chatgpt-prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    15 thg 12, 2023
    Authors
    Fatih Kadir Akın
    License

    https://choosealicense.com/licenses/cc0-1.0/https://choosealicense.com/licenses/cc0-1.0/

    Description

    🧠 Awesome ChatGPT Prompts [CSV dataset]

    This is a Dataset Repository of Awesome ChatGPT Prompts View All Prompts on GitHub

      License
    

    CC-0

  8. h

    JailbreakPrompts

    • huggingface.co
    Updated 26 thg 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Simon Knuts (2025). JailbreakPrompts [Dataset]. https://huggingface.co/datasets/Simsonsun/JailbreakPrompts
    Explore at:
    Dataset updated
    26 thg 6, 2025
    Authors
    Simon Knuts
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Independent Jailbreak Datasets for LLM Guardrail Evaluation

    Constructed for the thesis:“Contamination Effects: How Training Data Leakage Affects Red Team Evaluation of LLM Jailbreak Detection” The effectiveness of LLM guardrails is commonly evaluated using open-source red teaming tools. However, this study reveals that significant data contamination exists between the training sets of binary jailbreak classifiers (ProtectAI, Katanemo, TestSavantAI, etc.) and the test prompts used in… See the full description on the dataset page: https://huggingface.co/datasets/Simsonsun/JailbreakPrompts.

  9. h

    in-the-wild-jailbreak-prompts

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    TrustAIRLab, in-the-wild-jailbreak-prompts [Dataset]. https://huggingface.co/datasets/TrustAIRLab/in-the-wild-jailbreak-prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset authored and provided by
    TrustAIRLab
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    In-The-Wild Jailbreak Prompts on LLMs

    This is the official repository for the ACM CCS 2024 paper "Do Anything Now'': Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models by Xinyue Shen, Zeyuan Chen, Michael Backes, Yun Shen, and Yang Zhang. In this project, employing our new framework JailbreakHub, we conduct the first measurement study on jailbreak prompts in the wild, with 15,140 prompts collected from December 2022 to December 2023 (including 1,405… See the full description on the dataset page: https://huggingface.co/datasets/TrustAIRLab/in-the-wild-jailbreak-prompts.

  10. h

    self-align-prompts

    • huggingface.co
    Updated 31 thg 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    mii-llm (2024). self-align-prompts [Dataset]. https://huggingface.co/datasets/mii-llm/self-align-prompts
    Explore at:
    Dataset updated
    31 thg 7, 2024
    Dataset authored and provided by
    mii-llm
    Description

    mii-llm/self-align-prompts dataset hosted on Hugging Face and contributed by the HF Datasets community

  11. h

    System-Prompt-Library-030825

    • huggingface.co
    Updated 17 thg 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daniel Rosehill (2025). System-Prompt-Library-030825 [Dataset]. http://doi.org/10.57967/hf/6319
    Explore at:
    Dataset updated
    17 thg 10, 2025
    Authors
    Daniel Rosehill
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    System Prompts Dataset - August 2025

    Point-in-time export from Daniel Rosehill's system prompt library as of August 3rd, 2025

      Overview
    

    This repository contains a comprehensive collection of 944 system prompts designed for various AI applications, agent workflows, and conversational AI systems. While many of these prompts now serve as the foundation for more complex agent-based workflows, they continue to provide essential building blocks for AI system design and… See the full description on the dataset page: https://huggingface.co/datasets/danielrosehill/System-Prompt-Library-030825.

  12. h

    llm-judge-eval-n5-new-prompt

    • huggingface.co
    Updated 14 thg 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ST Projects (2025). llm-judge-eval-n5-new-prompt [Dataset]. https://huggingface.co/datasets/st-projects/llm-judge-eval-n5-new-prompt
    Explore at:
    Dataset updated
    14 thg 9, 2025
    Dataset authored and provided by
    ST Projects
    Description

    st-projects/llm-judge-eval-n5-new-prompt dataset hosted on Hugging Face and contributed by the HF Datasets community

  13. h

    deepseek-r1-reasoning-prompts

    • huggingface.co
    Updated 27 thg 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    umar igan (2025). deepseek-r1-reasoning-prompts [Dataset]. https://huggingface.co/datasets/umarigan/deepseek-r1-reasoning-prompts
    Explore at:
    Dataset updated
    27 thg 1, 2025
    Authors
    umar igan
    Description

    I created a reasoning prompt dataset from deepseek-r1 model with the purpose of fine-tuning small language models to use them to generate better reasoning prompt to use with bigger llm models.

      Metadata
    

    The metadata is made available through a series of parquet files with the following schema:

    id: A unique identifier for the qa. question: answer: Answer from deepseek-r1 think model. reasoning: Reasoning from deepseek-r1 model.

  14. h

    llm-prompt-recovery

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tuhina Tripathi, llm-prompt-recovery [Dataset]. https://huggingface.co/datasets/tuhinatripathi/llm-prompt-recovery
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Authors
    Tuhina Tripathi
    License

    https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/

    Description

    tuhinatripathi/llm-prompt-recovery dataset hosted on Hugging Face and contributed by the HF Datasets community

  15. h

    speedrender-llm-prompt

    • huggingface.co
    Updated 12 thg 2, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kaique Pereira (2025). speedrender-llm-prompt [Dataset]. https://huggingface.co/datasets/kaiquedu/speedrender-llm-prompt
    Explore at:
    Dataset updated
    12 thg 2, 2025
    Authors
    Kaique Pereira
    Description

    kaiquedu/speedrender-llm-prompt dataset hosted on Hugging Face and contributed by the HF Datasets community

  16. h

    vigil-instruction-bypass-all-MiniLM-L6-v2

    • huggingface.co
    Updated 16 thg 10, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Adam Swanda (2023). vigil-instruction-bypass-all-MiniLM-L6-v2 [Dataset]. https://huggingface.co/datasets/deadbits/vigil-instruction-bypass-all-MiniLM-L6-v2
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    16 thg 10, 2023
    Authors
    Adam Swanda
    Description

    Vigil: LLM Instruction Bypass all-MiniLM-L6-v2

    Repo: github.com/deadbits/vigil-llm

    Vigil is a Python framework and REST API for assessing Large Language Model (LLM) prompts against a set of scanners to detect prompt injections, jailbreaks, and other potentially risky inputs. This repository contains all-MiniLM-L6-v2 embeddings for all Instruction Bypass style prompts ("Ignore instructions ...") used by Vigil. You can use the parquet2vdb.py utility to load the embeddings in the… See the full description on the dataset page: https://huggingface.co/datasets/deadbits/vigil-instruction-bypass-all-MiniLM-L6-v2.

  17. h

    prompt-injections

    • huggingface.co
    Updated 4 thg 5, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    HSE LLM @ Saint Petersburg (2025). prompt-injections [Dataset]. https://huggingface.co/datasets/hse-llm/prompt-injections
    Explore at:
    Dataset updated
    4 thg 5, 2025
    Dataset authored and provided by
    HSE LLM @ Saint Petersburg
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    hse-llm/prompt-injections dataset hosted on Hugging Face and contributed by the HF Datasets community

  18. h

    prompt-safety-scores

    • huggingface.co
    Updated 28 thg 8, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alan Tseng (2025). prompt-safety-scores [Dataset]. https://huggingface.co/datasets/agentlans/prompt-safety-scores
    Explore at:
    Dataset updated
    28 thg 8, 2025
    Authors
    Alan Tseng
    License

    https://choosealicense.com/licenses/odc-by/https://choosealicense.com/licenses/odc-by/

    Description

    Composite Safety Scoring for Prompts Using Multiple LLM Annotations

      Introduction
    

    Evaluating the safety of prompts is essential but challenging. Existing approaches often depend on predefined categories, which can be circumvented by new jailbreaks or attacks. Additionally, different tasks may require different safety thresholds. This study explores using large language models (LLMs) themselves to annotate prompt safety. By combining these annotations, a continuous safety… See the full description on the dataset page: https://huggingface.co/datasets/agentlans/prompt-safety-scores.

  19. h

    OpenEndedLLMPrompts

    • huggingface.co
    Updated 6 thg 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shreyan (2024). OpenEndedLLMPrompts [Dataset]. https://huggingface.co/datasets/shreyanmitra/OpenEndedLLMPrompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    6 thg 7, 2024
    Authors
    Shreyan
    Description

    Dataset Card for OpenEndedLLMPrompts

    A cleaned and consolidated set of questions (without context) and answers for LLM hallucination detection. Each question-answer pair is not the work of the author, but was selected from OpenAssistant/oasst2. If you use any of the data provided, please cite this source in addition to the following paper Shreyan Mitra and Leilani Gilpin. Detecting LLM Hallucinations Pre-generation (paper pending) The original dataset was provided in a tree… See the full description on the dataset page: https://huggingface.co/datasets/shreyanmitra/OpenEndedLLMPrompts.

  20. h

    HALoGEN-prompts

    • huggingface.co
    Updated 20 thg 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abhilasha Ravichander (2025). HALoGEN-prompts [Dataset]. https://huggingface.co/datasets/lasha-nlp/HALoGEN-prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    20 thg 1, 2025
    Authors
    Abhilasha Ravichander
    Description

    HALOGEN🔦: Fantastic LLM Hallucinations and Where to Find Them

    This repository contains the prompts of HALOGEN🔦: Fantastic LLM Hallucinations and Where to Find Them by *Abhilasha Ravichander, *Shrusti Ghela, David Wadden, and Yejin Choi Website | Paper | HALoGEN prompts | LLM Hallucinations | Decomposers and Verifiers | Scoring Functions

      Overview
    

    Despite their impressive ability to generate high-quality and fluent text, generative large language models (LLMs) also… See the full description on the dataset page: https://huggingface.co/datasets/lasha-nlp/HALoGEN-prompts.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Naomi Bashkansky (2024). llm-system-prompts-benchmark [Dataset]. https://huggingface.co/datasets/Naomibas/llm-system-prompts-benchmark

llm-system-prompts-benchmark

Naomibas/llm-system-prompts-benchmark

100 system prompts for benchmarking large language models

Explore at:
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
10 thg 1, 2024
Authors
Naomi Bashkansky
License

Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically

Description

Dataset Card for Dataset Name

This datset is a collection of 100 system prompts for large language models.

  Dataset Details





  Dataset Description

These 100 system prompts test a model's ability to follow grammatical patterns; answer basic multiple choice questions; act according to a particular persona; memorize information; and speak in French. Files:

hundred_system_prompts.py: refer to this to see the (prompt, probe, function) triplets, as well as the… See the full description on the dataset page: https://huggingface.co/datasets/Naomibas/llm-system-prompts-benchmark.

Search
Clear search
Close search
Google apps
Main menu