34 datasets found
  1. h

    LLM-Sec-Evaluation

    • huggingface.co
    Updated Jul 17, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Max (2023). LLM-Sec-Evaluation [Dataset]. https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation
    Explore at:
    Dataset updated
    Jul 17, 2023
    Authors
    Max
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    LLM Security Evaluation

    This repo contains scripts for evaluating LLM security abilities. We gathered hundreds of questions cover different ascepts of security, such as vulnerablities, pentest, threat intelligence, etc. All the questions can be viewed at https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation.

      Suppoted LLM
    

    ChatGLM Baichuan Vicuna (GGML format)

      Usage
    

    Because of different LLM requires for different running environment, we highly recommended… See the full description on the dataset page: https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation.

  2. h

    combine-llm-security-benchmark

    • huggingface.co
    Updated Oct 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dũng Võ (2025). combine-llm-security-benchmark [Dataset]. https://huggingface.co/datasets/tuandunghcmut/combine-llm-security-benchmark
    Explore at:
    Dataset updated
    Oct 9, 2025
    Authors
    Dũng Võ
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Combined LLM Security Benchmark 🔐

    A comprehensive, unified benchmark dataset for evaluating Large Language Models (LLMs) on cybersecurity tasks. This dataset combines 10 security benchmarks into a standardized format with 18,059 examples across 5 task types.

      📊 Dataset Summary
    

    This dataset consolidates multiple security-focused benchmarks into a single, easy-to-use format for comprehensive LLM evaluation across various cybersecurity domains:

    Total Examples: 18,059 Total… See the full description on the dataset page: https://huggingface.co/datasets/tuandunghcmut/combine-llm-security-benchmark.

  3. h

    llama-security-llm

    • huggingface.co
    Updated Sep 22, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Maximilian Kenfenheuer (2025). llama-security-llm [Dataset]. https://huggingface.co/datasets/mkenfenheuer/llama-security-llm
    Explore at:
    Dataset updated
    Sep 22, 2025
    Authors
    Maximilian Kenfenheuer
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    mkenfenheuer/llama-security-llm dataset hosted on Hugging Face and contributed by the HF Datasets community

  4. h

    DecodingTrust

    • huggingface.co
    Updated Aug 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Secure Learning Lab (2024). DecodingTrust [Dataset]. https://huggingface.co/datasets/AI-Secure/DecodingTrust
    Explore at:
    Dataset updated
    Aug 11, 2024
    Dataset authored and provided by
    Secure Learning Lab
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models

      Overview
    

    This repo contains the source code of DecodingTrust. This research endeavor is designed to help researchers better understand the capabilities, limitations, and potential risks associated with deploying these state-of-the-art Large Language Models (LLMs). See our paper for details. DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models Boxin Wang, Weixin Chen, Hengzhi… See the full description on the dataset page: https://huggingface.co/datasets/AI-Secure/DecodingTrust.

  5. f

    LLMS_Application

    • figshare.com
    csv
    Updated Jul 22, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dilly Rabbit (2025). LLMS_Application [Dataset]. http://doi.org/10.6084/m9.figshare.29610356.v1
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jul 22, 2025
    Dataset provided by
    figshare
    Authors
    Dilly Rabbit
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This folder contains the meta-information of LLM applications scraped from huggingface, as well as security issue information

  6. L

    Large Language Model(LLM) Cloud Service Report

    • datainsightsmarket.com
    doc, pdf, ppt
    Updated Jun 8, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Data Insights Market (2025). Large Language Model(LLM) Cloud Service Report [Dataset]. https://www.datainsightsmarket.com/reports/large-language-modelllm-cloud-service-1401545
    Explore at:
    doc, pdf, pptAvailable download formats
    Dataset updated
    Jun 8, 2025
    Dataset authored and provided by
    Data Insights Market
    License

    https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The Large Language Model (LLM) cloud service market is experiencing explosive growth, driven by increasing demand for AI-powered applications across diverse sectors. The market's substantial size, estimated at $20 billion in 2025, reflects the significant investment and adoption of LLMs by businesses seeking to leverage their capabilities in natural language processing, machine learning, and other AI-related tasks. A Compound Annual Growth Rate (CAGR) of 35% is projected from 2025 to 2033, indicating a substantial market expansion to an estimated $150 billion by 2033. Key drivers include advancements in LLM technology, decreasing computational costs, and rising demand for personalized user experiences. Trends such as the increasing adoption of hybrid cloud deployments and the integration of LLMs into various software-as-a-service (SaaS) offerings are further fueling market growth. While data security and privacy concerns present some restraints, the overall market outlook remains exceptionally positive. The competitive landscape is dynamic, with major players like Amazon Web Services (AWS), Google Cloud Platform (GCP), and Microsoft Azure vying for market share alongside emerging players like OpenAI and Hugging Face. The market is segmented by deployment model (cloud, on-premise), application (chatbots, machine translation, sentiment analysis), and industry (healthcare, finance, retail). Geographical expansion into emerging markets will further contribute to the overall growth trajectory. The success of LLMs hinges on their ability to handle large datasets and complex computations, requiring robust cloud infrastructure. This necessitates partnerships and collaborations between LLM developers and cloud providers, leading to a synergistic relationship that is accelerating innovation. The market is likely to see further consolidation as smaller players are acquired by larger cloud providers or face challenges in competing on cost and scalability. Ongoing advancements in model architectures, such as improvements in efficiency and reduced latency, will continue to drive down costs and enhance accessibility. Moreover, increasing regulatory scrutiny regarding data privacy and ethical considerations will shape the development and deployment of LLMs, requiring robust security measures and responsible AI practices. This evolution will ultimately refine the LLM landscape, resulting in more sophisticated, reliable, and ethically responsible AI solutions.

  7. Information Security Web-Clips (10y) FAISS-Index

    • kaggle.com
    zip
    Updated Apr 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marius Ciepluch (2024). Information Security Web-Clips (10y) FAISS-Index [Dataset]. https://www.kaggle.com/datasets/mariusciepluch/faiss-text-db-infosec-archive
    Explore at:
    zip(2780475660 bytes)Available download formats
    Dataset updated
    Apr 5, 2024
    Authors
    Marius Ciepluch
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Private RnD project: Bookworm

    Using:

    Couple of tutorials, work in progress, learning by doing.

  8. llm-security-leaderboard-contents

    • huggingface.co
    Updated Mar 4, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Stacklok, Inc (2025). llm-security-leaderboard-contents [Dataset]. https://huggingface.co/datasets/stacklok/llm-security-leaderboard-contents
    Explore at:
    Dataset updated
    Mar 4, 2025
    Dataset provided by
    Stacklok, Inc.
    Authors
    Stacklok, Inc
    Description

    stacklok/llm-security-leaderboard-contents dataset hosted on Hugging Face and contributed by the HF Datasets community

  9. llm-security-leaderboard-data

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Stacklok, Inc, llm-security-leaderboard-data [Dataset]. https://huggingface.co/datasets/stacklok/llm-security-leaderboard-data
    Explore at:
    Dataset provided by
    Stacklok, Inc.
    Authors
    Stacklok, Inc
    Description

    LLM Security Leaderboard Evaluation Data

    This dataset contains the packages, CVEs and code snippets that are used to evaluate models in the LLM Security Leaderboard.

  10. hr-policies-qa-dataset

    • kaggle.com
    zip
    Updated Sep 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Syncora_ai (2025). hr-policies-qa-dataset [Dataset]. https://www.kaggle.com/datasets/syncoraai/hr-policies-qa-dataset
    Explore at:
    zip(54895 bytes)Available download formats
    Dataset updated
    Sep 11, 2025
    Authors
    Syncora_ai
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    🏢 HR Policies Q&A Synthetic Dataset

    This synthetic dataset for LLM training captures realistic employee–assistant interactions about HR and compliance policies.
    Generated using Syncora.ai's synthetic data generation engine, it provides privacy-safe, high-quality conversations for training Large Language Models (LLMs) to handle HR-related queries.

    Perfect for researchers, HR tech startups, and AI developers building chatbots, compliance assistants, or policy QA systems — without exposing sensitive employee data.

    🧠 Context & Applications

    HR departments handle countless queries on policies, compliance, and workplace practices.
    This dataset simulates those Q&A flows, making it a powerful dataset for LLM training and research.

    You can use it for:

    • HR chatbot prototyping
    • Policy compliance assistants
    • Internal knowledge base fine-tuning
    • Generative AI experimentation
    • Synthetic benchmarking in enterprise QA systems

    📊 Dataset Features

    ColumnDescription
    roleRole of the message author (system, user, or assistant)
    contentActual text of the message
    messagesGrouped sequence of role–content exchanges (conversation turns)

    Each entry represents a self-contained dialogue snippet designed to reflect natural HR conversations, ideal for synthetic data generation research.

    📦 This Repo Contains

    • HR Policies QA Dataset – JSON format, ready to use for LLM training or evaluation
    • Jupyter Notebook – Explore the dataset structure and basic preprocessing
    • Synthetic Data Tools – Generate your own datasets using Syncora.ai
    • Generate Synthetic Data
      Need more? Use Syncora.ai’s synthetic data generation tool to create custom HR/compliance datasets. Our process is simple, reliable, and ensures privacy.

    🧪 ML & Research Use Cases

    • Policy Chatbots — Train assistants to answer compliance and HR questions
    • Knowledge Management — Fine-tune models for consistent responses
    • Synthetic Data Research — Explore structured dialogue datasets without legal risks
    • Evaluation Benchmarks — Test enterprise AI assistants on HR-related queries
    • Dataset Expansion — Combine this dataset with your own data using synthetic generation

    🔒 Why Syncora.ai Synthetic Data?

    • Zero real-user data → Zero privacy liability
    • High realism → Actionable insights for LLM training
    • Fully customizable → Generate synthetic data tailored to your domain
    • Ethically aligned → Safe and responsible dataset creation

    Whether you're building an HR assistant, compliance bot, or experimenting with enterprise LLMs, Syncora.ai synthetic datasets give you trustworthy, free datasets to start with — and scalable tools to grow further.

    💬 Questions or Contributions?

    Got feedback, research use cases, or want to collaborate?
    Open an issue or reach out — we’re excited to work with AI researchers, HR tech builders, and compliance innovators.

    BOOK A DEMO

    ⚠️ Disclaimer

    This dataset is 100% synthetic and does not represent real employees or organizations.
    It is intended solely for research, educational, and experimental use in HR analytics, compliance automation, and machine learning.

  11. h

    llm-trustworthy-leaderboard-results

    • huggingface.co
    Updated Jan 19, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Secure Learning Lab (2024). llm-trustworthy-leaderboard-results [Dataset]. https://huggingface.co/datasets/AI-Secure/llm-trustworthy-leaderboard-results
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 19, 2024
    Dataset authored and provided by
    Secure Learning Lab
    Description

    AI-Secure/llm-trustworthy-leaderboard-results dataset hosted on Hugging Face and contributed by the HF Datasets community

  12. h

    security_steerability

    • huggingface.co
    Updated Aug 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Itay H (2025). security_steerability [Dataset]. https://huggingface.co/datasets/itayhf/security_steerability
    Explore at:
    Dataset updated
    Aug 23, 2025
    Authors
    Itay H
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Security Steerability & the VeganRibs Benchmark

    Security steerability is defined as an LLM's ability to stick to the specific rules and boundaries set by a system prompt, particularly for content that isn't typically considered prohibited. To evaluate this, we developed the VeganRibs benchmark. The benchmark tests an LLM's skill at handling conflicts by seeing if it can follow system-level instructions even when a user's input tries to contradict them. VeganRibs works by presenting… See the full description on the dataset page: https://huggingface.co/datasets/itayhf/security_steerability.

  13. h

    S-Eval

    • huggingface.co
    Updated Oct 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Intelligent System Security Lab (2025). S-Eval [Dataset]. https://huggingface.co/datasets/IS2Lab/S-Eval
    Explore at:
    Dataset updated
    Oct 9, 2025
    Dataset authored and provided by
    Intelligent System Security Lab
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    S-Eval: Towards Automated and Comprehensive Safety Evaluation for Large Language Models

    🏆 Leaderboard

    🔔 Updates

    📣 [2025/10/09]: We update the evaluation for the latest LLMs in 🏆 LeaderBoard, and further release Octopus, an automated LLM safety evaluator, to meet the community’s need for accurate and reproducible safety assessment tools. You can download the model from HuggingFace or ModelScope. 📣 [2025/03/30]: 🎉 Our paper has been accepted by ISSTA 2025. To meet… See the full description on the dataset page: https://huggingface.co/datasets/IS2Lab/S-Eval.

  14. Primus-Instruct

    • huggingface.co
    Updated Aug 9, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Trend Cybertron (Trend Micro) (2025). Primus-Instruct [Dataset]. https://huggingface.co/datasets/trend-cybertron/Primus-Instruct
    Explore at:
    Dataset updated
    Aug 9, 2025
    Dataset provided by
    Trend Microhttp://trendmicro.com/
    Authors
    Trend Cybertron (Trend Micro)
    License

    https://choosealicense.com/licenses/odc-by/https://choosealicense.com/licenses/odc-by/

    Description

    ⭐ Please download the dataset from here.

      PRIMUS: A Pioneering Collection of Open-Source Datasets for Cybersecurity LLM Training
    
    
    
    
    
      🤗 Primus-Instruct
    

    The Primus-Instruct dataset contains hundreds of expert-curated cybersecurity business scenario use case instructions, with responses generated by GPT-4o. It includes tasks such as:

    Explaining detected alerts
    Answering questions about retrieved security documents
    Analyzing executed suspicious commands
    Generating query… See the full description on the dataset page: https://huggingface.co/datasets/trend-cybertron/Primus-Instruct.

  15. h

    security-en

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    John1604 -- Applied LLM, AML2 Lab, security-en [Dataset]. https://huggingface.co/datasets/John1604/security-en
    Explore at:
    Authors
    John1604 -- Applied LLM, AML2 Lab
    Description

    aml2lab product under International Inventor's License

      Description
    

    This dataset was created using the Easy Dataset tool.

      Format
    

    This dataset is in alpaca format.

      International Inventor's License
    

    If the use is not commercial, it is free to use without any fees. For commercial use, if the company or individual does not make any profit, no fees are required. For commercial use, if the company or individual has a net profit, they should pay 1% of the net… See the full description on the dataset page: https://huggingface.co/datasets/John1604/security-en.

  16. h

    viettelsecurity-ai_security-llama3.2-3b-details

    • huggingface.co
    Updated Jul 30, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Open LLM Leaderboard (2025). viettelsecurity-ai_security-llama3.2-3b-details [Dataset]. https://huggingface.co/datasets/open-llm-leaderboard/viettelsecurity-ai_security-llama3.2-3b-details
    Explore at:
    Dataset updated
    Jul 30, 2025
    Dataset authored and provided by
    Open LLM Leaderboard
    Description

    Dataset Card for Evaluation run of viettelsecurity-ai/security-llama3.2-3b

    Dataset automatically created during the evaluation run of model viettelsecurity-ai/security-llama3.2-3b The dataset is composed of 38 configuration(s), each one corresponding to one of the evaluated task. The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to… See the full description on the dataset page: https://huggingface.co/datasets/open-llm-leaderboard/viettelsecurity-ai_security-llama3.2-3b-details.

  17. h

    robust-test-llm-response

    • huggingface.co
    Updated Feb 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Raft Security Lab (2025). robust-test-llm-response [Dataset]. https://huggingface.co/datasets/raft-security-lab/robust-test-llm-response
    Explore at:
    Dataset updated
    Feb 7, 2025
    Dataset authored and provided by
    Raft Security Lab
    Description

    Robust test, LLM responses

    Состоит из 2-х сплитов safe и unsafe

    unsafe состоит из одной категории :

    harmful conent

    safe состоит из нескольких категорий:

    Выполнение задания по инструкции, Консультирование в роли эксперта, Поиск информации, Обучение, Рассуждение на тему, Генерация контента, Общение, Ответ на нерелевантный запрос

      Дополнительные файлы
    

    Файл labeled_unsafe_response-00000-of-00001.parquet включает ручную разметку качества ответов в контексте… See the full description on the dataset page: https://huggingface.co/datasets/raft-security-lab/robust-test-llm-response.

  18. h

    llama-security-log-analysis

    • huggingface.co
    Updated Oct 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dũng Võ (2025). llama-security-log-analysis [Dataset]. https://huggingface.co/datasets/tuandunghcmut/llama-security-log-analysis
    Explore at:
    Dataset updated
    Oct 7, 2025
    Authors
    Dũng Võ
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    LLaMA Security Log Analysis (Clean Format)

    A security log analysis dataset converted from mkenfenheuer/llama-security-llm with all LLaMA special tokens removed for clean GPT/ShareGPT format compatibility.

      Dataset Description
    

    This dataset contains 4,189 examples of security log analysis conversations. The original dataset had LLaMA 3 formatting tokens (<|begin_of_text|>, <|start_header_id|>, etc.) which have been cleanly removed to create a universal conversation format.… See the full description on the dataset page: https://huggingface.co/datasets/tuandunghcmut/llama-security-log-analysis.

  19. h

    echr-date

    • huggingface.co
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    LLM Privacy Benchmark, echr-date [Dataset]. https://huggingface.co/datasets/LLM-PBE/echr-date
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset authored and provided by
    LLM Privacy Benchmark
    Description

    LLM-PBE/echr-date dataset hosted on Hugging Face and contributed by the HF Datasets community

  20. h

    DiSCo

    • huggingface.co
    Updated Oct 3, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    LUNR lab at Stony Brook University (2014). DiSCo [Dataset]. https://huggingface.co/datasets/StonyBrookNLP/DiSCo
    Explore at:
    Dataset updated
    Oct 3, 2014
    Dataset authored and provided by
    LUNR lab at Stony Brook University
    Description

    DiSCo: Distilled Secure Code Preference Dataset

    DiSCo (Distilled Secure Code) is a preference dataset of insecure and secure code pairs, along with security reasoning that explains the issues and fixes. It is introduced in the paper Teaching an Old LLM Secure Coding: Localized Preference Optimization on Distilled Preferences. This dataset is designed to address challenges in improving secure code generation by providing high-quality training data covering a broad set of security… See the full description on the dataset page: https://huggingface.co/datasets/StonyBrookNLP/DiSCo.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Max (2023). LLM-Sec-Evaluation [Dataset]. https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation

LLM-Sec-Evaluation

c01dsnap/LLM-Sec-Evaluation

Explore at:
Dataset updated
Jul 17, 2023
Authors
Max
License

Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically

Description

LLM Security Evaluation

This repo contains scripts for evaluating LLM security abilities. We gathered hundreds of questions cover different ascepts of security, such as vulnerablities, pentest, threat intelligence, etc. All the questions can be viewed at https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation.

  Suppoted LLM

ChatGLM Baichuan Vicuna (GGML format)

  Usage

Because of different LLM requires for different running environment, we highly recommended… See the full description on the dataset page: https://huggingface.co/datasets/c01dsnap/LLM-Sec-Evaluation.

Search
Clear search
Close search
Google apps
Main menu