4 datasets found
  1. h

    ChatML-truthy-dpo-v0.1

    • huggingface.co
    Updated Feb 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Victor Nogueira (2024). ChatML-truthy-dpo-v0.1 [Dataset]. https://huggingface.co/datasets/Felladrin/ChatML-truthy-dpo-v0.1
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 16, 2024
    Authors
    Victor Nogueira
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    jondurbin/truthy-dpo-v0.1 in ChatML format, ready to use in HuggingFace TRL's DPO Trainer. Python code used for conversion: from datasets import load_dataset

    dataset = load_dataset("jondurbin/truthy-dpo-v0.1", split="train")

    def format(columns): prompt = f"<|im_start|>user {columns['prompt']}<|im_end|> <|im_start|>assistant "

    if (columns['system']):
      prompt = f"<|im_start|>system
    

    {columns['system']}<|im_end|> {prompt}"

    return {
      "prompt": prompt… See the full description on the dataset page: https://huggingface.co/datasets/Felladrin/ChatML-truthy-dpo-v0.1.
    
  2. h

    GAIL-gsm8k-preference-small

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Genesis AI Labs, GAIL-gsm8k-preference-small [Dataset]. https://huggingface.co/datasets/Genesis-AI-Labs/GAIL-gsm8k-preference-small
    Explore at:
    Dataset authored and provided by
    Genesis AI Labs
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    How to load the dataset (Usage code): from datasets import load_from_disk

    saved_dataset_path = "path/to/dataset/folder" loaded_preference_dataset = load_from_disk(saved_dataset_path) print(f" Successfully loaded dataset from {saved_dataset_path}:") print(loaded_preference_dataset) # Now 'loaded_preference_dataset' can be used directly with DPOTrainer # Example: # dpo_trainer = DPOTrainer(..., train_dataset=loaded_preference_dataset, ...) except Exception as e:… See the full description on the dataset page: https://huggingface.co/datasets/Genesis-AI-Labs/GAIL-gsm8k-preference-small.

  3. h

    ChatML-Neural-DPO

    • huggingface.co
    Updated Feb 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Victor Nogueira (2024). ChatML-Neural-DPO [Dataset]. https://huggingface.co/datasets/Felladrin/ChatML-Neural-DPO
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 25, 2024
    Authors
    Victor Nogueira
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    NeuralNovel/Neural-DPO in ChatML format, ready to use in HuggingFace TRL's DPO Trainer. Python code used for conversion: from datasets import load_dataset

    dataset = load_dataset("NeuralNovel/Neural-DPO", split="train")

    def format(columns): prompt = f"<|im_start|>user {columns['question']}<|im_end|> <|im_start|>assistant "

    if (columns['system']):
      prompt = f"<|im_start|>system
    

    {columns['system']}<|im_end|> {prompt}"

    return {
      "prompt": prompt… See the full description on the dataset page: https://huggingface.co/datasets/Felladrin/ChatML-Neural-DPO.
    
  4. h

    Justin-Wing-Chung-Law_DPO

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Justin Law, Justin-Wing-Chung-Law_DPO [Dataset]. https://huggingface.co/datasets/justinthelaw/Justin-Wing-Chung-Law_DPO
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Authors
    Justin Law
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Justin Law's Q&A DPO Dataset

    This is a very small dataset meant to provide guidance for further DPO data generation or for direct fine-tuning using Hugging Face's DPO trainer. This initial dataset was generated using OpenAI's GPT4o, along with Justin Law's publicly available resume and cover letter.

  5. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Victor Nogueira (2024). ChatML-truthy-dpo-v0.1 [Dataset]. https://huggingface.co/datasets/Felladrin/ChatML-truthy-dpo-v0.1

ChatML-truthy-dpo-v0.1

Felladrin/ChatML-truthy-dpo-v0.1

Explore at:
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Feb 16, 2024
Authors
Victor Nogueira
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

jondurbin/truthy-dpo-v0.1 in ChatML format, ready to use in HuggingFace TRL's DPO Trainer. Python code used for conversion: from datasets import load_dataset

dataset = load_dataset("jondurbin/truthy-dpo-v0.1", split="train")

def format(columns): prompt = f"<|im_start|>user {columns['prompt']}<|im_end|> <|im_start|>assistant "

if (columns['system']):
  prompt = f"<|im_start|>system

{columns['system']}<|im_end|> {prompt}"

return {
  "prompt": prompt… See the full description on the dataset page: https://huggingface.co/datasets/Felladrin/ChatML-truthy-dpo-v0.1.
Search
Clear search
Close search
Google apps
Main menu