13 datasets found
  1. h

    ragbench-sentence-relevance-balanced

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weights and Biases, ragbench-sentence-relevance-balanced [Dataset]. https://huggingface.co/datasets/wandb/ragbench-sentence-relevance-balanced
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset authored and provided by
    Weights and Biases
    Description

    wandb/ragbench-sentence-relevance-balanced dataset hosted on Hugging Face and contributed by the HF Datasets community

  2. h

    RAGTruth-processed

    • huggingface.co
    Updated Sep 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weights and Biases (2025). RAGTruth-processed [Dataset]. https://huggingface.co/datasets/wandb/RAGTruth-processed
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 28, 2025
    Dataset authored and provided by
    Weights and Biases
    Description

    RAGTruth Dataset

      Dataset Description
    
    
    
    
    
      Dataset Summary
    

    The RAGTruth dataset is designed for evaluating hallucinations in text generation models, particularly in retrieval-augmented generation (RAG) contexts. It contains examples of model outputs along with expert annotations indicating whether the outputs contain hallucinations.

      Dataset Structure
    

    Each example contains:

    A query/question Context passages Model output Hallucination labels (evident… See the full description on the dataset page: https://huggingface.co/datasets/wandb/RAGTruth-processed.

  3. h

    ADARFT-wandb

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Liu, ADARFT-wandb [Dataset]. https://huggingface.co/datasets/happynew111/ADARFT-wandb
    Explore at:
    Authors
    Liu
    Description

    happynew111/ADARFT-wandb dataset hosted on Hugging Face and contributed by the HF Datasets community

  4. h

    finqa-data-processed

    • huggingface.co
    Updated Dec 31, 2004
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weights and Biases (2004). finqa-data-processed [Dataset]. https://huggingface.co/datasets/wandb/finqa-data-processed
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 31, 2004
    Dataset authored and provided by
    Weights and Biases
    Description

    FinQA Dataset (Processed)

      Dataset Description
    
    
    
    
    
      Dataset Summary
    

    The FinQA dataset is designed for numerical reasoning over financial data, containing questions that require complex reasoning over tables and text from financial reports.

      Dataset Statistics
    

    Total examples: 8281 Training set size: 6624 examples Test set size: 1657 examples

      Dataset Structure
    

    Each example contains:

    Required columns: query: The question to be answered (derived… See the full description on the dataset page: https://huggingface.co/datasets/wandb/finqa-data-processed.

  5. rl-algo-impls v0.0.12

    • kaggle.com
    zip
    Updated Apr 25, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Scott Goodfriend (2023). rl-algo-impls v0.0.12 [Dataset]. https://www.kaggle.com/sgoodfriend/rl-algo-impls-v0012
    Explore at:
    zip(168254 bytes)Available download formats
    Dataset updated
    Apr 25, 2023
    Authors
    Scott Goodfriend
    Description

    rl-algo-impls

    Implementations of reinforcement learning algorithms.

    Prerequisites: Weights & Biases (WandB)

    Training and benchmarking assumes you have a Weights & Biases project to upload runs to. By default training goes to a rl-algo-impls project while benchmarks go to rl-algo-impls-benchmarks. During training and benchmarking runs, videos of the best models and the model weights are uploaded to WandB.

    Before doing anything below, you'll need to create a wandb account and run wandb login.

    Setup and Usage

    Lambda Labs instance for benchmarking

    Benchmark runs are uploaded to WandB, which can be made into reports ("https://api.wandb.ai/links/sgoodfriend/6p2sjqtn">for example). So far I've found Lambda Labs A10 instances to be a good balance of performance (14 hours to train PPO in 14 environments [5 basic gym, 4 PyBullet, CarRacing-v0, and 4 Atari] across 3 seeds) vs cost ($0.60/hr).

    git clone https://github.com/sgoodfriend/rl-algo-impls.git
    cd rl-algo-impls
    # git checkout BRANCH_NAME if running on non-main branch
    bash ./scripts/setup.sh
    wandb login
    bash ./scripts/benchmark.sh [-a {"ppo"}] [-e ENVS] [-j {6}] [-p {rl-algo-impls-benchmarks}] [-s {"1 2 3"}]
    

    Benchmarking runs are by default upload to a rl-algo-impls-benchmarks project. Runs upload videos of the running best model and the weights of the best and last model. Benchmarking runs are tagged with a shorted commit hash (i.e., benchmark_5598ebc) and hostname (i.e., host_192-9-145-26)

    Publishing models to Huggingface

    Publishing benchmarks to Huggingface requires logging into Huggingface with a write-capable API token:

    git config --global credential.helper store
    huggingface-cli login
    # For example: python benchmark_publish.py --wandb-tags host_192-9-147-166 benchmark_1d4094f --wandb-report-url https://api.wandb.ai/links/sgoodfriend/099h4lvj
    # --virtual-display likely must be specified if running on a remote machine.
    python benchmark_publish.py --wandb-tags HOST_TAG COMMIT_TAG --wandb-report-url WANDB_REPORT_URL [--virtual-display]
    

    Hyperparameter tuning with Optuna

    Hyperparameter tuning can be done with the tuning/tuning.sh script, which runs multiple processes of optimize.py. Start by doing all the setup meant for training before running tuning/tuning.sh:

    # Setup similar to training above
    wandb login
    bash scripts/tuning.sh -a ALGO -e ENV -j N_JOBS -s NUM_SEEDS
    

    Google Colab Pro+

    3 notebooks in the colab directory are setup to be used with Google Colab:

    • colab_benchmark.ipynb: Even with a Google Colab Pro+ subscription you'd need to only run parts of the benchmark. The file recommends 4 splits (basic+pybullet, carcarcing, atari1, atari2) because it would otherwise exceed the 24-hour session limit. This mostly comes from being unable to get pool_size above 1 because of WandB errors.
    • colab_train.ipynb: Train models while being able to specify the env, seeds, and algo. By default training runs are uploaded to the rl-algo-impls project.
    • colab_enjoy.ipynb: Download models from WandB and evaluate them. Training is likely to be more interesting given videos are uploaded.

    macOS

    Installation

    My local development has been on an M1 Mac. These instructions might not be complete, but these are the approximate setup and usage I've been using:

    1. Install libraries with homebrew
    brew install swig
    brew install --cask xquartz
    
    1. Download and install Miniconda for arm64
    curl -O https://repo.anaconda.com/miniconda/Miniconda3-latest-MacOSX-arm64.sh
    sh Miniconda3-latest-MacOSX-arm64.sh
    
    1. Create a conda environment from this repo's environment.yml
    conda env create -f environment.yml -n rl_algo_impls
    conda activate rl_algo_impls
    
    1. Install other dependencies with poetry
    poetry install
    

    Usage

    Training, benchmarking,...

  6. h

    deita-10k-v0-sft-latin

    • huggingface.co
    Updated Mar 20, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weights and Biases (2024). deita-10k-v0-sft-latin [Dataset]. https://huggingface.co/datasets/wandb/deita-10k-v0-sft-latin
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 20, 2024
    Dataset authored and provided by
    Weights and Biases
    Description

    Same as HuggingFaceH4/deita-10k-v0-sft but without non-latin text.

  7. chaii-finetuned-baseline-models

    • kaggle.com
    zip
    Updated Aug 30, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Saurav Maheshkar ☕️ (2021). chaii-finetuned-baseline-models [Dataset]. https://www.kaggle.com/sauravmaheshkar/chaiifinetunedbaselinemodels
    Explore at:
    zip(13240431286 bytes)Available download formats
    Dataset updated
    Aug 30, 2021
    Authors
    Saurav Maheshkar ☕️
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Link to Weights and Biases 🔥 Interactive Dashboard

    All models were taken from Huggingface Question Answering Models trained using @rhtsingh's processed dataset [ External Data - MLQA, XQUAD Preprocessing ] using huggingface/transformers inbuilt weights and biases logger.

    https://raw.githubusercontent.com/SauravMaheshkar/chaii-Hindi-Tamil-QA/500ff923d44525d25d28a7b299995200b36c76cd/assets/Evaluation%20Loss.svg" alt="">

    The Model Weights can be found here

    NameTraining LossEvaluation Loss
    electra-base-squad21.98232.27
    distilbert-base-cased-distilled-squad1.16941.31
    bert-base-cased-squad21.09921.26
    distilbert-base-uncased-distilled-squad1.06421.19
    bert-large-uncased-whole-word-masking-squad20.92061.02
    bert-large-uncased-whole-word-masking-finetuned-squad0.90681.01
    xlm-roberta-base-squad20.79080.90
    distilbert-**multi**-finetuned-for-xqua-on-tydiqa0.78270.89
    bert-**multi**-uncased-finetuned-xquadv10.70720.93
    bert-**multi**-cased-finetuned-xquadv10.65170.74
    bert-base-**multilingual**-cased-finetuned-squad0.62570.73
    xlm-**multi**-roberta-large-squad20.62090.74
    bert-**multi**-cased-finedtuned-xquad-tydiqa-goldp0.61560.70
    roberta-large-squad20.24880.36
    roberta-base-squad20.2360.35
  8. R

    Custom Yolov7 On Kaggle On Custom Dataset

    • universe.roboflow.com
    zip
    Updated Jan 29, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Owais Ahmad (2023). Custom Yolov7 On Kaggle On Custom Dataset [Dataset]. https://universe.roboflow.com/owais-ahmad/custom-yolov7-on-kaggle-on-custom-dataset-rakiq/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 29, 2023
    Dataset authored and provided by
    Owais Ahmad
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Person Car Bounding Boxes
    Description

    Custom Training with YOLOv7 🔥

    Some Important links

    Contact Information

    Objective

    To Showcase custom Object Detection on the Given Dataset to train and Infer the Model using newly launched YoloV7.

    Data Acquisition

    The goal of this task is to train a model that can localize and classify each instance of Person and Car as accurately as possible.

    from IPython.display import Markdown, display
    
    display(Markdown("../input/Car-Person-v2-Roboflow/README.roboflow.txt"))
    

    Custom Training with YOLOv7 🔥

    In this Notebook, I have processed the images with RoboFlow because in COCO formatted dataset was having different dimensions of image and Also data set was not splitted into different Format. To train a custom YOLOv7 model we need to recognize the objects in the dataset. To do so I have taken the following steps:

    • Export the dataset to YOLOv7
    • Train YOLOv7 to recognize the objects in our dataset
    • Evaluate our YOLOv7 model's performance
    • Run test inference to view performance of YOLOv7 model at work

    📦 YOLOv7

    https://raw.githubusercontent.com/Owaiskhan9654/Yolo-V7-Custom-Dataset-Train-on-Kaggle/main/car-person-2.PNG" width=800>

    Image Credit - jinfagang

    Step 1: Install Requirements

    !git clone https://github.com/WongKinYiu/yolov7 # Downloading YOLOv7 repository and installing requirements
    %cd yolov7
    !pip install -qr requirements.txt
    !pip install -q roboflow
    

    Downloading YOLOV7 starting checkpoint

    !wget "https://github.com/WongKinYiu/yolov7/releases/download/v0.1/yolov7.pt"
    
    import os
    import glob
    import wandb
    import torch
    from roboflow import Roboflow
    from kaggle_secrets import UserSecretsClient
    from IPython.display import Image, clear_output, display # to display images
    
    
    
    print(f"Setup complete. Using torch {torch._version_} ({torch.cuda.get_device_properties(0).name if torch.cuda.is_available() else 'CPU'})")
    

    https://camo.githubusercontent.com/dd842f7b0be57140e68b2ab9cb007992acd131c48284eaf6b1aca758bfea358b/68747470733a2f2f692e696d6775722e636f6d2f52557469567a482e706e67">

    I will be integrating W&B for visualizations and logging artifacts and comparisons of different models!

    YOLOv7-Car-Person-Custom

    try:
      user_secrets = UserSecretsClient()
      wandb_api_key = user_secrets.get_secret("wandb_api")
      wandb.login(key=wandb_api_key)
      anonymous = None
    except:
      wandb.login(anonymous='must')
      print('To use your W&B account,
    Go to Add-ons -> Secrets and provide your W&B access token. Use the Label name as WANDB. 
    Get your W&B access token from here: https://wandb.ai/authorize')
      
      
      
    wandb.init(project="YOLOvR",name=f"7. YOLOv7-Car-Person-Custom-Run-7")
    

    Step 2: Assemble Our Dataset

    https://uploads-ssl.webflow.com/5f6bc60e665f54545a1e52a5/615627e5824c9c6195abfda9_computer-vision-cycle.png" alt="">

    In order to train our custom model, we need to assemble a dataset of representative images with bounding box annotations around the objects that we want to detect. And we need our dataset to be in YOLOv7 format.

    In Roboflow, We can choose between two paths:

    Version v2 Aug 12, 2022 Looks like this.

    https://raw.githubusercontent.com/Owaiskhan9654/Yolo-V7-Custom-Dataset-Train-on-Kaggle/main/Roboflow.PNG" alt="">

    user_secrets = UserSecretsClient()
    roboflow_api_key = user_secrets.get_secret("roboflow_api")
    
    rf = Roboflow(api_key=roboflow_api_key)
    project = rf.workspace("owais-ahmad").project("custom-yolov7-on-kaggle-on-custom-dataset-rakiq")
    dataset = project.version(2).download("yolov7")
    

    Step 3: Training Custom pretrained YOLOv7 model

    Here, I am able to pass a number of arguments: - img: define input image size - batch: determine

  9. h

    fava-data-processed

    • huggingface.co
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weights and Biases (2024). fava-data-processed [Dataset]. https://huggingface.co/datasets/wandb/fava-data-processed
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 2, 2024
    Dataset authored and provided by
    Weights and Biases
    Description

    FAVA Dataset (Processed)

      Dataset Description
    
    
    
    
    
      Dataset Summary
    

    The FAVA (Factual Association and Verification Annotations) dataset is designed for evaluating hallucinations in language model outputs. This processed version contains binary hallucination labels derived from detailed span-level annotations in the original dataset.

      Dataset Structure
    

    Each example contains:

    Required columns: query: The prompt given to the model context: Empty field (for… See the full description on the dataset page: https://huggingface.co/datasets/wandb/fava-data-processed.

  10. h

    openvalidators-test

    • huggingface.co
    Updated Mar 23, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pedro Ferreira (2024). openvalidators-test [Dataset]. https://huggingface.co/datasets/pedroferreira/openvalidators-test
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 23, 2024
    Authors
    Pedro Ferreira
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset Card for Openvalidators dataset

      Dataset Summary
    

    The OpenValidators dataset, created by the OpenTensor Foundation, is a continuously growing collection of data generated by the OpenValidators project in W&B. It contains hundreds of thousands of records and serves researchers, data scientists, and miners in the Bittensor network. The dataset provides information on network performance, node behaviors, and wandb run details. Researchers can gain insights and detect… See the full description on the dataset page: https://huggingface.co/datasets/pedroferreira/openvalidators-test.

  11. h

    damaged-media

    • huggingface.co
    Updated May 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daniela Ivanova (2024). damaged-media [Dataset]. https://huggingface.co/datasets/danielaivanova/damaged-media
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 6, 2024
    Authors
    Daniela Ivanova
    License

    https://choosealicense.com/licenses/afl-3.0/https://choosealicense.com/licenses/afl-3.0/

    Description

    Dataset Card for "ARTeFACT"

    ARTeFACT: Benchmarking Segmentation Models on Diverse Analogue Media Damage

    Here we provide example code for downloading the data, loading it as a PyTorch dataset, splitting by material and/or content, and visualising examples.

      Housekeeping
    

    !pip install datasets !pip install -qqqU wandb transformers pytorch-lightning==1.9.2 albumentations torchmetrics torchinfo !pip install -qqq requests gradio

    import os from glob import glob

    import cv2… See the full description on the dataset page: https://huggingface.co/datasets/danielaivanova/damaged-media.

  12. h

    real_cartpole_200k

    • huggingface.co
    Updated Aug 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Armand (2024). real_cartpole_200k [Dataset]. https://huggingface.co/datasets/armandpl/real_cartpole_200k
    Explore at:
    Dataset updated
    Aug 5, 2024
    Authors
    Armand
    Description

    This dataset contains sequences of actions, motor angles and pendulum angles as well as velocities for a rotary inverted pendulum robot. The dataset was collected while training the robot to swing up and balance (wandb run).Angles are in radian. Velocities were computed from the angles and fed to the policy. Control frequency is 75Hz.The action maps to the motor voltage with: deadzone = 0.1 center = 0.05 max_act = 0.9 if abs(action) > center: V = np.sign(action) * (… See the full description on the dataset page: https://huggingface.co/datasets/armandpl/real_cartpole_200k.

  13. h

    autoeval-eval-acronym_identification-default-641c5d-40516105295

    • huggingface.co
    Updated Feb 24, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Evaluation on the Hub (2023). autoeval-eval-acronym_identification-default-641c5d-40516105295 [Dataset]. https://huggingface.co/datasets/autoevaluate/autoeval-eval-acronym_identification-default-641c5d-40516105295
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 24, 2023
    Dataset authored and provided by
    Evaluation on the Hub
    Description

    Dataset Card for AutoTrain Evaluator

    This repository contains model predictions generated by AutoTrain for the following task and dataset:

    Task: Token Classification Model: lewtun/autotrain-acronym-identification-7324788 Dataset: acronym_identification Config: default Split: validation

    To run new evaluation jobs, visit Hugging Face's automatic model evaluator.

      Contributions
    

    Thanks to @wandb.init(project=PROJECT for evaluating this model.

  14. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Weights and Biases, ragbench-sentence-relevance-balanced [Dataset]. https://huggingface.co/datasets/wandb/ragbench-sentence-relevance-balanced

ragbench-sentence-relevance-balanced

wandb/ragbench-sentence-relevance-balanced

Explore at:
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset authored and provided by
Weights and Biases
Description

wandb/ragbench-sentence-relevance-balanced dataset hosted on Hugging Face and contributed by the HF Datasets community

Search
Clear search
Close search
Google apps
Main menu