3 datasets found
  1. h

    AlphaNLI

    • huggingface.co
    Updated Jun 21, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Massive Text Embedding Benchmark (2025). AlphaNLI [Dataset]. https://huggingface.co/datasets/mteb/AlphaNLI
    Explore at:
    Dataset updated
    Jun 21, 2025
    Dataset authored and provided by
    Massive Text Embedding Benchmark
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    AlphaNLI An MTEB dataset Massive Text Embedding Benchmark

    Measuring the ability to retrieve the groundtruth answers to reasoning task queries on AlphaNLI.

    Task category t2t

    Domains Encyclopaedic, Written

    Reference https://leaderboard.allenai.org/anli/submissions/get-started

      How to evaluate on this task
    

    You can evaluate an embedding model on this dataset using the following code: import mteb

    task = mteb.get_task("AlphaNLI") evaluator =… See the full description on the dataset page: https://huggingface.co/datasets/mteb/AlphaNLI.

  2. h

    SIQA

    • huggingface.co
    Updated Jun 21, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Massive Text Embedding Benchmark (2025). SIQA [Dataset]. https://huggingface.co/datasets/mteb/SIQA
    Explore at:
    Dataset updated
    Jun 21, 2025
    Dataset authored and provided by
    Massive Text Embedding Benchmark
    License

    https://choosealicense.com/licenses/unknown/https://choosealicense.com/licenses/unknown/

    Description

    SIQA An MTEB dataset Massive Text Embedding Benchmark

    Measuring the ability to retrieve the groundtruth answers to reasoning task queries on SIQA.

    Task category t2t

    Domains Encyclopaedic, Written

    Reference https://leaderboard.allenai.org/socialiqa/submissions/get-started

      How to evaluate on this task
    

    You can evaluate an embedding model on this dataset using the following code: import mteb

    task = mteb.get_task("SIQA") evaluator = mteb.MTEB([task])… See the full description on the dataset page: https://huggingface.co/datasets/mteb/SIQA.

  3. h

    CrosslingualMultiDomainsDataset

    • huggingface.co
    Updated Jan 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lei Shen (2024). CrosslingualMultiDomainsDataset [Dataset]. https://huggingface.co/datasets/maidalun1020/CrosslingualMultiDomainsDataset
    Explore at:
    Dataset updated
    Jan 3, 2024
    Authors
    Lei Shen
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Leadboard

    BCEmbedding: Bilingual and Crosslingual Embedding for RAG

    GitHub

    Click to Open Contents

    🌐 Bilingual and Crosslingual Superiority πŸ’‘ Key Features πŸš€ Latest Updates 🍎 Model List πŸ“– Manual Installation Quick Start

    βš™οΈ Evaluation Evaluate Semantic Representation by MTEB Evaluate RAG by LlamaIndex

    πŸ“ˆ Leaderboard Semantic Representation Evaluations in MTEB RAG Evaluations in LlamaIndex

    πŸ›  Youdao's BCEmbedding API 🧲 WeChat Group ✏️ Citation πŸ”β€¦ See the full description on the dataset page: https://huggingface.co/datasets/maidalun1020/CrosslingualMultiDomainsDataset.

  4. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Massive Text Embedding Benchmark (2025). AlphaNLI [Dataset]. https://huggingface.co/datasets/mteb/AlphaNLI

AlphaNLI

mteb/AlphaNLI

Explore at:
27 scholarly articles cite this dataset (View in Google Scholar)
Dataset updated
Jun 21, 2025
Dataset authored and provided by
Massive Text Embedding Benchmark
License

Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically

Description

AlphaNLI An MTEB dataset Massive Text Embedding Benchmark

Measuring the ability to retrieve the groundtruth answers to reasoning task queries on AlphaNLI.

Task category t2t

Domains Encyclopaedic, Written

Reference https://leaderboard.allenai.org/anli/submissions/get-started

  How to evaluate on this task

You can evaluate an embedding model on this dataset using the following code: import mteb

task = mteb.get_task("AlphaNLI") evaluator =… See the full description on the dataset page: https://huggingface.co/datasets/mteb/AlphaNLI.

Search
Clear search
Close search
Google apps
Main menu