Dataset Card for Evaluation run of Nekochu/Llama-3.1-8B-German-ORPO
Dataset automatically created during the evaluation run of model Nekochu/Llama-3.1-8B-German-ORPO The dataset is composed of 38 configuration(s), each one corresponding to one of the evaluated task. The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest… See the full description on the dataset page: https://huggingface.co/datasets/open-llm-leaderboard/Nekochu_Llama-3.1-8B-German-ORPO-details.
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
8.5k+13k+5k rows (v1, v2, v3) of multilingual prompts and responses. Prompts taken from lmsys 1m dataset. Same format as host-provided dataset.
Model | Count |
---|---|
mistralai/Mistral-Nemo-Instruct-2407 | 1867 |
meta-llama/Meta-Llama-3-8B-Instruct | 1702 |
mistralai/Mixtral-8x7B-Instruct-v0.1 | 1506 |
mistralai/Mistral-7B-Instruct-v0.3 | 1424 |
NousResearch/Hermes-3-Llama-3.1-8B | 1408 |
meta-llama/Llama-3.3-70B-Instruct | 1344 |
Qwen/Qwen2.5-72B-Instruct | 1322 |
01-ai/Yi-1.5-34B-Chat | 1322 |
HuggingFaceH4/starchat2-15b-v0.1 | 1302 |
microsoft/Phi-3.5-mini-instruct | 1294 |
google/gemma-2-27b-it | 1230 |
Qwen/QwQ-32B-Preview | 1117 |
Language | Count |
---|---|
Portuguese | 1079 |
Russian | 966 |
Chinese | 909 |
English | 883 |
Spanish | 779 |
German | 615 |
French | 585 |
Italian | 493 |
unknown | 383 |
Japanese | 319 |
Korean | 201 |
Polish | 132 |
Indonesian | 104 |
Arabic | 75 |
Vietnamese | 57 |
Turkish | 57 |
Dutch | 50 |
Latin | 40 |
Hungarian | 37 |
Ukrainian | 36 |
Persian | 34 |
Danish | 33 |
Greek | 33 |
Czech | 29 |
Swedish | 25 |
Romanian | 24 |
Galician | 22 |
Hebrew | 19 |
Serbian | 18 |
Scots | 17 |
Norwegian | 17 |
Bulgarian | 15 |
Finnish | 14 |
Catalan | 14 |
Hawaiian | 13 |
Corsican | 13 |
Malay | 12 |
Slovak | 11 |
Thai | 10 |
Occitan | 9 |
Norwegian Nynorsk | 8 |
Afrikaans | 8 |
Haitian Creole | 8 |
Quechua | 8 |
Samoan | 7 |
Breton | 7 |
Uzbek | 7 |
Bangla | 7 |
Hausa | 6 |
Luxembourgish | 6 |
Tsonga | 6 |
Esperanto | 6 |
Interlingua | 5 |
Somali | 5 |
Basque | 5 |
Aymara | 5 |
Tatar | 5 |
Nauru | 4 |
Tagalog | 4 |
Tswana | 4 |
Wolof | 4 |
Guarani | 4 |
Faroese | 4 |
Croatian | 4 |
Malagasy | 4 |
Estonian | 4 |
Lithuanian | 3 |
Khasi | 3 |
Tongan | 3 |
Akan | 3 |
Manx | 3 |
Javanese | 3 |
Swahili | 3 |
Seselwa Creole French | 3 |
Oromo | 3 |
Latvian | 3 |
Lingala | 2 |
Interlingue | 2 |
Bosnian | 2 |
Yoruba | 2 |
Kazakh | 2 |
zzp | 2 |
Macedonian | 2 |
Tajik | 2 |
Southern Sotho | 2 |
Welsh | 2 |
Scottish Gaelic | 2 |
Northern Sotho | 2 |
Kinyarwanda | 2 |
Irish | 2 |
Fijian | 2 |
Amharic | 2 |
Bislama | 2 |
Hmong | 2 |
Hindi | 2 |
Waray | 2 |
Volapük | 2 |
Marathi | 1 |
Sundanese | 1 |
Kalaallisut | 1 |
Ganda | 1 |
Afar | 1 |
Rundi | 1 |
Sanskrit | 1 |
Bashkir | 1 |
Cebuano | 1 |
Zulu | 1 |
Sinhala | 1 |
Romansh | 1 |
Nepali | 1 |
Xhosa | 1 |
Tamil | 1 |
Māori | 1 |
Albanian | 1 |
Icelandic | 1 |
Slovenian | 1 |
xx | 1 |
Model Name | Count |
---|---|
google/gemma-2-9b-it | 1242 |
01-ai/Yi-1.5-34B-Chat | 1229 |
microsoft/phi-4 | 1195 |
microsoft/Phi-3.5-mini-instruct | 1187 |
NousResearch/Hermes-3-Llama-3.1-8B | 1179 |
meta-llama/Llama-2-7b-chat-hf | 1179 |
mistralai/Mixtral-8x7B-Instruct-v0.1 | 1177 |
mistralai/Mistral-Nemo-Instruct-2407 | 1163 |
meta-llama/Meta-Llama-3-8B-Instruct | 1158 |
meta-llama/Llama-3.1-70B-Instruct | 1146 |
meta-llama/Llama-3.3-70B-Instruct | 1142 |
microsoft/Phi-3-mini-4k-instruct | 1141 |
Qwen/Qwen2.5-0.5B-Instruct | 1138 |
google/gemma-2-2b-it | 1133 |
google/gemma-1.1-7b-it | 1130 |
meta-llama/Llama-3.2-1B-Instruct | 1115 |
mistralai/Mistral-7B-Instruct-v0.3 | 1115 |
HuggingFaceH4/starchat2-15b-v0.1 | 1112 |
meta-llama/Llama-3.2-3B-Instruct | 1097 |
HuggingFaceTB/SmolLM2-1.7B-Instruct | 1092 |
Qwen/Qwen2.5-72B-Instruct | 1088 |
tiiuae/falcon-7b-instruct | 1064 |
Qwen/QwQ-32B-Preview | 964 |
Language | Count |
---|---|
English | 2724 |
Portuguese | 1482 |
Russian | 1410 |
Chinese | 1121 |
Spanish | 1088 |
French | 859 |
German | 814 |
Italian | 725 |
unknown | 502 |
Japanese | 378 |
Korean | 270 |
Polish | 151 |
Indonesian | 132 |
Arabic | 114 |
Vietnamese | 98 |
Latin ... |
A reformatted version of the DRXD1000/Dolly-15k-German dataset. Available for finetuning in hiyouga/LLaMA-Factory.
https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/
AlephAlphaGermanWeb
Aleph-Alpha-GermanWeb is a new German-language dataset that combines heuristic and model-based filtering techniques with synthetic data generation to achieve SOTA performance in German-language benchmarks. The dataset draws from three sources: (1) Common Crawl web data, (2) FineWeb2, and (3) synthetically-generated data conditioned on actual, organic web data. In our accompanying paper, we evaluated our dataset by training both a 1B Llama-style model and an 8B… See the full description on the dataset page: https://huggingface.co/datasets/Aleph-Alpha/Aleph-Alpha-GermanWeb.
A german translation for the wiki_qa dataset. Extracted from seedboxventures/multitask_german_examples_32k. Translation created by seedbox ai for KafkaLM ❤️. Available for finetuning in hiyouga/LLaMA-Factory.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Dataset Name: Deutsche Bahn FAQ in Llama 3 Format Dataset Description: This dataset contains 1000 question-answer pairs extracted from the official Deutsche Bahn (German Railways) FAQ section. The data has been specifically formatted to be compatible with the Llama 3 instruct models for supervised fine-tuning (SFT). Dataset Purpose: The primary purpose of this dataset is to facilitate the fine-tuning of Llama 3 instruct models for tasks related to customer service and information retrieval in… See the full description on the dataset page: https://huggingface.co/datasets/islam-hajosman/deutsche_bahn_faq_128.
Notebook based on Sarah Oberbichler's (oberbichler@ieg-mainz.de) Notebook 'Researching German Historical Newspapers with Llama AI Model' (https://github.com/soberbichler/Notebooks4Historical_Newspapers/blob/main/Llama3_OCR.ipynb) Edited by Christian Wachter (christian.wachter@uni-bielefeld.de) and Patrick Jentsch (p.jentsch@uni-bielefeld.de) This notebook shows how LLMs can be used to support research with historical newspapers. In this example, the Llama 3.1 model is used to correct OCR of previously OCR'd historical newspaper pages. OCR quality has been a long-standing issue in digitization efforts. Historical newspapers are particularly affected due their complexity, historical fonts, or degradation. Additionally, OCR technology faced limitations when dealing with historical scripts.
License: GNU GPLv3
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
HARD-REASONING-DE
The original dataset was obtained from German-RAG LLM-HARD BENCHMARK and was further cleaned, filtered and re-evaluated.
Methodology: Reasoning-DE
Providing Persona Descriptions and rewriting in a similar style with a different focus area and name in german/english language Generating Simple Logical Problems out of Persona-specific Views & Language. Generating Approaches, Thinking-Steps & Solutions separately verified by Llama-3.1-405B-Instruct Quality… See the full description on the dataset page: https://huggingface.co/datasets/embraceableAI/HARD-REASONING-DE.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Dataset Card for Evaluation run of Nekochu/Llama-3.1-8B-German-ORPO
Dataset automatically created during the evaluation run of model Nekochu/Llama-3.1-8B-German-ORPO The dataset is composed of 38 configuration(s), each one corresponding to one of the evaluated task. The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest… See the full description on the dataset page: https://huggingface.co/datasets/open-llm-leaderboard/Nekochu_Llama-3.1-8B-German-ORPO-details.