MATH-500 test set with the remaining 12000 examples in train. import datasets
from math_utils import last_boxed_only_string, remove_boxed
math = datasets.load_dataset('DigitalLearningGmbH/MATH-lighteval', 'default') math500 = datasets.load_dataset('HuggingFaceH4/MATH-500')
def map_to_500(example): return { 'problem':โฆ See the full description on the dataset page: https://huggingface.co/datasets/ricdomolm/MATH-500.
Source: HuggingFaceH4/MATH-500 Modification:
Remove redundent columns
Keys:
problem answer
Size
test: 500
Comparison of Independently conducted by Artificial Analysis by Model
VLM-Reasoner/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
yinjiewang/MATH500 dataset hosted on Hugging Face and contributed by the HF Datasets community
FlippyDora/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
MATH-500 Chain of Thought Experiment Dataset
This dataset contains the results of an experiment testing different prompting strategies (standard, chain of thought, and gibberish chain of thought) on the MATH-500 benchmark using the Llama-3.2-1B-Instruct model.
Dataset Structure
The dataset is split into three parts:
standard: Direct prompting with no reasoning steps (500 examples) cot: Chain of thought prompting with structured reasoning (500 examples) gibberish:โฆ See the full description on the dataset page: https://huggingface.co/datasets/codelion/math500-cot-experiment.
Dataset Summary
This dataset contains mathematical problem-solving responses generated using two decoding methods: Greedy and Best-of-N for 20 problems spanning three difficulty levels (1, 2, 3) from the [MATH500 dataset](https://huggingface.co/datasets/HuggingFaceH4/MATH-500.
Languages
The dataset content is entirely in English.
Dataset Structure
Each instance contains the following fields:
problem: math problem given as an input answer: a ground truth answerโฆ See the full description on the dataset page: https://huggingface.co/datasets/Jakh0103/math500_best_of_n.
Wloner0809/MATH500 dataset hosted on Hugging Face and contributed by the HF Datasets community
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Gen-Verse/MATH500 dataset hosted on Hugging Face and contributed by the HF Datasets community
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Math500 Enhanced Dataset
This dataset contains LLM-enhanced versions of mathematical problems with step-by-step reasoning solutions.
Dataset Statistics
Examples: 500 (500 enhanced with LLM) Enhancement Rate: 100.0%
Data Fields
question: The mathematical problem statement solution: LLM-enhanced step-by-step solution original_solution: Original solution text (for reference) answer: Final numerical answer level: Problem difficulty level type: Problemโฆ See the full description on the dataset page: https://huggingface.co/datasets/rachitbansal-harvard/math500-enhanced.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
DongfuJiang/MATH-500 dataset hosted on Hugging Face and contributed by the HF Datasets community
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Dataset Card for "math500-deepseek-r1-distill-qwen-1.5b"
Dataset Summary
This dataset is a distilled version of the MATH500 dataset, augmented with reasoning-based responses generated by the deepseek-r1-distill-qwen-1.5b language model. The dataset is designed to evaluate and improve the mathematical reasoning capabilities of LLMs through step-by-step solutions and final answers. Each example consists of:
The original problem statement from MATH500 The reference solutionโฆ See the full description on the dataset page: https://huggingface.co/datasets/jsm0424/math500-deepseek-r1-distill-qwen-1.5b.
Dataset Card for Dataset Name
Forked from HuggingFaceH4/MATH-500
Dataset Details
Dataset Description
Curated by: [More Information Needed] Funded by [optional]: [More Information Needed] Shared by [optional]: [More Information Needed] Language(s) (NLP): [More Information Needed] License: [More Information Needed]
Dataset Sources [optional]
Repository: [More Information Needed] Paper [optional]: [More Information Needed] Demo [optional]:โฆ See the full description on the dataset page: https://huggingface.co/datasets/smoorsmith/math500.
ShuoZheLi/MATH-train-MATH500-test dataset hosted on Hugging Face and contributed by the HF Datasets community
sparkle-reasoning/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
mikasenghaas/math500-math-rubric dataset hosted on Hugging Face and contributed by the HF Datasets community
weqweasdas/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
dsrtrain/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
ScaleFrontierData/math500 dataset hosted on Hugging Face and contributed by the HF Datasets community
MATH-500 test set with the remaining 12000 examples in train. import datasets
from math_utils import last_boxed_only_string, remove_boxed
math = datasets.load_dataset('DigitalLearningGmbH/MATH-lighteval', 'default') math500 = datasets.load_dataset('HuggingFaceH4/MATH-500')
def map_to_500(example): return { 'problem':โฆ See the full description on the dataset page: https://huggingface.co/datasets/ricdomolm/MATH-500.