🧠 Math Reasoning SFT Model

This model is a Supervised Fine-Tuned (SFT) version of a base language model on a high-quality mathematics dataset.
It has been fine-tuned for 3 epochs on a 46k-sample dataset using the Hugging Face πŸ€— transformers library.

πŸ“Š Evaluation Results

Evaluated on a held-out validation set covering a variety of mathematical domains:

Subset Accuracy
Olympiad 0.484
Minerva 0.460
Math 0.874
AMC 0.610
AIME24 0.332
AIME25 0.263
AVG 0.504

Accuracy is measured via exact match on extracted final answers using rule-based labeling functions.

πŸ› οΈ Model Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("your-org/math-sft-model")
tokenizer = AutoTokenizer.from_pretrained("your-org/math-sft-model")

prompt = "If 2x+3=72x + 3 = 7, what is the value of xx?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=128)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
6
Safetensors
Model size
7.62B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support