Reasoning-Small-1B

A compact and efficient 1B parameter model fine-tuned for reasoning tasks using extracted Chain-of-Thought (CoT) data.

Training Details

  • Base Model: LLaMA 3.1 (1B parameters)
  • Dataset: 2.67M extracted Chain-of-Thought examples

Benchmarking

Benchmark Performance
GSM8K (Math) 70.8%
ARC (Reasoning) 69.4%
Math(CoT) 40.8%

License

This repository is released under the MIT License.


library_name: transformers tags: - text-generation - causal-lm

Downloads last month
15
Safetensors
Model size
1.24B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Kameshr/reasoning-small-1B

Quantizations
1 model