Tina - Ablation Studies
Collection
12 items
•
Updated
•
1
Tina (Tiny Reasoning Models via LoRA) models are all fine-tuned adapters on the base model deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B. This LoRA adapter in this repo is fine-tuned with the dataset knoveleng/open-rs. Please refer to our paper Tina: Tiny Reasoning Models via LoRA for more training details.
The Tina model is meant to be used in combination with the base model as a standard adapter. Particularly, we release all checkpoints we have for each Tina model and one could select different checkpoint to use by specifying the subfolder
.
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
base_model = AutoModelForCausalLM.from_pretrained(
"deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(
"deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
)
model = PeftModel.from_pretrained(
base_model,
"Tina-Yi/Tina-Open-RS3-format-only",
subfolder="checkpoint-850" # checkpoint 850 is the best
)
Base model
deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B