RRT1-3B

A fine-tuned 3B parameter model specialized for reasoning and chain-of-thought tasks

Model Details

This model is a fine-tuned version of unsloth/Qwen2.5-3B-Instruct-bnb-4bit using the Unsloth framework with LoRA (Low-Rank Adaptation) for efficient training.

  • Developed by: theprint
  • Model type: Causal Language Model (Fine-tuned with LoRA)
  • Language: en
  • License: apache-2.0
  • Base model: unsloth/Qwen2.5-3B-Instruct-bnb-4bit
  • Fine-tuning method: LoRA with rank 128

Intended Use

Reasoning, chain-of-thought, and general instruction following

Training Details

Training Data

ShareGPT conversations with chain-of-thought reasoning examples

  • Dataset: AiCloser/sharegpt_cot_dataset
  • Format: sharegpt

Training Procedure

  • Training epochs: 3
  • LoRA rank: 128
  • Learning rate: 0.0002
  • Batch size: 4
  • Framework: Unsloth + transformers + PEFT
  • Hardware: NVIDIA RTX 5090

Usage

from unsloth import FastLanguageModel
import torch

# Load model and tokenizer
model, tokenizer = FastLanguageModel.from_pretrained(
    model_name="theprint/RRT1-3B",
    max_seq_length=4096,
    dtype=None,
    load_in_4bit=True,
)

# Enable inference mode
FastLanguageModel.for_inference(model)

# Example usage
inputs = tokenizer(["Your prompt here"], return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

GGUF Quantized Versions

Quantized GGUF versions are available in the gguf/ directory for use with llama.cpp:

  • RRT1-3B-q4_k_m.gguf - 4-bit quantization (recommended for most use cases)
  • RRT1-3B-q5_k_m.gguf - 5-bit quantization (higher quality)
  • RRT1-3B-q8_0.gguf - 8-bit quantization (highest quality)

Limitations

May hallucinate or provide incorrect information. Not suitable for critical decision making.

Citation

If you use this model, please cite:

@misc{rrt1_3b,
  title={RRT1-3B: Fine-tuned Qwen2.5-3B-Instruct-bnb-4bit},
  author={theprint},
  year={2025},
  publisher={Hugging Face},
  url={https://huggingface.co/theprint/RRT1-3B}
}
Downloads last month
71
GGUF
Model size
3.09B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for theprint/RRT1-3B

Base model

Qwen/Qwen2.5-3B
Adapter
(1)
this model

Dataset used to train theprint/RRT1-3B