Model Card for Qwen2.5-1.5B-Medical-Az

Model Description

This model is a fine-tuned version of Qwen2.5-1.5B-Instruct on an Azerbaijani medical reasoning dataset. It is designed to understand complex medical instructions, interpret clinical cases, and generate informed answers in Azerbaijani.

  • Developed by: Rustam Shiriyev
  • Model type: Causal Language Model
  • Language(s) (NLP): Azerbaijani
  • License: MIT
  • Finetuned from model: unsloth/Qwen2.5-1.5B-Instruct
  • Fine-tuning Method: Supervised Fine-Tuning (SFT) using Unsloth + LoRA
  • Domain: Medical Question Answering / Reasoning
  • Dataset: The training data consists of ~19,696 rows, translated from the FreedomIntelligence/medical-o1-reasoning-SFT dataset

Uses

Direct Use

You can use this model directly for:

  • Medical QA tasks in Azerbaijani
  • Evaluating LLMs' ability to reason about clinical data in low-resource languages
  • Generating educational prompts or tutoring-style medical answers
  • Research on instruction tuning and localization of medical language models

Out-of-Scope Use

  • Use in life-critical medical applications
  • Any application where incorrect answers could cause harm
  • Use by patients or non-medical professionals for self-diagnosis
  • Deployment in commercial healthcare systems without regulatory oversight or expert validation

Bias, Risks, and Limitations

The model has not been clinically validated and must not be used for real medical decision-making. Trained only on a single-source dataset, so it may not generalize to all medical topics. Performance in zero-shot generalisation (e.g., English → Azerbaijani medical transfer) has not been tested.

How to Get Started with the Model

login(token="")  

tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen2.5-1.5B-Instruct",)
base_model = AutoModelForCausalLM.from_pretrained(
    "unsloth/Qwen2.5-1.5B-Instruct",
    device_map={"": 0}, token=""
)

model = PeftModel.from_pretrained(base_model,"Rustamshry/Qwen2.5-1.5B-Medical-Az")

question = "45 yaşlı kişi qəfil danışıqda pozulma, yeriyişində dəyişiklik və titrəmə meydana gəlir. Ən ehtimal diaqnoz nədir?"
prompt = f"""### Question:\n{question}\n\n### Response:\n"""

input_ids = tokenizer(prompt, return_tensors="pt").to(model.device)

outputs = model.generate(
    **input_ids, 
    max_new_tokens=2000,
    #temperature=0.6,
    #top_p=0.95,
    #do_sample=True,
    #eos_token_id=tokenizer.eos_token_id
)

print(tokenizer.decode(outputs[0]))

Training Details

Training Data

The model was fine-tuned on a translated and cleaned version of FreedomIntelligence/medical-o1-reasoning-SFT, which was manually converted into Azerbaijani. All examples were filtered for translation quality and medical relevance.

  • Dataset (Translated): Rustamshry/medical_o1_reasoning_SFT_az
  • Link of Original Dataset: huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT

Training Procedure

The model was trained using supervised fine-tuning (SFT) with parameter-efficient fine-tuning (PEFT) via LoRA, using the Unsloth library for memory-optimized training.

  • Training regime: fp16
  • Epochs: 2
  • Batch size: 2
  • Gradient accumulation steps: 4
  • Max sequence lenght: 2000
  • Learning rate: 2e-5
  • Optimizer: adamw_torch
  • fp16: True
  • LoRa rank: 6
  • Aplha: 16
  • Target Modules: 28 layers with 28 QKV, 28 O, and 28 MLP.

Speeds, Sizes, Times

  • Training speed: 0.12 steps/sec
  • Total training time: 11 hours, 26 minutes
  • Total training steps: 4924

Hardware

  • GPUs Used:. NVIDIA Tesla T4 GPUs via Kaggle Notebook

Result

  • Training loss: 2.68 → 1.63

Framework versions

  • PEFT 0.14.0
Downloads last month
37
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Rustamshry/Qwen2.5-1.5B-Medical-Az

Base model

Qwen/Qwen2.5-1.5B
Adapter
(286)
this model

Dataset used to train Rustamshry/Qwen2.5-1.5B-Medical-Az