π§ Qwen3-1.7B-MedicalDataset-GGUF
A quantized GGUF-format version of Qwen3-1.7B, fine-tuned on a medical dataset to assist with healthcare-related tasks. Packaged in GGUF format for use with efficient inference engines like llama.cpp
. Released by XformAI-India.
π Model Details
- Base Model: Qwen3-1.7B
- Format: GGUF (quantized)
- Quantization Types: Multiple
- Precision: 2-8 bit quantized
- Use Case: Low-resource and edge device inference for medical AI applications
π§ͺ Intended Use
This quantized model is intended for:
- Medical Q&A on low-resource devices
- Offline chatbot usage in healthcare education
- Mobile inference for healthcare reasoning
π« Limitations & Disclaimer
β οΈ This model is not intended for clinical use.
- Not suitable for real-time diagnostics or emergency decisions.
- May produce inaccurate or hallucinated medical information.
- Use for research and prototyping only.
π How to Use
Run with llama.cpp
:
./main -m qwen3-1.7b-medical-q4_k_m.gguf -p "Explain symptoms of hypertension."
Or from Python using llama-cpp-python
:
from llama_cpp import Llama
llm = Llama(model_path="qwen3-1.7b-medical-q4_k_m.gguf")
output = llm("What are treatment options for Type 2 Diabetes?", max_tokens=200)
print(output)
π Training Info (Base Fine-Tuning)
- Dataset: FreedomIntelligence/medical-o1-reasoning-SFT
- Epochs: 3
- Batch Size: 8
- Learning Rate: 2e-5
- Framework: PyTorch + Transformers
π§ Citation
If you use this model, please cite:
@misc{qwen3medicalgguf2025,
title={Qwen3-1.7B-MedicalDataset-GGUF: A Quantized Medical AI Model},
author={XformAI-India},
year={2025},
url={https://huggingface.co/XformAI-india/Qwen3-1.7B-medicaldataset-gguf}
}
- Downloads last month
- 38
Hardware compatibility
Log In
to view the estimation
1-bit
2-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support