Update README.md
Browse files
README.md
CHANGED
|
@@ -3,10 +3,10 @@ license: llama3.1
|
|
| 3 |
---
|
| 4 |
# π Introduction
|
| 5 |
|
| 6 |
-
This model is based on the LLaMA3.1-8B-Chinese
|
| 7 |
|
| 8 |
Data: We collected and cleaned high-quality medical knowledge data. With the help of commercial large models, we expanded the training set to about 8,000 high-quality instruction samples, covering key medical subfields such as treatment and pharmacology.
|
| 9 |
|
| 10 |
Framework: DeepSpeed
|
| 11 |
|
| 12 |
-
If you are interested in technologies such as DeepSpeed distributed training, LoRA fine-tuning, VLLM-based high-concurrency inference service deployment, or model quantization and compression, feel free to check out my open-source project,
|
|
|
|
| 3 |
---
|
| 4 |
# π Introduction
|
| 5 |
|
| 6 |
+
This model is based on the [LLaMA3.1-8B-Chinese](https://huggingface.co/shenzhi-wang/Llama3.1-8B-Chinese-Chat), with a focus on medical question answering tasks. By combining DeepSpeed distributed training and LoRA fine-tuning (including both SFT and DPO), the model achieves significant improvements in accuracy and professionalism when handling specialized medical questions. The training was conducted on 4 NVIDIA RTX A6000 GPUs.
|
| 7 |
|
| 8 |
Data: We collected and cleaned high-quality medical knowledge data. With the help of commercial large models, we expanded the training set to about 8,000 high-quality instruction samples, covering key medical subfields such as treatment and pharmacology.
|
| 9 |
|
| 10 |
Framework: DeepSpeed
|
| 11 |
|
| 12 |
+
If you are interested in technologies such as DeepSpeed distributed training, LoRA fine-tuning, VLLM-based high-concurrency inference service deployment, or model quantization and compression, feel free to check out my [open-source project](https://github.com/RyanZxucheng/deepspeed-sft), provided for everyone to learn from.
|