Ryyyyyyyan commited on
Commit
627220c
Β·
verified Β·
1 Parent(s): 0796fb2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -3,10 +3,10 @@ license: llama3.1
3
  ---
4
  # πŸš€ Introduction
5
 
6
- This model is based on the LLaMA3.1-8B-Chinese open-source foundation model, with a focus on medical question answering tasks. By combining DeepSpeed distributed training and LoRA fine-tuning (including both SFT and DPO), the model achieves significant improvements in accuracy and professionalism when handling specialized medical questions. The training was conducted on 4 NVIDIA RTX A6000 GPUs.
7
 
8
  Data: We collected and cleaned high-quality medical knowledge data. With the help of commercial large models, we expanded the training set to about 8,000 high-quality instruction samples, covering key medical subfields such as treatment and pharmacology.
9
 
10
  Framework: DeepSpeed
11
 
12
- If you are interested in technologies such as DeepSpeed distributed training, LoRA fine-tuning, VLLM-based high-concurrency inference service deployment, or model quantization and compression, feel free to check out my open-source project,
 
3
  ---
4
  # πŸš€ Introduction
5
 
6
+ This model is based on the [LLaMA3.1-8B-Chinese](https://huggingface.co/shenzhi-wang/Llama3.1-8B-Chinese-Chat), with a focus on medical question answering tasks. By combining DeepSpeed distributed training and LoRA fine-tuning (including both SFT and DPO), the model achieves significant improvements in accuracy and professionalism when handling specialized medical questions. The training was conducted on 4 NVIDIA RTX A6000 GPUs.
7
 
8
  Data: We collected and cleaned high-quality medical knowledge data. With the help of commercial large models, we expanded the training set to about 8,000 high-quality instruction samples, covering key medical subfields such as treatment and pharmacology.
9
 
10
  Framework: DeepSpeed
11
 
12
+ If you are interested in technologies such as DeepSpeed distributed training, LoRA fine-tuning, VLLM-based high-concurrency inference service deployment, or model quantization and compression, feel free to check out my [open-source project](https://github.com/RyanZxucheng/deepspeed-sft), provided for everyone to learn from.