ykarout commited on
Commit
3312fd9
·
verified ·
1 Parent(s): 1bef6ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -36,7 +36,7 @@ A reasoning-enhanced version of Mixtral-8x7B-Instruct-v0.1, fine-tuned on reason
36
 
37
  This model is a fine-tuned version of Mixtral-8x7B-Instruct-v0.1 that has been trained on reasoning-rich datasets to improve its step-by-step thinking and problem-solving capabilities. The model learns to generate explicit reasoning traces similar to those produced by advanced reasoning models like DeepSeek-R1.
38
 
39
- - **Developed by:** Yehya Karout
40
  - **Model type:** Mixture of Experts (MoE) Language Model
41
  - **Language(s) (NLP):** English, Arabic, French, Spanish (inherited from base model)
42
  - **License:** Apache 2.0
@@ -211,16 +211,13 @@ The model exhibits improved reasoning capabilities compared to the base Mixtral
211
  ```bibtex
212
  @model{mixtral-deepseek-r1-distill,
213
  title={Mixtral-8x7B-DeepSeek-R1-Distill: Reasoning-Enhanced Mixture of Experts},
214
- author={Karout, Yehya},
215
  year={2025},
216
  publisher={Hugging Face},
217
  url={https://huggingface.co/ykarout/Mixtral-8x7B-DeepSeek-R1-Distill-16bit}
218
  }
219
  ```
220
 
221
- ## Model Card Authors
222
-
223
- Yehya Karout
224
 
225
  ## Model Card Contact
226
 
 
36
 
37
  This model is a fine-tuned version of Mixtral-8x7B-Instruct-v0.1 that has been trained on reasoning-rich datasets to improve its step-by-step thinking and problem-solving capabilities. The model learns to generate explicit reasoning traces similar to those produced by advanced reasoning models like DeepSeek-R1.
38
 
39
+ - **Developed by:** ykarout
40
  - **Model type:** Mixture of Experts (MoE) Language Model
41
  - **Language(s) (NLP):** English, Arabic, French, Spanish (inherited from base model)
42
  - **License:** Apache 2.0
 
211
  ```bibtex
212
  @model{mixtral-deepseek-r1-distill,
213
  title={Mixtral-8x7B-DeepSeek-R1-Distill: Reasoning-Enhanced Mixture of Experts},
214
+ author={ykarout},
215
  year={2025},
216
  publisher={Hugging Face},
217
  url={https://huggingface.co/ykarout/Mixtral-8x7B-DeepSeek-R1-Distill-16bit}
218
  }
219
  ```
220
 
 
 
 
221
 
222
  ## Model Card Contact
223