Update README.md
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ A reasoning-enhanced version of Mixtral-8x7B-Instruct-v0.1, fine-tuned on reason
|
|
36 |
|
37 |
This model is a fine-tuned version of Mixtral-8x7B-Instruct-v0.1 that has been trained on reasoning-rich datasets to improve its step-by-step thinking and problem-solving capabilities. The model learns to generate explicit reasoning traces similar to those produced by advanced reasoning models like DeepSeek-R1.
|
38 |
|
39 |
-
- **Developed by:**
|
40 |
- **Model type:** Mixture of Experts (MoE) Language Model
|
41 |
- **Language(s) (NLP):** English, Arabic, French, Spanish (inherited from base model)
|
42 |
- **License:** Apache 2.0
|
@@ -211,16 +211,13 @@ The model exhibits improved reasoning capabilities compared to the base Mixtral
|
|
211 |
```bibtex
|
212 |
@model{mixtral-deepseek-r1-distill,
|
213 |
title={Mixtral-8x7B-DeepSeek-R1-Distill: Reasoning-Enhanced Mixture of Experts},
|
214 |
-
author={
|
215 |
year={2025},
|
216 |
publisher={Hugging Face},
|
217 |
url={https://huggingface.co/ykarout/Mixtral-8x7B-DeepSeek-R1-Distill-16bit}
|
218 |
}
|
219 |
```
|
220 |
|
221 |
-
## Model Card Authors
|
222 |
-
|
223 |
-
Yehya Karout
|
224 |
|
225 |
## Model Card Contact
|
226 |
|
|
|
36 |
|
37 |
This model is a fine-tuned version of Mixtral-8x7B-Instruct-v0.1 that has been trained on reasoning-rich datasets to improve its step-by-step thinking and problem-solving capabilities. The model learns to generate explicit reasoning traces similar to those produced by advanced reasoning models like DeepSeek-R1.
|
38 |
|
39 |
+
- **Developed by:** ykarout
|
40 |
- **Model type:** Mixture of Experts (MoE) Language Model
|
41 |
- **Language(s) (NLP):** English, Arabic, French, Spanish (inherited from base model)
|
42 |
- **License:** Apache 2.0
|
|
|
211 |
```bibtex
|
212 |
@model{mixtral-deepseek-r1-distill,
|
213 |
title={Mixtral-8x7B-DeepSeek-R1-Distill: Reasoning-Enhanced Mixture of Experts},
|
214 |
+
author={ykarout},
|
215 |
year={2025},
|
216 |
publisher={Hugging Face},
|
217 |
url={https://huggingface.co/ykarout/Mixtral-8x7B-DeepSeek-R1-Distill-16bit}
|
218 |
}
|
219 |
```
|
220 |
|
|
|
|
|
|
|
221 |
|
222 |
## Model Card Contact
|
223 |
|