Update README.md
Browse files
README.md
CHANGED
@@ -39,7 +39,7 @@ We provide 6 variants (of which 3 base and 3 instruction-tuned models):
|
|
39 |
- **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
|
40 |
- **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
|
41 |
|
42 |
-
For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](
|
43 |
|
44 |
### Model Description
|
45 |
|
@@ -51,8 +51,8 @@ For benchmark results for all models, including compared to their base models an
|
|
51 |
|
52 |
### Model Sources
|
53 |
|
54 |
-
- **Repository:**
|
55 |
-
- **Paper:**
|
56 |
|
57 |
## Uses
|
58 |
|
|
|
39 |
- **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
|
40 |
- **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
|
41 |
|
42 |
+
For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](https://arxiv.org/pdf/2412.07633).
|
43 |
|
44 |
### Model Description
|
45 |
|
|
|
51 |
|
52 |
### Model Sources
|
53 |
|
54 |
+
- **Repository:** [on Github here](https://github.com/ChocoLlamaModel/ChocoLlama).
|
55 |
+
- **Paper:** [on ArXiv here](https://arxiv.org/pdf/2412.07633).
|
56 |
|
57 |
## Uses
|
58 |
|