matthieumeeus97 commited on
Commit
929af7d
·
verified ·
1 Parent(s): 2fab3cc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -39,7 +39,7 @@ We provide 6 variants (of which 3 base and 3 instruction-tuned models):
39
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
40
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
41
 
42
- For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](some_url).
43
 
44
  ### Model Description
45
 
@@ -51,8 +51,8 @@ For benchmark results for all models, including compared to their base models an
51
 
52
  ### Model Sources
53
 
54
- - **Repository:** Will be released soon.
55
- - **Paper:** Will be released soon.
56
 
57
  ## Uses
58
 
 
39
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
40
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
41
 
42
+ For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](https://arxiv.org/pdf/2412.07633).
43
 
44
  ### Model Description
45
 
 
51
 
52
  ### Model Sources
53
 
54
+ - **Repository:** [on Github here](https://github.com/ChocoLlamaModel/ChocoLlama).
55
+ - **Paper:** [on ArXiv here](https://arxiv.org/pdf/2412.07633).
56
 
57
  ## Uses
58