fearlessdots
commited on
Commit
•
7ab2b01
1
Parent(s):
0d31bb3
Update README.md
Browse files
README.md
CHANGED
@@ -34,12 +34,19 @@ The LoRA used to create this model is available at [https://huggingface.co/fearl
|
|
34 |
|
35 |
## Fine Tuning
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
### - PEFT Parameters
|
38 |
|
39 |
-
- lora_alpha=64
|
40 |
-
- lora_dropout=0.05
|
41 |
-
- r=128
|
42 |
-
- bias="none"
|
43 |
|
44 |
### - Training Arguments
|
45 |
|
@@ -62,6 +69,7 @@ The LoRA used to create this model is available at [https://huggingface.co/fearl
|
|
62 |
## Credits
|
63 |
|
64 |
- Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
|
|
|
65 |
- failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
|
66 |
- NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
|
67 |
- Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
|
|
|
34 |
|
35 |
## Fine Tuning
|
36 |
|
37 |
+
### - Quantization Configuration
|
38 |
+
|
39 |
+
- load_in_4bit=True
|
40 |
+
- bnb_4bit_quant_type="fp4"
|
41 |
+
- bnb_4bit_compute_dtype=compute_dtype
|
42 |
+
- bnb_4bit_use_double_quant=False
|
43 |
+
-
|
44 |
### - PEFT Parameters
|
45 |
|
46 |
+
- lora_alpha=64
|
47 |
+
- lora_dropout=0.05
|
48 |
+
- r=128
|
49 |
+
- bias="none"
|
50 |
|
51 |
### - Training Arguments
|
52 |
|
|
|
69 |
## Credits
|
70 |
|
71 |
- Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
|
72 |
+
- HuggingFace: for hosting this model and for creating the fine-tuning tools used;
|
73 |
- failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
|
74 |
- NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
|
75 |
- Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
|