fearlessdots commited on
Commit
7ab2b01
1 Parent(s): 0d31bb3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -4
README.md CHANGED
@@ -34,12 +34,19 @@ The LoRA used to create this model is available at [https://huggingface.co/fearl
34
 
35
  ## Fine Tuning
36
 
 
 
 
 
 
 
 
37
  ### - PEFT Parameters
38
 
39
- - lora_alpha=64,
40
- - lora_dropout=0.05,
41
- - r=128,
42
- - bias="none",
43
 
44
  ### - Training Arguments
45
 
@@ -62,6 +69,7 @@ The LoRA used to create this model is available at [https://huggingface.co/fearl
62
  ## Credits
63
 
64
  - Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
 
65
  - failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
66
  - NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
67
  - Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
 
34
 
35
  ## Fine Tuning
36
 
37
+ ### - Quantization Configuration
38
+
39
+ - load_in_4bit=True
40
+ - bnb_4bit_quant_type="fp4"
41
+ - bnb_4bit_compute_dtype=compute_dtype
42
+ - bnb_4bit_use_double_quant=False
43
+ -
44
  ### - PEFT Parameters
45
 
46
+ - lora_alpha=64
47
+ - lora_dropout=0.05
48
+ - r=128
49
+ - bias="none"
50
 
51
  ### - Training Arguments
52
 
 
69
  ## Credits
70
 
71
  - Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
72
+ - HuggingFace: for hosting this model and for creating the fine-tuning tools used;
73
  - failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
74
  - NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
75
  - Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]