kevin009 commited on
Commit
640b720
·
verified ·
1 Parent(s): e8907b0

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/meta-llama-3.1-8b-instruct
3
+ tags:
4
+ - llama adapter
5
+ - trl
6
+ - llama3.1 8b
7
+ license: apache-2.0
8
+ language:
9
+ - en
10
+ ---
11
+ ## Model Overview
12
+ A LoRA (Low-Rank Adaptation) decomposed from base model from base to instruct
13
+
14
+ ## Model Details
15
+ - Base Model: (meta-llama/Llama-3.1-8B) - (meta-llama/Llama-3.1-8B-instruct)
16
+ - Adaptation Method: LoRA
17
+
18
+ ## Training Configuration
19
+ ### Training Hyperparameters
20
+ - Rank (r): 16
21
+ - Alpha: 16
22
+
23
+ ### LoRA Configuration
24
+ - Rank (r): 16
25
+ - Alpha: 16
26
+ - Target Modules:
27
+ - `q_proj` (Query projection)
28
+ - `k_proj` (Key projection)
29
+ - `v_proj` (Value projection)
30
+ - `o_proj` (Output projection)
31
+ - `up_proj` (Upsampling projection)
32
+ - `down_proj` (Downsampling projection)
33
+ - `gate_proj` (Gate projection)
34
+
35
+ ## Usage
36
+ This adapter must be used in conjunction with the base Llama-3.1-8B-instruct model.
37
+
38
+ ### Loading the Model
39
+ ```python
40
+ from peft import PeftModel, PeftConfig
41
+ from transformers import AutoModelForCausalLM, AutoTokenizer
42
+
43
+ # Load base model
44
+ base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.1-8B-instruct")
45
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B-instruct")
46
+
47
+ # Load LoRA adapter
48
+ model = PeftModel.from_pretrained(base_model, "path_to_adapter")
49
+ ```
50
+
51
+ ## Limitations and Biases
52
+ - This adapter might inherits some limitations and biases present in the base Llama-3.1-8B-instruct model