TheBloke commited on
Commit
ac17f3c
·
1 Parent(s): d652c81

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -0
README.md CHANGED
@@ -1,3 +1,41 @@
1
  ---
2
  license: other
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: other
3
+ language:
4
+ - en
5
+ pipeline_tag: text2text-generation
6
+ tags:
7
+ - alpaca
8
+ - llama
9
+ - chat
10
+ - gpt4
11
  ---
12
+ # GPT4 Alpaca LoRA 30B HF
13
+
14
+ This is a pre-merged version of the [Chansung GPT4 Alpaca 30B LoRA model](https://huggingface.co/chansung/gpt4-alpaca-lora-30b).
15
+
16
+ It was created by merging the LoRA provided in the above repo with the original Llama 30B model.
17
+
18
+ You will need at least 60GB VRAM to use this model.
19
+
20
+ For a [GPTQ](https://github.com/qwopqwop200/GPTQ-for-LLaMa) quantized 4bit model, usable on a 24GB GPU, see: [GPT4-Alpaca-LoRA-30B-HF](https://huggingface.co/TheBloke/gpt4-alpaca-lora-30B-GPTQ-4bit-128g)
21
+
22
+ # Original GPT4 Alpaca Lora model card
23
+
24
+ This repository comes with LoRA checkpoint to make LLaMA into a chatbot like language model. The checkpoint is the output of instruction following fine-tuning process with the following settings on 8xA100(40G) DGX system.
25
+ - Training script: borrowed from the official [Alpaca-LoRA](https://github.com/tloen/alpaca-lora) implementation
26
+ - Training script:
27
+ ```shell
28
+ python finetune.py \
29
+ --base_model='decapoda-research/llama-30b-hf' \
30
+ --data_path='alpaca_data_gpt4.json' \
31
+ --num_epochs=10 \
32
+ --cutoff_len=512 \
33
+ --group_by_length \
34
+ --output_dir='./gpt4-alpaca-lora-30b' \
35
+ --lora_target_modules='[q_proj,k_proj,v_proj,o_proj]' \
36
+ --lora_r=16 \
37
+ --batch_size=... \
38
+ --micro_batch_size=...
39
+ ```
40
+
41
+ You can find how the training went from W&B report [here](https://wandb.ai/chansung18/gpt4_alpaca_lora/runs/w3syd157?workspace=user-chansung18).