JiunYi's picture
Update README.md
e36f019
|
raw
history blame
1.03 kB
metadata
license: apache-2.0
datasets:
  - shibing624/alpaca-zh
language:
  - zh
tags:
  - LoRA
  - LLaMA
  - Alpaca
  - PEFT
  - int8

Model Card for llama-7b-alpaca-zh-20k

Uses

Direct Use

from peft import PeftModel
from transformers import GenerationConfig, LlamaForCausalLM, LlamaTokenizer


max_memory = {i: "15GIB" for i in range(torch.cuda.device_count())}
tokenizer = LlamaTokenizer.from_pretrained(base_model)
model = LlamaForCausalLM.from_pretrained(
    base_model,
    load_in_8bit=True,
    torch_dtype=torch.float16,
    device_map="auto"
    max_memory=max_memory
)
model = PeftModel.from_pretrained(
    model,
    lora_weights,
    torch_dtype=torch.float16,
    max_memory=max_memory
)