wasertech commited on
Commit
8f34b7d
·
verified ·
1 Parent(s): 1cb672e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -3
README.md CHANGED
@@ -68,14 +68,11 @@ pip install -U peft transformers
68
  ```
69
 
70
  ```python
71
- from peft import PeftModel, PeftConfig
72
  from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, pipeline
73
 
74
  max_tokens = 8096
75
 
76
  print("Loading...")
77
- config = PeftConfig.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora")
78
- # base_model = AutoModelForCausalLM.from_pretrained("cognitivecomputations/dolphin-2.2.1-mistral-7b")
79
  model = AutoModelForCausalLM.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", quantization_config=BitsAndBytesConfig(load_in_4bit=True), torch_dtype="auto")
80
  tokenizer = AutoTokenizer.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", torch_dtype="auto")
81
 
 
68
  ```
69
 
70
  ```python
 
71
  from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, pipeline
72
 
73
  max_tokens = 8096
74
 
75
  print("Loading...")
 
 
76
  model = AutoModelForCausalLM.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", quantization_config=BitsAndBytesConfig(load_in_4bit=True), torch_dtype="auto")
77
  tokenizer = AutoTokenizer.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", torch_dtype="auto")
78