Update README.md
Browse files
README.md
CHANGED
@@ -68,14 +68,11 @@ pip install -U peft transformers
|
|
68 |
```
|
69 |
|
70 |
```python
|
71 |
-
from peft import PeftModel, PeftConfig
|
72 |
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, pipeline
|
73 |
|
74 |
max_tokens = 8096
|
75 |
|
76 |
print("Loading...")
|
77 |
-
config = PeftConfig.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora")
|
78 |
-
# base_model = AutoModelForCausalLM.from_pretrained("cognitivecomputations/dolphin-2.2.1-mistral-7b")
|
79 |
model = AutoModelForCausalLM.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", quantization_config=BitsAndBytesConfig(load_in_4bit=True), torch_dtype="auto")
|
80 |
tokenizer = AutoTokenizer.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", torch_dtype="auto")
|
81 |
|
|
|
68 |
```
|
69 |
|
70 |
```python
|
|
|
71 |
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, pipeline
|
72 |
|
73 |
max_tokens = 8096
|
74 |
|
75 |
print("Loading...")
|
|
|
|
|
76 |
model = AutoModelForCausalLM.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", quantization_config=BitsAndBytesConfig(load_in_4bit=True), torch_dtype="auto")
|
77 |
tokenizer = AutoTokenizer.from_pretrained("wasertech/assistant-dolphin-2.2.1-mistral-7b-e1-qlora", torch_dtype="auto")
|
78 |
|