Update bit4-chat.py
Browse files- bit4-chat.py +1 -1
bit4-chat.py
CHANGED
@@ -10,7 +10,7 @@ quantization_config = BitsAndBytesConfig(
|
|
10 |
)
|
11 |
|
12 |
# Define the model name and path for the quantized model
|
13 |
-
model_name = "
|
14 |
|
15 |
# Load the quantized model with the specified configuration
|
16 |
model = AutoModelForCausalLM.from_pretrained(
|
|
|
10 |
)
|
11 |
|
12 |
# Define the model name and path for the quantized model
|
13 |
+
model_name = "./Llama-3.1-Nemotron-Nano-8B-v1-bnb-4bit"
|
14 |
|
15 |
# Load the quantized model with the specified configuration
|
16 |
model = AutoModelForCausalLM.from_pretrained(
|