Update app.py
Browse files
app.py
CHANGED
@@ -120,8 +120,8 @@ import transformers
|
|
120 |
# model = load_checkpoint_and_dispatch(model, model_id, device_map=device_map, no_split_module_classes=["GPTJBlock"])
|
121 |
# model.half()
|
122 |
|
123 |
-
MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
124 |
-
|
125 |
|
126 |
CHAT_TEMPLATE = "َAuto"
|
127 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
@@ -134,7 +134,7 @@ quantization_config = BitsAndBytesConfig(
|
|
134 |
bnb_4bit_compute_dtype=torch.bfloat16
|
135 |
)
|
136 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
137 |
-
model =
|
138 |
MODEL_ID,
|
139 |
device_map="auto",
|
140 |
low_cpu_mem_usage=True,
|
|
|
120 |
# model = load_checkpoint_and_dispatch(model, model_id, device_map=device_map, no_split_module_classes=["GPTJBlock"])
|
121 |
# model.half()
|
122 |
|
123 |
+
# MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
124 |
+
MODEL_ID = "microsoft/phi-4"
|
125 |
|
126 |
CHAT_TEMPLATE = "َAuto"
|
127 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
|
|
134 |
bnb_4bit_compute_dtype=torch.bfloat16
|
135 |
)
|
136 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
137 |
+
model = AutoModelForCausalLM.from_pretrained(
|
138 |
MODEL_ID,
|
139 |
device_map="auto",
|
140 |
low_cpu_mem_usage=True,
|