quantization config from the arguments has no `quant_method` attribute

#1
by endremoen - opened

Im trying to run the provided 4 bit quantized model (Ourse/AM-Thinking-v1-mlx-4Bit) on ubuntu, but get an error:
"model = AutoModelForCausalLM.from_pretrained("Ourse/AM-Thinking-v1-mlx-4Bit", device_map='auto')
...ValueError: The model's quantization config from the arguments has no quant_method attribute. Make sure that the model has been correctly quantized". The code I run is the provided for transformers:

Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("Ourse/AM-Thinking-v1-mlx-4Bit")
model = AutoModelForCausalLM.from_pretrained("Ourse/AM-Thinking-v1-mlx-4Bit")

Sign up or log in to comment