runtime error

Exit code: 1. Reason: tokenizer_config.json: 0%| | 0.00/365k [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 365k/365k [00:00<00:00, 90.1MB/s] tokenizer.json: 0%| | 0.00/9.78M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 9.78M/9.78M [00:00<00:00, 38.5MB/s] special_tokens_map.json: 0%| | 0.00/826 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 826/826 [00:00<00:00, 6.24MB/s] config.json: 0%| | 0.00/909 [00:00<?, ?B/s] config.json: 100%|██████████| 909/909 [00:00<00:00, 7.13MB/s] You have loaded a BitNet model on CPU and have a CUDA device available, make sure to set your model on a GPU device in order to run your model. `low_cpu_mem_usage` was None, now default to True since model is quantized. Traceback (most recent call last): File "/home/user/app/app.py", line 32, in <module> model = AutoModelForCausalLM.from_pretrained( File "/home/user/.pyenv/versions/3.10.16/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/home/user/.pyenv/versions/3.10.16/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3880, in from_pretrained raise EnvironmentError( OSError: tiiuae/Falcon3-7B-Instruct-1.58bit does not appear to have a file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt or flax_model.msgpack.

Container logs:

Fetching error logs...