Failed to server the model with VLLM 0.8.2

#1
by MRU4913 - opened

llm8017 | File "/usr/local/lib/python3.12/dist-packages/vllm/model_executor/model_loader/loader.py", line 1278, in load_model
llm8017 | self._load_weights(model_config, model)
llm8017 | File "/usr/local/lib/python3.12/dist-packages/vllm/model_executor/model_loader/loader.py", line 1188, in _load_weights
llm8017 | loaded_weights = model.load_weights(qweight_iterator)
llm8017 | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
llm8017 | File "/usr/local/lib/python3.12/dist-packages/vllm/model_executor/models/commandr.py", line 468, in load_weights
llm8017 | weight_loader(param, loaded_weight)
llm8017 | File "/usr/local/lib/python3.12/dist-packages/vllm/model_executor/layers/linear.py", line 1228, in weight_loader
llm8017 | assert param_data.shape == loaded_weight.shape
llm8017 | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
llm8017 | AssertionError
Loading safetensors checkpoint shards: 0% Completed | 0/6 [00:00<?, ?it/s]
llm8017 |


As above

MRU4913 changed discussion status to closed
This comment has been hidden (marked as Resolved)

Sign up or log in to comment