text stringlengths 0 3.84k |
|---|
File "/tmp/Kwai-Keye_Keye-VL-1_5-8B_0q0ge30.py", line 12, in <module> |
model = AutoModel.from_pretrained("Kwai-Keye/Keye-VL-1_5-8B", trust_remote_code=True, torch_dtype="auto") |
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 586, in from_pretrained |
model_class = get_class_from_dynamic_module( |
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **hub_kwargs, **kwargs |
) |
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 569, in get_class_from_dynamic_module |
final_module = get_cached_module_file( |
repo_id, |
...<8 lines>... |
repo_type=repo_type, |
) |
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 392, in get_cached_module_file |
modules_needed = check_imports(resolved_module_file) |
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 224, in check_imports |
raise ImportError( |
...<2 lines>... |
) |
ImportError: This modeling file requires the following packages that were not found in your environment: einops, flash_attn. Run `pip install einops flash_attn` |
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Base | 111.91 GB VRAM requirement |
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Base | 111.91 GB VRAM requirement |
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Instruct | 111.91 GB VRAM requirement |
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Instruct | 111.91 GB VRAM requirement |
Traceback (most recent call last): |
File "/tmp/LiquidAI_LFM2-8B-A1B_0wIkn8Q.py", line 16, in <module> |
pipe = pipeline("text-generation", model="LiquidAI/LFM2-8B-A1B", trust_remote_code=True) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 922, in pipeline |
config = AutoConfig.from_pretrained( |
model, _from_pipeline=task, code_revision=code_revision, **hub_kwargs, **model_kwargs |
) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1347, in from_pretrained |
config_class = get_class_from_dynamic_module( |
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **kwargs |
) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 604, in get_class_from_dynamic_module |
final_module = get_cached_module_file( |
repo_id, |
...<8 lines>... |
repo_type=repo_type, |
) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 406, in get_cached_module_file |
resolved_module_file = cached_file( |
pretrained_model_name_or_path, |
...<9 lines>... |
_commit_hash=_commit_hash, |
) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file |
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs) |
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/utils/hub.py", line 583, in cached_files |
raise OSError( |
...<2 lines>... |
) |
OSError: LiquidAI/LFM2-8B-A1B does not appear to have a file named configuration_lfm2_moe.py. Checkout 'https://huggingface.co/LiquidAI/LFM2-8B-A1B/tree/main' for available files. |
Traceback (most recent call last): |
File "/tmp/LiquidAI_LFM2-8B-A1B_1wsNwgk.py", line 17, in <module> |
model = AutoModelForCausalLM.from_pretrained("LiquidAI/LFM2-8B-A1B", trust_remote_code=True) |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 549, in from_pretrained |
config, kwargs = AutoConfig.from_pretrained( |
~~~~~~~~~~~~~~~~~~~~~~~~~~^ |
pretrained_model_name_or_path, |
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ |
...<4 lines>... |
**kwargs, |
^^^^^^^^^ |
) |
^ |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1347, in from_pretrained |
config_class = get_class_from_dynamic_module( |
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **kwargs |
) |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 604, in get_class_from_dynamic_module |
final_module = get_cached_module_file( |
repo_id, |
...<8 lines>... |
repo_type=repo_type, |
) |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 406, in get_cached_module_file |
resolved_module_file = cached_file( |
pretrained_model_name_or_path, |
...<9 lines>... |
_commit_hash=_commit_hash, |
) |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file |
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs) |
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/utils/hub.py", line 583, in cached_files |
raise OSError( |
...<2 lines>... |
) |
OSError: LiquidAI/LFM2-8B-A1B does not appear to have a file named configuration_lfm2_moe.py. Checkout 'https://huggingface.co/LiquidAI/LFM2-8B-A1B/tree/main' for available files. |
Traceback (most recent call last): |
File "/tmp/LiquidAI_LFM2-VL-1.6B_0X7Okei.py", line 13, in <module> |
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-1.6B", trust_remote_code=True) |
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1198, in pipeline |
raise e |
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1189, in pipeline |
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs) |
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 382, in from_pretrained |
processor_class = get_class_from_dynamic_module( |
processor_auto_map, pretrained_model_name_or_path, **kwargs |
) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.