text
stringlengths
0
3.84k
File "/tmp/Kwai-Keye_Keye-VL-1_5-8B_0q0ge30.py", line 12, in <module>
model = AutoModel.from_pretrained("Kwai-Keye/Keye-VL-1_5-8B", trust_remote_code=True, torch_dtype="auto")
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 586, in from_pretrained
model_class = get_class_from_dynamic_module(
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **hub_kwargs, **kwargs
)
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 569, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 392, in get_cached_module_file
modules_needed = check_imports(resolved_module_file)
File "/tmp/.cache/uv/environments-v2/d016abb5417bd7d5/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 224, in check_imports
raise ImportError(
...<2 lines>...
)
ImportError: This modeling file requires the following packages that were not found in your environment: einops, flash_attn. Run `pip install einops flash_attn`
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Base | 111.91 GB VRAM requirement
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Base | 111.91 GB VRAM requirement
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Instruct | 111.91 GB VRAM requirement
No suitable GPU found for Kwai-Klear/Klear-46B-A2.5B-Instruct | 111.91 GB VRAM requirement
Traceback (most recent call last):
File "/tmp/LiquidAI_LFM2-8B-A1B_0wIkn8Q.py", line 16, in <module>
pipe = pipeline("text-generation", model="LiquidAI/LFM2-8B-A1B", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 922, in pipeline
config = AutoConfig.from_pretrained(
model, _from_pipeline=task, code_revision=code_revision, **hub_kwargs, **model_kwargs
)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1347, in from_pretrained
config_class = get_class_from_dynamic_module(
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **kwargs
)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 604, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 406, in get_cached_module_file
resolved_module_file = cached_file(
pretrained_model_name_or_path,
...<9 lines>...
_commit_hash=_commit_hash,
)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
File "/tmp/.cache/uv/environments-v2/1b57f0f089e40716/lib/python3.13/site-packages/transformers/utils/hub.py", line 583, in cached_files
raise OSError(
...<2 lines>...
)
OSError: LiquidAI/LFM2-8B-A1B does not appear to have a file named configuration_lfm2_moe.py. Checkout 'https://huggingface.co/LiquidAI/LFM2-8B-A1B/tree/main' for available files.
Traceback (most recent call last):
File "/tmp/LiquidAI_LFM2-8B-A1B_1wsNwgk.py", line 17, in <module>
model = AutoModelForCausalLM.from_pretrained("LiquidAI/LFM2-8B-A1B", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 549, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
~~~~~~~~~~~~~~~~~~~~~~~~~~^
pretrained_model_name_or_path,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
...<4 lines>...
**kwargs,
^^^^^^^^^
)
^
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1347, in from_pretrained
config_class = get_class_from_dynamic_module(
class_ref, pretrained_model_name_or_path, code_revision=code_revision, **kwargs
)
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 604, in get_class_from_dynamic_module
final_module = get_cached_module_file(
repo_id,
...<8 lines>...
repo_type=repo_type,
)
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/dynamic_module_utils.py", line 406, in get_cached_module_file
resolved_module_file = cached_file(
pretrained_model_name_or_path,
...<9 lines>...
_commit_hash=_commit_hash,
)
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
File "/tmp/.cache/uv/environments-v2/dcce77d1c9c002ff/lib/python3.13/site-packages/transformers/utils/hub.py", line 583, in cached_files
raise OSError(
...<2 lines>...
)
OSError: LiquidAI/LFM2-8B-A1B does not appear to have a file named configuration_lfm2_moe.py. Checkout 'https://huggingface.co/LiquidAI/LFM2-8B-A1B/tree/main' for available files.
Traceback (most recent call last):
File "/tmp/LiquidAI_LFM2-VL-1.6B_0X7Okei.py", line 13, in <module>
pipe = pipeline("image-text-to-text", model="LiquidAI/LFM2-VL-1.6B", trust_remote_code=True)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1198, in pipeline
raise e
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1189, in pipeline
processor = AutoProcessor.from_pretrained(processor, _from_pipeline=task, **hub_kwargs, **model_kwargs)
File "/tmp/.cache/uv/environments-v2/6e3e46311085f7ee/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 382, in from_pretrained
processor_class = get_class_from_dynamic_module(
processor_auto_map, pretrained_model_name_or_path, **kwargs
)