runtime error
Exit code: 1. Reason: use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. @torch.library.impl_abstract("xformers_flash::flash_fwd") /usr/local/lib/python3.10/site-packages/xformers/ops/fmha/flash.py:344: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. @torch.library.impl_abstract("xformers_flash::flash_bwd") Loading pipeline components...: 0%| | 0/7 [00:00<?, ?it/s][A Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s][A Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] Loading pipeline components...: 0%| | 0/7 [00:01<?, ?it/s] Traceback (most recent call last): File "/home/user/app/app.py", line 16, in <module> pipe = FluxPipeline.from_pretrained("enhanceaiteam/kalpana", torch_dtype=torch.bfloat16) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/diffusers/pipelines/pipeline_utils.py", line 876, in from_pretrained loaded_sub_model = load_sub_model( File "/usr/local/lib/python3.10/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 700, in load_sub_model loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3941, in from_pretrained ) = cls._load_pretrained_model( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4395, in _load_pretrained_model state_dict = load_state_dict(shard_file, is_quantized=is_quantized) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 549, in load_state_dict with safe_open(checkpoint_file, framework="pt") as f: safetensors_rust.SafetensorError: Error while deserializing header: MetadataIncompleteBuffer
Container logs:
Fetching error logs...