runtime error

Exit code: 1. Reason: ==15.0.1 xxhash==3.6.0 yarl==1.22.0 zipp==3.23.0 /usr/local/lib/python3.10/site-packages/torch/amp/autocast_mode.py:270: UserWarning: User provided device_type of 'cuda', but CUDA is not available. Disabling warnings.warn( model_index.json: 0%| | 0.00/83.0 [00:00<?, ?B/s] model_index.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 83.0/83.0 [00:00<00:00, 430kB/s] config.json: 0%| | 0.00/730 [00:00<?, ?B/s] config.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 730/730 [00:00<00:00, 4.10MB/s] vae/diffusion_pytorch_model.safetensors: 0%| | 0.00/254M [00:00<?, ?B/s] vae/diffusion_pytorch_model.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 254M/254M [00:01<00:00, 246MB/s] vae/diffusion_pytorch_model.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 254M/254M [00:01<00:00, 246MB/s] config.json: 0%| | 0.00/339 [00:00<?, ?B/s] config.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 339/339 [00:00<00:00, 2.07MB/s] Traceback (most recent call last): File "/app/app.py", line 24, in <module> vae, transformer, embeds_dict, processing_resolution = load_network(uri_base, uri_lora, device) File "/app/windowseat_inference.py", line 103, in load_network transformer = load_qwen_transformer(uri_base, device) File "/app/windowseat_inference.py", line 67, in load_qwen_transformer transformer = QwenImageTransformer2DModel.from_pretrained( File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/diffusers/models/modeling_utils.py", line 1097, in from_pretrained hf_quantizer.validate_environment(torch_dtype=torch_dtype, from_flax=from_flax, device_map=device_map) File "/usr/local/lib/python3.10/site-packages/diffusers/quantizers/bitsandbytes/bnb_quantizer.py", line 65, in validate_environment raise RuntimeError("No GPU found. A GPU is needed for quantization.") RuntimeError: No GPU found. A GPU is needed for quantization.

Container logs:

Fetching error logs...