runtime error

Exit code: 1. Reason: | 608M/4.78G [00:01<00:07, 590MB/s] model-00009-of-00010.safetensors: 43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 2.08G/4.78G [00:02<00:02, 1.10GB/s] model-00009-of-00010.safetensors: 74%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 3.52G/4.78G [00:03<00:01, 1.25GB/s] model-00009-of-00010.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 4.78G/4.78G [00:03<00:00, 1.27GB/s] Downloading shards: 90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 9/10 [00:39<00:04, 4.16s/it] model-00010-of-00010.safetensors: 0%| | 0.00/3.90G [00:00<?, ?B/s] model-00010-of-00010.safetensors: 12%|β–ˆβ– | 472M/3.90G [00:01<00:07, 471MB/s] model-00010-of-00010.safetensors: 50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 1.96G/3.90G [00:02<00:01, 1.06GB/s] model-00010-of-00010.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 3.90G/3.90G [00:03<00:00, 1.29GB/s] Downloading shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 10/10 [00:42<00:00, 3.89s/it] Downloading shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 10/10 [00:42<00:00, 4.30s/it] Loading checkpoint shards: 0%| | 0/10 [00:00<?, ?it/s] Loading checkpoint shards: 50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 5/10 [00:01<00:01, 4.21it/s] Loading checkpoint shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 10/10 [00:02<00:00, 4.40it/s] Loading checkpoint shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 10/10 [00:02<00:00, 4.37it/s] generation_config.json: 0%| | 0.00/154 [00:00<?, ?B/s] generation_config.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 154/154 [00:00<00:00, 958kB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 203, in <module> model, tokenizer = initialize_model() File "/home/user/app/app.py", line 147, in initialize_model ).to("cuda") File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3149, in to raise ValueError( ValueError: `.to` is not supported for `8-bit` bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct `dtype`.

Container logs:

Fetching error logs...