runtime error

Exit code: 1. Reason: e arr[i32,262144] = [3, 3, 3, 3, 3, 4, 3, 3, 3, 3, 3, 3, ... llama_model_loader: - kv 29: tokenizer.ggml.bos_token_id u32 = 2 llama_model_loader: - kv 30: tokenizer.ggml.eos_token_id u32 = 1 llama_model_loader: - kv 31: tokenizer.ggml.unknown_token_id u32 = 3 llama_model_loader: - kv 32: tokenizer.ggml.padding_token_id u32 = 0 llama_model_loader: - kv 33: tokenizer.ggml.add_bos_token bool = true llama_model_loader: - kv 34: tokenizer.ggml.add_eos_token bool = false llama_model_loader: - kv 35: tokenizer.chat_template str = {{ bos_token }}\n{%- if messages[0]['r... llama_model_loader: - kv 36: tokenizer.ggml.add_space_prefix bool = false llama_model_loader: - kv 37: general.quantization_version u32 = 2 llama_model_loader: - type f32: 157 tensors llama_model_loader: - type q8_0: 183 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q8_0 print_info: file size = 1013.54 MiB (8.50 BPW) llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'gemma3' llama_model_load_from_file_impl: failed to load model Traceback (most recent call last): File "/home/user/app/app.py", line 6, in <module> llm = Llama.from_pretrained( File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 2357, in from_pretrained return cls( File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 372, in __init__ internals.LlamaModel( File "/usr/local/lib/python3.10/site-packages/llama_cpp/_internals.py", line 56, in __init__ raise ValueError(f"Failed to load model from file: {path_model}") ValueError: Failed to load model from file: /home/user/.cache/huggingface/hub/models--ggml-org--gemma-3-1b-it-GGUF/snapshots/f9c28bcd85737ffc5aef028638d3341d49869c27/./gemma-3-1b-it-Q8_0.gguf

Container logs:

Fetching error logs...