Getting error when deploying on HF Inference with 2 A100 GPU's of AWS on region us-east
Endpoint encountered an error.
You can try restarting it using the "retry" button above. Check logs for more details.
[Server message]Endpoint failed to start
Exit code: 1. Reason: eturn _run_code(code, main_globals, None,
File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
exec(code, run_globals)
File "/sgl-workspace/sglang/python/sglang/launch_server.py", line 11, in
server_args = prepare_server_args(sys.argv[1:])
File "/sgl-workspace/sglang/python/sglang/srt/server_args.py", line 2003, in prepare_server_args
server_args = ServerArgs.from_cli_args(raw_args)
File "/sgl-workspace/sglang/python/sglang/srt/server_args.py", line 1815, in from_cli_args
return cls(**{attr: getattr(args, attr) for attr in attrs})
File "", line 183, in init
File "/sgl-workspace/sglang/python/sglang/srt/server_args.py", line 333, in post_init
model_config = ModelConfig.from_server_args(self)
File "/sgl-workspace/sglang/python/sglang/srt/configs/model_config.py", line 284, in from_server_args
return ModelConfig(
File "/sgl-workspace/sglang/python/sglang/srt/configs/model_config.py", line 79, in init
self.hf_config = get_config(
File "/sgl-workspace/sglang/python/sglang/srt/utils.py", line 2745, in wrapper
result = func(*args, **kwargs)
File "/sgl-workspace/sglang/python/sglang/srt/hf_transformers_utils.py", line 123, in get_config
config = AutoConfig.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 1267, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type seed_oss
but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
You can update Transformers with the command pip install --upgrade transformers
. If this does not work, and the checkpoint is very new, then there may not be a release version that supports this model yet. In this case, you can get the most up-to-date code by installing Transformers from source with the command pip install git+https://github.com/huggingface/transformers.git