VLLM 0.7.2 can start the model normally, but there is no output when simulating a request using Curl, it blocks!
1
#2 opened 7 days ago
by
JZMALi
sglang inference issue
7
#1 opened 10 days ago
by
su400