hzhwcmhf commited on
Commit
61082d4
·
verified ·
1 Parent(s): 3ffd1f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -220,9 +220,9 @@ After updating the config, proceed with either **vLLM** or **SGLang** for servin
220
  To run Qwen with 1M context support:
221
 
222
  ```bash
223
- git clone https://github.com/vllm-project/vllm.git
224
- cd vllm
225
- pip install -e .
226
  ```
227
 
228
  Then launch the server with Dual Chunk Flash Attention enabled:
 
220
  To run Qwen with 1M context support:
221
 
222
  ```bash
223
+ pip install -U vllm \
224
+ --torch-backend=auto \
225
+ --extra-index-url https://wheels.vllm.ai/nightly
226
  ```
227
 
228
  Then launch the server with Dual Chunk Flash Attention enabled: