Spaces:
Sleeping
Sleeping
| services: | |
| backend: | |
| container_name: mindsearch-backend | |
| build: | |
| context: . | |
| dockerfile: backend.dockerfile | |
| image: mindsearch/backend:latest | |
| restart: unless-stopped | |
| # Uncomment the following line to force using local build | |
| # pull: never | |
| ports: | |
| - "8002:8002" | |
| environment: | |
| - PYTHONUNBUFFERED=1 | |
| # - OPENAI_API_KEY=${OPENAI_API_KEY:-} | |
| - OPENAI_API_BASE=${OPENAI_API_BASE:-https://api.openai.com/v1} | |
| # - QWEN_API_KEY=${QWEN_API_KEY:-} | |
| # - SILICON_API_KEY=${SILICON_API_KEY:-} | |
| command: python -m mindsearch.app --lang ${LANG:-cn} --model_format ${MODEL_FORMAT:-internlm_server} | |
| volumes: | |
| - /root/.cache:/root/.cache | |
| deploy: | |
| resources: | |
| reservations: | |
| devices: | |
| - driver: nvidia | |
| count: 1 | |
| capabilities: [gpu] | |
| # GPU support explanation: | |
| # The current configuration has been tested with NVIDIA GPUs. If you use other types of GPUs, you may need to adjust the configuration. | |
| # For AMD GPUs, you can try using the ROCm driver by modifying the configuration as follows: | |
| # deploy: | |
| # resources: | |
| # reservations: | |
| # devices: | |
| # - driver: amd | |
| # count: 1 | |
| # capabilities: [gpu] | |
| # | |
| # For other GPU types, you may need to consult the respective Docker GPU support documentation. | |
| # In theory, any GPU supported by PyTorch should be configurable here. | |
| # If you encounter issues, try the following steps: | |
| # 1. Ensure the correct GPU drivers are installed on the host | |
| # 2. Check if your Docker version supports your GPU type | |
| # 3. Install necessary GPU-related libraries in the Dockerfile | |
| # 4. Adjust the deploy configuration here to match your GPU type | |
| # | |
| # Note: After changing GPU configuration, you may need to rebuild the image. | |
| frontend: | |
| container_name: mindsearch-frontend | |
| build: | |
| context: . | |
| dockerfile: frontend.dockerfile | |
| image: mindsearch/frontend:latest | |
| restart: unless-stopped | |
| # Uncomment the following line to force using local build | |
| # pull: never | |
| ports: | |
| - "8080:8080" | |
| depends_on: | |
| - backend | |