Spaces:
Running
on
Zero
Running
on
Zero
Upload step03_chatbot.py with huggingface_hub
Browse files- step03_chatbot.py +6 -1
step03_chatbot.py
CHANGED
|
@@ -8,6 +8,9 @@ import os
|
|
| 8 |
import json
|
| 9 |
import numpy as np
|
| 10 |
import gradio as gr
|
|
|
|
|
|
|
|
|
|
| 11 |
from gradio import ChatMessage
|
| 12 |
from typing import List, Dict, Optional, Tuple
|
| 13 |
import time
|
|
@@ -255,13 +258,15 @@ class Qwen3Reranker:
|
|
| 255 |
if self.is_mps:
|
| 256 |
self.device = torch.device("mps")
|
| 257 |
self.model = self.model.to(self.device)
|
| 258 |
-
elif self.is_cuda:
|
|
|
|
| 259 |
if hasattr(self.model, 'device'):
|
| 260 |
self.device = next(self.model.parameters()).device
|
| 261 |
else:
|
| 262 |
self.device = torch.device("cuda")
|
| 263 |
self.model = self.model.to(self.device)
|
| 264 |
else:
|
|
|
|
| 265 |
self.device = torch.device("cpu")
|
| 266 |
self.model = self.model.to(self.device)
|
| 267 |
|
|
|
|
| 8 |
import json
|
| 9 |
import numpy as np
|
| 10 |
import gradio as gr
|
| 11 |
+
|
| 12 |
+
# Désactiver le warning tokenizers sur ZeroGPU
|
| 13 |
+
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
| 14 |
from gradio import ChatMessage
|
| 15 |
from typing import List, Dict, Optional, Tuple
|
| 16 |
import time
|
|
|
|
| 258 |
if self.is_mps:
|
| 259 |
self.device = torch.device("mps")
|
| 260 |
self.model = self.model.to(self.device)
|
| 261 |
+
elif self.is_cuda and not os.getenv("SPACE_ID"):
|
| 262 |
+
# Utiliser CUDA seulement si pas sur ZeroGPU Spaces
|
| 263 |
if hasattr(self.model, 'device'):
|
| 264 |
self.device = next(self.model.parameters()).device
|
| 265 |
else:
|
| 266 |
self.device = torch.device("cuda")
|
| 267 |
self.model = self.model.to(self.device)
|
| 268 |
else:
|
| 269 |
+
# Forcer CPU sur ZeroGPU pour éviter l'erreur CUDA init
|
| 270 |
self.device = torch.device("cpu")
|
| 271 |
self.model = self.model.to(self.device)
|
| 272 |
|