VincentGOURBIN commited on
Commit
72e5939
·
verified ·
1 Parent(s): 4448628

Upload step03_chatbot.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. step03_chatbot.py +6 -1
step03_chatbot.py CHANGED
@@ -8,6 +8,9 @@ import os
8
  import json
9
  import numpy as np
10
  import gradio as gr
 
 
 
11
  from gradio import ChatMessage
12
  from typing import List, Dict, Optional, Tuple
13
  import time
@@ -255,13 +258,15 @@ class Qwen3Reranker:
255
  if self.is_mps:
256
  self.device = torch.device("mps")
257
  self.model = self.model.to(self.device)
258
- elif self.is_cuda:
 
259
  if hasattr(self.model, 'device'):
260
  self.device = next(self.model.parameters()).device
261
  else:
262
  self.device = torch.device("cuda")
263
  self.model = self.model.to(self.device)
264
  else:
 
265
  self.device = torch.device("cpu")
266
  self.model = self.model.to(self.device)
267
 
 
8
  import json
9
  import numpy as np
10
  import gradio as gr
11
+
12
+ # Désactiver le warning tokenizers sur ZeroGPU
13
+ os.environ["TOKENIZERS_PARALLELISM"] = "false"
14
  from gradio import ChatMessage
15
  from typing import List, Dict, Optional, Tuple
16
  import time
 
258
  if self.is_mps:
259
  self.device = torch.device("mps")
260
  self.model = self.model.to(self.device)
261
+ elif self.is_cuda and not os.getenv("SPACE_ID"):
262
+ # Utiliser CUDA seulement si pas sur ZeroGPU Spaces
263
  if hasattr(self.model, 'device'):
264
  self.device = next(self.model.parameters()).device
265
  else:
266
  self.device = torch.device("cuda")
267
  self.model = self.model.to(self.device)
268
  else:
269
+ # Forcer CPU sur ZeroGPU pour éviter l'erreur CUDA init
270
  self.device = torch.device("cpu")
271
  self.model = self.model.to(self.device)
272