Update app.py
Browse files
app.py
CHANGED
@@ -7,18 +7,10 @@ import time
|
|
7 |
from fastapi.responses import HTMLResponse
|
8 |
from fastapi.staticfiles import StaticFiles
|
9 |
|
10 |
-
print(f"Is CUDA available: {torch.cuda.is_available()}")
|
11 |
-
# True
|
12 |
-
print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
|
13 |
-
|
14 |
-
model = load_pytorch_model()
|
15 |
-
model = model.to("cuda")
|
16 |
-
|
17 |
# Define FastAPI app
|
18 |
app = FastAPI()
|
19 |
|
20 |
-
|
21 |
-
device = torch.device("cuda")
|
22 |
asr_pipeline = pipeline(model="openai/whisper-large", device=device) # Initialize Whisper model
|
23 |
# asr_pipeline = pipeline( model="openai/whisper-small", device=device, language="pt")
|
24 |
|
|
|
7 |
from fastapi.responses import HTMLResponse
|
8 |
from fastapi.staticfiles import StaticFiles
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
# Define FastAPI app
|
11 |
app = FastAPI()
|
12 |
|
13 |
+
device = 0 if torch.cuda.is_available() else -1
|
|
|
14 |
asr_pipeline = pipeline(model="openai/whisper-large", device=device) # Initialize Whisper model
|
15 |
# asr_pipeline = pipeline( model="openai/whisper-small", device=device, language="pt")
|
16 |
|