Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -32,18 +32,6 @@ MAX_IMAGE_SIZE = 2048
|
|
| 32 |
dtype = torch.bfloat16
|
| 33 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 34 |
|
| 35 |
-
# ---- CUDA Check ----
|
| 36 |
-
print("CUDA_VISIBLE_DEVICES=", os.environ.get("CUDA_VISIBLE_DEVICES"))
|
| 37 |
-
print("torch.__version__ =", torch.__version__)
|
| 38 |
-
print("torch.version.cuda =", torch.version.cuda)
|
| 39 |
-
print("cuda available:", torch.cuda.is_available())
|
| 40 |
-
print("cuda device count:", torch.cuda.device_count())
|
| 41 |
-
if torch.cuda.is_available():
|
| 42 |
-
print("current device:", torch.cuda.current_device())
|
| 43 |
-
print("device name:", torch.cuda.get_device_name(torch.cuda.current_device()))
|
| 44 |
-
|
| 45 |
-
print("Using device:", device)
|
| 46 |
-
|
| 47 |
# --- Model Loading ---
|
| 48 |
pipe_qwen = DiffusionPipeline.from_pretrained("Qwen/Qwen-Image", torch_dtype=dtype).to(device)
|
| 49 |
|
|
@@ -290,4 +278,4 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
| 290 |
)
|
| 291 |
|
| 292 |
if __name__ == "__main__":
|
| 293 |
-
demo.queue(max_size=50).launch(share=
|
|
|
|
| 32 |
dtype = torch.bfloat16
|
| 33 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
# --- Model Loading ---
|
| 36 |
pipe_qwen = DiffusionPipeline.from_pretrained("Qwen/Qwen-Image", torch_dtype=dtype).to(device)
|
| 37 |
|
|
|
|
| 278 |
)
|
| 279 |
|
| 280 |
if __name__ == "__main__":
|
| 281 |
+
demo.queue(max_size=50).launch(share=False, mcp_server=True, ssr_mode=False, show_error=True)
|