Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -21,7 +21,7 @@ if not torch.cuda.is_available():
|
|
| 21 |
MAX_SEED = np.iinfo(np.int32).max
|
| 22 |
CACHE_EXAMPLES = torch.cuda.is_available() and os.getenv("CACHE_EXAMPLES") == "1"
|
| 23 |
MAX_IMAGE_SIZE = int(os.getenv("MAX_IMAGE_SIZE", "1536"))
|
| 24 |
-
USE_TORCH_COMPILE =
|
| 25 |
ENABLE_CPU_OFFLOAD = os.getenv("ENABLE_CPU_OFFLOAD") == "1"
|
| 26 |
|
| 27 |
dtype = torch.float16
|
|
@@ -203,10 +203,10 @@ with gr.Blocks() as demo:
|
|
| 203 |
)
|
| 204 |
prior_num_inference_steps = gr.Slider(
|
| 205 |
label="Prior Inference Steps",
|
| 206 |
-
minimum=
|
| 207 |
maximum=30,
|
| 208 |
step=1,
|
| 209 |
-
value=
|
| 210 |
)
|
| 211 |
|
| 212 |
decoder_guidance_scale = gr.Slider(
|
|
@@ -221,7 +221,7 @@ with gr.Blocks() as demo:
|
|
| 221 |
minimum=4,
|
| 222 |
maximum=12,
|
| 223 |
step=1,
|
| 224 |
-
value=
|
| 225 |
)
|
| 226 |
|
| 227 |
gr.Examples(
|
|
|
|
| 21 |
MAX_SEED = np.iinfo(np.int32).max
|
| 22 |
CACHE_EXAMPLES = torch.cuda.is_available() and os.getenv("CACHE_EXAMPLES") == "1"
|
| 23 |
MAX_IMAGE_SIZE = int(os.getenv("MAX_IMAGE_SIZE", "1536"))
|
| 24 |
+
USE_TORCH_COMPILE = True
|
| 25 |
ENABLE_CPU_OFFLOAD = os.getenv("ENABLE_CPU_OFFLOAD") == "1"
|
| 26 |
|
| 27 |
dtype = torch.float16
|
|
|
|
| 203 |
)
|
| 204 |
prior_num_inference_steps = gr.Slider(
|
| 205 |
label="Prior Inference Steps",
|
| 206 |
+
minimum=10,
|
| 207 |
maximum=30,
|
| 208 |
step=1,
|
| 209 |
+
value=25,
|
| 210 |
)
|
| 211 |
|
| 212 |
decoder_guidance_scale = gr.Slider(
|
|
|
|
| 221 |
minimum=4,
|
| 222 |
maximum=12,
|
| 223 |
step=1,
|
| 224 |
+
value=10,
|
| 225 |
)
|
| 226 |
|
| 227 |
gr.Examples(
|