Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -41,7 +41,7 @@ torch.cuda.empty_cache()
|
|
41 |
MAX_SEED = np.iinfo(np.int32).max
|
42 |
MAX_IMAGE_SIZE = 2048
|
43 |
|
44 |
-
pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
|
45 |
|
46 |
@spaces.GPU
|
47 |
def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=3.5, num_inference_steps=28, progress=gr.Progress(track_tqdm=True)):
|
@@ -49,7 +49,7 @@ def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, g
|
|
49 |
seed = random.randint(0, MAX_SEED)
|
50 |
generator = torch.Generator().manual_seed(seed)
|
51 |
|
52 |
-
|
53 |
prompt=prompt,
|
54 |
guidance_scale=guidance_scale,
|
55 |
num_inference_steps=num_inference_steps,
|
@@ -58,8 +58,10 @@ def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, g
|
|
58 |
generator=generator,
|
59 |
output_type="pil",
|
60 |
good_vae=good_vae,
|
61 |
-
)
|
62 |
-
|
|
|
|
|
63 |
|
64 |
@spaces.GPU
|
65 |
def llama_gen_fragrance(scene):
|
|
|
41 |
MAX_SEED = np.iinfo(np.int32).max
|
42 |
MAX_IMAGE_SIZE = 2048
|
43 |
|
44 |
+
#pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
|
45 |
|
46 |
@spaces.GPU
|
47 |
def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=3.5, num_inference_steps=28, progress=gr.Progress(track_tqdm=True)):
|
|
|
49 |
seed = random.randint(0, MAX_SEED)
|
50 |
generator = torch.Generator().manual_seed(seed)
|
51 |
|
52 |
+
img = pipe(
|
53 |
prompt=prompt,
|
54 |
guidance_scale=guidance_scale,
|
55 |
num_inference_steps=num_inference_steps,
|
|
|
58 |
generator=generator,
|
59 |
output_type="pil",
|
60 |
good_vae=good_vae,
|
61 |
+
).images[0]
|
62 |
+
|
63 |
+
img.save("flux-dev.png")
|
64 |
+
return "flux-dev.png"
|
65 |
|
66 |
@spaces.GPU
|
67 |
def llama_gen_fragrance(scene):
|