fffiloni commited on
Commit
fe118aa
·
verified ·
1 Parent(s): 535be05

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -4
app.py CHANGED
@@ -41,7 +41,7 @@ torch.cuda.empty_cache()
41
  MAX_SEED = np.iinfo(np.int32).max
42
  MAX_IMAGE_SIZE = 2048
43
 
44
- pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
45
 
46
  @spaces.GPU
47
  def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=3.5, num_inference_steps=28, progress=gr.Progress(track_tqdm=True)):
@@ -49,7 +49,7 @@ def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, g
49
  seed = random.randint(0, MAX_SEED)
50
  generator = torch.Generator().manual_seed(seed)
51
 
52
- for img in pipe.flux_pipe_call_that_returns_an_iterable_of_images(
53
  prompt=prompt,
54
  guidance_scale=guidance_scale,
55
  num_inference_steps=num_inference_steps,
@@ -58,8 +58,10 @@ def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, g
58
  generator=generator,
59
  output_type="pil",
60
  good_vae=good_vae,
61
- ):
62
- yield img
 
 
63
 
64
  @spaces.GPU
65
  def llama_gen_fragrance(scene):
 
41
  MAX_SEED = np.iinfo(np.int32).max
42
  MAX_IMAGE_SIZE = 2048
43
 
44
+ #pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
45
 
46
  @spaces.GPU
47
  def infer_flux(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=3.5, num_inference_steps=28, progress=gr.Progress(track_tqdm=True)):
 
49
  seed = random.randint(0, MAX_SEED)
50
  generator = torch.Generator().manual_seed(seed)
51
 
52
+ img = pipe(
53
  prompt=prompt,
54
  guidance_scale=guidance_scale,
55
  num_inference_steps=num_inference_steps,
 
58
  generator=generator,
59
  output_type="pil",
60
  good_vae=good_vae,
61
+ ).images[0]
62
+
63
+ img.save("flux-dev.png")
64
+ return "flux-dev.png"
65
 
66
  @spaces.GPU
67
  def llama_gen_fragrance(scene):