Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -21,7 +21,8 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 21 |
taef1 = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=dtype).to(device)
|
| 22 |
good_vae = AutoencoderKL.from_pretrained("black-forest-labs/FLUX.1-dev", subfolder="vae", torch_dtype=dtype).to(device)
|
| 23 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", torch_dtype=dtype, vae=taef1).to(device)
|
| 24 |
-
|
|
|
|
| 25 |
pipe.fuse_lora()
|
| 26 |
torch.cuda.empty_cache()
|
| 27 |
|
|
|
|
| 21 |
taef1 = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=dtype).to(device)
|
| 22 |
good_vae = AutoencoderKL.from_pretrained("black-forest-labs/FLUX.1-dev", subfolder="vae", torch_dtype=dtype).to(device)
|
| 23 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", torch_dtype=dtype, vae=taef1).to(device)
|
| 24 |
+
# srpo_128_base_oficial_model_fp16.safetensors
|
| 25 |
+
pipe.load_lora_weights('Alissonerdx/flux.1-dev-SRPO-LoRas', weight_name='srpo_16_base_oficial_model_fp16.safetensors')
|
| 26 |
pipe.fuse_lora()
|
| 27 |
torch.cuda.empty_cache()
|
| 28 |
|