gaparmar
commited on
Commit
·
4248464
1
Parent(s):
f197aad
use 4 bit if possible
Browse files
app.py
CHANGED
|
@@ -16,9 +16,14 @@ from my_utils.default_values import apply_defaults
|
|
| 16 |
|
| 17 |
import argparse
|
| 18 |
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.bfloat16).to("cuda")
|
| 23 |
|
| 24 |
m_clip = CLIPModel.from_pretrained("multimodalart/clip-vit-base-patch32").to("cuda")
|
|
|
|
| 16 |
|
| 17 |
import argparse
|
| 18 |
|
| 19 |
+
try:
|
| 20 |
+
precision = get_precision()
|
| 21 |
+
transformer = NunchakuFluxTransformer2dModel.from_pretrained(f"nunchaku-tech/nunchaku-flux.1-schnell/svdq-{precision}_r32-flux.1-schnell.safetensors")
|
| 22 |
+
pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", transformer=transformer, torch_dtype=torch.bfloat16).to("cuda")
|
| 23 |
+
except Exception as e:
|
| 24 |
+
print(f"Error loading quantized FluxPipeline: {e}")
|
| 25 |
+
pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", torch_dtype=torch.bfloat16).to("cuda")
|
| 26 |
+
|
| 27 |
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.bfloat16).to("cuda")
|
| 28 |
|
| 29 |
m_clip = CLIPModel.from_pretrained("multimodalart/clip-vit-base-patch32").to("cuda")
|