gaparmar commited on
Commit
4248464
·
1 Parent(s): f197aad

use 4 bit if possible

Browse files
Files changed (1) hide show
  1. app.py +8 -3
app.py CHANGED
@@ -16,9 +16,14 @@ from my_utils.default_values import apply_defaults
16
 
17
  import argparse
18
 
19
- precision = get_precision()
20
- transformer = NunchakuFluxTransformer2dModel.from_pretrained(f"nunchaku-tech/nunchaku-flux.1-schnell/svdq-{precision}_r32-flux.1-schnell.safetensors")
21
- pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", transformer=transformer, torch_dtype=torch.bfloat16).to("cuda")
 
 
 
 
 
22
  pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.bfloat16).to("cuda")
23
 
24
  m_clip = CLIPModel.from_pretrained("multimodalart/clip-vit-base-patch32").to("cuda")
 
16
 
17
  import argparse
18
 
19
+ try:
20
+ precision = get_precision()
21
+ transformer = NunchakuFluxTransformer2dModel.from_pretrained(f"nunchaku-tech/nunchaku-flux.1-schnell/svdq-{precision}_r32-flux.1-schnell.safetensors")
22
+ pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", transformer=transformer, torch_dtype=torch.bfloat16).to("cuda")
23
+ except Exception as e:
24
+ print(f"Error loading quantized FluxPipeline: {e}")
25
+ pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", torch_dtype=torch.bfloat16).to("cuda")
26
+
27
  pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.bfloat16).to("cuda")
28
 
29
  m_clip = CLIPModel.from_pretrained("multimodalart/clip-vit-base-patch32").to("cuda")