linoyts HF Staff commited on
Commit
5c99054
·
verified ·
1 Parent(s): cea20cd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -64,14 +64,14 @@ pipe.load_lora_weights(
64
  )
65
  kwargs = {}
66
  kwargs["load_into_transformer_2"] = True
67
- pipe.load_lora_weights(
68
- "vrgamedevgirl84/Wan14BT2VFusioniX",
69
- weight_name="FusionX_LoRa/Phantom_Wan_14B_FusionX_LoRA.safetensors",
70
- adapter_name="phantom_2", **kwargs
71
- )
72
- pipe.set_adapters(["phantom", "phantom_2"], adapter_weights=[1., 1.])
73
  pipe.fuse_lora(adapter_names=["phantom"], lora_scale=3., components=["transformer"])
74
- pipe.fuse_lora(adapter_names=["phantom_2"], lora_scale=1., components=["transformer_2"])
75
  pipe.unload_lora_weights()
76
 
77
  optimize_pipeline_(pipe,
 
64
  )
65
  kwargs = {}
66
  kwargs["load_into_transformer_2"] = True
67
+ # pipe.load_lora_weights(
68
+ # "vrgamedevgirl84/Wan14BT2VFusioniX",
69
+ # weight_name="FusionX_LoRa/Phantom_Wan_14B_FusionX_LoRA.safetensors",
70
+ # adapter_name="phantom_2", **kwargs
71
+ # )
72
+ # pipe.set_adapters(["phantom", "phantom_2"], adapter_weights=[1., 1.])
73
  pipe.fuse_lora(adapter_names=["phantom"], lora_scale=3., components=["transformer"])
74
+ # pipe.fuse_lora(adapter_names=["phantom_2"], lora_scale=1., components=["transformer_2"])
75
  pipe.unload_lora_weights()
76
 
77
  optimize_pipeline_(pipe,