multimodalart HF Staff commited on
Commit
eb6b15e
·
verified ·
1 Parent(s): 380ee19
Files changed (1) hide show
  1. app.py +6 -5
app.py CHANGED
@@ -24,8 +24,8 @@ def hf_hub_download_local(repo_id, filename, local_dir, **kwargs):
24
  print("Downloading models from Hugging Face Hub...")
25
  text_encoder_repo = hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors", local_dir="models/text_encoders")
26
  print(text_encoder_repo)
27
- hf_hub_download_local(repo_id="Comfy-Org/Wan_2.2_ComfyUI_Repackaged", filename="split_files/diffusion_models/wan2.2_i2v_low_noise_14B_fp8_scaled.safetensors", local_dir="models/unet")
28
- hf_hub_download_local(repo_id="Comfy-Org/Wan_2.2_ComfyUI_Repackaged", filename="split_files/diffusion_models/wan2.2_i2v_high_noise_14B_fp8_scaled.safetensors", local_dir="models/unet")
29
  hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/vae/wan_2.1_vae.safetensors", local_dir="models/vae")
30
  hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/clip_vision/clip_vision_h.safetensors", local_dir="models/clip_vision")
31
  hf_hub_download_local(repo_id="Kijai/WanVideo_comfy", filename="Wan22-Lightning/Wan2.2-Lightning_I2V-A14B-4steps-lora_HIGH_fp16.safetensors", local_dir="models/loras")
@@ -78,8 +78,8 @@ createvideo = nodes.NODE_CLASS_MAPPINGS["CreateVideo"]()
78
  savevideo = nodes.NODE_CLASS_MAPPINGS["SaveVideo"]()
79
 
80
  cliploader_38 = cliploader.load_clip(clip_name="umt5_xxl_fp8_e4m3fn_scaled.safetensors", type="wan", device="cpu")
81
- unetloader_37_low_noise = unetloader.load_unet(unet_name="wan2.2_i2v_low_noise_14B_fp8_scaled.safetensors", weight_dtype="default")
82
- unetloader_91_high_noise = unetloader.load_unet(unet_name="wan2.2_i2v_high_noise_14B_fp8_scaled.safetensors", weight_dtype="default")
83
  vaeloader_39 = vaeloader.load_vae(vae_name="wan_2.1_vae.safetensors")
84
  clipvisionloader_49 = clipvisionloader.load_clip(clip_name="clip_vision_h.safetensors")
85
 
@@ -92,6 +92,7 @@ pathchsageattentionkj_96_high = pathchsageattentionkj.patch(sage_attention="auto
92
 
93
  model_loaders = [cliploader_38, unetloader_37_low_noise, unetloader_91_high_noise, vaeloader_39, clipvisionloader_49, loraloadermodelonly_94_high, loraloadermodelonly_95_low]
94
  valid_models = [getattr(loader[0], 'patcher', loader[0]) for loader in model_loaders if not isinstance(loader[0], dict) and not isinstance(getattr(loader[0], 'patcher', None), dict)]
 
95
 
96
  # --- App Logic ---
97
  def calculate_dimensions(width, height):
@@ -107,7 +108,7 @@ def generate_video(prompt, first_image_path, last_image_path, duration_seconds):
107
  os.makedirs(temp_dir, exist_ok=True)
108
 
109
  with torch.inference_mode():
110
- model_management.load_models_gpu(valid_models)
111
  # --- Python Image Preprocessing using Pillow ---
112
  print("Preprocessing images with Pillow...")
113
  with Image.open(first_image_path) as img:
 
24
  print("Downloading models from Hugging Face Hub...")
25
  text_encoder_repo = hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors", local_dir="models/text_encoders")
26
  print(text_encoder_repo)
27
+ hf_hub_download_local(repo_id="Comfy-Org/Wan_2.2_ComfyUI_Repackaged", filename="split_files/diffusion_models/wan2.2_i2v_low_noise_14B_fp16.safetensors", local_dir="models/unet")
28
+ hf_hub_download_local(repo_id="Comfy-Org/Wan_2.2_ComfyUI_Repackaged", filename="split_files/diffusion_models/wan2.2_i2v_high_noise_14B_fp16.safetensors", local_dir="models/unet")
29
  hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/vae/wan_2.1_vae.safetensors", local_dir="models/vae")
30
  hf_hub_download_local(repo_id="Comfy-Org/Wan_2.1_ComfyUI_repackaged", filename="split_files/clip_vision/clip_vision_h.safetensors", local_dir="models/clip_vision")
31
  hf_hub_download_local(repo_id="Kijai/WanVideo_comfy", filename="Wan22-Lightning/Wan2.2-Lightning_I2V-A14B-4steps-lora_HIGH_fp16.safetensors", local_dir="models/loras")
 
78
  savevideo = nodes.NODE_CLASS_MAPPINGS["SaveVideo"]()
79
 
80
  cliploader_38 = cliploader.load_clip(clip_name="umt5_xxl_fp8_e4m3fn_scaled.safetensors", type="wan", device="cpu")
81
+ unetloader_37_low_noise = unetloader.load_unet(unet_name="wan2.2_i2v_low_noise_14B_fp16.safetensors", weight_dtype="default")
82
+ unetloader_91_high_noise = unetloader.load_unet(unet_name="wan2.2_i2v_high_noise_14B_fp16.safetensors", weight_dtype="default")
83
  vaeloader_39 = vaeloader.load_vae(vae_name="wan_2.1_vae.safetensors")
84
  clipvisionloader_49 = clipvisionloader.load_clip(clip_name="clip_vision_h.safetensors")
85
 
 
92
 
93
  model_loaders = [cliploader_38, unetloader_37_low_noise, unetloader_91_high_noise, vaeloader_39, clipvisionloader_49, loraloadermodelonly_94_high, loraloadermodelonly_95_low]
94
  valid_models = [getattr(loader[0], 'patcher', loader[0]) for loader in model_loaders if not isinstance(loader[0], dict) and not isinstance(getattr(loader[0], 'patcher', None), dict)]
95
+ model_management.load_models_gpu(valid_models)
96
 
97
  # --- App Logic ---
98
  def calculate_dimensions(width, height):
 
108
  os.makedirs(temp_dir, exist_ok=True)
109
 
110
  with torch.inference_mode():
111
+
112
  # --- Python Image Preprocessing using Pillow ---
113
  print("Preprocessing images with Pillow...")
114
  with Image.open(first_image_path) as img: