Spaces:
Running
on
A100
Running
on
A100
Commit
·
d67e4c8
1
Parent(s):
bc207eb
let's have some fun
Browse files
app.py
CHANGED
@@ -185,10 +185,13 @@ logger.info("All required model files are ready")
|
|
185 |
|
186 |
args = create_args()
|
187 |
logger.info(f"Created args, val_disable_autocast: {hasattr(args, 'val_disable_autocast')} = {getattr(args, 'val_disable_autocast', 'NOT SET')}")
|
|
|
|
|
|
|
188 |
hunyuan_video_sampler = HunyuanVideoSampler.from_pretrained(
|
189 |
args.ckpt,
|
190 |
args=args,
|
191 |
-
device=
|
192 |
)
|
193 |
logger.info(f"After from_pretrained, sampler.args has val_disable_autocast: {hasattr(hunyuan_video_sampler.args, 'val_disable_autocast')} = {getattr(hunyuan_video_sampler.args, 'val_disable_autocast', 'NOT SET')}")
|
194 |
args = hunyuan_video_sampler.args
|
@@ -204,6 +207,15 @@ if args.cpu_offload:
|
|
204 |
num_blocks_per_group=1
|
205 |
)
|
206 |
logger.info("Enabled CPU offloading for transformer blocks")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
207 |
|
208 |
logger.info("Model loaded successfully!")
|
209 |
|
|
|
185 |
|
186 |
args = create_args()
|
187 |
logger.info(f"Created args, val_disable_autocast: {hasattr(args, 'val_disable_autocast')} = {getattr(args, 'val_disable_autocast', 'NOT SET')}")
|
188 |
+
# Load model to CPU if offloading is enabled, otherwise load to GPU
|
189 |
+
model_device = torch.device("cpu") if args.cpu_offload else torch.device("cuda")
|
190 |
+
logger.info(f"Loading model to device: {model_device}")
|
191 |
hunyuan_video_sampler = HunyuanVideoSampler.from_pretrained(
|
192 |
args.ckpt,
|
193 |
args=args,
|
194 |
+
device=model_device
|
195 |
)
|
196 |
logger.info(f"After from_pretrained, sampler.args has val_disable_autocast: {hasattr(hunyuan_video_sampler.args, 'val_disable_autocast')} = {getattr(hunyuan_video_sampler.args, 'val_disable_autocast', 'NOT SET')}")
|
197 |
args = hunyuan_video_sampler.args
|
|
|
207 |
num_blocks_per_group=1
|
208 |
)
|
209 |
logger.info("Enabled CPU offloading for transformer blocks")
|
210 |
+
else:
|
211 |
+
# Ensure all model components are on GPU when not using CPU offload
|
212 |
+
hunyuan_video_sampler.pipeline.transformer.to('cuda')
|
213 |
+
hunyuan_video_sampler.vae.to('cuda')
|
214 |
+
if hunyuan_video_sampler.text_encoder:
|
215 |
+
hunyuan_video_sampler.text_encoder.model.to('cuda')
|
216 |
+
if hunyuan_video_sampler.text_encoder_2:
|
217 |
+
hunyuan_video_sampler.text_encoder_2.model.to('cuda')
|
218 |
+
logger.info("Model components moved to GPU")
|
219 |
|
220 |
logger.info("Model loaded successfully!")
|
221 |
|