Spaces:
Running
on
L40S
Running
on
L40S
| # Open Source Model Licensed under the Apache License Version 2.0 and Other Licenses of the Third-Party Components therein: | |
| # The below Model in this distribution may have been modified by THL A29 Limited ("Tencent Modifications"). All Tencent Modifications are Copyright (C) 2024 THL A29 Limited. | |
| # Copyright (C) 2024 THL A29 Limited, a Tencent company. All rights reserved. | |
| # The below software and/or models in this distribution may have been | |
| # modified by THL A29 Limited ("Tencent Modifications"). | |
| # All Tencent Modifications are Copyright (C) THL A29 Limited. | |
| # Hunyuan 3D is licensed under the TENCENT HUNYUAN NON-COMMERCIAL LICENSE AGREEMENT | |
| # except for the third-party components listed below. | |
| # Hunyuan 3D does not impose any additional limitations beyond what is outlined | |
| # in the repsective licenses of these third-party components. | |
| # Users must comply with all terms and conditions of original licenses of these third-party | |
| # components and must ensure that the usage of the third party components adheres to | |
| # all relevant laws and regulations. | |
| # For avoidance of doubts, Hunyuan 3D means the large language models and | |
| # their software and algorithms, including trained model weights, parameters (including | |
| # optimizer states), machine-learning model code, inference-enabling code, training-enabling code, | |
| # fine-tuning enabling code and other elements of the foregoing made publicly available | |
| # by Tencent in accordance with TENCENT HUNYUAN COMMUNITY LICENSE AGREEMENT.l | |
| import os | |
| import warnings | |
| import torch | |
| from PIL import Image | |
| import argparse | |
| from infer import Text2Image, Removebg, Image2Views, Views2Mesh, GifRenderer | |
| warnings.simplefilter('ignore', category=UserWarning) | |
| warnings.simplefilter('ignore', category=FutureWarning) | |
| warnings.simplefilter('ignore', category=DeprecationWarning) | |
| def get_args(): | |
| parser = argparse.ArgumentParser() | |
| parser.add_argument( | |
| "--use_lite", default=False, action="store_true" | |
| ) | |
| parser.add_argument( | |
| "--mv23d_cfg_path", default="./svrm/configs/svrm.yaml", type=str | |
| ) | |
| parser.add_argument( | |
| "--mv23d_ckt_path", default="weights/svrm/svrm.safetensors", type=str | |
| ) | |
| parser.add_argument( | |
| "--text2image_path", default="weights/hunyuanDiT", type=str | |
| ) | |
| parser.add_argument( | |
| "--save_folder", default="./outputs/test/", type=str | |
| ) | |
| parser.add_argument( | |
| "--text_prompt", default="", type=str, | |
| ) | |
| parser.add_argument( | |
| "--image_prompt", default="", type=str | |
| ) | |
| parser.add_argument( | |
| "--device", default="cuda:0", type=str | |
| ) | |
| parser.add_argument( | |
| "--t2i_seed", default=0, type=int | |
| ) | |
| parser.add_argument( | |
| "--t2i_steps", default=25, type=int | |
| ) | |
| parser.add_argument( | |
| "--gen_seed", default=0, type=int | |
| ) | |
| parser.add_argument( | |
| "--gen_steps", default=50, type=int | |
| ) | |
| parser.add_argument( | |
| "--max_faces_num", default=80000, type=int, | |
| help="max num of face, suggest 80000 for effect, 10000 for speed" | |
| ) | |
| parser.add_argument( | |
| "--save_memory", default=False, action="store_true" | |
| ) | |
| parser.add_argument( | |
| "--do_texture_mapping", default=False, action="store_true" | |
| ) | |
| parser.add_argument( | |
| "--do_render", default=False, action="store_true" | |
| ) | |
| return parser.parse_args() | |
| if __name__ == "__main__": | |
| args = get_args() | |
| assert not (args.text_prompt and args.image_prompt), "Text and image can only be given to one" | |
| assert args.text_prompt or args.image_prompt, "Text and image can only be given to one" | |
| # init model | |
| rembg_model = Removebg() | |
| image_to_views_model = Image2Views( | |
| device=args.device, | |
| use_lite=args.use_lite, | |
| save_memory=args.save_memory | |
| ) | |
| views_to_mesh_model = Views2Mesh( | |
| args.mv23d_cfg_path, | |
| args.mv23d_ckt_path, | |
| args.device, | |
| use_lite=args.use_lite, | |
| save_memory=args.save_memory | |
| ) | |
| if args.text_prompt: | |
| text_to_image_model = Text2Image( | |
| pretrain = args.text2image_path, | |
| device = args.device, | |
| save_memory = args.save_memory | |
| ) | |
| if args.do_render: | |
| gif_renderer = GifRenderer(device=args.device) | |
| # ---- ----- ---- ---- ---- ---- | |
| os.makedirs(args.save_folder, exist_ok=True) | |
| # stage 1, text to image | |
| if args.text_prompt: | |
| res_rgb_pil = text_to_image_model( | |
| args.text_prompt, | |
| seed=args.t2i_seed, | |
| steps=args.t2i_steps | |
| ) | |
| res_rgb_pil.save(os.path.join(args.save_folder, "img.jpg")) | |
| elif args.image_prompt: | |
| res_rgb_pil = Image.open(args.image_prompt) | |
| # stage 2, remove back ground | |
| res_rgba_pil = rembg_model(res_rgb_pil) | |
| res_rgb_pil.save(os.path.join(args.save_folder, "img_nobg.png")) | |
| # stage 3, image to views | |
| (views_grid_pil, cond_img), view_pil_list = image_to_views_model( | |
| res_rgba_pil, | |
| seed = args.gen_seed, | |
| steps = args.gen_steps | |
| ) | |
| views_grid_pil.save(os.path.join(args.save_folder, "views.jpg")) | |
| # stage 4, views to mesh | |
| views_to_mesh_model( | |
| views_grid_pil, | |
| cond_img, | |
| seed = args.gen_seed, | |
| target_face_count = args.max_faces_num, | |
| save_folder = args.save_folder, | |
| do_texture_mapping = args.do_texture_mapping | |
| ) | |
| # stage 5, render gif | |
| if args.do_render: | |
| gif_renderer( | |
| os.path.join(args.save_folder, 'mesh.obj'), | |
| gif_dst_path = os.path.join(args.save_folder, 'output.gif'), | |
| ) | |