Spaces:
Running
Running
Commit
·
9aaf024
1
Parent(s):
8f28809
Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +23 -0
- .gitignore +174 -0
- .ipynb_checkpoints/Untitled-checkpoint.ipynb +6 -0
- LICENSE +21 -0
- README.md +268 -13
- Untitled.ipynb +533 -0
- app_sadtalker.py +111 -0
- checkpoints/SadTalker_V0.0.2_256.safetensors +3 -0
- checkpoints/SadTalker_V0.0.2_512.safetensors +3 -0
- checkpoints/mapping_00109-model.pth.tar +3 -0
- checkpoints/mapping_00229-model.pth.tar +3 -0
- cog.yaml +35 -0
- docs/FAQ.md +46 -0
- docs/best_practice.md +94 -0
- docs/changlelog.md +29 -0
- docs/example_crop.gif +3 -0
- docs/example_crop_still.gif +3 -0
- docs/example_full.gif +3 -0
- docs/example_full_crop.gif +0 -0
- docs/example_full_enhanced.gif +3 -0
- docs/face3d.md +48 -0
- docs/free_view_result.gif +3 -0
- docs/install.md +47 -0
- docs/resize_good.gif +3 -0
- docs/resize_no.gif +3 -0
- docs/sadtalker_logo.png +0 -0
- docs/using_ref_video.gif +3 -0
- docs/webui_extension.md +50 -0
- examples/driven_audio/RD_Radio31_000.wav +0 -0
- examples/driven_audio/RD_Radio34_002.wav +0 -0
- examples/driven_audio/RD_Radio36_000.wav +0 -0
- examples/driven_audio/RD_Radio40_000.wav +0 -0
- examples/driven_audio/bus_chinese.wav +0 -0
- examples/driven_audio/chinese_news.wav +3 -0
- examples/driven_audio/chinese_poem1.wav +0 -0
- examples/driven_audio/chinese_poem2.wav +0 -0
- examples/driven_audio/deyu.wav +3 -0
- examples/driven_audio/eluosi.wav +3 -0
- examples/driven_audio/fayu.wav +3 -0
- examples/driven_audio/imagine.wav +3 -0
- examples/driven_audio/itosinger1.wav +0 -0
- examples/driven_audio/japanese.wav +3 -0
- examples/ref_video/WDA_AlexandriaOcasioCortez_000.mp4 +3 -0
- examples/ref_video/WDA_KatieHill_000.mp4 +3 -0
- examples/source_image/art_0.png +0 -0
- examples/source_image/art_1.png +0 -0
- examples/source_image/art_10.png +0 -0
- examples/source_image/art_11.png +0 -0
- examples/source_image/art_12.png +0 -0
- examples/source_image/art_13.png +0 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,26 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
docs/example_crop.gif filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
docs/example_crop_still.gif filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
docs/example_full.gif filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
docs/example_full_enhanced.gif filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
docs/free_view_result.gif filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
docs/resize_good.gif filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
docs/resize_no.gif filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
docs/using_ref_video.gif filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
examples/driven_audio/chinese_news.wav filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
examples/driven_audio/deyu.wav filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
examples/driven_audio/eluosi.wav filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
examples/driven_audio/fayu.wav filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
examples/driven_audio/imagine.wav filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
examples/driven_audio/japanese.wav filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
examples/ref_video/WDA_AlexandriaOcasioCortez_000.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
examples/ref_video/WDA_KatieHill_000.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
examples/source_image/art_16.png filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
examples/source_image/art_17.png filter=lfs diff=lfs merge=lfs -text
|
| 54 |
+
examples/source_image/art_3.png filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
examples/source_image/art_4.png filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
examples/source_image/art_5.png filter=lfs diff=lfs merge=lfs -text
|
| 57 |
+
examples/source_image/art_8.png filter=lfs diff=lfs merge=lfs -text
|
| 58 |
+
examples/source_image/art_9.png filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
|
@@ -0,0 +1,174 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Byte-compiled / optimized / DLL files
|
| 2 |
+
__pycache__/
|
| 3 |
+
*.py[cod]
|
| 4 |
+
*$py.class
|
| 5 |
+
|
| 6 |
+
# C extensions
|
| 7 |
+
*.so
|
| 8 |
+
|
| 9 |
+
# Distribution / packaging
|
| 10 |
+
.Python
|
| 11 |
+
build/
|
| 12 |
+
develop-eggs/
|
| 13 |
+
dist/
|
| 14 |
+
downloads/
|
| 15 |
+
eggs/
|
| 16 |
+
.eggs/
|
| 17 |
+
lib/
|
| 18 |
+
lib64/
|
| 19 |
+
parts/
|
| 20 |
+
sdist/
|
| 21 |
+
var/
|
| 22 |
+
wheels/
|
| 23 |
+
share/python-wheels/
|
| 24 |
+
*.egg-info/
|
| 25 |
+
.installed.cfg
|
| 26 |
+
*.egg
|
| 27 |
+
MANIFEST
|
| 28 |
+
|
| 29 |
+
# PyInstaller
|
| 30 |
+
# Usually these files are written by a python script from a template
|
| 31 |
+
# before PyInstaller builds the exe, so as to inject date/other infos into it.
|
| 32 |
+
*.manifest
|
| 33 |
+
*.spec
|
| 34 |
+
|
| 35 |
+
# Installer logs
|
| 36 |
+
pip-log.txt
|
| 37 |
+
pip-delete-this-directory.txt
|
| 38 |
+
|
| 39 |
+
# Unit test / coverage reports
|
| 40 |
+
htmlcov/
|
| 41 |
+
.tox/
|
| 42 |
+
.nox/
|
| 43 |
+
.coverage
|
| 44 |
+
.coverage.*
|
| 45 |
+
.cache
|
| 46 |
+
nosetests.xml
|
| 47 |
+
coverage.xml
|
| 48 |
+
*.cover
|
| 49 |
+
*.py,cover
|
| 50 |
+
.hypothesis/
|
| 51 |
+
.pytest_cache/
|
| 52 |
+
cover/
|
| 53 |
+
|
| 54 |
+
# Translations
|
| 55 |
+
*.mo
|
| 56 |
+
*.pot
|
| 57 |
+
|
| 58 |
+
# Django stuff:
|
| 59 |
+
*.log
|
| 60 |
+
local_settings.py
|
| 61 |
+
db.sqlite3
|
| 62 |
+
db.sqlite3-journal
|
| 63 |
+
|
| 64 |
+
# Flask stuff:
|
| 65 |
+
instance/
|
| 66 |
+
.webassets-cache
|
| 67 |
+
|
| 68 |
+
# Scrapy stuff:
|
| 69 |
+
.scrapy
|
| 70 |
+
|
| 71 |
+
# Sphinx documentation
|
| 72 |
+
docs/_build/
|
| 73 |
+
|
| 74 |
+
# PyBuilder
|
| 75 |
+
.pybuilder/
|
| 76 |
+
target/
|
| 77 |
+
|
| 78 |
+
# Jupyter Notebook
|
| 79 |
+
.ipynb_checkpoints
|
| 80 |
+
|
| 81 |
+
# IPython
|
| 82 |
+
profile_default/
|
| 83 |
+
ipython_config.py
|
| 84 |
+
|
| 85 |
+
# pyenv
|
| 86 |
+
# For a library or package, you might want to ignore these files since the code is
|
| 87 |
+
# intended to run in multiple environments; otherwise, check them in:
|
| 88 |
+
# .python-version
|
| 89 |
+
|
| 90 |
+
# pipenv
|
| 91 |
+
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
|
| 92 |
+
# However, in case of collaboration, if having platform-specific dependencies or dependencies
|
| 93 |
+
# having no cross-platform support, pipenv may install dependencies that don't work, or not
|
| 94 |
+
# install all needed dependencies.
|
| 95 |
+
#Pipfile.lock
|
| 96 |
+
|
| 97 |
+
# poetry
|
| 98 |
+
# Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control.
|
| 99 |
+
# This is especially recommended for binary packages to ensure reproducibility, and is more
|
| 100 |
+
# commonly ignored for libraries.
|
| 101 |
+
# https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control
|
| 102 |
+
#poetry.lock
|
| 103 |
+
|
| 104 |
+
# pdm
|
| 105 |
+
# Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control.
|
| 106 |
+
#pdm.lock
|
| 107 |
+
# pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it
|
| 108 |
+
# in version control.
|
| 109 |
+
# https://pdm.fming.dev/#use-with-ide
|
| 110 |
+
.pdm.toml
|
| 111 |
+
|
| 112 |
+
# PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm
|
| 113 |
+
__pypackages__/
|
| 114 |
+
|
| 115 |
+
# Celery stuff
|
| 116 |
+
celerybeat-schedule
|
| 117 |
+
celerybeat.pid
|
| 118 |
+
|
| 119 |
+
# SageMath parsed files
|
| 120 |
+
*.sage.py
|
| 121 |
+
|
| 122 |
+
# Environments
|
| 123 |
+
.env
|
| 124 |
+
.venv
|
| 125 |
+
env/
|
| 126 |
+
venv/
|
| 127 |
+
ENV/
|
| 128 |
+
env.bak/
|
| 129 |
+
venv.bak/
|
| 130 |
+
|
| 131 |
+
# Spyder project settings
|
| 132 |
+
.spyderproject
|
| 133 |
+
.spyproject
|
| 134 |
+
|
| 135 |
+
# Rope project settings
|
| 136 |
+
.ropeproject
|
| 137 |
+
|
| 138 |
+
# mkdocs documentation
|
| 139 |
+
/site
|
| 140 |
+
|
| 141 |
+
# mypy
|
| 142 |
+
.mypy_cache/
|
| 143 |
+
.dmypy.json
|
| 144 |
+
dmypy.json
|
| 145 |
+
|
| 146 |
+
# Pyre type checker
|
| 147 |
+
.pyre/
|
| 148 |
+
|
| 149 |
+
# pytype static type analyzer
|
| 150 |
+
.pytype/
|
| 151 |
+
|
| 152 |
+
# Cython debug symbols
|
| 153 |
+
cython_debug/
|
| 154 |
+
|
| 155 |
+
# PyCharm
|
| 156 |
+
# JetBrains specific template is maintained in a separate JetBrains.gitignore that can
|
| 157 |
+
# be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
|
| 158 |
+
# and can be added to the global gitignore or merged into this file. For a more nuclear
|
| 159 |
+
# option (not recommended) you can uncomment the following to ignore the entire idea folder.
|
| 160 |
+
.idea/
|
| 161 |
+
|
| 162 |
+
examples/results/*
|
| 163 |
+
gfpgan/*
|
| 164 |
+
checkpoints/*
|
| 165 |
+
assets/*
|
| 166 |
+
results/*
|
| 167 |
+
Dockerfile
|
| 168 |
+
start_docker.sh
|
| 169 |
+
start.sh
|
| 170 |
+
|
| 171 |
+
checkpoints
|
| 172 |
+
|
| 173 |
+
# Mac
|
| 174 |
+
.DS_Store
|
.ipynb_checkpoints/Untitled-checkpoint.ipynb
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [],
|
| 3 |
+
"metadata": {},
|
| 4 |
+
"nbformat": 4,
|
| 5 |
+
"nbformat_minor": 5
|
| 6 |
+
}
|
LICENSE
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
MIT License
|
| 2 |
+
|
| 3 |
+
Copyright (c) 2023 Tencent AI Lab
|
| 4 |
+
|
| 5 |
+
Permission is hereby granted, free of charge, to any person obtaining a copy
|
| 6 |
+
of this software and associated documentation files (the "Software"), to deal
|
| 7 |
+
in the Software without restriction, including without limitation the rights
|
| 8 |
+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
| 9 |
+
copies of the Software, and to permit persons to whom the Software is
|
| 10 |
+
furnished to do so, subject to the following conditions:
|
| 11 |
+
|
| 12 |
+
The above copyright notice and this permission notice shall be included in all
|
| 13 |
+
copies or substantial portions of the Software.
|
| 14 |
+
|
| 15 |
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
| 16 |
+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
| 17 |
+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
| 18 |
+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
| 19 |
+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| 20 |
+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
| 21 |
+
SOFTWARE.
|
README.md
CHANGED
|
@@ -1,13 +1,268 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<div align="center">
|
| 2 |
+
|
| 3 |
+
<img src='https://user-images.githubusercontent.com/4397546/229094115-862c747e-7397-4b54-ba4a-bd368bfe2e0f.png' width='500px'/>
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
<!--<h2> 😭 SadTalker: <span style="font-size:12px">Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation </span> </h2> -->
|
| 7 |
+
|
| 8 |
+
<a href='https://arxiv.org/abs/2211.12194'><img src='https://img.shields.io/badge/ArXiv-PDF-red'></a> <a href='https://sadtalker.github.io'><img src='https://img.shields.io/badge/Project-Page-Green'></a> [](https://colab.research.google.com/github/Winfredy/SadTalker/blob/main/quick_demo.ipynb) [](https://huggingface.co/spaces/vinthony/SadTalker) [](https://colab.research.google.com/github/camenduru/stable-diffusion-webui-colab/blob/main/video/stable/stable_diffusion_1_5_video_webui_colab.ipynb) [](https://replicate.com/cjwbw/sadtalker)
|
| 9 |
+
|
| 10 |
+
<div>
|
| 11 |
+
<a target='_blank'>Wenxuan Zhang <sup>*,1,2</sup> </a> 
|
| 12 |
+
<a href='https://vinthony.github.io/' target='_blank'>Xiaodong Cun <sup>*,2</a> 
|
| 13 |
+
<a href='https://xuanwangvc.github.io/' target='_blank'>Xuan Wang <sup>3</sup></a> 
|
| 14 |
+
<a href='https://yzhang2016.github.io/' target='_blank'>Yong Zhang <sup>2</sup></a> 
|
| 15 |
+
<a href='https://xishen0220.github.io/' target='_blank'>Xi Shen <sup>2</sup></a>  </br>
|
| 16 |
+
<a href='https://yuguo-xjtu.github.io/' target='_blank'>Yu Guo<sup>1</sup> </a> 
|
| 17 |
+
<a href='https://scholar.google.com/citations?hl=zh-CN&user=4oXBp9UAAAAJ' target='_blank'>Ying Shan <sup>2</sup> </a> 
|
| 18 |
+
<a target='_blank'>Fei Wang <sup>1</sup> </a> 
|
| 19 |
+
</div>
|
| 20 |
+
<br>
|
| 21 |
+
<div>
|
| 22 |
+
<sup>1</sup> Xi'an Jiaotong University   <sup>2</sup> Tencent AI Lab   <sup>3</sup> Ant Group  
|
| 23 |
+
</div>
|
| 24 |
+
<br>
|
| 25 |
+
<i><strong><a href='https://arxiv.org/abs/2211.12194' target='_blank'>CVPR 2023</a></strong></i>
|
| 26 |
+
<br>
|
| 27 |
+
<br>
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+

|
| 31 |
+
|
| 32 |
+
<b>TL;DR: single portrait image 🙎♂️ + audio 🎤 = talking head video 🎞.</b>
|
| 33 |
+
|
| 34 |
+
<br>
|
| 35 |
+
|
| 36 |
+
</div>
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
## 🔥 Highlight
|
| 41 |
+
|
| 42 |
+
- 🔥 The extension of the [stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) is online. Checkout more details [here](docs/webui_extension.md).
|
| 43 |
+
|
| 44 |
+
https://user-images.githubusercontent.com/4397546/231495639-5d4bb925-ea64-4a36-a519-6389917dac29.mp4
|
| 45 |
+
|
| 46 |
+
- 🔥 `full image mode` is online! checkout [here](https://github.com/Winfredy/SadTalker#full-bodyimage-generation) for more details.
|
| 47 |
+
|
| 48 |
+
| still+enhancer in v0.0.1 | still + enhancer in v0.0.2 | [input image @bagbag1815](https://twitter.com/bagbag1815/status/1642754319094108161) |
|
| 49 |
+
|:--------------------: |:--------------------: | :----: |
|
| 50 |
+
| <video src="https://user-images.githubusercontent.com/48216707/229484996-5d7be64f-2553-4c9e-a452-c5cf0b8ebafe.mp4" type="video/mp4"> </video> | <video src="https://user-images.githubusercontent.com/4397546/230717873-355b7bf3-d3de-49f9-a439-9220e623fce7.mp4" type="video/mp4"> </video> | <img src='./examples/source_image/full_body_2.png' width='380'>
|
| 51 |
+
|
| 52 |
+
- 🔥 Several new mode, eg, `still mode`, `reference mode`, `resize mode` are online for better and custom applications.
|
| 53 |
+
|
| 54 |
+
- 🔥 Happy to see more community demos at [bilibili](https://search.bilibili.com/all?keyword=sadtalker&from_source=webtop_search&spm_id_from=333.1007&search_source=3
|
| 55 |
+
), [Youtube](https://www.youtube.com/results?search_query=sadtalker&sp=CAM%253D) and [twitter #sadtalker](https://twitter.com/search?q=%23sadtalker&src=typed_query).
|
| 56 |
+
|
| 57 |
+
## 📋 Changelog (Previous changelog can be founded [here](docs/changlelog.md))
|
| 58 |
+
|
| 59 |
+
- __[2023.06.12]__: add more new features in WEBUI extension, see the discussion [here](https://github.com/OpenTalker/SadTalker/discussions/386).
|
| 60 |
+
|
| 61 |
+
- __[2023.06.05]__: release a new 512 beta face model. Fixed some bugs and improve the performance.
|
| 62 |
+
|
| 63 |
+
- __[2023.04.15]__: Adding automatic1111 colab by @camenduru, thanks for this awesome colab: [](https://colab.research.google.com/github/camenduru/stable-diffusion-webui-colab/blob/main/video/stable/stable_diffusion_1_5_video_webui_colab.ipynb).
|
| 64 |
+
|
| 65 |
+
- __[2023.04.12]__: adding a more detailed sd-webui installation document, fixed reinstallation problem.
|
| 66 |
+
|
| 67 |
+
- __[2023.04.12]__: Fixed the sd-webui safe issues becasue of the 3rd packages, optimize the output path in `sd-webui-extension`.
|
| 68 |
+
|
| 69 |
+
- __[2023.04.08]__: ❗️❗️❗️ In v0.0.2, we add a logo watermark to the generated video to prevent abusing since it is very realistic.
|
| 70 |
+
|
| 71 |
+
- __[2023.04.08]__: v0.0.2, full image animation, adding baidu driver for download checkpoints. Optimizing the logic about enhancer.
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
## 🚧 TODO: See the Discussion https://github.com/OpenTalker/SadTalker/issues/280
|
| 75 |
+
|
| 76 |
+
## If you have any problem, please view our [FAQ](docs/FAQ.md) before opening an issue.
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
## ⚙️ 1. Installation.
|
| 81 |
+
|
| 82 |
+
Tutorials from communities: [中文windows教程](https://www.bilibili.com/video/BV1Dc411W7V6/) | [日本語コース](https://br-d.fanbox.cc/posts/5685086?utm_campaign=manage_post_page&utm_medium=share&utm_source=twitter)
|
| 83 |
+
|
| 84 |
+
### Linux:
|
| 85 |
+
|
| 86 |
+
1. Installing [anaconda](https://www.anaconda.com/), python and git.
|
| 87 |
+
|
| 88 |
+
2. Creating the env and install the requirements.
|
| 89 |
+
```bash
|
| 90 |
+
git clone https://github.com/Winfredy/SadTalker.git
|
| 91 |
+
|
| 92 |
+
cd SadTalker
|
| 93 |
+
|
| 94 |
+
conda create -n sadtalker python=3.8
|
| 95 |
+
|
| 96 |
+
conda activate sadtalker
|
| 97 |
+
|
| 98 |
+
pip install torch==1.12.1+cu113 torchvision==0.13.1+cu113 torchaudio==0.12.1 --extra-index-url https://download.pytorch.org/whl/cu113
|
| 99 |
+
|
| 100 |
+
conda install ffmpeg
|
| 101 |
+
|
| 102 |
+
pip install -r requirements.txt
|
| 103 |
+
|
| 104 |
+
### tts is optional for gradio demo.
|
| 105 |
+
### pip install TTS
|
| 106 |
+
|
| 107 |
+
```
|
| 108 |
+
### Windows ([中文windows教程](https://www.bilibili.com/video/BV1Dc411W7V6/)):
|
| 109 |
+
|
| 110 |
+
1. Install [Python 3.10.6](https://www.python.org/downloads/windows/), checking "Add Python to PATH".
|
| 111 |
+
2. Install [git](https://git-scm.com/download/win) manually (OR `scoop install git` via [scoop](https://scoop.sh/)).
|
| 112 |
+
3. Install `ffmpeg`, following [this instruction](https://www.wikihow.com/Install-FFmpeg-on-Windows) (OR using `scoop install ffmpeg` via [scoop](https://scoop.sh/)).
|
| 113 |
+
4. Download our SadTalker repository, for example by running `git clone https://github.com/Winfredy/SadTalker.git`.
|
| 114 |
+
5. Download the `checkpoint` and `gfpgan` [below↓](https://github.com/Winfredy/SadTalker#-2-download-trained-models).
|
| 115 |
+
5. Run `start.bat` from Windows Explorer as normal, non-administrator, user, a gradio WebUI demo will be started.
|
| 116 |
+
|
| 117 |
+
### Macbook:
|
| 118 |
+
|
| 119 |
+
More tips about installnation on Macbook and the Docker file can be founded [here](docs/install.md)
|
| 120 |
+
|
| 121 |
+
## 📥 2. Download Trained Models.
|
| 122 |
+
|
| 123 |
+
You can run the following script to put all the models in the right place.
|
| 124 |
+
|
| 125 |
+
```bash
|
| 126 |
+
bash scripts/download_models.sh
|
| 127 |
+
```
|
| 128 |
+
|
| 129 |
+
Other alternatives:
|
| 130 |
+
> we also provide an offline patch (`gfpgan/`), thus, no model will be downloaded when generating.
|
| 131 |
+
|
| 132 |
+
**Google Driver**: download our pre-trained model from [ this link (main checkpoints)](https://drive.google.com/file/d/1gwWh45pF7aelNP_P78uDJL8Sycep-K7j/view?usp=sharing) and [ gfpgan (offline patch)](https://drive.google.com/file/d/19AIBsmfcHW6BRJmeqSFlG5fL445Xmsyi?usp=sharing)
|
| 133 |
+
|
| 134 |
+
**Github Release Page**: download all the files from the [lastest github release page](https://github.com/Winfredy/SadTalker/releases), and then, put it in ./checkpoints.
|
| 135 |
+
|
| 136 |
+
**百度云盘**: we provided the downloaded model in [checkpoints, 提取码: sadt.](https://pan.baidu.com/s/1P4fRgk9gaSutZnn8YW034Q?pwd=sadt) And [gfpgan, 提取码: sadt.](https://pan.baidu.com/s/1kb1BCPaLOWX1JJb9Czbn6w?pwd=sadt)
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
<details><summary>Model Details</summary>
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
Model explains:
|
| 144 |
+
|
| 145 |
+
##### New version
|
| 146 |
+
| Model | Description
|
| 147 |
+
| :--- | :----------
|
| 148 |
+
|checkpoints/mapping_00229-model.pth.tar | Pre-trained MappingNet in Sadtalker.
|
| 149 |
+
|checkpoints/mapping_00109-model.pth.tar | Pre-trained MappingNet in Sadtalker.
|
| 150 |
+
|checkpoints/SadTalker_V0.0.2_256.safetensors | packaged sadtalker checkpoints of old version, 256 face render).
|
| 151 |
+
|checkpoints/SadTalker_V0.0.2_512.safetensors | packaged sadtalker checkpoints of old version, 512 face render).
|
| 152 |
+
|gfpgan/weights | Face detection and enhanced models used in `facexlib` and `gfpgan`.
|
| 153 |
+
|
| 154 |
+
|
| 155 |
+
##### Old version
|
| 156 |
+
| Model | Description
|
| 157 |
+
| :--- | :----------
|
| 158 |
+
|checkpoints/auido2exp_00300-model.pth | Pre-trained ExpNet in Sadtalker.
|
| 159 |
+
|checkpoints/auido2pose_00140-model.pth | Pre-trained PoseVAE in Sadtalker.
|
| 160 |
+
|checkpoints/mapping_00229-model.pth.tar | Pre-trained MappingNet in Sadtalker.
|
| 161 |
+
|checkpoints/mapping_00109-model.pth.tar | Pre-trained MappingNet in Sadtalker.
|
| 162 |
+
|checkpoints/facevid2vid_00189-model.pth.tar | Pre-trained face-vid2vid model from [the reappearance of face-vid2vid](https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis).
|
| 163 |
+
|checkpoints/epoch_20.pth | Pre-trained 3DMM extractor in [Deep3DFaceReconstruction](https://github.com/microsoft/Deep3DFaceReconstruction).
|
| 164 |
+
|checkpoints/wav2lip.pth | Highly accurate lip-sync model in [Wav2lip](https://github.com/Rudrabha/Wav2Lip).
|
| 165 |
+
|checkpoints/shape_predictor_68_face_landmarks.dat | Face landmark model used in [dilb](http://dlib.net/).
|
| 166 |
+
|checkpoints/BFM | 3DMM library file.
|
| 167 |
+
|checkpoints/hub | Face detection models used in [face alignment](https://github.com/1adrianb/face-alignment).
|
| 168 |
+
|gfpgan/weights | Face detection and enhanced models used in `facexlib` and `gfpgan`.
|
| 169 |
+
|
| 170 |
+
The final folder will be shown as:
|
| 171 |
+
|
| 172 |
+
<img width="331" alt="image" src="https://user-images.githubusercontent.com/4397546/232511411-4ca75cbf-a434-48c5-9ae0-9009e8316484.png">
|
| 173 |
+
|
| 174 |
+
|
| 175 |
+
</details>
|
| 176 |
+
|
| 177 |
+
## 🔮 3. Quick Start ([Best Practice](docs/best_practice.md)).
|
| 178 |
+
|
| 179 |
+
### WebUI Demos:
|
| 180 |
+
|
| 181 |
+
**Online**: [Huggingface](https://huggingface.co/spaces/vinthony/SadTalker) | [SDWebUI-Colab](https://colab.research.google.com/github/camenduru/stable-diffusion-webui-colab/blob/main/video/stable/stable_diffusion_1_5_video_webui_colab.ipynb) | [Colab](https://colab.research.google.com/github/Winfredy/SadTalker/blob/main/quick_demo.ipynb)
|
| 182 |
+
|
| 183 |
+
**Local Autiomatic1111 stable-diffusion webui extension**: please refer to [Autiomatic1111 stable-diffusion webui docs](docs/webui_extension.md).
|
| 184 |
+
|
| 185 |
+
**Local gradio demo(highly recommanded!)**: Similar to our [hugging-face demo](https://huggingface.co/spaces/vinthony/SadTalker) can be run by:
|
| 186 |
+
|
| 187 |
+
```bash
|
| 188 |
+
## you need manually install TTS(https://github.com/coqui-ai/TTS) via `pip install tts` in advanced.
|
| 189 |
+
python app.py
|
| 190 |
+
```
|
| 191 |
+
|
| 192 |
+
**Local gradio demo(highly recommanded!)**:
|
| 193 |
+
|
| 194 |
+
- windows: just double click `webui.bat`, the requirements will be installed automatically.
|
| 195 |
+
- Linux/Mac OS: run `bash webui.sh` to start the webui.
|
| 196 |
+
|
| 197 |
+
|
| 198 |
+
### Manually usages:
|
| 199 |
+
|
| 200 |
+
##### Animating a portrait image from default config:
|
| 201 |
+
```bash
|
| 202 |
+
python inference.py --driven_audio <audio.wav> \
|
| 203 |
+
--source_image <video.mp4 or picture.png> \
|
| 204 |
+
--enhancer gfpgan
|
| 205 |
+
```
|
| 206 |
+
The results will be saved in `results/$SOME_TIMESTAMP/*.mp4`.
|
| 207 |
+
|
| 208 |
+
##### Full body/image Generation:
|
| 209 |
+
|
| 210 |
+
Using `--still` to generate a natural full body video. You can add `enhancer` to improve the quality of the generated video.
|
| 211 |
+
|
| 212 |
+
```bash
|
| 213 |
+
python inference.py --driven_audio <audio.wav> \
|
| 214 |
+
--source_image <video.mp4 or picture.png> \
|
| 215 |
+
--result_dir <a file to store results> \
|
| 216 |
+
--still \
|
| 217 |
+
--preprocess full \
|
| 218 |
+
--enhancer gfpgan
|
| 219 |
+
```
|
| 220 |
+
|
| 221 |
+
More examples and configuration and tips can be founded in the [ >>> best practice documents <<<](docs/best_practice.md).
|
| 222 |
+
|
| 223 |
+
## 🛎 Citation
|
| 224 |
+
|
| 225 |
+
If you find our work useful in your research, please consider citing:
|
| 226 |
+
|
| 227 |
+
```bibtex
|
| 228 |
+
@article{zhang2022sadtalker,
|
| 229 |
+
title={SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation},
|
| 230 |
+
author={Zhang, Wenxuan and Cun, Xiaodong and Wang, Xuan and Zhang, Yong and Shen, Xi and Guo, Yu and Shan, Ying and Wang, Fei},
|
| 231 |
+
journal={arXiv preprint arXiv:2211.12194},
|
| 232 |
+
year={2022}
|
| 233 |
+
}
|
| 234 |
+
```
|
| 235 |
+
|
| 236 |
+
|
| 237 |
+
|
| 238 |
+
## 💗 Acknowledgements
|
| 239 |
+
|
| 240 |
+
Facerender code borrows heavily from [zhanglonghao's reproduction of face-vid2vid](https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis) and [PIRender](https://github.com/RenYurui/PIRender). We thank the authors for sharing their wonderful code. In training process, We also use the model from [Deep3DFaceReconstruction](https://github.com/microsoft/Deep3DFaceReconstruction) and [Wav2lip](https://github.com/Rudrabha/Wav2Lip). We thank for their wonderful work.
|
| 241 |
+
|
| 242 |
+
See also these wonderful 3rd libraries we use:
|
| 243 |
+
|
| 244 |
+
- **Face Utils**: https://github.com/xinntao/facexlib
|
| 245 |
+
- **Face Enhancement**: https://github.com/TencentARC/GFPGAN
|
| 246 |
+
- **Image/Video Enhancement**:https://github.com/xinntao/Real-ESRGAN
|
| 247 |
+
|
| 248 |
+
## 🥂 Extensions:
|
| 249 |
+
|
| 250 |
+
- [SadTalker-Video-Lip-Sync](https://github.com/Zz-ww/SadTalker-Video-Lip-Sync) from [@Zz-ww](https://github.com/Zz-ww): SadTalker for Video Lip Editing
|
| 251 |
+
|
| 252 |
+
## 🥂 Related Works
|
| 253 |
+
- [StyleHEAT: One-Shot High-Resolution Editable Talking Face Generation via Pre-trained StyleGAN (ECCV 2022)](https://github.com/FeiiYin/StyleHEAT)
|
| 254 |
+
- [CodeTalker: Speech-Driven 3D Facial Animation with Discrete Motion Prior (CVPR 2023)](https://github.com/Doubiiu/CodeTalker)
|
| 255 |
+
- [VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild (SIGGRAPH Asia 2022)](https://github.com/vinthony/video-retalking)
|
| 256 |
+
- [DPE: Disentanglement of Pose and Expression for General Video Portrait Editing (CVPR 2023)](https://github.com/Carlyx/DPE)
|
| 257 |
+
- [3D GAN Inversion with Facial Symmetry Prior (CVPR 2023)](https://github.com/FeiiYin/SPI/)
|
| 258 |
+
- [T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations (CVPR 2023)](https://github.com/Mael-zys/T2M-GPT)
|
| 259 |
+
|
| 260 |
+
## 📢 Disclaimer
|
| 261 |
+
|
| 262 |
+
This is not an official product of Tencent. This repository can only be used for personal/research/non-commercial purposes.
|
| 263 |
+
|
| 264 |
+
LOGO: color and font suggestion: [ChatGPT](ai.com), logo font:[Montserrat Alternates
|
| 265 |
+
](https://fonts.google.com/specimen/Montserrat+Alternates?preview.text=SadTalker&preview.text_type=custom&query=mont).
|
| 266 |
+
|
| 267 |
+
All the copyright of the demo images and audio are from communities users or the geneartion from stable diffusion. Free free to contact us if you feel uncomfortable.
|
| 268 |
+
|
Untitled.ipynb
ADDED
|
@@ -0,0 +1,533 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [
|
| 3 |
+
{
|
| 4 |
+
"cell_type": "code",
|
| 5 |
+
"execution_count": 1,
|
| 6 |
+
"id": "c69a901b-8c3f-4188-97bd-5594c4496ec5",
|
| 7 |
+
"metadata": {},
|
| 8 |
+
"outputs": [],
|
| 9 |
+
"source": [
|
| 10 |
+
"from huggingface_hub import HfApi\n",
|
| 11 |
+
"api = HfApi()"
|
| 12 |
+
]
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"cell_type": "code",
|
| 16 |
+
"execution_count": 3,
|
| 17 |
+
"id": "8b249adc-ccd0-4145-86ce-64509ad276cf",
|
| 18 |
+
"metadata": {},
|
| 19 |
+
"outputs": [
|
| 20 |
+
{
|
| 21 |
+
"data": {
|
| 22 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 23 |
+
"model_id": "faf7b41b81e54705bae8921f2a86e9fd",
|
| 24 |
+
"version_major": 2,
|
| 25 |
+
"version_minor": 0
|
| 26 |
+
},
|
| 27 |
+
"text/plain": [
|
| 28 |
+
"VBox(children=(HTML(value='<center> <img\\nsrc=https://huggingface.co/front/assets/huggingface_logo-noborder.sv…"
|
| 29 |
+
]
|
| 30 |
+
},
|
| 31 |
+
"metadata": {},
|
| 32 |
+
"output_type": "display_data"
|
| 33 |
+
}
|
| 34 |
+
],
|
| 35 |
+
"source": [
|
| 36 |
+
"from huggingface_hub import login\n",
|
| 37 |
+
"login()"
|
| 38 |
+
]
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"cell_type": "code",
|
| 42 |
+
"execution_count": null,
|
| 43 |
+
"id": "41815b66-b12b-45da-9a5e-92c527fe2b4a",
|
| 44 |
+
"metadata": {},
|
| 45 |
+
"outputs": [
|
| 46 |
+
{
|
| 47 |
+
"data": {
|
| 48 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 49 |
+
"model_id": "6bb4c00dda664acd9f70e029e15dbd57",
|
| 50 |
+
"version_major": 2,
|
| 51 |
+
"version_minor": 0
|
| 52 |
+
},
|
| 53 |
+
"text/plain": [
|
| 54 |
+
"SadTalker_V0.0.2_512.safetensors: 0%| | 0.00/725M [00:00<?, ?B/s]"
|
| 55 |
+
]
|
| 56 |
+
},
|
| 57 |
+
"metadata": {},
|
| 58 |
+
"output_type": "display_data"
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"data": {
|
| 62 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 63 |
+
"model_id": "b9a78d85c20642ac86da6b6d0ab97031",
|
| 64 |
+
"version_major": 2,
|
| 65 |
+
"version_minor": 0
|
| 66 |
+
},
|
| 67 |
+
"text/plain": [
|
| 68 |
+
"mapping_00109-model.pth.tar: 0%| | 0.00/156M [00:00<?, ?B/s]"
|
| 69 |
+
]
|
| 70 |
+
},
|
| 71 |
+
"metadata": {},
|
| 72 |
+
"output_type": "display_data"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"data": {
|
| 76 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 77 |
+
"model_id": "d396b1372ea04ae79dd82677c1c33352",
|
| 78 |
+
"version_major": 2,
|
| 79 |
+
"version_minor": 0
|
| 80 |
+
},
|
| 81 |
+
"text/plain": [
|
| 82 |
+
"Upload 31 LFS files: 0%| | 0/31 [00:00<?, ?it/s]"
|
| 83 |
+
]
|
| 84 |
+
},
|
| 85 |
+
"metadata": {},
|
| 86 |
+
"output_type": "display_data"
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"data": {
|
| 90 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 91 |
+
"model_id": "7b924c6f560b40b78152a87f5c388d6b",
|
| 92 |
+
"version_major": 2,
|
| 93 |
+
"version_minor": 0
|
| 94 |
+
},
|
| 95 |
+
"text/plain": [
|
| 96 |
+
"mapping_00229-model.pth.tar: 0%| | 0.00/156M [00:00<?, ?B/s]"
|
| 97 |
+
]
|
| 98 |
+
},
|
| 99 |
+
"metadata": {},
|
| 100 |
+
"output_type": "display_data"
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"data": {
|
| 104 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 105 |
+
"model_id": "4245c014a4d04d6daacd47d7d81c6249",
|
| 106 |
+
"version_major": 2,
|
| 107 |
+
"version_minor": 0
|
| 108 |
+
},
|
| 109 |
+
"text/plain": [
|
| 110 |
+
"SadTalker_V0.0.2_256.safetensors: 0%| | 0.00/725M [00:00<?, ?B/s]"
|
| 111 |
+
]
|
| 112 |
+
},
|
| 113 |
+
"metadata": {},
|
| 114 |
+
"output_type": "display_data"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"data": {
|
| 118 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 119 |
+
"model_id": "d7f2b0a6b3ed4d5e895da21237f42577",
|
| 120 |
+
"version_major": 2,
|
| 121 |
+
"version_minor": 0
|
| 122 |
+
},
|
| 123 |
+
"text/plain": [
|
| 124 |
+
"example_crop.gif: 0%| | 0.00/1.55M [00:00<?, ?B/s]"
|
| 125 |
+
]
|
| 126 |
+
},
|
| 127 |
+
"metadata": {},
|
| 128 |
+
"output_type": "display_data"
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"data": {
|
| 132 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 133 |
+
"model_id": "24ea098c821c4fc79ffa38b2fe5a1575",
|
| 134 |
+
"version_major": 2,
|
| 135 |
+
"version_minor": 0
|
| 136 |
+
},
|
| 137 |
+
"text/plain": [
|
| 138 |
+
"example_crop_still.gif: 0%| | 0.00/1.25M [00:00<?, ?B/s]"
|
| 139 |
+
]
|
| 140 |
+
},
|
| 141 |
+
"metadata": {},
|
| 142 |
+
"output_type": "display_data"
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"data": {
|
| 146 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 147 |
+
"model_id": "b828fea9017846748cbe423f1b5fada7",
|
| 148 |
+
"version_major": 2,
|
| 149 |
+
"version_minor": 0
|
| 150 |
+
},
|
| 151 |
+
"text/plain": [
|
| 152 |
+
"example_full.gif: 0%| | 0.00/1.46M [00:00<?, ?B/s]"
|
| 153 |
+
]
|
| 154 |
+
},
|
| 155 |
+
"metadata": {},
|
| 156 |
+
"output_type": "display_data"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"data": {
|
| 160 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 161 |
+
"model_id": "a0b664fbf3f8465d9b286290e5f14d8f",
|
| 162 |
+
"version_major": 2,
|
| 163 |
+
"version_minor": 0
|
| 164 |
+
},
|
| 165 |
+
"text/plain": [
|
| 166 |
+
"example_full_enhanced.gif: 0%| | 0.00/5.78M [00:00<?, ?B/s]"
|
| 167 |
+
]
|
| 168 |
+
},
|
| 169 |
+
"metadata": {},
|
| 170 |
+
"output_type": "display_data"
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"data": {
|
| 174 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 175 |
+
"model_id": "e7483527580845429ce33f4f987e7ddb",
|
| 176 |
+
"version_major": 2,
|
| 177 |
+
"version_minor": 0
|
| 178 |
+
},
|
| 179 |
+
"text/plain": [
|
| 180 |
+
"free_view_result.gif: 0%| | 0.00/5.61M [00:00<?, ?B/s]"
|
| 181 |
+
]
|
| 182 |
+
},
|
| 183 |
+
"metadata": {},
|
| 184 |
+
"output_type": "display_data"
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"data": {
|
| 188 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 189 |
+
"model_id": "7662c74dc37e4e8a9e756da854c7267a",
|
| 190 |
+
"version_major": 2,
|
| 191 |
+
"version_minor": 0
|
| 192 |
+
},
|
| 193 |
+
"text/plain": [
|
| 194 |
+
"resize_good.gif: 0%| | 0.00/1.73M [00:00<?, ?B/s]"
|
| 195 |
+
]
|
| 196 |
+
},
|
| 197 |
+
"metadata": {},
|
| 198 |
+
"output_type": "display_data"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"data": {
|
| 202 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 203 |
+
"model_id": "0e6e3f1dbb26419c895a7fed7a7f31ac",
|
| 204 |
+
"version_major": 2,
|
| 205 |
+
"version_minor": 0
|
| 206 |
+
},
|
| 207 |
+
"text/plain": [
|
| 208 |
+
"resize_no.gif: 0%| | 0.00/2.14M [00:00<?, ?B/s]"
|
| 209 |
+
]
|
| 210 |
+
},
|
| 211 |
+
"metadata": {},
|
| 212 |
+
"output_type": "display_data"
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"data": {
|
| 216 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 217 |
+
"model_id": "bdf7d2a9e5784350af4be3d57cc72fa4",
|
| 218 |
+
"version_major": 2,
|
| 219 |
+
"version_minor": 0
|
| 220 |
+
},
|
| 221 |
+
"text/plain": [
|
| 222 |
+
"using_ref_video.gif: 0%| | 0.00/8.11M [00:00<?, ?B/s]"
|
| 223 |
+
]
|
| 224 |
+
},
|
| 225 |
+
"metadata": {},
|
| 226 |
+
"output_type": "display_data"
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"data": {
|
| 230 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 231 |
+
"model_id": "3348de1636bc4d85b2674eeed42d03d9",
|
| 232 |
+
"version_major": 2,
|
| 233 |
+
"version_minor": 0
|
| 234 |
+
},
|
| 235 |
+
"text/plain": [
|
| 236 |
+
"chinese_news.wav: 0%| | 0.00/1.54M [00:00<?, ?B/s]"
|
| 237 |
+
]
|
| 238 |
+
},
|
| 239 |
+
"metadata": {},
|
| 240 |
+
"output_type": "display_data"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"data": {
|
| 244 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 245 |
+
"model_id": "40d1dc6faf634304a13e18feab5816af",
|
| 246 |
+
"version_major": 2,
|
| 247 |
+
"version_minor": 0
|
| 248 |
+
},
|
| 249 |
+
"text/plain": [
|
| 250 |
+
"deyu.wav: 0%| | 0.00/2.69M [00:00<?, ?B/s]"
|
| 251 |
+
]
|
| 252 |
+
},
|
| 253 |
+
"metadata": {},
|
| 254 |
+
"output_type": "display_data"
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"data": {
|
| 258 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 259 |
+
"model_id": "2c91050e17e34841ae74d4a9b414eaf7",
|
| 260 |
+
"version_major": 2,
|
| 261 |
+
"version_minor": 0
|
| 262 |
+
},
|
| 263 |
+
"text/plain": [
|
| 264 |
+
"eluosi.wav: 0%| | 0.00/1.79M [00:00<?, ?B/s]"
|
| 265 |
+
]
|
| 266 |
+
},
|
| 267 |
+
"metadata": {},
|
| 268 |
+
"output_type": "display_data"
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"data": {
|
| 272 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 273 |
+
"model_id": "35ed59910dd04972b65bd0684dd497c1",
|
| 274 |
+
"version_major": 2,
|
| 275 |
+
"version_minor": 0
|
| 276 |
+
},
|
| 277 |
+
"text/plain": [
|
| 278 |
+
"fayu.wav: 0%| | 0.00/1.94M [00:00<?, ?B/s]"
|
| 279 |
+
]
|
| 280 |
+
},
|
| 281 |
+
"metadata": {},
|
| 282 |
+
"output_type": "display_data"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"data": {
|
| 286 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 287 |
+
"model_id": "6cc44214a3224b9fa952d8fdd69b262d",
|
| 288 |
+
"version_major": 2,
|
| 289 |
+
"version_minor": 0
|
| 290 |
+
},
|
| 291 |
+
"text/plain": [
|
| 292 |
+
"imagine.wav: 0%| | 0.00/1.62M [00:00<?, ?B/s]"
|
| 293 |
+
]
|
| 294 |
+
},
|
| 295 |
+
"metadata": {},
|
| 296 |
+
"output_type": "display_data"
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"data": {
|
| 300 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 301 |
+
"model_id": "21b4c8e2efbe4205b961d849c52c0c66",
|
| 302 |
+
"version_major": 2,
|
| 303 |
+
"version_minor": 0
|
| 304 |
+
},
|
| 305 |
+
"text/plain": [
|
| 306 |
+
"japanese.wav: 0%| | 0.00/2.62M [00:00<?, ?B/s]"
|
| 307 |
+
]
|
| 308 |
+
},
|
| 309 |
+
"metadata": {},
|
| 310 |
+
"output_type": "display_data"
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"data": {
|
| 314 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 315 |
+
"model_id": "8df2d5959a394789b9defb2d8dea9680",
|
| 316 |
+
"version_major": 2,
|
| 317 |
+
"version_minor": 0
|
| 318 |
+
},
|
| 319 |
+
"text/plain": [
|
| 320 |
+
"WDA_AlexandriaOcasioCortez_000.mp4: 0%| | 0.00/2.26M [00:00<?, ?B/s]"
|
| 321 |
+
]
|
| 322 |
+
},
|
| 323 |
+
"metadata": {},
|
| 324 |
+
"output_type": "display_data"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"data": {
|
| 328 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 329 |
+
"model_id": "bee85421a93345f88e1ba92bdc627a92",
|
| 330 |
+
"version_major": 2,
|
| 331 |
+
"version_minor": 0
|
| 332 |
+
},
|
| 333 |
+
"text/plain": [
|
| 334 |
+
"WDA_KatieHill_000.mp4: 0%| | 0.00/3.55M [00:00<?, ?B/s]"
|
| 335 |
+
]
|
| 336 |
+
},
|
| 337 |
+
"metadata": {},
|
| 338 |
+
"output_type": "display_data"
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"data": {
|
| 342 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 343 |
+
"model_id": "1dfceecdadf2443bb935b58c06116ac3",
|
| 344 |
+
"version_major": 2,
|
| 345 |
+
"version_minor": 0
|
| 346 |
+
},
|
| 347 |
+
"text/plain": [
|
| 348 |
+
"art_16.png: 0%| | 0.00/1.48M [00:00<?, ?B/s]"
|
| 349 |
+
]
|
| 350 |
+
},
|
| 351 |
+
"metadata": {},
|
| 352 |
+
"output_type": "display_data"
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"data": {
|
| 356 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 357 |
+
"model_id": "7b054b4faa49444786ac49f10bc2ca27",
|
| 358 |
+
"version_major": 2,
|
| 359 |
+
"version_minor": 0
|
| 360 |
+
},
|
| 361 |
+
"text/plain": [
|
| 362 |
+
"art_17.png: 0%| | 0.00/2.09M [00:00<?, ?B/s]"
|
| 363 |
+
]
|
| 364 |
+
},
|
| 365 |
+
"metadata": {},
|
| 366 |
+
"output_type": "display_data"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"data": {
|
| 370 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 371 |
+
"model_id": "78db6e2bdf3f4e8289cda18189d2f9bd",
|
| 372 |
+
"version_major": 2,
|
| 373 |
+
"version_minor": 0
|
| 374 |
+
},
|
| 375 |
+
"text/plain": [
|
| 376 |
+
"art_3.png: 0%| | 0.00/1.35M [00:00<?, ?B/s]"
|
| 377 |
+
]
|
| 378 |
+
},
|
| 379 |
+
"metadata": {},
|
| 380 |
+
"output_type": "display_data"
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"data": {
|
| 384 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 385 |
+
"model_id": "ef5a284a239548c4964ea3a924975cbc",
|
| 386 |
+
"version_major": 2,
|
| 387 |
+
"version_minor": 0
|
| 388 |
+
},
|
| 389 |
+
"text/plain": [
|
| 390 |
+
"art_4.png: 0%| | 0.00/3.63M [00:00<?, ?B/s]"
|
| 391 |
+
]
|
| 392 |
+
},
|
| 393 |
+
"metadata": {},
|
| 394 |
+
"output_type": "display_data"
|
| 395 |
+
},
|
| 396 |
+
{
|
| 397 |
+
"data": {
|
| 398 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 399 |
+
"model_id": "fbdf097df06247c98da1a3f7b4322c92",
|
| 400 |
+
"version_major": 2,
|
| 401 |
+
"version_minor": 0
|
| 402 |
+
},
|
| 403 |
+
"text/plain": [
|
| 404 |
+
"art_5.png: 0%| | 0.00/1.23M [00:00<?, ?B/s]"
|
| 405 |
+
]
|
| 406 |
+
},
|
| 407 |
+
"metadata": {},
|
| 408 |
+
"output_type": "display_data"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"data": {
|
| 412 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 413 |
+
"model_id": "1099a9330eae49178d53d35827f9fa05",
|
| 414 |
+
"version_major": 2,
|
| 415 |
+
"version_minor": 0
|
| 416 |
+
},
|
| 417 |
+
"text/plain": [
|
| 418 |
+
"art_8.png: 0%| | 0.00/3.12M [00:00<?, ?B/s]"
|
| 419 |
+
]
|
| 420 |
+
},
|
| 421 |
+
"metadata": {},
|
| 422 |
+
"output_type": "display_data"
|
| 423 |
+
},
|
| 424 |
+
{
|
| 425 |
+
"data": {
|
| 426 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 427 |
+
"model_id": "f53f1a45946f4a8db6c7a48f525b8f7c",
|
| 428 |
+
"version_major": 2,
|
| 429 |
+
"version_minor": 0
|
| 430 |
+
},
|
| 431 |
+
"text/plain": [
|
| 432 |
+
"art_9.png: 0%| | 0.00/1.26M [00:00<?, ?B/s]"
|
| 433 |
+
]
|
| 434 |
+
},
|
| 435 |
+
"metadata": {},
|
| 436 |
+
"output_type": "display_data"
|
| 437 |
+
},
|
| 438 |
+
{
|
| 439 |
+
"data": {
|
| 440 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 441 |
+
"model_id": "de038d6f5aa84c63b2335af75c82e337",
|
| 442 |
+
"version_major": 2,
|
| 443 |
+
"version_minor": 0
|
| 444 |
+
},
|
| 445 |
+
"text/plain": [
|
| 446 |
+
"GFPGANv1.4.pth: 0%| | 0.00/349M [00:00<?, ?B/s]"
|
| 447 |
+
]
|
| 448 |
+
},
|
| 449 |
+
"metadata": {},
|
| 450 |
+
"output_type": "display_data"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"data": {
|
| 454 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 455 |
+
"model_id": "69daa05b13a745fa8a27e6447282fdff",
|
| 456 |
+
"version_major": 2,
|
| 457 |
+
"version_minor": 0
|
| 458 |
+
},
|
| 459 |
+
"text/plain": [
|
| 460 |
+
"alignment_WFLW_4HG.pth: 0%| | 0.00/194M [00:00<?, ?B/s]"
|
| 461 |
+
]
|
| 462 |
+
},
|
| 463 |
+
"metadata": {},
|
| 464 |
+
"output_type": "display_data"
|
| 465 |
+
},
|
| 466 |
+
{
|
| 467 |
+
"data": {
|
| 468 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 469 |
+
"model_id": "ee0b32bbe61647359a6e8324a092b95e",
|
| 470 |
+
"version_major": 2,
|
| 471 |
+
"version_minor": 0
|
| 472 |
+
},
|
| 473 |
+
"text/plain": [
|
| 474 |
+
"detection_Resnet50_Final.pth: 0%| | 0.00/109M [00:00<?, ?B/s]"
|
| 475 |
+
]
|
| 476 |
+
},
|
| 477 |
+
"metadata": {},
|
| 478 |
+
"output_type": "display_data"
|
| 479 |
+
},
|
| 480 |
+
{
|
| 481 |
+
"data": {
|
| 482 |
+
"application/vnd.jupyter.widget-view+json": {
|
| 483 |
+
"model_id": "709b3e709e704234826dd1050f260688",
|
| 484 |
+
"version_major": 2,
|
| 485 |
+
"version_minor": 0
|
| 486 |
+
},
|
| 487 |
+
"text/plain": [
|
| 488 |
+
"parsing_parsenet.pth: 0%| | 0.00/85.3M [00:00<?, ?B/s]"
|
| 489 |
+
]
|
| 490 |
+
},
|
| 491 |
+
"metadata": {},
|
| 492 |
+
"output_type": "display_data"
|
| 493 |
+
}
|
| 494 |
+
],
|
| 495 |
+
"source": [
|
| 496 |
+
"api.upload_folder(\n",
|
| 497 |
+
" folder_path=\"/workspace/SadTalker\",\n",
|
| 498 |
+
" repo_id=\"kevinwang676/SadTalker\",\n",
|
| 499 |
+
" repo_type=\"space\",\n",
|
| 500 |
+
")"
|
| 501 |
+
]
|
| 502 |
+
},
|
| 503 |
+
{
|
| 504 |
+
"cell_type": "code",
|
| 505 |
+
"execution_count": null,
|
| 506 |
+
"id": "c15c0fad-c4ac-4fd5-b886-89d2f97bc285",
|
| 507 |
+
"metadata": {},
|
| 508 |
+
"outputs": [],
|
| 509 |
+
"source": []
|
| 510 |
+
}
|
| 511 |
+
],
|
| 512 |
+
"metadata": {
|
| 513 |
+
"kernelspec": {
|
| 514 |
+
"display_name": "Python 3 (ipykernel)",
|
| 515 |
+
"language": "python",
|
| 516 |
+
"name": "python3"
|
| 517 |
+
},
|
| 518 |
+
"language_info": {
|
| 519 |
+
"codemirror_mode": {
|
| 520 |
+
"name": "ipython",
|
| 521 |
+
"version": 3
|
| 522 |
+
},
|
| 523 |
+
"file_extension": ".py",
|
| 524 |
+
"mimetype": "text/x-python",
|
| 525 |
+
"name": "python",
|
| 526 |
+
"nbconvert_exporter": "python",
|
| 527 |
+
"pygments_lexer": "ipython3",
|
| 528 |
+
"version": "3.10.11"
|
| 529 |
+
}
|
| 530 |
+
},
|
| 531 |
+
"nbformat": 4,
|
| 532 |
+
"nbformat_minor": 5
|
| 533 |
+
}
|
app_sadtalker.py
ADDED
|
@@ -0,0 +1,111 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, sys
|
| 2 |
+
import gradio as gr
|
| 3 |
+
from src.gradio_demo import SadTalker
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
try:
|
| 7 |
+
import webui # in webui
|
| 8 |
+
in_webui = True
|
| 9 |
+
except:
|
| 10 |
+
in_webui = False
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
def toggle_audio_file(choice):
|
| 14 |
+
if choice == False:
|
| 15 |
+
return gr.update(visible=True), gr.update(visible=False)
|
| 16 |
+
else:
|
| 17 |
+
return gr.update(visible=False), gr.update(visible=True)
|
| 18 |
+
|
| 19 |
+
def ref_video_fn(path_of_ref_video):
|
| 20 |
+
if path_of_ref_video is not None:
|
| 21 |
+
return gr.update(value=True)
|
| 22 |
+
else:
|
| 23 |
+
return gr.update(value=False)
|
| 24 |
+
|
| 25 |
+
def sadtalker_demo(checkpoint_path='checkpoints', config_path='src/config', warpfn=None):
|
| 26 |
+
|
| 27 |
+
sad_talker = SadTalker(checkpoint_path, config_path, lazy_load=True)
|
| 28 |
+
|
| 29 |
+
with gr.Blocks(analytics_enabled=False) as sadtalker_interface:
|
| 30 |
+
gr.Markdown("<div align='center'> <h2> 😭 SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation (CVPR 2023) </span> </h2> \
|
| 31 |
+
<a style='font-size:18px;color: #efefef' href='https://arxiv.org/abs/2211.12194'>Arxiv</a> \
|
| 32 |
+
<a style='font-size:18px;color: #efefef' href='https://sadtalker.github.io'>Homepage</a> \
|
| 33 |
+
<a style='font-size:18px;color: #efefef' href='https://github.com/Winfredy/SadTalker'> Github </div>")
|
| 34 |
+
|
| 35 |
+
with gr.Row().style(equal_height=False):
|
| 36 |
+
with gr.Column(variant='panel'):
|
| 37 |
+
with gr.Tabs(elem_id="sadtalker_source_image"):
|
| 38 |
+
with gr.TabItem('Upload image'):
|
| 39 |
+
with gr.Row():
|
| 40 |
+
source_image = gr.Image(label="Source image", source="upload", type="filepath", elem_id="img2img_image").style(width=512)
|
| 41 |
+
|
| 42 |
+
with gr.Tabs(elem_id="sadtalker_driven_audio"):
|
| 43 |
+
with gr.TabItem('Upload OR TTS'):
|
| 44 |
+
with gr.Column(variant='panel'):
|
| 45 |
+
driven_audio = gr.Audio(label="Input audio", source="upload", type="filepath")
|
| 46 |
+
|
| 47 |
+
if sys.platform != 'win32' and not in_webui:
|
| 48 |
+
from src.utils.text2speech import TTSTalker
|
| 49 |
+
tts_talker = TTSTalker()
|
| 50 |
+
with gr.Column(variant='panel'):
|
| 51 |
+
input_text = gr.Textbox(label="Generating audio from text", lines=5, placeholder="please enter some text here, we genreate the audio from text using @Coqui.ai TTS.")
|
| 52 |
+
tts = gr.Button('Generate audio',elem_id="sadtalker_audio_generate", variant='primary')
|
| 53 |
+
tts.click(fn=tts_talker.test, inputs=[input_text], outputs=[driven_audio])
|
| 54 |
+
|
| 55 |
+
with gr.Column(variant='panel'):
|
| 56 |
+
with gr.Tabs(elem_id="sadtalker_checkbox"):
|
| 57 |
+
with gr.TabItem('Settings'):
|
| 58 |
+
gr.Markdown("need help? please visit our [best practice page](https://github.com/OpenTalker/SadTalker/blob/main/docs/best_practice.md) for more detials")
|
| 59 |
+
with gr.Column(variant='panel'):
|
| 60 |
+
# width = gr.Slider(minimum=64, elem_id="img2img_width", maximum=2048, step=8, label="Manually Crop Width", value=512) # img2img_width
|
| 61 |
+
# height = gr.Slider(minimum=64, elem_id="img2img_height", maximum=2048, step=8, label="Manually Crop Height", value=512) # img2img_width
|
| 62 |
+
pose_style = gr.Slider(minimum=0, maximum=46, step=1, label="Pose style", value=0) #
|
| 63 |
+
size_of_image = gr.Radio([256, 512], value=256, label='face model resolution', info="use 256/512 model?") #
|
| 64 |
+
preprocess_type = gr.Radio(['crop', 'resize','full', 'extcrop', 'extfull'], value='crop', label='preprocess', info="How to handle input image?")
|
| 65 |
+
is_still_mode = gr.Checkbox(label="Still Mode (fewer hand motion, works with preprocess `full`)")
|
| 66 |
+
batch_size = gr.Slider(label="batch size in generation", step=1, maximum=10, value=2)
|
| 67 |
+
enhancer = gr.Checkbox(label="GFPGAN as Face enhancer")
|
| 68 |
+
submit = gr.Button('Generate', elem_id="sadtalker_generate", variant='primary')
|
| 69 |
+
|
| 70 |
+
with gr.Tabs(elem_id="sadtalker_genearted"):
|
| 71 |
+
gen_video = gr.Video(label="Generated video", format="mp4").style(width=256)
|
| 72 |
+
|
| 73 |
+
if warpfn:
|
| 74 |
+
submit.click(
|
| 75 |
+
fn=warpfn(sad_talker.test),
|
| 76 |
+
inputs=[source_image,
|
| 77 |
+
driven_audio,
|
| 78 |
+
preprocess_type,
|
| 79 |
+
is_still_mode,
|
| 80 |
+
enhancer,
|
| 81 |
+
batch_size,
|
| 82 |
+
size_of_image,
|
| 83 |
+
pose_style
|
| 84 |
+
],
|
| 85 |
+
outputs=[gen_video]
|
| 86 |
+
)
|
| 87 |
+
else:
|
| 88 |
+
submit.click(
|
| 89 |
+
fn=sad_talker.test,
|
| 90 |
+
inputs=[source_image,
|
| 91 |
+
driven_audio,
|
| 92 |
+
preprocess_type,
|
| 93 |
+
is_still_mode,
|
| 94 |
+
enhancer,
|
| 95 |
+
batch_size,
|
| 96 |
+
size_of_image,
|
| 97 |
+
pose_style
|
| 98 |
+
],
|
| 99 |
+
outputs=[gen_video]
|
| 100 |
+
)
|
| 101 |
+
|
| 102 |
+
return sadtalker_interface
|
| 103 |
+
|
| 104 |
+
|
| 105 |
+
if __name__ == "__main__":
|
| 106 |
+
|
| 107 |
+
demo = sadtalker_demo()
|
| 108 |
+
demo.queue()
|
| 109 |
+
demo.launch()
|
| 110 |
+
|
| 111 |
+
|
checkpoints/SadTalker_V0.0.2_256.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c211f5d6de003516bf1bbda9f47049a4c9c99133b1ab565c6961e5af16477bff
|
| 3 |
+
size 725066984
|
checkpoints/SadTalker_V0.0.2_512.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e063f7ff5258240bdb0f7690783a7b1374e6a4a81ce8fa33456f4cd49694340
|
| 3 |
+
size 725066984
|
checkpoints/mapping_00109-model.pth.tar
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:84a8642468a3fcfdd9ab6be955267043116c2bec2284686a5262f1eaf017f64c
|
| 3 |
+
size 155779231
|
checkpoints/mapping_00229-model.pth.tar
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:62a1e06006cc963220f6477438518ed86e9788226c62ae382ddc42fbcefb83f1
|
| 3 |
+
size 155521183
|
cog.yaml
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
build:
|
| 2 |
+
gpu: true
|
| 3 |
+
cuda: "11.3"
|
| 4 |
+
python_version: "3.8"
|
| 5 |
+
system_packages:
|
| 6 |
+
- "ffmpeg"
|
| 7 |
+
- "libgl1-mesa-glx"
|
| 8 |
+
- "libglib2.0-0"
|
| 9 |
+
python_packages:
|
| 10 |
+
- "torch==1.12.1"
|
| 11 |
+
- "torchvision==0.13.1"
|
| 12 |
+
- "torchaudio==0.12.1"
|
| 13 |
+
- "joblib==1.1.0"
|
| 14 |
+
- "scikit-image==0.19.3"
|
| 15 |
+
- "basicsr==1.4.2"
|
| 16 |
+
- "facexlib==0.3.0"
|
| 17 |
+
- "resampy==0.3.1"
|
| 18 |
+
- "pydub==0.25.1"
|
| 19 |
+
- "scipy==1.10.1"
|
| 20 |
+
- "kornia==0.6.8"
|
| 21 |
+
- "face_alignment==1.3.5"
|
| 22 |
+
- "imageio==2.19.3"
|
| 23 |
+
- "imageio-ffmpeg==0.4.7"
|
| 24 |
+
- "librosa==0.9.2" #
|
| 25 |
+
- "tqdm==4.65.0"
|
| 26 |
+
- "yacs==0.1.8"
|
| 27 |
+
- "gfpgan==1.3.8"
|
| 28 |
+
- "dlib-bin==19.24.1"
|
| 29 |
+
- "av==10.0.0"
|
| 30 |
+
- "trimesh==3.9.20"
|
| 31 |
+
run:
|
| 32 |
+
- mkdir -p /root/.cache/torch/hub/checkpoints/ && wget --output-document "/root/.cache/torch/hub/checkpoints/s3fd-619a316812.pth" "https://www.adrianbulat.com/downloads/python-fan/s3fd-619a316812.pth"
|
| 33 |
+
- mkdir -p /root/.cache/torch/hub/checkpoints/ && wget --output-document "/root/.cache/torch/hub/checkpoints/2DFAN4-cd938726ad.zip" "https://www.adrianbulat.com/downloads/python-fan/2DFAN4-cd938726ad.zip"
|
| 34 |
+
|
| 35 |
+
predict: "predict.py:Predictor"
|
docs/FAQ.md
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
## Frequency Asked Question
|
| 3 |
+
|
| 4 |
+
**Q: `ffmpeg` is not recognized as an internal or external command**
|
| 5 |
+
|
| 6 |
+
In Linux, you can install the ffmpeg via `conda install ffmpeg`. Or on Mac OS X, try to install ffmpeg via `brew install ffmpeg`. On windows, make sure you have `ffmpeg` in the `%PATH%` as suggested in [#54](https://github.com/Winfredy/SadTalker/issues/54), then, following [this](https://www.geeksforgeeks.org/how-to-install-ffmpeg-on-windows/) installation to install `ffmpeg`.
|
| 7 |
+
|
| 8 |
+
**Q: Running Requirments.**
|
| 9 |
+
|
| 10 |
+
Please refer to the discussion here: https://github.com/Winfredy/SadTalker/issues/124#issuecomment-1508113989
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
**Q: ModuleNotFoundError: No module named 'ai'**
|
| 14 |
+
|
| 15 |
+
please check the checkpoint's size of the `epoch_20.pth`. (https://github.com/Winfredy/SadTalker/issues/167, https://github.com/Winfredy/SadTalker/issues/113)
|
| 16 |
+
|
| 17 |
+
**Q: Illegal Hardware Error: Mac M1**
|
| 18 |
+
|
| 19 |
+
please reinstall the `dlib` by `pip install dlib` individually. (https://github.com/Winfredy/SadTalker/issues/129, https://github.com/Winfredy/SadTalker/issues/109)
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
**Q: FileNotFoundError: [Errno 2] No such file or directory: checkpoints\BFM_Fitting\similarity_Lm3D_all.mat**
|
| 23 |
+
|
| 24 |
+
Make sure you have downloaded the checkpoints and gfpgan as [here](https://github.com/Winfredy/SadTalker#-2-download-trained-models) and placed them in the right place.
|
| 25 |
+
|
| 26 |
+
**Q: RuntimeError: unexpected EOF, expected 237192 more bytes. The file might be corrupted.**
|
| 27 |
+
|
| 28 |
+
The files are not automatically downloaded. Please update the code and download the gfpgan folders as [here](https://github.com/Winfredy/SadTalker#-2-download-trained-models).
|
| 29 |
+
|
| 30 |
+
**Q: CUDA out of memory error**
|
| 31 |
+
|
| 32 |
+
please refer to https://stackoverflow.com/questions/73747731/runtimeerror-cuda-out-of-memory-how-setting-max-split-size-mb
|
| 33 |
+
|
| 34 |
+
```
|
| 35 |
+
# windows
|
| 36 |
+
set PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:128
|
| 37 |
+
python inference.py ...
|
| 38 |
+
|
| 39 |
+
# linux
|
| 40 |
+
export PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:128
|
| 41 |
+
python inference.py ...
|
| 42 |
+
```
|
| 43 |
+
|
| 44 |
+
**Q: Error while decoding stream #0:0: Invalid data found when processing input [mp3float @ 0000015037628c00] Header missing**
|
| 45 |
+
|
| 46 |
+
Our method only support wav or mp3 files as input, please make sure the feeded audios are in these formats.
|
docs/best_practice.md
ADDED
|
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Best Practice and Tips for configuration
|
| 2 |
+
|
| 3 |
+
> Our model only works on REAL person's photo or the portrait image similar to REAL person. The anime talking head genreation method will be released in future.
|
| 4 |
+
|
| 5 |
+
Advanced confiurations for `inference.py`:
|
| 6 |
+
|
| 7 |
+
| Name | Configuration | default | Explaination |
|
| 8 |
+
|:------------- |:------------- |:----- | :------------- |
|
| 9 |
+
| Enhance Mode | `--enhancer` | None | Using `gfpgan` or `RestoreFormer` to enhance the generated face via face restoration network
|
| 10 |
+
| Background Enhancer | `--background_enhancer` | None | Using `realesrgan` to enhance the full video.
|
| 11 |
+
| Still Mode | ` --still` | False | Using the same pose parameters as the original image, fewer head motion.
|
| 12 |
+
| Expressive Mode | `--expression_scale` | 1.0 | a larger value will make the expression motion stronger.
|
| 13 |
+
| save path | `--result_dir` |`./results` | The file will be save in the newer location.
|
| 14 |
+
| preprocess | `--preprocess` | `crop` | Run and produce the results in the croped input image. Other choices: `resize`, where the images will be resized to the specific resolution. `full` Run the full image animation, use with `--still` to get better results.
|
| 15 |
+
| ref Mode (eye) | `--ref_eyeblink` | None | A video path, where we borrow the eyeblink from this reference video to provide more natural eyebrow movement.
|
| 16 |
+
| ref Mode (pose) | `--ref_pose` | None | A video path, where we borrow the pose from the head reference video.
|
| 17 |
+
| 3D Mode | `--face3dvis` | False | Need additional installation. More details to generate the 3d face can be founded [here](docs/face3d.md).
|
| 18 |
+
| free-view Mode | `--input_yaw`,<br> `--input_pitch`,<br> `--input_roll` | None | Genearting novel view or free-view 4D talking head from a single image. More details can be founded [here](https://github.com/Winfredy/SadTalker#generating-4d-free-view-talking-examples-from-audio-and-a-single-image).
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
### About `--preprocess`
|
| 22 |
+
|
| 23 |
+
Our method automatically handle the input images via `crop`, `resize` and `full`.
|
| 24 |
+
|
| 25 |
+
In `crop` mode, we only generate the croped image via the facial keypoints and generated the facial anime avator. The animation of both expression and head pose are realistic.
|
| 26 |
+
|
| 27 |
+
> still mode will stop the eyeblink and head pose movement.
|
| 28 |
+
|
| 29 |
+
| [input image @bagbag1815](https://twitter.com/bagbag1815/status/1642754319094108161) | crop | crop w/still |
|
| 30 |
+
|:--------------------: |:--------------------: | :----: |
|
| 31 |
+
| <img src='../examples/source_image/full_body_2.png' width='380'> |  |  |
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
In `resize` mode, we resize the whole images to generate the fully talking head video. Thus, an image similar to the ID photo can be produced. ⚠️ It will produce bad results for full person images.
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
| <img src='../examples/source_image/full_body_2.png' width='380'> | <img src='../examples/source_image/full4.jpeg' width='380'> |
|
| 40 |
+
|:--------------------: |:--------------------: |
|
| 41 |
+
| ❌ not suitable for resize mode | ✅ good for resize mode |
|
| 42 |
+
| <img src='resize_no.gif'> | <img src='resize_good.gif' width='380'> |
|
| 43 |
+
|
| 44 |
+
In `full` mode, our model will automatically process the croped region and paste back to the original image. Remember to use `--still` to keep the original head pose.
|
| 45 |
+
|
| 46 |
+
| input | `--still` | `--still` & `enhancer` |
|
| 47 |
+
|:--------------------: |:--------------------: | :--:|
|
| 48 |
+
| <img src='../examples/source_image/full_body_2.png' width='380'> | <img src='./example_full.gif' width='380'> | <img src='./example_full_enhanced.gif' width='380'>
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
### About `--enhancer`
|
| 52 |
+
|
| 53 |
+
For better facial quality, we intergate [gfpgan](https://github.com/TencentARC/GFPGAN) and [real-esrgan](https://github.com/xinntao/Real-ESRGAN) for different purpose. Just adding `--enhancer <gfpgan or RestoreFormer>` or `--background_enhancer <realesrgan>` for the enhancement of the face and the full image.
|
| 54 |
+
|
| 55 |
+
```bash
|
| 56 |
+
# make sure above packages are available:
|
| 57 |
+
pip install gfpgan
|
| 58 |
+
pip install realesrgan
|
| 59 |
+
```
|
| 60 |
+
|
| 61 |
+
### About `--face3dvis`
|
| 62 |
+
|
| 63 |
+
This flag indicate that we can generated the 3d-rendered face and it's 3d facial landmarks. More details can be founded [here](face3d.md).
|
| 64 |
+
|
| 65 |
+
| Input | Animated 3d face |
|
| 66 |
+
|:-------------: | :-------------: |
|
| 67 |
+
| <img src='../examples/source_image/art_0.png' width='200px'> | <video src="https://user-images.githubusercontent.com/4397546/226856847-5a6a0a4d-a5ec-49e2-9b05-3206db65e8e3.mp4"></video> |
|
| 68 |
+
|
| 69 |
+
> Kindly ensure to activate the audio as the default audio playing is incompatible with GitHub.
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
#### reference eye-link mode.
|
| 74 |
+
|
| 75 |
+
| Input, w/ reference video , reference video |
|
| 76 |
+
|:-------------: |
|
| 77 |
+
| |
|
| 78 |
+
| If the reference video is shorter than the input audio, we will loop the reference video .
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
|
| 82 |
+
#### Generating 4D free-view talking examples from audio and a single image
|
| 83 |
+
|
| 84 |
+
We use `input_yaw`, `input_pitch`, `input_roll` to control head pose. For example, `--input_yaw -20 30 10` means the input head yaw degree changes from -20 to 30 and then changes from 30 to 10.
|
| 85 |
+
```bash
|
| 86 |
+
python inference.py --driven_audio <audio.wav> \
|
| 87 |
+
--source_image <video.mp4 or picture.png> \
|
| 88 |
+
--result_dir <a file to store results> \
|
| 89 |
+
--input_yaw -20 30 10
|
| 90 |
+
```
|
| 91 |
+
|
| 92 |
+
| Results, Free-view results, Novel view results |
|
| 93 |
+
|:-------------: |
|
| 94 |
+
| |
|
docs/changlelog.md
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
## changelogs
|
| 2 |
+
|
| 3 |
+
|
| 4 |
+
- __[2023.04.06]__: stable-diffiusion webui extension is release.
|
| 5 |
+
|
| 6 |
+
- __[2023.04.03]__: Enable TTS in huggingface and gradio local demo.
|
| 7 |
+
|
| 8 |
+
- __[2023.03.30]__: Launch beta version of the full body mode.
|
| 9 |
+
|
| 10 |
+
- __[2023.03.30]__: Launch new feature: through using reference videos, our algorithm can generate videos with more natural eye blinking and some eyebrow movement.
|
| 11 |
+
|
| 12 |
+
- __[2023.03.29]__: `resize mode` is online by `python infererence.py --preprocess resize`! Where we can produce a larger crop of the image as discussed in https://github.com/Winfredy/SadTalker/issues/35.
|
| 13 |
+
|
| 14 |
+
- __[2023.03.29]__: local gradio demo is online! `python app.py` to start the demo. New `requirments.txt` is used to avoid the bugs in `librosa`.
|
| 15 |
+
|
| 16 |
+
- __[2023.03.28]__: Online demo is launched in [](https://huggingface.co/spaces/vinthony/SadTalker), thanks AK!
|
| 17 |
+
|
| 18 |
+
- __[2023.03.22]__: Launch new feature: generating the 3d face animation from a single image. New applications about it will be updated.
|
| 19 |
+
|
| 20 |
+
- __[2023.03.22]__: Launch new feature: `still mode`, where only a small head pose will be produced via `python inference.py --still`.
|
| 21 |
+
|
| 22 |
+
- __[2023.03.18]__: Support `expression intensity`, now you can change the intensity of the generated motion: `python inference.py --expression_scale 1.3 (some value > 1)`.
|
| 23 |
+
|
| 24 |
+
- __[2023.03.18]__: Reconfig the data folders, now you can download the checkpoint automatically using `bash scripts/download_models.sh`.
|
| 25 |
+
- __[2023.03.18]__: We have offically integrate the [GFPGAN](https://github.com/TencentARC/GFPGAN) for face enhancement, using `python inference.py --enhancer gfpgan` for better visualization performance.
|
| 26 |
+
- __[2023.03.14]__: Specify the version of package `joblib` to remove the errors in using `librosa`, [](https://colab.research.google.com/github/Winfredy/SadTalker/blob/main/quick_demo.ipynb) is online!
|
| 27 |
+
- __[2023.03.06]__: Solve some bugs in code and errors in installation
|
| 28 |
+
- __[2023.03.03]__: Release the test code for audio-driven single image animation!
|
| 29 |
+
- __[2023.02.28]__: SadTalker has been accepted by CVPR 2023!
|
docs/example_crop.gif
ADDED
|
Git LFS Details
|
docs/example_crop_still.gif
ADDED
|
Git LFS Details
|
docs/example_full.gif
ADDED
|
Git LFS Details
|
docs/example_full_crop.gif
ADDED
|
docs/example_full_enhanced.gif
ADDED
|
Git LFS Details
|
docs/face3d.md
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
## 3D Face visualization
|
| 2 |
+
|
| 3 |
+
We use pytorch3d to visualize the produced 3d face from a single image.
|
| 4 |
+
|
| 5 |
+
Since it is not easy to install, we produce a new install guidence here:
|
| 6 |
+
|
| 7 |
+
```bash
|
| 8 |
+
git clone https://github.com/Winfredy/SadTalker.git
|
| 9 |
+
cd SadTalker
|
| 10 |
+
conda create -n sadtalker3d python=3.8
|
| 11 |
+
source activate sadtalker3d
|
| 12 |
+
|
| 13 |
+
conda install ffmpeg
|
| 14 |
+
conda install -c fvcore -c iopath -c conda-forge fvcore iopath
|
| 15 |
+
conda install libgcc gmp
|
| 16 |
+
|
| 17 |
+
pip install torch==1.11.0+cu113 torchvision==0.12.0+cu113 torchaudio==0.11.0 --extra-index-url https://download.pytorch.org/whl/cu113
|
| 18 |
+
|
| 19 |
+
# insintall pytorch3d
|
| 20 |
+
pip install --no-index --no-cache-dir pytorch3d -f https://dl.fbaipublicfiles.com/pytorch3d/packaging/wheels/py38_cu113_pyt1110/download.html
|
| 21 |
+
|
| 22 |
+
pip install -r requirements3d.txt
|
| 23 |
+
|
| 24 |
+
### install gpfgan for enhancer
|
| 25 |
+
pip install git+https://github.com/TencentARC/GFPGAN
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
### when occurs gcc version problem `from pytorch import _C` from pytorch3d, add the anaconda path to LD_LIBRARY_PATH
|
| 29 |
+
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/$YOUR_ANACONDA_PATH/lib/
|
| 30 |
+
|
| 31 |
+
```
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
Then, generating the result via:
|
| 35 |
+
|
| 36 |
+
```bash
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
python inference.py --driven_audio <audio.wav> \
|
| 40 |
+
--source_image <video.mp4 or picture.png> \
|
| 41 |
+
--result_dir <a file to store results> \
|
| 42 |
+
--face3dvis
|
| 43 |
+
|
| 44 |
+
```
|
| 45 |
+
|
| 46 |
+
Then, the results will be given in the folders with the file name of `face3d.mp4`.
|
| 47 |
+
|
| 48 |
+
More applications about 3d face will be released.
|
docs/free_view_result.gif
ADDED
|
Git LFS Details
|
docs/install.md
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
### Mac (Tested on M1 Mac OS 13.3)
|
| 3 |
+
|
| 4 |
+
```
|
| 5 |
+
git clone https://github.com/Winfredy/SadTalker.git
|
| 6 |
+
|
| 7 |
+
cd SadTalker
|
| 8 |
+
|
| 9 |
+
conda create -n sadtalker python=3.8
|
| 10 |
+
|
| 11 |
+
conda activate sadtalker
|
| 12 |
+
|
| 13 |
+
# install pytorch 2.0
|
| 14 |
+
pip install torch torchvision torchaudio
|
| 15 |
+
|
| 16 |
+
conda install ffmpeg
|
| 17 |
+
|
| 18 |
+
pip install -r requirements.txt
|
| 19 |
+
|
| 20 |
+
pip install dlib # mac need to install the original dlib.
|
| 21 |
+
|
| 22 |
+
```
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
### Windows Native
|
| 27 |
+
|
| 28 |
+
- Make sure you have `ffmpeg` in the `%PATH%` as suggested in [#54](https://github.com/Winfredy/SadTalker/issues/54), following [this](https://www.geeksforgeeks.org/how-to-install-ffmpeg-on-windows/) installation to install `ffmpeg`.
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
### Windows WSL
|
| 32 |
+
- Make sure the environment: `export LD_LIBRARY_PATH=/usr/lib/wsl/lib:$LD_LIBRARY_PATH`
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
### Docker installnation
|
| 36 |
+
|
| 37 |
+
A dockerfile are also provided by [@thegenerativegeneration](https://github.com/thegenerativegeneration) in [docker hub](https://hub.docker.com/repository/docker/wawa9000/sadtalker), which can be used directly as:
|
| 38 |
+
|
| 39 |
+
```bash
|
| 40 |
+
docker run --gpus "all" --rm -v $(pwd):/host_dir wawa9000/sadtalker \
|
| 41 |
+
--driven_audio /host_dir/deyu.wav \
|
| 42 |
+
--source_image /host_dir/image.jpg \
|
| 43 |
+
--expression_scale 1.0 \
|
| 44 |
+
--still \
|
| 45 |
+
--result_dir /host_dir
|
| 46 |
+
```
|
| 47 |
+
|
docs/resize_good.gif
ADDED
|
Git LFS Details
|
docs/resize_no.gif
ADDED
|
Git LFS Details
|
docs/sadtalker_logo.png
ADDED
|
docs/using_ref_video.gif
ADDED
|
Git LFS Details
|
docs/webui_extension.md
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
## Run SadTalker as a Stable Diffusion WebUI Extension.
|
| 3 |
+
|
| 4 |
+
1. Installing the lastest version of [stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) and install the sadtalker via `extension`.
|
| 5 |
+
<img width="726" alt="image" src="https://user-images.githubusercontent.com/4397546/230698519-267d1d1f-6e99-4dd4-81e1-7b889259efbd.png">
|
| 6 |
+
|
| 7 |
+
2. Download the checkpoints manually, for Linux and Mac:
|
| 8 |
+
|
| 9 |
+
```bash
|
| 10 |
+
|
| 11 |
+
cd SOMEWHERE_YOU_LIKE
|
| 12 |
+
|
| 13 |
+
bash <(wget -qO- https://raw.githubusercontent.com/Winfredy/SadTalker/main/scripts/download_models.sh)
|
| 14 |
+
```
|
| 15 |
+
|
| 16 |
+
For windows, you can download all the checkpoints from [google drive](https://drive.google.com/drive/folders/1Wd88VDoLhVzYsQ30_qDVluQr_Xm46yHT?usp=sharing) or [百度云盘](https://pan.baidu.com/s/1nXuVNd0exUl37ISwWqbFGA?pwd=sadt) 提取码: sadt.
|
| 17 |
+
|
| 18 |
+
3.1. options 1: put the checkpoint in `stable-diffusion-webui/models/SadTalker` or `stable-diffusion-webui/extensions/SadTalker/checkpoints/`, the checkpoints will be detected automatically.
|
| 19 |
+
|
| 20 |
+
3.2. Options 2: Set the path of `SADTALKTER_CHECKPOINTS` in `webui_user.sh`(linux) or `webui_user.bat`(windows) by:
|
| 21 |
+
|
| 22 |
+
> only works if you are directly starting webui from `webui_user.sh` or `webui_user.bat`.
|
| 23 |
+
|
| 24 |
+
```bash
|
| 25 |
+
# windows (webui_user.bat)
|
| 26 |
+
set SADTALKER_CHECKPOINTS=D:\SadTalker\checkpoints
|
| 27 |
+
|
| 28 |
+
# linux (webui_user.sh)
|
| 29 |
+
export SADTALKER_CHECKPOINTS=/path/to/SadTalker/checkpoints
|
| 30 |
+
```
|
| 31 |
+
|
| 32 |
+
4. Then, starting the webui via `webui.sh or webui_user.sh(linux)` or `webui_user.bat(windows)` or any other methods, the SadTalker can be used in stable-diffusion-webui directly.
|
| 33 |
+
|
| 34 |
+
<img width="726" alt="image" src="https://user-images.githubusercontent.com/4397546/230698614-58015182-2916-4240-b324-e69022ef75b3.png">
|
| 35 |
+
|
| 36 |
+
## Questsions
|
| 37 |
+
|
| 38 |
+
1. if you are running on CPU, you need to specific `--disable-safe-unpickle` in `webui_user.sh` or `webui_user.bat`.
|
| 39 |
+
|
| 40 |
+
```bash
|
| 41 |
+
# windows (webui_user.bat)
|
| 42 |
+
set COMMANDLINE_ARGS="--disable-safe-unpickle"
|
| 43 |
+
|
| 44 |
+
# linux (webui_user.sh)
|
| 45 |
+
export COMMANDLINE_ARGS="--disable-safe-unpickle"
|
| 46 |
+
```
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
(Some [important discussion](https://github.com/Winfredy/SadTalker/issues/78) if you are unable to use `full` mode).
|
examples/driven_audio/RD_Radio31_000.wav
ADDED
|
Binary file (512 kB). View file
|
|
|
examples/driven_audio/RD_Radio34_002.wav
ADDED
|
Binary file (512 kB). View file
|
|
|
examples/driven_audio/RD_Radio36_000.wav
ADDED
|
Binary file (512 kB). View file
|
|
|
examples/driven_audio/RD_Radio40_000.wav
ADDED
|
Binary file (512 kB). View file
|
|
|
examples/driven_audio/bus_chinese.wav
ADDED
|
Binary file (652 kB). View file
|
|
|
examples/driven_audio/chinese_news.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7b0f4d313a1ca671bc4831d60bcf0c12225efbffe6c0e93e54fbfe9bcd4021cb
|
| 3 |
+
size 1536078
|
examples/driven_audio/chinese_poem1.wav
ADDED
|
Binary file (263 kB). View file
|
|
|
examples/driven_audio/chinese_poem2.wav
ADDED
|
Binary file (461 kB). View file
|
|
|
examples/driven_audio/deyu.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba1839c57770a2ab0b593ce814344bfd4d750da02acc9be9e8cf5b9113a0f88a
|
| 3 |
+
size 2694784
|
examples/driven_audio/eluosi.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4a3593815dc7b68c256672baa61934c9479efa770af2065fb0886f02713606e
|
| 3 |
+
size 1786672
|
examples/driven_audio/fayu.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:16ebd13626ae4171030b4ea05cceef06078483c352e4b68d469fc2a52bfffceb
|
| 3 |
+
size 1940428
|
examples/driven_audio/imagine.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2db410217e074d91ae6011e1c5dc0b94f02d05d381c50af8e54253eeacad17d2
|
| 3 |
+
size 1618510
|
examples/driven_audio/itosinger1.wav
ADDED
|
Binary file (500 kB). View file
|
|
|
examples/driven_audio/japanese.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3db5426d0b158799e2be4f609b11f75bfbd4affffe18e9a1c8e6f241fcdedcfc
|
| 3 |
+
size 2622712
|
examples/ref_video/WDA_AlexandriaOcasioCortez_000.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a85242c3fc4d50e2202cea393b9e7ee59019759b68e78e26a254d528c22615a7
|
| 3 |
+
size 2257667
|
examples/ref_video/WDA_KatieHill_000.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1fbb4cfd64eedc49b170c441714a9c4fd5e2c2f8a11592070ad89fbd257f2817
|
| 3 |
+
size 3548230
|
examples/source_image/art_0.png
ADDED
|
examples/source_image/art_1.png
ADDED
|
examples/source_image/art_10.png
ADDED
|
examples/source_image/art_11.png
ADDED
|
examples/source_image/art_12.png
ADDED
|
examples/source_image/art_13.png
ADDED
|