🚀 LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control
LivePortrait is an official PyTorch implementation for portrait animation. It supports various features like portrait video editing, pose editing, and more, with continuous updates and improvements.
🚀 Quick Start
1. Clone the code and prepare the environment
git clone https://github.com/KwaiVGI/LivePortrait
cd LivePortrait
conda create -n LivePortrait python==3.9
conda activate LivePortrait
pip install -r requirements.txt
pip install -r requirements_macOS.txt
⚠️ Important Note
Make sure your system has FFmpeg installed, including both ffmpeg
and ffprobe
!
2. Download pretrained weights
The easiest way to download the pretrained weights is from HuggingFace:
git lfs install
git clone https://huggingface.co/KwaiVGI/LivePortrait temp_pretrained_weights
mv temp_pretrained_weights/* pretrained_weights/
rm -rf temp_pretrained_weights
Alternatively, you can download all pretrained weights from Google Drive or Baidu Yun. Unzip and place them in ./pretrained_weights
.
Ensuring the directory structure is as follows, or contains:
pretrained_weights
├── insightface
│ └── models
│ └── buffalo_l
│ ├── 2d106det.onnx
│ └── det_10g.onnx
└── liveportrait
├── base_models
│ ├── appearance_feature_extractor.pth
│ ├── motion_extractor.pth
│ ├── spade_generator.pth
│ └── warping_module.pth
├── landmark.onnx
└── retargeting_models
└── stitching_retargeting_module.pth
3. Inference 🚀
Basic Usage
python inference.py
PYTORCH_ENABLE_MPS_FALLBACK=1 python inference.py
If the script runs successfully, you will get an output mp4 file named animations/s6--d0_concat.mp4
. This file includes the following results: driving video, input image or video, and generated result.
Advanced Usage
You can change the input by specifying the -s
and -d
arguments:
python inference.py -s assets/examples/source/s9.jpg -d assets/examples/driving/d0.mp4
python inference.py -s assets/examples/source/s13.mp4 -d assets/examples/driving/d0.mp4
python inference.py -h
Driving video auto-cropping 📢📢📢
To use your own driving video, we recommend: ⬇️
- Crop it to a 1:1 aspect ratio (e.g., 512x512 or 256x256 pixels), or enable auto-cropping by
--flag_crop_driving_video
.
- Focus on the head area, similar to the example videos.
- Minimize shoulder movement.
- Make sure the first frame of driving video is a frontal face with neutral expression.
Below is a auto-cropping case by --flag_crop_driving_video
:
python inference.py -s assets/examples/source/s9.jpg -d assets/examples/driving/d13.mp4 --flag_crop_driving_video
If you find the results of auto-cropping is not well, you can modify the --scale_crop_driving_video
, --vy_ratio_crop_driving_video
options to adjust the scale and offset, or do it manually.
Motion template making
You can also use the auto-generated motion template files ending with .pkl
to speed up inference, and protect privacy, such as:
python inference.py -s assets/examples/source/s9.jpg -d assets/examples/driving/d5.pkl
python inference.py -s assets/examples/source/s13.mp4 -d assets/examples/driving/d5.pkl
4. Gradio interface 🤗
We also provide a Gradio
interface for a better experience, just run by:
python app.py
PYTORCH_ENABLE_MPS_FALLBACK=1 python app.py
You can specify the --server_port
, --share
, --server_name
arguments to satisfy your needs!
🚀 We also provide an acceleration option --flag_do_torch_compile
. The first-time inference triggers an optimization process (about one minute), making subsequent inferences 20-30% faster. Performance gains may vary with different CUDA versions.
python app.py --flag_do_torch_compile
⚠️ Important Note
This method is not supported on Windows and macOS.
Or, try it out effortlessly on HuggingFace 🤗
5. Inference speed evaluation 🚀🚀🚀
We have also provided a script to evaluate the inference speed of each module:
python speed.py
Below are the results of inferring one frame on an RTX 4090 GPU using the native PyTorch framework with torch.compile
:
Property |
Details |
Appearance Feature Extractor |
Parameters: 0.84M, Model Size: 3.3MB, Inference: 0.82ms |
Motion Extractor |
Parameters: 28.12M, Model Size: 108MB, Inference: 0.84ms |
Spade Generator |
Parameters: 55.37M, Model Size: 212MB, Inference: 7.59ms |
Warping Module |
Parameters: 45.53M, Model Size: 174MB, Inference: 5.21ms |
Stitching and Retargeting Modules |
Parameters: 0.23M, Model Size: 2.3MB, Inference: 0.31ms |
Note: The values for the Stitching and Retargeting Modules represent the combined parameter counts and total inference time of three sequential MLP networks.
✨ Features
This repo, named LivePortrait, contains the official PyTorch implementation of our paper LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control.
We are actively updating and improving this repository. If you find any bugs or have suggestions, welcome to raise issues or submit pull requests (PR) 💖.
📚 Documentation
Updates
2024/08/02
: 😸 We released a version of the Animals model, along with several other updates and improvements. Check out the details here!
2024/07/25
: 📦 Windows users can now download the package from HuggingFace or BaiduYun. Simply unzip and double-click run_windows.bat
to enjoy!
2024/07/24
: 🎨 We support pose editing for source portraits in the Gradio interface. We’ve also lowered the default detection threshold to increase recall. Have fun!
2024/07/19
: ✨ We support 🎞️ portrait video editing (aka v2v)! More to see here.
2024/07/17
: 🍎 We support macOS with Apple Silicon, modified from jeethu's PR #143.
2024/07/10
: 💪 We support audio and video concatenating, driving video auto-cropping, and template making to protect privacy. More to see here.
2024/07/09
: 🤗 We released the HuggingFace Space, thanks to the HF team and Gradio!
2024/07/04
: 😊 We released the initial version of the inference code and models. Continuous updates, stay tuned!
2024/07/04
: 🔥 We released the homepage and technical report on arXiv.
Community Resources
Discover the invaluable resources contributed by our community to enhance your LivePortrait experience:
And many more amazing contributions from our community!
Acknowledgements
We would like to thank the contributors of FOMM, Open Facevid2vid, SPADE, InsightFace repositories, for their open research and contributions.
Citation
If you find LivePortrait useful for your research, welcome to 🌟 this repo and cite our work using the following BibTeX:
@article{guo2024liveportrait,
title = {LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control},
author = {Guo, Jianzhu and Zhang, Dingyun and Liu, Xiaoqiang and Zhong, Zhizhou and Zhang, Yuan and Wan, Pengfei and Zhang, Di},
journal = {arXiv preprint arXiv:2407.03168},
year = {2024}
}
Long live in arXiv.
Contact
Jianzhu Guo (郭建珠); guojianzhu1994@gmail.com
📄 License
This project is licensed under the MIT license.