HunyuanPortrait is a diffusion-based framework for generating lifelike, temporally consistent portrait animations by decoupling identity and motion using pre-trained encoders. It encodes driving video expressions/poses into implicit control signals, injects them via attention-based adapters into a stabilized diffusion backbone, enabling detailed and style-flexible animation from a single reference image. The method outperforms existing approaches in controllability and coherence.
Some results of portrait animation using HunyuanPortrait.
More results can be found on our Project page.
f1.mp4 |
f2.mp4 |
f4.mp4 |
f3.mp4 |
sing_1.mp4
act_1.mp4
emoji_1.mp4
If you think this project is helpful, please feel free to leave a star⭐️⭐️⭐️ and cite our paper:
@inproceedings{xu2025hunyuanportrait,
title={Hunyuanportrait: Implicit condition control for enhanced portrait animation},
author={Xu, Zunnan and Yu, Zhentao and Zhou, Zixiang and Zhou, Jun and Jin, Xiaoyu and Hong, Fa-Ting and Ji, Xiaozhong and Zhu, Junwei and Cai, Chengfei and Tang, Shiyu and others},
booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
pages={15909--15919},
year={2025}
}