Skip to content

ProgrammerManstein/FACIAL

Repository files navigation

FACIAL

part of digital human, synthesis of head poses and expression.
You can recurrent the project by using the jupter notebook facial_train_install.ipynb. We need mtcnn to extract image facial features, and Openface to extract video facial features. And please unzip BFM_model_front. You can train the module by using the jupter notebook facial_train.ipynb after having recurrented it.
After training, you can get any video that is instructed by your text by using the jupter notebook facial_test.ipynb.
You can also consult the original project FACIAL: Synthesizing Dynamic Talking Face With Implicit Attribute Learning. ICCV, 2021. for more information.

Requirements

Python environment

conda create -n audio_face  
conda activate audio_face  

ffmpeg

sudo apt-get install ffmpeg  

python packages

pip install -r requirements.txt  

you may add opencv by conda.

conda install opencv  

Citation

@inproceedings{zhang2021facial,
  title={FACIAL: Synthesizing Dynamic Talking Face with Implicit Attribute Learning},
  author={Zhang, Chenxu and Zhao, Yifan and Huang, Yifei and Zeng, Ming and Ni, Saifeng and Budagavi, Madhukar and Guo, Xiaohu},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
  pages={3867--3876},
  year={2021}
}
Acknowledgments
We use Deep3DFaceReconstruction for face reconstruction, DeepSpeech and VOCA for audio feature extraction, and 3dface for face rendering. Rendering-to-video module borrows heavily from everybody-dance-now.

About

part of digital human, synthesis of head poses and expression.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages