by Takahiro Ishiguro and Alan lin
This repository contains the pytorch implementation emotional talking face generation prefesented in paper : Emotional Talking Facial Video Generation using Single Image and Sentence.
Our published paper can be found here.
Python 3.6
- ffmpeg:
sudo apt-get install ffmpeg
- Install necessary packages using
pip install -r requirements.txt
. - ( Weights & Biases for experiment tracking )
.
├── docs # Documentation files
│ └── evaluation.md
├── TalkingFaceGeneration # Source files
│ ├── model
│ ├── data
│ ├── DeepFaceLab
│ ├── StyleGAN
│ ├── dnnlib # Helper functions for StyleGAN generator
│ ├── filelists # describe the directory of train & validation datasets
│ ├── face_detection # module used to crop face image
│ ├── preprocess.py
│ ├── train.py
│ └── inference.py
├── demo # demo notebooks
│ ├── DeepFaceLab.ipynb # notebook for deepfacelab
│ └── SPT-Wav2Lip.ipynb # notebook for SPT-Wav2Lip with Google TTS
├── LICENSE
├── README.md
└── requirement.txt
image_to_video.mp4
video_to_video.mp4
EmoFaceGeneration is released under the Apache 2.0 license.
This code borrows heavily from Wav2Lip, StyleGAN and DeepFaceLab. We thank the authors for releasing their models and codebase. We also like to thank BBC for allowing us to use thier VoxCeleb2 dataset.