A jittor implementation for ECCV2022 paper DFRF "Learning Dynamic Facial Radiance Fields for Few-Shot Talking Head Synthesis". The code is based on the authors' Pytorch implementation here. To see our results, we provide some video demos here.
First, install the requirements following install_script.
We conduct the experiments with a 24G RTX3090.
-
Download 3DMM model from Basel Face Model 2009:
cp 01_MorphableModel.mat data_util/face_tracking/3DMM/ cd data_util/face_tracking python convert_BFM.py
Put the video ${id}.mp4 to dataset/vids/, then run the following command for data preprocess.
sh process_data.sh ${id}
Specially, to process data for training base model:
- Download the videos for training base model provided by the author here.
- Put the videos to
dataset/vids/. - Run the following command for data preprocess.
sh process_base_train_data.sh 0
sh process_base_train_data.sh 1
sh process_base_train_data.sh 2
To train on a new dataset for fine-tuning, run the following command. Some pre-trained models are here.
sh run.sh ${id}
Specially, to train the base model:
sh train_base.sh
Change the configurations in the rendering.sh, including the iters, names, datasets, near and far.
sh rendering.sh
This code is built upon the publicly available code DFRF. Thanks the authors of DFRF for making their excellent work and codes publicly available.