Skip to content

Official PyTorch implementation for the paper Generalizable Face Landmarking Guided by Conditional Face Warping (CVPR 2024).

Notifications You must be signed in to change notification settings

plustwo0/generalized-face-landmarker

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Generalizable Face Landmarking Guided by Conditional Face Warping (CVPR 2024)

This is the official repository for the following paper:

Generalizable Face Landmarking Guided by Conditional Face Warping [paper] [arxiv] [project page]

Jiayi Liang*, Haotian Liu*, Hongteng Xu, Dixin Luo
Accepted by CVPR 2024.

Scheme

Install

pip install -r requirements.txt

Model

Our proposed framework mainly contains two parts: face warper and landmark detector. They are trained in an alternative optimization framework.

  • The face warper aims to deform real human faces according to stylized facial images, generating warped faces and corresponding warping fields.
  • The face landmarker, as the key module of the warper, predicts the ending points of the warping fields and thus provides us with pseudo landmarks for the stylized facial images.

In our implmentation, we employ SLPT as our backbone and locate the model in Landmark2 folder. For the reproduction on other detectors, substitute the Landmark2 folder with target model and make modifications in train.py.

Data Preparation

Source Domain

Download images and annotations of 300-W from ibug.

We select frontal faces from the trainset of 300W as our training data, and list of image path 300W_frontal_train_list.txt can be downloaded from Google Drive.

Your directory should be like:

  Dataset
  │
  └──300W
     │
     └───300W_frontal_train_list.txt
     └───frontal_train
         └───261068_1.jpg
         │
         └───...
     └───frontal_train_label
         └───261068_1.jpg.npy
         │
         └───...
     └───test_list.txt
     └───test_list_common.txt
     └───test_list_challenge.txt
     └───lfpw
         └───trainset
         └───testset
             └───image_0001.png
             └───image_0001.pts
             │
             └───...
     └───helen
         │
         └───...
     └───ibug
         │
         └───...

Target Domain

  • Download CariFace according to CariFace Dataset. The split of training and testing set can be downloaded from Google Drive.

  • Other domains like Artistic-Faces can be retrieved from Link. The split of training and testing set can be downloaded from Google Drive.

Please note that ArtiFace contains a total of 160 images. Under the GZSL (Unseen ArtiFace) scenario, the test set size of ArtiFace is 160 images; under the DA (300W->ArtiFace) and GZSL (Unseen CariFace) scenarios, the test set of ArtiFace only contains 32 images.
Therefore, please modify the specified test set list in Artistic.py to test_list.txt (32 images) or test_list_all.txt (160 images) according to the different circumstances.

Your directory should be like:

  Dataset
  │
  └──CariFace_dataset
     │
     └───images
         └───00005.jpg
         │
         └───...
     └───landmarks
         └───00005.jpg.npy
         │
         └───...
     └───train_list.txt
     └───test_list.txt
  │
  └──AF_dataset
     │
     └───images
         └───0.png
         │
         └───...
     └───landmarks
         └───0.png.npy
         │
         └───...
     └───train_list.txt
     └───test_list.txt
     └───test_list_all.txt

Train

Load Pretrained Model

Download source-pretrained weights model_best.pth from Google Drive and move it into folder Landmark2.

Training Begin!

python train.py --src_data path/to/source/data --tgt_data path/to/target/data --pretrain_path path/to/pretrained/checkpoint

Inference

Download our model and test on the CariFace by running:

python test.py --checkpoint path/to/model/weights

Further, to test on ArtiFace, download checkpoint and inference:

python test_Artistic.py --checkpoint path/to/model/weights

Citation

If our work is helpful for your research, please cite our paper:

@InProceedings{Liang_2024_CVPR,
    author    = {Liang, Jiayi and Liu, Haotian and Xu, Hongteng and Luo, Dixin},
    title     = {Generalizable Face Landmarking Guided by Conditional Face Warping},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2024},
    pages     = {2425-2435}
}

About

Official PyTorch implementation for the paper Generalizable Face Landmarking Guided by Conditional Face Warping (CVPR 2024).

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%