-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A robust kinship verification scheme using face age transformation #59
Comments
OverviewThis paper, published in 2023, aims to solidify and continue the author’s work, that started in “A Cross-age Kinship Verification Scheme Using Face Age Transfer Model, 2022 IEEE”. They present a kinship verification scheme that involves generating face images from various age groups to make more robust training in order to achieve better kinship accuracy. They do so by basically generating this images, extracting unique facial features that do not change even with age changes from it, and train the kinship verification model using these features. The main difference between the two papers is that in the first one, they use HRFAE as the age transformation model, and in this paper, they create their own. Face Age Transformation ModelThere are four main components in the face transformation model: Age encoder, Age classifier, Generator and Descriminator. The age encoder is a simple FC layer with sigmoid. The age classifier is a pretrained VGG-16 network architecture finetuned on IMDB-WIKI dataset. The generator itself has an encoder and a decoder: the enconder which contains three convolution layers and four residual blocks, extracts the encoded feature map from the image. The decoder network, which contains two nearest-neighbor up-sampling layers and three convolution layers, decodes the encoded feature map back into the facial image of the target age. The discriminator consists of six convolution layers, with batch normalization and LeakyReLU, except in the first convolution layer. Loss functions
Kinship Verification ModelThis model consists of two networks: Facial feature extractor that extracts facial features from the generated image groups and a classifier that determines whether they are kin or not. For feature extraction, the authors finetune the Inception Resnet model using FIW. In the training process, triplet loss is used to ensure that kin-related images are closer together in the embedding space. Authors also report that in training, they use hard samples to increase learning efficiency, which means the hard samples in the positive set are far from the anchor, whereas those in the negative set are near the anchor. As for classification, they construct a network with convolution and residual blocks. Based on these blocks, the fully connected and softmax function layers perform kinship classification and calculate the cross-entropy loss. Experiments and ResultsTo train and evaluate the age transformation model, the authors train it on Cross-Age Celebrity Dataset (CACD) which contains 163,446 face images of 2000 celebrities with their respective age. For the kinship verification, the model was trained on FIW, as described above, and evaluated on KinFaceW-I and KinFaceW-II using 5-fold cross validation. Testing the impact of the strategy proposed, it is clear that aging effect created by training with face age transformed images is relevant, even if the overall result is not SOTA ( this one belongs to #68 ). Looking at the ROC curves, all kinship types show improvement with the use of aging, specially Father-Son, which indicates that this method is probably effective on diminishing the effects of age in kinship verification and might be useful for building age-invariant kinship models. |
Bom resumo, @matheuslevi11. Curioso para saber mais detalhes sobre o modelo. Muito bom que é recente. Há código? Estranho usarem backbones antigos, como VGG-16, sendo que os últimos trabalhos de kinship usam ArcFace ou AdaFace. Outra coisa "chata" é não haver avaliação dos resultados no próprio FIW... Enquanto isso, no KFW há resultados claros, mas não chegam a ser SOTA. Fico pensando no quanto a perda contrastiva ajudaria aqui. Além disso, o que um sampler adaptado, similar ao proposto em #80, traria de resultados. Nos meus experimentos, ainda a serem registrados aqui, o uso do sampler melhorou significativamente o baseline. E se fosse adaptado para considerar também a idade? Fica a ideia para avaliarmos em breve. |
Infelizmente não há código disponível. Apesar de no artigo conter informações detalhadas sobre a arquitetura dos modelos e detalhes de implementação como batch size e learning rate, não sei estimar com precisão se seria fácil reproduzi-lo. Também gostaria de ter visto resultados no FIW, acredito que seria interessante. Com relação a perda contrastiva, eu fiquei pensando nisso também, já que eles utilizam triplet loss e tem bons resultados. Quanto aos outros questionamentos, eu acho a ideia promissora, é a direção que estou indo no TCC atualmente, na sexta podemos falar mais sobre isso. |
Realizei umas pesquisas no assunto e encontrei um paper interessante The paper "Synthetic Face Ageing: Evaluation, Analysis and Facilitation of Age-Robust Facial Recognition Algorithms" by Wang Yao et al. focuses on enhancing age-invariant face recognition (AIFR) systems using synthetic ageing data. Key points include:
Pertinente também a existência de outro dataset, que não foi usado pelo paper dessa issue, o B3FD:
|
No description provided.
The text was updated successfully, but these errors were encountered: