Families In the Wild: A Kinship Recognition Toolbox. Visit FIW project page to download and learn more!
https://web.northeastern.edu/smilelab/fiw/
Version 1.0
This document must be completed, i.e., work in progress, as is the API itself.
This API serves as the main codebase for the kinship effort with the FIW database. In addition, below is a detailed description of the database (i.e., data and label) structure.
This documentation describes FIW DB and (working) development kit. This is a work in progress (i.e., still to come are FIW-CNN models, updated benchmarks, more in README (this), and more).
Check out FIW project page
Download here
See publications below. A more complete list of references can be found here
@unknown{robinson2021families,
author = {Robinson, Joseph and Qin, Can and Shao, Ming and Turk, Matthew and Chellappa, Rama and Fu, Yun},
year = {2021},
month = {10},
pages = {},
title = {Recognizing Families In the Wild (RFIW): The 5th Edition},
doi = {10.48550/arXiv.2111.00598}
}
@ARTICLE{8337841,
author={Robinson, Joseph P. and Shao, Ming and Wu, Yue and Liu, Hongfu and Gillis, Timothy and Fu, Yun},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Visual Kinship Recognition of Families in the Wild},
year={2018},
volume={40},
number={11},
pages={2624-2637},
keywords={Labeling;Visualization;Machine learning;Benchmark testing;Databases;Task analysis;Face recognition;Large-scale image dataset;kinship verification;family classification;semi-supervised clustering;deep learning},
doi={10.1109/TPAMI.2018.2826549}}
@InProceedings{kinFG2017,
author = "Wang, Shuyang and Robinson, Joseph P and Fu, Yun",
title = "Kinship Verification on Families In The Wild with Marginalized Denoising Metric Learning",
booktitle = "Automatic Face and Gesture Recognition (FG), 2017 12th IEEE International Conference and Workshops on",
year = "2017",
}
@InProceedings{robinson2016families,
author = "Robinson, Joseph P. and Shao, Ming and Wu, Yue and Fu, Yun",
title = "Families In the Wild (FIW): Large-Scale Kinship Image Database and Benchmarks",
booktitle = "Proceedings of the 2016 ACM on Multimedia Conference",
pages = "242--246",
publisher = "ACM",
year = "2016"
}
The following provides a comprehensive survey on visual kinship recognition, with details on SOA and experimental protocols.
@ARTICLE{9367013,
author={Robinson, Joseph P. and Shao, Ming and Fu, Yun},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Survey on the Analysis and Modeling of Visual Kinship: A Decade in the Making},
year={2022},
volume={44},
number={8},
pages={4432-4453},
keywords={Visualization;Face recognition;Task analysis;Measurement;Tutorials;Protocols;Streaming media;Survey;facial recognition;benchmarks and evaluation;deep learning;data challenges;visual kinship recognition},
doi={10.1109/TPAMI.2021.3063078}}
-
FIW_PIDs.csv: Photo lookup table. Each row is an image instance containing the following fields:
- PID: Photo ID
- Name: Surname.firstName (root reference for given family)
- URL: Photo URL on web
- Metadata: Text caption for photo
-
FIW_FIDs.csv: FID (family)/ Surname lookup table.
- FID: Unique ID key assigned to each family.
- Surname: Family Name corresponding to FID key.
-
FIW_RIDs.csv: Relationship lookup table with keys [1-9] assigned to relationship types.
-
FIDs/
- FID####/ Contains labels and cropped facial images for members of the family (1-1000)
-
MID#/ Face images of family member with ID key , i.e., MID #.
-
F####.csv: File containing member information of each family:
- relationships matrix representing relationship
- names First name of family member.
- gender ender of family member
-
- FID####/ Contains labels and cropped facial images for members of the family (1-1000)
For example:
FID0001.csv
0 1 2 3 Name Gender
1 0 4 5 name1 female
2 1 0 1 name2 female
3 5 4 0 name3 male
Here, we have three family members, as seen by the MIDs across columns and down rows.
We can see that MID1 is related to MID2 by 4->1 (Parent->Sibling), which, of course, can be viewed as the inverse, i.e., MID2->MID1 is 1->4. It can also be seen that MID1 and MID3 are Spouses of one another, i.e., 5->5. And so on.
├── LICENSE
├── README.md <- The top-level README for developers using this project.
├── data
│ ├── external <- Data from third party sources.
│ ├── interim <- Intermediate data that has been transformed.
│ ├── processed <- The final, canonical data sets for modeling.
│ └── raw <- The original, immutable data dump.
│
├── docs <- A default Sphinx project; see sphinx-doc.org for details
│
├── models <- Trained and serialized models, model predictions, or model summaries
│
├── notebooks <- Jupyter notebooks.
│
├── references <- Data dictionaries, manuals, and all other explanatory materials.
│
├── reports <- Generated analysis as HTML, PDF, LaTeX, etc.
│ └── figures <- Generated graphics and figures to be used in reporting
│
├── requirements.txt <- The requirements file for reproducing the analysis environment, e.g.
│ generated with `pip freeze > requirements.txt`
│
├── src <- Source code for use in this project.
│ ├── __init__.py <- Makes src a Python module
│ │
│ ├── data <- Scripts to download or generate data
│ │ └── make_dataset.py
│ │
│ ├── features <- Scripts to turn raw data into features for modeling
│ │ └── build_features.py
│ │
│ ├── models <- Scripts to train models and then use trained models to make
│ │ │ predictions
│ │ ├── predict_model.py
│ │ └── train_model.py
│ │
│ └── visualization <- Scripts to create exploratory and results oriented visualizations
│ └── visualize.py
│
└── <OPEN> <- TBD
- Feature Extraction Module
- Metric Learning Module
- Fine-tune module
- Eval module
- End-to-end scripts demoing usage of modules
- VGG Center-Face
- ResNet with Metric Learning
- ResNet Triplet-loss?
- Cross Dataset Eval
- Features from different layers of VGG
By downloading the image data, you agree to the following terms:
- You will use the data only for non-commercial research and educational purposes.
- You won't be able to distribute the above images.
- Northeastern University makes no representations or warranties regarding the data, including but not limited to warranties of non-infringement or fitness for a particular purpose.
- You accept full responsibility for your use of the data. You shall defend and indemnify Northeastern University, including its employees, officers, and agents, against any claims arising from your use of the data, including but not limited to your use of any copies of copyrighted images that you may create from the data.
Please get in touch with Joseph Robinson (robinson.jo@husky.neu.edu) for questions, comments, or bug reports.