Skip to content

ece324-2019/FearNet

Repository files navigation

FearNet

FearNet is an image classification model capable of identifying images likely to trigger phobias. The architecture leverages 8 transfer-learned pre-trained models ensembled to feed into a MLP architecture. The input is any image (re-sized to 128x128). The output is a 17-long tensor, w/ each element describing the probability of one of the 16 phobias (1 class for no phobia) being detected in the input image.

Metrics

Recall/ Sensitivity is used to gauge the performance of the model due to the significance of false negatives over false positives in the classification problem.

Transfer Model Experimentation (Torchvision):

Hyper Params: 128x128img, LR = 0.001, Epoch = 20, Batch Size = 64, No BN (unless model inherent, i.e. VGG)

Baseline: 13.67%
3-DCNN Ensemble: 14.48%
Resnet152 (1 unfrozen fc layer): ~69%
Resnet152 (1 unfrozen fc layer + 1 ext. fc layer): ~75%
Resnet152 (1 unfrozen fc layer + 2 ext. fc layers): ~75%
VGG19_BN (1 unfrozen fc layer): ~73%
VGG19_BN (1 unfrozen fc layer + 1 ext. fc layer): ~73-74%
Densenet161 (1 unfrozen fc layer): 75%
Densenet161 (1 unfrozen fc layer + 1 ext. fc layer): ~75-76%

Using Updated (Cleaned) Dataset (~5% overall val. acc. improvement)

Densenet161 (1 unfrozen fc layer): ~79%
Densenet161 (1 unfrozen fc layer + 1 ext. fc layer): ~80-81%
Densenet161 (1 unfrozen fc layer + 2 ext. fc layer): ~80%
Resnext101 (1 unfrozen fc layer): ~78%
Resnext101 (1 unfrozen fc layer + 1 ext. fc layer): ~78-79%
Resnext101 (1 unfrozen fc layer + 2 ext. fc layer): ~78-79%
Wres101 (1 unfrozen fc layer): ~75%
Wres101 (1 unfrozen fc layer + 1 ext. fc layer): ~76%
Alexnet (1 unfrozen fc layer): ~70%
Googlenet (1 unfrozen fc layer): ~72%
Shufflenet (1 unfrozen fc layer): ~73%

W/ Batch Norm on modified linear layers (~2% overall val. acc. improvement)

Densenet161 (1 unfrozen fc layer + 1 ext. fc layer): ~80.5%
Resnext101 (1 unfrozen fc layer + 1 ext. fc layer): ~80-81%
Resnext101 (1 unfrozen fc layer + 2 ext. fc layer): ~80-81%
Wres101 (1 unfrozen fc layer + 1 ext. fc layer): ~77%
Alexnet (1 unfrozen fc layer + 1 ext. fc layer): ~73-74%
Googlenet (1 unfrozen fc layer + 1 ext. fc layer): ~73%
Shufflenet (1 unfrozen fc layer + 1 ext. fc layer): ~75% (was 77 up to 50 epochs)

Ensembling Results

8-Transfer-Learned-Model Ensemble w/ Averaged Output: ~86-87%
8-TL Ensemble w/ 1 layer MLP: ~84% 8-TL Ensemble w/ 2 layer MLP: ~86%

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages