Ultimate goal is to develop a deep NN that is able to predict labels for the samples from taraget domain
Besides the 1) standard source classification loss, we also employ 2)domain adversarial loss(domain confusion loss) 3)semantic loss (pseudo-labeled semantic loss)
- G : Feature Extractor
- D : Domain Discriminator (whether features from G arise from source or target domain)
- F : Classifier
- CNN architecture : AlexNet Architecture, a bottleneck layer fcb with 256 units is added after fc7 layer for safer transfer representation learning.
- fcb as inputs to the discriminator as well as centroid computation
- discriminator: x-> 1024 -> 1024 -> 1, dropout is used
- Hyper-parameters tuning : weight balance parameter lamda, and moving average coefficient theta. (reverse validation)
- Stochastic Gradient Descent with 0.9 momentum is used.
- feature extraction
- align distributions (moving average)
- centroid alignment
- we follow the protocol in (Tzeng et al. 2017 "Adversarial discriminative domain adaptation")
- we use the training set of SVHN and test set of MNIST for evaluation
- During testing, target images are mapped with the target encoder to the shared feature space and classified by the source classifier.
- digit images are also cast to 28x28x1 in all experiements
- two convolution layers followed by max pool layers and two fully connected layers are placed behind
- for discriminator, x->500->500->1
- Batch normalization is inserted in convolutional layers