k = 5
epochs = 100
verbose_epoch = 95
learning_rate = 0.3
weight_decay = 3.0
net.add(gluon.nn.Dense(128))
net.add(gluon.nn.BatchNorm(),
gluon.nn.Activation('relu'))
net.add(gluon.nn.Dense(1))
k = 5
epochs = 50
verbose_epoch = 45
learning_rate = 0.03
weight_decay = 170
net.add(gluon.nn.Dense(1024, activation='relu'))
net.add(gluon.nn.Dropout(0.5))
net.add(gluon.nn.Dense(1))
##3.1 CIFAR10 param
num_epochs = 300
learning_rate = 0.1
weight_decay = 0.0005
lr_period = 40
learning_rate = 0.1
lr_decay = 0.5
num_epochs = 300
learning_rate = 0.1
weight_decay = 0.001
lr_period = 40
lr_decay = 0.5
if e > 150 and e % 20 == 0:
trainer.set_learning_rate(trainer.learning_rate * lr_decay) # decrease lr
Epoch 299. Train Loss: 0.286618, Train acc 0.904632, Valid acc 0.933200, lr=0.00078125,
the same params with 3.1.1 .Score is 0.9439 that only add data augmentation
step: 1) pad to (40,40)
2) horizontal flip to image with probability 0.5
3) random cropping other than center crop
4) normalization
https://github.com/SinyerAtlantis/deep_learning_gluon/tree/master/2.%20cnn_cifar10