Skip to content

Commit

Permalink
Merge pull request #21 from fzliu/develop
Browse files Browse the repository at this point in the history
Bugfix and other minor changes.
  • Loading branch information
fzliu committed Nov 10, 2015
2 parents df070c5 + 2130e54 commit 3d77007
Show file tree
Hide file tree
Showing 3 changed files with 143 additions and 142 deletions.
124 changes: 62 additions & 62 deletions models/vgg16/VGG_ILSVRC_16_layers_deploy.prototxt
Expand Up @@ -7,276 +7,276 @@ input_shape {
dim: 224
dim: 224
}
layers {
layer {
bottom: "data"
top: "conv1_1"
name: "conv1_1"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 64
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv1_1"
top: "conv1_1"
name: "relu1_1"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv1_1"
top: "conv1_2"
name: "conv1_2"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 64
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv1_2"
top: "conv1_2"
name: "relu1_2"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv1_2"
top: "pool1"
name: "pool1"
type: POOLING
type: "Pooling"
pooling_param {
pool: AVE
kernel_size: 2
stride: 2
}
}
layers {
layer {
bottom: "pool1"
top: "conv2_1"
name: "conv2_1"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 128
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv2_1"
top: "conv2_1"
name: "relu2_1"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv2_1"
top: "conv2_2"
name: "conv2_2"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 128
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv2_2"
top: "conv2_2"
name: "relu2_2"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv2_2"
top: "pool2"
name: "pool2"
type: POOLING
type: "Pooling"
pooling_param {
pool: AVE
kernel_size: 2
stride: 2
}
}
layers {
layer {
bottom: "pool2"
top: "conv3_1"
name: "conv3_1"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv3_1"
top: "conv3_1"
name: "relu3_1"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv3_1"
top: "conv3_2"
name: "conv3_2"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv3_2"
top: "conv3_2"
name: "relu3_2"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv3_2"
top: "conv3_3"
name: "conv3_3"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv3_3"
top: "conv3_3"
name: "relu3_3"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv3_3"
top: "pool3"
name: "pool3"
type: POOLING
type: "Pooling"
pooling_param {
pool: AVE
kernel_size: 2
stride: 2
}
}
layers {
layer {
bottom: "pool3"
top: "conv4_1"
name: "conv4_1"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv4_1"
top: "conv4_1"
name: "relu4_1"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv4_1"
top: "conv4_2"
name: "conv4_2"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv4_2"
top: "conv4_2"
name: "relu4_2"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv4_2"
top: "conv4_3"
name: "conv4_3"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv4_3"
top: "conv4_3"
name: "relu4_3"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv4_3"
top: "pool4"
name: "pool4"
type: POOLING
type: "Pooling"
pooling_param {
pool: AVE
kernel_size: 2
stride: 2
}
}
layers {
layer {
bottom: "pool4"
top: "conv5_1"
name: "conv5_1"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv5_1"
top: "conv5_1"
name: "relu5_1"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv5_1"
top: "conv5_2"
name: "conv5_2"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv5_2"
top: "conv5_2"
name: "relu5_2"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv5_2"
top: "conv5_3"
name: "conv5_3"
type: CONVOLUTION
type: "Convolution"
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
layer {
bottom: "conv5_3"
top: "conv5_3"
name: "relu5_3"
type: RELU
type: "ReLU"
}
layers {
layer {
bottom: "conv5_3"
top: "pool5"
name: "pool5"
type: POOLING
type: "Pooling"
pooling_param {
pool: AVE
kernel_size: 2
Expand Down

0 comments on commit 3d77007

Please sign in to comment.