/
ResNet.py
110 lines (85 loc) · 3.49 KB
/
ResNet.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
import tensorflow as tf
from tensorflow.keras import layers, Sequential, Model
class BasicBlock(layers.Layer):
def __init__(self, kernels, stride=1):
super(BasicBlock, self).__init__()
self.features = Sequential([
layers.Conv2D(kernels, (3, 3), strides=stride, padding='same'),
layers.BatchNormalization(),
layers.ReLU(),
layers.Conv2D(kernels, (3, 3), strides=1, padding='same'),
layers.BatchNormalization()
])
if stride != 1:
shortcut = [
layers.Conv2D(kernels, (1, 1), strides=stride),
layers.BatchNormalization()
]
else:
shortcut = []
self.shorcut = Sequential(shortcut)
def call(self, inputs, training=False):
residual = self.shorcut(inputs, training=training)
x = self.features(inputs, training=training)
x = tf.nn.relu(layers.add([residual, x]))
return x
class BottleNeckBlock(layers.Layer):
def __init__(self, kernels, stride=1):
super(BottleNeckBlock, self).__init__()
self.features = Sequential([
layers.Conv2D(kernels, (1, 1), strides=1, padding='same'),
layers.BatchNormalization(),
layers.Conv2D(kernels, (3, 3), strides=stride, padding='same'),
layers.BatchNormalization(),
layers.Conv2D(kernels * 4, (1, 1), strides=1, padding='same'),
layers.BatchNormalization(),
])
self.shorcut = Sequential([
layers.Conv2D(kernels * 4, (1, 1), strides=stride),
layers.BatchNormalization()
])
def call(self, inputs, training=False):
residual = self.shorcut(inputs, training=training)
x = self.features(inputs, training=training)
x = tf.nn.relu(x + residual)
return x
class ResNet(Model):
def __init__(self, block, num_blocks, num_classes, input_shape=(32, 32, 3)):
super(ResNet, self).__init__()
self.conv1 = Sequential([
layers.Input(input_shape),
layers.Conv2D(64, (3, 3), padding='same', use_bias=False),
layers.BatchNormalization(),
layers.ReLU()
])
self.conv2_x = self._make_layer(block, 64, num_blocks[0], 1)
self.conv3_x = self._make_layer(block, 128, num_blocks[1], 2)
self.conv4_x = self._make_layer(block, 256, num_blocks[2], 2)
self.conv5_x = self._make_layer(block, 512, num_blocks[3], 2)
self.gap = layers.GlobalAveragePooling2D()
self.fc = layers.Dense(num_classes, activation='softmax')
def _make_layer(self, block, kernels, num_blocks, stride):
strides = [stride] + [1] * (num_blocks - 1)
nets = []
for stride in strides:
nets.append(block(kernels, stride))
return Sequential(nets)
def call(self, inputs):
x = self.conv1(inputs)
x = self.conv2_x(x)
x = self.conv3_x(x)
x = self.conv4_x(x)
x = self.conv5_x(x)
x = self.gap(x)
x = self.fc(x)
return x
def ResNet18(num_classes):
return ResNet(BasicBlock, [2, 2, 2, 2], num_classes)
def ResNet34(num_classes):
return ResNet(BasicBlock, [3, 4, 6, 3], num_classes)
def ResNet50(num_classes):
return ResNet(BottleNeckBlock, [3, 4, 6, 3], num_classes)
def ResNet101(num_classes):
return ResNet(BottleNeckBlock, [3, 4, 23, 3], num_classes)
def ResNet152(num_classes):
return ResNet(BottleNeckBlock, [3, 8, 36, 3], num_classes)