forked from foamliu/Face-Attributes-Mobile
-
Notifications
You must be signed in to change notification settings - Fork 0
/
utils.py
166 lines (130 loc) · 5.23 KB
/
utils.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
import argparse
import logging
import torch
from config import expression_dict, face_shape_dict, face_type_dict, gender_dict, glasses_dict, race_dict
def clip_gradient(optimizer, grad_clip):
"""
Clips gradients computed during backpropagation to avoid explosion of gradients.
:param optimizer: optimizer with the gradients to be clipped
:param grad_clip: clip value
"""
for group in optimizer.param_groups:
for param in group['params']:
if param.grad is not None:
param.grad.data.clamp_(-grad_clip, grad_clip)
def save_checkpoint(epoch, epochs_since_improvement, model, optimizer, loss, is_best):
state = {'epoch': epoch,
'epochs_since_improvement': epochs_since_improvement,
'loss': loss,
'model': model,
'optimizer': optimizer}
# filename = 'checkpoint_' + str(epoch) + '_' + str(loss) + '.tar'
filename = 'checkpoint.tar'
torch.save(state, filename)
# If this checkpoint is the best so far, store a copy so it doesn't get overwritten by a worse checkpoint
if is_best:
torch.save(state, 'BEST_checkpoint.tar')
class AverageMeter(object):
"""
Keeps track of most recent, average, sum, and count of a metric.
"""
def __init__(self):
self.reset()
def reset(self):
self.val = 0
self.avg = 0
self.sum = 0
self.count = 0
def update(self, val, n=1):
self.val = val
self.sum += val * n
self.count += n
self.avg = self.sum / self.count
class LossMeterBag(object):
def __init__(self, name_list):
self.meter_dict = dict()
self.name_list = name_list
for name in self.name_list:
self.meter_dict[name] = AverageMeter()
def update(self, val_list):
for i, name in enumerate(self.name_list):
val = val_list[i]
self.meter_dict[name].update(val)
def __str__(self):
ret = ''
for name in self.name_list:
ret += '{0}:\t {1:.4f}({2:.4f})\t'.format(name, self.meter_dict[name].val, self.meter_dict[name].avg)
return ret
def adjust_learning_rate(optimizer, shrink_factor):
"""
Shrinks learning rate by a specified factor.
:param optimizer: optimizer whose learning rate must be shrunk.
:param shrink_factor: factor in interval (0, 1) to multiply learning rate with.
"""
print("\nDECAYING learning rate.")
for param_group in optimizer.param_groups:
param_group['lr'] = param_group['lr'] * shrink_factor
print("The new learning rate is %f\n" % (optimizer.param_groups[0]['lr'],))
def accuracy(scores, targets, k=1):
batch_size = targets.size(0)
_, ind = scores.topk(k, 1, True, True)
correct = ind.eq(targets.view(-1, 1).expand_as(ind))
correct_total = correct.view(-1).float().sum() # 0D tensor
return correct_total.item() * (100.0 / batch_size)
def parse_args():
parser = argparse.ArgumentParser(description='Train face network')
# general
parser.add_argument('--end-epoch', type=int, default=1000, help='training epoch size.')
parser.add_argument('--lr', type=float, default=0.001, help='start learning rate')
parser.add_argument('--lr-step', type=int, default=10, help='period of learning rate decay')
parser.add_argument('--optimizer', default='sgd', help='optimizer')
parser.add_argument('--weight-decay', type=float, default=0.0005, help='weight decay')
parser.add_argument('--mom', type=float, default=0.9, help='momentum')
parser.add_argument('--batch-size', type=int, default=32, help='batch size in each context')
parser.add_argument('--checkpoint', type=str, default=None, help='checkpoint')
parser.add_argument('--pretrained', type=bool, default=False, help='pretrained model')
args = parser.parse_args()
return args
def get_logger():
logger = logging.getLogger()
handler = logging.StreamHandler()
formatter = logging.Formatter("%(asctime)s %(levelname)s \t%(message)s")
handler.setFormatter(formatter)
logger.addHandler(handler)
logger.setLevel(logging.DEBUG)
return logger
def idx2name(idx, tag):
name = None
if tag == 'expression':
name = expression_dict[idx]
elif tag == 'face_shape':
name = face_shape_dict[idx]
elif tag == 'face_type':
name = face_type_dict[idx]
elif tag == 'gender':
name = gender_dict[idx]
elif tag == 'glasses':
name = glasses_dict[idx]
elif tag == 'race':
name = race_dict[idx]
return name
def name2idx(name):
lookup_table = {'none': 0, 'smile': 1, 'laugh': 2,
'square': 0, 'oval': 1, 'heart': 2, 'round': 3, 'triangle': 4,
'human': 0, 'cartoon': 1,
'female': 0, 'male': 1,
'sun': 1, 'common': 2,
'yellow': 0, 'white': 1, 'black': 2, 'arabs': 3}
return lookup_table[name]
def crop_image(img, bbox):
# print(bbox.shape)
x1 = max(int(round(bbox[0])), 0)
y1 = max(int(round(bbox[1])), 0)
x2 = max(int(round(bbox[2])), 0)
y2 = max(int(round(bbox[3])), 0)
crop_img = img[y1:y2, x1:x2]
return crop_img
def ensure_folder(folder):
import os
if not os.path.isdir(folder):
os.mkdir(folder)