You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Great thanks for the opensource code. I wonder how you calculate the size of the models?
Actually, I have a very fundamental question the BERT-base model (fp32) is
embedding_param = 23835648
num_param = 85526016
(embedding_param + num_param) / 1e6 * 4 =437.4MB
instead of 418MB?
for n, p in model.items():
if 'Norm' in n:
continue
if len(p.size()) ==2 :
if 'embedding' in n:
embedding_param += p.numel()
else:
num_param += p.numel()
Dear authors,
Great thanks for the opensource code. I wonder how you calculate the size of the models?
Actually, I have a very fundamental question the BERT-base model (fp32) is
embedding_param = 23835648
num_param = 85526016
(embedding_param + num_param) / 1e6 * 4 =437.4MB
instead of 418MB?
It would be great if you can clarify this. (similar repohttps://github.com/huawei-noah/Pretrained-Language-Model/issues/184)
The text was updated successfully, but these errors were encountered: