Skip to content

Commit

Permalink
fp16 turn on/off support
Browse files Browse the repository at this point in the history
  • Loading branch information
ivanvovk committed Oct 29, 2020
1 parent d90736a commit 6be2f4c
Show file tree
Hide file tree
Showing 2 changed files with 16 additions and 8 deletions.
1 change: 1 addition & 0 deletions configs/default.json
Original file line number Diff line number Diff line change
Expand Up @@ -39,6 +39,7 @@
"n_epoch": 100000000,
"n_samples_to_test": 4,
"test_interval": 1,
"use_fp16": true,

"training_noise_schedule": {
"n_iter": 1000,
Expand Down
23 changes: 15 additions & 8 deletions train.py
Original file line number Diff line number Diff line change
@@ -1,14 +1,12 @@
import os
import argparse
import json
import socketserver
import numpy as np

import torch
import torch.distributed as dist
import torch.multiprocessing as mp
from torch.utils.data.distributed import DistributedSampler
from torch.multiprocessing import Process
from torch.utils.data import DataLoader

from datetime import datetime
Expand Down Expand Up @@ -50,7 +48,8 @@ def run_training(rank, config, args):
step_size=config.training_config.scheduler_step_size,
gamma=config.training_config.scheduler_gamma
)
scaler = torch.cuda.amp.GradScaler()
if config.training_config.use_fp16:
scaler = torch.cuda.amp.GradScaler()

show_message('Initializing data loaders...', verbose=args.verbose, rank=rank)
train_dataset = AudioDataset(config, training=True)
Expand Down Expand Up @@ -113,17 +112,25 @@ def run_training(rank, config, args):
batch = batch.cuda()
mels = mel_fn(batch)

with torch.cuda.amp.autocast():
if config.training_config.use_fp16:
with torch.cuda.amp.autocast():
loss = (model if args.n_gpus == 1 else model.module).compute_loss(mels, batch)
scaler.scale(loss).backward()
scaler.unscale_(optimizer)
else:
loss = (model if args.n_gpus == 1 else model.module).compute_loss(mels, batch)
loss.backward()

scaler.scale(loss).backward()
scaler.unscale_(optimizer)
grad_norm = torch.nn.utils.clip_grad_norm_(
parameters=model.parameters(),
max_norm=config.training_config.grad_clip_threshold
)
scaler.step(optimizer)
scaler.update()

if config.training_config.use_fp16:
scaler.step(optimizer)
scaler.update()
else:
optimizer.step()

loss_stats = {
'total_loss': loss.item(),
Expand Down

0 comments on commit 6be2f4c

Please sign in to comment.