Skip to content

Commit

Permalink
flattening package for easier redistribution
Browse files Browse the repository at this point in the history
  • Loading branch information
drscotthawley committed Jul 24, 2019
1 parent 3d5c824 commit bfe5f9b
Show file tree
Hide file tree
Showing 9 changed files with 7 additions and 18 deletions.
2 changes: 1 addition & 1 deletion demo/bokeh_sliders.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
import os, sys
sys.path.append(os.path.abspath('../')) # for running from signaltrain/demo/
import signaltrain as st
from signaltrain.nn_modules import nn_proc
from signaltrain import nn_proc

if torch.cuda.is_available():
device = torch.device("cuda:0")
Expand Down
2 changes: 1 addition & 1 deletion signaltrain/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,6 @@
from . import io_methods
from . import learningrate
from . import loss_functions
from . import nn_modules
from . import nn_proc
from . import misc
from . import train
File renamed without changes.
File renamed without changes.
11 changes: 0 additions & 11 deletions signaltrain/nn_modules/__init__.py

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@

# -*- coding: utf-8 -*-
__author__ = 'S.I. Mimilakis'
__author__ = 'S.I. Mimilakis & S.H. Hawley'
__copyright__ = 'MacSeNet'

# imports
Expand Down
4 changes: 2 additions & 2 deletions signaltrain/train.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,8 @@
from torch.utils.data import DataLoader
import os, sys
import time
from signaltrain import audio, io_methods, learningrate, datasets, loss_functions, misc
from signaltrain.nn_modules import nn_proc as nn_proc
from signaltrain import audio, io_methods, learningrate, datasets, loss_functions, misc, nn_proc
#from signaltrain.nn_modules import nn_proc as nn_proc

# NVIDIA Apex for mixed-precision training
have_apex = False
Expand Down
2 changes: 1 addition & 1 deletion utils/lr_finder.py
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,7 @@ def lrfind(model, dataloader, optimizer, calc_loss, start=1e-6, stop=4e-3, num_l


# Initialize nn modules
model = st.nn_modules.nn_proc.st_model(scale_factor=args.scale, shrink_factor=args.shrink, num_knobs=len(effect.knob_names), sr=44100)
model = st.nn_proc.st_model(scale_factor=args.scale, shrink_factor=args.shrink, num_knobs=len(effect.knob_names), sr=44100)
chunk_size = model.in_chunk_size

optimizer = torch.optim.Adam(list(model.parameters()), lr=args.lrmax, weight_decay=0)
Expand Down
2 changes: 1 addition & 1 deletion utils/predict_long.py
Original file line number Diff line number Diff line change
Expand Up @@ -97,7 +97,7 @@ def calc_ct(signal, effect, knobs_wc, out_chunk_size, chunk_size, sr=44100):
## Can be run as standalone app for testing / eval purposes
import os
import argparse
from signaltrain.nn_modules import nn_proc
from signaltrain import nn_proc

# torch device
if torch.cuda.is_available():
Expand Down

0 comments on commit bfe5f9b

Please sign in to comment.