/
train_lm.py
41 lines (37 loc) · 1.17 KB
/
train_lm.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
import os
#from lm_worker import train
from lm_gen import train
def main(job_id, params):
print params
validerr = train(
saveto=params['model'][0],
reload_=params['reload'][0],
dim_word=params['dim_word'][0],
dim=params['dim'][0],
n_words=params['n-words'][0],
decay_c=params['decay-c'][0],
lrate=params['learning-rate'][0],
optimizer=params['optimizer'][0],
maxlen=30,
batch_size=32,
valid_batch_size=16,
validFreq=5000,
dispFreq=10,
saveFreq=1000,
sampleFreq=1000,
dataset='/data/lisatmp3/chokyun/wikipedia/extracted/wiki.tok.txt.gz',
valid_dataset='/data/lisatmp4/anirudhg/newstest2011.en.tok',
dictionary='/data/lisatmp4/anirudhg/wiki.tok.txt.gz.pkl',
use_dropout=params['use-dropout'][0])
return validerr
if __name__ == '__main__':
main(0, {
'model': ['./models/model_lm.npz'],
'dim_word': [512],
'dim': [1024],
'n-words': [30000],
'optimizer': ['adadelta'],
'decay-c': [0.],
'use-dropout': [False],
'learning-rate': [0.0001],
'reload': [True]})