Skip to content

Commit

Permalink
Switch back to dense layer size 32 from 16
Browse files Browse the repository at this point in the history
  • Loading branch information
timodonnell committed Nov 30, 2017
1 parent bdee833 commit d65f774
Show file tree
Hide file tree
Showing 11 changed files with 10 additions and 62 deletions.
4 changes: 1 addition & 3 deletions downloads-generation/models_class1/hyperparameters.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -42,9 +42,7 @@
],
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
],
"layer_sizes": [32],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
"dropout_probability": 0.0,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ time mhcflurry-class1-train-allele-specific-models \
--allele $ALLELES 2>&1 | tee -a LOG.standard.txt &

# Model variations on qualitative + quantitative
for mod in 0local_noL1 0local 2local widelocal dense8 dense32 dense64 noL1 onehot embedding
for mod in 0local_noL1 0local 2local widelocal dense16 dense64 noL1 onehot embedding
do
cp $SCRIPT_DIR/hyperparameters-${mod}.yaml .
mkdir models-${mod}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.0,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
8
16
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.0,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@
"activation": "relu",
"output_activation": "sigmoid",
"layer_sizes": [
16
32
],
"dense_layer_l1_regularization": 0.001,
"batch_normalization": false,
Expand Down

0 comments on commit d65f774

Please sign in to comment.