Skip to content

Commit

Permalink
[dynaml-tensorflow]: Learn
Browse files Browse the repository at this point in the history
  -- Added gradient clipping parameter to model building

Signed-off-by: mandar2812 <mandar2812@gmail.com>
  • Loading branch information
mandar2812 committed Nov 17, 2018
1 parent 7df9d9b commit a18093d
Showing 1 changed file with 9 additions and 6 deletions.
Expand Up @@ -21,7 +21,7 @@ package io.github.mandar2812.dynaml.tensorflow
import _root_.io.github.mandar2812.dynaml.pipes.DataPipe
import _root_.io.github.mandar2812.dynaml.models.TFModel
import io.github.mandar2812.dynaml.tensorflow.layers.{DynamicTimeStepCTRNN, FiniteHorizonCTRNN, FiniteHorizonLinear}
import org.platanios.tensorflow.api.learn.{Mode, StopCriteria}
import org.platanios.tensorflow.api.learn.{ClipGradients, Mode, NoClipGradients, StopCriteria}
import org.platanios.tensorflow.api.learn.layers.{Input, Layer}
import org.platanios.tensorflow.api.ops.NN.SameConvPadding
import org.platanios.tensorflow.api.ops.io.data.Dataset
Expand Down Expand Up @@ -492,7 +492,8 @@ private[tensorflow] object Learn {
stopCriteria: StopCriteria,
stepRateFreq: Int = 5000,
summarySaveFreq: Int = 5000,
checkPointFreq: Int = 5000)(
checkPointFreq: Int = 5000,
clipGradients: ClipGradients = NoClipGradients)(
training_data: Dataset[
(IT, TT), (IO, TO),
(ID, TD), (IS, TS)],
Expand All @@ -503,7 +504,8 @@ private[tensorflow] object Learn {
val model = tf.learn.Model.supervised(
input, architecture,
target, processTarget,
loss, optimizer)
loss, optimizer,
clipGradients)

println("\nTraining model.\n")

Expand Down Expand Up @@ -597,9 +599,10 @@ private[tensorflow] object Learn {
optimizer: Optimizer,
summariesDir: java.nio.file.Path,
stopCriteria: StopCriteria,
stepRateFreq: Int,
summarySaveFreq: Int,
checkPointFreq: Int)(
stepRateFreq: Int = 5000,
summarySaveFreq: Int = 5000,
checkPointFreq: Int = 5000,
clipGradients: ClipGradients = NoClipGradients)(
training_data: Dataset[IT, IO, ID, IS],
inMemory: Boolean): UnsupModelPair[IT, IO, ID, IS, I] = {

Expand Down

0 comments on commit a18093d

Please sign in to comment.