Skip to content

Commit

Permalink
Fix broken RRNN layer.
Browse files Browse the repository at this point in the history
  • Loading branch information
Leif Johnson committed Feb 21, 2016
1 parent d90eb62 commit ce65b3d
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions theanets/layers/recurrent.py
Expand Up @@ -380,12 +380,12 @@ def transform(self, inputs):
x = self._only_input(inputs).dimshuffle(1, 0, 2)

step = self._step_static
inputs = [TT.dot(x, self.find('xh')) + self.find('b')]
arrays = [TT.dot(x, self.find('xh')) + self.find('b')]
const = []
if self.rate == 'matrix':
step = self._step_dynamic
r = TT.nnet.sigmoid(TT.dot(x, self.find('xr')) + self.find('r'))
inputs.append(r)
arrays.append(r)
elif self.rate == 'vector':
r = TT.nnet.sigmoid(self.find('r'))
const.append(r)
Expand All @@ -396,7 +396,7 @@ def transform(self, inputs):
# output is: (time, batch, output)
# we want: (batch, time, output)
(p, h, o), updates = self._scan(
inputs, [None, None, inputs.get(self.h_0, x.shape[1])],
arrays, [None, None, inputs.get(self.h_0, x.shape[1])],
constants=const, step=step)
pre = p.dimshuffle(1, 0, 2)
hid = h.dimshuffle(1, 0, 2)
Expand Down

0 comments on commit ce65b3d

Please sign in to comment.