You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In 'NAO/cnn/encoder/encoder.py', the gradient of predictor is computed as follow: new_arch_outputs = self.encoder_outputs - self.params['predict_lambda'] * grads_on_outputs
However, I think the gradient should be computed as described in paper:
Is there an error with computing the gradient of predictor in code?
The text was updated successfully, but these errors were encountered:
@wenerhg In the paper, for the general purpose, we denote f as the performance which is the larger the better, so we move towards the positive direction of the gradient, which shows as '+'. In the code, we use error rate as f which is the lower the better, so we need to move towards the negative direction, which shows as '-'.
In 'NAO/cnn/encoder/encoder.py', the gradient of predictor is computed as follow:
new_arch_outputs = self.encoder_outputs - self.params['predict_lambda'] * grads_on_outputs
However, I think the gradient should be computed as described in paper:
![h_{t}^{'} = h_{t} + \eta \frac{\partial f}{\partial h_{t}}, e_{x^{'}}=\{h_{1}^{'},\dots,h_{T}^{'}\}](https://camo.githubusercontent.com/76c9adad995316c3eaccd85d8961c6a622e9ee0693528f7daac22b6d0ad7959e/68747470733a2f2f6c617465782e636f6465636f67732e636f6d2f6769662e6c617465783f685f7b747d5e7b277d2673706163653b3d2673706163653b685f7b747d2673706163653b2b2673706163653b5c6574612673706163653b5c667261637b5c7061727469616c2673706163653b667d7b5c7061727469616c2673706163653b685f7b747d7d2c2673706163653b655f7b785e7b277d7d3d5c7b685f7b317d5e7b277d2c5c646f74732c685f7b547d5e7b277d5c7d)
Is there an error with computing the gradient of predictor in code?
The text was updated successfully, but these errors were encountered: