You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why "freezing old network" of after_task is left out in file der.py, while other methods, such as podnet and WA execute this step as follows:
self._old_network = self._network.copy().freeze()?
The text was updated successfully, but these errors were encountered:
Hi, thanks for your interest. WA and PodNet are distillation-based methods, meaning they have to save the model in the last incremental session, while DER does not.
In original paper, DER noticed that "we freeze the previously learned representation" for several times. If the previous convs in self.convnets are not freezed, which is also not executed in the code, does it means that those previous "resnet" modules will be affected by the back propagation? How to understand "freeze the previously learned representation"?
Why "freezing old network" of after_task is left out in file der.py, while other methods, such as podnet and WA execute this step as follows:
self._old_network = self._network.copy().freeze()?
The text was updated successfully, but these errors were encountered: