You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I read the code, I found there is only the discriminator loss and no generator loss. In other words, there is no adversarial training in MEALv2, which is different from my intuition. I want to know what is the advantage of just using the discriminator.
The text was updated successfully, but these errors were encountered:
Hi @PeterouZh, generator loss is the similarity loss in our framework to produce the same distribution as teachers', i.e., the KL divergence loss. Conventional adversarial training uses alternate updating but since the input and output of our discriminator and generator (student) are differentiable (not images), we can train the whole pipeline jointly.
Hi,
thanks for your great job.
When I read the code, I found there is only the discriminator loss and no generator loss. In other words, there is no adversarial training in MEALv2, which is different from my intuition. I want to know what is the advantage of just using the discriminator.
The text was updated successfully, but these errors were encountered: