Join GitHub today
why is detach necessary #116
Hi, I am wondering why is detach necessary in this line:
I understand that we want to update the gradients of netD without changin the ones of netG. But if the optimizer is only using the parameters of netD, then only its weight will be updated. Am I missing something here?
@soumith This is not true. Detaching