You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For both models of confusion, the preferred usage is softmax probabilities (obtained after feeding the logits through a softmax). You can, alternatively, try the logits for pairwise confusion, but the loss weight will have to be scaled to a very small value to prevent oscillations, and hence we recommend operating on the softmax probabilities themselves.
In the paper,
So for the PairwiseConfusion, we are using logits? which is the direct output of pytorch models.
But for EntropicConfusion, obviously we should use softmax probabilities, which is obtained by feeding logits through a softmax function.
Am I right?
Thank you
The text was updated successfully, but these errors were encountered: