- 
                Notifications
    
You must be signed in to change notification settings  - Fork 37
 
Closed
Labels
enhancementNew feature or requestNew feature or requestgood first issueGood for newcomersGood for newcomers
Description
In the quiz and judge section, particularly in the judge part, I’ve noticed that the yes/no probabilities are not normalized.
For example, suppose the trainee model predicts Token("yes", p=0.1) while the ground truth is "no". In this case, we assume that the model’s probability for "no" is p("no") = 0.9, and we compute the loss as loss = -log(0.9). This results in a small loss, indicating that the classifier assigned a high probability (0.9) to the correct class, but actually the final verdict was incorrect.
Could you please comment on this behavior?
tpoisonooo
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or requestgood first issueGood for newcomersGood for newcomers