In the example of per-sample gradient, the following line can be misleading since the `predictions` of a net are logits: https://github.com/pytorch/tutorials/blob/08a61b7cae9d00312d0029b1f86a248ec1253a83/intermediate_source/per_sample_grads.py#L49 The correct way should be: ``` python return F.nll_loss(F.log_softmax(predictions, dim=-1), targets) ``` Would appreciate if this can be corrected.