From ed984c1f26be73d19af5cd0df5ca2938ba946112 Mon Sep 17 00:00:00 2001 From: Eduardo Patrocinio Date: Sat, 29 Nov 2025 12:23:09 -0500 Subject: [PATCH] Fix backpropagation order in optimization tutorial Reorder optimizer.zero_grad(), loss.backward(), and optimizer.step() to match the recommended best practice documented in the tutorial. Fixes #3507 --- beginner_source/basics/optimization_tutorial.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/beginner_source/basics/optimization_tutorial.py b/beginner_source/basics/optimization_tutorial.py index 82bfaa8f07c..63b1e1eff52 100644 --- a/beginner_source/basics/optimization_tutorial.py +++ b/beginner_source/basics/optimization_tutorial.py @@ -158,9 +158,9 @@ def train_loop(dataloader, model, loss_fn, optimizer): loss = loss_fn(pred, y) # Backpropagation + optimizer.zero_grad() loss.backward() optimizer.step() - optimizer.zero_grad() if batch % 100 == 0: loss, current = loss.item(), batch * batch_size + len(X)