-
Notifications
You must be signed in to change notification settings - Fork 0
/
gradient-descent.gyp
74 lines (57 loc) · 1.72 KB
/
gradient-descent.gyp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
import torch
import numpy as np
import torch.nn as nn
import torch.nn.functional as F
from torch.utils.data import TensorDataset
inputs = np.array([
[800, 1200, 700],
[900, 1500, 800],
[450, 800, 300],
[650, 1050, 500]], dtype="float32")
targets = np.array([
[3000],
[3500],
[1500],
[2000]], dtype="float32")
inputs = torch.from_numpy(inputs)
targets = torch.from_numpy(targets)
#we create a dataset with the inputs and the targets
train_ds = TensorDataset(inputs, targets)
#we randomly initliasie a set of weights and biases
model = nn.Linear(3,1)
#print(model.weight)
#print(model.bias)
#we try to get some predictions with the weights we initialised randomly
preds = model(inputs)
#print(preds)
#when nn.Linear it generates a set of weights and biases
weight = model.weight
bias = model.bias
#we compute the loss function
loss_fn = F.mse_loss
loss = loss_fn(preds, targets)
print("Std. loss", loss)
#when we call loss.backward PyTorch computes the gradients of the loss with respect to the weights
loss.backward()
#after we compute the weights and biases we subtract to them
#their partial derivatives, times a small amount that we call the llearning rate
with torch.no_grad():
weight -= weight.grad * 1e-7
bias -= bias.grad * 1e-7
weight.grad.zero_()
bias.grad.zero_()
preds = model(inputs)
loss = loss_fn(preds, targets)
print("Adjusted loss:", loss)
for i in range(10000):
preds = model(inputs)
loss = loss_fn(preds, targets)
loss.backward()
with torch.no_grad():
weight -= weight.grad * 1e-7
bias -= bias.grad * 1e-7
weight.grad.zero_()
bias.grad.zero_()
preds = model(inputs)
loss = loss_fn(preds, targets)
print("Final loss:", loss)