forked from GPflow/GPflow
-
Notifications
You must be signed in to change notification settings - Fork 0
/
test_model.py
126 lines (99 loc) · 3.75 KB
/
test_model.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
from __future__ import print_function
import GPflow
import tensorflow as tf
import numpy as np
import unittest
class TestOptimize(unittest.TestCase):
def setUp(self):
tf.reset_default_graph()
rng = np.random.RandomState(0)
class Quadratic(GPflow.model.Model):
def __init__(self):
GPflow.model.Model.__init__(self)
self.x = GPflow.param.Param(rng.randn(10))
def build_likelihood(self):
return -tf.reduce_sum(tf.square(self.x))
self.m = Quadratic()
def test_adam(self):
o = tf.train.AdamOptimizer()
self.m.optimize(o, maxiter=5000)
self.assertTrue(self.m.x.value.max() < 1e-2)
def test_lbfgsb(self):
self.m.optimize(disp=False)
self.assertTrue(self.m.x.value.max() < 1e-6)
class TestNeedsRecompile(unittest.TestCase):
def setUp(self):
self.m = GPflow.model.Model()
self.m.p = GPflow.param.Param(1.0)
def test_fix(self):
self.m._needs_recompile = False
self.m.p.fixed = True
self.assertTrue(self.m._needs_recompile)
def test_replace_param(self):
self.m._needs_recompile = False
new_p = GPflow.param.Param(3.0)
self.m.p = new_p
self.assertTrue(self.m._needs_recompile)
def test_set_prior(self):
self.m._needs_recompile = False
self.m.p.prior = GPflow.priors.Gaussian(0, 1)
self.assertTrue(self.m._needs_recompile)
def test_set_transform(self):
self.m._needs_recompile = False
self.m.p.transform = GPflow.transforms.Identity()
self.assertTrue(self.m._needs_recompile)
class KeyboardRaiser:
"""
This wraps a function and makes it raise a KeyboardInterrupt after some number of calls
"""
def __init__(self, iters_to_raise, f):
self.iters_to_raise, self.f = iters_to_raise, f
self.count = 0
def __call__(self, *a, **kw):
self.count += 1
if self.count >= self.iters_to_raise:
raise KeyboardInterrupt
return self.f(*a, **kw)
class TestKeyboardCatching(unittest.TestCase):
def setUp(self):
tf.reset_default_graph()
X = np.random.randn(1000, 3)
Y = np.random.randn(1000, 3)
Z = np.random.randn(100, 3)
self.m = GPflow.sgpr.SGPR(X, Y, Z=Z, kern=GPflow.kernels.RBF(3))
def test_optimize_np(self):
x0 = self.m.get_free_state()
self.m._compile()
self.m._objective = KeyboardRaiser(15, self.m._objective)
self.m.optimize(disp=0, maxiter=10000, ftol=0, gtol=0)
x1 = self.m.get_free_state()
self.assertFalse(np.allclose(x0, x1))
def test_optimize_tf(self):
x0 = self.m.get_free_state()
callback = KeyboardRaiser(5, lambda x: None)
o = tf.train.AdamOptimizer()
self.m.optimize(o, maxiter=15, callback=callback)
x1 = self.m.get_free_state()
self.assertFalse(np.allclose(x0, x1))
class TestLikelihoodAutoflow(unittest.TestCase):
def setUp(self):
tf.reset_default_graph()
X = np.random.randn(1000, 3)
Y = np.random.randn(1000, 3)
Z = np.random.randn(100, 3)
self.m = GPflow.sgpr.SGPR(X, Y, Z=Z, kern=GPflow.kernels.RBF(3))
def test_lik_and_prior(self):
l0 = self.m.compute_log_likelihood()
p0 = self.m.compute_log_prior()
self.m.kern.variance.prior = GPflow.priors.Gamma(1.4, 1.6)
l1 = self.m.compute_log_likelihood()
p1 = self.m.compute_log_prior()
self.assertTrue(p0 == 0.0)
self.assertFalse(p0 == p1)
self.assertTrue(l0 == l1)
class TestName(unittest.TestCase):
def test_name(self):
m = GPflow.model.Model(name='foo')
assert m.name == 'foo'
if __name__ == "__main__":
unittest.main()