-
Notifications
You must be signed in to change notification settings - Fork 363
/
fixtures.go
466 lines (423 loc) · 15.3 KB
/
fixtures.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
// Copyright 2019 Antrea Authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package e2e
import (
"fmt"
"log"
"net"
"os"
"path/filepath"
"strings"
"sync"
"testing"
"time"
"k8s.io/component-base/featuregate"
"antrea.io/antrea/pkg/agent/config"
)
func skipIfNotBenchmarkTest(tb testing.TB) {
if !testOptions.withBench {
tb.Skipf("Skipping benchmark test: %s", tb.Name())
}
}
func skipIfProviderIs(tb testing.TB, name string, reason string) {
if testOptions.providerName == name {
tb.Skipf("Skipping test for the '%s' provider: %s", name, reason)
}
}
func skipIfNotRequired(tb testing.TB, keys ...string) {
for _, v := range keys {
if strings.Contains(testOptions.skipCases, v) {
tb.Skipf("Skipping test as %s is in skip list %s", v, testOptions.skipCases)
}
}
}
func skipIfNumNodesLessThan(tb testing.TB, required int) {
if clusterInfo.numNodes < required {
tb.Skipf("Skipping test as it requires %d different Nodes but cluster only has %d", required, clusterInfo.numNodes)
}
}
func skipIfRunCoverage(tb testing.TB, reason string) {
if testOptions.enableCoverage {
tb.Skipf("Skipping test for the '%s' when run coverage: %s", tb.Name(), reason)
}
}
func skipIfNotIPv4Cluster(tb testing.TB) {
if clusterInfo.podV4NetworkCIDR == "" {
tb.Skipf("Skipping test as it requires IPv4 addresses but the IPv4 network CIDR is not set")
}
}
func skipIfIPv6Cluster(tb testing.TB) {
if clusterInfo.podV6NetworkCIDR != "" {
tb.Skipf("Skipping test as it is not supported in IPv6 cluster")
}
}
func skipIfNotIPv6Cluster(tb testing.TB) {
if clusterInfo.podV6NetworkCIDR == "" {
tb.Skipf("Skipping test as it requires IPv6 addresses but the IPv6 network CIDR is not set")
}
}
func skipIfMissingKernelModule(tb testing.TB, nodeName string, requiredModules []string) {
for _, module := range requiredModules {
// modprobe with "--dry-run" does not require root privileges
cmd := fmt.Sprintf("modprobe --dry-run %s", module)
rc, stdout, stderr, err := RunCommandOnNode(nodeName, cmd)
if err != nil {
tb.Skipf("Skipping test as modprobe could not be run to confirm the presence of module '%s': %v", module, err)
}
if rc != 0 {
tb.Skipf("Skipping test as modprobe exited with an error when trying to confirm the presence of module '%s' - stdout: %s - stderr: %s", module, stdout, stderr)
}
}
tb.Logf("The following modules have been found on Node '%s': %v", nodeName, requiredModules)
}
func skipIfEncapModeIsNot(tb testing.TB, data *TestData, encapMode config.TrafficEncapModeType) {
currentEncapMode, err := data.GetEncapMode()
if err != nil {
tb.Fatalf("Failed to get encap mode: %v", err)
}
if currentEncapMode != encapMode {
tb.Skipf("Skipping test for encap mode '%s', test requires '%s'", currentEncapMode.String(), encapMode.String())
}
}
func skipIfHasWindowsNodes(tb testing.TB) {
if len(clusterInfo.windowsNodes) != 0 {
tb.Skipf("Skipping test as the cluster has Windows Nodes")
}
}
func skipIfNoWindowsNodes(tb testing.TB) {
if len(clusterInfo.windowsNodes) == 0 {
tb.Skipf("Skipping test as the cluster has no Windows Nodes")
}
}
func skipIfFeatureDisabled(tb testing.TB, feature featuregate.Feature, checkAgent bool, checkController bool) {
if checkAgent {
if featureGate, err := GetAgentFeatures(); err != nil {
tb.Fatalf("Cannot determine if %s is enabled in the Agent: %v", feature, err)
} else if !featureGate.Enabled(feature) {
tb.Skipf("Skipping test because %s is not enabled in the Agent", feature)
}
}
if checkController {
if featureGate, err := GetControllerFeatures(); err != nil {
tb.Fatalf("Cannot determine if %s is enabled in the Controller: %v", feature, err)
} else if !featureGate.Enabled(feature) {
tb.Skipf("Skipping test because %s is not enabled in the Controller", feature)
}
}
}
func ensureAntreaRunning(data *TestData) error {
log.Println("Applying Antrea YAML")
if err := data.deployAntrea(deployAntreaDefault); err != nil {
return err
}
log.Println("Waiting for all Antrea DaemonSet Pods")
if err := data.waitForAntreaDaemonSetPods(defaultTimeout); err != nil {
return err
}
log.Println("Checking CoreDNS deployment")
if err := data.checkCoreDNSPods(defaultTimeout); err != nil {
return err
}
return nil
}
func createDirectory(path string) error {
return os.Mkdir(path, 0700)
}
func (data *TestData) setupLogDirectoryForTest(testName string) error {
path := filepath.Join(testOptions.logsExportDir, testName)
// remove directory if it already exists. This ensures that we start with an empty
// directory
_ = os.RemoveAll(path)
err := createDirectory(path)
if err != nil {
return err
}
data.logsDirForTestCase = path
return nil
}
func setupTest(tb testing.TB) (*TestData, error) {
if err := testData.setupLogDirectoryForTest(tb.Name()); err != nil {
tb.Errorf("Error creating logs directory '%s': %v", testData.logsDirForTestCase, err)
return nil, err
}
success := false
defer func() {
if !success {
tb.Fail()
exportLogs(tb, testData, "afterSetupTest", true)
}
}()
tb.Logf("Creating '%s' K8s Namespace", testNamespace)
if err := ensureAntreaRunning(testData); err != nil {
return nil, err
}
if err := testData.createTestNamespace(); err != nil {
return nil, err
}
success = true
return testData, nil
}
func setupTestWithIPFIXCollector(tb testing.TB) (*TestData, bool, bool, error) {
v4Enabled := clusterInfo.podV4NetworkCIDR != ""
v6Enabled := clusterInfo.podV6NetworkCIDR != ""
testData, err := setupTest(tb)
if err != nil {
return testData, v4Enabled, v6Enabled, err
}
// Create pod using ipfix collector image
if err = testData.createPodOnNode("ipfix-collector", testNamespace, "", ipfixCollectorImage, nil, nil, nil, nil, true, nil); err != nil {
tb.Errorf("Error when creating the ipfix collector Pod: %v", err)
}
ipfixCollectorIP, err := testData.podWaitForIPs(defaultTimeout, "ipfix-collector", testNamespace)
if err != nil || len(ipfixCollectorIP.ipStrings) == 0 {
tb.Errorf("Error when waiting to get ipfix collector Pod IP: %v", err)
return nil, v4Enabled, v6Enabled, err
}
var ipStr string
if v6Enabled && ipfixCollectorIP.ipv6 != nil {
ipStr = ipfixCollectorIP.ipv6.String()
} else {
ipStr = ipfixCollectorIP.ipv4.String()
}
ipfixCollectorAddr := fmt.Sprintf("%s:tcp", net.JoinHostPort(ipStr, ipfixCollectorPort))
faClusterIPAddr := ""
tb.Logf("Applying flow aggregator YAML with ipfix collector address: %s", ipfixCollectorAddr)
faClusterIP, err := testData.deployFlowAggregator(ipfixCollectorAddr)
if err != nil {
return testData, v4Enabled, v6Enabled, err
}
if testOptions.providerName == "kind" {
// In Kind cluster, there are issues with DNS name resolution on worker nodes.
// Please note that CoreDNS services are forced on to control-plane Node.
faClusterIPAddr = fmt.Sprintf("%s:%s:tls", faClusterIP, ipfixCollectorPort)
}
tb.Logf("Deploying flow exporter with collector address: %s", faClusterIPAddr)
if err = testData.deployAntreaFlowExporter(faClusterIPAddr); err != nil {
return testData, v4Enabled, v6Enabled, err
}
tb.Logf("Checking CoreDNS deployment")
if err = testData.checkCoreDNSPods(defaultTimeout); err != nil {
return testData, v4Enabled, v6Enabled, err
}
return testData, v4Enabled, v6Enabled, nil
}
func exportLogs(tb testing.TB, data *TestData, logsSubDir string, writeNodeLogs bool) {
if tb.Skipped() {
return
}
// if test was successful and --logs-export-on-success was not provided, we do not export
// any logs.
if !tb.Failed() && !testOptions.logsExportOnSuccess {
return
}
const timeFormat = "Jan02-15-04-05"
timeStamp := time.Now().Format(timeFormat)
logsDir := filepath.Join(data.logsDirForTestCase, fmt.Sprintf("%s.%s", logsSubDir, timeStamp))
err := createDirectory(logsDir)
if err != nil {
tb.Errorf("Error when creating logs directory '%s': %v", logsDir, err)
return
}
tb.Logf("Exporting test logs to '%s'", logsDir)
// for now we just retrieve the logs for the Antrea Pods, but maybe we can find a good way to
// retrieve the logs for the test Pods in the future (before deleting them) if it is useful
// for debugging.
// getPodWriter creates the file with name nodeName-podName-suffix. It returns nil if the
// file cannot be created. File must be closed by the caller.
getPodWriter := func(nodeName, podName, suffix string) *os.File {
logFile := filepath.Join(logsDir, fmt.Sprintf("%s-%s-%s", nodeName, podName, suffix))
f, err := os.Create(logFile)
if err != nil {
tb.Errorf("Error when creating log file '%s': '%v'", logFile, err)
return nil
}
return f
}
// runKubectl runs the provided kubectl command on the control-plane Node and returns the
// output. It returns an empty string in case of error.
runKubectl := func(cmd string) string {
rc, stdout, _, err := RunCommandOnNode(controlPlaneNodeName(), cmd)
if err != nil || rc != 0 {
tb.Errorf("Error when running this kubectl command on control-plane Node: %s", cmd)
return ""
}
return stdout
}
// dump the logs for Antrea Pods to disk.
writePodLogs := func(nodeName, podName, nsName string) error {
w := getPodWriter(nodeName, podName, "logs")
if w == nil {
return nil
}
defer w.Close()
cmd := fmt.Sprintf("kubectl -n %s logs --all-containers %s", nsName, podName)
stdout := runKubectl(cmd)
if stdout == "" {
return nil
}
w.WriteString(stdout)
return nil
}
data.forAllMatchingPodsInNamespace("k8s-app=kube-proxy", kubeNamespace, writePodLogs)
data.forAllMatchingPodsInNamespace("app=antrea", antreaNamespace, writePodLogs)
// dump the logs for monitoring Pods to disk.
data.forAllMatchingPodsInNamespace("", monitoringNamespace, writePodLogs)
// dump the logs for flow-aggregator Pods to disk.
data.forAllMatchingPodsInNamespace("", flowAggregatorNamespace, writePodLogs)
// dump the output of "kubectl describe" for Antrea pods to disk.
data.forAllMatchingPodsInNamespace("app=antrea", antreaNamespace, func(nodeName, podName, nsName string) error {
w := getPodWriter(nodeName, podName, "describe")
if w == nil {
return nil
}
defer w.Close()
cmd := fmt.Sprintf("kubectl -n %s describe pod %s", nsName, podName)
stdout := runKubectl(cmd)
if stdout == "" {
return nil
}
w.WriteString(stdout)
return nil
})
if !writeNodeLogs {
return
}
// getNodeWriter creates the file with name nodeName-suffix. It returns nil if the file
// cannot be created. File must be closed by the caller.
getNodeWriter := func(nodeName, suffix string) *os.File {
logFile := filepath.Join(logsDir, fmt.Sprintf("%s-%s", nodeName, suffix))
f, err := os.Create(logFile)
if err != nil {
tb.Errorf("Error when creating log file '%s': '%v'", logFile, err)
return nil
}
return f
}
// export kubelet logs with journalctl for each Node. If the Nodes do not use journalctl we
// print a log message. If kubelet is not run with systemd, the log file will be empty.
if err := forAllNodes(func(nodeName string) error {
const numLines = 100
// --no-pager ensures the command does not hang.
cmd := fmt.Sprintf("journalctl -u kubelet -n %d --no-pager", numLines)
if clusterInfo.nodesOS[nodeName] == "windows" {
cmd = "Get-EventLog -LogName \"System\" -Source \"Service Control Manager\" | grep kubelet ; Get-EventLog -LogName \"Application\" -Source \"nssm\" | grep kubelet"
}
rc, stdout, _, err := RunCommandOnNode(nodeName, cmd)
if err != nil || rc != 0 {
// return an error and skip subsequent Nodes
return fmt.Errorf("error when running journalctl on Node '%s', is it available? Error: %v", nodeName, err)
}
w := getNodeWriter(nodeName, "kubelet")
if w == nil {
// move on to the next Node
return nil
}
defer w.Close()
w.WriteString(stdout)
return nil
}); err != nil {
tb.Logf("Error when exporting kubelet logs: %v", err)
}
}
func teardownFlowAggregator(tb testing.TB, data *TestData) {
if testOptions.enableCoverage {
if err := testData.gracefulExitFlowAggregator(testOptions.coverageDir); err != nil {
tb.Fatalf("Error when gracefully exiting Flow Aggregator: %v", err)
}
}
tb.Logf("Deleting '%s' K8s Namespace", flowAggregatorNamespace)
if err := data.deleteNamespace(flowAggregatorNamespace, defaultTimeout); err != nil {
tb.Logf("Error when tearing down flow aggregator: %v", err)
}
}
func teardownTest(tb testing.TB, data *TestData) {
exportLogs(tb, data, "beforeTeardown", true)
if empty, _ := IsDirEmpty(data.logsDirForTestCase); empty {
_ = os.Remove(data.logsDirForTestCase)
}
tb.Logf("Deleting '%s' K8s Namespace", testNamespace)
if err := data.deleteTestNamespace(defaultTimeout); err != nil {
tb.Logf("Error when tearing down test: %v", err)
}
}
func deletePodWrapper(tb testing.TB, data *TestData, name string) {
tb.Logf("Deleting Pod '%s'", name)
if err := data.deletePod(testNamespace, name); err != nil {
tb.Logf("Error when deleting Pod: %v", err)
}
}
// createTestBusyboxPods creates the desired number of busybox Pods and wait for their IP address to
// become available. This is a common patter in our tests, so having this helper function makes
// sense. It calls Fatalf in case of error, so it must be called from the goroutine running the test
// or benchmark function. You can create all the Pods on the same Node by setting nodeName. If
// nodeName is the empty string, each Pod will be created on an arbitrary
// Node. createTestBusyboxPods returns the cleanupFn function which can be used to delete the
// created Pods. Pods are created in parallel to reduce the time required to run the tests.
func createTestBusyboxPods(tb testing.TB, data *TestData, num int, ns string, nodeName string) (
podNames []string, podIPs []*PodIPs, cleanupFn func(),
) {
cleanupFn = func() {
var wg sync.WaitGroup
for _, podName := range podNames {
wg.Add(1)
go func(name string) {
deletePodWrapper(tb, data, name)
wg.Done()
}(podName)
}
wg.Wait()
}
type podData struct {
podName string
podIP *PodIPs
err error
}
createPodAndGetIP := func() (string, *PodIPs, error) {
podName := randName("test-pod-")
tb.Logf("Creating a busybox test Pod '%s' and waiting for IP", podName)
if err := data.createBusyboxPodOnNode(podName, ns, nodeName); err != nil {
tb.Errorf("Error when creating busybox test Pod '%s': %v", podName, err)
return "", nil, err
}
podIP, err := data.podWaitForIPs(defaultTimeout, podName, ns)
if err != nil {
tb.Errorf("Error when waiting for IP for Pod '%s': %v", podName, err)
return podName, nil, err
}
return podName, podIP, nil
}
podsCh := make(chan podData, num)
for i := 0; i < num; i++ {
go func() {
podName, podIP, err := createPodAndGetIP()
podsCh <- podData{podName, podIP, err}
}()
}
errCnt := 0
for i := 0; i < num; i++ {
pod := <-podsCh
if pod.podName != "" {
podNames = append(podNames, pod.podName)
podIPs = append(podIPs, pod.podIP)
}
if pod.err != nil {
errCnt++
}
}
if errCnt > 0 {
defer cleanupFn()
tb.Fatalf("%d / %d Pods could not be created successfully", errCnt, num)
}
return podNames, podIPs, cleanupFn
}