Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

clean up test code #81974

Merged
merged 1 commit into from
Sep 11, 2019
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
9 changes: 4 additions & 5 deletions test/e2e/examples.go
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@ import (
commonutils "k8s.io/kubernetes/test/e2e/common"
"k8s.io/kubernetes/test/e2e/framework"
"k8s.io/kubernetes/test/e2e/framework/auth"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
"k8s.io/kubernetes/test/e2e/framework/testfiles"

Expand Down Expand Up @@ -83,14 +82,14 @@ var _ = framework.KubeDescribe("[Feature:Example]", func() {
pod, err := c.CoreV1().Pods(ns).Get(podName, metav1.GetOptions{})
framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", podName))
stat := podutil.GetExistingContainerStatus(pod.Status.ContainerStatuses, podName)
e2elog.Logf("Pod: %s, restart count:%d", stat.Name, stat.RestartCount)
framework.Logf("Pod: %s, restart count:%d", stat.Name, stat.RestartCount)
if stat.RestartCount > 0 {
e2elog.Logf("Saw %v restart, succeeded...", podName)
framework.Logf("Saw %v restart, succeeded...", podName)
wg.Done()
return
}
}
e2elog.Logf("Failed waiting for %v restart! ", podName)
framework.Logf("Failed waiting for %v restart! ", podName)
passed = false
wg.Done()
}
Expand All @@ -106,7 +105,7 @@ var _ = framework.KubeDescribe("[Feature:Example]", func() {
}
wg.Wait()
if !passed {
e2elog.Failf("At least one liveness example failed. See the logs above.")
framework.Failf("At least one liveness example failed. See the logs above.")
}
})
})
Expand Down
9 changes: 4 additions & 5 deletions test/e2e/gke_local_ssd.go
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,6 @@ import (
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/util/uuid"
"k8s.io/kubernetes/test/e2e/framework"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"

"github.com/onsi/ginkgo"
)
Expand All @@ -38,23 +37,23 @@ var _ = framework.KubeDescribe("GKE local SSD [Feature:GKELocalSSD]", func() {
})

ginkgo.It("should write and read from node local SSD [Feature:GKELocalSSD]", func() {
e2elog.Logf("Start local SSD test")
framework.Logf("Start local SSD test")
createNodePoolWithLocalSsds("np-ssd")
doTestWriteAndReadToLocalSsd(f)
})
})

func createNodePoolWithLocalSsds(nodePoolName string) {
e2elog.Logf("Create node pool: %s with local SSDs in cluster: %s ",
framework.Logf("Create node pool: %s with local SSDs in cluster: %s ",
nodePoolName, framework.TestContext.CloudConfig.Cluster)
out, err := exec.Command("gcloud", "alpha", "container", "node-pools", "create",
nodePoolName,
fmt.Sprintf("--cluster=%s", framework.TestContext.CloudConfig.Cluster),
"--local-ssd-count=1").CombinedOutput()
if err != nil {
e2elog.Failf("Failed to create node pool %s: Err: %v\n%v", nodePoolName, err, string(out))
framework.Failf("Failed to create node pool %s: Err: %v\n%v", nodePoolName, err, string(out))
}
e2elog.Logf("Successfully created node pool %s:\n%v", nodePoolName, string(out))
framework.Logf("Successfully created node pool %s:\n%v", nodePoolName, string(out))
}

func doTestWriteAndReadToLocalSsd(f *framework.Framework) {
Expand Down
1 change: 0 additions & 1 deletion test/e2e/node/BUILD
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,6 @@ go_library(
"//test/e2e/framework:go_default_library",
"//test/e2e/framework/job:go_default_library",
"//test/e2e/framework/kubelet:go_default_library",
"//test/e2e/framework/log:go_default_library",
"//test/e2e/framework/metrics:go_default_library",
"//test/e2e/framework/node:go_default_library",
"//test/e2e/framework/perf:go_default_library",
Expand Down
3 changes: 1 addition & 2 deletions test/e2e/node/apparmor.go
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,6 @@ package node
import (
"k8s.io/kubernetes/test/e2e/common"
"k8s.io/kubernetes/test/e2e/framework"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"

"github.com/onsi/ginkgo"
)
Expand All @@ -36,7 +35,7 @@ var _ = SIGDescribe("AppArmor", func() {
if !ginkgo.CurrentGinkgoTestDescription().Failed {
return
}
framework.LogFailedContainers(f.ClientSet, f.Namespace.Name, e2elog.Logf)
framework.LogFailedContainers(f.ClientSet, f.Namespace.Name, framework.Logf)
})

ginkgo.It("should enforce an AppArmor profile", func() {
Expand Down
9 changes: 4 additions & 5 deletions test/e2e/node/crictl.go
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@ import (
"strings"

"k8s.io/kubernetes/test/e2e/framework"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2essh "k8s.io/kubernetes/test/e2e/framework/ssh"

"github.com/onsi/ginkgo"
Expand All @@ -42,7 +41,7 @@ var _ = SIGDescribe("crictl", func() {
ginkgo.By("Getting all nodes' SSH-able IP addresses")
hosts, err := e2essh.NodeSSHHosts(f.ClientSet)
if err != nil {
e2elog.Failf("Error getting node hostnames: %v", err)
framework.Failf("Error getting node hostnames: %v", err)
}

testCases := []struct {
Expand All @@ -60,15 +59,15 @@ var _ = SIGDescribe("crictl", func() {
result, err := e2essh.SSH(testCase.cmd, host, framework.TestContext.Provider)
stdout, stderr := strings.TrimSpace(result.Stdout), strings.TrimSpace(result.Stderr)
if err != nil {
e2elog.Failf("Ran %q on %q, got error %v", testCase.cmd, host, err)
framework.Failf("Ran %q on %q, got error %v", testCase.cmd, host, err)
}
// Log the stdout/stderr output.
// TODO: Verify the output.
if len(stdout) > 0 {
e2elog.Logf("Got stdout from %q:\n %s\n", host, strings.TrimSpace(stdout))
framework.Logf("Got stdout from %q:\n %s\n", host, strings.TrimSpace(stdout))
}
if len(stderr) > 0 {
e2elog.Logf("Got stderr from %q:\n %s\n", host, strings.TrimSpace(stderr))
framework.Logf("Got stderr from %q:\n %s\n", host, strings.TrimSpace(stderr))
}
}
})
Expand Down
11 changes: 5 additions & 6 deletions test/e2e/node/events.go
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,6 @@ import (
"k8s.io/apimachinery/pkg/util/uuid"
"k8s.io/apimachinery/pkg/util/wait"
"k8s.io/kubernetes/test/e2e/framework"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"

"github.com/onsi/ginkgo"
)
Expand Down Expand Up @@ -73,7 +72,7 @@ var _ = SIGDescribe("Events", func() {
podClient.Delete(pod.Name, nil)
}()
if _, err := podClient.Create(pod); err != nil {
e2elog.Failf("Failed to create pod: %v", err)
framework.Failf("Failed to create pod: %v", err)
}

framework.ExpectNoError(f.WaitForPodRunning(pod.Name))
Expand All @@ -87,9 +86,9 @@ var _ = SIGDescribe("Events", func() {
ginkgo.By("retrieving the pod")
podWithUID, err := podClient.Get(pod.Name, metav1.GetOptions{})
if err != nil {
e2elog.Failf("Failed to get pod: %v", err)
framework.Failf("Failed to get pod: %v", err)
}
e2elog.Logf("%+v\n", podWithUID)
framework.Logf("%+v\n", podWithUID)
var events *v1.EventList
// Check for scheduler event about the pod.
ginkgo.By("checking for scheduler event about the pod")
Expand All @@ -106,7 +105,7 @@ var _ = SIGDescribe("Events", func() {
return false, err
}
if len(events.Items) > 0 {
e2elog.Logf("Saw scheduler event for our pod.")
framework.Logf("Saw scheduler event for our pod.")
return true, nil
}
return false, nil
Expand All @@ -126,7 +125,7 @@ var _ = SIGDescribe("Events", func() {
return false, err
}
if len(events.Items) > 0 {
e2elog.Logf("Saw kubelet event for our pod.")
framework.Logf("Saw kubelet event for our pod.")
return true, nil
}
return false, nil
Expand Down
15 changes: 7 additions & 8 deletions test/e2e/node/kubelet.go
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,6 @@ import (
clientset "k8s.io/client-go/kubernetes"
"k8s.io/kubernetes/test/e2e/framework"
e2ekubelet "k8s.io/kubernetes/test/e2e/framework/kubelet"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
e2essh "k8s.io/kubernetes/test/e2e/framework/ssh"
"k8s.io/kubernetes/test/e2e/framework/volume"
Expand All @@ -53,10 +52,10 @@ const (
// podNamePrefix and namespace.
func getPodMatches(c clientset.Interface, nodeName string, podNamePrefix string, namespace string) sets.String {
matches := sets.NewString()
e2elog.Logf("Checking pods on node %v via /runningpods endpoint", nodeName)
framework.Logf("Checking pods on node %v via /runningpods endpoint", nodeName)
runningPods, err := e2ekubelet.GetKubeletPods(c, nodeName)
if err != nil {
e2elog.Logf("Error checking running pods on %v: %v", nodeName, err)
framework.Logf("Error checking running pods on %v: %v", nodeName, err)
return matches
}
for _, pod := range runningPods.Items {
Expand Down Expand Up @@ -93,7 +92,7 @@ func waitTillNPodsRunningOnNodes(c clientset.Interface, nodeNames sets.String, p
if seen.Len() == targetNumPods {
return true, nil
}
e2elog.Logf("Waiting for %d pods to be running on the node; %d are currently running;", targetNumPods, seen.Len())
framework.Logf("Waiting for %d pods to be running on the node; %d are currently running;", targetNumPods, seen.Len())
return false, nil
})
}
Expand Down Expand Up @@ -214,7 +213,7 @@ func checkPodCleanup(c clientset.Interface, pod *v1.Pod, expectClean bool) {
}

for _, test := range tests {
e2elog.Logf("Wait up to %v for host's (%v) %q to be %v", timeout, nodeIP, test.feature, condMsg)
framework.Logf("Wait up to %v for host's (%v) %q to be %v", timeout, nodeIP, test.feature, condMsg)
err = wait.Poll(poll, timeout, func() (bool, error) {
result, err := e2essh.NodeExec(nodeIP, test.cmd, framework.TestContext.Provider)
framework.ExpectNoError(err)
Expand All @@ -232,9 +231,9 @@ func checkPodCleanup(c clientset.Interface, pod *v1.Pod, expectClean bool) {
}

if expectClean {
e2elog.Logf("Pod's host has been cleaned up")
framework.Logf("Pod's host has been cleaned up")
} else {
e2elog.Logf("Pod's host has not been cleaned up (per expectation)")
framework.Logf("Pod's host has not been cleaned up (per expectation)")
}
}

Expand Down Expand Up @@ -349,7 +348,7 @@ var _ = SIGDescribe("kubelet", func() {
start := time.Now()
err = waitTillNPodsRunningOnNodes(f.ClientSet, nodeNames, rcName, ns, 0, itArg.timeout)
framework.ExpectNoError(err)
e2elog.Logf("Deleting %d pods on %d nodes completed in %v after the RC was deleted", totalPods, len(nodeNames),
framework.Logf("Deleting %d pods on %d nodes completed in %v after the RC was deleted", totalPods, len(nodeNames),
time.Since(start))
if resourceMonitor != nil {
resourceMonitor.LogCPUSummary()
Expand Down
23 changes: 11 additions & 12 deletions test/e2e/node/kubelet_perf.go
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,6 @@ import (
kubeletstatsv1alpha1 "k8s.io/kubernetes/pkg/kubelet/apis/stats/v1alpha1"
"k8s.io/kubernetes/test/e2e/framework"
e2ekubelet "k8s.io/kubernetes/test/e2e/framework/kubelet"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2emetrics "k8s.io/kubernetes/test/e2e/framework/metrics"
e2eperf "k8s.io/kubernetes/test/e2e/framework/perf"
"k8s.io/kubernetes/test/e2e/perftype"
Expand Down Expand Up @@ -58,10 +57,10 @@ func logPodsOnNodes(c clientset.Interface, nodeNames []string) {
for _, n := range nodeNames {
podList, err := e2ekubelet.GetKubeletRunningPods(c, n)
if err != nil {
e2elog.Logf("Unable to retrieve kubelet pods for node %v", n)
framework.Logf("Unable to retrieve kubelet pods for node %v", n)
continue
}
e2elog.Logf("%d pods are running on node %v", len(podList.Items), n)
framework.Logf("%d pods are running on node %v", len(podList.Items), n)
}
}

Expand Down Expand Up @@ -95,7 +94,7 @@ func runResourceTrackingTest(f *framework.Framework, podsPerNode int, nodeNames
deadline := time.Now().Add(monitoringTime)
for time.Now().Before(deadline) {
timeLeft := deadline.Sub(time.Now())
e2elog.Logf("Still running...%v left", timeLeft)
framework.Logf("Still running...%v left", timeLeft)
if timeLeft < reportingPeriod {
time.Sleep(timeLeft)
} else {
Expand All @@ -109,13 +108,13 @@ func runResourceTrackingTest(f *framework.Framework, podsPerNode int, nodeNames
usageSummary, err := rm.GetLatest()
framework.ExpectNoError(err)
// TODO(random-liu): Remove the original log when we migrate to new perfdash
e2elog.Logf("%s", rm.FormatResourceUsage(usageSummary))
framework.Logf("%s", rm.FormatResourceUsage(usageSummary))
// Log perf result
printPerfData(e2eperf.ResourceUsageToPerfData(rm.GetMasterNodeLatest(usageSummary)))
verifyMemoryLimits(f.ClientSet, expectedMemory, usageSummary)

cpuSummary := rm.GetCPUSummary()
e2elog.Logf("%s", rm.FormatCPUSummary(cpuSummary))
framework.Logf("%s", rm.FormatCPUSummary(cpuSummary))
// Log perf result
printPerfData(e2eperf.CPUUsageToPerfData(rm.GetMasterNodeCPUSummary(cpuSummary)))
verifyCPULimits(expectedCPU, cpuSummary)
Expand Down Expand Up @@ -149,14 +148,14 @@ func verifyMemoryLimits(c clientset.Interface, expected e2ekubelet.ResourceUsage
errList = append(errList, fmt.Sprintf("node %v:\n %s", nodeName, strings.Join(nodeErrs, ", ")))
heapStats, err := e2ekubelet.GetKubeletHeapStats(c, nodeName)
if err != nil {
e2elog.Logf("Unable to get heap stats from %q", nodeName)
framework.Logf("Unable to get heap stats from %q", nodeName)
} else {
e2elog.Logf("Heap stats on %q\n:%v", nodeName, heapStats)
framework.Logf("Heap stats on %q\n:%v", nodeName, heapStats)
}
}
}
if len(errList) > 0 {
e2elog.Failf("Memory usage exceeding limits:\n %s", strings.Join(errList, "\n"))
framework.Failf("Memory usage exceeding limits:\n %s", strings.Join(errList, "\n"))
}
}

Expand Down Expand Up @@ -190,7 +189,7 @@ func verifyCPULimits(expected e2ekubelet.ContainersCPUSummary, actual e2ekubelet
}
}
if len(errList) > 0 {
e2elog.Failf("CPU usage exceeding limits:\n %s", strings.Join(errList, "\n"))
framework.Failf("CPU usage exceeding limits:\n %s", strings.Join(errList, "\n"))
}
}

Expand All @@ -215,7 +214,7 @@ var _ = SIGDescribe("Kubelet [Serial] [Slow]", func() {
ginkgo.AfterEach(func() {
rm.Stop()
result := om.GetLatestRuntimeOperationErrorRate()
e2elog.Logf("runtime operation error metrics:\n%s", e2ekubelet.FormatRuntimeOperationErrorRate(result))
framework.Logf("runtime operation error metrics:\n%s", e2ekubelet.FormatRuntimeOperationErrorRate(result))
})
SIGDescribe("regular resource usage tracking [Feature:RegularResourceUsageTracking]", func() {
// We assume that the scheduler will make reasonable scheduling choices
Expand Down Expand Up @@ -287,6 +286,6 @@ var _ = SIGDescribe("Kubelet [Serial] [Slow]", func() {
func printPerfData(p *perftype.PerfData) {
// Notice that we must make sure the perftype.PerfResultEnd is in a new line.
if str := e2emetrics.PrettyPrintJSON(p); str != "" {
e2elog.Logf("%s %s\n%s", perftype.PerfResultTag, str, perftype.PerfResultEnd)
framework.Logf("%s %s\n%s", perftype.PerfResultTag, str, perftype.PerfResultEnd)
}
}
3 changes: 1 addition & 2 deletions test/e2e/node/mount_propagation.go
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,6 @@ import (
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/util/sets"
"k8s.io/kubernetes/test/e2e/framework"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2essh "k8s.io/kubernetes/test/e2e/framework/ssh"
imageutils "k8s.io/kubernetes/test/utils/image"

Expand Down Expand Up @@ -166,7 +165,7 @@ var _ = SIGDescribe("Mount propagation", func() {
for _, mountName := range dirNames {
cmd := fmt.Sprintf("cat /mnt/test/%s/file", mountName)
stdout, stderr, err := f.ExecShellInPodWithFullOutput(podName, cmd)
e2elog.Logf("pod %s mount %s: stdout: %q, stderr: %q error: %v", podName, mountName, stdout, stderr, err)
framework.Logf("pod %s mount %s: stdout: %q, stderr: %q error: %v", podName, mountName, stdout, stderr, err)
msg := fmt.Sprintf("When checking pod %s and directory %s", podName, mountName)
shouldBeVisible := mounts.Has(mountName)
if shouldBeVisible {
Expand Down
3 changes: 1 addition & 2 deletions test/e2e/node/node_problem_detector.go
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,6 @@ import (
"k8s.io/apimachinery/pkg/fields"
"k8s.io/kubernetes/test/e2e/framework"
e2ekubelet "k8s.io/kubernetes/test/e2e/framework/kubelet"
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
e2essh "k8s.io/kubernetes/test/e2e/framework/ssh"
testutils "k8s.io/kubernetes/test/utils"
Expand Down Expand Up @@ -188,7 +187,7 @@ var _ = SIGDescribe("NodeProblemDetector [DisabledForLargeClusters]", func() {
workingSetStatsMsg += fmt.Sprintf(" %s[%.1f|%.1f|%.1f];", nodes.Items[i].Name,
workingSetStats[host][0], workingSetStats[host][len(workingSetStats[host])/2], workingSetStats[host][len(workingSetStats[host])-1])
}
e2elog.Logf("Node-Problem-Detector CPU and Memory Stats:\n\t%s\n\t%s\n\t%s", cpuStatsMsg, rssStatsMsg, workingSetStatsMsg)
framework.Logf("Node-Problem-Detector CPU and Memory Stats:\n\t%s\n\t%s\n\t%s", cpuStatsMsg, rssStatsMsg, workingSetStatsMsg)
})
})

Expand Down