mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			512 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			512 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// +build cgo,linux
 | 
						|
 | 
						|
/*
 | 
						|
Copyright 2016 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package e2enode
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"fmt"
 | 
						|
	"os"
 | 
						|
	"path"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/onsi/ginkgo"
 | 
						|
	"github.com/onsi/gomega"
 | 
						|
	v1 "k8s.io/api/core/v1"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/fields"
 | 
						|
	"k8s.io/apimachinery/pkg/labels"
 | 
						|
	"k8s.io/apimachinery/pkg/types"
 | 
						|
	"k8s.io/apimachinery/pkg/util/uuid"
 | 
						|
	clientset "k8s.io/client-go/kubernetes"
 | 
						|
	coreclientset "k8s.io/client-go/kubernetes/typed/core/v1"
 | 
						|
 | 
						|
	"k8s.io/kubernetes/pkg/kubelet/util"
 | 
						|
	"k8s.io/kubernetes/test/e2e/framework"
 | 
						|
	e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
 | 
						|
	testutils "k8s.io/kubernetes/test/utils"
 | 
						|
)
 | 
						|
 | 
						|
var _ = SIGDescribe("NodeProblemDetector [NodeFeature:NodeProblemDetector] [Serial]", func() {
 | 
						|
	const (
 | 
						|
		pollInterval   = 1 * time.Second
 | 
						|
		pollConsistent = 5 * time.Second
 | 
						|
		pollTimeout    = 1 * time.Minute
 | 
						|
	)
 | 
						|
	f := framework.NewDefaultFramework("node-problem-detector")
 | 
						|
	var c clientset.Interface
 | 
						|
	var uid string
 | 
						|
	var ns, name, configName, eventNamespace string
 | 
						|
	var bootTime, nodeTime time.Time
 | 
						|
	var image string
 | 
						|
 | 
						|
	ginkgo.BeforeEach(func() {
 | 
						|
		c = f.ClientSet
 | 
						|
		ns = f.Namespace.Name
 | 
						|
		uid = string(uuid.NewUUID())
 | 
						|
		name = "node-problem-detector-" + uid
 | 
						|
		configName = "node-problem-detector-config-" + uid
 | 
						|
		// There is no namespace for Node, event recorder will set default namespace for node events.
 | 
						|
		eventNamespace = metav1.NamespaceDefault
 | 
						|
		image = getNodeProblemDetectorImage()
 | 
						|
		ginkgo.By(fmt.Sprintf("Using node-problem-detector image: %s", image))
 | 
						|
	})
 | 
						|
 | 
						|
	// Test system log monitor. We may add other tests if we have more problem daemons in the future.
 | 
						|
	ginkgo.Describe("SystemLogMonitor", func() {
 | 
						|
		const (
 | 
						|
			// Use test condition to avoid changing the real node condition in use.
 | 
						|
			// TODO(random-liu): Now node condition could be arbitrary string, consider whether we need to
 | 
						|
			// add TestCondition when switching to predefined condition list.
 | 
						|
			condition = v1.NodeConditionType("TestCondition")
 | 
						|
 | 
						|
			// File paths used in the test.
 | 
						|
			logFile        = "/log/test.log"
 | 
						|
			configFile     = "/config/testconfig.json"
 | 
						|
			kubeConfigFile = "/config/kubeconfig"
 | 
						|
			etcLocaltime   = "/etc/localtime"
 | 
						|
 | 
						|
			// Volumes used in the test.
 | 
						|
			configVolume    = "config"
 | 
						|
			logVolume       = "log"
 | 
						|
			localtimeVolume = "localtime"
 | 
						|
 | 
						|
			// Reasons and messages used in the test.
 | 
						|
			defaultReason  = "Default"
 | 
						|
			defaultMessage = "default message"
 | 
						|
			tempReason     = "Temporary"
 | 
						|
			tempMessage    = "temporary error"
 | 
						|
			permReason1    = "Permanent1"
 | 
						|
			permMessage1   = "permanent error 1"
 | 
						|
			permReason2    = "Permanent2"
 | 
						|
			permMessage2   = "permanent error 2"
 | 
						|
		)
 | 
						|
		var source, config, hostLogFile string
 | 
						|
		var lookback time.Duration
 | 
						|
		var eventListOptions metav1.ListOptions
 | 
						|
 | 
						|
		ginkgo.BeforeEach(func() {
 | 
						|
			ginkgo.By("Calculate Lookback duration")
 | 
						|
			var err error
 | 
						|
 | 
						|
			nodeTime = time.Now()
 | 
						|
			bootTime, err = util.GetBootTime()
 | 
						|
			framework.ExpectNoError(err)
 | 
						|
 | 
						|
			// Set lookback duration longer than node up time.
 | 
						|
			// Assume the test won't take more than 1 hour, in fact it usually only takes 90 seconds.
 | 
						|
			lookback = nodeTime.Sub(bootTime) + time.Hour
 | 
						|
 | 
						|
			// Randomize the source name
 | 
						|
			source = "kernel-monitor-" + uid
 | 
						|
			config = `
 | 
						|
			{
 | 
						|
				"plugin": "filelog",
 | 
						|
				"pluginConfig": {
 | 
						|
					"timestamp": "^.{15}",
 | 
						|
					"message": "kernel: \\[.*\\] (.*)",
 | 
						|
					"timestampFormat": "` + time.Stamp + `"
 | 
						|
				},
 | 
						|
				"logPath": "` + logFile + `",
 | 
						|
				"lookback": "` + lookback.String() + `",
 | 
						|
				"bufferSize": 10,
 | 
						|
				"source": "` + source + `",
 | 
						|
				"conditions": [
 | 
						|
				{
 | 
						|
					"type": "` + string(condition) + `",
 | 
						|
					"reason": "` + defaultReason + `",
 | 
						|
					"message": "` + defaultMessage + `"
 | 
						|
				}
 | 
						|
				],
 | 
						|
				"rules": [
 | 
						|
				{
 | 
						|
					"type": "temporary",
 | 
						|
					"reason": "` + tempReason + `",
 | 
						|
					"pattern": "` + tempMessage + `"
 | 
						|
				},
 | 
						|
				{
 | 
						|
					"type": "permanent",
 | 
						|
					"condition": "` + string(condition) + `",
 | 
						|
					"reason": "` + permReason1 + `",
 | 
						|
					"pattern": "` + permMessage1 + ".*" + `"
 | 
						|
				},
 | 
						|
				{
 | 
						|
					"type": "permanent",
 | 
						|
					"condition": "` + string(condition) + `",
 | 
						|
					"reason": "` + permReason2 + `",
 | 
						|
					"pattern": "` + permMessage2 + ".*" + `"
 | 
						|
				}
 | 
						|
				]
 | 
						|
			}`
 | 
						|
 | 
						|
			// This token is known to apiserver and its group is `system:masters`.
 | 
						|
			// See also the function `generateTokenFile` in `test/e2e_node/services/apiserver.go`.
 | 
						|
			kubeConfig := fmt.Sprintf(`
 | 
						|
apiVersion: v1
 | 
						|
kind: Config
 | 
						|
users:
 | 
						|
- name: node-problem-detector
 | 
						|
  user:
 | 
						|
    token: %s
 | 
						|
clusters:
 | 
						|
- cluster:
 | 
						|
    server: %s
 | 
						|
    insecure-skip-tls-verify: true
 | 
						|
  name: local
 | 
						|
contexts:
 | 
						|
- context:
 | 
						|
    cluster: local
 | 
						|
    user: node-problem-detector
 | 
						|
  name: local-context
 | 
						|
current-context: local-context
 | 
						|
`, framework.TestContext.BearerToken, framework.TestContext.Host)
 | 
						|
 | 
						|
			ginkgo.By("Generate event list options")
 | 
						|
			selector := fields.Set{
 | 
						|
				"involvedObject.kind":      "Node",
 | 
						|
				"involvedObject.name":      framework.TestContext.NodeName,
 | 
						|
				"involvedObject.namespace": metav1.NamespaceAll,
 | 
						|
				"source":                   source,
 | 
						|
			}.AsSelector().String()
 | 
						|
			eventListOptions = metav1.ListOptions{FieldSelector: selector}
 | 
						|
 | 
						|
			ginkgo.By("Create config map for the node problem detector")
 | 
						|
			_, err = c.CoreV1().ConfigMaps(ns).Create(context.TODO(), &v1.ConfigMap{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{Name: configName},
 | 
						|
				Data: map[string]string{
 | 
						|
					path.Base(configFile):     config,
 | 
						|
					path.Base(kubeConfigFile): kubeConfig,
 | 
						|
				},
 | 
						|
			}, metav1.CreateOptions{})
 | 
						|
			framework.ExpectNoError(err)
 | 
						|
 | 
						|
			ginkgo.By("Create the node problem detector")
 | 
						|
			hostPathType := new(v1.HostPathType)
 | 
						|
			*hostPathType = v1.HostPathFileOrCreate
 | 
						|
			pod := f.PodClient().CreateSync(&v1.Pod{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Name: name,
 | 
						|
				},
 | 
						|
				Spec: v1.PodSpec{
 | 
						|
					HostNetwork:        true,
 | 
						|
					SecurityContext:    &v1.PodSecurityContext{},
 | 
						|
					ServiceAccountName: name,
 | 
						|
					Volumes: []v1.Volume{
 | 
						|
						{
 | 
						|
							Name: configVolume,
 | 
						|
							VolumeSource: v1.VolumeSource{
 | 
						|
								ConfigMap: &v1.ConfigMapVolumeSource{
 | 
						|
									LocalObjectReference: v1.LocalObjectReference{Name: configName},
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
						{
 | 
						|
							Name: logVolume,
 | 
						|
							VolumeSource: v1.VolumeSource{
 | 
						|
								EmptyDir: &v1.EmptyDirVolumeSource{},
 | 
						|
							},
 | 
						|
						},
 | 
						|
						{
 | 
						|
							Name: localtimeVolume,
 | 
						|
							VolumeSource: v1.VolumeSource{
 | 
						|
								HostPath: &v1.HostPathVolumeSource{
 | 
						|
									Path: etcLocaltime,
 | 
						|
									Type: hostPathType,
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
					InitContainers: []v1.Container{
 | 
						|
						{
 | 
						|
							Name:    "init-log-file",
 | 
						|
							Image:   "debian",
 | 
						|
							Command: []string{"/bin/sh"},
 | 
						|
							Args: []string{
 | 
						|
								"-c",
 | 
						|
								fmt.Sprintf("touch %s", logFile),
 | 
						|
							},
 | 
						|
							VolumeMounts: []v1.VolumeMount{
 | 
						|
								{
 | 
						|
									Name:      logVolume,
 | 
						|
									MountPath: path.Dir(logFile),
 | 
						|
								},
 | 
						|
								{
 | 
						|
									Name:      localtimeVolume,
 | 
						|
									MountPath: etcLocaltime,
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
					Containers: []v1.Container{
 | 
						|
						{
 | 
						|
							Name:    name,
 | 
						|
							Image:   image,
 | 
						|
							Command: []string{"/node-problem-detector"},
 | 
						|
							Args: []string{
 | 
						|
								"--logtostderr",
 | 
						|
								fmt.Sprintf("--system-log-monitors=%s", configFile),
 | 
						|
								// `ServiceAccount` admission controller is disabled in node e2e tests, so we could not use
 | 
						|
								// inClusterConfig here.
 | 
						|
								fmt.Sprintf("--apiserver-override=%s?inClusterConfig=false&auth=%s", framework.TestContext.Host, kubeConfigFile),
 | 
						|
							},
 | 
						|
							Env: []v1.EnvVar{
 | 
						|
								{
 | 
						|
									Name: "NODE_NAME",
 | 
						|
									ValueFrom: &v1.EnvVarSource{
 | 
						|
										FieldRef: &v1.ObjectFieldSelector{
 | 
						|
											APIVersion: "v1",
 | 
						|
											FieldPath:  "spec.nodeName",
 | 
						|
										},
 | 
						|
									},
 | 
						|
								},
 | 
						|
							},
 | 
						|
							VolumeMounts: []v1.VolumeMount{
 | 
						|
								{
 | 
						|
									Name:      logVolume,
 | 
						|
									MountPath: path.Dir(logFile),
 | 
						|
								},
 | 
						|
								{
 | 
						|
									Name:      localtimeVolume,
 | 
						|
									MountPath: etcLocaltime,
 | 
						|
								},
 | 
						|
								{
 | 
						|
									Name:      configVolume,
 | 
						|
									MountPath: path.Dir(configFile),
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			})
 | 
						|
			// TODO: remove hardcoded kubelet volume directory path
 | 
						|
			// framework.TestContext.KubeVolumeDir is currently not populated for node e2e
 | 
						|
			hostLogFile = "/var/lib/kubelet/pods/" + string(pod.UID) + "/volumes/kubernetes.io~empty-dir" + logFile
 | 
						|
		})
 | 
						|
 | 
						|
		ginkgo.It("should generate node condition and events for corresponding errors", func() {
 | 
						|
			for _, test := range []struct {
 | 
						|
				description      string
 | 
						|
				timestamp        time.Time
 | 
						|
				message          string
 | 
						|
				messageNum       int
 | 
						|
				tempEvents       int // Events for temp errors
 | 
						|
				totalEvents      int // Events for both temp errors and condition changes
 | 
						|
				conditionReason  string
 | 
						|
				conditionMessage string
 | 
						|
				conditionType    v1.ConditionStatus
 | 
						|
			}{
 | 
						|
				{
 | 
						|
					description:      "should generate default node condition",
 | 
						|
					conditionReason:  defaultReason,
 | 
						|
					conditionMessage: defaultMessage,
 | 
						|
					conditionType:    v1.ConditionFalse,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should not generate events for too old log",
 | 
						|
					timestamp:        bootTime.Add(-1 * time.Minute),
 | 
						|
					message:          tempMessage,
 | 
						|
					messageNum:       3,
 | 
						|
					conditionReason:  defaultReason,
 | 
						|
					conditionMessage: defaultMessage,
 | 
						|
					conditionType:    v1.ConditionFalse,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should not change node condition for too old log",
 | 
						|
					timestamp:        bootTime.Add(-1 * time.Minute),
 | 
						|
					message:          permMessage1,
 | 
						|
					messageNum:       1,
 | 
						|
					conditionReason:  defaultReason,
 | 
						|
					conditionMessage: defaultMessage,
 | 
						|
					conditionType:    v1.ConditionFalse,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should generate event for old log within lookback duration",
 | 
						|
					timestamp:        nodeTime,
 | 
						|
					message:          tempMessage,
 | 
						|
					messageNum:       3,
 | 
						|
					tempEvents:       3,
 | 
						|
					totalEvents:      3,
 | 
						|
					conditionReason:  defaultReason,
 | 
						|
					conditionMessage: defaultMessage,
 | 
						|
					conditionType:    v1.ConditionFalse,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should change node condition for old log within lookback duration",
 | 
						|
					timestamp:        nodeTime,
 | 
						|
					message:          permMessage1,
 | 
						|
					messageNum:       1,
 | 
						|
					tempEvents:       3, // event number for temp errors should not change
 | 
						|
					totalEvents:      4, // add 1 event for condition change
 | 
						|
					conditionReason:  permReason1,
 | 
						|
					conditionMessage: permMessage1,
 | 
						|
					conditionType:    v1.ConditionTrue,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should generate event for new log",
 | 
						|
					timestamp:        nodeTime.Add(5 * time.Minute),
 | 
						|
					message:          tempMessage,
 | 
						|
					messageNum:       3,
 | 
						|
					tempEvents:       6, // add 3 events for temp errors
 | 
						|
					totalEvents:      7, // add 3 events for temp errors
 | 
						|
					conditionReason:  permReason1,
 | 
						|
					conditionMessage: permMessage1,
 | 
						|
					conditionType:    v1.ConditionTrue,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should not update node condition with the same reason",
 | 
						|
					timestamp:        nodeTime.Add(5 * time.Minute),
 | 
						|
					message:          permMessage1 + "different message",
 | 
						|
					messageNum:       1,
 | 
						|
					tempEvents:       6, // event number should not change
 | 
						|
					totalEvents:      7, // event number should not change
 | 
						|
					conditionReason:  permReason1,
 | 
						|
					conditionMessage: permMessage1,
 | 
						|
					conditionType:    v1.ConditionTrue,
 | 
						|
				},
 | 
						|
				{
 | 
						|
					description:      "should change node condition for new log",
 | 
						|
					timestamp:        nodeTime.Add(5 * time.Minute),
 | 
						|
					message:          permMessage2,
 | 
						|
					messageNum:       1,
 | 
						|
					tempEvents:       6, // event number for temp errors should not change
 | 
						|
					totalEvents:      8, // add 1 event for condition change
 | 
						|
					conditionReason:  permReason2,
 | 
						|
					conditionMessage: permMessage2,
 | 
						|
					conditionType:    v1.ConditionTrue,
 | 
						|
				},
 | 
						|
			} {
 | 
						|
				ginkgo.By(test.description)
 | 
						|
				if test.messageNum > 0 {
 | 
						|
					ginkgo.By(fmt.Sprintf("Inject %d logs: %q", test.messageNum, test.message))
 | 
						|
					err := injectLog(hostLogFile, test.timestamp, test.message, test.messageNum)
 | 
						|
					framework.ExpectNoError(err)
 | 
						|
				}
 | 
						|
 | 
						|
				ginkgo.By(fmt.Sprintf("Wait for %d temp events generated", test.tempEvents))
 | 
						|
				gomega.Eventually(func() error {
 | 
						|
					return verifyEvents(c.CoreV1().Events(eventNamespace), eventListOptions, test.tempEvents, tempReason, tempMessage)
 | 
						|
				}, pollTimeout, pollInterval).Should(gomega.Succeed())
 | 
						|
				ginkgo.By(fmt.Sprintf("Wait for %d total events generated", test.totalEvents))
 | 
						|
				gomega.Eventually(func() error {
 | 
						|
					return verifyTotalEvents(c.CoreV1().Events(eventNamespace), eventListOptions, test.totalEvents)
 | 
						|
				}, pollTimeout, pollInterval).Should(gomega.Succeed())
 | 
						|
				ginkgo.By(fmt.Sprintf("Make sure only %d total events generated", test.totalEvents))
 | 
						|
				gomega.Consistently(func() error {
 | 
						|
					return verifyTotalEvents(c.CoreV1().Events(eventNamespace), eventListOptions, test.totalEvents)
 | 
						|
				}, pollConsistent, pollInterval).Should(gomega.Succeed())
 | 
						|
 | 
						|
				ginkgo.By(fmt.Sprintf("Make sure node condition %q is set", condition))
 | 
						|
				gomega.Eventually(func() error {
 | 
						|
					return verifyNodeCondition(c.CoreV1().Nodes(), condition, test.conditionType, test.conditionReason, test.conditionMessage)
 | 
						|
				}, pollTimeout, pollInterval).Should(gomega.Succeed())
 | 
						|
				ginkgo.By(fmt.Sprintf("Make sure node condition %q is stable", condition))
 | 
						|
				gomega.Consistently(func() error {
 | 
						|
					return verifyNodeCondition(c.CoreV1().Nodes(), condition, test.conditionType, test.conditionReason, test.conditionMessage)
 | 
						|
				}, pollConsistent, pollInterval).Should(gomega.Succeed())
 | 
						|
			}
 | 
						|
		})
 | 
						|
 | 
						|
		ginkgo.AfterEach(func() {
 | 
						|
			if ginkgo.CurrentGinkgoTestDescription().Failed && framework.TestContext.DumpLogsOnFailure {
 | 
						|
				ginkgo.By("Get node problem detector log")
 | 
						|
				log, err := e2epod.GetPodLogs(c, ns, name, name)
 | 
						|
				gomega.Expect(err).ShouldNot(gomega.HaveOccurred())
 | 
						|
				framework.Logf("Node Problem Detector logs:\n %s", log)
 | 
						|
			}
 | 
						|
			ginkgo.By("Delete the node problem detector")
 | 
						|
			f.PodClient().Delete(context.TODO(), name, *metav1.NewDeleteOptions(0))
 | 
						|
			ginkgo.By("Wait for the node problem detector to disappear")
 | 
						|
			gomega.Expect(e2epod.WaitForPodToDisappear(c, ns, name, labels.Everything(), pollInterval, pollTimeout)).To(gomega.Succeed())
 | 
						|
			ginkgo.By("Delete the config map")
 | 
						|
			c.CoreV1().ConfigMaps(ns).Delete(context.TODO(), configName, metav1.DeleteOptions{})
 | 
						|
			ginkgo.By("Clean up the events")
 | 
						|
			gomega.Expect(c.CoreV1().Events(eventNamespace).DeleteCollection(context.TODO(), *metav1.NewDeleteOptions(0), eventListOptions)).To(gomega.Succeed())
 | 
						|
			ginkgo.By("Clean up the node condition")
 | 
						|
			patch := []byte(fmt.Sprintf(`{"status":{"conditions":[{"$patch":"delete","type":"%s"}]}}`, condition))
 | 
						|
			c.CoreV1().RESTClient().Patch(types.StrategicMergePatchType).Resource("nodes").Name(framework.TestContext.NodeName).SubResource("status").Body(patch).Do(context.TODO())
 | 
						|
		})
 | 
						|
	})
 | 
						|
})
 | 
						|
 | 
						|
// injectLog injects kernel log into specified file.
 | 
						|
func injectLog(file string, timestamp time.Time, log string, num int) error {
 | 
						|
	f, err := os.OpenFile(file, os.O_RDWR|os.O_APPEND, 0666)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	defer f.Close()
 | 
						|
	for i := 0; i < num; i++ {
 | 
						|
		_, err := f.WriteString(fmt.Sprintf("%s kernel: [0.000000] %s\n", timestamp.Format(time.Stamp), log))
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// verifyEvents verifies there are num specific events generated with given reason and message.
 | 
						|
func verifyEvents(e coreclientset.EventInterface, options metav1.ListOptions, num int, reason, message string) error {
 | 
						|
	events, err := e.List(context.TODO(), options)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	count := 0
 | 
						|
	for _, event := range events.Items {
 | 
						|
		if event.Reason != reason || event.Message != message {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		count += int(event.Count)
 | 
						|
	}
 | 
						|
	if count != num {
 | 
						|
		return fmt.Errorf("expected %d events with reason set to %s and message set to %s\nbut %d actual events occurred. Events : %v", num, reason, message, count, events.Items)
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// verifyTotalEvents verifies there are num events in total.
 | 
						|
func verifyTotalEvents(e coreclientset.EventInterface, options metav1.ListOptions, num int) error {
 | 
						|
	events, err := e.List(context.TODO(), options)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	count := 0
 | 
						|
	for _, event := range events.Items {
 | 
						|
		count += int(event.Count)
 | 
						|
	}
 | 
						|
	if count != num {
 | 
						|
		return fmt.Errorf("expected total number of events was %d, actual events counted was %d\nEvents  : %v", num, count, events.Items)
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// verifyNodeCondition verifies specific node condition is generated, if reason and message are empty, they will not be checked
 | 
						|
func verifyNodeCondition(n coreclientset.NodeInterface, condition v1.NodeConditionType, status v1.ConditionStatus, reason, message string) error {
 | 
						|
	node, err := n.Get(context.TODO(), framework.TestContext.NodeName, metav1.GetOptions{})
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	_, c := testutils.GetNodeCondition(&node.Status, condition)
 | 
						|
	if c == nil {
 | 
						|
		return fmt.Errorf("node condition %q not found", condition)
 | 
						|
	}
 | 
						|
	if c.Status != status || c.Reason != reason || c.Message != message {
 | 
						|
		return fmt.Errorf("unexpected node condition %q: %+v", condition, c)
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 |