mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			349 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			349 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2015 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
/*
 | 
						|
This soak tests places a specified number of pods on each node and then
 | 
						|
repeatedly sends queries to a service running on these pods via
 | 
						|
a serivce
 | 
						|
*/
 | 
						|
 | 
						|
package main
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"flag"
 | 
						|
	"fmt"
 | 
						|
	"os"
 | 
						|
	"path/filepath"
 | 
						|
	"time"
 | 
						|
 | 
						|
	v1 "k8s.io/api/core/v1"
 | 
						|
	apierrors "k8s.io/apimachinery/pkg/api/errors"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/runtime"
 | 
						|
	"k8s.io/apimachinery/pkg/util/intstr"
 | 
						|
	clientset "k8s.io/client-go/kubernetes"
 | 
						|
	restclient "k8s.io/client-go/rest"
 | 
						|
	"k8s.io/client-go/tools/clientcmd"
 | 
						|
	"k8s.io/kubernetes/pkg/api/legacyscheme"
 | 
						|
	e2e "k8s.io/kubernetes/test/e2e/framework"
 | 
						|
	"k8s.io/kubernetes/test/e2e/framework/service"
 | 
						|
 | 
						|
	"k8s.io/klog"
 | 
						|
)
 | 
						|
 | 
						|
var (
 | 
						|
	queriesAverage = flag.Int("queries", 100, "Number of hostname queries to make in each iteration per pod on average")
 | 
						|
	podsPerNode    = flag.Int("pods_per_node", 1, "Number of serve_hostname pods per node")
 | 
						|
	upTo           = flag.Int("up_to", 1, "Number of iterations or -1 for no limit")
 | 
						|
	maxPar         = flag.Int("max_par", 500, "Maximum number of queries in flight")
 | 
						|
	gke            = flag.String("gke_context", "", "Target GKE cluster with context gke_{project}_{zone}_{cluster-name}")
 | 
						|
)
 | 
						|
 | 
						|
const (
 | 
						|
	deleteTimeout          = 2 * time.Minute
 | 
						|
	endpointTimeout        = 5 * time.Minute
 | 
						|
	nodeListTimeout        = 2 * time.Minute
 | 
						|
	podCreateTimeout       = 2 * time.Minute
 | 
						|
	podStartTimeout        = 30 * time.Minute
 | 
						|
	serviceCreateTimeout   = 2 * time.Minute
 | 
						|
	namespaceDeleteTimeout = 5 * time.Minute
 | 
						|
)
 | 
						|
 | 
						|
func main() {
 | 
						|
	flag.Parse()
 | 
						|
 | 
						|
	klog.Infof("Starting serve_hostnames soak test with queries=%d and podsPerNode=%d upTo=%d",
 | 
						|
		*queriesAverage, *podsPerNode, *upTo)
 | 
						|
 | 
						|
	var spec string
 | 
						|
	if *gke != "" {
 | 
						|
		spec = filepath.Join(os.Getenv("HOME"), ".config", "gcloud", "kubernetes", "kubeconfig")
 | 
						|
	} else {
 | 
						|
		spec = filepath.Join(os.Getenv("HOME"), ".kube", "config")
 | 
						|
	}
 | 
						|
	settings, err := clientcmd.LoadFromFile(spec)
 | 
						|
	if err != nil {
 | 
						|
		klog.Fatalf("Error loading configuration: %v", err.Error())
 | 
						|
	}
 | 
						|
	if *gke != "" {
 | 
						|
		settings.CurrentContext = *gke
 | 
						|
	}
 | 
						|
	config, err := clientcmd.NewDefaultClientConfig(*settings, &clientcmd.ConfigOverrides{}).ClientConfig()
 | 
						|
	if err != nil {
 | 
						|
		klog.Fatalf("Failed to construct config: %v", err)
 | 
						|
	}
 | 
						|
 | 
						|
	client, err := clientset.NewForConfig(config)
 | 
						|
	if err != nil {
 | 
						|
		klog.Fatalf("Failed to make client: %v", err)
 | 
						|
	}
 | 
						|
 | 
						|
	var nodes *v1.NodeList
 | 
						|
	for start := time.Now(); time.Since(start) < nodeListTimeout; time.Sleep(2 * time.Second) {
 | 
						|
		nodes, err = client.CoreV1().Nodes().List(context.TODO(), metav1.ListOptions{})
 | 
						|
		if err == nil {
 | 
						|
			break
 | 
						|
		}
 | 
						|
		klog.Warningf("Failed to list nodes: %v", err)
 | 
						|
	}
 | 
						|
	if err != nil {
 | 
						|
		klog.Fatalf("Giving up trying to list nodes: %v", err)
 | 
						|
	}
 | 
						|
 | 
						|
	if len(nodes.Items) == 0 {
 | 
						|
		klog.Fatalf("Failed to find any nodes.")
 | 
						|
	}
 | 
						|
 | 
						|
	klog.Infof("Found %d nodes on this cluster:", len(nodes.Items))
 | 
						|
	for i, node := range nodes.Items {
 | 
						|
		klog.Infof("%d: %s", i, node.Name)
 | 
						|
	}
 | 
						|
 | 
						|
	queries := *queriesAverage * len(nodes.Items) * *podsPerNode
 | 
						|
 | 
						|
	// Create the namespace
 | 
						|
	got, err := client.CoreV1().Namespaces().Create(context.TODO(), &v1.Namespace{ObjectMeta: metav1.ObjectMeta{GenerateName: "serve-hostnames-"}}, metav1.CreateOptions{})
 | 
						|
	if err != nil {
 | 
						|
		klog.Fatalf("Failed to create namespace: %v", err)
 | 
						|
	}
 | 
						|
	ns := got.Name
 | 
						|
	defer func(ns string) {
 | 
						|
		if err := client.CoreV1().Namespaces().Delete(context.TODO(), ns, nil); err != nil {
 | 
						|
			klog.Warningf("Failed to delete namespace %s: %v", ns, err)
 | 
						|
		} else {
 | 
						|
			// wait until the namespace disappears
 | 
						|
			for i := 0; i < int(namespaceDeleteTimeout/time.Second); i++ {
 | 
						|
				if _, err := client.CoreV1().Namespaces().Get(context.TODO(), ns, metav1.GetOptions{}); err != nil {
 | 
						|
					if apierrors.IsNotFound(err) {
 | 
						|
						return
 | 
						|
					}
 | 
						|
				}
 | 
						|
				time.Sleep(time.Second)
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}(ns)
 | 
						|
	klog.Infof("Created namespace %s", ns)
 | 
						|
 | 
						|
	// Create a service for these pods.
 | 
						|
	klog.Infof("Creating service %s/serve-hostnames", ns)
 | 
						|
	// Make several attempts to create a service.
 | 
						|
	var svc *v1.Service
 | 
						|
	for start := time.Now(); time.Since(start) < serviceCreateTimeout; time.Sleep(2 * time.Second) {
 | 
						|
		t := time.Now()
 | 
						|
		svc, err = client.CoreV1().Services(ns).Create(context.TODO(), &v1.Service{
 | 
						|
			ObjectMeta: metav1.ObjectMeta{
 | 
						|
				Name: "serve-hostnames",
 | 
						|
				Labels: map[string]string{
 | 
						|
					"name": "serve-hostname",
 | 
						|
				},
 | 
						|
			},
 | 
						|
			Spec: v1.ServiceSpec{
 | 
						|
				Ports: []v1.ServicePort{{
 | 
						|
					Protocol:   "TCP",
 | 
						|
					Port:       9376,
 | 
						|
					TargetPort: intstr.FromInt(9376),
 | 
						|
				}},
 | 
						|
				Selector: map[string]string{
 | 
						|
					"name": "serve-hostname",
 | 
						|
				},
 | 
						|
			},
 | 
						|
		}, metav1.CreateOptions{})
 | 
						|
		klog.V(4).Infof("Service create %s/server-hostnames took %v", ns, time.Since(t))
 | 
						|
		if err == nil {
 | 
						|
			break
 | 
						|
		}
 | 
						|
		klog.Warningf("After %v failed to create service %s/serve-hostnames: %v", time.Since(start), ns, err)
 | 
						|
	}
 | 
						|
	if err != nil {
 | 
						|
		klog.Warningf("Unable to create service %s/%s: %v", ns, svc.Name, err)
 | 
						|
		return
 | 
						|
	}
 | 
						|
	// Clean up service
 | 
						|
	defer func() {
 | 
						|
		klog.Infof("Cleaning up service %s/serve-hostnames", ns)
 | 
						|
		// Make several attempts to delete the service.
 | 
						|
		for start := time.Now(); time.Since(start) < deleteTimeout; time.Sleep(1 * time.Second) {
 | 
						|
			if err := client.CoreV1().Services(ns).Delete(context.TODO(), svc.Name, nil); err == nil {
 | 
						|
				return
 | 
						|
			}
 | 
						|
			klog.Warningf("After %v unable to delete service %s/%s: %v", time.Since(start), ns, svc.Name, err)
 | 
						|
		}
 | 
						|
	}()
 | 
						|
 | 
						|
	// Put serve-hostname pods on each node.
 | 
						|
	podNames := []string{}
 | 
						|
	for i, node := range nodes.Items {
 | 
						|
		for j := 0; j < *podsPerNode; j++ {
 | 
						|
			podName := fmt.Sprintf("serve-hostname-%d-%d", i, j)
 | 
						|
			podNames = append(podNames, podName)
 | 
						|
			// Make several attempts
 | 
						|
			for start := time.Now(); time.Since(start) < podCreateTimeout; time.Sleep(2 * time.Second) {
 | 
						|
				klog.Infof("Creating pod %s/%s on node %s", ns, podName, node.Name)
 | 
						|
				t := time.Now()
 | 
						|
				_, err = client.CoreV1().Pods(ns).Create(context.TODO(), &v1.Pod{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{
 | 
						|
						Name: podName,
 | 
						|
						Labels: map[string]string{
 | 
						|
							"name": "serve-hostname",
 | 
						|
						},
 | 
						|
					},
 | 
						|
					Spec: v1.PodSpec{
 | 
						|
						Containers: []v1.Container{
 | 
						|
							{
 | 
						|
								Name:  "serve-hostname",
 | 
						|
								Image: e2e.ServeHostnameImage,
 | 
						|
								Ports: []v1.ContainerPort{{ContainerPort: 9376}},
 | 
						|
							},
 | 
						|
						},
 | 
						|
						NodeName: node.Name,
 | 
						|
					},
 | 
						|
				}, metav1.CreateOptions{})
 | 
						|
				klog.V(4).Infof("Pod create %s/%s request took %v", ns, podName, time.Since(t))
 | 
						|
				if err == nil {
 | 
						|
					break
 | 
						|
				}
 | 
						|
				klog.Warningf("After %s failed to create pod %s/%s: %v", time.Since(start), ns, podName, err)
 | 
						|
			}
 | 
						|
			if err != nil {
 | 
						|
				klog.Warningf("Failed to create pod %s/%s: %v", ns, podName, err)
 | 
						|
				return
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	// Clean up the pods
 | 
						|
	defer func() {
 | 
						|
		klog.Info("Cleaning up pods")
 | 
						|
		// Make several attempts to delete the pods.
 | 
						|
		for _, podName := range podNames {
 | 
						|
			for start := time.Now(); time.Since(start) < deleteTimeout; time.Sleep(1 * time.Second) {
 | 
						|
				if err = client.CoreV1().Pods(ns).Delete(context.TODO(), podName, nil); err == nil {
 | 
						|
					break
 | 
						|
				}
 | 
						|
				klog.Warningf("After %v failed to delete pod %s/%s: %v", time.Since(start), ns, podName, err)
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}()
 | 
						|
 | 
						|
	klog.Info("Waiting for the serve-hostname pods to be ready")
 | 
						|
	for _, podName := range podNames {
 | 
						|
		var pod *v1.Pod
 | 
						|
		for start := time.Now(); time.Since(start) < podStartTimeout; time.Sleep(5 * time.Second) {
 | 
						|
			pod, err = client.CoreV1().Pods(ns).Get(context.TODO(), podName, metav1.GetOptions{})
 | 
						|
			if err != nil {
 | 
						|
				klog.Warningf("Get pod %s/%s failed, ignoring for %v: %v", ns, podName, err, podStartTimeout)
 | 
						|
				continue
 | 
						|
			}
 | 
						|
			if pod.Status.Phase == v1.PodRunning {
 | 
						|
				break
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if pod.Status.Phase != v1.PodRunning {
 | 
						|
			klog.Warningf("Gave up waiting on pod %s/%s to be running (saw %v)", ns, podName, pod.Status.Phase)
 | 
						|
		} else {
 | 
						|
			klog.Infof("%s/%s is running", ns, podName)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	rclient, err := restclient.RESTClientFor(config)
 | 
						|
	if err != nil {
 | 
						|
		klog.Warningf("Failed to build restclient: %v", err)
 | 
						|
		return
 | 
						|
	}
 | 
						|
	proxyRequest, errProxy := service.GetServicesProxyRequest(client, rclient.Get())
 | 
						|
	if errProxy != nil {
 | 
						|
		klog.Warningf("Get services proxy request failed: %v", errProxy)
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Wait for the endpoints to propagate.
 | 
						|
	for start := time.Now(); time.Since(start) < endpointTimeout; time.Sleep(10 * time.Second) {
 | 
						|
		hostname, err := proxyRequest.
 | 
						|
			Namespace(ns).
 | 
						|
			Name("serve-hostnames").
 | 
						|
			DoRaw(context.TODO())
 | 
						|
		if err != nil {
 | 
						|
			klog.Infof("After %v while making a proxy call got error %v", time.Since(start), err)
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		var r metav1.Status
 | 
						|
		if err := runtime.DecodeInto(legacyscheme.Codecs.UniversalDecoder(), hostname, &r); err != nil {
 | 
						|
			break
 | 
						|
		}
 | 
						|
		if r.Status == metav1.StatusFailure {
 | 
						|
			klog.Infof("After %v got status %v", time.Since(start), string(hostname))
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		break
 | 
						|
	}
 | 
						|
 | 
						|
	// Repeatedly make requests.
 | 
						|
	for iteration := 0; iteration != *upTo; iteration++ {
 | 
						|
		responseChan := make(chan string, queries)
 | 
						|
		// Use a channel of size *maxPar to throttle the number
 | 
						|
		// of in-flight requests to avoid overloading the service.
 | 
						|
		inFlight := make(chan struct{}, *maxPar)
 | 
						|
		start := time.Now()
 | 
						|
		for q := 0; q < queries; q++ {
 | 
						|
			go func(i int, query int) {
 | 
						|
				inFlight <- struct{}{}
 | 
						|
				t := time.Now()
 | 
						|
				hostname, err := proxyRequest.
 | 
						|
					Namespace(ns).
 | 
						|
					Name("serve-hostnames").
 | 
						|
					DoRaw(context.TODO())
 | 
						|
				klog.V(4).Infof("Proxy call in namespace %s took %v", ns, time.Since(t))
 | 
						|
				if err != nil {
 | 
						|
					klog.Warningf("Call failed during iteration %d query %d : %v", i, query, err)
 | 
						|
					// If the query failed return a string which starts with a character
 | 
						|
					// that can't be part of a hostname.
 | 
						|
					responseChan <- fmt.Sprintf("!failed in iteration %d to issue query %d: %v", i, query, err)
 | 
						|
				} else {
 | 
						|
					responseChan <- string(hostname)
 | 
						|
				}
 | 
						|
				<-inFlight
 | 
						|
			}(iteration, q)
 | 
						|
		}
 | 
						|
		responses := make(map[string]int, *podsPerNode*len(nodes.Items))
 | 
						|
		missing := 0
 | 
						|
		for q := 0; q < queries; q++ {
 | 
						|
			r := <-responseChan
 | 
						|
			klog.V(4).Infof("Got response from %s", r)
 | 
						|
			responses[r]++
 | 
						|
			// If the returned hostname starts with '!' then it indicates
 | 
						|
			// an error response.
 | 
						|
			if len(r) > 0 && r[0] == '!' {
 | 
						|
				klog.V(3).Infof("Got response %s", r)
 | 
						|
				missing++
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if missing > 0 {
 | 
						|
			klog.Warningf("Missing %d responses out of %d", missing, queries)
 | 
						|
		}
 | 
						|
		// Report any nodes that did not respond.
 | 
						|
		for n, node := range nodes.Items {
 | 
						|
			for i := 0; i < *podsPerNode; i++ {
 | 
						|
				name := fmt.Sprintf("serve-hostname-%d-%d", n, i)
 | 
						|
				if _, ok := responses[name]; !ok {
 | 
						|
					klog.Warningf("No response from pod %s on node %s at iteration %d", name, node.Name, iteration)
 | 
						|
				}
 | 
						|
			}
 | 
						|
		}
 | 
						|
		klog.Infof("Iteration %d took %v for %d queries (%.2f QPS) with %d missing",
 | 
						|
			iteration, time.Since(start), queries-missing, float64(queries-missing)/time.Since(start).Seconds(), missing)
 | 
						|
	}
 | 
						|
}
 |