mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			287 lines
		
	
	
		
			9.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			287 lines
		
	
	
		
			9.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2015 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package metrics
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"fmt"
 | 
						|
	"sync"
 | 
						|
	"time"
 | 
						|
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/fields"
 | 
						|
	"k8s.io/apimachinery/pkg/util/wait"
 | 
						|
	clientset "k8s.io/client-go/kubernetes"
 | 
						|
	e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
 | 
						|
	"k8s.io/kubernetes/test/e2e/system"
 | 
						|
 | 
						|
	"k8s.io/klog"
 | 
						|
)
 | 
						|
 | 
						|
const (
 | 
						|
	// insecureSchedulerPort is the default port for the scheduler status server.
 | 
						|
	// May be overridden by a flag at startup.
 | 
						|
	// Deprecated: use the secure KubeSchedulerPort instead.
 | 
						|
	insecureSchedulerPort = 10251
 | 
						|
	// insecureKubeControllerManagerPort is the default port for the controller manager status server.
 | 
						|
	// May be overridden by a flag at startup.
 | 
						|
	// Deprecated: use the secure KubeControllerManagerPort instead.
 | 
						|
	insecureKubeControllerManagerPort = 10252
 | 
						|
)
 | 
						|
 | 
						|
// Collection is metrics collection of components
 | 
						|
type Collection struct {
 | 
						|
	APIServerMetrics         APIServerMetrics
 | 
						|
	ControllerManagerMetrics ControllerManagerMetrics
 | 
						|
	KubeletMetrics           map[string]KubeletMetrics
 | 
						|
	SchedulerMetrics         SchedulerMetrics
 | 
						|
	ClusterAutoscalerMetrics ClusterAutoscalerMetrics
 | 
						|
}
 | 
						|
 | 
						|
// Grabber provides functions which grab metrics from components
 | 
						|
type Grabber struct {
 | 
						|
	client                            clientset.Interface
 | 
						|
	externalClient                    clientset.Interface
 | 
						|
	grabFromAPIServer                 bool
 | 
						|
	grabFromControllerManager         bool
 | 
						|
	grabFromKubelets                  bool
 | 
						|
	grabFromScheduler                 bool
 | 
						|
	grabFromClusterAutoscaler         bool
 | 
						|
	masterName                        string
 | 
						|
	registeredMaster                  bool
 | 
						|
	waitForControllerManagerReadyOnce sync.Once
 | 
						|
}
 | 
						|
 | 
						|
// NewMetricsGrabber returns new metrics which are initialized.
 | 
						|
func NewMetricsGrabber(c clientset.Interface, ec clientset.Interface, kubelets bool, scheduler bool, controllers bool, apiServer bool, clusterAutoscaler bool) (*Grabber, error) {
 | 
						|
	registeredMaster := false
 | 
						|
	masterName := ""
 | 
						|
	nodeList, err := c.CoreV1().Nodes().List(context.TODO(), metav1.ListOptions{})
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
	if len(nodeList.Items) < 1 {
 | 
						|
		klog.Warning("Can't find any Nodes in the API server to grab metrics from")
 | 
						|
	}
 | 
						|
	for _, node := range nodeList.Items {
 | 
						|
		if system.DeprecatedMightBeMasterNode(node.Name) {
 | 
						|
			registeredMaster = true
 | 
						|
			masterName = node.Name
 | 
						|
			break
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if !registeredMaster {
 | 
						|
		scheduler = false
 | 
						|
		controllers = false
 | 
						|
		clusterAutoscaler = ec != nil
 | 
						|
		if clusterAutoscaler {
 | 
						|
			klog.Warningf("Master node is not registered. Grabbing metrics from Scheduler, ControllerManager is disabled.")
 | 
						|
		} else {
 | 
						|
			klog.Warningf("Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.")
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return &Grabber{
 | 
						|
		client:                    c,
 | 
						|
		externalClient:            ec,
 | 
						|
		grabFromAPIServer:         apiServer,
 | 
						|
		grabFromControllerManager: controllers,
 | 
						|
		grabFromKubelets:          kubelets,
 | 
						|
		grabFromScheduler:         scheduler,
 | 
						|
		grabFromClusterAutoscaler: clusterAutoscaler,
 | 
						|
		masterName:                masterName,
 | 
						|
		registeredMaster:          registeredMaster,
 | 
						|
	}, nil
 | 
						|
}
 | 
						|
 | 
						|
// HasRegisteredMaster returns if metrics grabber was able to find a master node
 | 
						|
func (g *Grabber) HasRegisteredMaster() bool {
 | 
						|
	return g.registeredMaster
 | 
						|
}
 | 
						|
 | 
						|
// GrabFromKubelet returns metrics from kubelet
 | 
						|
func (g *Grabber) GrabFromKubelet(nodeName string) (KubeletMetrics, error) {
 | 
						|
	nodes, err := g.client.CoreV1().Nodes().List(context.TODO(), metav1.ListOptions{FieldSelector: fields.Set{"metadata.name": nodeName}.AsSelector().String()})
 | 
						|
	if err != nil {
 | 
						|
		return KubeletMetrics{}, err
 | 
						|
	}
 | 
						|
	if len(nodes.Items) != 1 {
 | 
						|
		return KubeletMetrics{}, fmt.Errorf("Error listing nodes with name %v, got %v", nodeName, nodes.Items)
 | 
						|
	}
 | 
						|
	kubeletPort := nodes.Items[0].Status.DaemonEndpoints.KubeletEndpoint.Port
 | 
						|
	return g.grabFromKubeletInternal(nodeName, int(kubeletPort))
 | 
						|
}
 | 
						|
 | 
						|
func (g *Grabber) grabFromKubeletInternal(nodeName string, kubeletPort int) (KubeletMetrics, error) {
 | 
						|
	if kubeletPort <= 0 || kubeletPort > 65535 {
 | 
						|
		return KubeletMetrics{}, fmt.Errorf("Invalid Kubelet port %v. Skipping Kubelet's metrics gathering", kubeletPort)
 | 
						|
	}
 | 
						|
	output, err := g.getMetricsFromNode(nodeName, int(kubeletPort))
 | 
						|
	if err != nil {
 | 
						|
		return KubeletMetrics{}, err
 | 
						|
	}
 | 
						|
	return parseKubeletMetrics(output)
 | 
						|
}
 | 
						|
 | 
						|
// GrabFromScheduler returns metrics from scheduler
 | 
						|
func (g *Grabber) GrabFromScheduler() (SchedulerMetrics, error) {
 | 
						|
	if !g.registeredMaster {
 | 
						|
		return SchedulerMetrics{}, fmt.Errorf("Master's Kubelet is not registered. Skipping Scheduler's metrics gathering")
 | 
						|
	}
 | 
						|
	output, err := g.getMetricsFromPod(g.client, fmt.Sprintf("%v-%v", "kube-scheduler", g.masterName), metav1.NamespaceSystem, insecureSchedulerPort)
 | 
						|
	if err != nil {
 | 
						|
		return SchedulerMetrics{}, err
 | 
						|
	}
 | 
						|
	return parseSchedulerMetrics(output)
 | 
						|
}
 | 
						|
 | 
						|
// GrabFromClusterAutoscaler returns metrics from cluster autoscaler
 | 
						|
func (g *Grabber) GrabFromClusterAutoscaler() (ClusterAutoscalerMetrics, error) {
 | 
						|
	if !g.registeredMaster && g.externalClient == nil {
 | 
						|
		return ClusterAutoscalerMetrics{}, fmt.Errorf("Master's Kubelet is not registered. Skipping ClusterAutoscaler's metrics gathering")
 | 
						|
	}
 | 
						|
	var client clientset.Interface
 | 
						|
	var namespace string
 | 
						|
	if g.externalClient != nil {
 | 
						|
		client = g.externalClient
 | 
						|
		namespace = "kubemark"
 | 
						|
	} else {
 | 
						|
		client = g.client
 | 
						|
		namespace = metav1.NamespaceSystem
 | 
						|
	}
 | 
						|
	output, err := g.getMetricsFromPod(client, "cluster-autoscaler", namespace, 8085)
 | 
						|
	if err != nil {
 | 
						|
		return ClusterAutoscalerMetrics{}, err
 | 
						|
	}
 | 
						|
	return parseClusterAutoscalerMetrics(output)
 | 
						|
}
 | 
						|
 | 
						|
// GrabFromControllerManager returns metrics from controller manager
 | 
						|
func (g *Grabber) GrabFromControllerManager() (ControllerManagerMetrics, error) {
 | 
						|
	if !g.registeredMaster {
 | 
						|
		return ControllerManagerMetrics{}, fmt.Errorf("Master's Kubelet is not registered. Skipping ControllerManager's metrics gathering")
 | 
						|
	}
 | 
						|
 | 
						|
	var err error
 | 
						|
	podName := fmt.Sprintf("%v-%v", "kube-controller-manager", g.masterName)
 | 
						|
	g.waitForControllerManagerReadyOnce.Do(func() {
 | 
						|
		if readyErr := e2epod.WaitForPodsReady(g.client, metav1.NamespaceSystem, podName, 0); readyErr != nil {
 | 
						|
			err = fmt.Errorf("error waiting for controller manager pod to be ready: %w", readyErr)
 | 
						|
			return
 | 
						|
		}
 | 
						|
 | 
						|
		var lastMetricsFetchErr error
 | 
						|
		if metricsWaitErr := wait.PollImmediate(time.Second, time.Minute, func() (bool, error) {
 | 
						|
			_, lastMetricsFetchErr = g.getMetricsFromPod(g.client, podName, metav1.NamespaceSystem, insecureKubeControllerManagerPort)
 | 
						|
			return lastMetricsFetchErr == nil, nil
 | 
						|
		}); metricsWaitErr != nil {
 | 
						|
			err = fmt.Errorf("error waiting for controller manager pod to expose metrics: %v; %v", metricsWaitErr, lastMetricsFetchErr)
 | 
						|
			return
 | 
						|
		}
 | 
						|
	})
 | 
						|
	if err != nil {
 | 
						|
		return ControllerManagerMetrics{}, err
 | 
						|
	}
 | 
						|
 | 
						|
	output, err := g.getMetricsFromPod(g.client, podName, metav1.NamespaceSystem, insecureKubeControllerManagerPort)
 | 
						|
	if err != nil {
 | 
						|
		return ControllerManagerMetrics{}, err
 | 
						|
	}
 | 
						|
	return parseControllerManagerMetrics(output)
 | 
						|
}
 | 
						|
 | 
						|
// GrabFromAPIServer returns metrics from API server
 | 
						|
func (g *Grabber) GrabFromAPIServer() (APIServerMetrics, error) {
 | 
						|
	output, err := g.getMetricsFromAPIServer()
 | 
						|
	if err != nil {
 | 
						|
		return APIServerMetrics{}, nil
 | 
						|
	}
 | 
						|
	return parseAPIServerMetrics(output)
 | 
						|
}
 | 
						|
 | 
						|
// Grab returns metrics from corresponding component
 | 
						|
func (g *Grabber) Grab() (Collection, error) {
 | 
						|
	result := Collection{}
 | 
						|
	var errs []error
 | 
						|
	if g.grabFromAPIServer {
 | 
						|
		metrics, err := g.GrabFromAPIServer()
 | 
						|
		if err != nil {
 | 
						|
			errs = append(errs, err)
 | 
						|
		} else {
 | 
						|
			result.APIServerMetrics = metrics
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if g.grabFromScheduler {
 | 
						|
		metrics, err := g.GrabFromScheduler()
 | 
						|
		if err != nil {
 | 
						|
			errs = append(errs, err)
 | 
						|
		} else {
 | 
						|
			result.SchedulerMetrics = metrics
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if g.grabFromControllerManager {
 | 
						|
		metrics, err := g.GrabFromControllerManager()
 | 
						|
		if err != nil {
 | 
						|
			errs = append(errs, err)
 | 
						|
		} else {
 | 
						|
			result.ControllerManagerMetrics = metrics
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if g.grabFromClusterAutoscaler {
 | 
						|
		metrics, err := g.GrabFromClusterAutoscaler()
 | 
						|
		if err != nil {
 | 
						|
			errs = append(errs, err)
 | 
						|
		} else {
 | 
						|
			result.ClusterAutoscalerMetrics = metrics
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if g.grabFromKubelets {
 | 
						|
		result.KubeletMetrics = make(map[string]KubeletMetrics)
 | 
						|
		nodes, err := g.client.CoreV1().Nodes().List(context.TODO(), metav1.ListOptions{})
 | 
						|
		if err != nil {
 | 
						|
			errs = append(errs, err)
 | 
						|
		} else {
 | 
						|
			for _, node := range nodes.Items {
 | 
						|
				kubeletPort := node.Status.DaemonEndpoints.KubeletEndpoint.Port
 | 
						|
				metrics, err := g.grabFromKubeletInternal(node.Name, int(kubeletPort))
 | 
						|
				if err != nil {
 | 
						|
					errs = append(errs, err)
 | 
						|
				}
 | 
						|
				result.KubeletMetrics[node.Name] = metrics
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if len(errs) > 0 {
 | 
						|
		return result, fmt.Errorf("Errors while grabbing metrics: %v", errs)
 | 
						|
	}
 | 
						|
	return result, nil
 | 
						|
}
 | 
						|
 | 
						|
func (g *Grabber) getMetricsFromPod(client clientset.Interface, podName string, namespace string, port int) (string, error) {
 | 
						|
	rawOutput, err := client.CoreV1().RESTClient().Get().
 | 
						|
		Namespace(namespace).
 | 
						|
		Resource("pods").
 | 
						|
		SubResource("proxy").
 | 
						|
		Name(fmt.Sprintf("%v:%v", podName, port)).
 | 
						|
		Suffix("metrics").
 | 
						|
		Do(context.TODO()).Raw()
 | 
						|
	if err != nil {
 | 
						|
		return "", err
 | 
						|
	}
 | 
						|
	return string(rawOutput), nil
 | 
						|
}
 |