mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-10-31 22:01:06 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			309 lines
		
	
	
		
			9.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			309 lines
		
	
	
		
			9.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // +build integration,!no-etcd
 | |
| 
 | |
| /*
 | |
| Copyright 2015 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package scheduler
 | |
| 
 | |
| // This file tests scheduler extender.
 | |
| 
 | |
| import (
 | |
| 	"encoding/json"
 | |
| 	"fmt"
 | |
| 	"net/http"
 | |
| 	"net/http/httptest"
 | |
| 	"strings"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"k8s.io/apimachinery/pkg/api/resource"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 	"k8s.io/apimachinery/pkg/util/wait"
 | |
| 	v1core "k8s.io/client-go/kubernetes/typed/core/v1"
 | |
| 	clientv1 "k8s.io/client-go/pkg/api/v1"
 | |
| 	restclient "k8s.io/client-go/rest"
 | |
| 	"k8s.io/client-go/tools/record"
 | |
| 	"k8s.io/kubernetes/pkg/api"
 | |
| 	"k8s.io/kubernetes/pkg/api/v1"
 | |
| 	"k8s.io/kubernetes/pkg/client/clientset_generated/clientset"
 | |
| 	"k8s.io/kubernetes/plugin/pkg/scheduler"
 | |
| 	_ "k8s.io/kubernetes/plugin/pkg/scheduler/algorithmprovider"
 | |
| 	schedulerapi "k8s.io/kubernetes/plugin/pkg/scheduler/api"
 | |
| 	"k8s.io/kubernetes/plugin/pkg/scheduler/factory"
 | |
| 	e2e "k8s.io/kubernetes/test/e2e/framework"
 | |
| 	"k8s.io/kubernetes/test/integration/framework"
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	filter     = "filter"
 | |
| 	prioritize = "prioritize"
 | |
| )
 | |
| 
 | |
| type fitPredicate func(pod *v1.Pod, node *v1.Node) (bool, error)
 | |
| type priorityFunc func(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error)
 | |
| 
 | |
| type priorityConfig struct {
 | |
| 	function priorityFunc
 | |
| 	weight   int
 | |
| }
 | |
| 
 | |
| type Extender struct {
 | |
| 	name         string
 | |
| 	predicates   []fitPredicate
 | |
| 	prioritizers []priorityConfig
 | |
| }
 | |
| 
 | |
| func (e *Extender) serveHTTP(t *testing.T, w http.ResponseWriter, req *http.Request) {
 | |
| 	var args schedulerapi.ExtenderArgs
 | |
| 
 | |
| 	decoder := json.NewDecoder(req.Body)
 | |
| 	defer req.Body.Close()
 | |
| 
 | |
| 	if err := decoder.Decode(&args); err != nil {
 | |
| 		http.Error(w, "Decode error", http.StatusBadRequest)
 | |
| 		return
 | |
| 	}
 | |
| 
 | |
| 	encoder := json.NewEncoder(w)
 | |
| 
 | |
| 	if strings.Contains(req.URL.Path, filter) {
 | |
| 		resp := &schedulerapi.ExtenderFilterResult{}
 | |
| 		nodes, failedNodes, err := e.Filter(&args.Pod, &args.Nodes)
 | |
| 		if err != nil {
 | |
| 			resp.Error = err.Error()
 | |
| 		} else {
 | |
| 			resp.Nodes = *nodes
 | |
| 			resp.FailedNodes = failedNodes
 | |
| 		}
 | |
| 
 | |
| 		if err := encoder.Encode(resp); err != nil {
 | |
| 			t.Fatalf("Failed to encode %+v", resp)
 | |
| 		}
 | |
| 	} else if strings.Contains(req.URL.Path, prioritize) {
 | |
| 		// Prioritize errors are ignored. Default k8s priorities or another extender's
 | |
| 		// priorities may be applied.
 | |
| 		priorities, _ := e.Prioritize(&args.Pod, &args.Nodes)
 | |
| 
 | |
| 		if err := encoder.Encode(priorities); err != nil {
 | |
| 			t.Fatalf("Failed to encode %+v", priorities)
 | |
| 		}
 | |
| 	} else {
 | |
| 		http.Error(w, "Unknown method", http.StatusNotFound)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (e *Extender) Filter(pod *v1.Pod, nodes *v1.NodeList) (*v1.NodeList, schedulerapi.FailedNodesMap, error) {
 | |
| 	filtered := []v1.Node{}
 | |
| 	failedNodesMap := schedulerapi.FailedNodesMap{}
 | |
| 	for _, node := range nodes.Items {
 | |
| 		fits := true
 | |
| 		for _, predicate := range e.predicates {
 | |
| 			fit, err := predicate(pod, &node)
 | |
| 			if err != nil {
 | |
| 				return &v1.NodeList{}, schedulerapi.FailedNodesMap{}, err
 | |
| 			}
 | |
| 			if !fit {
 | |
| 				fits = false
 | |
| 				break
 | |
| 			}
 | |
| 		}
 | |
| 		if fits {
 | |
| 			filtered = append(filtered, node)
 | |
| 		} else {
 | |
| 			failedNodesMap[node.Name] = fmt.Sprintf("extender failed: %s", e.name)
 | |
| 		}
 | |
| 	}
 | |
| 	return &v1.NodeList{Items: filtered}, failedNodesMap, nil
 | |
| }
 | |
| 
 | |
| func (e *Extender) Prioritize(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | |
| 	result := schedulerapi.HostPriorityList{}
 | |
| 	combinedScores := map[string]int{}
 | |
| 	for _, prioritizer := range e.prioritizers {
 | |
| 		weight := prioritizer.weight
 | |
| 		if weight == 0 {
 | |
| 			continue
 | |
| 		}
 | |
| 		priorityFunc := prioritizer.function
 | |
| 		prioritizedList, err := priorityFunc(pod, nodes)
 | |
| 		if err != nil {
 | |
| 			return &schedulerapi.HostPriorityList{}, err
 | |
| 		}
 | |
| 		for _, hostEntry := range *prioritizedList {
 | |
| 			combinedScores[hostEntry.Host] += hostEntry.Score * weight
 | |
| 		}
 | |
| 	}
 | |
| 	for host, score := range combinedScores {
 | |
| 		result = append(result, schedulerapi.HostPriority{Host: host, Score: score})
 | |
| 	}
 | |
| 	return &result, nil
 | |
| }
 | |
| 
 | |
| func machine_1_2_3_Predicate(pod *v1.Pod, node *v1.Node) (bool, error) {
 | |
| 	if node.Name == "machine1" || node.Name == "machine2" || node.Name == "machine3" {
 | |
| 		return true, nil
 | |
| 	}
 | |
| 	return false, nil
 | |
| }
 | |
| 
 | |
| func machine_2_3_5_Predicate(pod *v1.Pod, node *v1.Node) (bool, error) {
 | |
| 	if node.Name == "machine2" || node.Name == "machine3" || node.Name == "machine5" {
 | |
| 		return true, nil
 | |
| 	}
 | |
| 	return false, nil
 | |
| }
 | |
| 
 | |
| func machine_2_Prioritizer(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | |
| 	result := schedulerapi.HostPriorityList{}
 | |
| 	for _, node := range nodes.Items {
 | |
| 		score := 1
 | |
| 		if node.Name == "machine2" {
 | |
| 			score = 10
 | |
| 		}
 | |
| 		result = append(result, schedulerapi.HostPriority{node.Name, score})
 | |
| 	}
 | |
| 	return &result, nil
 | |
| }
 | |
| 
 | |
| func machine_3_Prioritizer(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | |
| 	result := schedulerapi.HostPriorityList{}
 | |
| 	for _, node := range nodes.Items {
 | |
| 		score := 1
 | |
| 		if node.Name == "machine3" {
 | |
| 			score = 10
 | |
| 		}
 | |
| 		result = append(result, schedulerapi.HostPriority{node.Name, score})
 | |
| 	}
 | |
| 	return &result, nil
 | |
| }
 | |
| 
 | |
| func TestSchedulerExtender(t *testing.T) {
 | |
| 	_, s := framework.RunAMaster(nil)
 | |
| 	defer s.Close()
 | |
| 
 | |
| 	ns := framework.CreateTestingNamespace("scheduler-extender", s, t)
 | |
| 	defer framework.DeleteTestingNamespace(ns, s, t)
 | |
| 
 | |
| 	clientSet := clientset.NewForConfigOrDie(&restclient.Config{Host: s.URL, ContentConfig: restclient.ContentConfig{GroupVersion: &api.Registry.GroupOrDie(v1.GroupName).GroupVersion}})
 | |
| 
 | |
| 	extender1 := &Extender{
 | |
| 		name:         "extender1",
 | |
| 		predicates:   []fitPredicate{machine_1_2_3_Predicate},
 | |
| 		prioritizers: []priorityConfig{{machine_2_Prioritizer, 1}},
 | |
| 	}
 | |
| 	es1 := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
 | |
| 		extender1.serveHTTP(t, w, req)
 | |
| 	}))
 | |
| 	defer es1.Close()
 | |
| 
 | |
| 	extender2 := &Extender{
 | |
| 		name:         "extender2",
 | |
| 		predicates:   []fitPredicate{machine_2_3_5_Predicate},
 | |
| 		prioritizers: []priorityConfig{{machine_3_Prioritizer, 1}},
 | |
| 	}
 | |
| 	es2 := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
 | |
| 		extender2.serveHTTP(t, w, req)
 | |
| 	}))
 | |
| 	defer es2.Close()
 | |
| 
 | |
| 	policy := schedulerapi.Policy{
 | |
| 		ExtenderConfigs: []schedulerapi.ExtenderConfig{
 | |
| 			{
 | |
| 				URLPrefix:      es1.URL,
 | |
| 				FilterVerb:     filter,
 | |
| 				PrioritizeVerb: prioritize,
 | |
| 				Weight:         3,
 | |
| 				EnableHttps:    false,
 | |
| 			},
 | |
| 			{
 | |
| 				URLPrefix:      es2.URL,
 | |
| 				FilterVerb:     filter,
 | |
| 				PrioritizeVerb: prioritize,
 | |
| 				Weight:         4,
 | |
| 				EnableHttps:    false,
 | |
| 			},
 | |
| 		},
 | |
| 	}
 | |
| 	policy.APIVersion = api.Registry.GroupOrDie(v1.GroupName).GroupVersion.String()
 | |
| 
 | |
| 	schedulerConfigFactory := factory.NewConfigFactory(clientSet, v1.DefaultSchedulerName, v1.DefaultHardPodAffinitySymmetricWeight)
 | |
| 	schedulerConfig, err := schedulerConfigFactory.CreateFromConfig(policy)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Couldn't create scheduler config: %v", err)
 | |
| 	}
 | |
| 	eventBroadcaster := record.NewBroadcaster()
 | |
| 	schedulerConfig.Recorder = eventBroadcaster.NewRecorder(api.Scheme, clientv1.EventSource{Component: v1.DefaultSchedulerName})
 | |
| 	eventBroadcaster.StartRecordingToSink(&v1core.EventSinkImpl{Interface: v1core.New(clientSet.Core().RESTClient()).Events("")})
 | |
| 	scheduler.New(schedulerConfig).Run()
 | |
| 
 | |
| 	defer close(schedulerConfig.StopEverything)
 | |
| 
 | |
| 	DoTestPodScheduling(ns, t, clientSet)
 | |
| }
 | |
| 
 | |
| func DoTestPodScheduling(ns *v1.Namespace, t *testing.T, cs clientset.Interface) {
 | |
| 	// NOTE: This test cannot run in parallel, because it is creating and deleting
 | |
| 	// non-namespaced objects (Nodes).
 | |
| 	defer cs.Core().Nodes().DeleteCollection(nil, metav1.ListOptions{})
 | |
| 
 | |
| 	goodCondition := v1.NodeCondition{
 | |
| 		Type:              v1.NodeReady,
 | |
| 		Status:            v1.ConditionTrue,
 | |
| 		Reason:            fmt.Sprintf("schedulable condition"),
 | |
| 		LastHeartbeatTime: metav1.Time{time.Now()},
 | |
| 	}
 | |
| 	node := &v1.Node{
 | |
| 		Spec: v1.NodeSpec{Unschedulable: false},
 | |
| 		Status: v1.NodeStatus{
 | |
| 			Capacity: v1.ResourceList{
 | |
| 				v1.ResourcePods: *resource.NewQuantity(32, resource.DecimalSI),
 | |
| 			},
 | |
| 			Conditions: []v1.NodeCondition{goodCondition},
 | |
| 		},
 | |
| 	}
 | |
| 
 | |
| 	for ii := 0; ii < 5; ii++ {
 | |
| 		node.Name = fmt.Sprintf("machine%d", ii+1)
 | |
| 		if _, err := cs.Core().Nodes().Create(node); err != nil {
 | |
| 			t.Fatalf("Failed to create nodes: %v", err)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	pod := &v1.Pod{
 | |
| 		ObjectMeta: metav1.ObjectMeta{Name: "extender-test-pod"},
 | |
| 		Spec: v1.PodSpec{
 | |
| 			Containers: []v1.Container{{Name: "container", Image: e2e.GetPauseImageName(cs)}},
 | |
| 		},
 | |
| 	}
 | |
| 
 | |
| 	myPod, err := cs.Core().Pods(ns.Name).Create(pod)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Failed to create pod: %v", err)
 | |
| 	}
 | |
| 
 | |
| 	err = wait.Poll(time.Second, wait.ForeverTestTimeout, podScheduled(cs, myPod.Namespace, myPod.Name))
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Failed to schedule pod: %v", err)
 | |
| 	}
 | |
| 
 | |
| 	if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name, metav1.GetOptions{}); err != nil {
 | |
| 		t.Fatalf("Failed to get pod: %v", err)
 | |
| 	} else if myPod.Spec.NodeName != "machine3" {
 | |
| 		t.Fatalf("Failed to schedule using extender, expected machine3, got %v", myPod.Spec.NodeName)
 | |
| 	}
 | |
| 	t.Logf("Scheduled pod using extenders")
 | |
| }
 |