mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-10-31 05:40:42 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			207 lines
		
	
	
		
			7.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			207 lines
		
	
	
		
			7.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2015 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package namespace
 | |
| 
 | |
| import (
 | |
| 	"time"
 | |
| 
 | |
| 	"k8s.io/kubernetes/pkg/api"
 | |
| 	"k8s.io/kubernetes/pkg/api/unversioned"
 | |
| 	"k8s.io/kubernetes/pkg/client/cache"
 | |
| 	clientset "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset"
 | |
| 	"k8s.io/kubernetes/pkg/client/typed/dynamic"
 | |
| 	"k8s.io/kubernetes/pkg/controller"
 | |
| 	"k8s.io/kubernetes/pkg/controller/framework"
 | |
| 	"k8s.io/kubernetes/pkg/runtime"
 | |
| 	"k8s.io/kubernetes/pkg/util/metrics"
 | |
| 	utilruntime "k8s.io/kubernetes/pkg/util/runtime"
 | |
| 	"k8s.io/kubernetes/pkg/util/wait"
 | |
| 	"k8s.io/kubernetes/pkg/util/workqueue"
 | |
| 	"k8s.io/kubernetes/pkg/watch"
 | |
| 
 | |
| 	"github.com/golang/glog"
 | |
| )
 | |
| 
 | |
| // NamespaceController is responsible for performing actions dependent upon a namespace phase
 | |
| type NamespaceController struct {
 | |
| 	// client that purges namespace content, must have list/delete privileges on all content
 | |
| 	kubeClient clientset.Interface
 | |
| 	// clientPool manages a pool of dynamic clients
 | |
| 	clientPool dynamic.ClientPool
 | |
| 	// store that holds the namespaces
 | |
| 	store cache.Store
 | |
| 	// controller that observes the namespaces
 | |
| 	controller *framework.Controller
 | |
| 	// namespaces that have been queued up for processing by workers
 | |
| 	queue workqueue.RateLimitingInterface
 | |
| 	// list of preferred group versions and their corresponding resource set for namespace deletion
 | |
| 	groupVersionResources []unversioned.GroupVersionResource
 | |
| 	// opCache is a cache to remember if a particular operation is not supported to aid dynamic client.
 | |
| 	opCache operationNotSupportedCache
 | |
| 	// finalizerToken is the finalizer token managed by this controller
 | |
| 	finalizerToken api.FinalizerName
 | |
| }
 | |
| 
 | |
| // NewNamespaceController creates a new NamespaceController
 | |
| func NewNamespaceController(
 | |
| 	kubeClient clientset.Interface,
 | |
| 	clientPool dynamic.ClientPool,
 | |
| 	groupVersionResources []unversioned.GroupVersionResource,
 | |
| 	resyncPeriod time.Duration,
 | |
| 	finalizerToken api.FinalizerName) *NamespaceController {
 | |
| 
 | |
| 	// the namespace deletion code looks at the discovery document to enumerate the set of resources on the server.
 | |
| 	// it then finds all namespaced resources, and in response to namespace deletion, will call delete on all of them.
 | |
| 	// unfortunately, the discovery information does not include the list of supported verbs/methods.  if the namespace
 | |
| 	// controller calls LIST/DELETECOLLECTION for a resource, it will get a 405 error from the server and cache that that was the case.
 | |
| 	// we found in practice though that some auth engines when encountering paths they don't know about may return a 50x.
 | |
| 	// until we have verbs, we pre-populate resources that do not support list or delete for well-known apis rather than
 | |
| 	// probing the server once in order to be told no.
 | |
| 	opCache := operationNotSupportedCache{}
 | |
| 	ignoredGroupVersionResources := []unversioned.GroupVersionResource{
 | |
| 		{Group: "", Version: "v1", Resource: "bindings"},
 | |
| 	}
 | |
| 	for _, ignoredGroupVersionResource := range ignoredGroupVersionResources {
 | |
| 		opCache[operationKey{op: operationDeleteCollection, gvr: ignoredGroupVersionResource}] = true
 | |
| 		opCache[operationKey{op: operationList, gvr: ignoredGroupVersionResource}] = true
 | |
| 	}
 | |
| 
 | |
| 	// create the controller so we can inject the enqueue function
 | |
| 	namespaceController := &NamespaceController{
 | |
| 		kubeClient: kubeClient,
 | |
| 		clientPool: clientPool,
 | |
| 		queue:      workqueue.NewNamedRateLimitingQueue(workqueue.DefaultControllerRateLimiter(), "namespace"),
 | |
| 		groupVersionResources: groupVersionResources,
 | |
| 		opCache:               opCache,
 | |
| 		finalizerToken:        finalizerToken,
 | |
| 	}
 | |
| 
 | |
| 	if kubeClient != nil && kubeClient.Core().GetRESTClient().GetRateLimiter() != nil {
 | |
| 		metrics.RegisterMetricAndTrackRateLimiterUsage("namespace_controller", kubeClient.Core().GetRESTClient().GetRateLimiter())
 | |
| 	}
 | |
| 
 | |
| 	// configure the backing store/controller
 | |
| 	store, controller := framework.NewInformer(
 | |
| 		&cache.ListWatch{
 | |
| 			ListFunc: func(options api.ListOptions) (runtime.Object, error) {
 | |
| 				return kubeClient.Core().Namespaces().List(options)
 | |
| 			},
 | |
| 			WatchFunc: func(options api.ListOptions) (watch.Interface, error) {
 | |
| 				return kubeClient.Core().Namespaces().Watch(options)
 | |
| 			},
 | |
| 		},
 | |
| 		&api.Namespace{},
 | |
| 		resyncPeriod,
 | |
| 		framework.ResourceEventHandlerFuncs{
 | |
| 			AddFunc: func(obj interface{}) {
 | |
| 				namespace := obj.(*api.Namespace)
 | |
| 				namespaceController.enqueueNamespace(namespace)
 | |
| 			},
 | |
| 			UpdateFunc: func(oldObj, newObj interface{}) {
 | |
| 				namespace := newObj.(*api.Namespace)
 | |
| 				namespaceController.enqueueNamespace(namespace)
 | |
| 			},
 | |
| 		},
 | |
| 	)
 | |
| 
 | |
| 	namespaceController.store = store
 | |
| 	namespaceController.controller = controller
 | |
| 	return namespaceController
 | |
| }
 | |
| 
 | |
| // enqueueNamespace adds an object to the controller work queue
 | |
| // obj could be an *api.Namespace, or a DeletionFinalStateUnknown item.
 | |
| func (nm *NamespaceController) enqueueNamespace(obj interface{}) {
 | |
| 	key, err := controller.KeyFunc(obj)
 | |
| 	if err != nil {
 | |
| 		glog.Errorf("Couldn't get key for object %+v: %v", obj, err)
 | |
| 		return
 | |
| 	}
 | |
| 	nm.queue.Add(key)
 | |
| }
 | |
| 
 | |
| // worker processes the queue of namespace objects.
 | |
| // Each namespace can be in the queue at most once.
 | |
| // The system ensures that no two workers can process
 | |
| // the same namespace at the same time.
 | |
| func (nm *NamespaceController) worker() {
 | |
| 	workFunc := func() bool {
 | |
| 		key, quit := nm.queue.Get()
 | |
| 		if quit {
 | |
| 			return true
 | |
| 		}
 | |
| 		defer nm.queue.Done(key)
 | |
| 
 | |
| 		err := nm.syncNamespaceFromKey(key.(string))
 | |
| 		if err == nil {
 | |
| 			// no error, forget this entry and return
 | |
| 			nm.queue.Forget(key)
 | |
| 			return false
 | |
| 		}
 | |
| 
 | |
| 		if estimate, ok := err.(*contentRemainingError); ok {
 | |
| 			t := estimate.Estimate/2 + 1
 | |
| 			glog.V(4).Infof("Content remaining in namespace %s, waiting %d seconds", key, t)
 | |
| 			nm.queue.AddAfter(key, time.Duration(t)*time.Second)
 | |
| 		} else {
 | |
| 			// rather than wait for a full resync, re-add the namespace to the queue to be processed
 | |
| 			nm.queue.AddRateLimited(key)
 | |
| 			utilruntime.HandleError(err)
 | |
| 		}
 | |
| 		return false
 | |
| 	}
 | |
| 
 | |
| 	for {
 | |
| 		quit := workFunc()
 | |
| 
 | |
| 		if quit {
 | |
| 			return
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // syncNamespaceFromKey looks for a namespace with the specified key in its store and synchronizes it
 | |
| func (nm *NamespaceController) syncNamespaceFromKey(key string) (err error) {
 | |
| 	startTime := time.Now()
 | |
| 	defer glog.V(4).Infof("Finished syncing namespace %q (%v)", key, time.Now().Sub(startTime))
 | |
| 
 | |
| 	obj, exists, err := nm.store.GetByKey(key)
 | |
| 	if !exists {
 | |
| 		glog.Infof("Namespace has been deleted %v", key)
 | |
| 		return nil
 | |
| 	}
 | |
| 	if err != nil {
 | |
| 		glog.Errorf("Unable to retrieve namespace %v from store: %v", key, err)
 | |
| 		nm.queue.Add(key)
 | |
| 		return err
 | |
| 	}
 | |
| 	namespace := obj.(*api.Namespace)
 | |
| 	return syncNamespace(nm.kubeClient, nm.clientPool, nm.opCache, nm.groupVersionResources, namespace, nm.finalizerToken)
 | |
| }
 | |
| 
 | |
| // Run starts observing the system with the specified number of workers.
 | |
| func (nm *NamespaceController) Run(workers int, stopCh <-chan struct{}) {
 | |
| 	defer utilruntime.HandleCrash()
 | |
| 	go nm.controller.Run(stopCh)
 | |
| 	for i := 0; i < workers; i++ {
 | |
| 		go wait.Until(nm.worker, time.Second, stopCh)
 | |
| 	}
 | |
| 	<-stopCh
 | |
| 	glog.Infof("Shutting down NamespaceController")
 | |
| 	nm.queue.ShutDown()
 | |
| }
 |