diff --git a/pkg/client/cache/listwatch.go b/pkg/client/cache/listwatch.go index 3bd3b1e3353..22d4307c926 100644 --- a/pkg/client/cache/listwatch.go +++ b/pkg/client/cache/listwatch.go @@ -20,6 +20,7 @@ import ( "time" "k8s.io/apimachinery/pkg/api/meta" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/watch" @@ -32,16 +33,16 @@ import ( type ListerWatcher interface { // List should return a list type object; the Items field will be extracted, and the // ResourceVersion field will be used to start the watch in the right place. - List(options v1.ListOptions) (runtime.Object, error) + List(options metav1.ListOptions) (runtime.Object, error) // Watch should begin a watch at the specified version. - Watch(options v1.ListOptions) (watch.Interface, error) + Watch(options metav1.ListOptions) (watch.Interface, error) } // ListFunc knows how to list resources -type ListFunc func(options v1.ListOptions) (runtime.Object, error) +type ListFunc func(options metav1.ListOptions) (runtime.Object, error) // WatchFunc knows how to watch resources -type WatchFunc func(options v1.ListOptions) (watch.Interface, error) +type WatchFunc func(options metav1.ListOptions) (watch.Interface, error) // ListWatch knows how to list and watch a set of apiserver resources. It satisfies the ListerWatcher interface. // It is a convenience function for users of NewReflector, etc. @@ -58,28 +59,28 @@ type Getter interface { // NewListWatchFromClient creates a new ListWatch from the specified client, resource, namespace and field selector. func NewListWatchFromClient(c Getter, resource string, namespace string, fieldSelector fields.Selector) *ListWatch { - listFunc := func(options v1.ListOptions) (runtime.Object, error) { + listFunc := func(options metav1.ListOptions) (runtime.Object, error) { return c.Get(). Namespace(namespace). Resource(resource). - VersionedParams(&options, api.ParameterCodec). + VersionedParams(&options, metav1.ParameterCodec). FieldsSelectorParam(fieldSelector). Do(). Get() } - watchFunc := func(options v1.ListOptions) (watch.Interface, error) { + watchFunc := func(options metav1.ListOptions) (watch.Interface, error) { return c.Get(). Prefix("watch"). Namespace(namespace). Resource(resource). - VersionedParams(&options, api.ParameterCodec). + VersionedParams(&options, metav1.ParameterCodec). FieldsSelectorParam(fieldSelector). Watch() } return &ListWatch{ListFunc: listFunc, WatchFunc: watchFunc} } -func timeoutFromListOptions(options v1.ListOptions) time.Duration { +func timeoutFromListOptions(options metav1.ListOptions) time.Duration { if options.TimeoutSeconds != nil { return time.Duration(*options.TimeoutSeconds) * time.Second } @@ -87,12 +88,12 @@ func timeoutFromListOptions(options v1.ListOptions) time.Duration { } // List a set of apiserver resources -func (lw *ListWatch) List(options v1.ListOptions) (runtime.Object, error) { +func (lw *ListWatch) List(options metav1.ListOptions) (runtime.Object, error) { return lw.ListFunc(options) } // Watch a set of apiserver resources -func (lw *ListWatch) Watch(options v1.ListOptions) (watch.Interface, error) { +func (lw *ListWatch) Watch(options metav1.ListOptions) (watch.Interface, error) { return lw.WatchFunc(options) } @@ -102,7 +103,7 @@ func ListWatchUntil(timeout time.Duration, lw ListerWatcher, conditions ...watch return nil, nil } - list, err := lw.List(v1.ListOptions{}) + list, err := lw.List(metav1.ListOptions{}) if err != nil { return nil, err } @@ -154,7 +155,7 @@ func ListWatchUntil(timeout time.Duration, lw ListerWatcher, conditions ...watch } currResourceVersion := metaObj.GetResourceVersion() - watchInterface, err := lw.Watch(v1.ListOptions{ResourceVersion: currResourceVersion}) + watchInterface, err := lw.Watch(metav1.ListOptions{ResourceVersion: currResourceVersion}) if err != nil { return nil, err } diff --git a/pkg/client/cache/reflector.go b/pkg/client/cache/reflector.go index c638e97e96f..d3144e1a9ce 100644 --- a/pkg/client/cache/reflector.go +++ b/pkg/client/cache/reflector.go @@ -36,11 +36,11 @@ import ( "github.com/golang/glog" apierrs "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/meta" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime" utilruntime "k8s.io/apimachinery/pkg/util/runtime" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/apimachinery/pkg/watch" - "k8s.io/kubernetes/pkg/api/v1" ) // Reflector watches a specified resource and causes all changes to be reflected in the given store. @@ -239,7 +239,7 @@ func (r *Reflector) ListAndWatch(stopCh <-chan struct{}) error { // Explicitly set "0" as resource version - it's fine for the List() // to be served from cache and potentially be delayed relative to // etcd contents. Reflector framework will catch up via Watch() eventually. - options := v1.ListOptions{ResourceVersion: "0"} + options := metav1.ListOptions{ResourceVersion: "0"} list, err := r.listerWatcher.List(options) if err != nil { return fmt.Errorf("%s: Failed to list %v: %v", r.name, r.expectedType, err) @@ -282,7 +282,7 @@ func (r *Reflector) ListAndWatch(stopCh <-chan struct{}) error { for { timemoutseconds := int64(minWatchTimeout.Seconds() * (rand.Float64() + 1.0)) - options = v1.ListOptions{ + options = metav1.ListOptions{ ResourceVersion: resourceVersion, // We want to avoid situations of hanging watchers. Stop any wachers that do not // receive any events within the timeout window. diff --git a/pkg/controller/client_builder.go b/pkg/controller/client_builder.go index 3841a953aad..8299db986a0 100644 --- a/pkg/controller/client_builder.go +++ b/pkg/controller/client_builder.go @@ -122,11 +122,11 @@ func (b SAControllerClientBuilder) Config(name string) (*restclient.Config, erro } lw := &cache.ListWatch{ - ListFunc: func(options v1.ListOptions) (runtime.Object, error) { + ListFunc: func(options metav1.ListOptions) (runtime.Object, error) { options.FieldSelector = fields.SelectorFromSet(map[string]string{api.SecretTypeField: string(v1.SecretTypeServiceAccountToken)}).String() return b.CoreClient.Secrets(b.Namespace).List(options) }, - WatchFunc: func(options v1.ListOptions) (watch.Interface, error) { + WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) { options.FieldSelector = fields.SelectorFromSet(map[string]string{api.SecretTypeField: string(v1.SecretTypeServiceAccountToken)}).String() return b.CoreClient.Secrets(b.Namespace).Watch(options) }, diff --git a/pkg/controller/cronjob/cronjob_controller.go b/pkg/controller/cronjob/cronjob_controller.go index 850908e78a9..01bc1c4dc4e 100644 --- a/pkg/controller/cronjob/cronjob_controller.go +++ b/pkg/controller/cronjob/cronjob_controller.go @@ -97,7 +97,7 @@ func (jm *CronJobController) Run(stopCh <-chan struct{}) { // SyncAll lists all the CronJobs and Jobs and reconciles them. func (jm *CronJobController) SyncAll() { - sjl, err := jm.kubeClient.BatchV2alpha1().CronJobs(v1.NamespaceAll).List(v1.ListOptions{}) + sjl, err := jm.kubeClient.BatchV2alpha1().CronJobs(metav1.NamespaceAll).List(metav1.ListOptions{}) if err != nil { glog.Errorf("Error listing cronjobs: %v", err) return @@ -105,7 +105,7 @@ func (jm *CronJobController) SyncAll() { sjs := sjl.Items glog.V(4).Infof("Found %d cronjobs", len(sjs)) - jl, err := jm.kubeClient.BatchV2alpha1().Jobs(v1.NamespaceAll).List(v1.ListOptions{}) + jl, err := jm.kubeClient.BatchV2alpha1().Jobs(metav1.NamespaceAll).List(metav1.ListOptions{}) if err != nil { glog.Errorf("Error listing jobs") return @@ -238,7 +238,7 @@ func SyncOne(sj batch.CronJob, js []batch.Job, now time.Time, jc jobControlInter } // remove all pods... selector, _ := metav1.LabelSelectorAsSelector(job.Spec.Selector) - options := v1.ListOptions{LabelSelector: selector.String()} + options := metav1.ListOptions{LabelSelector: selector.String()} podList, err := pc.ListPods(job.Namespace, options) if err != nil { recorder.Eventf(&sj, v1.EventTypeWarning, "FailedList", "List job-pods: %v", err) diff --git a/pkg/controller/cronjob/injection.go b/pkg/controller/cronjob/injection.go index 5863cb739c4..a717d45e918 100644 --- a/pkg/controller/cronjob/injection.go +++ b/pkg/controller/cronjob/injection.go @@ -177,7 +177,7 @@ func (f *fakeJobControl) Clear() { // created as an interface to allow testing. type podControlInterface interface { // ListPods list pods - ListPods(namespace string, opts v1.ListOptions) (*v1.PodList, error) + ListPods(namespace string, opts metav1.ListOptions) (*v1.PodList, error) // DeleteJob deletes the pod identified by name. // TODO: delete by UID? DeletePod(namespace string, name string) error @@ -191,7 +191,7 @@ type realPodControl struct { var _ podControlInterface = &realPodControl{} -func (r realPodControl) ListPods(namespace string, opts v1.ListOptions) (*v1.PodList, error) { +func (r realPodControl) ListPods(namespace string, opts metav1.ListOptions) (*v1.PodList, error) { return r.KubeClient.Core().Pods(namespace).List(opts) } @@ -208,7 +208,7 @@ type fakePodControl struct { var _ podControlInterface = &fakePodControl{} -func (f *fakePodControl) ListPods(namespace string, opts v1.ListOptions) (*v1.PodList, error) { +func (f *fakePodControl) ListPods(namespace string, opts metav1.ListOptions) (*v1.PodList, error) { f.Lock() defer f.Unlock() return &v1.PodList{Items: f.Pods}, nil diff --git a/pkg/controller/podautoscaler/metrics/metrics_client.go b/pkg/controller/podautoscaler/metrics/metrics_client.go index 7b55d1ca2c2..8ba73323542 100644 --- a/pkg/controller/podautoscaler/metrics/metrics_client.go +++ b/pkg/controller/podautoscaler/metrics/metrics_client.go @@ -23,13 +23,14 @@ import ( "time" "github.com/golang/glog" + heapster "k8s.io/heapster/metrics/api/v1/types" + metricsapi "k8s.io/heapster/metrics/apis/metrics/v1alpha1" + + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/clientset" v1core "k8s.io/kubernetes/pkg/client/clientset_generated/clientset/typed/core/v1" - - heapster "k8s.io/heapster/metrics/api/v1/types" - metricsapi "k8s.io/heapster/metrics/apis/metrics/v1alpha1" ) // PodResourceInfo contains pod resourcemetric values as a map from pod names to @@ -128,7 +129,7 @@ func (h *HeapsterMetricsClient) GetResourceMetric(resource v1.ResourceName, name } func (h *HeapsterMetricsClient) GetRawMetric(metricName string, namespace string, selector labels.Selector) (PodMetricsInfo, time.Time, error) { - podList, err := h.podsGetter.Pods(namespace).List(v1.ListOptions{LabelSelector: selector.String()}) + podList, err := h.podsGetter.Pods(namespace).List(metav1.ListOptions{LabelSelector: selector.String()}) if err != nil { return nil, time.Time{}, fmt.Errorf("failed to get pod list while fetching metrics: %v", err) } diff --git a/pkg/kubelet/util/csr/csr.go b/pkg/kubelet/util/csr/csr.go index 9c543f81000..8390e796ce5 100644 --- a/pkg/kubelet/util/csr/csr.go +++ b/pkg/kubelet/util/csr/csr.go @@ -70,7 +70,7 @@ func RequestNodeCertificate(client certificatesclient.CertificateSigningRequestI // Make a default timeout = 3600s. var defaultTimeoutSeconds int64 = 3600 - resultCh, err := client.Watch(v1.ListOptions{ + resultCh, err := client.Watch(metav1.ListOptions{ Watch: true, TimeoutSeconds: &defaultTimeoutSeconds, FieldSelector: fields.OneTermEqualSelector("metadata.name", req.Name).String(), diff --git a/pkg/registry/core/pod/storage/eviction.go b/pkg/registry/core/pod/storage/eviction.go index 5b4852c262c..6205f874ad2 100644 --- a/pkg/registry/core/pod/storage/eviction.go +++ b/pkg/registry/core/pod/storage/eviction.go @@ -181,7 +181,7 @@ func (r *EvictionREST) getPodDisruptionBudgets(ctx genericapirequest.Context, po return nil, nil } - pdbList, err := r.podDisruptionBudgetClient.PodDisruptionBudgets(pod.Namespace).List(api.ListOptions{}) + pdbList, err := r.podDisruptionBudgetClient.PodDisruptionBudgets(pod.Namespace).List(metav1.ListOptions{}) if err != nil { return nil, err } diff --git a/pkg/registry/core/service/ipallocator/controller/repair.go b/pkg/registry/core/service/ipallocator/controller/repair.go index efd748dfdc7..d5c9e8258f3 100644 --- a/pkg/registry/core/service/ipallocator/controller/repair.go +++ b/pkg/registry/core/service/ipallocator/controller/repair.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/apimachinery/pkg/api/errors" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/runtime" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/kubernetes/pkg/api" @@ -119,7 +120,7 @@ func (c *Repair) runOnce() error { // the service collection. The caching layer keeps per-collection RVs, // and this is proper, since in theory the collections could be hosted // in separate etcd (or even non-etcd) instances. - list, err := c.serviceClient.Services(api.NamespaceAll).List(api.ListOptions{}) + list, err := c.serviceClient.Services(metav1.NamespaceAll).List(metav1.ListOptions{}) if err != nil { return fmt.Errorf("unable to refresh the service IP block: %v", err) } diff --git a/pkg/registry/rbac/rest/storage_rbac.go b/pkg/registry/rbac/rest/storage_rbac.go index 7c1f2333cbe..d3df1571937 100644 --- a/pkg/registry/rbac/rest/storage_rbac.go +++ b/pkg/registry/rbac/rest/storage_rbac.go @@ -23,11 +23,11 @@ import ( "github.com/golang/glog" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" utilruntime "k8s.io/apimachinery/pkg/util/runtime" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/apiserver/pkg/authorization/authorizer" - "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/apis/rbac" rbacapiv1alpha1 "k8s.io/kubernetes/pkg/apis/rbac/v1alpha1" rbacapiv1beta1 "k8s.io/kubernetes/pkg/apis/rbac/v1beta1" @@ -132,7 +132,7 @@ func PostStartHook(hookContext genericapiserver.PostStartHookContext) error { return false, nil } - existingClusterRoles, err := clientset.ClusterRoles().List(api.ListOptions{}) + existingClusterRoles, err := clientset.ClusterRoles().List(metav1.ListOptions{}) if err != nil { utilruntime.HandleError(fmt.Errorf("unable to initialize clusterroles: %v", err)) return false, nil @@ -149,7 +149,7 @@ func PostStartHook(hookContext genericapiserver.PostStartHookContext) error { } } - existingClusterRoleBindings, err := clientset.ClusterRoleBindings().List(api.ListOptions{}) + existingClusterRoleBindings, err := clientset.ClusterRoleBindings().List(metav1.ListOptions{}) if err != nil { utilruntime.HandleError(fmt.Errorf("unable to initialize clusterrolebindings: %v", err)) return false, nil diff --git a/pkg/volume/glusterfs/glusterfs.go b/pkg/volume/glusterfs/glusterfs.go index 105684bf994..9b838eb7fa1 100644 --- a/pkg/volume/glusterfs/glusterfs.go +++ b/pkg/volume/glusterfs/glusterfs.go @@ -456,7 +456,7 @@ func (p *glusterfsPlugin) collectGids(className string, gidTable *MinMaxAllocato if kubeClient == nil { return fmt.Errorf("glusterfs: failed to get kube client when collecting gids") } - pvList, err := kubeClient.Core().PersistentVolumes().List(v1.ListOptions{LabelSelector: labels.Everything().String()}) + pvList, err := kubeClient.Core().PersistentVolumes().List(metav1.ListOptions{LabelSelector: labels.Everything().String()}) if err != nil { glog.Errorf("glusterfs: failed to get existing persistent volumes") return err diff --git a/pkg/volume/util.go b/pkg/volume/util.go index ad3d5ffee3d..a3e98abd67e 100644 --- a/pkg/volume/util.go +++ b/pkg/volume/util.go @@ -178,7 +178,7 @@ func (c *realRecyclerClient) Event(eventtype, message string) { func (c *realRecyclerClient) WatchPod(name, namespace string, stopChannel chan struct{}) (<-chan watch.Event, error) { podSelector, _ := fields.ParseSelector("metadata.name=" + name) - options := v1.ListOptions{ + options := metav1.ListOptions{ FieldSelector: podSelector.String(), Watch: true, } @@ -189,7 +189,7 @@ func (c *realRecyclerClient) WatchPod(name, namespace string, stopChannel chan s } eventSelector, _ := fields.ParseSelector("involvedObject.name=" + name) - eventWatch, err := c.client.Core().Events(namespace).Watch(v1.ListOptions{ + eventWatch, err := c.client.Core().Events(namespace).Watch(metav1.ListOptions{ FieldSelector: eventSelector.String(), Watch: true, })