mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-24 12:15:52 +00:00
Merge pull request #25972 from luxas/remove_arch_constants
Automatic merge from submit-queue Use pause image depending on the server's platform when testing Removed all pause image constant strings, now the pause image is chosen by arch. Part of the effort of making e2e arch-agnostic. The pause image name and version is also now only in two places, and it's documented to bump both Also removed "amd64" constants in the code. Such constants should be replaced by `runtime.GOARCH` or by looking up the server platform Fixes: #22876 and #15140 Makes it easier for: #25730 Related: #17981 This is for `v1.3` @ixdy @thockin @vishh @kubernetes/sig-testing @andyzheng0831 @pensu
This commit is contained in:
commit
04bdd37bc4
@ -787,7 +787,7 @@ func runSchedulerNoPhantomPodsTest(client *client.Client) {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "c1",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: e2e.GetPauseImageName(client),
|
||||
Ports: []api.ContainerPort{
|
||||
{ContainerPort: 1234, HostPort: 9999},
|
||||
},
|
||||
@ -797,7 +797,7 @@ func runSchedulerNoPhantomPodsTest(client *client.Client) {
|
||||
},
|
||||
}
|
||||
|
||||
// Assuming we only have two kublets, the third pod here won't schedule
|
||||
// Assuming we only have two kubelets, the third pod here won't schedule
|
||||
// if the scheduler doesn't correctly handle the delete for the second
|
||||
// pod.
|
||||
pod.ObjectMeta.Name = "phantom.foo"
|
||||
|
@ -40,6 +40,7 @@ const (
|
||||
defaultRootDir = "/var/lib/kubelet"
|
||||
experimentalFlannelOverlay = false
|
||||
|
||||
// When these values are updated, also update test/e2e/util.go
|
||||
defaultPodInfraContainerImageName = "gcr.io/google_containers/pause"
|
||||
defaultPodInfraContainerImageVersion = "3.0"
|
||||
)
|
||||
|
@ -19,8 +19,11 @@
|
||||
# `make container` will build a container-- you must supply a tag.
|
||||
# `make push` will push the container-- you must supply a tag.
|
||||
|
||||
GOARCH?=$(shell go env GOARCH)
|
||||
GOOS?=$(shell go env GOOS)
|
||||
|
||||
kubectl:
|
||||
KUBE_STATIC_OVERRIDES="kubectl" ../../hack/build-go.sh cmd/kubectl; cp ../../_output/local/bin/linux/amd64/kubectl .
|
||||
KUBE_STATIC_OVERRIDES="kubectl" ../../hack/build-go.sh cmd/kubectl; cp ../../_output/local/bin/$(GOOS)/$(GOARCH)/kubectl .
|
||||
|
||||
.tag: kubectl
|
||||
./kubectl version -c | grep -o 'GitVersion:"[^"]*"' | cut -f 2 -d '"' > .tag
|
||||
|
@ -32,6 +32,7 @@ import (
|
||||
"k8s.io/kubernetes/plugin/pkg/scheduler"
|
||||
_ "k8s.io/kubernetes/plugin/pkg/scheduler/algorithmprovider"
|
||||
"k8s.io/kubernetes/plugin/pkg/scheduler/factory"
|
||||
e2e "k8s.io/kubernetes/test/e2e/framework"
|
||||
"k8s.io/kubernetes/test/integration/framework"
|
||||
)
|
||||
|
||||
@ -113,7 +114,7 @@ func makePodSpec() api.PodSpec {
|
||||
return api.PodSpec{
|
||||
Containers: []api.Container{{
|
||||
Name: "pause",
|
||||
Image: "gcr.io/google_containers/pause:1.0",
|
||||
Image: e2e.GetPauseImageNameForHostArch(),
|
||||
Ports: []api.ContainerPort{{ContainerPort: 80}},
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
|
@ -86,7 +86,7 @@ func CreateHostPortPods(f *framework.Framework, id string, replicas int, expectR
|
||||
Name: id,
|
||||
Namespace: f.Namespace.Name,
|
||||
Timeout: scaleTimeout,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: replicas,
|
||||
HostPorts: map[string]int{"port1": 4321},
|
||||
}
|
||||
@ -105,7 +105,7 @@ func ReserveCpu(f *framework.Framework, id string, replicas, millicores int) {
|
||||
Name: id,
|
||||
Namespace: f.Namespace.Name,
|
||||
Timeout: scaleTimeout,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: replicas,
|
||||
CpuRequest: request,
|
||||
}
|
||||
@ -120,7 +120,7 @@ func ReserveMemory(f *framework.Framework, id string, replicas, megabytes int, e
|
||||
Name: id,
|
||||
Namespace: f.Namespace.Name,
|
||||
Timeout: scaleTimeout,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: replicas,
|
||||
MemRequest: request,
|
||||
}
|
||||
|
@ -209,7 +209,7 @@ var _ = framework.KubeDescribe("DaemonRestart [Disruptive]", func() {
|
||||
Client: f.Client,
|
||||
Name: rcName,
|
||||
Namespace: ns,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: numPods,
|
||||
CreatedPods: &[]*api.Pod{},
|
||||
}
|
||||
|
@ -232,7 +232,7 @@ var _ = framework.KubeDescribe("Density", func() {
|
||||
for i := 0; i < numberOrRCs; i++ {
|
||||
RCName = "density" + strconv.Itoa(totalPods) + "-" + strconv.Itoa(i) + "-" + uuid
|
||||
RCConfigs[i] = framework.RCConfig{Client: c,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Name: RCName,
|
||||
Namespace: ns,
|
||||
Labels: map[string]string{"type": "densityPod"},
|
||||
@ -460,7 +460,7 @@ var _ = framework.KubeDescribe("Density", func() {
|
||||
}
|
||||
for i := 1; i <= nodeCount; i++ {
|
||||
name := additionalPodsPrefix + "-" + strconv.Itoa(i)
|
||||
go createRunningPodFromRC(&wg, c, name, ns, "gcr.io/google_containers/pause-amd64:3.0", additionalPodsPrefix, cpuRequest, memRequest)
|
||||
go createRunningPodFromRC(&wg, c, name, ns, framework.GetPauseImageName(f.Client), additionalPodsPrefix, cpuRequest, memRequest)
|
||||
time.Sleep(200 * time.Millisecond)
|
||||
}
|
||||
wg.Wait()
|
||||
|
@ -44,7 +44,7 @@ var _ = framework.KubeDescribe("Etcd failure [Disruptive]", func() {
|
||||
Client: f.Client,
|
||||
Name: "baz",
|
||||
Namespace: f.Namespace.Name,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: 1,
|
||||
})).NotTo(HaveOccurred())
|
||||
})
|
||||
|
@ -30,6 +30,7 @@ import (
|
||||
"os/exec"
|
||||
"path"
|
||||
"path/filepath"
|
||||
goRuntime "runtime"
|
||||
"sort"
|
||||
"strconv"
|
||||
"strings"
|
||||
@ -129,12 +130,42 @@ const (
|
||||
|
||||
// How long claims have to become dynamically provisioned
|
||||
ClaimProvisionTimeout = 5 * time.Minute
|
||||
|
||||
// When these values are updated, also update cmd/kubelet/app/options/options.go
|
||||
currentPodInfraContainerImageName = "gcr.io/google_containers/pause"
|
||||
currentPodInfraContainerImageVersion = "3.0"
|
||||
)
|
||||
|
||||
// Label allocated to the image puller static pod that runs on each node
|
||||
// before e2es.
|
||||
var ImagePullerLabels = map[string]string{"name": "e2e-image-puller"}
|
||||
|
||||
// GetServerArchitecture fetches the architecture of the target cluster.
|
||||
func GetServerArchitecture(c *client.Client) string {
|
||||
arch := ""
|
||||
sVer, err := c.Discovery().ServerVersion()
|
||||
if err != nil || sVer.Platform == "" {
|
||||
// If we failed to get the server version for some reason, default to amd64.
|
||||
arch = "amd64"
|
||||
} else {
|
||||
// Split the platform string into OS and Arch separately.
|
||||
// The platform string may for example be "linux/amd64", "linux/arm" or "windows/amd64".
|
||||
osArchArray := strings.Split(sVer.Platform, "/")
|
||||
arch = osArchArray[1]
|
||||
}
|
||||
return arch
|
||||
}
|
||||
|
||||
// GetPauseImageName fetches the architecture of the target cluster and chooses the pause image to use.
|
||||
func GetPauseImageName(c *client.Client) string {
|
||||
return currentPodInfraContainerImageName + "-" + GetServerArchitecture(c) + ":" + currentPodInfraContainerImageVersion
|
||||
}
|
||||
|
||||
// GetPauseImageNameForHostArch() fetches the pause image for the same architecture the machine is running on.
|
||||
func GetPauseImageNameForHostArch() string {
|
||||
return currentPodInfraContainerImageName + "-" + goRuntime.GOARCH + ":" + currentPodInfraContainerImageVersion
|
||||
}
|
||||
|
||||
// SubResource proxy should have been functional in v1.0.0, but SubResource
|
||||
// proxy via tunneling is known to be broken in v1.0. See
|
||||
// https://github.com/kubernetes/kubernetes/pull/15224#issuecomment-146769463
|
||||
|
@ -1448,37 +1448,6 @@ func streamingUpload(file *os.File, fileName string, postBodyWriter *multipart.W
|
||||
}
|
||||
}
|
||||
|
||||
var binPrefixes = []string{
|
||||
"_output/dockerized/bin",
|
||||
"_output/local/bin",
|
||||
"platforms",
|
||||
}
|
||||
|
||||
// findBinary searches through likely paths to find the specified binary. It
|
||||
// takes the one that has been built most recently. Platform should be
|
||||
// specified as '<os>/<arch>'. For example: 'linux/amd64'.
|
||||
func findBinary(binName string, platform string) (string, error) {
|
||||
var binTime time.Time
|
||||
var binPath string
|
||||
|
||||
for _, pre := range binPrefixes {
|
||||
tryPath := path.Join(framework.TestContext.RepoRoot, pre, platform, binName)
|
||||
fi, err := os.Stat(tryPath)
|
||||
if err != nil {
|
||||
continue
|
||||
}
|
||||
if fi.ModTime().After(binTime) {
|
||||
binPath = tryPath
|
||||
binTime = fi.ModTime()
|
||||
}
|
||||
}
|
||||
|
||||
if len(binPath) > 0 {
|
||||
return binPath, nil
|
||||
}
|
||||
return binPath, fmt.Errorf("Could not find %v for %v", binName, platform)
|
||||
}
|
||||
|
||||
func startLocalProxy() (srv *httptest.Server, logs *bytes.Buffer) {
|
||||
logs = &bytes.Buffer{}
|
||||
p := goproxy.NewProxyHttpServer()
|
||||
|
@ -128,7 +128,7 @@ var _ = framework.KubeDescribe("kubelet", func() {
|
||||
Client: f.Client,
|
||||
Name: rcName,
|
||||
Namespace: f.Namespace.Name,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: totalPods,
|
||||
})).NotTo(HaveOccurred())
|
||||
// Perform a sanity check so that we know all desired pods are
|
||||
|
@ -69,7 +69,7 @@ func runResourceTrackingTest(f *framework.Framework, podsPerNode int, nodeNames
|
||||
Client: f.Client,
|
||||
Name: rcName,
|
||||
Namespace: f.Namespace.Name,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Replicas: totalPods,
|
||||
})).NotTo(HaveOccurred())
|
||||
|
||||
|
@ -53,7 +53,7 @@ var _ = framework.KubeDescribe("LimitRange", func() {
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
By("Creating a Pod with no resource requirements")
|
||||
pod := newTestPod("pod-no-resources", api.ResourceList{}, api.ResourceList{})
|
||||
pod := newTestPod(f, "pod-no-resources", api.ResourceList{}, api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
@ -70,7 +70,7 @@ var _ = framework.KubeDescribe("LimitRange", func() {
|
||||
}
|
||||
|
||||
By("Creating a Pod with partial resource requirements")
|
||||
pod = newTestPod("pod-partial-resources", getResourceList("", "150Mi"), getResourceList("300m", ""))
|
||||
pod = newTestPod(f, "pod-partial-resources", getResourceList("", "150Mi"), getResourceList("300m", ""))
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
@ -91,12 +91,12 @@ var _ = framework.KubeDescribe("LimitRange", func() {
|
||||
}
|
||||
|
||||
By("Failing to create a Pod with less than min resources")
|
||||
pod = newTestPod(podName, getResourceList("10m", "50Mi"), api.ResourceList{})
|
||||
pod = newTestPod(f, podName, getResourceList("10m", "50Mi"), api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).To(HaveOccurred())
|
||||
|
||||
By("Failing to create a Pod with more than max resources")
|
||||
pod = newTestPod(podName, getResourceList("600m", "600Mi"), api.ResourceList{})
|
||||
pod = newTestPod(f, podName, getResourceList("600m", "600Mi"), api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).To(HaveOccurred())
|
||||
})
|
||||
@ -167,7 +167,7 @@ func newLimitRange(name string, limitType api.LimitType,
|
||||
}
|
||||
|
||||
// newTestPod returns a pod that has the specified requests and limits
|
||||
func newTestPod(name string, requests api.ResourceList, limits api.ResourceList) *api.Pod {
|
||||
func newTestPod(f *framework.Framework, name string, requests api.ResourceList, limits api.ResourceList) *api.Pod {
|
||||
return &api.Pod{
|
||||
ObjectMeta: api.ObjectMeta{
|
||||
Name: name,
|
||||
@ -175,8 +175,8 @@ func newTestPod(name string, requests api.ResourceList, limits api.ResourceList)
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "nginx",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Name: "pause",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Requests: requests,
|
||||
Limits: limits,
|
||||
|
@ -94,7 +94,7 @@ var _ = framework.KubeDescribe("Mesos", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "beta.gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
|
@ -97,7 +97,7 @@ func ensurePodsAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "nginx",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
|
@ -179,7 +179,7 @@ func createOutOfDiskPod(c *client.Client, ns, name string, milliCPU int64) {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "pause",
|
||||
Image: "beta.gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(c),
|
||||
Resources: api.ResourceRequirements{
|
||||
Requests: api.ResourceList{
|
||||
// Request enough CPU to fit only two pods on a given node.
|
||||
|
@ -219,7 +219,7 @@ var _ = framework.KubeDescribe("Pods", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "test",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -244,7 +244,7 @@ var _ = framework.KubeDescribe("Pods", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "nginx",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
api.ResourceCPU: *resource.NewMilliQuantity(100, resource.DecimalSI),
|
||||
@ -754,7 +754,7 @@ var _ = framework.KubeDescribe("Pods", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "run1",
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
api.ResourceCPU: *resource.NewMilliQuantity(100, resource.DecimalSI),
|
||||
@ -823,7 +823,7 @@ var _ = framework.KubeDescribe("Pods", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "run1",
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
api.ResourceCPU: *resource.NewMilliQuantity(100, resource.DecimalSI),
|
||||
|
@ -328,7 +328,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
requests := api.ResourceList{}
|
||||
requests[api.ResourceCPU] = resource.MustParse("500m")
|
||||
requests[api.ResourceMemory] = resource.MustParse("252Mi")
|
||||
pod := newTestPodForQuota(podName, requests, api.ResourceList{})
|
||||
pod := newTestPodForQuota(f, podName, requests, api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
podToUpdate := pod
|
||||
@ -345,7 +345,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
requests = api.ResourceList{}
|
||||
requests[api.ResourceCPU] = resource.MustParse("600m")
|
||||
requests[api.ResourceMemory] = resource.MustParse("100Mi")
|
||||
pod = newTestPodForQuota("fail-pod", requests, api.ResourceList{})
|
||||
pod = newTestPodForQuota(f, "fail-pod", requests, api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).To(HaveOccurred())
|
||||
|
||||
@ -509,7 +509,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
limits := api.ResourceList{}
|
||||
limits[api.ResourceCPU] = resource.MustParse("1")
|
||||
limits[api.ResourceMemory] = resource.MustParse("400Mi")
|
||||
pod := newTestPodForQuota(podName, requests, limits)
|
||||
pod := newTestPodForQuota(f, podName, requests, limits)
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
@ -546,7 +546,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
|
||||
By("Creating a terminating pod")
|
||||
podName = "terminating-pod"
|
||||
pod = newTestPodForQuota(podName, requests, limits)
|
||||
pod = newTestPodForQuota(f, podName, requests, limits)
|
||||
activeDeadlineSeconds := int64(3600)
|
||||
pod.Spec.ActiveDeadlineSeconds = &activeDeadlineSeconds
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
@ -604,7 +604,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
By("Creating a best-effort pod")
|
||||
pod := newTestPodForQuota(podName, api.ResourceList{}, api.ResourceList{})
|
||||
pod := newTestPodForQuota(f, podName, api.ResourceList{}, api.ResourceList{})
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
@ -634,7 +634,7 @@ var _ = framework.KubeDescribe("ResourceQuota", func() {
|
||||
limits := api.ResourceList{}
|
||||
limits[api.ResourceCPU] = resource.MustParse("1")
|
||||
limits[api.ResourceMemory] = resource.MustParse("400Mi")
|
||||
pod = newTestPodForQuota("burstable-pod", requests, limits)
|
||||
pod = newTestPodForQuota(f, "burstable-pod", requests, limits)
|
||||
pod, err = f.Client.Pods(f.Namespace.Name).Create(pod)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
|
||||
@ -697,7 +697,7 @@ func newTestResourceQuota(name string) *api.ResourceQuota {
|
||||
}
|
||||
|
||||
// newTestPodForQuota returns a pod that has the specified requests and limits
|
||||
func newTestPodForQuota(name string, requests api.ResourceList, limits api.ResourceList) *api.Pod {
|
||||
func newTestPodForQuota(f *framework.Framework, name string, requests api.ResourceList, limits api.ResourceList) *api.Pod {
|
||||
return &api.Pod{
|
||||
ObjectMeta: api.ObjectMeta{
|
||||
Name: name,
|
||||
@ -705,8 +705,8 @@ func newTestPodForQuota(name string, requests api.ResourceList, limits api.Resou
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "nginx",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Name: "pause",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Requests: requests,
|
||||
Limits: limits,
|
||||
|
@ -239,7 +239,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -258,7 +258,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -316,7 +316,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
"cpu": *resource.NewMilliQuantity(milliCpuPerPod, "DecimalSI"),
|
||||
@ -343,7 +343,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Resources: api.ResourceRequirements{
|
||||
Limits: api.ResourceList{
|
||||
"cpu": *resource.NewMilliQuantity(milliCpuPerPod, "DecimalSI"),
|
||||
@ -383,7 +383,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
NodeSelector: map[string]string{
|
||||
@ -426,7 +426,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -462,7 +462,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -500,7 +500,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
NodeSelector: map[string]string{
|
||||
@ -564,7 +564,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -599,7 +599,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -655,7 +655,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -693,7 +693,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -772,7 +772,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -823,7 +823,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -858,7 +858,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -911,7 +911,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -950,7 +950,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1003,7 +1003,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1038,7 +1038,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1099,7 +1099,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1138,7 +1138,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1202,7 +1202,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: labelPodName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1241,7 +1241,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause:2.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1305,7 +1305,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1370,7 +1370,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: tolerationPodName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1427,7 +1427,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podName,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -1482,7 +1482,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: podNameNoTolerations,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
|
@ -1138,8 +1138,8 @@ func createPodOrFail(c *client.Client, ns, name string, labels map[string]string
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "test",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Name: "pause",
|
||||
Image: framework.GetPauseImageName(c),
|
||||
Ports: containerPorts,
|
||||
// Add a dummy environment variable to work around a docker issue.
|
||||
// https://github.com/docker/docker/issues/14203
|
||||
|
@ -118,7 +118,7 @@ var _ = framework.KubeDescribe("Service endpoints latency", func() {
|
||||
func runServiceLatencies(f *framework.Framework, inParallel, total int) (output []time.Duration, err error) {
|
||||
cfg := framework.RCConfig{
|
||||
Client: f.Client,
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
Name: "svc-latency-rc",
|
||||
Namespace: f.Namespace.Name,
|
||||
Replicas: 1,
|
||||
|
@ -88,7 +88,7 @@ func SpreadServiceOrFail(f *framework.Framework, replicaCount int, image string)
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "test",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: framework.GetPauseImageName(f.Client),
|
||||
},
|
||||
},
|
||||
},
|
||||
|
@ -21,6 +21,8 @@ import (
|
||||
"time"
|
||||
|
||||
"github.com/golang/glog"
|
||||
|
||||
"k8s.io/kubernetes/test/e2e/framework"
|
||||
)
|
||||
|
||||
const (
|
||||
@ -45,7 +47,7 @@ var ImageRegistry = map[int]string{
|
||||
hostExecImage: "gcr.io/google_containers/hostexec:1.2",
|
||||
netExecImage: "gcr.io/google_containers/netexec:1.4",
|
||||
nginxImage: "gcr.io/google_containers/nginx:1.7.9",
|
||||
pauseImage: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
pauseImage: framework.GetPauseImageNameForHostArch(),
|
||||
}
|
||||
|
||||
// These are used by tests that explicitly test the ability to pull images
|
||||
|
@ -35,6 +35,7 @@ import (
|
||||
"k8s.io/kubernetes/pkg/util/wait"
|
||||
"k8s.io/kubernetes/pkg/version"
|
||||
"k8s.io/kubernetes/pkg/watch"
|
||||
e2e "k8s.io/kubernetes/test/e2e/framework"
|
||||
"k8s.io/kubernetes/test/integration/framework"
|
||||
)
|
||||
|
||||
@ -234,8 +235,8 @@ func TestMultiWatch(t *testing.T) {
|
||||
},
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{{
|
||||
Name: "nothing",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Name: "pause",
|
||||
Image: e2e.GetPauseImageName(client),
|
||||
}},
|
||||
},
|
||||
})
|
||||
@ -341,7 +342,7 @@ func TestMultiWatch(t *testing.T) {
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{{
|
||||
Name: "nothing",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: e2e.GetPauseImageName(client),
|
||||
}},
|
||||
},
|
||||
})
|
||||
@ -372,7 +373,7 @@ func TestMultiWatch(t *testing.T) {
|
||||
if err != nil {
|
||||
panic(fmt.Sprintf("Couldn't get %v: %v", name, err))
|
||||
}
|
||||
pod.Spec.Containers[0].Image = "gcr.io/google_containers/pause-amd64:3.0"
|
||||
pod.Spec.Containers[0].Image = e2e.GetPauseImageName(client)
|
||||
sentTimes <- timePair{time.Now(), name}
|
||||
if _, err := client.Pods(ns).Update(pod); err != nil {
|
||||
panic(fmt.Sprintf("Couldn't make %v: %v", name, err))
|
||||
|
@ -42,6 +42,7 @@ import (
|
||||
_ "k8s.io/kubernetes/plugin/pkg/scheduler/algorithmprovider"
|
||||
schedulerapi "k8s.io/kubernetes/plugin/pkg/scheduler/api"
|
||||
"k8s.io/kubernetes/plugin/pkg/scheduler/factory"
|
||||
e2e "k8s.io/kubernetes/test/e2e/framework"
|
||||
"k8s.io/kubernetes/test/integration/framework"
|
||||
)
|
||||
|
||||
@ -283,7 +284,7 @@ func DoTestPodScheduling(t *testing.T, restClient *client.Client) {
|
||||
pod := &api.Pod{
|
||||
ObjectMeta: api.ObjectMeta{Name: "extender-test-pod"},
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{{Name: "container", Image: "gcr.io/google_containers/pause-amd64:3.0"}},
|
||||
Containers: []api.Container{{Name: "container", Image: e2e.GetPauseImageName(restClient)}},
|
||||
},
|
||||
}
|
||||
|
||||
|
@ -41,6 +41,7 @@ import (
|
||||
"k8s.io/kubernetes/plugin/pkg/scheduler"
|
||||
_ "k8s.io/kubernetes/plugin/pkg/scheduler/algorithmprovider"
|
||||
"k8s.io/kubernetes/plugin/pkg/scheduler/factory"
|
||||
e2e "k8s.io/kubernetes/test/e2e/framework"
|
||||
"k8s.io/kubernetes/test/integration/framework"
|
||||
)
|
||||
|
||||
@ -232,7 +233,7 @@ func DoTestUnschedulableNodes(t *testing.T, restClient *client.Client, nodeStore
|
||||
pod := &api.Pod{
|
||||
ObjectMeta: api.ObjectMeta{Name: "node-scheduling-test-pod"},
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{{Name: "container", Image: "gcr.io/google_containers/pause-amd64:3.0"}},
|
||||
Containers: []api.Container{{Name: "container", Image: e2e.GetPauseImageName(restClient)}},
|
||||
},
|
||||
}
|
||||
myPod, err := restClient.Pods(api.NamespaceDefault).Create(pod)
|
||||
@ -339,21 +340,21 @@ func TestMultiScheduler(t *testing.T) {
|
||||
restClient.Nodes().Create(node)
|
||||
|
||||
// 3. create 3 pods for testing
|
||||
podWithNoAnnotation := createPod("pod-with-no-annotation", nil)
|
||||
podWithNoAnnotation := createPod(restClient, "pod-with-no-annotation", nil)
|
||||
testPodNoAnnotation, err := restClient.Pods(api.NamespaceDefault).Create(podWithNoAnnotation)
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create pod: %v", err)
|
||||
}
|
||||
|
||||
schedulerAnnotationFitsDefault := map[string]string{"scheduler.alpha.kubernetes.io/name": "default-scheduler"}
|
||||
podWithAnnotationFitsDefault := createPod("pod-with-annotation-fits-default", schedulerAnnotationFitsDefault)
|
||||
podWithAnnotationFitsDefault := createPod(restClient, "pod-with-annotation-fits-default", schedulerAnnotationFitsDefault)
|
||||
testPodWithAnnotationFitsDefault, err := restClient.Pods(api.NamespaceDefault).Create(podWithAnnotationFitsDefault)
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create pod: %v", err)
|
||||
}
|
||||
|
||||
schedulerAnnotationFitsFoo := map[string]string{"scheduler.alpha.kubernetes.io/name": "foo-scheduler"}
|
||||
podWithAnnotationFitsFoo := createPod("pod-with-annotation-fits-foo", schedulerAnnotationFitsFoo)
|
||||
podWithAnnotationFitsFoo := createPod(restClient, "pod-with-annotation-fits-foo", schedulerAnnotationFitsFoo)
|
||||
testPodWithAnnotationFitsFoo, err := restClient.Pods(api.NamespaceDefault).Create(podWithAnnotationFitsFoo)
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create pod: %v", err)
|
||||
@ -456,11 +457,11 @@ func TestMultiScheduler(t *testing.T) {
|
||||
*/
|
||||
}
|
||||
|
||||
func createPod(name string, annotation map[string]string) *api.Pod {
|
||||
func createPod(client *client.Client, name string, annotation map[string]string) *api.Pod {
|
||||
return &api.Pod{
|
||||
ObjectMeta: api.ObjectMeta{Name: name, Annotations: annotation},
|
||||
Spec: api.PodSpec{
|
||||
Containers: []api.Container{{Name: "container", Image: "gcr.io/google_containers/pause-amd64:3.0"}},
|
||||
Containers: []api.Container{{Name: "container", Image: e2e.GetPauseImageName(client)}},
|
||||
},
|
||||
}
|
||||
}
|
||||
@ -521,7 +522,7 @@ func TestAllocatable(t *testing.T) {
|
||||
Containers: []api.Container{
|
||||
{
|
||||
Name: "container",
|
||||
Image: "gcr.io/google_containers/pause-amd64:3.0",
|
||||
Image: e2e.GetPauseImageName(restClient),
|
||||
Resources: api.ResourceRequirements{
|
||||
Requests: api.ResourceList{
|
||||
api.ResourceCPU: *resource.NewMilliQuantity(20, resource.DecimalSI),
|
||||
|
Loading…
Reference in New Issue
Block a user