changed api to api/v1

This commit is contained in:
David Ashpole 2016-11-28 14:24:43 -08:00
parent a232c15a45
commit 679c503ae3

View File

@ -20,7 +20,7 @@ import (
"fmt" "fmt"
"time" "time"
"k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1"
"k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/framework"
. "github.com/onsi/ginkgo" . "github.com/onsi/ginkgo"
@ -41,11 +41,11 @@ var _ = framework.KubeDescribe("InodeEviction [Slow] [Serial] [Disruptive]", fun
podTestSpecs := []podTestSpec{ podTestSpecs := []podTestSpec{
{ {
evictionPriority: 1, // This pod should be evicted before the normal memory usage pod evictionPriority: 1, // This pod should be evicted before the normal memory usage pod
pod: api.Pod{ pod: v1.Pod{
ObjectMeta: api.ObjectMeta{Name: "container-inode-hog-pod"}, ObjectMeta: v1.ObjectMeta{Name: "container-inode-hog-pod"},
Spec: api.PodSpec{ Spec: v1.PodSpec{
RestartPolicy: api.RestartPolicyNever, RestartPolicy: v1.RestartPolicyNever,
Containers: []api.Container{ Containers: []v1.Container{
{ {
Image: "gcr.io/google_containers/busybox:1.24", Image: "gcr.io/google_containers/busybox:1.24",
Name: "container-inode-hog-pod", Name: "container-inode-hog-pod",
@ -61,11 +61,11 @@ var _ = framework.KubeDescribe("InodeEviction [Slow] [Serial] [Disruptive]", fun
}, },
{ {
evictionPriority: 1, // This pod should be evicted before the normal memory usage pod evictionPriority: 1, // This pod should be evicted before the normal memory usage pod
pod: api.Pod{ pod: v1.Pod{
ObjectMeta: api.ObjectMeta{Name: "volume-inode-hog-pod"}, ObjectMeta: v1.ObjectMeta{Name: "volume-inode-hog-pod"},
Spec: api.PodSpec{ Spec: v1.PodSpec{
RestartPolicy: api.RestartPolicyNever, RestartPolicy: v1.RestartPolicyNever,
Containers: []api.Container{ Containers: []v1.Container{
{ {
Image: "gcr.io/google_containers/busybox:1.24", Image: "gcr.io/google_containers/busybox:1.24",
Name: "volume-inode-hog-pod", Name: "volume-inode-hog-pod",
@ -74,24 +74,24 @@ var _ = framework.KubeDescribe("InodeEviction [Slow] [Serial] [Disruptive]", fun
"-c", // Make 100 billion small files (more than we have inodes) "-c", // Make 100 billion small files (more than we have inodes)
"i=0; while [[ $i -lt 100000000000 ]]; do touch /test-empty-dir-mnt/smallfile$i.txt; sleep 0.001; i=$((i+=1)); done;", "i=0; while [[ $i -lt 100000000000 ]]; do touch /test-empty-dir-mnt/smallfile$i.txt; sleep 0.001; i=$((i+=1)); done;",
}, },
VolumeMounts: []api.VolumeMount{ VolumeMounts: []v1.VolumeMount{
{MountPath: "/test-empty-dir-mnt", Name: "test-empty-dir"}, {MountPath: "/test-empty-dir-mnt", Name: "test-empty-dir"},
}, },
}, },
}, },
Volumes: []api.Volume{ Volumes: []v1.Volume{
{Name: "test-empty-dir", VolumeSource: api.VolumeSource{EmptyDir: &api.EmptyDirVolumeSource{}}}, {Name: "test-empty-dir", VolumeSource: v1.VolumeSource{EmptyDir: &v1.EmptyDirVolumeSource{}}},
}, },
}, },
}, },
}, },
{ {
evictionPriority: 0, // This pod should never be evicted evictionPriority: 0, // This pod should never be evicted
pod: api.Pod{ pod: v1.Pod{
ObjectMeta: api.ObjectMeta{Name: "normal-memory-usage-pod"}, ObjectMeta: v1.ObjectMeta{Name: "normal-memory-usage-pod"},
Spec: api.PodSpec{ Spec: v1.PodSpec{
RestartPolicy: api.RestartPolicyNever, RestartPolicy: v1.RestartPolicyNever,
Containers: []api.Container{ Containers: []v1.Container{
{ {
Image: "gcr.io/google_containers/busybox:1.24", Image: "gcr.io/google_containers/busybox:1.24",
Name: "normal-memory-usage-pod", Name: "normal-memory-usage-pod",
@ -118,7 +118,7 @@ type podTestSpec struct {
// If two are ranked at 1, either is permitted to fail before the other. // If two are ranked at 1, either is permitted to fail before the other.
// The test ends when all other than the 0 have been evicted // The test ends when all other than the 0 have been evicted
evictionPriority int evictionPriority int
pod api.Pod pod v1.Pod
} }
// runEvictionTest sets up a testing environment given the provided nodes, and checks a few things: // runEvictionTest sets up a testing environment given the provided nodes, and checks a few things:
@ -152,7 +152,7 @@ func runEvictionTest(f *framework.Framework, testCondition string, podTestSpecs
Eventually(func() error { Eventually(func() error {
// Gather current information // Gather current information
updatedPodList, err := f.ClientSet.Core().Pods(f.Namespace.Name).List(api.ListOptions{}) updatedPodList, err := f.ClientSet.Core().Pods(f.Namespace.Name).List(v1.ListOptions{})
updatedPods := updatedPodList.Items updatedPods := updatedPodList.Items
for _, p := range updatedPods { for _, p := range updatedPods {
framework.Logf("fetching pod %s; phase= %v", p.Name, p.Status.Phase) framework.Logf("fetching pod %s; phase= %v", p.Name, p.Status.Phase)
@ -163,7 +163,7 @@ func runEvictionTest(f *framework.Framework, testCondition string, podTestSpecs
By("checking eviction ordering and ensuring important pods dont fail") By("checking eviction ordering and ensuring important pods dont fail")
done := true done := true
for _, priorityPodSpec := range podTestSpecs { for _, priorityPodSpec := range podTestSpecs {
var priorityPod api.Pod var priorityPod v1.Pod
for _, p := range updatedPods { for _, p := range updatedPods {
if p.Name == priorityPodSpec.pod.Name { if p.Name == priorityPodSpec.pod.Name {
priorityPod = p priorityPod = p
@ -174,27 +174,27 @@ func runEvictionTest(f *framework.Framework, testCondition string, podTestSpecs
// Check eviction ordering. // Check eviction ordering.
// Note: it is alright for a priority 1 and priority 2 pod (for example) to fail in the same round // Note: it is alright for a priority 1 and priority 2 pod (for example) to fail in the same round
for _, lowPriorityPodSpec := range podTestSpecs { for _, lowPriorityPodSpec := range podTestSpecs {
var lowPriorityPod api.Pod var lowPriorityPod v1.Pod
for _, p := range updatedPods { for _, p := range updatedPods {
if p.Name == lowPriorityPodSpec.pod.Name { if p.Name == lowPriorityPodSpec.pod.Name {
lowPriorityPod = p lowPriorityPod = p
} }
} }
Expect(lowPriorityPod).NotTo(BeNil()) Expect(lowPriorityPod).NotTo(BeNil())
if priorityPodSpec.evictionPriority < lowPriorityPodSpec.evictionPriority && lowPriorityPod.Status.Phase == api.PodRunning { if priorityPodSpec.evictionPriority < lowPriorityPodSpec.evictionPriority && lowPriorityPod.Status.Phase == v1.PodRunning {
Expect(priorityPod.Status.Phase).NotTo(Equal(api.PodFailed), Expect(priorityPod.Status.Phase).NotTo(Equal(v1.PodFailed),
fmt.Sprintf("%s pod failed before %s pod", priorityPodSpec.pod.Name, lowPriorityPodSpec.pod.Name)) fmt.Sprintf("%s pod failed before %s pod", priorityPodSpec.pod.Name, lowPriorityPodSpec.pod.Name))
} }
} }
// EvictionPriority 0 pods should not fail // EvictionPriority 0 pods should not fail
if priorityPodSpec.evictionPriority == 0 { if priorityPodSpec.evictionPriority == 0 {
Expect(priorityPod.Status.Phase).NotTo(Equal(api.PodFailed), Expect(priorityPod.Status.Phase).NotTo(Equal(v1.PodFailed),
fmt.Sprintf("%s pod failed (and shouldn't have failed)", priorityPod.Name)) fmt.Sprintf("%s pod failed (and shouldn't have failed)", priorityPod.Name))
} }
// If a pod that is not evictionPriority 0 has not been evicted, we are not done // If a pod that is not evictionPriority 0 has not been evicted, we are not done
if priorityPodSpec.evictionPriority != 0 && priorityPod.Status.Phase != api.PodFailed { if priorityPodSpec.evictionPriority != 0 && priorityPod.Status.Phase != v1.PodFailed {
done = false done = false
} }
} }
@ -222,13 +222,13 @@ func runEvictionTest(f *framework.Framework, testCondition string, podTestSpecs
By("making sure we can start a new pod after the test") By("making sure we can start a new pod after the test")
podName := "test-admit-pod" podName := "test-admit-pod"
f.PodClient().Create(&api.Pod{ f.PodClient().Create(&v1.Pod{
ObjectMeta: api.ObjectMeta{ ObjectMeta: v1.ObjectMeta{
Name: podName, Name: podName,
}, },
Spec: api.PodSpec{ Spec: v1.PodSpec{
RestartPolicy: api.RestartPolicyNever, RestartPolicy: v1.RestartPolicyNever,
Containers: []api.Container{ Containers: []v1.Container{
{ {
Image: "gcr.io/google_containers/busybox:1.24", Image: "gcr.io/google_containers/busybox:1.24",
Name: podName, Name: podName,
@ -247,15 +247,15 @@ func runEvictionTest(f *framework.Framework, testCondition string, podTestSpecs
// Returns TRUE if the node has disk pressure due to inodes exists on the node, FALSE otherwise // Returns TRUE if the node has disk pressure due to inodes exists on the node, FALSE otherwise
func hasInodePressure(f *framework.Framework, testCondition string) (bool, error) { func hasInodePressure(f *framework.Framework, testCondition string) (bool, error) {
nodeList, err := f.ClientSet.Core().Nodes().List(api.ListOptions{}) nodeList, err := f.ClientSet.Core().Nodes().List(v1.ListOptions{})
framework.ExpectNoError(err, "getting node list") framework.ExpectNoError(err, "getting node list")
if len(nodeList.Items) != 1 { if len(nodeList.Items) != 1 {
return false, fmt.Errorf("expected 1 node, but see %d. List: %v", len(nodeList.Items), nodeList.Items) return false, fmt.Errorf("expected 1 node, but see %d. List: %v", len(nodeList.Items), nodeList.Items)
} }
_, pressure := api.GetNodeCondition(&nodeList.Items[0].Status, api.NodeDiskPressure) _, pressure := v1.GetNodeCondition(&nodeList.Items[0].Status, v1.NodeDiskPressure)
Expect(pressure).NotTo(BeNil()) Expect(pressure).NotTo(BeNil())
hasPressure := pressure.Status == api.ConditionTrue hasPressure := pressure.Status == v1.ConditionTrue
By(fmt.Sprintf("checking if pod has %s: %v", testCondition, hasPressure)) By(fmt.Sprintf("checking if pod has %s: %v", testCondition, hasPressure))
// Additional Logging relating to Inodes // Additional Logging relating to Inodes