Reduce pod usage for network e2e test

This commit is contained in:
Satnam Singh 2015-03-18 17:46:31 -07:00
parent 52e1ee9d5b
commit 4645d13e77
2 changed files with 66 additions and 49 deletions

View File

@ -23,7 +23,6 @@ import (
"github.com/GoogleCloudPlatform/kubernetes/pkg/api"
"github.com/GoogleCloudPlatform/kubernetes/pkg/client"
"github.com/GoogleCloudPlatform/kubernetes/pkg/kubectl"
"github.com/GoogleCloudPlatform/kubernetes/pkg/util"
. "github.com/onsi/ginkgo"
@ -48,6 +47,16 @@ var _ = Describe("Networking", func() {
return
}
// Obtain a list of nodes so we can place one webserver container on each node.
nodes, err := c.Nodes().List()
if err != nil {
Failf("Failed to list nodes: %v", err)
}
peers := len(nodes.Items)
if peers == 0 {
Failf("Failed to find any nodes")
}
// Test basic external connectivity.
resp, err := http.Get("http://google.com/")
if err != nil {
@ -87,53 +96,54 @@ var _ = Describe("Networking", func() {
}
}()
By("Creating a replication controller")
rc, err := c.ReplicationControllers(ns).Create(&api.ReplicationController{
ObjectMeta: api.ObjectMeta{
Name: name,
Labels: map[string]string{
"name": name,
},
},
Spec: api.ReplicationControllerSpec{
Replicas: 8,
Selector: map[string]string{
"name": name,
},
Template: &api.PodTemplateSpec{
ObjectMeta: api.ObjectMeta{
Labels: map[string]string{"name": name},
By("Creating a webserver pod on each node")
podNames := []string{}
for i, node := range nodes.Items {
podName := fmt.Sprintf("%s-%d", name, i)
podNames = append(podNames, podName)
Logf("Creating pod %s on node %s", podName, node.Name)
_, err := c.Pods(ns).Create(&api.Pod{
ObjectMeta: api.ObjectMeta{
Name: podName,
Labels: map[string]string{
"name": name,
},
Spec: api.PodSpec{
Containers: []api.Container{
{
Name: "webserver",
Image: "kubernetes/nettest:1.1",
Command: []string{"-service=" + name, "-namespace=" + ns},
Ports: []api.ContainerPort{{ContainerPort: 8080}},
},
},
Spec: api.PodSpec{
Containers: []api.Container{
{
Name: "webserver",
Image: "kubernetes/nettest:1.1",
Command: []string{
"-service=" + name,
fmt.Sprintf("-peers=%d", peers),
"-namespace=" + ns},
Ports: []api.ContainerPort{{ContainerPort: 8080}},
},
},
Host: node.Name,
RestartPolicy: api.RestartPolicyNever,
},
},
})
if err != nil {
Fail(fmt.Sprintf("unable to create test rc: %v", err))
})
Expect(err).NotTo(HaveOccurred())
}
// Clean up rc
// Clean up the pods
defer func() {
defer GinkgoRecover()
By("Cleaning up the replication controller")
// Resize the replication controller to zero to get rid of pods.
rcReaper, err := kubectl.ReaperFor("ReplicationController", c)
if err != nil {
Fail(fmt.Sprintf("unable to stop rc %v: %v", rc.Name, err))
}
if _, err = rcReaper.Stop(ns, rc.Name); err != nil {
Fail(fmt.Sprintf("unable to stop rc %v: %v", rc.Name, err))
By("Cleaning up the webserver pods")
for _, podName := range podNames {
if err = c.Pods(ns).Delete(podName); err != nil {
Logf("Failed to delete pod %s: %v", podName, err)
}
}
}()
By("Wait for the webserver pods to be ready")
for _, podName := range podNames {
err = waitForPodRunningInNamespace(c, podName, ns)
Expect(err).NotTo(HaveOccurred())
}
By("Waiting for connectivity to be verified")
const maxAttempts = 60
passed := false
@ -148,22 +158,26 @@ var _ = Describe("Networking", func() {
Suffix("status").
Do().Raw()
if err != nil {
fmt.Printf("Attempt %v/%v: service/pod still starting. (error: '%v')\n", i, maxAttempts, err)
Logf("Attempt %v/%v: service/pod still starting. (error: '%v')", i, maxAttempts, err)
continue
}
switch string(body) {
case "pass":
fmt.Printf("Passed on attempt %v. Cleaning up.\n", i)
Logf("Passed on attempt %v. Cleaning up.", i)
passed = true
break
case "running":
fmt.Printf("Attempt %v/%v: test still running\n", i, maxAttempts)
Logf("Attempt %v/%v: test still running", i, maxAttempts)
break
case "fail":
if body, err = c.Get().Namespace(ns).Prefix("proxy").Resource("services").Name(svc.Name).Suffix("read").Do().Raw(); err != nil {
if body, err = c.Get().
Namespace(ns).Prefix("proxy").
Resource("services").
Name(svc.Name).Suffix("read").
Do().Raw(); err != nil {
Fail(fmt.Sprintf("Failed on attempt %v. Cleaning up. Error reading details: %v", i, err))
} else {
Fail(fmt.Sprintf("Failed on attempt %v. Cleaning up. Details:\n%v", i, string(body)))
Fail(fmt.Sprintf("Failed on attempt %v. Cleaning up. Details:\n%s", i, string(body)))
}
break
}
@ -179,7 +193,7 @@ var _ = Describe("Networking", func() {
Do().Raw(); err != nil {
Fail(fmt.Sprintf("Timed out. Cleaning up. Error reading details: %v", err))
} else {
Fail(fmt.Sprintf("Timed out. Cleaning up. Details:\n%v", string(body)))
Fail(fmt.Sprintf("Timed out. Cleaning up. Details:\n%s", string(body)))
}
}
Expect(string(body)).To(Equal("pass"))
@ -198,8 +212,7 @@ var _ = Describe("Networking", func() {
data, err := c.RESTClient.Get().
Namespace(ns).
AbsPath(test.path).
Do().
Raw()
Do().Raw()
if err != nil {
Fail(fmt.Sprintf("Failed: %v\nBody: %s", err, string(data)))
}

View File

@ -71,17 +71,21 @@ func waitForPodCondition(c *client.Client, ns, podName, desc string, condition p
if done {
return err
}
Logf("Waiting for pod %s status to be %q (found %q) (%.2f seconds)", podName, desc, pod.Status.Phase, time.Since(start).Seconds())
Logf("Waiting for pod %s in namespace %s status to be %q (found %q) (%v)", podName, ns, desc, pod.Status.Phase, time.Since(start))
}
return fmt.Errorf("gave up waiting for pod %s to be %s after %.2f seconds", podName, desc, podStartTimeout.Seconds())
}
func waitForPodRunning(c *client.Client, podName string) error {
return waitForPodCondition(c, api.NamespaceDefault, podName, "running", func(pod *api.Pod) (bool, error) {
func waitForPodRunningInNamespace(c *client.Client, podName string, namespace string) error {
return waitForPodCondition(c, namespace, podName, "running", func(pod *api.Pod) (bool, error) {
return (pod.Status.Phase == api.PodRunning), nil
})
}
func waitForPodRunning(c *client.Client, podName string) error {
return waitForPodRunningInNamespace(c, podName, api.NamespaceDefault)
}
// waitForPodNotPending returns an error if it took too long for the pod to go out of pending state.
func waitForPodNotPending(c *client.Client, ns, podName string) error {
return waitForPodCondition(c, ns, podName, "!pending", func(pod *api.Pod) (bool, error) {