mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-23 19:56:01 +00:00
Merge pull request #19987 from gmarek/remove-test
Remove monitor_resources test
This commit is contained in:
commit
0b00928c74
@ -1,135 +0,0 @@
|
||||
/*
|
||||
Copyright 2015 The Kubernetes Authors All rights reserved.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package e2e
|
||||
|
||||
import (
|
||||
"time"
|
||||
|
||||
"k8s.io/kubernetes/pkg/api"
|
||||
client "k8s.io/kubernetes/pkg/client/unversioned"
|
||||
|
||||
. "github.com/onsi/ginkgo"
|
||||
. "github.com/onsi/gomega"
|
||||
)
|
||||
|
||||
const datapointAmount = 5
|
||||
|
||||
var systemContainers = []string{"/docker-daemon", "/kubelet", "/system"}
|
||||
|
||||
var allowedUsage = resourceUsagePerContainer{
|
||||
"/docker-daemon": &containerResourceUsage{
|
||||
CPUUsageInCores: 0.1,
|
||||
MemoryWorkingSetInBytes: 1500000000,
|
||||
},
|
||||
// TODO: Make Kubelet constraints sane again when #17774 is fixed.
|
||||
"/kubelet": &containerResourceUsage{
|
||||
CPUUsageInCores: 0.1,
|
||||
MemoryWorkingSetInBytes: 300000000,
|
||||
},
|
||||
"/system": &containerResourceUsage{
|
||||
CPUUsageInCores: 0.30,
|
||||
MemoryWorkingSetInBytes: 1500000000,
|
||||
},
|
||||
}
|
||||
|
||||
func computeAverage(sliceOfUsages []resourceUsagePerContainer) (result resourceUsagePerContainer) {
|
||||
result = make(resourceUsagePerContainer)
|
||||
for _, container := range systemContainers {
|
||||
result[container] = &containerResourceUsage{}
|
||||
}
|
||||
for _, usage := range sliceOfUsages {
|
||||
for _, container := range systemContainers {
|
||||
singleResult := &containerResourceUsage{
|
||||
CPUUsageInCores: result[container].CPUUsageInCores + usage[container].CPUUsageInCores,
|
||||
MemoryWorkingSetInBytes: result[container].MemoryWorkingSetInBytes + usage[container].MemoryWorkingSetInBytes,
|
||||
}
|
||||
result[container] = singleResult
|
||||
}
|
||||
}
|
||||
for _, container := range systemContainers {
|
||||
singleResult := &containerResourceUsage{
|
||||
CPUUsageInCores: result[container].CPUUsageInCores / float64(len(sliceOfUsages)),
|
||||
MemoryWorkingSetInBytes: result[container].MemoryWorkingSetInBytes / int64(len(sliceOfUsages)),
|
||||
}
|
||||
result[container] = singleResult
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
// This tests does nothing except checking current resource usage of containers defined in kubelet_stats systemContainers variable.
|
||||
// Test fails if an average container resource consumption over datapointAmount tries exceeds amount defined in allowedUsage.
|
||||
// Flaky issue #13931
|
||||
var _ = Describe("Resource usage of system containers [Serial] [Flaky]", func() {
|
||||
var c *client.Client
|
||||
BeforeEach(func() {
|
||||
var err error
|
||||
c, err = loadClient()
|
||||
expectNoError(err)
|
||||
})
|
||||
|
||||
It("should not exceed expected amount.", func() {
|
||||
By("Getting ResourceConsumption on all nodes")
|
||||
// It should be OK to list unschedulable Nodes here.
|
||||
nodeList, err := c.Nodes().List(api.ListOptions{})
|
||||
expectNoError(err)
|
||||
|
||||
resourceUsagePerNode := make(map[string][]resourceUsagePerContainer)
|
||||
|
||||
for i := 0; i < datapointAmount; i++ {
|
||||
for _, node := range nodeList.Items {
|
||||
resourceUsage, err := getOneTimeResourceUsageOnNode(c, node.Name, 15*time.Second, func() []string {
|
||||
if providerIs("gce", "gke") {
|
||||
return systemContainers
|
||||
} else {
|
||||
return []string{}
|
||||
}
|
||||
}, false)
|
||||
expectNoError(err)
|
||||
resourceUsagePerNode[node.Name] = append(resourceUsagePerNode[node.Name], resourceUsage)
|
||||
}
|
||||
time.Sleep(3 * time.Second)
|
||||
}
|
||||
|
||||
averageResourceUsagePerNode := make(map[string]resourceUsagePerContainer)
|
||||
for _, node := range nodeList.Items {
|
||||
averageResourceUsagePerNode[node.Name] = computeAverage(resourceUsagePerNode[node.Name])
|
||||
}
|
||||
|
||||
Logf("Observed resource usage:\n%v", averageResourceUsagePerNode)
|
||||
|
||||
violating := make(map[string]resourceUsagePerContainer)
|
||||
for node, usage := range averageResourceUsagePerNode {
|
||||
for container, cUsage := range usage {
|
||||
Logf("%v on %v usage: %#v", container, node, cUsage)
|
||||
if !allowedUsage[container].isStrictlyGreaterThan(cUsage) {
|
||||
if _, ok := violating[node]; !ok {
|
||||
violating[node] = make(resourceUsagePerContainer)
|
||||
}
|
||||
if allowedUsage[container].CPUUsageInCores < cUsage.CPUUsageInCores {
|
||||
Logf("CPU is too high for %s (%v)", container, cUsage.CPUUsageInCores)
|
||||
}
|
||||
if allowedUsage[container].MemoryWorkingSetInBytes < cUsage.MemoryWorkingSetInBytes {
|
||||
Logf("Working set is too high for %s (%v)", container, cUsage.MemoryWorkingSetInBytes)
|
||||
}
|
||||
violating[node][container] = usage[container]
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Expect(violating).To(BeEmpty())
|
||||
})
|
||||
})
|
Loading…
Reference in New Issue
Block a user