mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-24 20:24:09 +00:00
Merge pull request #97880 from justinsb/spelling
Fix spelling typos: scheduable -> schedulable
This commit is contained in:
commit
e1c1fd9edf
@ -184,7 +184,7 @@ func removeOldDNSDeploymentIfAnotherDNSIsUsed(cfg *kubeadmapi.ClusterConfigurati
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// If we're dry-running or there are no scheduable nodes available, we don't need to wait for the new DNS addon to become ready
|
// If we're dry-running or there are no schedulable nodes available, we don't need to wait for the new DNS addon to become ready
|
||||||
if !dryRun && len(nodes.Items) != 0 {
|
if !dryRun && len(nodes.Items) != 0 {
|
||||||
dnsDeployment, err := client.AppsV1().Deployments(metav1.NamespaceSystem).Get(context.TODO(), installedDeploymentName, metav1.GetOptions{})
|
dnsDeployment, err := client.AppsV1().Deployments(metav1.NamespaceSystem).Get(context.TODO(), installedDeploymentName, metav1.GetOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -137,7 +137,7 @@ type PriorityQueue struct {
|
|||||||
// when a pod is popped.
|
// when a pod is popped.
|
||||||
schedulingCycle int64
|
schedulingCycle int64
|
||||||
// moveRequestCycle caches the sequence number of scheduling cycle when we
|
// moveRequestCycle caches the sequence number of scheduling cycle when we
|
||||||
// received a move request. Unscheduable pods in and before this scheduling
|
// received a move request. Unschedulable pods in and before this scheduling
|
||||||
// cycle will be put back to activeQueue if we were trying to schedule them
|
// cycle will be put back to activeQueue if we were trying to schedule them
|
||||||
// when we received move request.
|
// when we received move request.
|
||||||
moveRequestCycle int64
|
moveRequestCycle int64
|
||||||
|
@ -27,7 +27,7 @@ const (
|
|||||||
TaintNodeUnreachable = "node.kubernetes.io/unreachable"
|
TaintNodeUnreachable = "node.kubernetes.io/unreachable"
|
||||||
|
|
||||||
// TaintNodeUnschedulable will be added when node becomes unschedulable
|
// TaintNodeUnschedulable will be added when node becomes unschedulable
|
||||||
// and removed when node becomes scheduable.
|
// and removed when node becomes schedulable.
|
||||||
TaintNodeUnschedulable = "node.kubernetes.io/unschedulable"
|
TaintNodeUnschedulable = "node.kubernetes.io/unschedulable"
|
||||||
|
|
||||||
// TaintNodeMemoryPressure will be added when node has memory pressure
|
// TaintNodeMemoryPressure will be added when node has memory pressure
|
||||||
|
@ -243,13 +243,13 @@ func getExpectReplicasFuncLinear(c clientset.Interface, params *DNSParamsLinear)
|
|||||||
replicasFromNodes = math.Ceil(float64(len(nodes.Items)) / params.nodesPerReplica)
|
replicasFromNodes = math.Ceil(float64(len(nodes.Items)) / params.nodesPerReplica)
|
||||||
}
|
}
|
||||||
if params.coresPerReplica > 0 {
|
if params.coresPerReplica > 0 {
|
||||||
replicasFromCores = math.Ceil(float64(getScheduableCores(nodes.Items)) / params.coresPerReplica)
|
replicasFromCores = math.Ceil(float64(getSchedulableCores(nodes.Items)) / params.coresPerReplica)
|
||||||
}
|
}
|
||||||
return int(math.Max(1.0, math.Max(replicasFromNodes, replicasFromCores)))
|
return int(math.Max(1.0, math.Max(replicasFromNodes, replicasFromCores)))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func getScheduableCores(nodes []v1.Node) int64 {
|
func getSchedulableCores(nodes []v1.Node) int64 {
|
||||||
var sc resource.Quantity
|
var sc resource.Quantity
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
if !node.Spec.Unschedulable {
|
if !node.Spec.Unschedulable {
|
||||||
|
@ -1022,7 +1022,7 @@ func getNodeEvents(c clientset.Interface, nodeName string) []v1.Event {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// WaitForAllNodesSchedulable waits up to timeout for all
|
// WaitForAllNodesSchedulable waits up to timeout for all
|
||||||
// (but TestContext.AllowedNotReadyNodes) to become scheduable.
|
// (but TestContext.AllowedNotReadyNodes) to become schedulable.
|
||||||
func WaitForAllNodesSchedulable(c clientset.Interface, timeout time.Duration) error {
|
func WaitForAllNodesSchedulable(c clientset.Interface, timeout time.Duration) error {
|
||||||
Logf("Waiting up to %v for all (but %d) nodes to be schedulable", timeout, TestContext.AllowedNotReadyNodes)
|
Logf("Waiting up to %v for all (but %d) nodes to be schedulable", timeout, TestContext.AllowedNotReadyNodes)
|
||||||
|
|
||||||
|
@ -959,7 +959,7 @@ func testSubpathReconstruction(f *framework.Framework, hostExec utils.HostExec,
|
|||||||
// Disruptive test run serially, we can cache all voluem global mount
|
// Disruptive test run serially, we can cache all voluem global mount
|
||||||
// points and verify after the test that we do not leak any global mount point.
|
// points and verify after the test that we do not leak any global mount point.
|
||||||
nodeList, err := e2enode.GetReadySchedulableNodes(f.ClientSet)
|
nodeList, err := e2enode.GetReadySchedulableNodes(f.ClientSet)
|
||||||
framework.ExpectNoError(err, "while listing scheduable nodes")
|
framework.ExpectNoError(err, "while listing schedulable nodes")
|
||||||
globalMountPointsByNode := make(map[string]sets.String, len(nodeList.Items))
|
globalMountPointsByNode := make(map[string]sets.String, len(nodeList.Items))
|
||||||
for _, node := range nodeList.Items {
|
for _, node := range nodeList.Items {
|
||||||
globalMountPointsByNode[node.Name] = utils.FindVolumeGlobalMountPoints(hostExec, &node)
|
globalMountPointsByNode[node.Name] = utils.FindVolumeGlobalMountPoints(hostExec, &node)
|
||||||
@ -993,7 +993,7 @@ func testSubpathReconstruction(f *framework.Framework, hostExec utils.HostExec,
|
|||||||
podNode = &nodeList.Items[i]
|
podNode = &nodeList.Items[i]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
framework.ExpectNotEqual(podNode, nil, "pod node should exist in scheduable nodes")
|
framework.ExpectNotEqual(podNode, nil, "pod node should exist in schedulable nodes")
|
||||||
|
|
||||||
utils.TestVolumeUnmountsFromDeletedPodWithForceOption(f.ClientSet, f, pod, forceDelete, true)
|
utils.TestVolumeUnmountsFromDeletedPodWithForceOption(f.ClientSet, f, pod, forceDelete, true)
|
||||||
|
|
||||||
|
@ -762,7 +762,7 @@ func GetReadySchedulableNodeInfos() []*NodeInfo {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// GetReadySchedulableRandomNodeInfo returns NodeInfo object for one of the Ready and Schedulable Node.
|
// GetReadySchedulableRandomNodeInfo returns NodeInfo object for one of the Ready and Schedulable Node.
|
||||||
// if multiple nodes are present with Ready and Scheduable state then one of the Node is selected randomly
|
// if multiple nodes are present with Ready and Schedulable state then one of the Node is selected randomly
|
||||||
// and it's associated NodeInfo object is returned.
|
// and it's associated NodeInfo object is returned.
|
||||||
func GetReadySchedulableRandomNodeInfo() *NodeInfo {
|
func GetReadySchedulableRandomNodeInfo() *NodeInfo {
|
||||||
nodesInfo := GetReadySchedulableNodeInfos()
|
nodesInfo := GetReadySchedulableNodeInfos()
|
||||||
|
@ -381,7 +381,7 @@ func TestVolumeBindingRescheduling(t *testing.T) {
|
|||||||
// Trigger
|
// Trigger
|
||||||
test.trigger(config)
|
test.trigger(config)
|
||||||
|
|
||||||
// Wait for pod is scheduled or unscheduable.
|
// Wait for pod is scheduled or unschedulable.
|
||||||
if !test.shouldFail {
|
if !test.shouldFail {
|
||||||
klog.Infof("Waiting for pod is scheduled")
|
klog.Infof("Waiting for pod is scheduled")
|
||||||
if err := waitForPodToSchedule(config.client, test.pod); err != nil {
|
if err := waitForPodToSchedule(config.client, test.pod); err != nil {
|
||||||
|
Loading…
Reference in New Issue
Block a user