mirror of
https://github.com/rancher/rke.git
synced 2025-07-17 17:01:08 +00:00
Merge pull request #58 from moelsayed/command_overhaul
Command overhaul
This commit is contained in:
commit
e21bf80950
@ -5,7 +5,7 @@ import (
|
|||||||
"github.com/rancher/rke/services"
|
"github.com/rancher/rke/services"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (c *Cluster) ClusterDown() error {
|
func (c *Cluster) ClusterRemove() error {
|
||||||
// Remove Worker Plane
|
// Remove Worker Plane
|
||||||
if err := services.RemoveWorkerPlane(c.ControlPlaneHosts, c.WorkerHosts); err != nil {
|
if err := services.RemoveWorkerPlane(c.ControlPlaneHosts, c.WorkerHosts); err != nil {
|
||||||
return err
|
return err
|
@ -1,67 +0,0 @@
|
|||||||
package cluster
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
|
|
||||||
"github.com/rancher/rke/hosts"
|
|
||||||
"github.com/rancher/rke/k8s"
|
|
||||||
"github.com/rancher/rke/services"
|
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
|
||||||
|
|
||||||
func (c *Cluster) ClusterUpgrade() error {
|
|
||||||
// make sure all nodes are Ready
|
|
||||||
logrus.Debugf("[upgrade] Checking node status")
|
|
||||||
if err := checkK8sNodesState(c.LocalKubeConfigPath); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// upgrade Contol Plane
|
|
||||||
logrus.Infof("[upgrade] Upgrading Control Plane Services")
|
|
||||||
if err := services.UpgradeControlPlane(c.ControlPlaneHosts, c.EtcdHosts, c.Services); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Infof("[upgrade] Control Plane Services updgraded successfully")
|
|
||||||
|
|
||||||
// upgrade Worker Plane
|
|
||||||
logrus.Infof("[upgrade] Upgrading Worker Plane Services")
|
|
||||||
if err := services.UpgradeWorkerPlane(c.ControlPlaneHosts, c.WorkerHosts, c.Services, c.LocalKubeConfigPath); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Infof("[upgrade] Worker Plane Services updgraded successfully")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func checkK8sNodesState(localConfigPath string) error {
|
|
||||||
k8sClient, err := k8s.NewClient(localConfigPath)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
nodeList, err := k8s.GetNodeList(k8sClient)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
for _, node := range nodeList.Items {
|
|
||||||
ready := k8s.IsNodeReady(node)
|
|
||||||
if !ready {
|
|
||||||
return fmt.Errorf("[upgrade] Node: %s is NotReady", node.Name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
logrus.Infof("[upgrade] All nodes are Ready")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func CheckHostsChangedOnUpgrade(kubeCluster, currentCluster *Cluster) error {
|
|
||||||
etcdChanged := hosts.IsHostListChanged(currentCluster.EtcdHosts, kubeCluster.EtcdHosts)
|
|
||||||
if etcdChanged {
|
|
||||||
return fmt.Errorf("Adding or removing Etcd nodes while upgrade is not supported")
|
|
||||||
}
|
|
||||||
cpChanged := hosts.IsHostListChanged(currentCluster.ControlPlaneHosts, kubeCluster.ControlPlaneHosts)
|
|
||||||
if cpChanged {
|
|
||||||
return fmt.Errorf("Adding or removing Control plane nodes while upgrade is not supported")
|
|
||||||
}
|
|
||||||
workerChanged := hosts.IsHostListChanged(currentCluster.WorkerHosts, kubeCluster.WorkerHosts)
|
|
||||||
if workerChanged {
|
|
||||||
return fmt.Errorf("Adding or removing Worker plane nodes while upgrade is not supported")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
286
cmd/cluster.go
286
cmd/cluster.go
@ -1,286 +0,0 @@
|
|||||||
package cmd
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bufio"
|
|
||||||
"fmt"
|
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
|
||||||
"path/filepath"
|
|
||||||
"strings"
|
|
||||||
|
|
||||||
"github.com/rancher/rke/cluster"
|
|
||||||
"github.com/rancher/rke/pki"
|
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
"github.com/urfave/cli"
|
|
||||||
"k8s.io/client-go/util/cert"
|
|
||||||
)
|
|
||||||
|
|
||||||
func ClusterCommand() cli.Command {
|
|
||||||
clusterUpDownUpgradeFlags := []cli.Flag{
|
|
||||||
cli.StringFlag{
|
|
||||||
Name: "cluster-file",
|
|
||||||
Usage: "Specify an alternate cluster YAML file",
|
|
||||||
Value: cluster.DefaultClusterConfig,
|
|
||||||
EnvVar: "CLUSTER_FILE",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
clusterForceRemove := []cli.Flag{
|
|
||||||
cli.BoolFlag{
|
|
||||||
Name: "force",
|
|
||||||
Usage: "Force removal of the cluster",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
return cli.Command{
|
|
||||||
Name: "cluster",
|
|
||||||
ShortName: "cluster",
|
|
||||||
Usage: "Operations on the cluster",
|
|
||||||
Flags: clusterUpDownUpgradeFlags,
|
|
||||||
Subcommands: []cli.Command{
|
|
||||||
cli.Command{
|
|
||||||
Name: "up",
|
|
||||||
Usage: "Bring the cluster up",
|
|
||||||
Action: clusterUpFromCli,
|
|
||||||
Flags: clusterUpDownUpgradeFlags,
|
|
||||||
},
|
|
||||||
cli.Command{
|
|
||||||
Name: "down",
|
|
||||||
Usage: "Teardown the cluster and clean cluster nodes",
|
|
||||||
Action: clusterDownFromCli,
|
|
||||||
Flags: append(clusterUpDownUpgradeFlags, clusterForceRemove...),
|
|
||||||
},
|
|
||||||
cli.Command{
|
|
||||||
Name: "version",
|
|
||||||
Usage: "Show Cluster Kubernetes version",
|
|
||||||
Action: getClusterVersion,
|
|
||||||
Flags: clusterUpDownUpgradeFlags,
|
|
||||||
},
|
|
||||||
cli.Command{
|
|
||||||
Name: "upgrade",
|
|
||||||
Usage: "Upgrade Cluster Kubernetes version",
|
|
||||||
Action: clusterUpgradeFromCli,
|
|
||||||
Flags: clusterUpDownUpgradeFlags,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func ClusterUp(clusterFile string) (string, string, string, string, error) {
|
|
||||||
logrus.Infof("Building Kubernetes cluster")
|
|
||||||
var APIURL, caCrt, clientCert, clientKey string
|
|
||||||
kubeCluster, err := cluster.ParseConfig(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.TunnelHosts()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
currentCluster, err := kubeCluster.GetClusterState()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := cluster.CheckEtcdHostsChanged(kubeCluster, currentCluster); err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = cluster.SetUpAuthentication(kubeCluster, currentCluster)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.SetUpHosts()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.DeployClusterPlanes()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = cluster.ReconcileCluster(kubeCluster, currentCluster)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.SaveClusterState(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.DeployNetworkPlugin()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.DeployK8sAddOns()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.DeployUserAddOns()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
APIURL = fmt.Sprintf("https://" + kubeCluster.ControlPlaneHosts[0].IP + ":6443")
|
|
||||||
caCrt = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.CACertName].Certificate))
|
|
||||||
clientCert = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Certificate))
|
|
||||||
clientKey = string(cert.EncodePrivateKeyPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Key))
|
|
||||||
|
|
||||||
logrus.Infof("Finished building Kubernetes cluster successfully")
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func ClusterUpgrade(clusterFile string) (string, string, string, string, error) {
|
|
||||||
logrus.Infof("Upgrading Kubernetes cluster")
|
|
||||||
var APIURL, caCrt, clientCert, clientKey string
|
|
||||||
kubeCluster, err := cluster.ParseConfig(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Debugf("Getting current cluster")
|
|
||||||
currentCluster, err := kubeCluster.GetClusterState()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if currentCluster == nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, fmt.Errorf("Failed to get the current state of Kubernetes cluster")
|
|
||||||
}
|
|
||||||
// check if user try to add/remove hosts during upgrade
|
|
||||||
if err := cluster.CheckHostsChangedOnUpgrade(kubeCluster, currentCluster); err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
/*
|
|
||||||
kubeCluster is the cluster.yaml definition. It should have updated configuration
|
|
||||||
currentCluster is the current state fetched from kubernetes
|
|
||||||
we add currentCluster certs to kubeCluster, kubeCluster would have the latest configuration from cluster.yaml and the certs to connect to k8s and apply the upgrade
|
|
||||||
*/
|
|
||||||
kubeCluster.Certificates = currentCluster.Certificates
|
|
||||||
logrus.Debugf("Setting up upgrade tunnels")
|
|
||||||
err = kubeCluster.TunnelHosts()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
logrus.Debugf("Starting cluster upgrade")
|
|
||||||
err = kubeCluster.ClusterUpgrade()
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.SaveClusterState(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Infof("Cluster upgraded successfully")
|
|
||||||
|
|
||||||
APIURL = fmt.Sprintf("https://" + kubeCluster.ControlPlaneHosts[0].IP + ":6443")
|
|
||||||
caCrt = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.CACertName].Certificate))
|
|
||||||
clientCert = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Certificate))
|
|
||||||
clientKey = string(cert.EncodePrivateKeyPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Key))
|
|
||||||
return APIURL, caCrt, clientCert, clientKey, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func ClusterDown(clusterFile string) error {
|
|
||||||
logrus.Infof("Tearing down Kubernetes cluster")
|
|
||||||
kubeCluster, err := cluster.ParseConfig(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = kubeCluster.TunnelHosts()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Debugf("Starting Cluster removal")
|
|
||||||
err = kubeCluster.ClusterDown()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Infof("Cluster removed successfully")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func clusterUpFromCli(ctx *cli.Context) error {
|
|
||||||
clusterFile, err := resolveClusterFile(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("Failed to resolve cluster file: %v", err)
|
|
||||||
}
|
|
||||||
_, _, _, _, err = ClusterUp(clusterFile)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func clusterUpgradeFromCli(ctx *cli.Context) error {
|
|
||||||
clusterFile, err := resolveClusterFile(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("Failed to resolve cluster file: %v", err)
|
|
||||||
}
|
|
||||||
_, _, _, _, err = ClusterUpgrade(clusterFile)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func clusterDownFromCli(ctx *cli.Context) error {
|
|
||||||
force := ctx.Bool("force")
|
|
||||||
if !force {
|
|
||||||
reader := bufio.NewReader(os.Stdin)
|
|
||||||
fmt.Printf("Are you sure you want to remove Kubernetes cluster [y/n]: ")
|
|
||||||
input, err := reader.ReadString('\n')
|
|
||||||
input = strings.TrimSpace(input)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if input != "y" && input != "Y" {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
clusterFile, err := resolveClusterFile(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("Failed to resolve cluster file: %v", err)
|
|
||||||
}
|
|
||||||
return ClusterDown(clusterFile)
|
|
||||||
}
|
|
||||||
|
|
||||||
func resolveClusterFile(ctx *cli.Context) (string, error) {
|
|
||||||
clusterFile := ctx.String("cluster-file")
|
|
||||||
fp, err := filepath.Abs(clusterFile)
|
|
||||||
if err != nil {
|
|
||||||
return "", fmt.Errorf("failed to lookup current directory name: %v", err)
|
|
||||||
}
|
|
||||||
file, err := os.Open(fp)
|
|
||||||
if err != nil {
|
|
||||||
return "", fmt.Errorf("Can not find cluster configuration file: %v", err)
|
|
||||||
}
|
|
||||||
defer file.Close()
|
|
||||||
buf, err := ioutil.ReadAll(file)
|
|
||||||
if err != nil {
|
|
||||||
return "", fmt.Errorf("failed to read file: %v", err)
|
|
||||||
}
|
|
||||||
clusterFileBuff := string(buf)
|
|
||||||
|
|
||||||
/*
|
|
||||||
This is a hacky way to add config path to cluster object without messing with
|
|
||||||
ClusterUp function and to avoid conflict with calls from kontainer-engine, basically
|
|
||||||
i add config path (cluster.yml by default) to a field into the config buffer
|
|
||||||
to be parsed later and added as ConfigPath field into cluster object.
|
|
||||||
*/
|
|
||||||
clusterFileBuff = fmt.Sprintf("%s\nconfig_path: %s\n", clusterFileBuff, clusterFile)
|
|
||||||
return clusterFileBuff, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func getClusterVersion(ctx *cli.Context) error {
|
|
||||||
localKubeConfig := cluster.GetLocalKubeConfig(ctx.String("cluster-file"))
|
|
||||||
serverVersion, err := cluster.GetK8sVersion(localKubeConfig)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
fmt.Printf("Server Version: %s\n", serverVersion)
|
|
||||||
return nil
|
|
||||||
}
|
|
37
cmd/common.go
Normal file
37
cmd/common.go
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
package cmd
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
|
||||||
|
"github.com/urfave/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
func resolveClusterFile(ctx *cli.Context) (string, error) {
|
||||||
|
clusterFile := ctx.String("config")
|
||||||
|
fp, err := filepath.Abs(clusterFile)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to lookup current directory name: %v", err)
|
||||||
|
}
|
||||||
|
file, err := os.Open(fp)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("Can not find cluster configuration file: %v", err)
|
||||||
|
}
|
||||||
|
defer file.Close()
|
||||||
|
buf, err := ioutil.ReadAll(file)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to read file: %v", err)
|
||||||
|
}
|
||||||
|
clusterFileBuff := string(buf)
|
||||||
|
|
||||||
|
/*
|
||||||
|
This is a hacky way to add config path to cluster object without messing with
|
||||||
|
ClusterUp function and to avoid conflict with calls from kontainer-engine, basically
|
||||||
|
i add config path (cluster.yml by default) to a field into the config buffer
|
||||||
|
to be parsed later and added as ConfigPath field into cluster object.
|
||||||
|
*/
|
||||||
|
clusterFileBuff = fmt.Sprintf("%s\nconfig_path: %s\n", clusterFileBuff, clusterFile)
|
||||||
|
return clusterFileBuff, nil
|
||||||
|
}
|
@ -179,7 +179,7 @@ func getHostConfig(reader *bufio.Reader, index int) (*v1.RKEConfigHost, error) {
|
|||||||
}
|
}
|
||||||
host.User = sshUser
|
host.User = sshUser
|
||||||
|
|
||||||
dockerSocketPath, err := getConfig(reader, fmt.Sprintf("Docker socker path on host (%s)", advertisedHostname), "/var/run/docker.sock")
|
dockerSocketPath, err := getConfig(reader, fmt.Sprintf("Docker socket path on host (%s)", advertisedHostname), "/var/run/docker.sock")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
76
cmd/remove.go
Normal file
76
cmd/remove.go
Normal file
@ -0,0 +1,76 @@
|
|||||||
|
package cmd
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/rancher/rke/cluster"
|
||||||
|
"github.com/sirupsen/logrus"
|
||||||
|
"github.com/urfave/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
func RemoveCommand() cli.Command {
|
||||||
|
removeFlags := []cli.Flag{
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "config",
|
||||||
|
Usage: "Specify an alternate cluster YAML file",
|
||||||
|
Value: cluster.DefaultClusterConfig,
|
||||||
|
EnvVar: "RKE_CONFIG",
|
||||||
|
},
|
||||||
|
cli.BoolFlag{
|
||||||
|
Name: "force",
|
||||||
|
Usage: "Force removal of the cluster",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
return cli.Command{
|
||||||
|
Name: "remove",
|
||||||
|
Usage: "Teardown the cluster and clean cluster nodes",
|
||||||
|
Action: clusterRemoveFromCli,
|
||||||
|
Flags: removeFlags,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func ClusterRemove(clusterFile string) error {
|
||||||
|
logrus.Infof("Tearing down Kubernetes cluster")
|
||||||
|
kubeCluster, err := cluster.ParseConfig(clusterFile)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.TunnelHosts()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
logrus.Debugf("Starting Cluster removal")
|
||||||
|
err = kubeCluster.ClusterRemove()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
logrus.Infof("Cluster removed successfully")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func clusterRemoveFromCli(ctx *cli.Context) error {
|
||||||
|
force := ctx.Bool("force")
|
||||||
|
if !force {
|
||||||
|
reader := bufio.NewReader(os.Stdin)
|
||||||
|
fmt.Printf("Are you sure you want to remove Kubernetes cluster [y/n]: ")
|
||||||
|
input, err := reader.ReadString('\n')
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if input != "y" && input != "Y" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
clusterFile, err := resolveClusterFile(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("Failed to resolve cluster file: %v", err)
|
||||||
|
}
|
||||||
|
return ClusterRemove(clusterFile)
|
||||||
|
}
|
108
cmd/up.go
Normal file
108
cmd/up.go
Normal file
@ -0,0 +1,108 @@
|
|||||||
|
package cmd
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/rancher/rke/cluster"
|
||||||
|
"github.com/rancher/rke/pki"
|
||||||
|
"github.com/sirupsen/logrus"
|
||||||
|
"github.com/urfave/cli"
|
||||||
|
"k8s.io/client-go/util/cert"
|
||||||
|
)
|
||||||
|
|
||||||
|
func UpCommand() cli.Command {
|
||||||
|
upFlags := []cli.Flag{
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "config",
|
||||||
|
Usage: "Specify an alternate cluster YAML file",
|
||||||
|
Value: cluster.DefaultClusterConfig,
|
||||||
|
EnvVar: "RKE_CONFIG",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
return cli.Command{
|
||||||
|
Name: "up",
|
||||||
|
Usage: "Bring the cluster up",
|
||||||
|
Action: clusterUpFromCli,
|
||||||
|
Flags: upFlags,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func ClusterUp(clusterFile string) (string, string, string, string, error) {
|
||||||
|
logrus.Infof("Building Kubernetes cluster")
|
||||||
|
var APIURL, caCrt, clientCert, clientKey string
|
||||||
|
kubeCluster, err := cluster.ParseConfig(clusterFile)
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.TunnelHosts()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
currentCluster, err := kubeCluster.GetClusterState()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := cluster.CheckEtcdHostsChanged(kubeCluster, currentCluster); err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = cluster.SetUpAuthentication(kubeCluster, currentCluster)
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.SetUpHosts()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.DeployClusterPlanes()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = cluster.ReconcileCluster(kubeCluster, currentCluster)
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.SaveClusterState(clusterFile)
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.DeployNetworkPlugin()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.DeployK8sAddOns()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
err = kubeCluster.DeployUserAddOns()
|
||||||
|
if err != nil {
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, err
|
||||||
|
}
|
||||||
|
|
||||||
|
APIURL = fmt.Sprintf("https://" + kubeCluster.ControlPlaneHosts[0].IP + ":6443")
|
||||||
|
caCrt = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.CACertName].Certificate))
|
||||||
|
clientCert = string(cert.EncodeCertPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Certificate))
|
||||||
|
clientKey = string(cert.EncodePrivateKeyPEM(kubeCluster.Certificates[pki.KubeAdminCommonName].Key))
|
||||||
|
|
||||||
|
logrus.Infof("Finished building Kubernetes cluster successfully")
|
||||||
|
return APIURL, caCrt, clientCert, clientKey, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func clusterUpFromCli(ctx *cli.Context) error {
|
||||||
|
clusterFile, err := resolveClusterFile(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("Failed to resolve cluster file: %v", err)
|
||||||
|
}
|
||||||
|
_, _, _, _, err = ClusterUp(clusterFile)
|
||||||
|
return err
|
||||||
|
}
|
35
cmd/version.go
Normal file
35
cmd/version.go
Normal file
@ -0,0 +1,35 @@
|
|||||||
|
package cmd
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/rancher/rke/cluster"
|
||||||
|
"github.com/urfave/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
func VersionCommand() cli.Command {
|
||||||
|
versionFlags := []cli.Flag{
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "config",
|
||||||
|
Usage: "Specify an alternate cluster YAML file",
|
||||||
|
Value: cluster.DefaultClusterConfig,
|
||||||
|
EnvVar: "RKE_CONFIG",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
return cli.Command{
|
||||||
|
Name: "version",
|
||||||
|
Usage: "Show cluster Kubernetes version",
|
||||||
|
Action: getClusterVersion,
|
||||||
|
Flags: versionFlags,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func getClusterVersion(ctx *cli.Context) error {
|
||||||
|
localKubeConfig := cluster.GetLocalKubeConfig(ctx.String("config"))
|
||||||
|
serverVersion, err := cluster.GetK8sVersion(localKubeConfig)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
fmt.Printf("Server Version: %s\n", serverVersion)
|
||||||
|
return nil
|
||||||
|
}
|
@ -20,9 +20,17 @@ func DoRunContainer(dClient *client.Client, imageCfg *container.Config, hostCfg
|
|||||||
}
|
}
|
||||||
if isRunning {
|
if isRunning {
|
||||||
logrus.Infof("[%s] Container %s is already running on host [%s]", plane, containerName, hostname)
|
logrus.Infof("[%s] Container %s is already running on host [%s]", plane, containerName, hostname)
|
||||||
|
isUpgradable, err := IsContainerUpgradable(dClient, imageCfg, containerName, hostname, plane)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if isUpgradable {
|
||||||
|
return DoRollingUpdateContainer(dClient, imageCfg, hostCfg, containerName, hostname, plane)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
logrus.Debugf("[%s] Pulling Image on host [%s]", plane, hostname)
|
|
||||||
|
logrus.Infof("[%s] Pulling Image on host [%s]", plane, hostname)
|
||||||
err = PullImage(dClient, hostname, imageCfg.Image)
|
err = PullImage(dClient, hostname, imageCfg.Image)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -50,6 +58,12 @@ func DoRollingUpdateContainer(dClient *client.Client, imageCfg *container.Config
|
|||||||
logrus.Infof("[%s] Container %s is not running on host [%s]", plane, containerName, hostname)
|
logrus.Infof("[%s] Container %s is not running on host [%s]", plane, containerName, hostname)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
logrus.Infof("[%s] Pulling Image on host [%s]", plane, hostname)
|
||||||
|
err = PullImage(dClient, hostname, imageCfg.Image)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
logrus.Infof("[%s] Successfully pulled %s image on host [%s]", plane, containerName, hostname)
|
||||||
logrus.Debugf("[%s] Stopping old container", plane)
|
logrus.Debugf("[%s] Stopping old container", plane)
|
||||||
oldContainerName := "old-" + containerName
|
oldContainerName := "old-" + containerName
|
||||||
if err := StopRenameContainer(dClient, hostname, containerName, oldContainerName); err != nil {
|
if err := StopRenameContainer(dClient, hostname, containerName, oldContainerName); err != nil {
|
||||||
@ -70,28 +84,28 @@ func DoRollingUpdateContainer(dClient *client.Client, imageCfg *container.Config
|
|||||||
}
|
}
|
||||||
|
|
||||||
func DoRemoveContainer(dClient *client.Client, containerName, hostname string) error {
|
func DoRemoveContainer(dClient *client.Client, containerName, hostname string) error {
|
||||||
logrus.Infof("[down/%s] Checking if container is running on host [%s]", containerName, hostname)
|
logrus.Infof("[remove/%s] Checking if container is running on host [%s]", containerName, hostname)
|
||||||
// not using the wrapper to check if the error is a NotFound error
|
// not using the wrapper to check if the error is a NotFound error
|
||||||
_, err := dClient.ContainerInspect(context.Background(), containerName)
|
_, err := dClient.ContainerInspect(context.Background(), containerName)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if client.IsErrNotFound(err) {
|
if client.IsErrNotFound(err) {
|
||||||
logrus.Infof("[down/%s] Container doesn't exist on host [%s]", containerName, hostname)
|
logrus.Infof("[remove/%s] Container doesn't exist on host [%s]", containerName, hostname)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
logrus.Infof("[down/%s] Stopping container on host [%s]", containerName, hostname)
|
logrus.Infof("[remove/%s] Stopping container on host [%s]", containerName, hostname)
|
||||||
err = StopContainer(dClient, hostname, containerName)
|
err = StopContainer(dClient, hostname, containerName)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
logrus.Infof("[down/%s] Removing container on host [%s]", containerName, hostname)
|
logrus.Infof("[remove/%s] Removing container on host [%s]", containerName, hostname)
|
||||||
err = RemoveContainer(dClient, hostname, containerName)
|
err = RemoveContainer(dClient, hostname, containerName)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
logrus.Infof("[down/%s] Sucessfully removed container on host [%s]", containerName, hostname)
|
logrus.Infof("[remove/%s] Sucessfully removed container on host [%s]", containerName, hostname)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -194,3 +208,19 @@ func WaitForContainer(dClient *client.Client, containerName string) error {
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func IsContainerUpgradable(dClient *client.Client, imageCfg *container.Config, containerName string, hostname string, plane string) (bool, error) {
|
||||||
|
logrus.Debugf("[%s] Checking if container %s is eligible for upgrade on host [%s]", plane, containerName, hostname)
|
||||||
|
// this should be moved to a higher layer.
|
||||||
|
|
||||||
|
containerInspect, err := InspectContainer(dClient, hostname, containerName)
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
if containerInspect.Config.Image == imageCfg.Image {
|
||||||
|
logrus.Debugf("[%s] Container %s is not eligible for updgrade on host [%s]", plane, containerName, hostname)
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
logrus.Debugf("[%s] Container %s is eligible for updgrade on host [%s]", plane, containerName, hostname)
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
4
main.go
4
main.go
@ -30,7 +30,9 @@ func mainErr() error {
|
|||||||
app.Author = "Rancher Labs, Inc."
|
app.Author = "Rancher Labs, Inc."
|
||||||
app.Email = ""
|
app.Email = ""
|
||||||
app.Commands = []cli.Command{
|
app.Commands = []cli.Command{
|
||||||
cmd.ClusterCommand(),
|
cmd.UpCommand(),
|
||||||
|
cmd.RemoveCommand(),
|
||||||
|
cmd.VersionCommand(),
|
||||||
cmd.ConfigCommand(),
|
cmd.ConfigCommand(),
|
||||||
}
|
}
|
||||||
app.Flags = []cli.Flag{
|
app.Flags = []cli.Flag{
|
||||||
|
@ -29,29 +29,6 @@ func RunControlPlane(controlHosts []hosts.Host, etcdHosts []hosts.Host, controlS
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func UpgradeControlPlane(controlHosts []hosts.Host, etcdHosts []hosts.Host, controlServices v1.RKEConfigServices) error {
|
|
||||||
logrus.Infof("[%s] Upgrading the Controller Plane..", ControlRole)
|
|
||||||
for _, host := range controlHosts {
|
|
||||||
// upgrade KubeAPI
|
|
||||||
if err := upgradeKubeAPI(host, etcdHosts, controlServices.KubeAPI); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// upgrade KubeController
|
|
||||||
if err := upgradeKubeController(host, controlServices.KubeController); err != nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// upgrade scheduler
|
|
||||||
err := upgradeScheduler(host, controlServices.Scheduler)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
logrus.Infof("[%s] Successfully upgraded Controller Plane..", ControlRole)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func RemoveControlPlane(controlHosts []hosts.Host) error {
|
func RemoveControlPlane(controlHosts []hosts.Host) error {
|
||||||
logrus.Infof("[%s] Tearing down the Controller Plane..", ControlRole)
|
logrus.Infof("[%s] Tearing down the Controller Plane..", ControlRole)
|
||||||
for _, host := range controlHosts {
|
for _, host := range controlHosts {
|
||||||
|
@ -9,7 +9,6 @@ import (
|
|||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/pki"
|
"github.com/rancher/rke/pki"
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func runKubeAPI(host hosts.Host, etcdHosts []hosts.Host, kubeAPIService v1.KubeAPIService) error {
|
func runKubeAPI(host hosts.Host, etcdHosts []hosts.Host, kubeAPIService v1.KubeAPIService) error {
|
||||||
@ -18,31 +17,6 @@ func runKubeAPI(host hosts.Host, etcdHosts []hosts.Host, kubeAPIService v1.KubeA
|
|||||||
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeAPIContainerName, host.AdvertisedHostname, ControlRole)
|
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeAPIContainerName, host.AdvertisedHostname, ControlRole)
|
||||||
}
|
}
|
||||||
|
|
||||||
func upgradeKubeAPI(host hosts.Host, etcdHosts []hosts.Host, kubeAPIService v1.KubeAPIService) error {
|
|
||||||
logrus.Debugf("[upgrade/KubeAPI] Checking for deployed version")
|
|
||||||
containerInspect, err := docker.InspectContainer(host.DClient, host.AdvertisedHostname, KubeAPIContainerName)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if containerInspect.Config.Image == kubeAPIService.Image {
|
|
||||||
logrus.Infof("[upgrade/KubeAPI] KubeAPI is already up to date")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/KubeAPI] Stopping old container")
|
|
||||||
oldContainerName := "old-" + KubeAPIContainerName
|
|
||||||
if err := docker.StopRenameContainer(host.DClient, host.AdvertisedHostname, KubeAPIContainerName, oldContainerName); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Container doesn't exist now!, lets deploy it!
|
|
||||||
logrus.Debugf("[upgrade/KubeAPI] Deploying new container")
|
|
||||||
if err := runKubeAPI(host, etcdHosts, kubeAPIService); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/KubeAPI] Removing old container")
|
|
||||||
err = docker.RemoveContainer(host.DClient, host.AdvertisedHostname, oldContainerName)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func removeKubeAPI(host hosts.Host) error {
|
func removeKubeAPI(host hosts.Host) error {
|
||||||
return docker.DoRemoveContainer(host.DClient, KubeAPIContainerName, host.AdvertisedHostname)
|
return docker.DoRemoveContainer(host.DClient, KubeAPIContainerName, host.AdvertisedHostname)
|
||||||
}
|
}
|
||||||
|
@ -8,7 +8,6 @@ import (
|
|||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/pki"
|
"github.com/rancher/rke/pki"
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func runKubeController(host hosts.Host, kubeControllerService v1.KubeControllerService) error {
|
func runKubeController(host hosts.Host, kubeControllerService v1.KubeControllerService) error {
|
||||||
@ -16,32 +15,6 @@ func runKubeController(host hosts.Host, kubeControllerService v1.KubeControllerS
|
|||||||
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeControllerContainerName, host.AdvertisedHostname, ControlRole)
|
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeControllerContainerName, host.AdvertisedHostname, ControlRole)
|
||||||
}
|
}
|
||||||
|
|
||||||
func upgradeKubeController(host hosts.Host, kubeControllerService v1.KubeControllerService) error {
|
|
||||||
logrus.Debugf("[upgrade/KubeController] Checking for deployed version")
|
|
||||||
containerInspect, err := docker.InspectContainer(host.DClient, host.AdvertisedHostname, KubeControllerContainerName)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if containerInspect.Config.Image == kubeControllerService.Image {
|
|
||||||
logrus.Infof("[upgrade/KubeController] KubeController is already up to date")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/KubeController] Stopping old container")
|
|
||||||
oldContainerName := "old-" + KubeControllerContainerName
|
|
||||||
if err := docker.StopRenameContainer(host.DClient, host.AdvertisedHostname, KubeControllerContainerName, oldContainerName); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Container doesn't exist now!, lets deploy it!
|
|
||||||
logrus.Debugf("[upgrade/KubeController] Deploying new container")
|
|
||||||
if err := runKubeController(host, kubeControllerService); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/KubeController] Removing old container")
|
|
||||||
err = docker.RemoveContainer(host.DClient, host.AdvertisedHostname, oldContainerName)
|
|
||||||
return err
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func removeKubeController(host hosts.Host) error {
|
func removeKubeController(host hosts.Host) error {
|
||||||
return docker.DoRemoveContainer(host.DClient, KubeControllerContainerName, host.AdvertisedHostname)
|
return docker.DoRemoveContainer(host.DClient, KubeControllerContainerName, host.AdvertisedHostname)
|
||||||
}
|
}
|
||||||
|
@ -9,7 +9,6 @@ import (
|
|||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/pki"
|
"github.com/rancher/rke/pki"
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func runKubelet(host hosts.Host, kubeletService v1.KubeletService, isMaster bool) error {
|
func runKubelet(host hosts.Host, kubeletService v1.KubeletService, isMaster bool) error {
|
||||||
@ -17,31 +16,6 @@ func runKubelet(host hosts.Host, kubeletService v1.KubeletService, isMaster bool
|
|||||||
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeletContainerName, host.AdvertisedHostname, WorkerRole)
|
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeletContainerName, host.AdvertisedHostname, WorkerRole)
|
||||||
}
|
}
|
||||||
|
|
||||||
func upgradeKubelet(host hosts.Host, kubeletService v1.KubeletService, isMaster bool) error {
|
|
||||||
logrus.Debugf("[upgrade/Kubelet] Checking for deployed version")
|
|
||||||
containerInspect, err := docker.InspectContainer(host.DClient, host.AdvertisedHostname, KubeletContainerName)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if containerInspect.Config.Image == kubeletService.Image {
|
|
||||||
logrus.Infof("[upgrade/Kubelet] Kubelet is already up to date")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Kubelet] Stopping old container")
|
|
||||||
oldContainerName := "old-" + KubeletContainerName
|
|
||||||
if err := docker.StopRenameContainer(host.DClient, host.AdvertisedHostname, KubeletContainerName, oldContainerName); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Container doesn't exist now!, lets deploy it!
|
|
||||||
logrus.Debugf("[upgrade/Kubelet] Deploying new container")
|
|
||||||
if err := runKubelet(host, kubeletService, isMaster); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Kubelet] Removing old container")
|
|
||||||
err = docker.RemoveContainer(host.DClient, host.AdvertisedHostname, oldContainerName)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func removeKubelet(host hosts.Host) error {
|
func removeKubelet(host hosts.Host) error {
|
||||||
return docker.DoRemoveContainer(host.DClient, KubeletContainerName, host.AdvertisedHostname)
|
return docker.DoRemoveContainer(host.DClient, KubeletContainerName, host.AdvertisedHostname)
|
||||||
}
|
}
|
||||||
|
@ -8,7 +8,6 @@ import (
|
|||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/pki"
|
"github.com/rancher/rke/pki"
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func runKubeproxy(host hosts.Host, kubeproxyService v1.KubeproxyService) error {
|
func runKubeproxy(host hosts.Host, kubeproxyService v1.KubeproxyService) error {
|
||||||
@ -16,31 +15,6 @@ func runKubeproxy(host hosts.Host, kubeproxyService v1.KubeproxyService) error {
|
|||||||
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeproxyContainerName, host.AdvertisedHostname, WorkerRole)
|
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, KubeproxyContainerName, host.AdvertisedHostname, WorkerRole)
|
||||||
}
|
}
|
||||||
|
|
||||||
func upgradeKubeproxy(host hosts.Host, kubeproxyService v1.KubeproxyService) error {
|
|
||||||
logrus.Debugf("[upgrade/Kubeproxy] Checking for deployed version")
|
|
||||||
containerInspect, err := docker.InspectContainer(host.DClient, host.AdvertisedHostname, KubeproxyContainerName)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if containerInspect.Config.Image == kubeproxyService.Image {
|
|
||||||
logrus.Infof("[upgrade/Kubeproxy] Kubeproxy is already up to date")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Kubeproxy] Stopping old container")
|
|
||||||
oldContainerName := "old-" + KubeproxyContainerName
|
|
||||||
if err := docker.StopRenameContainer(host.DClient, host.AdvertisedHostname, KubeproxyContainerName, oldContainerName); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Container doesn't exist now!, lets deploy it!
|
|
||||||
logrus.Debugf("[upgrade/Kubeproxy] Deploying new container")
|
|
||||||
if err := runKubeproxy(host, kubeproxyService); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Kubeproxy] Removing old container")
|
|
||||||
err = docker.RemoveContainer(host.DClient, host.AdvertisedHostname, oldContainerName)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func removeKubeproxy(host hosts.Host) error {
|
func removeKubeproxy(host hosts.Host) error {
|
||||||
return docker.DoRemoveContainer(host.DClient, KubeproxyContainerName, host.AdvertisedHostname)
|
return docker.DoRemoveContainer(host.DClient, KubeproxyContainerName, host.AdvertisedHostname)
|
||||||
}
|
}
|
||||||
|
@ -8,7 +8,6 @@ import (
|
|||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/pki"
|
"github.com/rancher/rke/pki"
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func runScheduler(host hosts.Host, schedulerService v1.SchedulerService) error {
|
func runScheduler(host hosts.Host, schedulerService v1.SchedulerService) error {
|
||||||
@ -16,31 +15,6 @@ func runScheduler(host hosts.Host, schedulerService v1.SchedulerService) error {
|
|||||||
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, SchedulerContainerName, host.AdvertisedHostname, ControlRole)
|
return docker.DoRunContainer(host.DClient, imageCfg, hostCfg, SchedulerContainerName, host.AdvertisedHostname, ControlRole)
|
||||||
}
|
}
|
||||||
|
|
||||||
func upgradeScheduler(host hosts.Host, schedulerService v1.SchedulerService) error {
|
|
||||||
logrus.Debugf("[upgrade/Scheduler] Checking for deployed version")
|
|
||||||
containerInspect, err := docker.InspectContainer(host.DClient, host.AdvertisedHostname, SchedulerContainerName)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if containerInspect.Config.Image == schedulerService.Image {
|
|
||||||
logrus.Infof("[upgrade/Scheduler] Scheduler is already up to date")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Scheduler] Stopping old container")
|
|
||||||
oldContainerName := "old-" + SchedulerContainerName
|
|
||||||
if err := docker.StopRenameContainer(host.DClient, host.AdvertisedHostname, SchedulerContainerName, oldContainerName); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Container doesn't exist now!, lets deploy it!
|
|
||||||
logrus.Debugf("[upgrade/Scheduler] Deploying new container")
|
|
||||||
if err := runScheduler(host, schedulerService); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
logrus.Debugf("[upgrade/Scheduler] Removing old container")
|
|
||||||
err = docker.RemoveContainer(host.DClient, host.AdvertisedHostname, oldContainerName)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func removeScheduler(host hosts.Host) error {
|
func removeScheduler(host hosts.Host) error {
|
||||||
return docker.DoRemoveContainer(host.DClient, SchedulerContainerName, host.AdvertisedHostname)
|
return docker.DoRemoveContainer(host.DClient, SchedulerContainerName, host.AdvertisedHostname)
|
||||||
}
|
}
|
||||||
|
@ -2,7 +2,6 @@ package services
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/rancher/rke/hosts"
|
"github.com/rancher/rke/hosts"
|
||||||
"github.com/rancher/rke/k8s"
|
|
||||||
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
"github.com/rancher/types/apis/cluster.cattle.io/v1"
|
||||||
"github.com/sirupsen/logrus"
|
"github.com/sirupsen/logrus"
|
||||||
)
|
)
|
||||||
@ -41,58 +40,6 @@ func RunWorkerPlane(controlHosts []hosts.Host, workerHosts []hosts.Host, workerS
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func UpgradeWorkerPlane(controlHosts []hosts.Host, workerHosts []hosts.Host, workerServices v1.RKEConfigServices, localConfigPath string) error {
|
|
||||||
logrus.Infof("[%s] Upgrading Worker Plane..", WorkerRole)
|
|
||||||
k8sClient, err := k8s.NewClient(localConfigPath)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
for _, host := range controlHosts {
|
|
||||||
// cordone the node
|
|
||||||
logrus.Debugf("[upgrade] Cordoning node: %s", host.AdvertisedHostname)
|
|
||||||
if err = k8s.CordonUncordon(k8sClient, host.AdvertisedHostname, true); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
err = upgradeKubelet(host, workerServices.Kubelet, true)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
err = upgradeKubeproxy(host, workerServices.Kubeproxy)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Debugf("[upgrade] Uncordoning node: %s", host.AdvertisedHostname)
|
|
||||||
if err = k8s.CordonUncordon(k8sClient, host.AdvertisedHostname, false); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for _, host := range workerHosts {
|
|
||||||
// cordone the node
|
|
||||||
logrus.Debugf("[upgrade] Cordoning node: %s", host.AdvertisedHostname)
|
|
||||||
if err = k8s.CordonUncordon(k8sClient, host.AdvertisedHostname, true); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// upgrade kubelet
|
|
||||||
err := upgradeKubelet(host, workerServices.Kubelet, false)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// upgrade kubeproxy
|
|
||||||
err = upgradeKubeproxy(host, workerServices.Kubeproxy)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Debugf("[upgrade] Uncordoning node: %s", host.AdvertisedHostname)
|
|
||||||
if err = k8s.CordonUncordon(k8sClient, host.AdvertisedHostname, false); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
logrus.Infof("[%s] Successfully upgraded Worker Plane..", WorkerRole)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func RemoveWorkerPlane(controlHosts []hosts.Host, workerHosts []hosts.Host) error {
|
func RemoveWorkerPlane(controlHosts []hosts.Host, workerHosts []hosts.Host) error {
|
||||||
logrus.Infof("[%s] Tearing down Worker Plane..", WorkerRole)
|
logrus.Infof("[%s] Tearing down Worker Plane..", WorkerRole)
|
||||||
for _, host := range controlHosts {
|
for _, host := range controlHosts {
|
||||||
|
Loading…
Reference in New Issue
Block a user