mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-31 07:20:13 +00:00
Make logdump for kubemark logs independent of KUBERNETES_PROVIDER
This commit is contained in:
parent
92db97dfcc
commit
aac1837218
@ -37,13 +37,13 @@ else
|
|||||||
readonly use_custom_instance_list=
|
readonly use_custom_instance_list=
|
||||||
fi
|
fi
|
||||||
|
|
||||||
readonly master_ssh_supported_providers="gce aws kubemark"
|
readonly master_ssh_supported_providers="gce aws"
|
||||||
readonly node_ssh_supported_providers="gce gke aws kubemark"
|
readonly node_ssh_supported_providers="gce gke aws"
|
||||||
|
|
||||||
readonly master_logfiles="kube-apiserver kube-scheduler rescheduler kube-controller-manager etcd etcd-events glbc cluster-autoscaler kube-addon-manager fluentd"
|
readonly master_logfiles="kube-apiserver kube-scheduler rescheduler kube-controller-manager etcd etcd-events glbc cluster-autoscaler kube-addon-manager fluentd"
|
||||||
readonly node_logfiles="kube-proxy fluentd node-problem-detector"
|
readonly node_logfiles="kube-proxy fluentd node-problem-detector"
|
||||||
readonly node_systemd_services="node-problem-detector"
|
readonly node_systemd_services="node-problem-detector"
|
||||||
readonly hollow_node_logfiles="kubelet-hollow-node-* kubeproxy-hollow-node-* npd-*"
|
readonly hollow_node_logfiles="kubelet-hollow-node-* kubeproxy-hollow-node-* npd-hollow-node-*"
|
||||||
readonly aws_logfiles="cloud-init-output"
|
readonly aws_logfiles="cloud-init-output"
|
||||||
readonly gce_logfiles="startupscript"
|
readonly gce_logfiles="startupscript"
|
||||||
readonly kern_logfile="kern"
|
readonly kern_logfile="kern"
|
||||||
@ -109,7 +109,7 @@ function copy-logs-from-node() {
|
|||||||
scp -oLogLevel=quiet -oConnectTimeout=30 -oStrictHostKeyChecking=no -i "${LOG_DUMP_SSH_KEY}" "${LOG_DUMP_SSH_USER}@${node}:${scp_files}" "${dir}" > /dev/null || true
|
scp -oLogLevel=quiet -oConnectTimeout=30 -oStrictHostKeyChecking=no -i "${LOG_DUMP_SSH_KEY}" "${LOG_DUMP_SSH_USER}@${node}:${scp_files}" "${dir}" > /dev/null || true
|
||||||
else
|
else
|
||||||
case "${KUBERNETES_PROVIDER}" in
|
case "${KUBERNETES_PROVIDER}" in
|
||||||
gce|gke|kubemark)
|
gce|gke)
|
||||||
# get-serial-port-output lets you ask for ports 1-4, but currently (11/21/2016) only port 1 contains useful information
|
# get-serial-port-output lets you ask for ports 1-4, but currently (11/21/2016) only port 1 contains useful information
|
||||||
gcloud compute instances get-serial-port-output --project "${PROJECT}" --zone "${ZONE}" --port 1 "${node}" > "${dir}/serial-1.log" || true
|
gcloud compute instances get-serial-port-output --project "${PROJECT}" --zone "${ZONE}" --port 1 "${node}" > "${dir}/serial-1.log" || true
|
||||||
gcloud compute scp --recurse --project "${PROJECT}" --zone "${ZONE}" "${node}:${scp_files}" "${dir}" > /dev/null || true
|
gcloud compute scp --recurse --project "${PROJECT}" --zone "${ZONE}" "${node}:${scp_files}" "${dir}" > /dev/null || true
|
||||||
@ -140,11 +140,8 @@ function save-logs() {
|
|||||||
fi
|
fi
|
||||||
else
|
else
|
||||||
case "${KUBERNETES_PROVIDER}" in
|
case "${KUBERNETES_PROVIDER}" in
|
||||||
gce|gke|kubemark)
|
gce|gke)
|
||||||
files="${files} ${gce_logfiles}"
|
files="${files} ${gce_logfiles}"
|
||||||
if [[ "${KUBERNETES_PROVIDER}" == "kubemark" && "${ENABLE_HOLLOW_NODE_LOGS:-}" == "true" ]]; then
|
|
||||||
files="${files} ${hollow_node_logfiles}"
|
|
||||||
fi
|
|
||||||
;;
|
;;
|
||||||
aws)
|
aws)
|
||||||
files="${files} ${aws_logfiles}"
|
files="${files} ${aws_logfiles}"
|
||||||
@ -240,6 +237,11 @@ function dump_nodes() {
|
|||||||
return
|
return
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
node_logfiles_all="${node_logfiles}"
|
||||||
|
if [[ "${ENABLE_HOLLOW_NODE_LOGS:-}" == "true" ]]; then
|
||||||
|
node_logfiles_all="${node_logfiles_all} ${hollow_node_logfiles}"
|
||||||
|
fi
|
||||||
|
|
||||||
nodes_selected_for_logs=()
|
nodes_selected_for_logs=()
|
||||||
if [[ -n "${LOGDUMP_ONLY_N_RANDOM_NODES:-}" ]]; then
|
if [[ -n "${LOGDUMP_ONLY_N_RANDOM_NODES:-}" ]]; then
|
||||||
# We randomly choose 'LOGDUMP_ONLY_N_RANDOM_NODES' many nodes for fetching logs.
|
# We randomly choose 'LOGDUMP_ONLY_N_RANDOM_NODES' many nodes for fetching logs.
|
||||||
@ -257,7 +259,7 @@ function dump_nodes() {
|
|||||||
mkdir -p "${node_dir}"
|
mkdir -p "${node_dir}"
|
||||||
# Save logs in the background. This speeds up things when there are
|
# Save logs in the background. This speeds up things when there are
|
||||||
# many nodes.
|
# many nodes.
|
||||||
save-logs "${node_name}" "${node_dir}" "${node_logfiles}" "${node_systemd_services}" &
|
save-logs "${node_name}" "${node_dir}" "${node_logfiles_all}" "${node_systemd_services}" &
|
||||||
|
|
||||||
# We don't want to run more than ${max_scp_processes} at a time, so
|
# We don't want to run more than ${max_scp_processes} at a time, so
|
||||||
# wait once we hit that many nodes. This isn't ideal, since one might
|
# wait once we hit that many nodes. This isn't ideal, since one might
|
||||||
|
Loading…
Reference in New Issue
Block a user