diff --git a/cluster/log-dump/log-dump.sh b/cluster/log-dump/log-dump.sh index 3874e05a540..ad1abe98c01 100755 --- a/cluster/log-dump/log-dump.sh +++ b/cluster/log-dump/log-dump.sh @@ -37,13 +37,13 @@ else readonly use_custom_instance_list= fi -readonly master_ssh_supported_providers="gce aws kubemark" -readonly node_ssh_supported_providers="gce gke aws kubemark" +readonly master_ssh_supported_providers="gce aws" +readonly node_ssh_supported_providers="gce gke aws" readonly master_logfiles="kube-apiserver kube-scheduler rescheduler kube-controller-manager etcd etcd-events glbc cluster-autoscaler kube-addon-manager fluentd" readonly node_logfiles="kube-proxy fluentd node-problem-detector" readonly node_systemd_services="node-problem-detector" -readonly hollow_node_logfiles="kubelet-hollow-node-* kubeproxy-hollow-node-* npd-*" +readonly hollow_node_logfiles="kubelet-hollow-node-* kubeproxy-hollow-node-* npd-hollow-node-*" readonly aws_logfiles="cloud-init-output" readonly gce_logfiles="startupscript" readonly kern_logfile="kern" @@ -109,7 +109,7 @@ function copy-logs-from-node() { scp -oLogLevel=quiet -oConnectTimeout=30 -oStrictHostKeyChecking=no -i "${LOG_DUMP_SSH_KEY}" "${LOG_DUMP_SSH_USER}@${node}:${scp_files}" "${dir}" > /dev/null || true else case "${KUBERNETES_PROVIDER}" in - gce|gke|kubemark) + gce|gke) # get-serial-port-output lets you ask for ports 1-4, but currently (11/21/2016) only port 1 contains useful information gcloud compute instances get-serial-port-output --project "${PROJECT}" --zone "${ZONE}" --port 1 "${node}" > "${dir}/serial-1.log" || true gcloud compute scp --recurse --project "${PROJECT}" --zone "${ZONE}" "${node}:${scp_files}" "${dir}" > /dev/null || true @@ -140,11 +140,8 @@ function save-logs() { fi else case "${KUBERNETES_PROVIDER}" in - gce|gke|kubemark) + gce|gke) files="${files} ${gce_logfiles}" - if [[ "${KUBERNETES_PROVIDER}" == "kubemark" && "${ENABLE_HOLLOW_NODE_LOGS:-}" == "true" ]]; then - files="${files} ${hollow_node_logfiles}" - fi ;; aws) files="${files} ${aws_logfiles}" @@ -240,6 +237,11 @@ function dump_nodes() { return fi + node_logfiles_all="${node_logfiles}" + if [[ "${ENABLE_HOLLOW_NODE_LOGS:-}" == "true" ]]; then + node_logfiles_all="${node_logfiles_all} ${hollow_node_logfiles}" + fi + nodes_selected_for_logs=() if [[ -n "${LOGDUMP_ONLY_N_RANDOM_NODES:-}" ]]; then # We randomly choose 'LOGDUMP_ONLY_N_RANDOM_NODES' many nodes for fetching logs. @@ -257,7 +259,7 @@ function dump_nodes() { mkdir -p "${node_dir}" # Save logs in the background. This speeds up things when there are # many nodes. - save-logs "${node_name}" "${node_dir}" "${node_logfiles}" "${node_systemd_services}" & + save-logs "${node_name}" "${node_dir}" "${node_logfiles_all}" "${node_systemd_services}" & # We don't want to run more than ${max_scp_processes} at a time, so # wait once we hit that many nodes. This isn't ideal, since one might