mirror of https://github.com/k3s-io/k3s
Merge pull request #48204 from shyamjvs/logdump-only-n-nodes
Automatic merge from submit-queue Allow log-dumping only N randomly-chosen nodes in the cluster This should let us save "lots" (~3-4 hours) of time in our 5000-node cluster scale tests as we copy logs from all the nodes to jenkins worker and then upload all of them to gcs (while we don't need too many). This will also prevent the jenkins container facing "No space left on device" error while dumping logs, that we saw in runs 12-13 of gce-enormous-cluster. The longterm fix will be to enable [logexporter](https://github.com/kubernetes/test-infra/tree/master/logexporter) for our tests. cc @kubernetes/sig-scalability-misc @kubernetes/test-infra-maintainers @gmarek @fejtapull/6/head
commit
7018479968
|
@ -224,8 +224,19 @@ function dump_nodes() {
|
|||
return
|
||||
fi
|
||||
|
||||
nodes_selected_for_logs=()
|
||||
if [[ -n "${LOGDUMP_ONLY_N_RANDOM_NODES:-}" ]]; then
|
||||
# We randomly choose 'LOGDUMP_ONLY_N_RANDOM_NODES' many nodes for fetching logs.
|
||||
for index in `shuf -i 0-$(( ${#node_names[*]} - 1 )) -n ${LOGDUMP_ONLY_N_RANDOM_NODES}`
|
||||
do
|
||||
nodes_selected_for_logs+=("${node_names[$index]}")
|
||||
done
|
||||
else
|
||||
nodes_selected_for_logs=( "${node_names[@]}" )
|
||||
fi
|
||||
|
||||
proc=${max_scp_processes}
|
||||
for node_name in "${node_names[@]}"; do
|
||||
for node_name in "${nodes_selected_for_logs[@]}"; do
|
||||
node_dir="${report_dir}/${node_name}"
|
||||
mkdir -p "${node_dir}"
|
||||
# Save logs in the background. This speeds up things when there are
|
||||
|
|
Loading…
Reference in New Issue