Batch etcdlog intervals to prevent hundreds of thousands slamming browser charts#30763
Batch etcdlog intervals to prevent hundreds of thousands slamming browser charts#30763dgoodwin wants to merge 1 commit intoopenshift:mainfrom
Conversation
|
Pipeline controller notification For optional jobs, comment This repository is configured in: automatic mode |
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: dgoodwin The full list of commands accepted by this bot can be found here. The pull request process is described here DetailsNeeds approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
Scheduling required tests: |
|
@dgoodwin: The following tests failed, say
Full PR test history. Your PR dashboard. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. I understand the commands that are listed here. |
|
PR needs rebase. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
|
Risk analysis has seen new tests most likely introduced by this PR. New tests seen in this PR at sha: 5accfa4
|
Runs like https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-master-nightly-4.22-e2e-rosa-sts-ovn/2018612074661285888 have 200k etcdlog intervals. Turns out etcd can log the messages we watch for a LOT. These intervals are in memory on any prow job page load, and make interval charts brutally slow to load if they do at all.
This change batches them on minute boundaries, we'll see the message, the locator, and a count within that minute, but it cuts hundreds of thousands of intervals down to less than 400 in this case.
Assisted-by: Claude