-
Notifications
You must be signed in to change notification settings - Fork 42.8k
5k node scalability tests are running on limited set of nodes #67597
Copy link
Copy link
Closed
Labels
lifecycle/rottenDenotes an issue or PR that has aged beyond stale and will be auto-closed.Denotes an issue or PR that has aged beyond stale and will be auto-closed.priority/critical-urgentHighest priority. Must be actively worked on as someone's top priority right now.Highest priority. Must be actively worked on as someone's top priority right now.sig/scalabilityCategorizes an issue or PR as relevant to SIG Scalability.Categorizes an issue or PR as relevant to SIG Scalability.
Description
Some attempts are using decreased number of nodes in tests, e.g.:
- In (green) https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/192 I see "Scheduling additional 3883 Pods to measure startup latencies"
- In (red) https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/195 I see "Scheduling additional 2175 Pods to measure startup latencies"
I'm not sure about the impact of this: whether all the test is using lower number of nodes or just latency measurements, but this looks suspicious and requires investigation.
/assign
/cc @shyamjvs
/sig scalability
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
lifecycle/rottenDenotes an issue or PR that has aged beyond stale and will be auto-closed.Denotes an issue or PR that has aged beyond stale and will be auto-closed.priority/critical-urgentHighest priority. Must be actively worked on as someone's top priority right now.Highest priority. Must be actively worked on as someone's top priority right now.sig/scalabilityCategorizes an issue or PR as relevant to SIG Scalability.Categorizes an issue or PR as relevant to SIG Scalability.