@@ -23,7 +23,7 @@ func DefaultClusterChecks() []ClusterCheck {
2323 func (cluster ClusterInfo ) conditions.Condition {
2424 return conditions .PollingCondition ("all k8s nodes to report ready" , func (ctx context.Context ) error {
2525 return K8sAllNodesReadyAssertion (ctx , cluster )
26- }, 10 * time . Minute , 5 * time .Second )
26+ }, 5 * time .Second )
2727 },
2828
2929 // wait for kube-proxy to report ready
@@ -39,7 +39,7 @@ func DefaultClusterChecks() []ClusterCheck {
3939 }
4040
4141 return K8sPodReadyAssertion (ctx , cluster , replicas , "kube-system" , "k8s-app=kube-proxy" )
42- }, 5 * time .Minute , 5 * time . Second )
42+ }, 5 * time .Second )
4343 },
4444
4545 // wait for coredns to report ready
@@ -55,14 +55,14 @@ func DefaultClusterChecks() []ClusterCheck {
5555 }
5656
5757 return K8sPodReadyAssertion (ctx , cluster , replicas , "kube-system" , "k8s-app=kube-dns" )
58- }, 5 * time .Minute , 5 * time . Second )
58+ }, 5 * time .Second )
5959 },
6060
6161 // wait for all the nodes to be schedulable
6262 func (cluster ClusterInfo ) conditions.Condition {
6363 return conditions .PollingCondition ("all k8s nodes to report schedulable" , func (ctx context.Context ) error {
6464 return K8sAllNodesSchedulableAssertion (ctx , cluster )
65- }, 5 * time .Minute , 5 * time . Second )
65+ }, 5 * time .Second )
6666 },
6767 },
6868 )
@@ -77,21 +77,21 @@ func K8sComponentsReadinessChecks() []ClusterCheck {
7777 func (cluster ClusterInfo ) conditions.Condition {
7878 return conditions .PollingCondition ("all k8s nodes to report" , func (ctx context.Context ) error {
7979 return K8sAllNodesReportedAssertion (ctx , cluster )
80- }, 5 * time . Minute , 30 * time .Second ) // give more time per each attempt, as this check is going to build and cache kubeconfig
80+ }, 30 * time .Second ) // give more time per each attempt, as this check is going to build and cache kubeconfig
8181 },
8282
8383 // wait for k8s control plane static pods
8484 func (cluster ClusterInfo ) conditions.Condition {
8585 return conditions .PollingCondition ("all control plane static pods to be running" , func (ctx context.Context ) error {
8686 return K8sControlPlaneStaticPods (ctx , cluster )
87- }, 5 * time .Minute , 5 * time . Second )
87+ }, 5 * time .Second )
8888 },
8989
9090 // wait for HA k8s control plane
9191 func (cluster ClusterInfo ) conditions.Condition {
9292 return conditions .PollingCondition ("all control plane components to be ready" , func (ctx context.Context ) error {
9393 return K8sFullControlPlaneAssertion (ctx , cluster )
94- }, 5 * time .Minute , 5 * time . Second )
94+ }, 5 * time .Second )
9595 },
9696 }
9797}
@@ -110,63 +110,63 @@ func PreBootSequenceChecks() []ClusterCheck {
110110 func (cluster ClusterInfo ) conditions.Condition {
111111 return conditions .PollingCondition ("etcd to be healthy" , func (ctx context.Context ) error {
112112 return ServiceHealthAssertion (ctx , cluster , "etcd" , WithNodeTypes (machine .TypeInit , machine .TypeControlPlane ))
113- }, 5 * time .Minute , 5 * time . Second )
113+ }, 5 * time .Second )
114114 },
115115
116116 // wait for etcd members to be consistent across nodes
117117 func (cluster ClusterInfo ) conditions.Condition {
118118 return conditions .PollingCondition ("etcd members to be consistent across nodes" , func (ctx context.Context ) error {
119119 return EtcdConsistentAssertion (ctx , cluster )
120- }, 5 * time .Minute , 5 * time . Second )
120+ }, 5 * time .Second )
121121 },
122122
123123 // wait for etcd members to be the control plane nodes
124124 func (cluster ClusterInfo ) conditions.Condition {
125125 return conditions .PollingCondition ("etcd members to be control plane nodes" , func (ctx context.Context ) error {
126126 return EtcdControlPlaneNodesAssertion (ctx , cluster )
127- }, 5 * time .Minute , 5 * time . Second )
127+ }, 5 * time .Second )
128128 },
129129
130130 // wait for apid to be ready on all the nodes
131131 func (cluster ClusterInfo ) conditions.Condition {
132132 return conditions .PollingCondition ("apid to be ready" , func (ctx context.Context ) error {
133133 return ApidReadyAssertion (ctx , cluster )
134- }, 5 * time .Minute , 5 * time . Second )
134+ }, 5 * time .Second )
135135 },
136136
137137 // wait for all nodes to report their memory size
138138 func (cluster ClusterInfo ) conditions.Condition {
139139 return conditions .PollingCondition ("all nodes memory sizes" , func (ctx context.Context ) error {
140140 return AllNodesMemorySizes (ctx , cluster )
141- }, 5 * time .Minute , 5 * time . Second )
141+ }, 5 * time .Second )
142142 },
143143
144144 // wait for all nodes to report their disk size
145145 func (cluster ClusterInfo ) conditions.Condition {
146146 return conditions .PollingCondition ("all nodes disk sizes" , func (ctx context.Context ) error {
147147 return AllNodesDiskSizes (ctx , cluster )
148- }, 5 * time .Minute , 5 * time . Second )
148+ }, 5 * time .Second )
149149 },
150150
151151 // check diagnostics
152152 func (cluster ClusterInfo ) conditions.Condition {
153153 return conditions .PollingCondition ("no diagnostics" , func (ctx context.Context ) error {
154154 return NoDiagnostics (ctx , cluster )
155- }, time . Minute , 5 * time .Second )
155+ }, 5 * time .Second )
156156 },
157157
158158 // wait for kubelet to be healthy on all
159159 func (cluster ClusterInfo ) conditions.Condition {
160160 return conditions .PollingCondition ("kubelet to be healthy" , func (ctx context.Context ) error {
161161 return ServiceHealthAssertion (ctx , cluster , "kubelet" , WithNodeTypes (machine .TypeInit , machine .TypeControlPlane ))
162- }, 5 * time .Minute , 5 * time . Second )
162+ }, 5 * time .Second )
163163 },
164164
165165 // wait for all nodes to finish booting
166166 func (cluster ClusterInfo ) conditions.Condition {
167167 return conditions .PollingCondition ("all nodes to finish boot sequence" , func (ctx context.Context ) error {
168168 return AllNodesBootedAssertion (ctx , cluster )
169- }, 5 * time .Minute , 5 * time . Second )
169+ }, 5 * time .Second )
170170 },
171171 }
172172}
0 commit comments