-
Notifications
You must be signed in to change notification settings - Fork 896
fix arm64 nodepool for kind #8156
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: upodroid The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
~ update in-place
+/- create replacement and then destroy
Terraform will perform the following actions:
# google_vmwareengine_network_peering.gvce_peering will be updated in-place
~ resource "google_vmwareengine_network_peering" "gvce_peering" {
~ export_custom_routes_with_public_ip = false -> true
id = "projects/k8s-infra-prow-build/locations/global/networkPeerings/peer-with-gcve-project"
~ import_custom_routes_with_public_ip = false -> true
name = "peer-with-gcve-project"
# (13 unchanged attributes hidden)
}
# module.prow_build_nodepool_c4d_highmem_8_localssd.google_container_node_pool.node_pool must be replaced
+/- resource "google_container_node_pool" "node_pool" {
~ id = "projects/k8s-infra-prow-build/locations/us-central1/clusters/prow-build/nodePools/pool7-20250528124554315100000001" -> (known after apply)
~ initial_node_count = 5 -> 1 # forces replacement
~ instance_group_urls = [
- "https://www.googleapis.com/compute/v1/projects/k8s-infra-prow-build/zones/us-central1-a/instanceGroupManagers/gke-prow-build-pool7-2025052812455431-48d85bd4-grp",
- "https://www.googleapis.com/compute/v1/projects/k8s-infra-prow-build/zones/us-central1-b/instanceGroupManagers/gke-prow-build-pool7-2025052812455431-f2ab4558-grp",
- "https://www.googleapis.com/compute/v1/projects/k8s-infra-prow-build/zones/us-central1-c/instanceGroupManagers/gke-prow-build-pool7-2025052812455431-2e4cd1ad-grp",
] -> (known after apply)
~ managed_instance_group_urls = [
- "https://www.googleapis.com/compute/beta/projects/k8s-infra-prow-build/zones/us-central1-a/instanceGroups/gke-prow-build-pool7-2025052812455431-48d85bd4-grp",
- "https://www.googleapis.com/compute/beta/projects/k8s-infra-prow-build/zones/us-central1-b/instanceGroups/gke-prow-build-pool7-2025052812455431-f2ab4558-grp",
- "https://www.googleapis.com/compute/beta/projects/k8s-infra-prow-build/zones/us-central1-c/instanceGroups/gke-prow-build-pool7-2025052812455431-2e4cd1ad-grp",
] -> (known after apply)
+ max_pods_per_node = (known after apply)
~ name = "pool7-20250528124554315100000001" -> (known after apply)
~ node_count = 5 -> (known after apply)
+ operation = (known after apply)
~ version = "1.32.3-gke.1927009" -> (known after apply)
# (5 unchanged attributes hidden)
~ autoscaling {
~ location_policy = "BALANCED" -> (known after apply)
~ min_node_count = 5 -> 10
- total_max_node_count = 0 -> null
- total_min_node_count = 0 -> null
# (1 unchanged attribute hidden)
}
~ network_config (known after apply)
~ node_config {
~ effective_taints = [] -> (known after apply)
- enable_confidential_storage = false -> null
~ labels = {} -> (known after apply)
~ local_ssd_count = 0 -> (known after apply)
~ logging_variant = "DEFAULT" -> (known after apply)
+ min_cpu_platform = (known after apply)
- resource_labels = {
- "goog-gke-node-pool-provisioning-model" = "on-demand"
} -> null
- resource_manager_tags = {} -> null
- storage_pools = [] -> null
- tags = [] -> null
# (13 unchanged attributes hidden)
~ confidential_nodes (known after apply)
- ephemeral_storage_local_ssd_config {
- local_ssd_count = 1 -> null
}
~ gcfs_config (known after apply)
~ guest_accelerator (known after apply)
~ kubelet_config (known after apply)
- kubelet_config {
- allowed_unsafe_sysctls = [] -> null
- container_log_max_files = 0 -> null
- cpu_cfs_quota = false -> null
- image_gc_high_threshold_percent = 0 -> null
- image_gc_low_threshold_percent = 0 -> null
- insecure_kubelet_readonly_port_enabled = "TRUE" -> null
- pod_pids_limit = 0 -> null
# (5 unchanged attributes hidden)
}
~ shielded_instance_config (known after apply)
- shielded_instance_config {
- enable_integrity_monitoring = true -> null
- enable_secure_boot = false -> null
}
~ windows_node_config (known after apply)
- windows_node_config {
# (1 unchanged attribute hidden)
}
# (1 unchanged block hidden)
}
~ upgrade_settings (known after apply)
- upgrade_settings {
- max_surge = 1 -> null
- max_unavailable = 0 -> null
- strategy = "SURGE" -> null
}
# (1 unchanged block hidden)
}
Plan: 1 to add, 1 to change, 1 to destroy.
Plan: 1 to add, 1 to change, 1 to destroy.
|
atlantis plan |
Ran Plan for dir: Plan Error Show Output
|
atlantis plan |
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
~ update in-place
Terraform will perform the following actions:
# google_vmwareengine_network_peering.gvce_peering will be updated in-place
~ resource "google_vmwareengine_network_peering" "gvce_peering" {
~ export_custom_routes_with_public_ip = false -> true
id = "projects/k8s-infra-prow-build/locations/global/networkPeerings/peer-with-gcve-project"
~ import_custom_routes_with_public_ip = false -> true
name = "peer-with-gcve-project"
# (13 unchanged attributes hidden)
}
# module.prow_build_nodepool_c4d_highmem_8_localssd.google_container_node_pool.node_pool will be updated in-place
~ resource "google_container_node_pool" "node_pool" {
id = "projects/k8s-infra-prow-build/locations/us-central1/clusters/prow-build/nodePools/pool7-20250528124554315100000001"
name = "pool7-20250528124554315100000001"
# (10 unchanged attributes hidden)
~ autoscaling {
~ min_node_count = 5 -> 10
# (4 unchanged attributes hidden)
}
# (3 unchanged blocks hidden)
}
Plan: 0 to add, 2 to change, 0 to destroy.
Plan: 0 to add, 2 to change, 0 to destroy.
|
/lgtm |
Locks and plans deleted for the projects and workspaces modified in this pull request:
|
thanks |
/cc @ameukam @aojea @BenTheElder
The kernel bug appeared in the kind jobs on arm64 nodes so I changed their os too.
c4d is in preview so I'm forcing the nodepool to always run atleast 10 nodes of this type per zone. Looking at the nodepool size metrics, it has consistently stayed above 20 per zone for weeks. The autoscaler seems to be preferring c4 pool over c4d pool, which isn't ideal.