Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

v1beta2 wishlist #768

Open
alculquicondor opened this issue May 12, 2023 · 17 comments
Open

v1beta2 wishlist #768

alculquicondor opened this issue May 12, 2023 · 17 comments
Labels
kind/cleanup Categorizes issue or PR as related to cleaning up code, process, or technical debt.

Comments

@alculquicondor
Copy link
Contributor

alculquicondor commented May 12, 2023

What would you like to be cleaned:

In a v1beta2 API, we would like to make the following breaking changes:

  • Workload
    • Report resource usage by pod, instead of the entire podset, in the admission struct.
    • Remove 'kueue.x-k8s.io/original-node-selectors' annotation
  • LocalQueue
    • Rename status.flavorUsage to status.flavorsUsage.

Why is this needed:

As we add more features, some field didn't age well.

@alculquicondor alculquicondor added the kind/cleanup Categorizes issue or PR as related to cleaning up code, process, or technical debt. label May 12, 2023
@tenzen-y
Copy link
Member

Remove 'kueue.x-k8s.io/original-node-selectors' annotation from workload

ref: #771

@lowang-bh
Copy link
Member

Has been removed in #834

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jan 30, 2024
@tenzen-y
Copy link
Member

/remove-lifecycle stale

@k8s-ci-robot k8s-ci-robot removed the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jan 30, 2024
@PBundyra
Copy link
Contributor

PBundyra commented Apr 5, 2024

Remove .spec.AdmissionChecks field in ClusterQueue and convert its content to .spec.AdmissionCheckStrategy
KEP: #1935
Issue: #1432

@IrvingMg
Copy link
Contributor

Change .spec.flavorFungibility, .spec.preemption, .spec.preemption.borrowWithinCohort type from pointer to value for ClusterQueueSpec.
Discussion: #1972 (comment)

@mimowo
Copy link
Contributor

mimowo commented May 22, 2024

Maybe we can consider to move the workload's spec.active to status.
Currently, the workload_controller needs write access to spec, which isn't recommended (see API conventions). Also, it makes tricky to set the reason for workload's deactivation reliably, see thread.

@alculquicondor
Copy link
Contributor Author

I'm not too sure about that one. active expresses a desired state, as such, it should be in the spec. It can come from the user or it can come from the system.

@tenzen-y
Copy link
Member

I'm not too sure about that one. active expresses a desired state, as such, it should be in the spec. It can come from the user or it can come from the system.

That is a good point.
In my point, I'm wondering if we should define the dedicated API for the kueue-controller-manager to deactivate the Workload in the status field.

@alculquicondor
Copy link
Contributor Author

Tricky... because we also want the users to be able to reset the active field back to true. And they definitely shouldn't have access to status.

@PBundyra
Copy link
Contributor

Deprecate the QueueVisibility API.
Issue: #2256

@tenzen-y
Copy link
Member

Tricky... because we also want the users to be able to reset the active field back to true. And they definitely shouldn't have access to status.

Yeah, I agree with you. My motivation is mitigating compexity of mechanism to deactivate workload by exceeding the backoffLimit in the workload controller.

Once we move or add active field in the Workload, that allows us to deactivate a workload by a single API call.

Current behavior needs to 3 reconciling: spec.active=false (workload controller) -> add Evicted condition to a workload (workload controller) -> stop Job (jobframework reconciler).
New .status.active needs to 2 reconciling: add Evicted condition to a workload (workload controller) -> stop Job (jobframework reconciler).

But, I'm sure your concerns. So, let us seek any alternative approach to mitigate complexity of mechanism of deactivation.

@PBundyra
Copy link
Contributor

PBundyra commented Jun 18, 2024

Deprecate or remove the retryDelayMinutes from the AdmissionCheck API
Issue: #2437

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Sep 16, 2024
@mimowo
Copy link
Contributor

mimowo commented Sep 16, 2024

/remove-lifecycle stale

@k8s-ci-robot k8s-ci-robot removed the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Sep 16, 2024
@mimowo
Copy link
Contributor

mimowo commented Oct 3, 2024

We may also revisit promoting domains for some labels (for example to support kueue.kubernetes.io/queue-name along with kueue.x-k8s.io/queue-name), see #2858

@tenzen-y
Copy link
Member

tenzen-y commented Oct 8, 2024

We may also revisit promoting domains for some labels (for example to support kueue.kubernetes.io/queue-name along with kueue.x-k8s.io/queue-name), see #2858

AFAIK, @kerthcet tried to add queueName field to the core PodSpec API. We may be able to retry it before adding the kueue.kuebernetes.io/queue-name. Although I'm not sure if the label is acceptable by core kube API.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/cleanup Categorizes issue or PR as related to cleaning up code, process, or technical debt.
Projects
None yet
Development

No branches or pull requests

8 participants