-
Notifications
You must be signed in to change notification settings - Fork 25.6k
[ML] Dynamically get of num allocations for ml node models #115233
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Pinging @elastic/ml-core (Team:ML) |
Hi @davidkyle, I've created a changelog YAML for you. |
jan-elastic
reviewed
Oct 22, 2024
...ava/org/elasticsearch/xpack/inference/services/elasticsearch/ElasticsearchInternalModel.java
Outdated
Show resolved
Hide resolved
jan-elastic
reviewed
Oct 22, 2024
...a/org/elasticsearch/xpack/inference/services/elasticsearch/ElasticsearchInternalService.java
Show resolved
Hide resolved
jan-elastic
reviewed
Oct 22, 2024
...a/org/elasticsearch/xpack/inference/services/elasticsearch/ElasticsearchInternalService.java
Outdated
Show resolved
Hide resolved
jan-elastic
approved these changes
Oct 22, 2024
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
💔 Backport failed
You can use sqren/backport to manually backport by running |
jgowdyelastic
added a commit
to elastic/kibana
that referenced
this pull request
Oct 24, 2024
Enables the previously commented out check for `num_allocations` when listing the inference endpoints. The adaptive allocation count can drop to 0, but it is still valid for use. Uploading a file will cause it to be re-deployed. Related to es PRs elastic/elasticsearch#115233 and elastic/elasticsearch#115095 Follow on from #196577
kibanamachine
pushed a commit
to kibanamachine/kibana
that referenced
this pull request
Oct 24, 2024
Enables the previously commented out check for `num_allocations` when listing the inference endpoints. The adaptive allocation count can drop to 0, but it is still valid for use. Uploading a file will cause it to be re-deployed. Related to es PRs elastic/elasticsearch#115233 and elastic/elasticsearch#115095 Follow on from elastic#196577 (cherry picked from commit 66b2447)
kibanamachine
pushed a commit
to kibanamachine/kibana
that referenced
this pull request
Oct 24, 2024
Enables the previously commented out check for `num_allocations` when listing the inference endpoints. The adaptive allocation count can drop to 0, but it is still valid for use. Uploading a file will cause it to be re-deployed. Related to es PRs elastic/elasticsearch#115233 and elastic/elasticsearch#115095 Follow on from elastic#196577 (cherry picked from commit 66b2447)
georgewallace
pushed a commit
to georgewallace/elasticsearch
that referenced
this pull request
Oct 25, 2024
…15233) The GET inference API which should dynamically update the num_allocations field with the actual number from the deployed model which is useful when adaptive allocations are used
jfreden
pushed a commit
to jfreden/elasticsearch
that referenced
this pull request
Nov 4, 2024
…15233) The GET inference API which should dynamically update the num_allocations field with the actual number from the deployed model which is useful when adaptive allocations are used
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
auto-backport
Automatically create backport pull requests when merged
backport pending
>enhancement
:ml
Machine learning
Team:ML
Meta label for the ML team
v8.17.0
v9.0.0
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Reinstates the ability to dynamically report the number of allocations for ml node models in the Inference API.
Dynamically reading the number of allocations in use on GET is useful for the case where adaptive allocations is enabled and if the model deployment is managed through ml trained models. The
num_allocations
field inservice_settings
is updated with the current value.The change was originally reverted in 2697f85 due to an error calling the GroupedActionListener with 0 request. That bug is fixed in cce77b9.
This RP also includes bug fixes forward ported from 8.16 added in #115095