-
Notifications
You must be signed in to change notification settings - Fork 6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[serve.llm] remove backoff dependency #50822
Conversation
Signed-off-by: Kourosh Hakhamaneshi <kourosh@anyscale.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Some small suggestions. Also would be nice to include a test. Let me know what you think
python/ray/llm/_internal/serve/deployments/llm/multiplex/lora_model_loader.py
Outdated
Show resolved
Hide resolved
python/ray/llm/_internal/serve/deployments/llm/multiplex/lora_model_loader.py
Outdated
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great tests 💯
def test_retry_success_first_try(): | ||
"""Test that the function works normally when no exceptions occur.""" | ||
# Create a mock function that doesn't raise exceptions | ||
mock_fn = Mock(return_value="success") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Non-blocker: Would prefer not to just those mock/ patch when can. It's really simply to just write a function to do exactly return success and such 🙃
It's me and my friend :) |
Remove backoff from ray.serve.llm dependency.