-
Notifications
You must be signed in to change notification settings - Fork 186
Ollama connector blueprint #4160
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Carlos Rolo <carlos.rolo@netapp.com>
Signed-off-by: Carlos Rolo <carlos.rolo@netapp.com>
| ## 4. Register the model | ||
|
|
||
| ```json | ||
| POST /_plugins/_ml/models/_register |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
maybe you can do register and deploy is true, which save one step, POST /_plugins/_ml/models/_register?deploy=true
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I had that originally and then removed because IMO POST /_plugins/_ml/models/_register? Deploy=true gives a sense of automagic things happening. And with the explicit POST readers have to understand that there is this step (which you can automate) of deploying the connector.
After your comment about the tutorial, I think adding all the options in the tutorial to keep the blueprint cleaner would be a good one. But I'm open to exactly the opposite: making the blueprint very clean, even with such automations in place (and a note saying that they are there) and make the tutorial more dense with explanations about the deploying (and others) of models.
What option do you prefer?
PS: Just for reference, my first commit didn't have any deploy because I have auto deploy in my test cluster, that is something I would like to avoid for new readers.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it would be nice to have both options, so the blueprint is friendly for starter and also giving a good tip for advanced users who wants to speed up.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I will do that.
|
this looks great! You nailed it. Just added a very minor comment. It would be even nice if you can add the tutorial here using Ollma with the RAG processor, from the nice demo you built for conversation chat. |
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #4160 +/- ##
=========================================
Coverage 81.91% 81.91%
- Complexity 8874 8875 +1
=========================================
Files 762 762
Lines 38166 38166
Branches 4264 4264
=========================================
+ Hits 31263 31264 +1
+ Misses 5078 5076 -2
- Partials 1825 1826 +1
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
|
Ok let me add that too this PR then. |
| ```json | ||
| POST /_plugins/_ml/models/WWQI44MBbzI2oUKAvNUt/_deploy | ||
| { | ||
| "node_ids": ["4PLK7KJWReyX0oWKnBA8nA"] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
To use this, you need to enable plugins.ml_commons.allow_custom_deployment_plan setting.
To make the doc easier to follow, I think you can just use POST /_plugins/_ml/models/<MODEL_ID>/_deploy
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That is a mistake on my part, let me fix that.
|
LGTM, the CI failure are not related to your code changes, I am going to retrigger one more time. |
Description
This change documents the creation and testing of local LLMs within opensearch (Ollama for example)
Related Issues
Resolves #4146
Check List
--signoff.By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
For more information on following Developer Certificate of Origin and signing off your commits, please check here.