Skip to content

Commit

Permalink
[Doc] add KubeAI to serving integrations (vllm-project#10837)
Browse files Browse the repository at this point in the history
Signed-off-by: Sam Stoelinga <sammiestoel@gmail.com>
  • Loading branch information
samos123 authored Dec 6, 2024
1 parent 8b59631 commit 7406274
Show file tree
Hide file tree
Showing 2 changed files with 18 additions and 0 deletions.
17 changes: 17 additions & 0 deletions docs/source/serving/deploying_with_kubeai.rst
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
.. _deploying_with_kubeai:

Deploying with KubeAI
=====================

`KubeAI <https://github.com/substratusai/kubeai>`_ is a Kubernetes operator that enables you to deploy and manage AI models on Kubernetes. It provides a simple and scalable way to deploy vLLM in production. Functionality such as scale-from-zero, load based autoscaling, model caching, and much more is provided out of the box with zero external dependencies.


Please see the Installation Guides for environment specific instructions:

* `Any Kubernetes Cluster <https://www.kubeai.org/installation/any/>`_
* `EKS <https://www.kubeai.org/installation/eks/>`_
* `GKE <https://www.kubeai.org/installation/gke/>`_

Once you have KubeAI installed, you can
`configure text generation models <https://www.kubeai.org/how-to/configure-text-generation-models/>`_
using vLLM.
1 change: 1 addition & 0 deletions docs/source/serving/integrations.rst
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ Integrations

run_on_sky
deploying_with_kserve
deploying_with_kubeai
deploying_with_triton
deploying_with_bentoml
deploying_with_cerebrium
Expand Down

0 comments on commit 7406274

Please sign in to comment.