forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
[Hardware][Intel CPU][DOC] Update docs for CPU backend (vllm-project#…
…6212) Signed-off-by: Yuan Zhou <yuan.zhou@intel.com> Co-authored-by: Rafael Vasquez <rafvasq21@gmail.com> Co-authored-by: Gubrud, Aaron D <aaron.d.gubrud@intel.com> Co-authored-by: adgubrud <96072084+adgubrud@users.noreply.github.com> Signed-off-by: NickLucche <nlucches@redhat.com>
- Loading branch information
1 parent
4255cf3
commit cae449a
Showing
3 changed files
with
165 additions
and
1 deletion.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,142 @@ | ||
.. _nginxloadbalancer: | ||
|
||
Deploying with Nginx Loadbalancer | ||
================================= | ||
|
||
This document shows how to launch multiple vLLM serving containers and use Nginx to act as a load balancer between the servers. | ||
|
||
Table of contents: | ||
|
||
#. :ref:`Build Nginx Container <nginxloadbalancer_nginx_build>` | ||
#. :ref:`Create Simple Nginx Config file <nginxloadbalancer_nginx_conf>` | ||
#. :ref:`Build vLLM Container <nginxloadbalancer_nginx_vllm_container>` | ||
#. :ref:`Create Docker Network <nginxloadbalancer_nginx_docker_network>` | ||
#. :ref:`Launch vLLM Containers <nginxloadbalancer_nginx_launch_container>` | ||
#. :ref:`Launch Nginx <nginxloadbalancer_nginx_launch_nginx>` | ||
#. :ref:`Verify That vLLM Servers Are Ready <nginxloadbalancer_nginx_verify_nginx>` | ||
|
||
.. _nginxloadbalancer_nginx_build: | ||
|
||
Build Nginx Container | ||
--------------------- | ||
|
||
This guide assumes that you have just cloned the vLLM project and you're currently in the vllm root directory. | ||
|
||
.. code-block:: console | ||
export vllm_root=`pwd` | ||
Create a file named ``Dockerfile.nginx``: | ||
|
||
.. code-block:: console | ||
FROM nginx:latest | ||
RUN rm /etc/nginx/conf.d/default.conf | ||
EXPOSE 80 | ||
CMD ["nginx", "-g", "daemon off;"] | ||
Build the container: | ||
|
||
.. code-block:: console | ||
docker build . -f Dockerfile.nginx --tag nginx-lb | ||
.. _nginxloadbalancer_nginx_conf: | ||
|
||
Create Simple Nginx Config file | ||
------------------------------- | ||
|
||
Create a file named ``nginx_conf/nginx.conf``. Note that you can add as many servers as you'd like. In the below example we'll start with two. To add more, add another ``server vllmN:8000 max_fails=3 fail_timeout=10000s;`` entry to ``upstream backend``. | ||
|
||
.. code-block:: console | ||
upstream backend { | ||
least_conn; | ||
server vllm0:8000 max_fails=3 fail_timeout=10000s; | ||
server vllm1:8000 max_fails=3 fail_timeout=10000s; | ||
} | ||
server { | ||
listen 80; | ||
location / { | ||
proxy_pass http://backend; | ||
proxy_set_header Host $host; | ||
proxy_set_header X-Real-IP $remote_addr; | ||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; | ||
proxy_set_header X-Forwarded-Proto $scheme; | ||
} | ||
} | ||
.. _nginxloadbalancer_nginx_vllm_container: | ||
|
||
Build vLLM Container | ||
-------------------- | ||
|
||
.. code-block:: console | ||
cd $vllm_root | ||
docker build -f Dockerfile . --tag vllm | ||
If you are behind proxy, you can pass the proxy settings to the docker build command as shown below: | ||
|
||
.. code-block:: console | ||
cd $vllm_root | ||
docker build -f Dockerfile . --tag vllm --build-arg http_proxy=$http_proxy --build-arg https_proxy=$https_proxy | ||
.. _nginxloadbalancer_nginx_docker_network: | ||
|
||
Create Docker Network | ||
--------------------- | ||
|
||
.. code-block:: console | ||
docker network create vllm_nginx | ||
.. _nginxloadbalancer_nginx_launch_container: | ||
|
||
Launch vLLM Containers | ||
---------------------- | ||
|
||
Notes: | ||
|
||
* If you have your HuggingFace models cached somewhere else, update ``hf_cache_dir`` below. | ||
* If you don't have an existing HuggingFace cache you will want to start ``vllm0`` and wait for the model to complete downloading and the server to be ready. This will ensure that ``vllm1`` can leverage the model you just downloaded and it won't have to be downloaded again. | ||
* The below example assumes GPU backend used. If you are using CPU backend, remove ``--gpus all``, add ``VLLM_CPU_KVCACHE_SPACE`` and ``VLLM_CPU_OMP_THREADS_BIND`` environment variables to the docker run command. | ||
* Adjust the model name that you want to use in your vLLM servers if you don't want to use ``Llama-2-7b-chat-hf``. | ||
|
||
.. code-block:: console | ||
mkdir -p ~/.cache/huggingface/hub/ | ||
hf_cache_dir=~/.cache/huggingface/ | ||
docker run -itd --ipc host --privileged --network vllm_nginx --gpus all --shm-size=10.24gb -v $hf_cache_dir:/root/.cache/huggingface/ -p 8081:8000 --name vllm0 vllm --model meta-llama/Llama-2-7b-chat-hf | ||
docker run -itd --ipc host --privileged --network vllm_nginx --gpus all --shm-size=10.24gb -v $hf_cache_dir:/root/.cache/huggingface/ -p 8082:8000 --name vllm1 vllm --model meta-llama/Llama-2-7b-chat-hf | ||
.. note:: | ||
If you are behind proxy, you can pass the proxy settings to the docker run command via ``-e http_proxy=$http_proxy -e https_proxy=$https_proxy``. | ||
|
||
.. _nginxloadbalancer_nginx_launch_nginx: | ||
|
||
Launch Nginx | ||
------------ | ||
|
||
.. code-block:: console | ||
docker run -itd -p 8000:80 --network vllm_nginx -v ./nginx_conf/:/etc/nginx/conf.d/ --name nginx-lb nginx-lb:latest | ||
.. _nginxloadbalancer_nginx_verify_nginx: | ||
|
||
Verify That vLLM Servers Are Ready | ||
---------------------------------- | ||
|
||
.. code-block:: console | ||
docker logs vllm0 | grep Uvicorn | ||
docker logs vllm1 | grep Uvicorn | ||
Both outputs should look like this: | ||
|
||
.. code-block:: console | ||
INFO: Uvicorn running on http://0.0.0.0:8000 (Press CTRL+C to quit) |