Skip to content
This repository has been archived by the owner on Oct 11, 2024. It is now read-only.

Commit

Permalink
[Doc]Add documentation to benchmarking script when running TGI (vllm-…
Browse files Browse the repository at this point in the history
  • Loading branch information
KuntaiDu authored and robertgshaw2-redhat committed Jul 14, 2024
1 parent ca46064 commit 38a41b9
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 1 deletion.
4 changes: 4 additions & 0 deletions benchmarks/benchmark_serving.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,10 @@
--dataset-path <path to dataset> \
--request-rate <request_rate> \ # By default <request_rate> is inf
--num-prompts <num_prompts> # By default <num_prompts> is 1000
when using tgi backend, add
--endpoint /generate_stream
to the end of the command above.
"""
import argparse
import asyncio
Expand Down
2 changes: 1 addition & 1 deletion benchmarks/launch_tgi_server.sh
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ PORT=8000
MODEL=$1
TOKENS=$2

docker run --gpus all --shm-size 1g -p $PORT:80 \
docker run -e HF_TOKEN=$HF_TOKEN --gpus all --shm-size 1g -p $PORT:80 \
-v $PWD/data:/data \
ghcr.io/huggingface/text-generation-inference:1.4.0 \
--model-id $MODEL \
Expand Down

0 comments on commit 38a41b9

Please sign in to comment.