Skip to content

Commit e79dc86

Browse files
authored
[Doc][Serving]serving doc update version to 1.0.0 (#755)
serving doc update version to 1.0.0
1 parent 8399092 commit e79dc86

File tree

10 files changed

+35
-25
lines changed

10 files changed

+35
-25
lines changed

examples/text/ernie-3.0/serving/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -51,10 +51,10 @@ models
5151
# GPU镜像
5252
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
5353
# CPU镜像
54-
docker pull paddlepaddle/fastdeploy:z.y.z-cpu-only-21.10
54+
docker pull paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10
5555

5656
# 运行
57-
docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v /path/serving/models:/models paddlepaddle/fastdeploy:0.6.0-cpu-only-21.10 bash
57+
docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v /path/serving/models:/models paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10 bash
5858
```
5959

6060
## 部署模型
@@ -67,7 +67,7 @@ token_cls_rpc_client.py # 序列标注任务发送pipeline预测请求的脚
6767
```
6868

6969
*注意*:启动服务时,Server的每个python后端进程默认申请`64M`内存,默认启动的docker无法启动多个python后端节点。有两个解决方案:
70-
- 1.启动容器时设置`shm-size`参数, 比如:`docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v /path/serving/models:/models paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash`
70+
- 1.启动容器时设置`shm-size`参数, 比如:`docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v /path/serving/models:/models paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash`
7171
- 2.启动服务时设置python后端的`shm-default-byte-size`参数, 设置python后端的默认内存为10M: `tritonserver --model-repository=/models --backend-config=python,shm-default-byte-size=10485760`
7272

7373
### 分类任务

examples/text/uie/serving/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,10 +34,10 @@ models
3434
# GPU镜像
3535
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
3636
# CPU镜像
37-
docker pull paddlepaddle/fastdeploy:z.y.z-cpu-only-21.10
37+
docker pull paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10
3838

3939
# 运行容器.容器名字为 fd_serving, 并挂载当前目录为容器的 /uie_serving 目录
40-
docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v `pwd`/:/uie_serving paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash
40+
docker run -it --net=host --name fastdeploy_server --shm-size="1g" -v `pwd`/:/uie_serving paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash
4141

4242
# 启动服务(不设置CUDA_VISIBLE_DEVICES环境变量,会拥有所有GPU卡的调度权限)
4343
CUDA_VISIBLE_DEVICES=0 fastdeployserver --model-repository=/uie_serving/models --backend-config=python,shm-default-byte-size=10485760

examples/vision/classification/paddleclas/serving/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,10 +28,10 @@ mv ResNet50_vd_infer/inference.pdiparams models/runtime/1/model.pdiparams
2828
# GPU镜像
2929
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
3030
# CPU镜像
31-
docker pull paddlepaddle/fastdeploy:z.y.z-cpu-only-21.10
31+
docker pull paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10
3232

3333
# 运行容器.容器名字为 fd_serving, 并挂载当前目录为容器的 /serving 目录
34-
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/serving paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash
34+
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/serving paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash
3535

3636
# 启动服务(不设置CUDA_VISIBLE_DEVICES环境变量,会拥有所有GPU卡的调度权限)
3737
CUDA_VISIBLE_DEVICES=0 fastdeployserver --model-repository=/serving/models --backend-config=python,shm-default-byte-size=10485760

examples/vision/detection/yolov5/serving/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,10 +22,10 @@ mv yolov5s.onnx models/runtime/1/model.onnx
2222
# GPU镜像
2323
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
2424
# CPU镜像
25-
docker pull paddlepaddle/fastdeploy:z.y.z-cpu-only-21.10
25+
docker pull paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10
2626

2727
# 运行容器.容器名字为 fd_serving, 并挂载当前目录为容器的 /yolov5_serving 目录
28-
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/yolov5_serving paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash
28+
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/yolov5_serving paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash
2929

3030
# 启动服务(不设置CUDA_VISIBLE_DEVICES环境变量,会拥有所有GPU卡的调度权限)
3131
CUDA_VISIBLE_DEVICES=0 fastdeployserver --model-repository=/yolov5_serving/models --backend-config=python,shm-default-byte-size=10485760

examples/vision/detection/yolov5/serving/README_EN.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ wget https://bj.bcebos.com/paddlehub/fastdeploy/yolov5s.onnx
99
# Save the model under models/infer/1 and rename it as model.onnx
1010
mv yolov5s.onnx models/infer/1/
1111

12-
# Pull fastdeploy image
13-
docker pull paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10
12+
# Pull fastdeploy image, x.y.z is FastDeploy version, example 1.0.0.
13+
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
1414

1515
# Run the docker. The docker name is fd_serving, and the current directory is mounted as the docker's /yolov5_serving directory
16-
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/yolov5_serving paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash
16+
nvidia-docker run -it --net=host --name fd_serving -v `pwd`/:/yolov5_serving paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash
1717

1818
# Start the service (Without setting the CUDA_VISIBLE_DEVICES environment variable, it will have scheduling privileges for all GPU cards)
1919
CUDA_VISIBLE_DEVICES=0 fastdeployserver --model-repository=models --backend-config=python,shm-default-byte-size=10485760

examples/vision/ocr/PP-OCRv3/serving/README.md

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,9 @@
11
# PP-OCR服务化部署示例
22

3+
在服务化部署前,需确认
4+
5+
- 1. 服务化镜像的软硬件环境要求和镜像拉取命令请参考[FastDeploy服务化部署](../../../../../serving/README_CN.md)
6+
37
## 介绍
48
本文介绍了使用FastDeploy搭建OCR文字识别服务的方法.
59

@@ -48,9 +52,9 @@ mv ppocr_keys_v1.txt models/rec_postprocess/1/
4852

4953
wget https://gitee.com/paddlepaddle/PaddleOCR/raw/release/2.6/doc/imgs/12.jpg
5054

51-
52-
docker pull paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10
53-
docker run -dit --net=host --name fastdeploy --shm-size="1g" -v $PWD:/ocr_serving paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 bash
55+
# x.y.z为镜像版本号,需参照serving文档替换为数字
56+
docker pull paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10
57+
docker run -dit --net=host --name fastdeploy --shm-size="1g" -v $PWD:/ocr_serving paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 bash
5458
docker exec -it -u root fastdeploy bash
5559
```
5660

serving/README_CN.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,13 +17,13 @@ FastDeploy基于[Triton Inference Server](https://github.com/triton-inference-se
1717
#### CPU镜像
1818
CPU镜像仅支持Paddle/ONNX模型在CPU上进行服务化部署,支持的推理后端包括OpenVINO、Paddle Inference和ONNX Runtime
1919
``` shell
20-
docker pull paddlepaddle/fastdeploy:0.6.0-cpu-only-21.10
20+
docker pull paddlepaddle/fastdeploy:1.0.0-cpu-only-21.10
2121
```
2222

2323
#### GPU镜像
2424
GPU镜像支持Paddle/ONNX模型在GPU/CPU上进行服务化部署,支持的推理后端包括OpenVINO、TensorRT、Paddle Inference和ONNX Runtime
2525
```
26-
docker pull paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10
26+
docker pull paddlepaddle/fastdeploy:1.0.0-gpu-cuda11.4-trt8.4-21.10
2727
```
2828

2929
用户也可根据自身需求,参考如下文档自行编译镜像

serving/README_EN.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,15 +20,15 @@ FastDeploy builds an end-to-end serving deployment based on [Triton Inference Se
2020
CPU images only support Paddle/ONNX models for serving deployment on CPUs, and supported inference backends include OpenVINO, Paddle Inference, and ONNX Runtime
2121

2222
```shell
23-
docker pull paddlepaddle/fastdeploy:0.6.0-cpu-only-21.10
23+
docker pull paddlepaddle/fastdeploy:1.0.0-cpu-only-21.10
2424
```
2525

2626
#### GPU Image
2727

2828
GPU images support Paddle/ONNX models for serving deployment on GPU and CPU, and supported inference backends including OpenVINO, TensorRT, Paddle Inference, and ONNX Runtime
2929

3030
```
31-
docker pull paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10
31+
docker pull paddlepaddle/fastdeploy:1.0.0-gpu-cuda11.4-trt8.4-21.10
3232
```
3333

3434
Users can also compile the image by themselves according to their own needs, referring to the following documents:

serving/docs/EN/compile-en.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,8 +12,9 @@ cd serving
1212
bash scripts/build.sh
1313

1414
# Exit to the FastDeploy home directory and create the image
15+
# x.y.z is FastDeploy version, example: 1.0.0
1516
cd ../
16-
docker build -t paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .
17+
docker build -t paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .
1718
```
1819

1920
## CPU Image
@@ -25,8 +26,9 @@ cd serving
2526
bash scripts/build.sh OFF
2627

2728
# Exit to the FastDeploy home directory and create the image
29+
# x.y.z is FastDeploy version, example: 1.0.0
2830
cd ../
29-
docker build -t paddlepaddle/fastdeploy:0.6.0-cpu-only-21.10 -f serving/Dockerfile_cpu .
31+
docker build -t paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10 -f serving/Dockerfile_cpu .
3032
```
3133

3234
## IPU Image
@@ -37,6 +39,7 @@ cd serving
3739
bash scripts/build_fd_ipu.sh
3840

3941
# Exit to the FastDeploy home directory and create the image
42+
# x.y.z is FastDeploy version, example: 1.0.0
4043
cd ../
41-
docker build -t paddlepaddle/fastdeploy:0.6.0-ipu-only-21.10 -f serving/Dockerfile_ipu .
44+
docker build -t paddlepaddle/fastdeploy:x.y.z-ipu-only-21.10 -f serving/Dockerfile_ipu .
4245
```

serving/docs/zh_CN/compile.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,8 +12,9 @@ cd serving
1212
bash scripts/build.sh
1313
1414
# 退出到FastDeploy主目录,制作镜像
15+
# x.y.z为FastDeploy版本号,可根据情况自己确定。比如: 1.0.0
1516
cd ../
16-
docker build -t paddlepaddle/fastdeploy:0.6.0-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .
17+
docker build -t paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .
1718
```
1819

1920
## 制作CPU镜像
@@ -24,8 +25,9 @@ cd serving
2425
bash scripts/build.sh OFF
2526
2627
# 退出到FastDeploy主目录,制作镜像
28+
# x.y.z为FastDeploy版本号,可根据情况自己确定。比如: 1.0.0
2729
cd ../
28-
docker build -t paddlepaddle/fastdeploy:0.6.0-cpu-only-21.10 -f serving/Dockerfile_cpu .
30+
docker build -t paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10 -f serving/Dockerfile_cpu .
2931
```
3032

3133
## 制作IPU镜像
@@ -36,6 +38,7 @@ cd serving
3638
bash scripts/build_fd_ipu.sh
3739
3840
# 退出到FastDeploy主目录,制作镜像
41+
# x.y.z为FastDeploy版本号,可根据情况自己确定。比如: 1.0.0
3942
cd ../
40-
docker build -t paddlepaddle/fastdeploy:0.6.0-ipu-only-21.10 -f serving/Dockerfile_ipu .
43+
docker build -t paddlepaddle/fastdeploy:x.y.z-ipu-only-21.10 -f serving/Dockerfile_ipu .
4144
```

0 commit comments

Comments
 (0)