Files
FastDeploy/serving/docs/EN/compile-en.md
heliqi e79dc863e4 [Doc][Serving]serving doc update version to 1.0.0 (#755)
serving doc update version to 1.0.0
2022-11-30 14:41:42 +08:00

46 lines
1.5 KiB
Markdown

# FastDeploy Serving Deployment Image Compilation
How to create a FastDploy image
## GPU Image
The GPU images published by FastDploy are based on version 21.10 of [Triton Inference Server](https://github.com/triton-inference-server/server). If developers need to use other CUDA versions, please refer to [ NVIDIA official website](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html) to modify the scripts in Dockerfile and scripts.
```shell
# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
bash scripts/build.sh
# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .
```
## CPU Image
```shell
# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
cd serving
bash scripts/build.sh OFF
# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10 -f serving/Dockerfile_cpu .
```
## IPU Image
```shell
# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
bash scripts/build_fd_ipu.sh
# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-ipu-only-21.10 -f serving/Dockerfile_ipu .
```