Skip to content

Latest commit

 

History

History
46 lines (35 loc) · 1.59 KB

compile-en.md

File metadata and controls

46 lines (35 loc) · 1.59 KB

English | 中文

FastDeploy Serving Deployment Image Compilation

This document is about how to create a FastDploy image.

GPU Image

The GPU images published by FastDploy are based on version 21.10 of Triton Inference Server. If developers need to use other CUDA versions, please refer to NVIDIA official website to modify the scripts in Dockerfile and scripts.

# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
bash scripts/build.sh

# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-gpu-cuda11.4-trt8.4-21.10 -f serving/Dockerfile .

CPU Image

# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
cd serving
bash scripts/build.sh OFF

# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-cpu-only-21.10 -f serving/Dockerfile_cpu .

IPU Image

# Enter the serving directory and execute the script to compile the FastDeploy and serving backend
cd serving
bash scripts/build_fd_ipu.sh

# Exit to the FastDeploy home directory and create the image
# x.y.z is FastDeploy version, example: 1.0.0
cd ../
docker build -t paddlepaddle/fastdeploy:x.y.z-ipu-only-21.10 -f serving/Dockerfile_ipu .