1.6 KiB
How to Build FastDeploy Library on Nvidia Jetson Platform
FastDeploy supports CPU inference with ONNX Runtime and GPU inference with Nvidia TensorRT on Nvidia Jetson platform
How to Build and Install FastDeploy C++ Library
Prerequisite for Compiling on NVIDIA Jetson:
- gcc/g++ >= 5.4 (8.2 is recommended)
- cmake >= 3.10.0
- jetpack >= 4.6.1
git clone https://github.com/PaddlePaddle/FastDeploy.git
cd FastDeploy
mkdir build && cd build
cmake .. -DBUILD_ON_JETSON=ON \
-DENABLE_VISION=ON \
-DCMAKE_INSTALL_PREFIX=${PWD}/installed_fastdeploy
make -j8
make install
Once compiled, the C++ inference library is generated in the directory specified by CMAKE_INSTALL_PREFIX
How to Build and Install FastDeploy Python Library
Prerequisite for Compiling on NVIDIA Jetson:
- gcc/g++ >= 5.4 (8.2 is recommended)
- cmake >= 3.10.0
- jetpack >= 4.6.1
- python >= 3.6
Notice the wheel
is required if you need to pack a wheel, execute pip install wheel
first.
All compilation options are imported via environment variables
git clone https://github.com/PaddlePaddle/FastDeploy.git
cd FastDeploy/python
export BUILD_ON_JETSON=ON
export ENABLE_VISION=ON
python setup.py build
python setup.py bdist_wheel
The compiled wheel
package will be generated in the FastDeploy/python/dist
directory once finished. Users can pip-install it directly.
During the compilation, if developers want to change the compilation parameters, it is advisable to delete the build
and .setuptools-cmake-build
subdirectories in the FastDeploy/python
to avoid the possible impact from cache, and then recompile.