DefTruth
99c2b6592d
[Backend] refactor paddle custom ops -> fastdeploy::paddle_custom_ops ( #2101 )
...
* [cmake] upgrade windows paddle inference -> 2.5.0
* [cmake] upgrade windows paddle inference -> 2.5.0
* fix paddle custom ops bug on windows
* [Backend] refactor paddle custom ops
2023-07-13 09:00:03 +08:00
DefTruth
b2426aefa9
[Backend] add paddle custom ops compatible policy ( #2070 )
...
* Add centerpoint
* fix postprocess op file name
* [Backend] add paddle custom ops compatible policy
* [Backend] add paddle custom ops compatible policy
* [Backend] add paddle custom ops compatible policy
* upgrade linx paddle gpu -> 2.5
* add custom op compatible policy for paddle 2.5
* add custom op compatible policy for paddle 2.5
* add custom op compatible policy for paddle 2.5
* add collect_trt_shape_by_device option for paddle backend
* add collect_trt_shape_by_device option for paddle backend
* add custom op option for python build
* fix python build bugs
* update paddle linux x86 cpu only lib
* update paddle linux gpu lib
* update patchelf cmake
* fix paddle backend option pybind
* update paddle_inference.cmake
* add cuda sm_80 support (A100)
---------
Co-authored-by: zengshao0622 <peter_z96@163.com >
Co-authored-by: qiuyanjun <qiuyanjun@baidu.com >
2023-06-29 22:32:14 +08:00
Wang Bojun
7b58387d3c
add option for inference precision and cutlass ( #2017 )
...
* add option for inference precision and cutlass
* code clean
* code refine
* code refine
2023-06-09 10:20:48 +08:00
DefTruth
284b1b46c8
[Serving] Support XPU encrypt & auth server ( #2007 )
...
* [patchelf] fix patchelf error for inference xpu
* [serving] add xpu dockerfile and support fd server
* [serving] add xpu dockerfile and support fd server
* [Serving] support XPU + Tritron
* [Serving] support XPU + Tritron
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] add comments for xpu tritron dockerfile
* [Doruntime] fix xpu infer error
* [Doruntime] fix xpu infer error
* [XPU] update xpu dockerfile
* add xpu triton server docs
* add xpu triton server docs
* add xpu triton server docs
* add xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* [XPU] Update XPU L3 Cache setting docs
* [XPU] Add Encryption and AUTH support for XPU Server
* [XPU] Add Encryption and AUTH support for XPU Server
* [Bug Fix] fix paddle reader error
* [Serving] Support XPU encrypt & auth server
* [Serving] Support XPU encrypt & auth server
* [Serving] Support XPU encrypt & auth server
* [Serving] Support XPU encrypt & auth server
* [Triton] switch TAG 22.12 -> TAG 21.10wq
* update xpu auth server script
2023-06-01 21:36:44 +08:00
DefTruth
387c5695b3
[XPU] Update XPU L3 Cache setting docs ( #2001 )
...
* [patchelf] fix patchelf error for inference xpu
* [serving] add xpu dockerfile and support fd server
* [serving] add xpu dockerfile and support fd server
* [Serving] support XPU + Tritron
* [Serving] support XPU + Tritron
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] update xpu tritron docker file -> paddle 0.0.0
* [Dockerfile] add comments for xpu tritron dockerfile
* [Doruntime] fix xpu infer error
* [Doruntime] fix xpu infer error
* [XPU] update xpu dockerfile
* add xpu triton server docs
* add xpu triton server docs
* add xpu triton server docs
* add xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* update xpu triton server docs
* [XPU] Update XPU L3 Cache setting docs
2023-05-30 11:21:04 +08:00
DefTruth
49c033a828
[XPU] Support XPU via Paddle Inference backend ( #1987 )
...
* [backend] Support XPU via Paddle Inference backend
* [backend] Support XPU via Paddle Inference backend
* [backend] Support XPU via Paddle Inference backend
* [XPU] support XPU benchmark via paddle inference
* [XPU] support XPU benchmark via paddle inference
* [benchmark] add xpu paddle h2d config files
2023-05-25 14:13:40 +08:00
DefTruth
652024d2f6
Revert "Remove Paddle Reader" ( #1860 )
...
Revert "Remove Paddle Reader (#1813 )"
This reverts commit f3d44785c4
.
2023-04-23 23:16:31 +08:00
Jason
f3d44785c4
Remove Paddle Reader ( #1813 )
...
* Remove Paddle Reader
* support pp-infer c++14
* disable trt cache
---------
Co-authored-by: wang-xinyu <wangxinyu_es@163.com >
2023-04-20 21:12:43 +08:00
WJJ1995
5c70db176f
[Backend]Add switch_ir_debug for Paddle Backend ( #1700 )
...
* avoid mem copy for cpp benchmark
* set CMAKE_BUILD_TYPE to Release
* Add SegmentationDiff
* change pointer to reference
* fixed bug
* cast uint8 to int32
* Add diff compare for OCR
* Add diff compare for OCR
* rm ppocr pipeline
* Add yolov5 diff compare
* Add yolov5 diff compare
* deal with comments
* deal with comments
* fixed bug
* fixed bug
* fixed thread nums
* Add Failed log
* optimizer x86 pipeline
* Add switch_ir_debug for paddle backend
* fixed for ci
2023-03-24 17:29:31 +08:00
Jason
6be2c0367b
[Example] Update runtime examples ( #1542 )
...
* Add notes for tensors
* Optimize some apis
* move some warnings
2023-03-08 16:56:04 +08:00
Jack Zhou
524c85745b
[Backend] Add fixed size optimization for transformer model ( #1430 )
...
Add enable_fixed_size_opt flag
2023-02-24 09:45:04 +08:00
Jason
dd527388bc
Update option.h
2023-02-22 11:28:04 +08:00
WJJ1995
b6658b8094
[Benchmark] Add ppocr && ppseg benchmark ( #1344 )
...
* add GPL lisence
* add GPL-3.0 lisence
* add GPL-3.0 lisence
* add GPL-3.0 lisence
* support yolov8
* add pybind for yolov8
* add yolov8 readme
* add cpp benchmark
* add cpu and gpu mem
* public part split
* add runtime mode
* fixed bugs
* add cpu_thread_nums
* deal with comments
* deal with comments
* deal with comments
* rm useless code
* add FASTDEPLOY_DECL
* add FASTDEPLOY_DECL
* fixed for windows
* mv rss to pss
* mv rss to pss
* Update utils.cc
* use thread to collect mem
* Add ResourceUsageMonitor
* rm useless code
* fixed bug
* fixed typo
* update ResourceUsageMonitor
* fixed bug
* fixed bug
* add note for ResourceUsageMonitor
* deal with comments
* add macros
* deal with comments
* deal with comments
* deal with comments
* re-lint
* rm pmap and use mem api
* rm pmap and use mem api
* add mem api
* Add PrintBenchmarkInfo func
* Add PrintBenchmarkInfo func
* Add PrintBenchmarkInfo func
* deal with comments
* fixed enable_paddle_to_trt
* add log for paddle_trt
* support ppcls benchmark
* use new trt option api
* update benchmark info
* simplify benchmark.cc
* simplify benchmark.cc
* deal with comments
* Add ppseg && ppocr benchmark
* add OCR rec img
* add ocr benchmark
* fixed trt shape
* add trt shape
* resolve conflict
* add ENABLE_BENCHMARK define
---------
Co-authored-by: DefTruth <31974251+DefTruth@users.noreply.github.com >
2023-02-20 15:59:27 +08:00
Jason
b8afb0d040
[Other] Move comments for deprecated functions ( #1275 )
...
Move comments for deprecated functions
2023-02-09 10:04:18 +08:00
Jason
a4b0565b9a
[Other] Optimize paddle backend ( #1265 )
...
* Optimize paddle backend
* optimize paddle backend
* add version support
2023-02-08 19:12:03 +08:00
huangjianhui
76df90afc3
[Other] FastDeploy TensorRT && ONNX backend support to load model form memory ( #1130 )
...
* Update all backends load model from buffer
* Delete redundant code
* Format code style
* Format code style
* Delete redundant code
* Delete redundant code
* Add some FDASSERTs
* Update load model form memory when cloning engine
* Update clone engine code
* Update set_model_buffer api parameters with char pointer
* Release memory buffer variables after finish init backends
* Fix conflict
* Fix bug
2023-02-01 11:36:09 +08:00
Jason
4aa4ebd7c3
[Other] [Part2] Upgrade runtime module ( #1080 )
...
[Other] Upgrade runtime module
2023-01-09 13:22:51 +08:00