mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-16 05:30:58 +08:00
Add paddle quantize model support for ORT, TRT and MKLDNN deploy backend (#257)
* add quantize model support for trt and paddle * fix bugs * fix * update paddle2onnx version * update version * add quantize test Co-authored-by: Jason <jiangjiajun@baidu.com>
This commit is contained in:
4
fastdeploy/runtime.h
Normal file → Executable file
4
fastdeploy/runtime.h
Normal file → Executable file
@@ -22,6 +22,7 @@
|
||||
|
||||
#include <map>
|
||||
#include <vector>
|
||||
#include <algorithm>
|
||||
|
||||
#include "fastdeploy/backends/backend.h"
|
||||
#include "fastdeploy/utils/perf.h"
|
||||
@@ -104,6 +105,9 @@ struct FASTDEPLOY_DECL RuntimeOption {
|
||||
*/
|
||||
void SetCpuThreadNum(int thread_num);
|
||||
|
||||
/// Use ORT graph opt level
|
||||
void SetOrtGraphOptLevel(int level = -1);
|
||||
|
||||
/// Set Paddle Inference as inference backend, support CPU/GPU
|
||||
void UsePaddleBackend();
|
||||
|
||||
|
Reference in New Issue
Block a user