mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-07 09:31:35 +08:00
[benchmark] add max_workspace_size flags for tensorrt/pptrt backend (#2058)
* [benchmark] fixed paddlex benchmark for picodet 320 * [Bug Fix] fixed paddlex ppseg pp-trt infer error * [Bug Fix] fixed paddlex dino benchmark trt shapes * [benchmark] support paddlex ppyoloe pptrt benchmark * [benchmark] adjust paddlex dino trt shapes * [benchmark] add max_workspace_size flags for tensorrt/pptrt backend * [benchmark] add max_workspace_size flags for tensorrt/pptrt backend * [benchmark] add max_workspace_size flags for tensorrt/pptrt backend --------- Co-authored-by: qiuyanjun <qiuyanjun@baidu.com>
This commit is contained in:
@@ -34,6 +34,11 @@ static void UpdateBaseCustomFlags(
|
||||
if (FLAGS_xpu_l3_cache >= 0) {
|
||||
config_info["xpu_l3_cache"] = std::to_string(FLAGS_xpu_l3_cache);
|
||||
}
|
||||
if (FLAGS_max_workspace_size > 0) {
|
||||
config_info["max_workspace_size"] = std::to_string(FLAGS_max_workspace_size);
|
||||
} else {
|
||||
config_info["max_workspace_size"] = "-1";
|
||||
}
|
||||
// update custom options for paddle backend
|
||||
if (FLAGS_enable_log_info) {
|
||||
config_info["enable_log_info"] = "true";
|
||||
@@ -60,6 +65,11 @@ static bool CreateRuntimeOption(fastdeploy::RuntimeOption* option,
|
||||
if (config_info["enable_log_info"] == "true") {
|
||||
option->paddle_infer_option.enable_log_info = true;
|
||||
}
|
||||
if (config_info["max_workspace_size"] != "-1") {
|
||||
int max_workspace_size = std::stoi(config_info["max_workspace_size"]);
|
||||
option->paddle_infer_option.trt_option.max_workspace_size = max_workspace_size;
|
||||
option->trt_option.max_workspace_size = max_workspace_size;
|
||||
}
|
||||
if (config_info["device"] == "gpu") {
|
||||
option->UseGpu(std::stoi(config_info["device_id"]));
|
||||
if (config_info["backend"] == "ort") {
|
||||
|
Reference in New Issue
Block a user