mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-06 09:07:10 +08:00
Rename PaddleBackend to PaddleInferBackend (#728)
This commit is contained in:
@@ -120,7 +120,7 @@ def create_paddle_inference_runtime(model_dir,
|
||||
use_fp16=False,
|
||||
device_id=0):
|
||||
option = fd.RuntimeOption()
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
if device_id == -1:
|
||||
option.use_cpu()
|
||||
else:
|
||||
|
@@ -24,7 +24,7 @@ int main(int argc, char* argv[]) {
|
||||
fd::RuntimeOption runtime_option;
|
||||
runtime_option.SetModelPath(model_file, params_file, fd::ModelFormat::PADDLE);
|
||||
// CPU
|
||||
runtime_option.UsePaddleBackend();
|
||||
runtime_option.UsePaddleInferBackend();
|
||||
runtime_option.SetCpuThreadNum(12);
|
||||
// GPU
|
||||
// runtime_option.UseGpu(0);
|
||||
|
@@ -26,7 +26,7 @@ option.set_model_path("mobilenetv2/inference.pdmodel",
|
||||
|
||||
# **** CPU 配置 ****
|
||||
option.use_cpu()
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
option.set_cpu_thread_num(12)
|
||||
|
||||
# **** GPU 配置 ***
|
||||
|
@@ -69,7 +69,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
if (FLAGS_backend == "onnx_runtime") {
|
||||
option->UseOrtBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "openvino") {
|
||||
option->UseOpenVINOBackend();
|
||||
} else if (FLAGS_backend == "tensorrt" ||
|
||||
|
@@ -99,7 +99,7 @@ class ErnieForSequenceClassificationPredictor(object):
|
||||
else:
|
||||
option.use_gpu()
|
||||
if args.backend == 'paddle':
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend == 'onnx_runtime':
|
||||
option.use_ort_backend()
|
||||
elif args.backend == 'openvino':
|
||||
|
@@ -50,7 +50,7 @@ int main(int argc, char* argv[]) {
|
||||
}
|
||||
switch (backend_type) {
|
||||
case 0:
|
||||
option.UsePaddleBackend();
|
||||
option.UsePaddleInferBackend();
|
||||
break;
|
||||
case 1:
|
||||
option.UseOrtBackend();
|
||||
|
@@ -57,7 +57,7 @@ def build_option(args):
|
||||
if args.backend == 'onnx_runtime':
|
||||
runtime_option.use_ort_backend()
|
||||
elif args.backend == 'paddle_inference':
|
||||
runtime_option.use_paddle_backend()
|
||||
runtime_option.use_paddle_infer_backend()
|
||||
elif args.backend == 'openvino':
|
||||
runtime_option.use_openvino_backend()
|
||||
runtime_option.set_cpu_thread_num(args.cpu_num_threads)
|
||||
|
@@ -78,7 +78,7 @@ class TritonPythonModel:
|
||||
|
||||
for option in options:
|
||||
if option['name'] == 'paddle':
|
||||
runtime_option.use_paddle_backend()
|
||||
runtime_option.use_paddle_infer_backend()
|
||||
elif option['name'] == 'onnxruntime':
|
||||
runtime_option.use_ort_backend()
|
||||
elif option['name'] == 'openvino':
|
||||
|
@@ -56,7 +56,7 @@ def build_option(args):
|
||||
elif args.backend.lower() == "ort":
|
||||
option.use_ort_backend()
|
||||
elif args.backend.lower() == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend.lower() == "openvino":
|
||||
assert args.device.lower(
|
||||
) == "cpu", "OpenVINO backend require inference on device CPU."
|
||||
|
@@ -78,7 +78,7 @@ int main(int argc, char* argv[]) {
|
||||
}
|
||||
else if (flag == 3) {
|
||||
option.UseCpu();
|
||||
option.UsePaddleBackend();
|
||||
option.UsePaddleInferBackend();
|
||||
}
|
||||
|
||||
std::string model_dir = argv[1];
|
||||
|
@@ -57,7 +57,7 @@ def build_option(args):
|
||||
elif args.backend.lower() == "ort":
|
||||
option.use_ort_backend()
|
||||
elif args.backend.lower() == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend.lower() == "openvino":
|
||||
assert args.device.lower(
|
||||
) == "cpu", "OpenVINO backend require inference on device CPU."
|
||||
|
@@ -40,7 +40,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
if (FLAGS_backend == "ort") {
|
||||
option->UseOrtBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "trt" ||
|
||||
FLAGS_backend == "paddle_trt") {
|
||||
option->UseTrtBackend();
|
||||
@@ -63,7 +63,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
} else if (FLAGS_backend == "ov") {
|
||||
option->UseOpenVINOBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "default") {
|
||||
return true;
|
||||
} else {
|
||||
|
@@ -38,7 +38,7 @@ def build_option(args):
|
||||
if backend == "ort":
|
||||
option.use_ort_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend in ["trt", "paddle_trt"]:
|
||||
option.use_trt_backend()
|
||||
option.set_trt_input_shape("input", [1, 3, 112, 112])
|
||||
@@ -58,7 +58,7 @@ def build_option(args):
|
||||
elif backend == "ov":
|
||||
option.use_openvino_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend == "default":
|
||||
return option
|
||||
else:
|
||||
|
@@ -40,7 +40,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
if (FLAGS_backend == "ort") {
|
||||
option->UseOrtBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "trt" ||
|
||||
FLAGS_backend == "paddle_trt") {
|
||||
option->UseTrtBackend();
|
||||
@@ -63,7 +63,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
} else if (FLAGS_backend == "ov") {
|
||||
option->UseOpenVINOBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "default") {
|
||||
return true;
|
||||
} else {
|
||||
|
@@ -37,7 +37,7 @@ def build_option(args):
|
||||
if backend == "ort":
|
||||
option.use_ort_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend in ["trt", "paddle_trt"]:
|
||||
option.use_trt_backend()
|
||||
option.set_trt_input_shape("input", [1, 3, 112, 112])
|
||||
@@ -57,7 +57,7 @@ def build_option(args):
|
||||
elif backend == "ov":
|
||||
option.use_openvino_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend == "default":
|
||||
return option
|
||||
else:
|
||||
|
@@ -41,7 +41,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
if (FLAGS_backend == "ort") {
|
||||
option->UseOrtBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "trt" ||
|
||||
FLAGS_backend == "paddle_trt") {
|
||||
option->UseTrtBackend();
|
||||
@@ -64,7 +64,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
} else if (FLAGS_backend == "ov") {
|
||||
option->UseOpenVINOBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "default") {
|
||||
return true;
|
||||
} else {
|
||||
|
@@ -42,7 +42,7 @@ def build_option(args):
|
||||
if backend == "ort":
|
||||
option.use_ort_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend in ["trt", "paddle_trt"]:
|
||||
option.use_trt_backend()
|
||||
option.set_trt_input_shape("input", [1, 3, 112, 112])
|
||||
@@ -62,7 +62,7 @@ def build_option(args):
|
||||
elif backend == "ov":
|
||||
option.use_openvino_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend == "default":
|
||||
return option
|
||||
else:
|
||||
|
@@ -40,7 +40,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
if (FLAGS_backend == "ort") {
|
||||
option->UseOrtBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "trt" ||
|
||||
FLAGS_backend == "paddle_trt") {
|
||||
option->UseTrtBackend();
|
||||
@@ -63,7 +63,7 @@ bool CreateRuntimeOption(fastdeploy::RuntimeOption* option) {
|
||||
} else if (FLAGS_backend == "ov") {
|
||||
option->UseOpenVINOBackend();
|
||||
} else if (FLAGS_backend == "paddle") {
|
||||
option->UsePaddleBackend();
|
||||
option->UsePaddleInferBackend();
|
||||
} else if (FLAGS_backend == "default") {
|
||||
return true;
|
||||
} else {
|
||||
|
@@ -37,7 +37,7 @@ def build_option(args):
|
||||
if backend == "ort":
|
||||
option.use_ort_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend in ["trt", "paddle_trt"]:
|
||||
option.use_trt_backend()
|
||||
option.set_trt_input_shape("input", [1, 3, 64, 64])
|
||||
@@ -57,7 +57,7 @@ def build_option(args):
|
||||
elif backend == "ov":
|
||||
option.use_openvino_backend()
|
||||
elif backend == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif backend == "default":
|
||||
return option
|
||||
else:
|
||||
|
@@ -60,7 +60,7 @@ void GpuInfer(const std::string& model_dir, const std::string& image_file,
|
||||
|
||||
auto option = fastdeploy::RuntimeOption();
|
||||
option.UseGpu();
|
||||
option.UsePaddleBackend();
|
||||
option.UsePaddleInferBackend();
|
||||
auto model = fastdeploy::vision::matting::PPMatting(model_file, params_file,
|
||||
config_file, option);
|
||||
if (!model.Initialized()) {
|
||||
|
@@ -34,7 +34,7 @@ def build_option(args):
|
||||
option = fd.RuntimeOption()
|
||||
if args.device.lower() == "gpu":
|
||||
option.use_gpu()
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
|
||||
if args.use_trt:
|
||||
option.use_trt_backend()
|
||||
|
@@ -81,7 +81,7 @@ def build_option(args):
|
||||
elif args.backend.lower() == "ort":
|
||||
option.use_ort_backend()
|
||||
elif args.backend.lower() == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend.lower() == "openvino":
|
||||
assert args.device.lower(
|
||||
) == "cpu", "OpenVINO backend require inference on device CPU."
|
||||
|
@@ -81,7 +81,7 @@ def build_option(args):
|
||||
elif args.backend.lower() == "ort":
|
||||
option.use_ort_backend()
|
||||
elif args.backend.lower() == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend.lower() == "openvino":
|
||||
assert args.device.lower(
|
||||
) == "cpu", "OpenVINO backend require inference on device CPU."
|
||||
|
@@ -57,7 +57,7 @@ void InitAndInfer(const std::string& model_dir, const std::string& image_file,
|
||||
|
||||
// fastdeploy::RuntimeOption option;
|
||||
// option.UseCpu();
|
||||
// option.UsePaddleBackend();
|
||||
// option.UsePaddleInferBackend();
|
||||
// std::cout<<"Xyy-debug, enable Paddle Backend==!";
|
||||
|
||||
// std::string model_dir = argv[1];
|
||||
@@ -89,7 +89,7 @@ int main(int argc, char* argv[]) {
|
||||
std::cout<<"Use ORT!"<<std::endl;
|
||||
} else if (flag == 1) {
|
||||
option.UseCpu();
|
||||
option.UsePaddleBackend();
|
||||
option.UsePaddleInferBackend();
|
||||
std::cout<<"Use PP!"<<std::endl;
|
||||
}
|
||||
|
||||
|
@@ -52,7 +52,7 @@ def build_option(args):
|
||||
elif args.backend.lower() == "ort":
|
||||
option.use_ort_backend()
|
||||
elif args.backend.lower() == "paddle":
|
||||
option.use_paddle_backend()
|
||||
option.use_paddle_infer_backend()
|
||||
elif args.backend.lower() == "openvino":
|
||||
assert args.device.lower(
|
||||
) == "cpu", "OpenVINO backend require inference on device CPU."
|
||||
|
Reference in New Issue
Block a user