mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-06 00:57:33 +08:00
[Backend] Add OCR、Seg、 KeypointDetection、Matting、 ernie-3.0 and adaface models for XPU Deploy (#960)
* [FlyCV] Bump up FlyCV -> official release 1.0.0 * add seg models for XPU * add ocr model for XPU * add matting * add matting python * fix infer.cc * add keypointdetection support for XPU * Add adaface support for XPU * add ernie-3.0 * fix doc Co-authored-by: DefTruth <qiustudent_r@163.com> Co-authored-by: DefTruth <31974251+DefTruth@users.noreply.github.com>
This commit is contained in:
2
examples/vision/ocr/PP-OCRv3/python/README.md
Normal file → Executable file
2
examples/vision/ocr/PP-OCRv3/python/README.md
Normal file → Executable file
@@ -33,6 +33,8 @@ python infer.py --det_model ch_PP-OCRv3_det_infer --cls_model ch_ppocr_mobile_v2
|
||||
python infer.py --det_model ch_PP-OCRv3_det_infer --cls_model ch_ppocr_mobile_v2.0_cls_infer --rec_model ch_PP-OCRv3_rec_infer --rec_label_file ppocr_keys_v1.txt --image 12.jpg --device gpu
|
||||
# GPU上使用TensorRT推理
|
||||
python infer.py --det_model ch_PP-OCRv3_det_infer --cls_model ch_ppocr_mobile_v2.0_cls_infer --rec_model ch_PP-OCRv3_rec_infer --rec_label_file ppocr_keys_v1.txt --image 12.jpg --device gpu --backend trt
|
||||
# XPU推理
|
||||
python infer.py --det_model ch_PP-OCRv3_det_infer --cls_model ch_ppocr_mobile_v2.0_cls_infer --rec_model ch_PP-OCRv3_rec_infer --rec_label_file ppocr_keys_v1.txt --image 12.jpg --device xpu
|
||||
```
|
||||
|
||||
运行完成可视化结果如下图所示
|
||||
|
6
examples/vision/ocr/PP-OCRv3/python/infer.py
Normal file → Executable file
6
examples/vision/ocr/PP-OCRv3/python/infer.py
Normal file → Executable file
@@ -41,7 +41,7 @@ def parse_arguments():
|
||||
"--device",
|
||||
type=str,
|
||||
default='cpu',
|
||||
help="Type of inference device, support 'cpu' or 'gpu'.")
|
||||
help="Type of inference device, support 'cpu', 'xpu' or 'gpu'.")
|
||||
parser.add_argument(
|
||||
"--backend",
|
||||
type=str,
|
||||
@@ -68,6 +68,10 @@ def build_option(args):
|
||||
|
||||
option.set_cpu_thread_num(args.cpu_thread_num)
|
||||
|
||||
if args.device.lower() == "xpu":
|
||||
option.use_xpu()
|
||||
return option
|
||||
|
||||
if args.backend.lower() == "trt":
|
||||
assert args.device.lower(
|
||||
) == "gpu", "TensorRT backend require inference on device GPU."
|
||||
|
Reference in New Issue
Block a user