mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-12-24 13:28:13 +08:00
Update README.md
This commit is contained in:
@@ -34,7 +34,7 @@ English | [简体中文](README_ch.md)
|
|||||||
- **New server-side deployment upgrade: faster inference performance, support more visual model**
|
- **New server-side deployment upgrade: faster inference performance, support more visual model**
|
||||||
- Release high-performance inference engine SDK based on x86 CPUs and NVIDIA GPUs, with significant increase in inference speed
|
- Release high-performance inference engine SDK based on x86 CPUs and NVIDIA GPUs, with significant increase in inference speed
|
||||||
- Integrate Paddle Inference, ONNXRuntime, TensorRT and other inference engines and provide a seamless deployment experience
|
- Integrate Paddle Inference, ONNXRuntime, TensorRT and other inference engines and provide a seamless deployment experience
|
||||||
- Supports full range of object detection models such as YOLOv7, YOLOv6, YOLOv5, PP-YOLOE and provides [End-To-End Deployment Demos]](examples/vision/detection/)
|
- Supports full range of object detection models such as YOLOv7, YOLOv6, YOLOv5, PP-YOLOE and provides [End-To-End Deployment Demos](examples/vision/detection/)
|
||||||
- Support over 40 key models and [Demo Examples](examples/vision/) including face detection, face recognition, real-time portrait matting, image segmentation.
|
- Support over 40 key models and [Demo Examples](examples/vision/) including face detection, face recognition, real-time portrait matting, image segmentation.
|
||||||
- Support deployment in both Python and C++
|
- Support deployment in both Python and C++
|
||||||
- **Supports Rexchip, Amlogic, NXP and other NPU chip deployment capabilities on end-side deployment**
|
- **Supports Rexchip, Amlogic, NXP and other NPU chip deployment capabilities on end-side deployment**
|
||||||
|
|||||||
Reference in New Issue
Block a user