mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-15 13:10:55 +08:00
[Example] Update runtime examples (#1542)
* Add notes for tensors * Optimize some apis * move some warnings
This commit is contained in:
@@ -25,27 +25,18 @@ namespace fastdeploy {
|
||||
/*! @brief Option object to configure ONNX Runtime backend
|
||||
*/
|
||||
struct OrtBackendOption {
|
||||
/*
|
||||
* @brief Level of graph optimization, -1: mean default(Enable all the optimization strategy)/0: disable all the optimization strategy/1: enable basic strategy/2:enable extend strategy/99: enable all
|
||||
*/
|
||||
/// Level of graph optimization, -1: mean default(Enable all the optimization strategy)/0: disable all the optimization strategy/1: enable basic strategy/2:enable extend strategy/99: enable all
|
||||
int graph_optimization_level = -1;
|
||||
/*
|
||||
* @brief Number of threads to execute the operator, -1: default
|
||||
*/
|
||||
/// Number of threads to execute the operator, -1: default
|
||||
int intra_op_num_threads = -1;
|
||||
/*
|
||||
* @brief Number of threads to execute the graph, -1: default. This parameter only will bring effects while the `OrtBackendOption::execution_mode` set to 1.
|
||||
*/
|
||||
/// Number of threads to execute the graph, -1: default. This parameter only will bring effects while the `OrtBackendOption::execution_mode` set to 1.
|
||||
int inter_op_num_threads = -1;
|
||||
/*
|
||||
* @brief Execution mode for the graph, -1: default(Sequential mode)/0: Sequential mode, execute the operators in graph one by one. /1: Parallel mode, execute the operators in graph parallelly.
|
||||
*/
|
||||
/// Execution mode for the graph, -1: default(Sequential mode)/0: Sequential mode, execute the operators in graph one by one. /1: Parallel mode, execute the operators in graph parallelly.
|
||||
int execution_mode = -1;
|
||||
/// Inference device, OrtBackend supports CPU/GPU
|
||||
// Inference device, OrtBackend supports CPU/GPU
|
||||
Device device = Device::CPU;
|
||||
/// Inference device id
|
||||
// Inference device id
|
||||
int device_id = 0;
|
||||
|
||||
void* external_stream_ = nullptr;
|
||||
};
|
||||
} // namespace fastdeploy
|
||||
|
Reference in New Issue
Block a user