mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-06 17:17:14 +08:00
Merge branch 'develop' into set_stream_infer-shareExData
This commit is contained in:
@@ -59,6 +59,12 @@ struct FASTDEPLOY_DECL RuntimeOption {
|
||||
const std::string& params_buffer = "",
|
||||
const ModelFormat& format = ModelFormat::PADDLE);
|
||||
|
||||
/** \brief When loading encrypted model, encryption_key is required to decrypte model
|
||||
*
|
||||
* \param[in] encryption_key The key for decrypting model
|
||||
*/
|
||||
void SetEncryptionKey(const std::string& encryption_key);
|
||||
|
||||
/// Use cpu to inference, the runtime will inference on CPU by default
|
||||
void UseCpu();
|
||||
/// Use Nvidia GPU to inference
|
||||
@@ -179,6 +185,8 @@ struct FASTDEPLOY_DECL RuntimeOption {
|
||||
/// format of input model
|
||||
ModelFormat model_format = ModelFormat::PADDLE;
|
||||
|
||||
std::string encryption_key_ = "";
|
||||
|
||||
// for cpu inference
|
||||
// default will let the backend choose their own default value
|
||||
int cpu_thread_num = -1;
|
||||
|
Reference in New Issue
Block a user