mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-06 00:57:33 +08:00

* add GPL lisence * add GPL-3.0 lisence * add GPL-3.0 lisence * add GPL-3.0 lisence * support yolov8 * add pybind for yolov8 * add yolov8 readme * add cpp benchmark * add cpu and gpu mem * public part split * add runtime mode * fixed bugs * add cpu_thread_nums * deal with comments * deal with comments * deal with comments * rm useless code * add FASTDEPLOY_DECL * add FASTDEPLOY_DECL * fixed for windows * mv rss to pss * mv rss to pss * Update utils.cc * use thread to collect mem * Add ResourceUsageMonitor * rm useless code * fixed bug * fixed typo * update ResourceUsageMonitor * fixed bug * fixed bug * add note for ResourceUsageMonitor * deal with comments --------- Co-authored-by: DefTruth <31974251+DefTruth@users.noreply.github.com>
85 lines
2.6 KiB
C++
Executable File
85 lines
2.6 KiB
C++
Executable File
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
#pragma once
|
|
|
|
#include <memory>
|
|
#include <thread> // NOLINT
|
|
#include "fastdeploy/utils/utils.h"
|
|
|
|
namespace fastdeploy {
|
|
namespace benchmark {
|
|
/*! @brief ResourceUsageMonitor object used when to collect memory info.
|
|
*/
|
|
class FASTDEPLOY_DECL ResourceUsageMonitor {
|
|
public:
|
|
/** \brief Set sampling_interval_ms and gpu_id for ResourceUsageMonitor.
|
|
*
|
|
* \param[in] sampling_interval_ms How often to collect memory info(ms).
|
|
* \param[in] gpu_id Device(gpu) id, default 0.
|
|
*/
|
|
explicit ResourceUsageMonitor(int sampling_interval_ms, int gpu_id = 0);
|
|
|
|
~ResourceUsageMonitor() { StopInternal(); }
|
|
|
|
/// Start memory info collect
|
|
void Start();
|
|
/// Stop memory info collect
|
|
void Stop();
|
|
/// Get maximum cpu memory usage
|
|
float GetMaxCpuMem() const {
|
|
if (!is_supported_ || check_memory_thd_ == nullptr) {
|
|
return -1.0f;
|
|
}
|
|
return max_cpu_mem_;
|
|
}
|
|
/// Get maximum gpu memory usage
|
|
float GetMaxGpuMem() const {
|
|
if (!is_supported_ || check_memory_thd_ == nullptr) {
|
|
return -1.0f;
|
|
}
|
|
return max_gpu_mem_;
|
|
}
|
|
/// Get maximum gpu util
|
|
float GetMaxGpuUtil() const {
|
|
if (!is_supported_ || check_memory_thd_ == nullptr) {
|
|
return -1.0f;
|
|
}
|
|
return max_gpu_util_;
|
|
}
|
|
|
|
ResourceUsageMonitor(ResourceUsageMonitor&) = delete;
|
|
ResourceUsageMonitor& operator=(const ResourceUsageMonitor&) = delete;
|
|
ResourceUsageMonitor(ResourceUsageMonitor&&) = delete;
|
|
ResourceUsageMonitor& operator=(const ResourceUsageMonitor&&) = delete;
|
|
|
|
private:
|
|
void StopInternal();
|
|
// Get current cpu memory info
|
|
std::string GetCurrentCpuMemoryInfo();
|
|
// Get current gpu memory info
|
|
std::string GetCurrentGpuMemoryInfo(int device_id);
|
|
|
|
bool is_supported_ = false;
|
|
bool stop_signal_ = false;
|
|
const int sampling_interval_;
|
|
float max_cpu_mem_ = 0.0f;
|
|
float max_gpu_mem_ = 0.0f;
|
|
float max_gpu_util_ = 0.0f;
|
|
const int gpu_id_ = 0;
|
|
std::unique_ptr<std::thread> check_memory_thd_ = nullptr;
|
|
};
|
|
|
|
} // namespace benchmark
|
|
} // namespace fastdeploy
|