mirror of
				https://github.com/PaddlePaddle/FastDeploy.git
				synced 2025-11-01 04:12:58 +08:00 
			
		
		
		
	
		
			
				
	
	
		
			242 lines
		
	
	
		
			6.8 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
			
		
		
	
	
			242 lines
		
	
	
		
			6.8 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2012 The Android Open Source Project
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *      http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| #include "benchmark.h"
 | |
| #include <inttypes.h>
 | |
| #include <regex.h>
 | |
| #include <stdio.h>
 | |
| #include <stdlib.h>
 | |
| #include <string.h>
 | |
| #include <time.h>
 | |
| #include <map>
 | |
| #include <string>
 | |
| 
 | |
| static int64_t g_flops_processed;
 | |
| static int64_t g_benchmark_total_time_ns;
 | |
| static int64_t g_benchmark_start_time_ns;
 | |
| typedef std::map<std::string, ::testing::Benchmark*> BenchmarkMap;
 | |
| typedef BenchmarkMap::iterator BenchmarkMapIt;
 | |
| 
 | |
| BenchmarkMap& gBenchmarks() {
 | |
|   static BenchmarkMap g_benchmarks;
 | |
|   return g_benchmarks;
 | |
| }
 | |
| 
 | |
| static int g_name_column_width = 20;
 | |
| 
 | |
| static int Round(int n) {
 | |
|   int base = 1;
 | |
|   while (base * 10 < n) {
 | |
|     base *= 10;
 | |
|   }
 | |
|   if (n < 2 * base) {
 | |
|     return 2 * base;
 | |
|   }
 | |
|   if (n < 5 * base) {
 | |
|     return 5 * base;
 | |
|   }
 | |
|   return 10 * base;
 | |
| }
 | |
| 
 | |
| #ifdef __APPLE__
 | |
| #include <mach/mach_time.h>
 | |
| static mach_timebase_info_data_t g_time_info;
 | |
| static void __attribute__((constructor)) init_info() {
 | |
|   mach_timebase_info(&g_time_info);
 | |
| }
 | |
| #endif
 | |
| 
 | |
| static int64_t NanoTime() {
 | |
| #if defined(__APPLE__)
 | |
|   uint64_t t = mach_absolute_time();
 | |
|   return t * g_time_info.numer / g_time_info.denom;
 | |
| #else
 | |
|   struct timespec t;
 | |
|   t.tv_sec = t.tv_nsec = 0;
 | |
|   clock_gettime(CLOCK_MONOTONIC, &t);
 | |
|   return static_cast<int64_t>(t.tv_sec) * 1000000000LL + t.tv_nsec;
 | |
| #endif
 | |
| }
 | |
| 
 | |
| namespace testing {
 | |
| Benchmark* Benchmark::Arg(int arg) {
 | |
|   args_.push_back(arg);
 | |
|   return this;
 | |
| }
 | |
| 
 | |
| Benchmark* Benchmark::Range(int lo, int hi) {
 | |
|   const int kRangeMultiplier = 8;
 | |
|   if (hi < lo) {
 | |
|     int temp = hi;
 | |
|     hi = lo;
 | |
|     lo = temp;
 | |
|   }
 | |
|   while (lo < hi) {
 | |
|     args_.push_back(lo);
 | |
|     lo *= kRangeMultiplier;
 | |
|   }
 | |
|   // We always run the hi number.
 | |
|   args_.push_back(hi);
 | |
|   return this;
 | |
| }
 | |
| 
 | |
| const char* Benchmark::Name() { return name_; }
 | |
| bool Benchmark::ShouldRun(int argc, char* argv[]) {
 | |
|   if (argc == 1) {
 | |
|     return true;  // With no arguments, we run all benchmarks.
 | |
|   }
 | |
|   // Otherwise, we interpret each argument as a regular expression and
 | |
|   // see if any of our benchmarks match.
 | |
|   for (int i = 1; i < argc; i++) {
 | |
|     regex_t re;
 | |
|     if (regcomp(&re, argv[i], 0) != 0) {
 | |
|       fprintf(stderr, "couldn't compile \"%s\" as a regular expression!\n",
 | |
|               argv[i]);
 | |
|       exit(EXIT_FAILURE);
 | |
|     }
 | |
|     int match = regexec(&re, name_, 0, NULL, 0);
 | |
|     regfree(&re);
 | |
|     if (match != REG_NOMATCH) {
 | |
|       return true;
 | |
|     }
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| void Benchmark::Register(const char* name, void (*fn)(int),
 | |
|                          void (*fn_range)(int, int)) {
 | |
|   name_ = name;
 | |
|   fn_ = fn;
 | |
|   fn_range_ = fn_range;
 | |
|   if (fn_ == NULL && fn_range_ == NULL) {
 | |
|     fprintf(stderr, "%s: missing function\n", name_);
 | |
|     exit(EXIT_FAILURE);
 | |
|   }
 | |
|   gBenchmarks().insert(std::make_pair(name, this));
 | |
| }
 | |
| void Benchmark::Run() {
 | |
|   if (fn_ != NULL) {
 | |
|     RunWithArg(0);
 | |
|   } else {
 | |
|     if (args_.empty()) {
 | |
|       fprintf(stderr, "%s: no args!\n", name_);
 | |
|       exit(EXIT_FAILURE);
 | |
|     }
 | |
|     for (size_t i = 0; i < args_.size(); ++i) {
 | |
|       RunWithArg(args_[i]);
 | |
|     }
 | |
|   }
 | |
| }
 | |
| void Benchmark::RunRepeatedlyWithArg(int iterations, int arg) {
 | |
|   g_flops_processed = 0;
 | |
|   g_benchmark_total_time_ns = 0;
 | |
|   g_benchmark_start_time_ns = NanoTime();
 | |
|   if (fn_ != NULL) {
 | |
|     fn_(iterations);
 | |
|   } else {
 | |
|     fn_range_(iterations, arg);
 | |
|   }
 | |
|   if (g_benchmark_start_time_ns != 0) {
 | |
|     g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns;
 | |
|   }
 | |
| }
 | |
| void Benchmark::RunWithArg(int arg) {
 | |
|   // run once in case it's expensive
 | |
|   int iterations = 1;
 | |
|   RunRepeatedlyWithArg(iterations, arg);
 | |
|   while (g_benchmark_total_time_ns < 1e9 && iterations < 1e9) {
 | |
|     int last = iterations;
 | |
|     if (g_benchmark_total_time_ns / iterations == 0) {
 | |
|       iterations = 1e9;
 | |
|     } else {
 | |
|       iterations = 1e9 / (g_benchmark_total_time_ns / iterations);
 | |
|     }
 | |
|     iterations =
 | |
|         std::max(last + 1, std::min(iterations + iterations / 2, 100 * last));
 | |
|     iterations = Round(iterations);
 | |
|     RunRepeatedlyWithArg(iterations, arg);
 | |
|   }
 | |
|   char throughput[100];
 | |
|   throughput[0] = '\0';
 | |
|   if (g_benchmark_total_time_ns > 0 && g_flops_processed > 0) {
 | |
|     double mflops_processed = static_cast<double>(g_flops_processed) / 1e6;
 | |
|     double seconds = static_cast<double>(g_benchmark_total_time_ns) / 1e9;
 | |
|     snprintf(throughput, sizeof(throughput), " %8.2f MFlops/s",
 | |
|              mflops_processed / seconds);
 | |
|   }
 | |
|   char full_name[100];
 | |
|   if (fn_range_ != NULL) {
 | |
|     if (arg >= (1 << 20)) {
 | |
|       snprintf(full_name, sizeof(full_name), "%s/%dM", name_, arg / (1 << 20));
 | |
|     } else if (arg >= (1 << 10)) {
 | |
|       snprintf(full_name, sizeof(full_name), "%s/%dK", name_, arg / (1 << 10));
 | |
|     } else {
 | |
|       snprintf(full_name, sizeof(full_name), "%s/%d", name_, arg);
 | |
|     }
 | |
|   } else {
 | |
|     snprintf(full_name, sizeof(full_name), "%s", name_);
 | |
|   }
 | |
|   printf("%-*s %10d %10" PRId64 "%s\n", g_name_column_width, full_name,
 | |
|          iterations, g_benchmark_total_time_ns / iterations, throughput);
 | |
|   fflush(stdout);
 | |
| }
 | |
| }  // namespace testing
 | |
| void SetBenchmarkFlopsProcessed(int64_t x) { g_flops_processed = x; }
 | |
| void StopBenchmarkTiming() {
 | |
|   if (g_benchmark_start_time_ns != 0) {
 | |
|     g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns;
 | |
|   }
 | |
|   g_benchmark_start_time_ns = 0;
 | |
| }
 | |
| void StartBenchmarkTiming() {
 | |
|   if (g_benchmark_start_time_ns == 0) {
 | |
|     g_benchmark_start_time_ns = NanoTime();
 | |
|   }
 | |
| }
 | |
| int main(int argc, char* argv[]) {
 | |
|   if (gBenchmarks().empty()) {
 | |
|     fprintf(stderr, "No benchmarks registered!\n");
 | |
|     exit(EXIT_FAILURE);
 | |
|   }
 | |
|   for (BenchmarkMapIt it = gBenchmarks().begin(); it != gBenchmarks().end();
 | |
|        ++it) {
 | |
|     int name_width = static_cast<int>(strlen(it->second->Name()));
 | |
|     g_name_column_width = std::max(g_name_column_width, name_width);
 | |
|   }
 | |
|   bool need_header = true;
 | |
|   for (BenchmarkMapIt it = gBenchmarks().begin(); it != gBenchmarks().end();
 | |
|        ++it) {
 | |
|     ::testing::Benchmark* b = it->second;
 | |
|     if (b->ShouldRun(argc, argv)) {
 | |
|       if (need_header) {
 | |
|         printf("%-*s %10s %10s\n", g_name_column_width, "", "iterations",
 | |
|                "ns/op");
 | |
|         fflush(stdout);
 | |
|         need_header = false;
 | |
|       }
 | |
|       b->Run();
 | |
|     }
 | |
|   }
 | |
|   if (need_header) {
 | |
|     fprintf(stderr, "No matching benchmarks!\n");
 | |
|     fprintf(stderr, "Available benchmarks:\n");
 | |
|     for (BenchmarkMapIt it = gBenchmarks().begin(); it != gBenchmarks().end();
 | |
|          ++it) {
 | |
|       fprintf(stderr, "  %s\n", it->second->Name());
 | |
|     }
 | |
|     exit(EXIT_FAILURE);
 | |
|   }
 | |
|   return 0;
 | |
| }
 | 
