mirror of
https://github.com/google/benchmark.git
synced 2025-02-14 13:20:13 +08:00
* Introduce warmup phase to BenchmarkRunner (#1130) In order to account for caching effects in user benchmarks introduce a new command line option "--benchmark_min_warmup_time" which allows to specify an amount of time for which the benchmark should be run before results are meaningful. * Adapt review suggestions regarding introduction of warmup phase (#1130) * Fix BM_CHECK call in MinWarmUpTime (#1130) * Fix comment on requirements of MinWarmUpTime (#1130) * Add basic description of warmup phase mechanism to user guide (#1130)
119 lines
3.4 KiB
C++
119 lines
3.4 KiB
C++
#include "benchmark_api_internal.h"
|
|
|
|
#include <cinttypes>
|
|
|
|
#include "string_util.h"
|
|
|
|
namespace benchmark {
|
|
namespace internal {
|
|
|
|
BenchmarkInstance::BenchmarkInstance(Benchmark* benchmark, int family_idx,
|
|
int per_family_instance_idx,
|
|
const std::vector<int64_t>& args,
|
|
int thread_count)
|
|
: benchmark_(*benchmark),
|
|
family_index_(family_idx),
|
|
per_family_instance_index_(per_family_instance_idx),
|
|
aggregation_report_mode_(benchmark_.aggregation_report_mode_),
|
|
args_(args),
|
|
time_unit_(benchmark_.GetTimeUnit()),
|
|
measure_process_cpu_time_(benchmark_.measure_process_cpu_time_),
|
|
use_real_time_(benchmark_.use_real_time_),
|
|
use_manual_time_(benchmark_.use_manual_time_),
|
|
complexity_(benchmark_.complexity_),
|
|
complexity_lambda_(benchmark_.complexity_lambda_),
|
|
statistics_(benchmark_.statistics_),
|
|
repetitions_(benchmark_.repetitions_),
|
|
min_time_(benchmark_.min_time_),
|
|
min_warmup_time_(benchmark_.min_warmup_time_),
|
|
iterations_(benchmark_.iterations_),
|
|
threads_(thread_count) {
|
|
name_.function_name = benchmark_.name_;
|
|
|
|
size_t arg_i = 0;
|
|
for (const auto& arg : args) {
|
|
if (!name_.args.empty()) {
|
|
name_.args += '/';
|
|
}
|
|
|
|
if (arg_i < benchmark->arg_names_.size()) {
|
|
const auto& arg_name = benchmark_.arg_names_[arg_i];
|
|
if (!arg_name.empty()) {
|
|
name_.args += StrFormat("%s:", arg_name.c_str());
|
|
}
|
|
}
|
|
|
|
name_.args += StrFormat("%" PRId64, arg);
|
|
++arg_i;
|
|
}
|
|
|
|
if (!IsZero(benchmark->min_time_)) {
|
|
name_.min_time = StrFormat("min_time:%0.3f", benchmark_.min_time_);
|
|
}
|
|
|
|
if (!IsZero(benchmark->min_warmup_time_)) {
|
|
name_.min_warmup_time =
|
|
StrFormat("min_warmup_time:%0.3f", benchmark_.min_warmup_time_);
|
|
}
|
|
|
|
if (benchmark_.iterations_ != 0) {
|
|
name_.iterations = StrFormat(
|
|
"iterations:%lu", static_cast<unsigned long>(benchmark_.iterations_));
|
|
}
|
|
|
|
if (benchmark_.repetitions_ != 0) {
|
|
name_.repetitions = StrFormat("repeats:%d", benchmark_.repetitions_);
|
|
}
|
|
|
|
if (benchmark_.measure_process_cpu_time_) {
|
|
name_.time_type = "process_time";
|
|
}
|
|
|
|
if (benchmark_.use_manual_time_) {
|
|
if (!name_.time_type.empty()) {
|
|
name_.time_type += '/';
|
|
}
|
|
name_.time_type += "manual_time";
|
|
} else if (benchmark_.use_real_time_) {
|
|
if (!name_.time_type.empty()) {
|
|
name_.time_type += '/';
|
|
}
|
|
name_.time_type += "real_time";
|
|
}
|
|
|
|
if (!benchmark_.thread_counts_.empty()) {
|
|
name_.threads = StrFormat("threads:%d", threads_);
|
|
}
|
|
|
|
setup_ = benchmark_.setup_;
|
|
teardown_ = benchmark_.teardown_;
|
|
}
|
|
|
|
State BenchmarkInstance::Run(
|
|
IterationCount iters, int thread_id, internal::ThreadTimer* timer,
|
|
internal::ThreadManager* manager,
|
|
internal::PerfCountersMeasurement* perf_counters_measurement) const {
|
|
State st(iters, args_, thread_id, threads_, timer, manager,
|
|
perf_counters_measurement);
|
|
benchmark_.Run(st);
|
|
return st;
|
|
}
|
|
|
|
void BenchmarkInstance::Setup() const {
|
|
if (setup_) {
|
|
State st(/*iters*/ 1, args_, /*thread_id*/ 0, threads_, nullptr, nullptr,
|
|
nullptr);
|
|
setup_(st);
|
|
}
|
|
}
|
|
|
|
void BenchmarkInstance::Teardown() const {
|
|
if (teardown_) {
|
|
State st(/*iters*/ 1, args_, /*thread_id*/ 0, threads_, nullptr, nullptr,
|
|
nullptr);
|
|
teardown_(st);
|
|
}
|
|
}
|
|
} // namespace internal
|
|
} // namespace benchmark
|