checked format before pull request
diff --git a/include/benchmark/benchmark_api.h b/include/benchmark/benchmark_api.h
index 8d5189a..49167ff 100644
--- a/include/benchmark/benchmark_api.h
+++ b/include/benchmark/benchmark_api.h
@@ -261,16 +261,16 @@
class State {
public:
State(size_t max_iters, bool has_x, int x, bool has_y, int y,
- int thread_i, int n_threads);
+ int thread_i, int n_threads);
- // Returns true iff the benchmark should continue through another iteration.
+ // Returns true if the benchmark should continue through another iteration.
// NOTE: A benchmark may not return from the test until KeepRunning() has
// returned false.
bool KeepRunning() {
if (BENCHMARK_BUILTIN_EXPECT(!started_, false)) {
- assert(!finished_);
- started_ = true;
- ResumeTiming();
+ assert(!finished_);
+ started_ = true;
+ ResumeTiming();
}
bool const res = total_iterations_++ < max_iterations;
if (BENCHMARK_BUILTIN_EXPECT(!res, false)) {
@@ -365,7 +365,7 @@
// represent the length of N.
BENCHMARK_ALWAYS_INLINE
void SetComplexityN(size_t complexity_n) {
- complexity_n_ = complexity_n;
+ complexity_n_ = complexity_n;
}
BENCHMARK_ALWAYS_INLINE
@@ -539,11 +539,11 @@
// to control how many iterations are run, and in the printing of items/second
// or MB/second values.
Benchmark* UseManualTime();
-
+
// Set the asymptotic computational complexity for the benchmark. If called
// the asymptotic computational complexity will be shown on the output.
Benchmark* Complexity(BigO complexity = benchmark::oAuto);
-
+
// Set the asymptotic computational complexity for the benchmark. If called
// the asymptotic computational complexity will be shown on the output.
Benchmark* Complexity(BigOFunc* complexity);
diff --git a/include/benchmark/reporter.h b/include/benchmark/reporter.h
index 4c7bff3..f37e0a3 100644
--- a/include/benchmark/reporter.h
+++ b/include/benchmark/reporter.h
@@ -83,12 +83,12 @@
// This is set to 0.0 if memory tracing is not enabled.
double max_heapbytes_used;
-
+
// Keep track of arguments to compute asymptotic complexity
BigO complexity;
BigOFunc* complexity_lambda;
size_t complexity_n;
-
+
// Inform print function whether the current run is a complexity report
bool report_big_o;
bool report_rms;
@@ -114,7 +114,7 @@
// 'reports' contains additional entries representing the asymptotic
// complexity and RMS of that benchmark family.
virtual void ReportRuns(const std::vector<Run>& report) = 0;
-
+
// Called once and only once after ever group of benchmarks is run and
// reported.
virtual void Finalize() {}
@@ -156,11 +156,11 @@
// Simple reporter that outputs benchmark data to the console. This is the
// default reporter used by RunSpecifiedBenchmarks().
class ConsoleReporter : public BenchmarkReporter {
- public:
+public:
virtual bool ReportContext(const Context& context);
virtual void ReportRuns(const std::vector<Run>& reports);
- protected:
+protected:
virtual void PrintRunData(const Run& report);
size_t name_field_width_;
diff --git a/src/benchmark.cc b/src/benchmark.cc
index 2b55ac7..c56faa9 100644
--- a/src/benchmark.cc
+++ b/src/benchmark.cc
@@ -140,7 +140,7 @@
manual_time_used_(0),
num_finalized_(0),
phase_number_(0),
- entered_(0)
+ entered_(0)
{
}
@@ -277,7 +277,7 @@
int phase_number_cp = phase_number_;
auto cb = [this, phase_number_cp]() {
return this->phase_number_ > phase_number_cp ||
- entered_ == running_threads_; // A thread has aborted in error
+ entered_ == running_threads_; // A thread has aborted in error
};
phase_condition_.wait(ml.native_handle(), cb);
if (phase_number_ > phase_number_cp)
@@ -731,7 +731,7 @@
} // end namespace internal
namespace {
-
+
// Execute one thread of benchmark b for the specified number of iterations.
// Adds the stats collected for the thread into *total.
void RunInThread(const benchmark::internal::Benchmark::Instance* b,
@@ -745,15 +745,15 @@
MutexLock l(GetBenchmarkLock());
total->bytes_processed += st.bytes_processed();
total->items_processed += st.items_processed();
- total->complexity_n += st.complexity_length_n();
+ total->complexity_n += st.complexity_length_n();
}
timer_manager->Finalize();
}
void RunBenchmark(const benchmark::internal::Benchmark::Instance& b,
- BenchmarkReporter* br,
- std::vector<BenchmarkReporter::Run>& complexity_reports)
+ BenchmarkReporter* br,
+ std::vector<BenchmarkReporter::Run>& complexity_reports)
EXCLUDES(GetBenchmarkLock()) {
size_t iters = 1;
@@ -764,7 +764,7 @@
pool.resize(b.threads);
const int repeats = b.repetitions != 0 ? b.repetitions
- : FLAGS_benchmark_repetitions;
+ : FLAGS_benchmark_repetitions;
for (int i = 0; i < repeats; i++) {
std::string mem;
for (;;) {
@@ -844,28 +844,28 @@
report.time_unit = b.time_unit;
if (!report.error_occurred) {
- double bytes_per_second = 0;
- if (total.bytes_processed > 0 && seconds > 0.0) {
- bytes_per_second = (total.bytes_processed / seconds);
- }
- double items_per_second = 0;
- if (total.items_processed > 0 && seconds > 0.0) {
- items_per_second = (total.items_processed / seconds);
- }
+ double bytes_per_second = 0;
+ if (total.bytes_processed > 0 && seconds > 0.0) {
+ bytes_per_second = (total.bytes_processed / seconds);
+ }
+ double items_per_second = 0;
+ if (total.items_processed > 0 && seconds > 0.0) {
+ items_per_second = (total.items_processed / seconds);
+ }
- if (b.use_manual_time) {
- report.real_accumulated_time = manual_accumulated_time;
- } else {
- report.real_accumulated_time = real_accumulated_time;
- }
- report.cpu_accumulated_time = cpu_accumulated_time;
- report.bytes_per_second = bytes_per_second;
- report.items_per_second = items_per_second;
- report.complexity_n = total.complexity_n;
- report.complexity = b.complexity;
- report.complexity_lambda = b.complexity_lambda;
- if(report.complexity != oNone)
- complexity_reports.push_back(report);
+ if (b.use_manual_time) {
+ report.real_accumulated_time = manual_accumulated_time;
+ } else {
+ report.real_accumulated_time = real_accumulated_time;
+ }
+ report.cpu_accumulated_time = cpu_accumulated_time;
+ report.bytes_per_second = bytes_per_second;
+ report.items_per_second = items_per_second;
+ report.complexity_n = total.complexity_n;
+ report.complexity = b.complexity;
+ report.complexity_lambda = b.complexity_lambda;
+ if(report.complexity != oNone)
+ complexity_reports.push_back(report);
}
reports.push_back(report);
@@ -893,17 +893,17 @@
}
std::vector<BenchmarkReporter::Run> additional_run_stats = ComputeStats(reports);
reports.insert(reports.end(), additional_run_stats.begin(),
- additional_run_stats.end());
+ additional_run_stats.end());
if((b.complexity != oNone) && b.last_benchmark_instance) {
additional_run_stats = ComputeBigO(complexity_reports);
reports.insert(reports.end(), additional_run_stats.begin(),
- additional_run_stats.end());
+ additional_run_stats.end());
complexity_reports.clear();
}
br->ReportRuns(reports);
-
+
if (b.multithreaded) {
for (std::thread& thread : pool)
thread.join();
@@ -964,56 +964,56 @@
}
namespace internal {
- namespace {
+namespace {
- void RunMatchingBenchmarks(const std::vector<Benchmark::Instance>& benchmarks,
- BenchmarkReporter* reporter) {
- CHECK(reporter != nullptr);
+void RunMatchingBenchmarks(const std::vector<Benchmark::Instance>& benchmarks,
+ BenchmarkReporter* reporter) {
+ CHECK(reporter != nullptr);
- // Determine the width of the name field using a minimum width of 10.
- bool has_repetitions = FLAGS_benchmark_repetitions > 1;
- size_t name_field_width = 10;
- for (const Benchmark::Instance& benchmark : benchmarks) {
- name_field_width =
- std::max<size_t>(name_field_width, benchmark.name.size());
- has_repetitions |= benchmark.repetitions > 1;
- }
- if (has_repetitions)
- name_field_width += std::strlen("_stddev");
+ // Determine the width of the name field using a minimum width of 10.
+ bool has_repetitions = FLAGS_benchmark_repetitions > 1;
+ size_t name_field_width = 10;
+ for (const Benchmark::Instance& benchmark : benchmarks) {
+ name_field_width =
+ std::max<size_t>(name_field_width, benchmark.name.size());
+ has_repetitions |= benchmark.repetitions > 1;
+ }
+ if (has_repetitions)
+ name_field_width += std::strlen("_stddev");
- // Print header here
- BenchmarkReporter::Context context;
- context.num_cpus = NumCPUs();
- context.mhz_per_cpu = CyclesPerSecond() / 1000000.0f;
+ // Print header here
+ BenchmarkReporter::Context context;
+ context.num_cpus = NumCPUs();
+ context.mhz_per_cpu = CyclesPerSecond() / 1000000.0f;
- context.cpu_scaling_enabled = CpuScalingEnabled();
- context.name_field_width = name_field_width;
+ context.cpu_scaling_enabled = CpuScalingEnabled();
+ context.name_field_width = name_field_width;
- // Keep track of runing times of all instances of current benchmark
- std::vector<BenchmarkReporter::Run> complexity_reports;
+ // Keep track of runing times of all instances of current benchmark
+ std::vector<BenchmarkReporter::Run> complexity_reports;
- if (reporter->ReportContext(context)) {
- for (const auto& benchmark : benchmarks) {
- RunBenchmark(benchmark, reporter, complexity_reports);
- }
- }
+ if (reporter->ReportContext(context)) {
+ for (const auto& benchmark : benchmarks) {
+ RunBenchmark(benchmark, reporter, complexity_reports);
}
+ }
+}
- std::unique_ptr<BenchmarkReporter> GetDefaultReporter() {
- typedef std::unique_ptr<BenchmarkReporter> PtrType;
- if (FLAGS_benchmark_format == "console") {
- return PtrType(new ConsoleReporter);
- } else if (FLAGS_benchmark_format == "json") {
- return PtrType(new JSONReporter);
- } else if (FLAGS_benchmark_format == "csv") {
- return PtrType(new CSVReporter);
- } else {
- std::cerr << "Unexpected format: '" << FLAGS_benchmark_format << "'\n";
- std::exit(1);
- }
- }
+std::unique_ptr<BenchmarkReporter> GetDefaultReporter() {
+ typedef std::unique_ptr<BenchmarkReporter> PtrType;
+ if (FLAGS_benchmark_format == "console") {
+ return PtrType(new ConsoleReporter);
+ } else if (FLAGS_benchmark_format == "json") {
+ return PtrType(new JSONReporter);
+ } else if (FLAGS_benchmark_format == "csv") {
+ return PtrType(new CSVReporter);
+ } else {
+ std::cerr << "Unexpected format: '" << FLAGS_benchmark_format << "'\n";
+ std::exit(1);
+ }
+}
- } // end namespace
+} // end namespace
} // end namespace internal
size_t RunSpecifiedBenchmarks() {
diff --git a/src/complexity.cc b/src/complexity.cc
index 97f86d8..0d6f90b 100644
--- a/src/complexity.cc
+++ b/src/complexity.cc
@@ -194,9 +194,9 @@
mean_data.benchmark_name = reports[0].benchmark_name + "_mean";
mean_data.iterations = run_iterations;
mean_data.real_accumulated_time = real_accumulated_time_stat.Mean() *
- run_iterations;
+ run_iterations;
mean_data.cpu_accumulated_time = cpu_accumulated_time_stat.Mean() *
- run_iterations;
+ run_iterations;
mean_data.bytes_per_second = bytes_per_second_stat.Mean();
mean_data.items_per_second = items_per_second_stat.Mean();
diff --git a/src/complexity.h b/src/complexity.h
index 798154a..85cc125 100644
--- a/src/complexity.h
+++ b/src/complexity.h
@@ -26,15 +26,15 @@
namespace benchmark {
- // Return a vector containing the mean and standard devation information for
- // the specified list of reports. If 'reports' contains less than two
- // non-errored runs an empty vector is returned
- std::vector<BenchmarkReporter::Run> ComputeStats(
+// Return a vector containing the mean and standard devation information for
+// the specified list of reports. If 'reports' contains less than two
+// non-errored runs an empty vector is returned
+std::vector<BenchmarkReporter::Run> ComputeStats(
const std::vector<BenchmarkReporter::Run>& reports);
- // Return a vector containing the bigO and RMS information for the specified
- // list of reports. If 'reports.size() < 2' an empty vector is returned.
- std::vector<BenchmarkReporter::Run> ComputeBigO(
+// Return a vector containing the bigO and RMS information for the specified
+// list of reports. If 'reports.size() < 2' an empty vector is returned.
+std::vector<BenchmarkReporter::Run> ComputeBigO(
const std::vector<BenchmarkReporter::Run>& reports);
// This data structure will contain the result returned by MinimalLeastSq
diff --git a/src/json_reporter.cc b/src/json_reporter.cc
index 04cb490..da88355 100644
--- a/src/json_reporter.cc
+++ b/src/json_reporter.cc
@@ -155,7 +155,7 @@
} else if(run.report_rms) {
out << indent
<< FormatKV("rms", RoundDouble(run.GetAdjustedCPUTime()*100))
- << "%";
+ << '%';
}
if (run.bytes_per_second > 0.0) {
out << ",\n" << indent