This directory contains the Swift Benchmark Suite.
To run Swift benchmarks, pass the --benchmark flag to build-script. The current benchmark results will be compared to the previous run's results if available. Results for each benchmark run are logged for future comparison.
For branch based development, take a baseline benchmark on the Swift master branch, switch to a development branch containing potentially performance impacting changes, and run the benchmarks again. Upon benchmark completion, the benchmark results for the development branch will be compared to the most recent benchmark results for master.
By default, Swift benchmarks for OS X are compiled during the Swift build process. To build Swift benchmarks for additional platforms, pass the following flags:
$ swift/utils/build-script --ios --watchos --tvos
OS X benchmark driver binaries are placed in bin alongside swiftc. Additional platform binaries are placed in the benchmark/bin build directory.
To build the Swift benchmarks using only an Xcode installation: install an Xcode version with Swift support, install cmake 2.8.12, and ensure Xcode is selected with xcode-select.
The following build options are available:
-DSWIFT_EXECswiftc) to use to compile the benchmarks (default: Xcode's swiftc)-DSWIFT_LIBRARY_PATHswiftc_directory/../lib/swift)-DONLY_PLATFORMS-DSWIFT_OPTIMIZATION_LEVELS-DSWIFT_BENCHMARK_EMIT_SIBThe following build targets are available:
swift-benchmark-macosx-x86_64swift-benchmark-iphoneos-arm64swift-benchmark-iphoneos-armv7swift-benchmark-appletvos-arm64swift-benchmark-watchos-armv7kBuild steps (with example options):
$ cd benchmark$ mkdir build$ cd build$ cmake ..$ make -j8 swift-benchmark-macosx-x86_64Benchmark driver binaries are placed in build/bin and the required Swift standard library dylibs are placed in build/lib. The drivers dynamically link Swift standard library dylibs from a path relative to their location (../lib/swift) so the standard library should be distributed alongside them.
./Driver [ test_name [ test_name ] ] [ option [ option ] ]
--num-iters--num-samples--list$ ./Benchmark_O --num-iters=1 --num-samples=1$ ./Benchmark_Onone --list$ ./Benchmark_Ounchecked AckermannAs a shortcut, you can also refer to benchmarks by their ordinal numbers. The regular --list option does not provide these, but you can run:
$ ./Benchmark_O --list --run-all | tail -n +2 | nl You can use ordinal numbers instead of test names like this:$ ./Benchmark_O 1 42$ ./Benchmark_Driver run 1 42scripts/generate_harness/generate_harness.py generates and replaces CMakeLists.txt and utils/main.swift from single and multiple file tests contained in the directories single-source and multi-source. It gathers information about the tests and then generates the files from templates using gyb. The motivation for creating this script was to eliminate the need to manually add at least three lines to harness files (one to CMakeLists.txt and two to utils/main.swift) for every new benchmark added.
Warning:
Since CMakeLists.txt and utils/main.swift are now generated from templates, they should not be directly modified. Work may be lost if the harness is executed after making changes to derived files. Instead, modifications should be made to the *.gyb template files.
To generate CMakeLists.txt and utils/main.swift from test sources, run the command:
$ scripts/generate_harness/generate_harness.py
To make changes to CMakeLists.txt or utils/main.swift, modify the template files CMakeLists.txt.gyb and main.swift.gyb These are templates, rendered by gyb in generate_harness.py, so ensure static changes don't interfere with the template portions. Test changes by regenerating the harness and rebuilding the repository with build-script.
The harness generator supports both single and multiple file tests.
To add a new single file test:
YourTestNameHere.swift), built according to the template below, to the single-source directory.To add a new multiple file test:
Add a new directory and files under the multi-source directory as specified below:
+-- multi-source | +-- YourTestName | | +-- TestFile1.swift | | +-- TestFile2.swift | | +-- TestFile3.swift
At least one run function (specified in the template below) must exist in the files.
Regenerate harness files by following the directions in Generating harness files before committing changes.
Note:
The generator script looks for functions prefixed with run_ in order to populate utils/main.swift.
The benchmark driver will measure the time taken for N = 1 and automatically calculate the necessary number of iterations N to run each benchmark in approximately one second, so the test should ideally run in a few milliseconds for N = 1. If the test contains any setup code before the loop, ensure the time spent on setup is insignificant compared to the time spent inside the loop (for N = 1) -- otherwise the automatic calculation of N might be significantly off and any performance gains/regressions will be masked by the fixed setup time. If needed you can multiply N by a fixed amount (e.g. 1...100*N) to achieve this.
Performance Test Template
// YourTestNameHere benchmark // // rdar://problem/00000000 import Foundation import TestsUtils @inline(never) public func run_YourTestNameHere(N: Int) { # Declare variables for i in 1...N { # Perform work # Verify work was done; break otherwise } # Assert with CheckResults that work was done }