You can run benchmarks in Continuous Integration (CI) to track performance over time, and recognize performance regressions (or improvements) before your app even releases. This page provides basic information about benchmarking in CI.
Before getting started with benchmarking in CI, consider how capturing and evaluating results differs from regular tests.
Fuzzy results
Although benchmarks are instrumented tests, results aren't just a pass/fail. Benchmarks provide timing measurements for the given device they run on. Graphing results over time lets you monitor change and observe noise in the measurement system.
Use real devices
Benchmarks should run on physical Android devices. While they can run on emulators, it’s strongly discouraged as it won't represent a realistic user experience and will instead provide numbers tied to the host OS and hardware capabilities. Consider using real devices or a service that lets you run tests on real devices, such as Firebase Test Lab.
Run the benchmarks
Running the benchmarks as part of your CI pipeline may be different than running
it locally from Android Studio. Locally, you’d usually run the Android
integration tests with one Gradle connectedCheck
task. This task automatically
builds your APK and test APK and runs the tests on the currently connected
device(s). When running in CI, this flow usually needs to be split into separate
phases.
Build
For the Microbenchmark library, run the Gradle task
assemble[VariantName]AndroidTest
, which creates your test APK that contains
both your application code as well as your tested code.
Alternatively, Macrobenchmark library requires you to build your target APK and
test APK separately. Therefore run :app:assemble[VariantName]
and
:macrobenchmark:assemble[VariantName]
Gradle tasks.
Install and run
These steps are typically done without needing to run Gradle tasks. Note, they may be abstracted depending on whether you use a service that lets you run tests on real devices.
For installation, use adb install
command and specify the test APK (or the
target APK).
Run adb shell am
instrument command to run all the benchmarks.
adb shell am instrument -w com.example.benchmark/androidx.benchmark.junit4.AndroidBenchmarkRunner
Note, that when using the Macrobenchmark library, use regular
androidx.test.runner.AndroidJUnitRunner
as instrumentation runner.
You can pass the same instrumentation arguments as in the Gradle configuration
using -e
argument. For all the instrumentation arguments options, check the
pages for Microbenchmark or Macrobenchmark.
For example, you can set the dryRunMode
argument to run microbenchmarks as
part of your pull request verification process. With this flag enabled,
microbenchmarks run only in single loop allowing to verify they are running
correctly, but not taking too long to execute.
adb shell am instrument -w -e "androidx.benchmark.dryRunMode.enable" "true" com.example.benchmark/androidx.benchmark.junit4.AndroidBenchmarkRunner
Check Running tests with ADB for more information on how to run instrumentation tests from the command line.
Lock clocks
The Microbenchmark Gradle plugin provides the command ./gradlew lockClocks
to
lock the CPU clocks of a rooted device. This is useful for ensuring stability
when you have access to rooted devices, such as "userdebug" builds. You can
replicate this with the lockClocks.sh
shell script, available in the
library's source.
You can either run the script directly from a Linux or Mac host, or you can push to the device with a few adb commands:
adb push path/lockClocks.sh /data/local/tmp/lockClocks.sh adb shell /data/local/tmp/lockClocks.sh adb shell rm /data/local/tmp/lockClocks.sh
If you run the shell script directly on a host, it dispatches these commands to a connected device.
For more information on why it is helpful to lock the CPU clocks, check how to obtain consistent benchmarks.
Collect the results
The benchmarking libraries output measurements in
JSON, along with profiling traces to a directory on
the Android device after each benchmark run. Macrobenchmark library outputs
multiple perfetto trace files: one per measured iteration of each
MacrobenchmarkRule.measureRepeated
loop. Microbenchmark, however, creates
just one trace file for all the iterations of each
BenchmarkRule.measureRepeated
. Profiling trace files are also output to
this same directory.
Save and locate the files
If you run the benchmarks with Gradle, these files are automatically copied to
your host computer's outputs directory under
build/outputs/connected_android_test_additional_output/debugAndroidTest/connected/
.
If running directly with adb
command, you need to pull them
manually. By default, the reports are saved on-device in the media directory of
the tested app's external storage. For convenience, the library prints the path
of the file into Logcat. Note, that the output folder may be different depending
on which Android version the benchmarks are running on.
Benchmark: writing results to /storage/emulated/0/Android/media/com.example.macrobenchmark/com.example.macrobenchmark-benchmarkData.json
You can also configure the location where benchmark reports are saved on the
device using the instrumentation argument additionalTestOutputDir
. This
folder must be writable by your application.
adb shell am instrument -w -e additionalTestOutputDir /sdcard/Download/ com.example.benchmark/androidx.benchmark.junit4.AndroidBenchmarkRunner
On Android 10 (API level 29) and higher, your app's tests run in a storage
sandbox by default which prevents your app from accessing files outside of the
app-specific directory. In order to be able to save into some global directory
(such as /sdcard/Download
), you can pass the following instrumentation
argument:
-e no-isolated-storage true
You also must explicitly allow legacy storage options in your benchmark's manifest:
<application android:requestLegacyExternalStorage="true" ... >
For more information, see Temporarily opt-out of scoped storage.
Retrieve the files
In order to get the generated files from the device, you can use adb pull
command, which pulls the specified file into the current directory on your host.
adb pull /storage/emulated/0/Android/media/com.example.macrobenchmark/com.example.macrobenchmark-benchmarkData.json
For retrieving all of the benchmarkData
from a specified folder, check the
following snippet:
# The following command pulls all files ending in -benchmarkData.json from the directory
# hierarchy starting at the root /storage/emulated/0/Android.
adb shell find /sdcard/Download -name "*-benchmarkData.json" | tr -d '\r' | xargs -n1 adb pull
Note, that the trace files (.trace
or .perfetto-trace
) are saved in the same
folder as the benchmarkData.json
, thus you can collect them in the same way.
Benchmark data example
The benchmark libraries generate JSON files containing information about the device it was running the benchmarks on and the actual benchmarks it ran. The following snippet represents the generated JSON file.
{
"context": {
"build": {
"brand": "google",
"device": "blueline",
"fingerprint": "google/blueline/blueline:12/SP1A.210812.015/7679548:user/release-keys",
"model": "Pixel 3",
"version": {
"sdk": 31
}
},
"cpuCoreCount": 8,
"cpuLocked": false,
"cpuMaxFreqHz": 2803200000,
"memTotalBytes": 3753299968,
"sustainedPerformanceModeEnabled": false
},
"benchmarks": [
{
"name": "startup",
"params": {},
"className": "com.example.macrobenchmark.startup.SampleStartupBenchmark",
"totalRunTimeNs": 4975598256,
"metrics": {
"timeToInitialDisplayMs": {
"minimum": 347.881076,
"maximum": 347.881076,
"median": 347.881076,
"runs": [
347.881076
]
}
},
"sampledMetrics": {},
"warmupIterations": 0,
"repeatIterations": 3,
"thermalThrottleSleepSeconds": 0
}
]
}
Additional resources
- For guidance in how to detect performance regressions, see Fighting Regressions with Benchmarks in CI.
- To see how to setup Github Actions with Firebase Test Lab, see Setting up Jetpack Macrobenchmarks for CI