Performance testing matchers for RSpec to set expectations on speed, resources usage and scalability.
RSpec::Benchmark is powered by:
- benchmark-perf for measuring execution time and iterations per second.
- benchmark-trend for estimating computation complexity.
- benchmark-malloc for measuring object and memory allocations.
Integration and unit tests ensure that changing code maintains expected functionality. What is not guaranteed is the code changes impact on library performance. It is easy to refactor your way out of fast to slow code.
If you are new to performance testing you may find Caveats section helpful.
Add this line to your application's Gemfile:
gem 'rspec-benchmark'
And then execute:
$ bundle
Or install it yourself as:
$ gem install rspec-benchmark
For matchers to be available globally, in spec_helper.rb
do:
require 'rspec-benchmark'
RSpec.configure do |config|
config.include RSpec::Benchmark::Matchers
end
This will add the following matchers:
perform_under
to see how fast your code runsperform_at_least
to see how many iteration per second your code can doperform_(faster|slower)_than
to compare implementationsperform_(constant|linear|logarithmic|power|exponential)
to see how your code scales with timeperform_allocation
to limit object and memory allocations
These will help you express expected performance benchmark for an evaluated code.
Alternatively, you can add matchers for particular example:
RSpec.describe "Performance testing" do
include RSpec::Benchmark::Matchers
end
Then you're good to start setting performance expectations:
expect {
...
}.to perform_under(6).ms
The perform_under
matcher answers the question of how long does it take to perform a given block of code on average. The measurements are taken executing the block of code in a child process for accurate CPU times.
expect { ... }.to perform_under(0.01).sec
All measurements are assumed to be expressed as seconds. However, you can also provide time in ms
, us
and ns
. The equivalent example in ms
would be:
expect { ... }.to perform_under(10).ms
expect { ... }.to perform_under(10000).us
By default the above code will be sampled only once but you can change this by using the sample
matcher like so:
expect { ... }.to perform_under(0.01).sample(10) # repeats measurements 10 times
For extra expressiveness you can use times
:
expect { ... }.to perform_under(0.01).sample(10).times
You can also use warmup
matcher that can run your code before the actual samples are taken to reduce erratic execution times.
For example, you can execute code twice before you take 10 actual measurements:
expect { ... }.to perform_under(0.01).sec.warmup(2).times.sample(10).times
The perform_at_least
matcher allows you to establish performance benchmark of how many iterations per second a given block of code should perform. For example, to expect a given code to perform at least 10K iterations per second do:
expect { ... }.to perform_at_least(10000).ips
The ips
part is optional but its usage clarifies the intent.
The performance timing of this matcher can be tweaked using the within
and warmup
matchers. These are expressed as seconds.
By default within
matcher is set to 0.2
second and warmup
matcher to 0.1
respectively. To change how long measurements are taken, for example, to double the time amount do:
expect { ... }.to perform_at_least(10000).within(0.4).warmup(0.2).ips
The higher values for within
and warmup
the more accurate average readings and more stable tests at the cost of longer test suite overall runtime.
The perform_faster_than
and perform_slower_than
matchers allow you to test performance of your code compared with other. For example:
expect { ... }.to perform_faster_than { ... }
expect { ... }.to perform_slower_than { ... }
And if you want to compare how much faster or slower your code is do:
expect { ... }.to perform_faster_than { ... }.once
expect { ... }.to perform_faster_than { ... }.twice
expect { ... }.to perform_faster_than { ... }.exactly(5).times
expect { ... }.to perform_faster_than { ... }.at_least(5).times
expect { ... }.to perform_faster_than { ... }.at_most(5).times
expect { ... }.to perform_slower_than { ... }.once
expect { ... }.to perform_slower_than { ... }.twice
expect { ... }.to perform_slower_than { ... }.at_least(5).times
expect { ... }.to perform_slower_than { ... }.at_most(5).times
expect { ... }.to perform_slower_than { ... }.exactly(5).times
The times
part is also optional.
The performance timing of each matcher can be tweaked using the within
and warmup
matchers. These are expressed as seconds. By default within
matcher is set to 0.2
and warmup
matcher to 0.1
second respectively. To change these matchers values do:
expect { ... }.to perform_faster_than { ... }.within(0.4).warmup(0.2)
The higher values for within
and warmup
the more accurate average readings and more stable tests at the cost of longer test suite overall runtime.
The perform_constant
, perform_logarithmic
, perform_linear
, perform_power
and perform_exponential
matchers are useful for estimating the asymptotic behaviour of a given block of code. The most basic way to use the expectations to test how your code scales is to use the matchers:
expect { ... }.to perform_constant
expect { ... }.to perform_logarithmic/perform_log
expect { ... }.to perform_linear
expect { ... }.to perform_power
expect { ... }.to perform_exponential/perform_exp
To test performance in terms of computation complexity you can follow the algorithm:
- Choose a method to profile.
- Choose workloads for the method.
- Describe workloads with input features.
- Assert the performance in terms of Big-O notation.
Often, before expectation can be set you need to setup some workloads. To create a range of inputs use the bench_range
helper method.
For example, to create a power range of inputs from 8
to 100_000
do:
sizes = bench_range(8, 100_000) # => [8, 64, 512, 4096, 32768, 100000]
Then you can use the sizes to create test data, for example to check Ruby's max
performance create array of number arrays.
number_arrays = sizes.map { |n| Array.new(n) { rand(n) } }
Using in_range
matcher you can inform the expectation about the inputs. Each range value together with its index will be yielded as arguments to the evaluated block.
You can either specify the range limits:
expect { |n, i|
number_arrays[i].max
}.to perform_linear.in_range(8, 100_000)
Or use previously generated sizes
array:
expect { |n, i|
number_arrays[i].max
}.to perform_linear.in_range(sizes)
This example will generate and yield input n
and index i
pairs [8, 0]
, [64, 1]
, [512, 2]
, [4K, 3]
, [32K, 4]
and [100K, 5]
respectively.
By default the range will be generated using ratio of 8
. You can change this using ratio
matcher:
expect { |n, i|
number_arrays[i].max
}.to perform_linear.in_range(8, 100_000).ratio(2)
The performance measurements for a code block are taken only once per range input. You can increase the stability of your performance test by using the sample
matcher. For example, to repeat measurements 100 times for each range input do:
expect { |n, i|
number_arrays[i].max
}.to perform_linear.in_range(8, 100_000).ratio(2).sample(100).times
The overall quality of the performance trend is assessed using a threshold value where 0
means a poor fit and 1
a perfect fit. By default this value is configured to 0.9
as a 'good enough' threshold. To change this use threshold
matcher:
expect { |n, i|
number_arrays[i].max
}.to perform_linear.in_range(8, 100_000).threshold(0.95)
The perform_allocation
matcher checks how much memory or objects have been allocated during a piece of Ruby code execution.
By default the matcher verifies the number of object allocations. The specified number serves as the upper limit of allocations, so your tests won't become brittle as different Ruby versions change internally how many objects are allocated for some operations.
Note that you can also check for memory allocation using the bytes
matcher.
To check number of objects allocated do:
expect {
["foo", "bar", "baz"].sort[1]
}.to perform_allocation(3)
You can also be more granular with your object allocations and specify which object types you're interested in:
expect {
_a = [Object.new]
_b = {Object.new => 'foo'}
}.to perform_allocation({Array => 1, Object => 2}).objects
And you can also check how many objects are left when expectation finishes to ensure that GC
is able to collect them.
expect {
["foo", "bar", "baz"].sort[1]
}.to perform_allocation(3).and_retain(3)
You can also set expectations on the memory size. In this case the memory size will serve as upper limit for the expectation:
expect {
_a = [Object.new]
_b = {Object.new => 'foo'}
}.to perform_allocation({Array => 40, Hash => 384, Object => 80}).bytes
All the matchers can be used in compound expressions via and/or
. For example, if you wish to check if a computation performs under certain time boundary and iterates at least a given number do:
expect {
...
}.to perform_under(6).ms and perform_at_least(10000).ips
By default the following configuration is used:
RSpec::Benchmark.configure do |config|
config.run_in_subprocess = false
config.disable_gc = false
end
By default all tests are run with GC
enabled. We want to measure real performance or Ruby code. However, disabling GC
may lead to much quicker test execution. You can change this setting:
RSpec::Benchmark.configure do |config|
config.disable_gc = true
end
The perform_under
matcher can run all the measurements in the subprocess. This will increase isolation from other processes activity. However, the rspec-rails
gem runs all tests in transactions. Unfortunately, when running tests in child process, database connections are used from connection pool and no data can be accessed. This is only a problem when running specs in Rails. Any other Ruby project can run specs using subprocesses. To enable this behaviour do:
RSpec::Benchmark.configure do |config|
config.run_in_subprocess = true
end
The perform_under
and computational complexity matchers allow to specify how many times to repeat measurements. You configure it globally for all matchers using the :samples
option which defaults to 1
:
RSpec::Benchmark.configure do |config|
config.samples = 10
end
The perform_at_least
matcher uses the :format
option to format the number of iterations when a failure message gets displayed. By default, the :human
values is used to make numbers more readable. For example, the 12300 i/s
gets turned into 12.3k i/s
. If you rather have an exact numbers presented do:
RSpec::Benchmark.configure do |config|
config.format = :raw
end
Usually performance tests are best left for CI or occasional runs that do not affect TDD/BDD cycle.
To achieve isolation you can use RSpec filters to exclude performance tests from regular runs. For example, in spec_helper
:
RSpec.config do |config|
config.filter_run_excluding perf: true
end
And then in your example group do:
RSpec.describe ..., :perf do
...
end
Then you can run groups or examples tagged with perf
:
rspec --tag perf
Another option is to simply isolate the performance specs in separate directory such as spec/performance/...
and add custom rake task to run them.
When writing performance tests things to be mindful are:
- The tests may potentially be flaky thus its best to use sensible boundaries:
- too strict boundaries may cause false positives, making tests fail
- too relaxed boundaries may also lead to false positives missing actual performance regressions
- Generally performance tests will be slow, but you may try to avoid unnecessarily slow tests by choosing smaller maximum value for sampling
If you have any other observations please share them!
- Fork it ( https://github.com/piotrmurach/rspec-benchmark/fork )
- Create your feature branch (
git checkout -b my-new-feature
) - Commit your changes (
git commit -am 'Add some feature'
) - Push to the branch (
git push origin my-new-feature
) - Create a new Pull Request
Everyone interacting in the Strings project’s codebases, issue trackers, chat rooms and mailing lists is expected to follow the code of conduct.
Copyright (c) 2016 Piotr Murach. See LICENSE for further details.