196 lines
6.3 KiB
Text
196 lines
6.3 KiB
Text
|
Git performance tests
|
||
|
=====================
|
||
|
|
||
|
This directory holds performance testing scripts for git tools. The
|
||
|
first part of this document describes the various ways in which you
|
||
|
can run them.
|
||
|
|
||
|
When fixing the tools or adding enhancements, you are strongly
|
||
|
encouraged to add tests in this directory to cover what you are
|
||
|
trying to fix or enhance. The later part of this short document
|
||
|
describes how your test scripts should be organized.
|
||
|
|
||
|
|
||
|
Running Tests
|
||
|
-------------
|
||
|
|
||
|
The easiest way to run tests is to say "make". This runs all
|
||
|
the tests on the current git repository.
|
||
|
|
||
|
=== Running 2 tests in this tree ===
|
||
|
[...]
|
||
|
Test this tree
|
||
|
---------------------------------------------------------
|
||
|
0001.1: rev-list --all 0.54(0.51+0.02)
|
||
|
0001.2: rev-list --all --objects 6.14(5.99+0.11)
|
||
|
7810.1: grep worktree, cheap regex 0.16(0.16+0.35)
|
||
|
7810.2: grep worktree, expensive regex 7.90(29.75+0.37)
|
||
|
7810.3: grep --cached, cheap regex 3.07(3.02+0.25)
|
||
|
7810.4: grep --cached, expensive regex 9.39(30.57+0.24)
|
||
|
|
||
|
You can compare multiple repositories and even git revisions with the
|
||
|
'run' script:
|
||
|
|
||
|
$ ./run . origin/next /path/to/git-tree p0001-rev-list.sh
|
||
|
|
||
|
where . stands for the current git tree. The full invocation is
|
||
|
|
||
|
./run [<revision|directory>...] [--] [<test-script>...]
|
||
|
|
||
|
A '.' argument is implied if you do not pass any other
|
||
|
revisions/directories.
|
||
|
|
||
|
You can also manually test this or another git build tree, and then
|
||
|
call the aggregation script to summarize the results:
|
||
|
|
||
|
$ ./p0001-rev-list.sh
|
||
|
[...]
|
||
|
$ ./run /path/to/other/git -- ./p0001-rev-list.sh
|
||
|
[...]
|
||
|
$ ./aggregate.perl . /path/to/other/git ./p0001-rev-list.sh
|
||
|
|
||
|
aggregate.perl has the same invocation as 'run', it just does not run
|
||
|
anything beforehand.
|
||
|
|
||
|
You can set the following variables (also in your config.mak):
|
||
|
|
||
|
GIT_PERF_REPEAT_COUNT
|
||
|
Number of times a test should be repeated for best-of-N
|
||
|
measurements. Defaults to 3.
|
||
|
|
||
|
GIT_PERF_MAKE_OPTS
|
||
|
Options to use when automatically building a git tree for
|
||
|
performance testing. E.g., -j6 would be useful. Passed
|
||
|
directly to make as "make $GIT_PERF_MAKE_OPTS".
|
||
|
|
||
|
GIT_PERF_MAKE_COMMAND
|
||
|
An arbitrary command that'll be run in place of the make
|
||
|
command, if set the GIT_PERF_MAKE_OPTS variable is
|
||
|
ignored. Useful in cases where source tree changes might
|
||
|
require issuing a different make command to different
|
||
|
revisions.
|
||
|
|
||
|
This can be (ab)used to monkeypatch or otherwise change the
|
||
|
tree about to be built. Note that the build directory can be
|
||
|
re-used for subsequent runs so the make command might get
|
||
|
executed multiple times on the same tree, but don't count on
|
||
|
any of that, that's an implementation detail that might change
|
||
|
in the future.
|
||
|
|
||
|
GIT_PERF_REPO
|
||
|
GIT_PERF_LARGE_REPO
|
||
|
Repositories to copy for the performance tests. The normal
|
||
|
repo should be at least git.git size. The large repo should
|
||
|
probably be about linux.git size for optimal results.
|
||
|
Both default to the git.git you are running from.
|
||
|
|
||
|
You can also pass the options taken by ordinary git tests; the most
|
||
|
useful one is:
|
||
|
|
||
|
--root=<directory>::
|
||
|
Create "trash" directories used to store all temporary data during
|
||
|
testing under <directory>, instead of the t/ directory.
|
||
|
Using this option with a RAM-based filesystem (such as tmpfs)
|
||
|
can massively speed up the test suite.
|
||
|
|
||
|
|
||
|
Naming Tests
|
||
|
------------
|
||
|
|
||
|
The performance test files are named as:
|
||
|
|
||
|
pNNNN-commandname-details.sh
|
||
|
|
||
|
where N is a decimal digit. The same conventions for choosing NNNN as
|
||
|
for normal tests apply.
|
||
|
|
||
|
|
||
|
Writing Tests
|
||
|
-------------
|
||
|
|
||
|
The perf script starts much like a normal test script, except it
|
||
|
sources perf-lib.sh:
|
||
|
|
||
|
#!/bin/sh
|
||
|
#
|
||
|
# Copyright (c) 2005 Junio C Hamano
|
||
|
#
|
||
|
|
||
|
test_description='xxx performance test'
|
||
|
. ./perf-lib.sh
|
||
|
|
||
|
After that you will want to use some of the following:
|
||
|
|
||
|
test_perf_fresh_repo # sets up an empty repository
|
||
|
test_perf_default_repo # sets up a "normal" repository
|
||
|
test_perf_large_repo # sets up a "large" repository
|
||
|
|
||
|
test_perf_default_repo sub # ditto, in a subdir "sub"
|
||
|
|
||
|
test_checkout_worktree # if you need the worktree too
|
||
|
|
||
|
At least one of the first two is required!
|
||
|
|
||
|
You can use test_expect_success as usual. In both test_expect_success
|
||
|
and in test_perf, running "git" points to the version that is being
|
||
|
perf-tested. The $MODERN_GIT variable points to the git wrapper for the
|
||
|
currently checked-out version (i.e., the one that matches the t/perf
|
||
|
scripts you are running). This is useful if your setup uses commands
|
||
|
that only work with newer versions of git than what you might want to
|
||
|
test (but obviously your new commands must still create a state that can
|
||
|
be used by the older version of git you are testing).
|
||
|
|
||
|
For actual performance tests, use
|
||
|
|
||
|
test_perf 'descriptive string' '
|
||
|
command1 &&
|
||
|
command2
|
||
|
'
|
||
|
|
||
|
test_perf spawns a subshell, for lack of better options. This means
|
||
|
that
|
||
|
|
||
|
* you _must_ export all variables that you need in the subshell
|
||
|
|
||
|
* you _must_ flag all variables that you want to persist from the
|
||
|
subshell with 'test_export':
|
||
|
|
||
|
test_perf 'descriptive string' '
|
||
|
foo=$(git rev-parse HEAD) &&
|
||
|
test_export foo
|
||
|
'
|
||
|
|
||
|
The so-exported variables are automatically marked for export in the
|
||
|
shell executing the perf test. For your convenience, test_export is
|
||
|
the same as export in the main shell.
|
||
|
|
||
|
This feature relies on a bit of magic using 'set' and 'source'.
|
||
|
While we have tried to make sure that it can cope with embedded
|
||
|
whitespace and other special characters, it will not work with
|
||
|
multi-line data.
|
||
|
|
||
|
Rather than tracking the performance by run-time as `test_perf` does, you
|
||
|
may also track output size by using `test_size`. The stdout of the
|
||
|
function should be a single numeric value, which will be captured and
|
||
|
shown in the aggregated output. For example:
|
||
|
|
||
|
test_perf 'time foo' '
|
||
|
./foo >foo.out
|
||
|
'
|
||
|
|
||
|
test_size 'output size'
|
||
|
wc -c <foo.out
|
||
|
'
|
||
|
|
||
|
might produce output like:
|
||
|
|
||
|
Test origin HEAD
|
||
|
-------------------------------------------------------------
|
||
|
1234.1 time foo 0.37(0.79+0.02) 0.26(0.51+0.02) -29.7%
|
||
|
1234.2 output size 4.3M 3.6M -14.7%
|
||
|
|
||
|
The item being measured (and its units) is up to the test; the context
|
||
|
and the test title should make it clear to the user whether bigger or
|
||
|
smaller numbers are better. Unlike test_perf, the test code will only be
|
||
|
run once, since output sizes tend to be more deterministic than timings.
|