1% Testing the JDK 2 3## Using "make test" (the run-test framework) 4 5This new way of running tests is developer-centric. It assumes that you have 6built a JDK locally and want to test it. Running common test targets is simple, 7and more complex ad-hoc combination of tests is possible. The user interface is 8forgiving, and clearly report errors it cannot resolve. 9 10The main target `test` uses the jdk-image as the tested product. There is 11also an alternate target `exploded-test` that uses the exploded image 12instead. Not all tests will run successfully on the exploded image, but using 13this target can greatly improve rebuild times for certain workflows. 14 15Previously, `make test` was used invoke an old system for running test, and 16`make run-test` was used for the new test framework. For backward compatibility 17with scripts and muscle memory, `run-test` (and variants like 18`exploded-run-test` or `run-test-tier1`) are kept as aliases. The old system 19can still be accessed for some time using `cd test && make`. 20 21Some example command-lines: 22 23 $ make test-tier1 24 $ make test-jdk_lang JTREG="JOBS=8" 25 $ make test TEST=jdk_lang 26 $ make test-only TEST="gtest:LogTagSet gtest:LogTagSetDescriptions" GTEST="REPEAT=-1" 27 $ make test TEST="hotspot:hotspot_gc" JTREG="JOBS=1;TIMEOUT=8;VM_OPTIONS=-XshowSettings -Xlog:gc+ref=debug" 28 $ make test TEST="jtreg:test/hotspot:hotspot_gc test/hotspot/jtreg/native_sanity/JniVersion.java" 29 $ make test TEST="micro:java.lang.reflect" MICRO="FORK=1;WARMUP_ITER=2" 30 $ make exploded-test TEST=tier2 31 32### Configuration 33 34To be able to run JTReg tests, `configure` needs to know where to find the 35JTReg test framework. If it is not picked up automatically by configure, use 36the `--with-jtreg=<path to jtreg home>` option to point to the JTReg framework. 37Note that this option should point to the JTReg home, i.e. the top directory, 38containing `lib/jtreg.jar` etc. (An alternative is to set the `JT_HOME` 39environment variable to point to the JTReg home before running `configure`.) 40 41To be able to run microbenchmarks, `configure` needs to know where to find 42the JMH dependency. Use `--with-jmh=<path to JMH jars>` to point to a directory 43containing the core JMH and transitive dependencies. The recommended dependencies 44can be retrieved by running `sh make/devkit/createJMHBundle.sh`, after which 45`--with-jmh=build/jmh/jars` should work. 46 47## Test selection 48 49All functionality is available using the `test` make target. In this use case, 50the test or tests to be executed is controlled using the `TEST` variable. To 51speed up subsequent test runs with no source code changes, `test-only` can be 52used instead, which do not depend on the source and test image build. 53 54For some common top-level tests, direct make targets have been generated. This 55includes all JTReg test groups, the hotspot gtest, and custom tests (if 56present). This means that `make test-tier1` is equivalent to `make test 57TEST="tier1"`, but the latter is more tab-completion friendly. For more complex 58test runs, the `test TEST="x"` solution needs to be used. 59 60The test specifications given in `TEST` is parsed into fully qualified test 61descriptors, which clearly and unambigously show which tests will be run. As an 62example, `:tier1` will expand to `jtreg:$(TOPDIR)/test/hotspot/jtreg:tier1 63jtreg:$(TOPDIR)/test/jdk:tier1 jtreg:$(TOPDIR)/test/langtools:tier1 64jtreg:$(TOPDIR)/test/nashorn:tier1 jtreg:$(TOPDIR)/test/jaxp:tier1`. You can 65always submit a list of fully qualified test descriptors in the `TEST` variable 66if you want to shortcut the parser. 67 68### JTReg 69 70JTReg tests can be selected either by picking a JTReg test group, or a selection 71of files or directories containing JTReg tests. 72 73JTReg test groups can be specified either without a test root, e.g. `:tier1` 74(or `tier1`, the initial colon is optional), or with, e.g. `hotspot:tier1`, 75`test/jdk:jdk_util` or `$(TOPDIR)/test/hotspot/jtreg:hotspot_all`. The test 76root can be specified either as an absolute path, or a path relative to the 77JDK top directory, or the `test` directory. For simplicity, the hotspot 78JTReg test root, which really is `hotspot/jtreg` can be abbreviated as 79just `hotspot`. 80 81When specified without a test root, all matching groups from all test roots 82will be added. Otherwise, only the group from the specified test root will be 83added. 84 85Individual JTReg tests or directories containing JTReg tests can also be 86specified, like `test/hotspot/jtreg/native_sanity/JniVersion.java` or 87`hotspot/jtreg/native_sanity`. Just like for test root selection, you can 88either specify an absolute path (which can even point to JTReg tests outside 89the source tree), or a path relative to either the JDK top directory or the 90`test` directory. `hotspot` can be used as an alias for `hotspot/jtreg` here as 91well. 92 93As long as the test groups or test paths can be uniquely resolved, you do not 94need to enter the `jtreg:` prefix. If this is not possible, or if you want to 95use a fully qualified test descriptor, add `jtreg:`, e.g. 96`jtreg:test/hotspot/jtreg/native_sanity`. 97 98### Gtest 99 100Since the Hotspot Gtest suite is so quick, the default is to run all tests. 101This is specified by just `gtest`, or as a fully qualified test descriptor 102`gtest:all`. 103 104If you want, you can single out an individual test or a group of tests, for 105instance `gtest:LogDecorations` or `gtest:LogDecorations.level_test_vm`. This 106can be particularly useful if you want to run a shaky test repeatedly. 107 108For Gtest, there is a separate test suite for each JVM variant. The JVM variant 109is defined by adding `/<variant>` to the test descriptor, e.g. 110`gtest:Log/client`. If you specify no variant, gtest will run once for each JVM 111variant present (e.g. server, client). So if you only have the server JVM 112present, then `gtest:all` will be equivalent to `gtest:all/server`. 113 114### Microbenchmarks 115 116Which microbenchmarks to run is selected using a regular expression 117following the `micro:` test descriptor, e.g., `micro:java.lang.reflect`. This 118delegates the test selection to JMH, meaning package name, class name and even 119benchmark method names can be used to select tests. 120 121Using special characters like `|` in the regular expression is possible, but 122needs to be escaped multiple times: `micro:ArrayCopy\\\\\|reflect`. 123 124### Special tests 125 126A handful of odd tests that are not covered by any other testing framework are 127accessible using the `special:` test descriptor. Currently, this includes 128`failure-handler` and `make`. 129 130 * Failure handler testing is run using `special:failure-handler` or just 131 `failure-handler` as test descriptor. 132 133 * Tests for the build system, including both makefiles and related 134 functionality, is run using `special:make` or just `make` as test 135 descriptor. This is equivalent to `special:make:all`. 136 137 A specific make test can be run by supplying it as argument, e.g. 138 `special:make:idea`. As a special syntax, this can also be expressed as 139 `make-idea`, which allows for command lines as `make test-make-idea`. 140 141## Test results and summary 142 143At the end of the test run, a summary of all tests run will be presented. This 144will have a consistent look, regardless of what test suites were used. This is 145a sample summary: 146 147 ============================== 148 Test summary 149 ============================== 150 TEST TOTAL PASS FAIL ERROR 151 >> jtreg:jdk/test:tier1 1867 1865 2 0 << 152 jtreg:langtools/test:tier1 4711 4711 0 0 153 jtreg:nashorn/test:tier1 133 133 0 0 154 ============================== 155 TEST FAILURE 156 157Tests where the number of TOTAL tests does not equal the number of PASSed tests 158will be considered a test failure. These are marked with the `>> ... <<` marker 159for easy identification. 160 161The classification of non-passed tests differs a bit between test suites. In 162the summary, ERROR is used as a catch-all for tests that neither passed nor are 163classified as failed by the framework. This might indicate test framework 164error, timeout or other problems. 165 166In case of test failures, `make test` will exit with a non-zero exit value. 167 168All tests have their result stored in `build/$BUILD/test-results/$TEST_ID`, 169where TEST_ID is a path-safe conversion from the fully qualified test 170descriptor, e.g. for `jtreg:jdk/test:tier1` the TEST_ID is 171`jtreg_jdk_test_tier1`. This path is also printed in the log at the end of the 172test run. 173 174Additional work data is stored in `build/$BUILD/test-support/$TEST_ID`. For 175some frameworks, this directory might contain information that is useful in 176determining the cause of a failed test. 177 178## Test suite control 179 180It is possible to control various aspects of the test suites using make control 181variables. 182 183These variables use a keyword=value approach to allow multiple values to be 184set. So, for instance, `JTREG="JOBS=1;TIMEOUT=8"` will set the JTReg 185concurrency level to 1 and the timeout factor to 8. This is equivalent to 186setting `JTREG_JOBS=1 JTREG_TIMEOUT=8`, but using the keyword format means that 187the `JTREG` variable is parsed and verified for correctness, so 188`JTREG="TMIEOUT=8"` would give an error, while `JTREG_TMIEOUT=8` would just 189pass unnoticed. 190 191To separate multiple keyword=value pairs, use `;` (semicolon). Since the shell 192normally eats `;`, the recommended usage is to write the assignment inside 193qoutes, e.g. `JTREG="...;..."`. This will also make sure spaces are preserved, 194as in `JTREG="VM_OPTIONS=-XshowSettings -Xlog:gc+ref=debug"`. 195 196(Other ways are possible, e.g. using backslash: `JTREG=JOBS=1\;TIMEOUT=8`. 197Also, as a special technique, the string `%20` will be replaced with space for 198certain options, e.g. `JTREG=VM_OPTIONS=-XshowSettings%20-Xlog:gc+ref=debug`. 199This can be useful if you have layers of scripts and have trouble getting 200proper quoting of command line arguments through.) 201 202As far as possible, the names of the keywords have been standardized between 203test suites. 204 205### JTReg keywords 206 207#### JOBS 208The test concurrency (`-concurrency`). 209 210Defaults to TEST_JOBS (if set by `--with-test-jobs=`), otherwise it defaults to 211JOBS, except for Hotspot, where the default is *number of CPU cores/2*, but 212never more than 12. 213 214#### TIMEOUT 215The timeout factor (`-timeoutFactor`). 216 217Defaults to 4. 218 219#### TEST_MODE 220The test mode (`-agentvm`, `-samevm` or `-othervm`). 221 222Defaults to `-agentvm`. 223 224#### ASSERT 225Enable asserts (`-ea -esa`, or none). 226 227Set to `true` or `false`. If true, adds `-ea -esa`. Defaults to true, except 228for hotspot. 229 230#### VERBOSE 231The verbosity level (`-verbose`). 232 233Defaults to `fail,error,summary`. 234 235#### RETAIN 236What test data to retain (`-retain`). 237 238Defaults to `fail,error`. 239 240#### MAX_MEM 241Limit memory consumption (`-Xmx` and `-vmoption:-Xmx`, or none). 242 243Limit memory consumption for JTReg test framework and VM under test. Set to 0 244to disable the limits. 245 246Defaults to 512m, except for hotspot, where it defaults to 0 (no limit). 247 248#### OPTIONS 249Additional options to the JTReg test framework. 250 251Use `JTREG="OPTIONS=--help all"` to see all available JTReg options. 252 253#### JAVA_OPTIONS 254Additional Java options to JTReg (`-javaoption`). 255 256#### VM_OPTIONS 257Additional VM options to JTReg (`-vmoption`). 258 259### Gtest keywords 260 261#### REPEAT 262The number of times to repeat the tests (`--gtest_repeat`). 263 264Default is 1. Set to -1 to repeat indefinitely. This can be especially useful 265combined with `OPTIONS=--gtest_break_on_failure` to reproduce an intermittent 266problem. 267 268#### OPTIONS 269Additional options to the Gtest test framework. 270 271Use `GTEST="OPTIONS=--help"` to see all available Gtest options. 272 273### Microbenchmark keywords 274 275#### FORK 276Override the number of benchmark forks to spawn. Same as specifying `-f <num>`. 277 278#### ITER 279Number of measurement iterations per fork. Same as specifying `-i <num>`. 280 281#### TIME 282Amount of time to spend in each measurement iteration, in seconds. Same as 283specifying `-r <num>` 284 285#### WARMUP_ITER 286Number of warmup iterations to run before the measurement phase in each fork. 287Same as specifying `-wi <num>`. 288 289#### WARMUP_TIME 290Amount of time to spend in each warmup iteration. Same as specifying `-w <num>`. 291 292#### RESULTS_FORMAT 293Specify to have the test run save a log of the values. Accepts the same values 294as `-rff`, i.e., `text`, `csv`, `scsv`, `json`, or `latex`. 295 296#### VM_OPTIONS 297Additional VM arguments to provide to forked off VMs. Same as `-jvmArgs <args>` 298 299#### OPTIONS 300Additional arguments to send to JMH. 301 302--- 303# Override some definitions in the global css file that are not optimal for 304# this document. 305header-includes: 306 - '<style type="text/css">pre, code, tt { color: #1d6ae5; }</style>' 307--- 308