Skip to content

Commit f832ded

Browse files
committed
Build: Remove shadowing from benchmarks (#32475)
Removes shadowing from the benchmarks. It isn't *strictly* needed. We do have to rework the documentation on how to run the benchmark, but it still seems to work if you run everything through gradle.
1 parent 8bb8b93 commit f832ded

File tree

4 files changed

+45
-59
lines changed

4 files changed

+45
-59
lines changed

benchmarks/README.md

+20-17
Original file line numberDiff line numberDiff line change
@@ -4,36 +4,39 @@ This directory contains the microbenchmark suite of Elasticsearch. It relies on
44

55
## Purpose
66

7-
We do not want to microbenchmark everything but the kitchen sink and should typically rely on our
8-
[macrobenchmarks](https://elasticsearch-benchmarks.elastic.co/app/kibana#/dashboard/Nightly-Benchmark-Overview) with
9-
[Rally](http://github.com/elastic/rally). Microbenchmarks are intended to spot performance regressions in performance-critical components.
7+
We do not want to microbenchmark everything but the kitchen sink and should typically rely on our
8+
[macrobenchmarks](https://elasticsearch-benchmarks.elastic.co/app/kibana#/dashboard/Nightly-Benchmark-Overview) with
9+
[Rally](http://github.com/elastic/rally). Microbenchmarks are intended to spot performance regressions in performance-critical components.
1010
The microbenchmark suite is also handy for ad-hoc microbenchmarks but please remove them again before merging your PR.
1111

1212
## Getting Started
1313

14-
Just run `gradle :benchmarks:jmh` from the project root directory. It will build all microbenchmarks, execute them and print the result.
14+
Just run `gradlew -p benchmarks run` from the project root
15+
directory. It will build all microbenchmarks, execute them and print
16+
the result.
1517

1618
## Running Microbenchmarks
1719

18-
Benchmarks are always run via Gradle with `gradle :benchmarks:jmh`.
19-
20-
Running via an IDE is not supported as the results are meaningless (we have no control over the JVM running the benchmarks).
20+
Running via an IDE is not supported as the results are meaningless
21+
because we have no control over the JVM running the benchmarks.
2122

22-
If you want to run a specific benchmark class, e.g. `org.elasticsearch.benchmark.MySampleBenchmark` or have special requirements
23-
generate the uberjar with `gradle :benchmarks:jmhJar` and run it directly with:
23+
If you want to run a specific benchmark class like, say,
24+
`MemoryStatsBenchmark`, you can use `--args`:
2425

2526
```
26-
java -jar benchmarks/build/distributions/elasticsearch-benchmarks-*.jar
27+
gradlew -p benchmarks run --args ' AllocationBenchmark'
2728
```
2829

29-
JMH supports lots of command line parameters. Add `-h` to the command above to see the available command line options.
30+
Everything in the `'` gets sent on the command line to JMH. The leading ` `
31+
inside the `'`s is important. Without it parameters are sometimes sent to
32+
gradle.
3033

3134
## Adding Microbenchmarks
3235

33-
Before adding a new microbenchmark, make yourself familiar with the JMH API. You can check our existing microbenchmarks and also the
36+
Before adding a new microbenchmark, make yourself familiar with the JMH API. You can check our existing microbenchmarks and also the
3437
[JMH samples](http://hg.openjdk.java.net/code-tools/jmh/file/tip/jmh-samples/src/main/java/org/openjdk/jmh/samples/).
3538

36-
In contrast to tests, the actual name of the benchmark class is not relevant to JMH. However, stick to the naming convention and
39+
In contrast to tests, the actual name of the benchmark class is not relevant to JMH. However, stick to the naming convention and
3740
end the class name of a benchmark with `Benchmark`. To have JMH execute a benchmark, annotate the respective methods with `@Benchmark`.
3841

3942
## Tips and Best Practices
@@ -42,15 +45,15 @@ To get realistic results, you should exercise care when running benchmarks. Here
4245

4346
### Do
4447

45-
* Ensure that the system executing your microbenchmarks has as little load as possible. Shutdown every process that can cause unnecessary
48+
* Ensure that the system executing your microbenchmarks has as little load as possible. Shutdown every process that can cause unnecessary
4649
runtime jitter. Watch the `Error` column in the benchmark results to see the run-to-run variance.
4750
* Ensure to run enough warmup iterations to get the benchmark into a stable state. If you are unsure, don't change the defaults.
4851
* Avoid CPU migrations by pinning your benchmarks to specific CPU cores. On Linux you can use `taskset`.
49-
* Fix the CPU frequency to avoid Turbo Boost from kicking in and skewing your results. On Linux you can use `cpufreq-set` and the
52+
* Fix the CPU frequency to avoid Turbo Boost from kicking in and skewing your results. On Linux you can use `cpufreq-set` and the
5053
`performance` CPU governor.
5154
* Vary the problem input size with `@Param`.
5255
* Use the integrated profilers in JMH to dig deeper if benchmark results to not match your hypotheses:
53-
* Run the generated uberjar directly and use `-prof gc` to check whether the garbage collector runs during a microbenchmarks and skews
56+
* Run the generated uberjar directly and use `-prof gc` to check whether the garbage collector runs during a microbenchmarks and skews
5457
your results. If so, try to force a GC between runs (`-gc true`) but watch out for the caveats.
5558
* Use `-prof perf` or `-prof perfasm` (both only available on Linux) to see hotspots.
5659
* Have your benchmarks peer-reviewed.
@@ -59,4 +62,4 @@ To get realistic results, you should exercise care when running benchmarks. Here
5962

6063
* Blindly believe the numbers that your microbenchmark produces but verify them by measuring e.g. with `-prof perfasm`.
6164
* Run more threads than your number of CPU cores (in case you run multi-threaded microbenchmarks).
62-
* Look only at the `Score` column and ignore `Error`. Instead take countermeasures to keep `Error` low / variance explainable.
65+
* Look only at the `Score` column and ignore `Error`. Instead take countermeasures to keep `Error` low / variance explainable.

benchmarks/build.gradle

+3-25
Original file line numberDiff line numberDiff line change
@@ -18,11 +18,8 @@
1818
*/
1919

2020
apply plugin: 'elasticsearch.build'
21-
22-
// order of this section matters, see: https://github.com/johnrengelman/shadow/issues/336
23-
apply plugin: 'application' // have the shadow plugin provide the runShadow task
21+
apply plugin: 'application'
2422
mainClassName = 'org.openjdk.jmh.Main'
25-
apply plugin: 'com.github.johnrengelman.shadow' // build an uberjar with all benchmarks
2623

2724
// Not published so no need to assemble
2825
tasks.remove(assemble)
@@ -51,10 +48,8 @@ compileJava.options.compilerArgs << "-Xlint:-cast,-deprecation,-rawtypes,-try,-u
5148
// needs to be added separately otherwise Gradle will quote it and javac will fail
5249
compileJava.options.compilerArgs.addAll(["-processor", "org.openjdk.jmh.generators.BenchmarkProcessor"])
5350

54-
forbiddenApis {
55-
// classes generated by JMH can use all sorts of forbidden APIs but we have no influence at all and cannot exclude these classes
56-
ignoreFailures = true
57-
}
51+
// classes generated by JMH can use all sorts of forbidden APIs but we have no influence at all and cannot exclude these classes
52+
forbiddenApisMain.enabled = false
5853

5954
// No licenses for our benchmark deps (we don't ship benchmarks)
6055
dependencyLicenses.enabled = false
@@ -70,20 +65,3 @@ thirdPartyAudit.excludes = [
7065
'org.openjdk.jmh.profile.HotspotRuntimeProfiler',
7166
'org.openjdk.jmh.util.Utils'
7267
]
73-
74-
runShadow {
75-
executable = new File(project.runtimeJavaHome, 'bin/java')
76-
}
77-
78-
// alias the shadowJar and runShadow tasks to abstract from the concrete plugin that we are using and provide a more consistent interface
79-
task jmhJar(
80-
dependsOn: shadowJar,
81-
description: 'Generates an uberjar with the microbenchmarks and all dependencies',
82-
group: 'Benchmark'
83-
)
84-
85-
task jmh(
86-
dependsOn: runShadow,
87-
description: 'Runs all microbenchmarks',
88-
group: 'Benchmark'
89-
)

client/benchmark/README.md

+22-13
Original file line numberDiff line numberDiff line change
@@ -2,43 +2,54 @@
22

33
1. Build `client-benchmark-noop-api-plugin` with `gradle :client:client-benchmark-noop-api-plugin:assemble`
44
2. Install it on the target host with `bin/elasticsearch-plugin install file:///full/path/to/client-benchmark-noop-api-plugin.zip`
5-
3. Start Elasticsearch on the target host (ideally *not* on the same machine)
6-
4. Build an uberjar with `gradle :client:benchmark:shadowJar` and execute it.
5+
3. Start Elasticsearch on the target host (ideally *not* on the machine
6+
that runs the benchmarks)
7+
4. Run the benchmark with
8+
```
9+
./gradlew -p client/benchmark run --args ' params go here'
10+
```
711

8-
Repeat all steps above for the other benchmark candidate.
12+
Everything in the `'` gets sent on the command line to JMH. The leading ` `
13+
inside the `'`s is important. Without it parameters are sometimes sent to
14+
gradle.
15+
16+
See below for some example invocations.
917

1018
### Example benchmark
1119

1220
In general, you should define a few GC-related settings `-Xms8192M -Xmx8192M -XX:+UseConcMarkSweepGC -verbose:gc -XX:+PrintGCDetails` and keep an eye on GC activity. You can also define `-XX:+PrintCompilation` to see JIT activity.
1321

1422
#### Bulk indexing
1523

16-
Download benchmark data from http://benchmarks.elastic.co/corpora/geonames/documents.json.bz2 and decompress them.
24+
Download benchmark data from http://benchmarks.elasticsearch.org.s3.amazonaws.com/corpora/geonames and decompress them.
1725

18-
Example command line parameters:
26+
Example invocation:
1927

2028
```
21-
rest bulk 192.168.2.2 ./documents.json geonames type 8647880 5000
29+
wget http://benchmarks.elasticsearch.org.s3.amazonaws.com/corpora/geonames/documents-2.json.bz2
30+
bzip2 -d documents-2.json.bz2
31+
mv documents-2.json client/benchmark/build
32+
gradlew -p client/benchmark run --args ' rest bulk localhost build/documents-2.json geonames type 8647880 5000'
2233
```
2334

24-
The parameters are in order:
35+
The parameters are all in the `'`s and are in order:
2536

2637
* Client type: Use either "rest" or "transport"
2738
* Benchmark type: Use either "bulk" or "search"
2839
* Benchmark target host IP (the host where Elasticsearch is running)
2940
* full path to the file that should be bulk indexed
3041
* name of the index
31-
* name of the (sole) type in the index
42+
* name of the (sole) type in the index
3243
* number of documents in the file
3344
* bulk size
3445

3546

36-
#### Bulk indexing
47+
#### Search
3748

38-
Example command line parameters:
49+
Example invocation:
3950

4051
```
41-
rest search 192.168.2.2 geonames "{ \"query\": { \"match_phrase\": { \"name\": \"Sankt Georgen\" } } }\"" 500,1000,1100,1200
52+
gradlew -p client/benchmark run --args ' rest search localhost geonames {"query":{"match_phrase":{"name":"Sankt Georgen"}}} 500,1000,1100,1200'
4253
```
4354

4455
The parameters are in order:
@@ -49,5 +60,3 @@ The parameters are in order:
4960
* name of the index
5061
* a search request body (remember to escape double quotes). The `TransportClientBenchmark` uses `QueryBuilders.wrapperQuery()` internally which automatically adds a root key `query`, so it must not be present in the command line parameter.
5162
* A comma-separated list of target throughput rates
52-
53-

client/benchmark/build.gradle

-4
Original file line numberDiff line numberDiff line change
@@ -18,9 +18,6 @@
1818
*/
1919

2020
apply plugin: 'elasticsearch.build'
21-
// build an uberjar with all benchmarks
22-
apply plugin: 'com.github.johnrengelman.shadow'
23-
// have the shadow plugin provide the runShadow task
2421
apply plugin: 'application'
2522

2623
group = 'org.elasticsearch.client'
@@ -32,7 +29,6 @@ build.dependsOn.remove('assemble')
3229
archivesBaseName = 'client-benchmarks'
3330
mainClassName = 'org.elasticsearch.client.benchmark.BenchmarkMain'
3431

35-
3632
// never try to invoke tests on the benchmark project - there aren't any
3733
check.dependsOn.remove(test)
3834
// explicitly override the test task too in case somebody invokes 'gradle test' so it won't trip

0 commit comments

Comments
 (0)