forked from mozilla/neqo
-
Notifications
You must be signed in to change notification settings - Fork 1
388 lines (356 loc) · 16.4 KB
/
bench.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
name: Bench
on:
workflow_call:
workflow_dispatch:
schedule:
# Run at minute 0 past every 8th hour, so there is a `main`-branch baseline in the cache.
- cron: '0 */8 * * *'
env:
CARGO_PROFILE_BENCH_BUILD_OVERRIDE_DEBUG: true
CARGO_PROFILE_RELEASE_DEBUG: true
TOOLCHAIN: stable
PERF_OPT: record -F4999 --call-graph fp -g
SCCACHE_CACHE_SIZE: 128G
SCCACHE_DIRECT: true
MTU: 1504 # https://github.com/microsoft/msquic/issues/4618
permissions:
contents: read
jobs:
bench:
name: Benchmark
runs-on: self-hosted # zizmor: ignore[self-hosted-runner]
defaults:
run:
shell: bash
steps:
- name: Checkout neqo
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
with:
persist-credentials: false
- name: Checkout msquic
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
with:
repository: microsoft/msquic
ref: main
path: msquic
submodules: true
persist-credentials: false
- name: Checkout google/quiche
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
with:
repository: google/quiche
ref: main
path: gquiche
submodules: true
persist-credentials: false
- name: Set PATH and environment
run: |
echo "/home/bench/.cargo/bin" >> "${GITHUB_PATH}"
- name: Install Rust
uses: ./.github/actions/rust
with:
version: $TOOLCHAIN
tools: hyperfine, flamegraph
token: ${{ secrets.GITHUB_TOKEN }}
- name: Get minimum NSS version
id: nss-version
run: echo "minimum=$(cat neqo-crypto/min_version.txt)" >> "$GITHUB_OUTPUT"
- name: Install NSS
id: nss
uses: ./.github/actions/nss
with:
minimum-version: ${{ steps.nss-version.outputs.minimum }}
- name: Build neqo
run: |
cargo "+$TOOLCHAIN" bench --workspace --features bench --no-run
# See https://github.com/flamegraph-rs/flamegraph for why we append to RUSTFLAGS here.
export RUSTFLAGS="-C link-arg=-Wl,--no-rosegment, -C force-frame-pointers=yes $RUSTFLAGS"
cargo "+$TOOLCHAIN" build --locked --release --bin neqo-client --bin neqo-server
- name: Build msquic
run: |
mkdir -p msquic/build
cd msquic/build
cmake -GNinja -DCMAKE_BUILD_TYPE=RelWithDebInfo -DQUIC_BUILD_TOOLS=1 -DQUIC_BUILD_PERF=1 ..
cmake --build .
- name: Build google/quiche
run: |
cd gquiche
bazel build -c opt --sandbox_writable_path=/home/bench/.cache/sccache quiche:quic_server quiche:quic_client
- name: Download cached main-branch results
uses: actions/cache/restore@1bd1e32a3bdc45362d1e726936510720a7c30a57 # v4.2.0
with:
path: |
target/criterion
hyperfine
key: bench-results-${{ runner.name }}-${{ github.sha }}
restore-keys: bench-results-${{ runner.name }}-
- name: Move cached hyperfine results
run: |
mv hyperfine hyperfine-main || true
mkdir -p hyperfine
# Disable turboboost, hyperthreading and use performance governor.
- name: Prepare machine
run: sudo /root/bin/prep.sh
- name: Run cargo bench
run: |
# Pin all but neqo-bin benchmarks to CPU 0. neqo-bin benchmarks run
# both a client and a server, thus benefiting from multiple CPU cores.
#
# Run all benchmarks at elevated priority.
taskset -c 0 nice -n -20 setarch --addr-no-randomize \
cargo "+$TOOLCHAIN" bench --workspace --exclude neqo-bin --features bench -- --noplot | tee results.txt
sudo ip link set dev lo mtu "$MTU"
nice -n -20 setarch --addr-no-randomize \
cargo "+$TOOLCHAIN" bench --package neqo-bin --features bench -- --noplot | tee -a results.txt
# Compare various configurations of neqo against msquic and google/quiche, and gather perf data
# during the hyperfine runs.
- name: Compare neqo, msquic and google/quiche
env:
HOST: 127.0.0.1
PORT: 4433
SIZE: 33554432 # 32 MB
RUNS: 30
run: |
TMP=$(mktemp -d)
# Make a cert and key for msquic and google.
openssl req -nodes -new -x509 -keyout "$TMP/key" -out "$TMP/cert" -subj "/CN=DOMAIN" 2>/dev/null
# Make test files for msquic to serve.
truncate -s "$SIZE" "$TMP/$SIZE"
BIGSIZE=$(bc -l <<< "$SIZE * $RUNS")
truncate -s "$BIGSIZE" "$TMP/$BIGSIZE"
# Define the commands to run for each client and server.
declare -A client_cmd=(
["neqo"]="target/release/neqo-client _cc _pacing --output-dir . _flags -Q 1 https://$HOST:$PORT/$SIZE"
["msquic"]="msquic/build/bin/Release/quicinterop -test:D -custom:$HOST -port:$PORT -urls:https://$HOST:$PORT/$SIZE"
["google"]="gquiche/bazel-bin/quiche/quic_client --disable_certificate_verification https://$HOST:$PORT/$SIZE > $SIZE"
)
declare -A server_cmd=(
["neqo"]="target/release/neqo-server _cc _pacing _flags -Q 1 $HOST:$PORT"
["msquic"]="msquic/build/bin/Release/quicinteropserver -root:$TMP -listen:$HOST -port:$PORT -file:$TMP/cert -key:$TMP/key -noexit"
["google"]="gquiche/bazel-bin/quiche/quic_server --generate_dynamic_responses --port $PORT --certificate_file $TMP/cert --key_file $TMP/key"
)
# Flags to pass to neqo when it runs against another implementation.
declare -A neqo_flags=(
["neqo"]=""
["msquic"]="-a hq-interop"
["google"]=""
)
# Replace various placeholders in the commands with the actual values.
# Also generate an extension to append to the file name.
function transmogrify {
CMD=$1
local cc=$2
local pacing=$3
local flags=$4
if [[ "$cc" != "" ]]; then
CMD=${CMD//_cc/--cc $cc}
EXT="-$cc"
fi
if [[ "$pacing" == "on" ]]; then
CMD=${CMD//_pacing/}
EXT="$EXT-pacing"
else
CMD=${CMD//_pacing/--no-pacing}
EXT="$EXT-nopacing"
fi
CMD=${CMD//_flags/$flags}
}
# A Welch's t-test to determine if a performance change is statistically significant.
# We use this later to highlight significant changes in the results.
cat <<EOF > welch.R
args <- commandArgs(trailingOnly = TRUE)
baseline <- scan(args[1], what = numeric())
result <- scan(args[2], what = numeric())
t_result <- t.test(baseline, result, alternative = "two.sided")
p_value <- t_result\$p.value
alpha <- 0.05
quit(status = as.integer(p_value < alpha))
EOF
# See https://github.com/microsoft/msquic/issues/4618#issuecomment-2422611592
sudo ip link set dev lo mtu "$MTU"
for server in neqo google msquic; do
for client in neqo google msquic; do
# Do not run msquic against google-quiche; the latter only supports H3.
# Also, we are not interested in google as the server, or msquic as the client, except against themselves.
if [[ "$client" == "google" && "$server" == "msquic" ||
"$client" == "msquic" && "$server" == "google" ||
"$client" != "google" && "$server" == "google" ||
"$client" == "msquic" && "$server" != "msquic" ]]; then
continue
fi
# google and msquic don't let us configure the congestion control or pacing.
if [[ "$client" != "neqo" && "$server" != "neqo" ]]; then
cc_opt=("")
pacing_opt=("")
else
cc_opt=("reno" "cubic")
pacing_opt=("on" "")
fi
for cc in "${cc_opt[@]}"; do
for pacing in "${pacing_opt[@]}"; do
# Make a tag string for this test, for the results. Highlight lines we care about.
if [[ "$client" == "neqo" && "$server" == "neqo" && "$cc" == "cubic" && "$pacing" == "on" ||
"$client" == "msquic" && "$server" == "msquic" ||
"$client" == "google" && "$server" == "google" ]]; then
TAG="**$client**,**$server**,${cc:+**}$cc${cc:+**},${pacing:+**}$pacing${pacing:+**}"
else
TAG="$client,$server,$cc,$pacing"
fi
echo "Running benchmarks for $TAG" | tee -a comparison.txt
transmogrify "${server_cmd[$server]}" "$cc" "$pacing" "${neqo_flags[$client]}"
FILENAME="$client-$server$EXT"
# shellcheck disable=SC2086
taskset -c 0 nice -n -20 setarch --addr-no-randomize \
perf $PERF_OPT -o "$FILENAME.server.perf" $CMD &
PID=$!
transmogrify "${client_cmd[$client]}" "$cc" "$pacing" "${neqo_flags[$server]}"
# shellcheck disable=SC2086
taskset -c 1 nice -n -20 setarch --addr-no-randomize \
hyperfine --command-name "$TAG" --time-unit millisecond \
--export-json "hyperfine/$FILENAME.json" \
--export-markdown "hyperfine/$FILENAME.md" \
--output null --warmup 5 --runs $RUNS --prepare "sleep 1" "$CMD" |
tee -a comparison.txt
echo >> comparison.txt
# Sanity check the size of the last retrieved file.
# google/quiche outputs the HTTP header, too, so we can't just check for -eq.
[ "$(wc -c <"$SIZE")" -ge "$SIZE" ] || exit 1
# Do a longer client run with perf separately. We used to just wrap the hyperfine command above in perf,
# but that uses different processes for the individual runs, and there is apparently no way to merge
# the perf profiles of those different runs.
CMD=${CMD//$SIZE/$BIGSIZE}
# shellcheck disable=SC2086
taskset -c 1 nice -n -20 setarch --addr-no-randomize \
perf $PERF_OPT -o "$FILENAME.client.perf" $CMD > /dev/null 2>&1
kill $PID
grep -Ev '^\|(:| Command)' < "hyperfine/$FILENAME.md" | \
sed -E 's/`//g; s/,/ \| /g;' | cut -f1-8 -d\| | tr -d '\n' >> steps.md
# Figure out if any performance difference to `main` is statistically relevant, and indicate that.
BASELINE="hyperfine-main/$FILENAME.json"
if [ -e "$BASELINE" ]; then
RESULT="hyperfine/$FILENAME.json"
BASELINE_MEAN=$(jq -r '.results[0].mean' "$BASELINE")
MEAN=$(jq -r '.results[0].mean' "$RESULT")
# Even though we tell hyperfine to use milliseconds, it still outputs in seconds when dumping to JSON.
DELTA=$(bc -l <<< "($MEAN - $BASELINE_MEAN) * 1000")
PERCENT=$(bc -l <<< "($MEAN - $BASELINE_MEAN) / ($BASELINE_MEAN + $MEAN)/2 * 100")
# If a performance change is statistically significant, highlight it.
jq -r '.results[0].times[]' "$BASELINE" > baseline.txt
jq -r '.results[0].times[]' "$RESULT" > result.txt
if ! Rscript welch.R baseline.txt result.txt 2> /dev/null; then
if (( $(bc -l <<< "$DELTA > 0") )); then
echo "Performance has regressed: $BASELINE_MEAN -> $MEAN"
SYMBOL=":broken_heart:"
FORMAT='**'
else
echo "Performance has improved: $BASELINE_MEAN -> $MEAN"
SYMBOL=":green_heart:"
FORMAT='**'
fi
else
echo "No statistically significant change: $BASELINE_MEAN -> $MEAN"
SYMBOL=""
FORMAT=""
fi
printf "| %s %s%.1f%s | %s%.1f%%%s |\n" \
"$SYMBOL" "$FORMAT" "$DELTA" "$FORMAT" "$FORMAT" "$PERCENT" "$FORMAT" >> steps.md
else
echo No cached baseline from main found.
echo '| :question: | :question: |' >> steps.md
fi
done
done
done
done
# Make a single results table.
{
echo "Transfer of $SIZE bytes over loopback, $RUNS runs. All unit-less numbers are in milliseconds."
echo
# shellcheck disable=SC2016
echo '| Client | Server | CC | Pacing | Mean ± σ | Min | Max | Δ `main` | Δ `main` |'
echo '|:---|:---|:---|---|---:|---:|---:|---:|---:|'
cat steps.md
} > comparison.md
rm -r "$TMP"
# Re-enable turboboost, hyperthreading and use powersave governor.
- name: Restore machine
run: |
sudo /root/bin/unprep.sh
# In case the previous test failed:
sudo ip link set dev lo mtu 65536
if: ${{ success() || failure() || cancelled() }}
- name: Post-process perf data
run: |
for f in *.perf; do
# Convert for profiler.firefox.com
perf script -i "$f" -F +pid > "$f.fx" &
# Generate perf reports
perf report -i "$f" --no-children --stdio > "$f.txt" &
# Generate flamegraphs
flamegraph --perfdata "$f" --palette rust -o "${f//.perf/.svg}" &
done
wait
rm neqo.svg
- name: Format results as Markdown
id: results
run: |
{
echo "### Benchmark results"
echo
} > results.md
SHA=$(cat target/criterion/baseline-sha.txt || true)
if [ -n "$SHA" ]; then
{
echo "Performance differences relative to $SHA."
echo
} | tee sha.md >> results.md
fi
sed -E -e 's/^ //gi' \
-e 's/((change|time|thrpt):[^%]*% )([^%]*%)(.*)/\1<b>\3<\/b>\4/gi' results.txt |\
perl -p -0777 -e 's/(.*?)\n(.*?)(((No change|Change within|Performance has).*?)(\nFound .*?)?)?\n\n/<details><summary>$1: $4<\/summary><pre>\n$2$6<\/pre><\/details>\n/gs' |\
sed -E -e 's/(Performance has regressed.)/:broken_heart: <b>\1<\/b>/gi' \
-e 's/(Performance has improved.)/:green_heart: <b>\1<\/b>/gi' \
-e 's/^ +((<\/pre>|Found).*)/\1/gi' \
-e 's/^<details>(.*Performance has.*)/<details open>\1/gi' >> results.md
{
echo
echo "### Client/server transfer results"
SHA=$(cat target/criterion/baseline-sha.txt || true)
if [ -n "$SHA" ]; then
cat sha.md >> results.md
fi
cat comparison.md
} >> results.md
cat results.md > "$GITHUB_STEP_SUMMARY"
- name: Remember main-branch push URL
if: ${{ github.ref == 'refs/heads/main' }}
run: echo "${{ github.sha }}" > target/criterion/baseline-sha.txt
- name: Cache main-branch results
if: ${{ github.ref == 'refs/heads/main' }}
uses: actions/cache/save@1bd1e32a3bdc45362d1e726936510720a7c30a57 # v4.2.0
with:
path: |
target/criterion
hyperfine
key: bench-results-${{ runner.name }}-${{ github.sha }}
- name: Export perf data
id: export
uses: actions/upload-artifact@65c4c4a1ddee5b72f698fdd19549f0f0fb45cf08 # v4.6.0
with:
name: ${{ github.event.repository.name }}-${{ github.sha }}
path: |
*.svg
*.perf
*.perf.fx
*.txt
*.md
results.*
target/criterion
hyperfine
compression-level: 9
- name: Export PR comment data
uses: ./.github/actions/pr-comment-data-export
with:
name: ${{ github.workflow }}
contents: results.md
log-url: ${{ steps.export.outputs.artifact-url }}