Wu-Wei (non-effort) is a benchmarking toolkit developed in the Sable Lab at McGill University with the objective of simplifying the study of the performance of programming languages implementations and tools.
We aim to make the toolkit and the benchmark suites built with it:
- Consistent and Correct by supporting correctness checks for every language implementation of benchmarks that automatically ensure that the computation result of the benchmarks are consistent across all language implementations and correct with regard to the algorithm for known inputs;
- Extensible across numerical languages, benchmarks, compilers, run-time environments;
- Friendly to language implementation research by automating all tasks for compiler and virtual-machine research and encouraging a writing style for benchmarks that factors the core computation from the runners to minimize the non-core functions necessary to validate the output of compilers;
- Easy to use by automating the deployment of benchmarks, their test on virtual (web browser and others) and native platforms, as well as the gathering and reporting of relative performance data;
- Fast by making the setup (data generation and loading) and teardown as quick as possible so that most of the time is spent in the core computation in every language;
- Small by minimizing the amount of data needed to use the suite;
- Simple by minimizing the amount of external dependencies and tools required to run the suite;
Although we tried our best to minimize external dependencies, the toolkit still depends on the following external tools:
- Node.js
- Python
Individual artifacts may have more dependencies. Refer to their documentation for more details.
Please read the handbook for more details on how to use it.
Copyright (c) 2016, Erick Lavoie, Hanfeng Chen