A distributed storage benchmark for files, objects & blocks with support for GPUs
elbencho was inspired by traditional storage benchmark tools like fio, mdtest and ior, but was written from scratch to replace them with a modern and easy to use unified tool for file systems, object stores & block devices.
(click to expand)
- Unified latency, throughput, IOPS benchmark for file, object & block storage
- Supports local and shared storage through distributed service mode
- For modern NVMe storage or classic spinning disk storage
- GPU storage access performance testing through Nvidia CUDA or GPUDirect Storage (GDS)
- Live statistics show how the system behaves under load
- Multi-threaded and async I/O support through libaio
- Results by first and by last finished thread
- CSV file output to easily create graphs in spreadsheet apps or via elbencho-chart tool
- Data integrity verification option
The built-in help (elbencho --help
) provides simple examples to get started.
You can get elbencho pre-built for Linux & Windows from the Releases section and from Docker Hub.
Building elbencho requires a C++17 compatible compiler, such as gcc version 7.x or higher.
sudo apt install build-essential debhelper devscripts fakeroot git libaio-dev libboost-filesystem-dev libboost-program-options-dev libboost-thread-dev libncurses-dev libnuma-dev lintian
sudo yum install boost-devel gcc-c++ git libaio-devel make ncurses-devel numactl-devel rpm-build
Skip these steps on RHEL / CentOS 8.0 or newer.
sudo yum install centos-release-scl # for CentOS
# ...or alternatively for RHEL: yum-config-manager --enable rhel-server-rhscl-7-rpms
sudo yum install devtoolset-8
scl enable devtoolset-8 bash # alternatively: source /opt/rh/devtoolset-8/enable
The scl enable
command enters a shell in which the environment variables are pointing to a newer gcc version. (The standard gcc version of the system remains unchanged.) Use this shell to run make
later. The resulting executable can run outside of this shell.
Start by cloning the main repository:
git clone https://github.com/breuner/elbencho.git
cd elbencho
make help
will show you all build & install options.
(Note that S3 support is not enabled by default due to longer build time, but can easily be enabled. See the additional build info below.)
This is the standard build command:
make -j $(nproc)
You can run elbencho directly from the bin subdir (bin/elbencho
), but you probably want to run make rpm
or make deb
now to build a package and install it. On Ubuntu, run this:
make deb
sudo apt install ./packaging/elbencho*.deb
There you go. Happy benchmarking!
Now comes the fun part: It's time to find out what your system can deliver.
The built-in help (elbencho --help
) provides many usage examples. You will be interested in throughput and IOPS, typically for a single client and also for multiple clients. For the latter, see --hosts
& --service
.
If your cluster is using Slurm to allocate nodes, you can find examples here.
If GPU data transfer performance is critical for you, e.g. because you are running DeepLearning applications, you will also want to include GPUs in your read/write benchmarks (--gpuids
).
Instead of giving benchmark parameters on the command line, you can also use a config file (--configfile myconfig.conf
). You can find an example here.
Elbencho presents two result columns: One showing the aggregate status at the point in time when the fastest I/O thread finished its work ("First Done") and one for the aggregate end result at the point in time when the slowest thread finished its work ("Last Done"). Ideally, both are close together.
To generate charts from your result series, e.g. based on different block sizes or different thread counts, use elbencho's csv file option (--csvfile
) and the elbencho-chart
tool to easily generate a chart from your csv file. Alternatively, spreadsheet tools like Microsoft Excel or Google Sheets can be used for this.
In case of questions, comments, if something is missing to make elbencho more useful or if you would just like to share your thoughts, feel free to contact me: sven.breuner[at]gmail.com
elbencho --version
shows which optional features are included in an executable.
CUDA support for GPU data transfers will automatically be enabled when CUDA development files (cuda_runtime.h
and libcudart.so
) are installed on the build system. Alternatively, elbencho CUDA support can be manually enabled or disabled. See make help
for details.
GPUDirect Storage (GDS) support through the cuFile API will automatically be enabled when GDS development files (cufile.h
and libcufile.so
) are installed on the build system. Alternatively, elbencho GDS support can be manually enabled or disabled. See make help
for details.
Enabling S3 Object Storage support will automatically download a AWS SDK git repository of over 1GB size and increases build time from a few seconds to a few minutes. Thus, S3 support is not enabled by default, but it can easily be enabled as described below.
sudo yum install cmake libarchive libcurl-devel openssl-devel libuuid-devel zlib zlib-devel
sudo apt install cmake libcurl4-openssl-dev libssl-dev uuid-dev zlib1g-dev
To build elbencho with S3 support, just add the S3_SUPPORT=1
parameter to the make command. (If you previously built elbencho without S3 support, then run make clean-all
before this.)
make S3_SUPPORT=1 -j $(nproc)