Some utility functions to generate HDR images from a sequence of exposure time or gain modulated images. You can find a separate README describing some functinos for realistic noise simulations here.
Table of contents
To download HDRUtils, use Pypi via pip:
pip install HDRutils
If you prefer cloning this repository, install the dependencies using pip:
pip clone https://github.com/gfxdisp/HDRutils.git
cd HDRutils
pip install -e .
You will need the FreeImage plugin for reading and writing OpenEXR images:
imageio_download_bin freeimage
If you wish to capture HDR stacks using a DSLR, you will need gphoto2:
sudo apt install gphoto2
Simple wrapper functions for imageio's imread
and imwrite
are provided to set appropriate flags for HDR data. You can even call imread
on RAW file formats:
import HDRutils
raw_file = 'example_raw.arw'
img_RGB = HDRutils.imread(raw_file)
hdr_file = 'example.exr'
img = HDRutils.imread(raw_file)
HDRutils.imwrite('rgb.png', img_RGB)
HDRutils.imwrite('output_filename.exr', img)
Make sure gphoto2 is installed. Additionally, set camera to manual mode and disable autofocus on the lens. Then, decide valid exposure times (by scrolling on the camera) and run:
from HDRutils.capture import DSLR
camera = DSLR(ext='.arw')
exposures = ['10', '1', '1/10', '1/100']
camera.capture_HDR_stack('image', exposures)
The rawpy wrapper is used to read RAW images. Noise-aware merging is performed using the Poisson-noise optimal estimator. The generated HDR image is linearly related to the scene radiance
files = ['`image_0.arw`', '`image_1.arw`', '`image_2.arw`'] # RAW input files
HDR_img = HDRutils.merge(files)[0]
HDRutils.imwrite('merged.exr', HDR_img)
Sometimes the shortest exposure may contain saturated pixels. These cause artifacts when manual white-balance/color calibration is performed. Thus, HDRutils.merge()
returns an unsaturated mask in addition to the merged image. The saturated pixels can be clipped after manual white-balance/color calibration.
This function can also be accessed from the command line. Run HDRmerge -h
for usage.
The default function processes each image individually using libraw and then merges the RGB images. This result relies on the robust camera pipeline (including black-level subtraction, demosaicing, white-balance) provided by libraw, and should be suitable for most projects.
If you need finer control over the exact radiance values, this behaviour can be overriden to merge RAW bayer images by setting the flag demosaic_first=False
. This mode is useful when the camera is custom-calibrated and you have an exact correspondance between camera pixels with the scene luminance and/or color. Moreover, saturated pixels can be precisely identified before demosaicing. In this mode, a basic camera pipeline is reproduced with the following steps:
Subtract black level -> Merge image stack -> Color transformation -> White-balance
Demosaicing algorithms that are currently supported can be found at this page. Change the algorithm using HDRutils.merge(..., demosaic_first=False, demosaic=*algorithm*)
If your camera provides RAW frames in a non-standard format, you can still merge them in the camera color-space without libraw processing
files = ['file1.png', 'file2.png', 'file3.png'] # PNG bayer input files
HDR_img = HDRutils.merge(files, demosaic_first=False, color_space='raw')[0]
HDRutils.imwrite('merged.exr', HDR_img)
While merging, some ghosting artifacts can be removed by setting HDRutils.merge(..., align=True)
. This attempts homography alignment and corrects camera motion for still scenes. Unfortunately non-rigid motion requiring dense optical flow is not yet implemented.
Exposure metadata from EXIF may be inaccurate and it may be benificial to estimate relative exposures directly from the image stack. Please see our paper for details.
This feauture is currently disabled, and EXIF values are used by default. To enable exposure estimation, run HDRutils.merge(..., estimate_exp='mst')
.
Camera lens aberrations produce attenuations of spatial frequencies, which have a strong effect on the image sharpness. The Modulation Transfer Function (MTF) models the response of the camera as a function of the spatial frequency (in cycles/pixel), and can be used to increase the sharpness of images captured by the camera, by applying a Fourier space deconvolution [Chen et al. 2023].
The procedure to compute a camera's MTF is described in detail here. The output of this procedure is a JSON file describing the MTF via the coefficients of a Gaussian Mixture Model. This file can be used in HDRutils to perform the deglaring, after merging RAW files and before demosaicing. To perform merge+deglare+demosaic on a set of RAW images, run the merge function with the following arguments:
HDRutils.merge(..., color_space="raw", demosaic_first=False, mtf_json=<mtf.json>)
Generating realistic camera noise using calibrated parameters of real-world cameras is described here.
If you find this package useful, we would be grateful if you cite
@inproceedings{hanji2020noise,
author = {Hanji, Param and Zhong, Fangcheng and Mantiuk, Rafa{\l} K.},
title = {Noise-Aware Merging of High Dynamic Range Image Stacks without Camera Calibration},
booktitle = {Advances in Image Manipulation (ECCV workshop)},
year = {2020},
publisher = {Springer},
pages = {376--391},
url = {http://www.cl.cam.ac.uk/research/rainbow/projects/noise-aware-merging/},
}
@ARTICLE{hanji2023exposures,
author = {Hanji, Param and and Mantiuk, Rafa{\l} K.},
journal = {IEEE Transactions on Computational Imaging},
title = {Robust estimation of exposure ratios in multi-exposure image stacks},
year = {2023},
volume = {9},
number = {},
pages = {721-731},
doi = {10.1109/TCI.2023.3301338},
url = {https://www.cl.cam.ac.uk/research/rainbow/projects/exposure-estimation/},
}