mirror of
https://github.com/opencv/opencv.git
synced 2024-12-14 08:59:11 +08:00
205 lines
7.7 KiB
Markdown
205 lines
7.7 KiB
Markdown
High Dynamic Range Imaging {#tutorial_hdr_imaging}
|
||
==========================
|
||
|
||
@tableofcontents
|
||
|
||
@next_tutorial{tutorial_stitcher}
|
||
|
||
| | |
|
||
| -: | :- |
|
||
| Original author | Fedor Morozov |
|
||
| Compatibility | OpenCV >= 3.0 |
|
||
|
||
Introduction
|
||
------------
|
||
|
||
Today most digital images and imaging devices use 8 bits per channel thus limiting the dynamic range
|
||
of the device to two orders of magnitude (actually 256 levels), while human eye can adapt to
|
||
lighting conditions varying by ten orders of magnitude. When we take photographs of a real world
|
||
scene bright regions may be overexposed, while the dark ones may be underexposed, so we can’t
|
||
capture all details using a single exposure. HDR imaging works with images that use more that 8 bits
|
||
per channel (usually 32-bit float values), allowing much wider dynamic range.
|
||
|
||
There are different ways to obtain HDR images, but the most common one is to use photographs of the
|
||
scene taken with different exposure values. To combine this exposures it is useful to know your
|
||
camera’s response function and there are algorithms to estimate it. After the HDR image has been
|
||
blended it has to be converted back to 8-bit to view it on usual displays. This process is called
|
||
tonemapping. Additional complexities arise when objects of the scene or camera move between shots,
|
||
since images with different exposures should be registered and aligned.
|
||
|
||
In this tutorial we show how to generate and display HDR image from an exposure sequence. In our
|
||
case images are already aligned and there are no moving objects. We also demonstrate an alternative
|
||
approach called exposure fusion that produces low dynamic range image. Each step of HDR pipeline can
|
||
be implemented using different algorithms so take a look at the reference manual to see them all.
|
||
|
||
Exposure sequence
|
||
-----------------
|
||
|
||
![](images/memorial.png)
|
||
|
||
Source Code
|
||
-----------
|
||
|
||
@add_toggle_cpp
|
||
This tutorial code's is shown lines below. You can also download it from
|
||
[here](https://github.com/opencv/opencv/tree/5.x/samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp)
|
||
@include samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
This tutorial code's is shown lines below. You can also download it from
|
||
[here](https://github.com/opencv/opencv/tree/5.x/samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java)
|
||
@include samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
This tutorial code's is shown lines below. You can also download it from
|
||
[here](https://github.com/opencv/opencv/tree/5.x/samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py)
|
||
@include samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py
|
||
@end_toggle
|
||
|
||
Sample images
|
||
-------------
|
||
|
||
Data directory that contains images, exposure times and `list.txt` file can be downloaded from
|
||
[here](https://github.com/opencv/opencv_extra/tree/5.x/testdata/cv/hdr/exposures).
|
||
|
||
Explanation
|
||
-----------
|
||
|
||
- **Load images and exposure times**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Load images and exposure times
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Load images and exposure times
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Load images and exposure times
|
||
@end_toggle
|
||
|
||
Firstly we load input images and exposure times from user-defined folder. The folder should
|
||
contain images and *list.txt* - file that contains file names and inverse exposure times.
|
||
|
||
For our image sequence the list is following:
|
||
@code{.none}
|
||
memorial00.png 0.03125
|
||
memorial01.png 0.0625
|
||
...
|
||
memorial15.png 1024
|
||
@endcode
|
||
|
||
- **Estimate camera response**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Estimate camera response
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Estimate camera response
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Estimate camera response
|
||
@end_toggle
|
||
|
||
It is necessary to know camera response function (CRF) for a lot of HDR construction algorithms.
|
||
We use one of the calibration algorithms to estimate inverse CRF for all 256 pixel values.
|
||
|
||
- **Make HDR image**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Make HDR image
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Make HDR image
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Make HDR image
|
||
@end_toggle
|
||
|
||
We use Debevec's weighting scheme to construct HDR image using response calculated in the previous
|
||
item.
|
||
|
||
- **Tonemap HDR image**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Tonemap HDR image
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Tonemap HDR image
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Tonemap HDR image
|
||
@end_toggle
|
||
|
||
Since we want to see our results on common LDR display we have to map our HDR image to 8-bit range
|
||
preserving most details. It is the main goal of tonemapping methods. We use tonemapper with
|
||
bilateral filtering and set 2.2 as the value for gamma correction.
|
||
|
||
- **Perform exposure fusion**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Perform exposure fusion
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Perform exposure fusion
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Perform exposure fusion
|
||
@end_toggle
|
||
|
||
There is an alternative way to merge our exposures in case when we don't need HDR image. This
|
||
process is called exposure fusion and produces LDR image that doesn't require gamma correction. It
|
||
also doesn't use exposure values of the photographs.
|
||
|
||
- **Write results**
|
||
|
||
@add_toggle_cpp
|
||
@snippet samples/cpp/tutorial_code/photo/hdr_imaging/hdr_imaging.cpp Write results
|
||
@end_toggle
|
||
|
||
@add_toggle_java
|
||
@snippet samples/java/tutorial_code/photo/hdr_imaging/HDRImagingDemo.java Write results
|
||
@end_toggle
|
||
|
||
@add_toggle_python
|
||
@snippet samples/python/tutorial_code/photo/hdr_imaging/hdr_imaging.py Write results
|
||
@end_toggle
|
||
|
||
Now it's time to look at the results. Note that HDR image can't be stored in one of common image
|
||
formats, so we save it to Radiance image (.hdr). Also all HDR imaging functions return results in
|
||
[0, 1] range so we should multiply result by 255.
|
||
|
||
You can try other tonemap algorithms: cv::TonemapDrago, cv::TonemapMantiuk and cv::TonemapReinhard
|
||
You can also adjust the parameters in the HDR calibration and tonemap methods for your own photos.
|
||
|
||
Results
|
||
-------
|
||
|
||
### Tonemapped image
|
||
|
||
![](images/ldr.png)
|
||
|
||
### Exposure fusion
|
||
|
||
![](images/fusion.png)
|
||
|
||
Additional Resources
|
||
--------------------
|
||
|
||
1. Paul E Debevec and Jitendra Malik. Recovering high dynamic range radiance maps from photographs. In ACM SIGGRAPH 2008 classes, page 31. ACM, 2008. @cite DM97
|
||
2. Mark A Robertson, Sean Borman, and Robert L Stevenson. Dynamic range improvement through multiple exposures. In Image Processing, 1999. ICIP 99. Proceedings. 1999 International Conference on, volume 3, pages 159–163. IEEE, 1999. @cite RB99
|
||
3. Tom Mertens, Jan Kautz, and Frank Van Reeth. Exposure fusion. In Computer Graphics and Applications, 2007. PG'07. 15th Pacific Conference on, pages 382–390. IEEE, 2007. @cite MK07
|
||
4. [Wikipedia-HDR](https://en.wikipedia.org/wiki/High-dynamic-range_imaging)
|
||
5. [Recovering High Dynamic Range Radiance Maps from Photographs (webpage)](http://www.pauldebevec.com/Research/HDR/)
|