Chapter 6

Video Tone Mapping

R. Boitard*,; R. Cozot*; K. Bouatouch*    * IRISA, Rennes, France
Technicolor, Cesson-Sevigne, France

Abstract

To display high dynamic range videos on a low dynamic range display, a tone mapping operation is needed. Tone mapping independently each frame of a video sequence leads to temporal artifacts that impair the visual quality of the resulting tone-mapped video. These temporal artifacts are classified into six categories: global flickering, local flickering, temporal noise, temporal brightness incoherency, temporal object incoherency, and hue coherency. We explain in detail the situations in which these artifacts may occur. We review existing video tone mapping operators (techniques that take into account more than a single frame) and show which artifacts are handled by these operators.

Keywords

Video; High dynamic range; Tone mapping; Temporal coherence; Visual artifacts

Each pixel of a low dynamic range (LDR) image is stored as color components, usually three. The way LDR displays interpret these components to shape color is defined through a display-dependent color space — for example, BT.709 (ITU, 1998) or BT.2020 (ITU, 2012). In contrast, high dynamic range (HDR) pixels represent, in floating point values, the captured physical intensity of light in candelas per square meter. They can also represent relative floating point values. Hence, adapting an HDR image to an LDR display amounts to retargeting physical values, with a virtually unlimited bit depth, to a constrained space (22n chromaticity values over 2n tonal level, n being the targeted bit depth). This operation, which ensures backward compatibility between HDR content and LDR displays, is called tone mapping. The bit-depth limitation means that many similar HDR values will be tone-mapped to the same LDR value. Consequently, contrast between neighboring pixels as well as between spatially distant areas will be reduced. Furthermore, LDR displays have a low peak luminance value when compared with the luminance that a real scene can achieve. Consequently, captured color information will have to be reproduced at different luminance levels.

In a nutshell, tone mapping an HDR image amounts to finding a balance between the preservation of details, the spatial coherency of the scene, and the fidelity of reproduction. One usually achieves this balance by taking advantage of the many weaknesses of the human visual system. Furthermore, the reproduction of a scene can sometimes be constrained by an artist or application intent. That is why a lot of tone mapping operators (TMOs) have been designed with different intents, from simulating human vision to achieving the best subjective quality (Reinhard et al., 2010; Myszkowski et al., 2008; Banterle et al., 2011).

In the early 1990s, the main goal of tone mapping was to display computer-generated HDR images on a traditional display. Indeed, use of a simple gamma mapping was not enough to reproduce all the information embedded in HDR images. Although throughout the years TMOs addressed different types of applications, most of them still focused on finding the optimal subjective quality, as the many subjective evaluations attest (Drago et al., 2003; Kuang et al., 2007; Čadík et al., 2008). However, because of the lack of high-quality HDR video content, the temporal aspect of tone mapping has been dismissed for a long time. Thanks to recent developments in the HDR video acquisition field (Tocci et al., 2011; Kronander et al., 2013, 2014), more and more HDR video content is now becoming publicly available (Unger, 2013; Krawczyk, 2006; Digital Multimedia Laboratory, 2014; Lasserre et al., 2013; IRISA, 2015). Soon many applications such as real-time TV broadcasting, cinema movies, and user-generated videos will require video tone mapping.

In this chapter, we propose to evaluate the status of the video tone mapping field when trying to achieve a defined subjective quality level. Indeed, naively applying a TMO to each frame of an HDR video sequence leads to temporal artifacts. That is why we describe, in Section 6.1, different types of temporal artifacts found through experimentation. Then, Section 6.2 introduces state-of-the-art video TMOs — that is to say, TMOs that rely on information from frames other than the frame currently being tone-mapped. In Section 6.3, we present two new types of temporal artifact that are introduced by video TMOs: temporal contrast adaptation (TCA) and ghosting artifacts (GAs). Finally, Section 6.4 presents in more detail two recently published video TMOs.

6.1 Temporal Artifacts

Through experimentation with different HDR video sequences, we encountered several types of temporal artifact. In this section we focus only on those occurring when a TMO is applied naively to each frame of an HDR video sequence. We propose classifying these artifacts into six categories:

1. Global flickering artifacts (GFAs; Section 6.1.1),

2. Local flickering artifacts (LFAs; Section 6.1.2),

3. Temporal noise (Section 6.1.3),

4. Temporal brightness incoherency (TBI; Section 6.1.4),

5. Temporal object incoherency (TOI; Section 6.1.5),

6. Temporal hue incoherency (THI; Section 6.1.6).

This section provides a description of those artifacts along with some examples. Note that all the results are provided with TMOs that do not handle time dependency — namely, TMOs that rely only on statistics of the current frame for tone mapping.

6.1.1 Global Flickering Artifacts

GFAs are well known in the video tone mapping literature and are characterized by abrupt changes, in successive frames, of the overall brightness of a tone-mapped video sequence. These artifacts appear because TMOs adapt their mapping using image statistics that tend to be unstable over time. Analysis of the overall brightness of each frame over time is usually sufficient to detect those artifacts. An overall brightness metric can be, for example, the mean luma value of an image. Note that if it is computed on HDR images, the luminance channel must first be perceptually encoded, with use of, for example, a log transform as proposed in Reinhard et al. (2002), before averaging is done.

To illustrate this type of artifact, we plot in Fig. 6.1 the overall brightness indication for both the HDR sequence and the tone-mapped sequence. Note how the evolution of the overall brightness is stable over time in the HDR sequence, while abrupt peaks occur in the LDR sequence. These artifacts appear because one of the TMO’s parameters, which adapts to each frame, varies over time. Fig. 6.2 illustrates such an artifact occurring in two successive frames of a tone-mapped video sequence. The overall brightness has changed because the relative area of the sky in the second frame is smaller, hence reducing the chosen normalization factor (99th percentile).

f06-01-9780081004128
Figure 6.1 Example of GFAs. The overall brightness in the HDR sequence is stable over time, while many abrupt variations occur in the LDR sequence. As luminance and luma have different dynamic ranges, they have been scaled to achieve a meaningful comparison.
f06-02-9780081004128
Figure 6.2 GFAs due to the use of the 99th percentile on two successive frames of the Desert sequence.

To summarize, GFAs mostly occur when one is using TMOs that rely on content-adaptive parameters that are unstable over time. They are usually considered as the most disturbing of the artifacts presented in this section, which is why they have received a lot of attention, as will be seen in Section 6.2.

6.1.2 Local Flickering Artifacts

LFAs correspond to the same phenomenon as their global counterpart but on a reduced area. They appear mostly when one is using TMOs that map a pixel on the basis of its neighborhood — namely, local TMOs. Small changes of this neighborhood, in consecutive frames, may result in a different mapping. Edge-aware TMOs are particularly prone to such artifacts as they decompose an HDR image into a base layer and one or more detail layers. As each layer is tone-mapped independently, a difference in the filtering in successive frames results in LFAs.

The top row in Fig. 6.3 represents a zoom on a portion of the computed base layer of three successive frames. Note how the edges are less filtered out in the middle frame compared with the other two frames. Application of the bilateral filter (Durand and Dorsey, 2002) operator results in an LFA in the tone-mapped result (bottom row). Although LFAs are visible in a video sequence, it is tenuous to represent LFAs by means of successive frames. A side effect of LFAs is that they modify the saliency of the tone-mapped sequence as the eye is attracted by these changes of brightness on small areas.

f06-03-9780081004128
Figure 6.3 Example of LFAs when a bilateral filtering TMO (Durand and Dorsey, 2002) is applied to three consecutive frames. Top row: Base layer computed with the bilateral filter. Bottom row: The corresponding tone-mapped result.

6.1.3 Temporal Noise

Temporal noise is a common artifact occurring in digital video sequences. Noise in digital imaging is mostly due to the camera and is particularly noticeable in low-light conditions. On images, camera noise has a small impact on the subjective quality; however, for video sequences its variation over time makes it more noticeable. This is why denoising algorithms (Brailean et al., 1995) are commonly applied to video sequences to increase their subjective quality.

As most TMOs aim to reproduce minute details, they struggle to distinguish information from noise. Consequently, most current TMOs increase the noise rather than reducing it. Local TMOs are particularly prone to such artifacts as they aim to preserve details even in dark areas, which tend to be quite noisy. Furthermore, noise is usually reproduced at a luma level higher than that of a native LDR image, which makes the noise more visible. An example of temporal noise enhanced by the application of local TMOs is illustrated in Fig. 6.4.

f06-04-9780081004128
Figure 6.4 Example of temporal noise amplification due to the application of a local edge-aware TMO (Gastal and Oliveira, 2011) (left) compared with the global photographic tone reproduction TMO (Reinhard et al., 2002) (right).

6.1.4 Temporal Brightness Incoherency

TBI artifacts occur when the relative brightness between two frames of an HDR sequence is not preserved during the tone mapping. As a TMO uses for each frame all its available range, the temporal brightness relationship between frames is not preserved throughout the tone mapping operation. Consequently, a frame perceived as the brightest in the HDR sequence is not necessarily the brightest in the LDR sequence.

For example, TBI artifacts occur when a change of illumination condition in the HDR sequence is not preserved during the tone mapping. Consequently, temporal information (ie, the change of condition) is lost, which changes the perception of the scene (along with its artistic intent). Fig. 6.5 illustrates a TBI artifact, where the overall brightness of both the HDR sequence and the LDR sequence is plotted. Note that although the mean value varies greatly in the HDR sequence, it remains stable in the LDR one. This is because a TMO searches for the best exposure for each frame. As it has no information on temporally close frames, the change of illumination is simply dismissed and the best exposure is defined independently (usually in the middle of the available range). Fig. 6.6 illustrates an example of TBI occurring in consecutive frames of a tone-mapped video sequence. The top row displays the HDR luminance of these frames in false color. The change of illumination conditions occurs when the disco ball light source is turned off. When a TMO is applied, this change of illumination condition is lost (bottom row).

f06-05-9780081004128
Figure 6.5 Example of TBI. The change of the illumination condition (represented by the mean value) in the HDR sequence is not preserved in the tone-mapped result.
f06-06-9780081004128
Figure 6.6 Example of TBI when a change of illumination occurs. False color luminance (top row) and tone-mapped results obtained with a photographic tone reproduction operator (Reinhard et al., 2002) (bottom row). Both frames appear at the same level of brightness although the false color representations indicate otherwise. The color bar indicates the value in candelas per square meter.

TBI artifacts can appear even if no change of illumination condition occurs — that is to say, when the tone mapping adapts to the content. When this adaptation occurs abruptly on successive frames, it gives rise to flickering artifacts as seen previously. However, when this adaptation is smoother — say, over a longer time — the brightness relationship between the HDR and LDR sequences will be slowly disrupted. These artifacts are similar to those that occur when commercial cameras adapt their exposure during a recording (Farbman and Lischinski, 2011). Such an artifact is shown in Fig. 6.7 as the brightest HDR frame (rightmost) is the dimmest one in the LDR sequence. This second cause of TBI artifacts is also a common cause of TOI, which is presented next.

f06-07-9780081004128
Figure 6.7 Example of TBI and TOI artifacts. False color luminance (top row) and tone-mapped result obtained with a photographic tone reproduction operator (Reinhard et al., 2002) (bottom row). The TBI is represented by the overall brightness of each frame that is not coherent between the HDR and LDR frames. The TOI is represented by the brightness of the underside of the bridge, which is similar in the HDR sequence but varies greatly in the LDR sequence. From left to right, frames 50, 100, 150, and 200.

6.1.5 Temporal Object Incoherency

TOI occurs when an object’s brightness, stable in the HDR sequence, varies in the LDR sequence. Fig. 6.8 plots the HDR and LDR overall brightness along with the value of a single pixel over several frames. Note that the HDR pixel’s value is constant over time, while the overall brightness changes. As the TMO adapts to each frame, the LDR pixel’s value changes, resulting in a TOI artifact. Fig. 6.7 illustrates visually such an artifact. When looking at the false color representation of the HDR luminance (Fig. 6.7, top row), one sees the level of brightness of the underside of the bridge to be stable over time. However, after application of a TMO (bottom row), the bridge, which appears relatively bright at the beginning of the sequence, is almost dark at the end. The temporal coherency of the bridge in the HDR sequence has not been preserved in the LDR sequence. The adaptation of a TMO to a scene is the source of TBI and TOI artifacts. However, TBI artifacts are of a global nature (difference in overall brightness between frames), while TOI artifacts are of a local nature (difference in brightness between a reduced area over time).

f06-08-9780081004128
Figure 6.8 Illustration of TBI. A pixel’s value that is constant in the HDR sequence varies greatly in the LDR sequence. The pixel and mean value have been computed on the UnderBridgeHigh sequence shown in Fig. 6.7.

6.1.6 Temporal Hue Incoherency

THI is closely related to TBI as it corresponds to the variation of the color perception of an object rather than its brightness. Such artifacts occur when the balance between tristimulus values in successive frames is not temporally preserved by the tone mapping. The main reason for this imbalance is color clipping. Color clipping corresponds to the saturation of one or more of the tone-mapped color channels (eg, red, green, or blue). Color clipping is a common artifact inherent in tone mapping of still images when one aims to reproduce as well as possible the HDR color (Xu et al., 2011; Pouli et al., 2013). When color clipping is considered as a temporal artifact, it is not the difference between the HDR and LDR reproduction that is important but rather the LDR coherency from frame to frame. Indeed, variations in the tone mapping may saturate one color channel of an area which was not in the previous frame.

To illustrate such an artifact, we generated an HDR sequence with the following characteristics:

 A square area of constant luminance (100 cd/m2) with two gradients along the CIE usi1_e and vsi1_e chrominances. The chrominance gradient ranges from −0.25 to 0.25 around the D65 white point.

 A neutral gray border area with a temporally varying luminance ranging from 0.005 to 10,000 cd/m2.

Fig. 6.9 illustrates a THI due to the clipping of one or more color channels by a TMO. Note the shift in hue illustrated both in Fig. 6.9A (right) and in a zoom on a portion of the tone-mapped frames (Fig. 6.9B).

f06-09-9780081004128
Figure 6.9 Example of THI due to color clipping. Each color channel desaturates at different temporal positions. (A) Tone-mapped frame 100 of the ColorSquare sequence, the perceptual brightness reproduction operator (Tumblin and Rushmeier, 1993) (left) and the temporal evolution of the central pixel of the square, where RGB denote the three color channels, and H the hue (right). (B) Zoom on the area outlined by the rectangle in frames 25, 50, 75, and 100 (from left to right).

6.2 Video TMOs

Applying a TMO naively to each frame of a video sequence leads to temporal artifacts. The aim of video TMOs is to prevent or reduce those artifacts. Video TMOs rely on information outside the current frame to perform their mapping. Most current video TMOs extend or postprocess TMOs designed for still images. We have sorted these techniques into three categories depending on the type of filtering:

1. Global temporal filtering (Section 6.2.1),

2. Local temporal filtering (Section 6.2.2),

3. Iterative filtering (Section 6.2.3).

For each category, we provide a description of the general technique along with different state-of-the-art references.

6.2.1 Global Temporal Filtering

Global temporal filtering aims to reduce GFAs when global TMOs are used. Indeed, global operators compute a monotonously increasing tone mapping curve that usually adapts to the image statistics of the frame to be tone-mapped. However, abrupt changes of this curve in successive frames result in GFAs. Two main approaches have been formulated so far to reduce those artifacts: filtering temporally either the tone mapping curve or the image statistics.

By application of a temporal filter to successive tone mapping curves, GFAs can be reduced. Such tone mapping curves are usually filtered during a second pass as a first pass is required to compute a tone mapping curve per frame. The display adaptive operator (Mantiuk et al., 2008) is able to perform such temporal filtering on the nodes of a computed piecewise tone mapping curve. The efficiency of this filtering is illustrated in Fig. 6.10. The top row provides the independently tone-mapped version of three successive frames of an HDR video sequence. The second row displays the corresponding piecewise tone mapping curves on top of their histogram. Note how the tone mapping curve of the middle frame is different from the other two, resulting in a change of overall brightness (GFA) in the tone-mapped result. The third row shows the temporally filtered version of the piecewise tone mapping curves. Finally, the bottom row provides the tone-mapped frames after the GFA has been reduced.

f06-10-9780081004128
Figure 6.10 Reduction of GFAs by temporally filtering the tone mapping curves. The pfstmo implementation of the display adaptive operator (Mantiuk et al., 2008) was used with options -d pd=lcd_office. From left to right, frames 153, 154, and 154 of the Temple sequence. (A) The display adaptive TMO (Mantiuk et al., 2008) without temporal filtering. (B) Histogram and piecewise tone mapping curves without temporal filtering. (C) Histogram and piecewise tone mapping curves with temporal filtering. (D) The display adaptive TMO (Mantiuk et al., 2008) with temporal filtering.

Image statistics can be unstable over time (eg, the 99th percentile, mean value, histogram of the luminance (Ward, 1994), etc.). For example, the photographic tone reproduction operator (Reinhard et al., 2002) relies on the geometric mean of an HDR image to scale it to the best exposure. One temporal extension of this operator filters this statistic along a set of previous frames (Kang et al., 2003). As a consequence, this method smooths abrupt variations of the frame geometric mean throughout the video sequence. This technique is capable of reducing flickering for sequences with slow illumination variations. However, for high variations it fails because it considers a fixed number of previous frames. That is why, Ramsey et al. (2004) proposed a method that adapts this number dynamically. The adaptation process depends on the variation of the current frame key value and that of the previous frame. Moreover, the adaptation discards outliers using a min/max threshold. This solution performs better than that of Kang et al. (2003) and for a wider range of video sequences. The computed geometric mean for these techniques and the original algorithm are plotted in Fig. 6.11. The green curve (Kang et al., 2003) smooths every peak but also propagates the resulting smoothed peaks to successive computed key values. The red curve (Ramsey et al., 2004), however, reduces the abrupt changes of the key value without propagating it to successive frames.

f06-11-9780081004128
Figure 6.11 Evolution of the frame geometric mean (key value) computed for every frame of a video sequence. An offset is added to avoid an overlap between the curves. The smoothing effects of the HDR video operator (Kang et al., 2003) and the adaptive temporal operator (Ramsey et al., 2004) are compared with that of the photographic tone reproduction operator (Reinhard et al., 2002).

Another temporal extension of the photographic tone reproduction operator was proposed in Kiser et al. (2012). The temporal filtering consists of a leaky integrator applied to three variables (a, A, and B) that modify the scaling of the HDR frame:

Ls=ϵ22(BA)/(A+B)kLw=akLw,

si3_e  (6.1)

where A=Lmaxksi4_e and B=kLminsi5_e, where Lmaxsi6_e and Lminsi7_e are the maximum and minimum values of Lw, which is the HDR luminance. k corresponds to the geometric mean and the leaky integrator is computed as

vt=(1αv)v(t1)+αvvt,

si8_e  (6.2)

where vt represents any of the three variables a, A, and B at time t and αv is a time constant giving the strength (leakiness) of the temporal filtering.

Many other TMOs filter their parameters temporally, including those in Pattanaik et al. (2000), Durand and Dorsey (2000), Irawan et al. (2005), and Van Hateren (2006). Most of them aim either to simulate the temporal adaptation of the human visual system or to reduce GFAs.

6.2.2 Local Temporal Filtering

Local temporal filtering consists in performing a pixelwise temporal filtering with or without motion compensation. Indeed, global temporal filtering cannot apply to local TMOs as such operators rely on a spatially varying mapping function. As outlined previously, local changes in a spatial neighborhood cause LFAs. To prevent these local variations of the mapping along successive frames, video TMOs can rely on pixelwise temporal filtering. For example, the gradient domain compression operator (Fattal et al., 2002) has been extended by Lee and Kim (2007) to cope with videos. This TMO computes an LDR result by finding the output image whose gradient field is the closest to a modified gradient field. Lee and Kim (2007) proposed adding a regularization term which includes a temporal coherency relying on a motion estimation:

x,yΔLd(x,y,t)G(x,y)2+λx,yLd(x,y,t)Ld(x+δx,y+δy,t1)2,

si9_e  (6.3)

where Ld is the output LDR luma at the preceding or current frame (t − 1 or t) and G is the modified gradient field. The pairs (x, y) and (δx, δy) represent, respectively, the pixel location of a considered pixel and its associated motion vectors. The parameter λ balances the distortion to the modified gradient field and to the previous tone-mapped frame.

Another operator (local model of eye adaptation (Ledda et al., 2004)) performs a pixelwise temporal filtering. However, the goal of this operator is to simulate the temporal adaptation of the human eye on a per-pixel basis. Besides increasing the temporal coherency, pixelwise temporal filtering also has denoising properties. Indeed, many denoising operators rely on temporal filtering to reduce noise (Brailean et al., 1995). Performing such a filtering during the tone mapping allows one to keep the noise level relatively low.

6.2.3 Iterative Filtering

The techniques presented so far in this section focus on preventing temporal artifacts (mostly flickering) when one is tone mapping video sequences. These a priori approaches consist in either preprocessing parameters or modifying the TMO to include a temporal filtering step. Another trend analyzes a posteriori the output of a TMO to detect and reduce temporal artifacts, the reduction consisting in iterative filtering.

One of these techniques (Guthier et al., 2011) aims at reducing GFAs. Such an artifact is detected if the overall brightness difference between two successive frames of a video sequence is greater than a brightness threshold (defined with either Weber’s law (Ferwerda, 2001) or Steven’s power law (Stevens and Stevens, 1963)). As soon as an artifact is located, it is reduced by an iterative brightness adjustment until the chosen brightness threshold is reached. Note that this technique performs an iterative brightness adjustment on the unquantized luma to avoid loss of signal due to clipping and quantization. Consequently, the TMO’s implementation needs to embed and apply the iterative filter before the quantization step. This technique relies only on the output of a TMO and hence can be applied to any TMO. Fig. 6.12 illustrates the reduction of a GFA when this postprocessing is applied.

f06-12-9780081004128
Figure 6.12 Results of the multiscale operator (Farbman et al., 2008) without (left) and with (right) flicker reduction postprocessing (Guthier et al., 2011). Each image represents two successive frames. Note the flickering artifact on the left image, while it has been removed on the right image after application of flicker reduction postprocessing (Guthier et al., 2011).

6.3 Temporal Artifacts Caused by Video TMOs

In the previous section, we presented solutions to reduce temporal artifacts when one is performing video tone mapping. These techniques target mostly flickering artifacts as they are considered as one of the most disturbing artifacts. However, these techniques can generate two new types of temporal artifact — temporal contrast adaptation (TCA) and ghosting artifacts (GAs) — which we describe in this section.

6.3.1 Temporal Contrast Adaptation

To reduce GFAs, many TMOs rely on global temporal filtering. Depending on the TMO used, the filter is either applied to the computed tone mapping curve (Mantiuk et al., 2008) or to the parameter that adapts the mapping to the image (Ramsey et al., 2004; Kiser et al., 2012). However, when a change of illumination occurs, as shown in Fig. 6.6, it also undergoes temporal filtering. Consequently, the resulting mapping does not correspond to any of the conditions but corresponds rather to a transition state. We refer to this artifact as temporal contrast adaptation (TCA). Fig. 6.13 illustrates the behavior of the temporal filtering when a change of illumination occurs. Note how the tone mapping curve, plotted on top of the histograms, shifts from the first illumination condition (frame 130) toward the second state of illumination (frame 150; see Fig. 6.6 for the false color luminance). As the tone mapping curve has anticipated this change of illumination, frames neighboring the change of illumination are tone-mapped incoherently.

f06-13-9780081004128
Figure 6.13 Example of temporal filtering of tone mapping curves when a change of illumination occurs. Top row: Tone-mapped result (frames 130, 140, 149, and 150) obtained with the display adaptive operator (Mantiuk et al., 2008) with the temporal filtering active (pfsTMO implementation Krawczyk and Mantiuk, 2007). Bottom row: Histograms of frame 130 (left) and frame 150 (right) along with the corresponding tone mapping curves for frames 130, 140, and 150.

These artifacts also occur when one is performing postprocessing to detect and reduce artifacts as in Guthier et al. (2011). Indeed, this technique relies only on the LDR results to detect and reduce artifacts. If one has no information related to the HDR video, then a change of illumination suppressed by a TMO cannot be anticipated or predicted.

6.3.2 Ghosting Artifacts

Similarly to global temporal filtering, local temporal filtering generates undesired temporal artifacts. Indeed, pixelwise temporal filtering relies on a motion field estimation which is not robust to a change of illumination conditions and object occlusions. When the motion model fails, the temporal filtering is computed along invalid motion trajectories, which results in GAs.

Fig. 6.14 illustrates a GA in two successive frames resulting from the application of the operator of Lee and Kim (2007). This artifact proves that pixelwise temporal filtering is efficient only for accurate motion vectors. A GA occurs when a motion vector associates pixels without a temporal relationship. Those “incoherent” motion vectors should be accounted for to prevent GAs, as these are the most disturbing artifacts (Eilertsen et al., 2013).

f06-14-9780081004128
Figure 6.14 Example of GAs appearing on two successive frames. They are most noticeable around the two forefront columns (red squares). The bottom row shows zooms on the rightmost column.

6.4 Recent Video TMOs

Recently, two novel contributions have been proposed in the field of video tone mapping. The first one, called zonal brightness coherency (ZBC) (Boitard et al., 2014), aims to reduce TBI and TOI artifacts through a postprocessing operation which relies on a video analysis performed before the tone mapping. The second one (Aydin et al., 2014) performs a ghost-free pixelwise spatiotemporal filtering to achieve high reproduction of contrast while preserving the temporal stability of the video.

6.4.1 Zonal Brightness Coherency

The ZBC algorithm (Boitard et al., 2014) aims to preserve the HDR relative brightness coherency between every object over the whole sequence. Effectively, it should reduce TBI, TOI, and TCA artifacts and in some cases THI artifacts. It is an iterative method based on the brightness coherency (Boitard et al., 2012) technique which considered only overall brightness coherency. This method consists of two steps: a video analysis and postprocessing.

The video analysis relies on a histogram-based segmentation as shown in Fig. 6.15. A first segmentation on a per frame basis provides several segments per frame. The geometric mean (called “key value” in the article) of each segment of each HDR frame is computed and used to build a second histogram, which is in turn segmented to compute zone boundaries. The key value kz(Lw) is computed for each zone of each frame. An anchor is then chosen either automatically or by the user to provide an intent for the rendering.

f06-15-9780081004128
Figure 6.15 Details of the video analysis. The Frame Segmentation function segments each frame of the sequence and computes each segment’s key value. The Video Segmentation determines the video zone’s boundaries and their corresponding key values. The Anchor function determines the anchor zone in the HDR sequence kvz(Lw) and computes its corresponding LDR key values kvz(Lm).

Once the video analysis has been performed, each frame is tone-mapped with any TMO. Then a scale ratio sz is applied to each pixel luminance Lm,z of each video zone z to ensure that the brightness ratio between the anchor and the current zone in the HDR sequence is preserved in the LDR sequence (Eq. 6.4):

Lzbc,z=szLm,z,sz=ζ+1ζkvz(Lm)kvz(Lw)kz(Lw)kz(Lm),

si10_e  (6.4)

where Lzbc,z is the scaled luminance, ζ is a user-defined parameter, kvz(Lw) is the anchor zone HDR key value, kvz(Lm) is the anchor zone LDR key value, kz(Lw) is the z zone HDR key value and kz(Lm) is the z zone LDR key value. Note that the subscript z stands for zone.

At the boundaries between two zones, an alpha blending is used to prevent abrupt spatial variations. The whole workflow of this technique is depicted in Fig. 6.16.

f06-16-9780081004128
Figure 6.16 Complete ZBC workflow with details of the scaling phase. The Zone′s function determines for each tone-mapped frame and each pixel, the corresponding video zone zj as well as the video blending zone bzj,j+1si11_e. Their respective scaling ratios sj and sj,j+1 are computed. The ZoneScaling function applies the scale ratios to the tone-mapped frames. Finally, Q quantizes linearly floating point values in the range [0;1] to integer values on a defined bit depth n (values in the range [0;2n − 1]).

Fig. 6.17 presents some results obtained when ZBC postprocessing is used on tone-mapped video sequences where temporal artifacts occurred. The left plot provides results regarding the reduction of the TOI artifact that was illustrated in Fig. 6.8. Thanks to the ZBC technique, the value of the pixel, which was constant in the HDR sequence, is much stabler over time. Note also that the LDR mean value is quite low at the beginning of the sequence, which will most likely result in a loss of spatial contrast in the tone-mapped frames. That is why Boitard et al. (2014) have provided a user-defined parameter which effectively trades off temporal coherency for an increase in spatial reproduction capabilities (see ζ in Eq. 6.4). In the right plot, we show some results regarding the reduction of TBI artifacts. This plot is to be compared with the one in Fig. 6.5. Use of ZBC postprocessing on the Disco sequence allows the change of illumination present in the HDR sequence to be preserved.

f06-17-9780081004128
Figure 6.17 Reduction of TOI and TBI artifacts using the ZBC post-processing. Left plot is to be compared with Fig. 6.8 while the right one with Fig. 6.5.

More results are available in Boitard et al. (2014) and Boitard (2014), especially regarding the preservation of fade effects and the impact of the different parameters of this technique.

6.4.2 Temporally Coherent Local Tone Mapping

In Section 6.3.2, we explained why pixelwise temporal filtering can cause GAs. However, this type of filtering is the only known solution to prevent LFAs that can arise when local TMOs are used. Consequently, Aydin et al. (2014) proposed a spatial filtering process to ensure high reproduction of contrast when ghost-free pixelwise temporal filtering is performed. Fig. 6.18 illustrates the workflow of this technique.

f06-18-9780081004128
Figure 6.18 General workflow of the temporally coherent local operator (Aydin et al., 2014). The spatial decomposition is illustrated in top right corner, while the base and detail filtering are illustrated in the bottom left and right corners, respectively.

This technique considers a temporal neighborhood composed of a center frame Ik and temporally close frames Ik ± i. In a first step, each frame is decomposed into a base and detail layer by use of a permeability map (spatial diffusion weights). Both subbands are then motion compensated (warped) with a previously computed motion flow. Note that no warping is necessary for the subbands associated with the central frame.

The second step consists of two temporal filters which allow separate filtering of the base and detail layers. To prevent GAs, the filtering relies on confidence weights composed of a photoconstancy permeability map and a penalization on pixels associated with high gradient flow vectors. The photoconstancy permeability map is computed between successive frames and corresponds to a temporal transposition of the spatial diffusion weight on which the spatial decomposition relies. Aydin et al. (2014) observed that this photoconstancy measure can be tuned to stop temporal filtering at most warping errors, hence preventing the appearance of GAs. However, it also defeats the purpose of temporal filtering, which is to smooth medium to low temporal variations. That is why the penalization term has been introduced as it is a good indication of complex motion where the flow estimation tends to be erroneous. This step provides two images, the spatiotemporally filtered base layer Bk and a temporally filtered detail layer Dk.

To obtain the tone-mapped frame IkTMOsi12_e, the base layer Bk can be fed to any TMO and then combined with the detail layer Dk, a process similar to that detailed in Durand and Dorsey (2002). This method addresses several of the artifacts presented in this chapter. First, the temporal noise is reduced as the details are filtered temporally. Second, LFAs are minimized thanks to the pixelwise temporal filtering. Finally, GAs are prevented by adaptation of the temporal filtering to a motion flow confidence metric. A simplified illustrative workflow is depicted in Fig. 6.19.

f06-19-9780081004128
Figure 6.19 Illustrative workflow of the temporally coherent local operator (Aydin et al., 2014). PGF, permeability-guided filtering. Source: Courtesy of Tunç Ozan Aydin.

As this technique is fairly new, extended results with more HDR sequences could help detect new types of artifacts. In particular, it would be interesting to test this method when changes of illumination or cut occur in a sequence. Furthermore, most of the results provided with this technique rely on user interaction to achieve the best trade-off between temporal and spatial contrast. This is not achievable for many applications, such as live broadcasts and tone mapping embedded in set-top boxes.

6.5 Summary

In this chapter, we have described known types of temporal artifact that occur when HDR video sequences are tone-mapped. We have categorized video TMOs with respect to how they handle temporal information, and we have shown that although these solutions can deal with certain types of temporal artifact, they can also be a source of new ones. An evaluation of video TMOs (Eilertsen et al., 2013) reported that none of the current solutions can handle a wide range of sequences. However, this study was performed before the publication of the two video TMOs presented in Section 6.4. These two techniques, albeit significantly different, provide solutions to types of artifact not dealt with before.

Table 6.1 gives an overview of the temporal artifacts presented in this chapter, along with possible solutions. From this table, we can see that all of the artifacts described in this chapter have a solution. However, none of the video TMOs presented here encompass all of the tools needed to deal with all different types of artifact. Furthermore, the appearance of more HDR video sequences and applications for video tone mapping will likely result in new types of temporal artifact. Although the two recent contributions have significantly advanced the field of video tone mapping, more work still lies ahead.

Table 6.1

Summary of Temporal Artifacts Along With Their Main Causes and Possible Solutions

Temporal ArtifactPossible CausePossible Solutions
Global flickerTemporal instability of parametersGlobal temporal filtering
Local flickerDifferent spatial filtering in successive framesPixelwise temporal filtering
Temporal noiseCamera noiseSpatial and/or temporal filtering (pixelwise)
TBI (brightness)Change of illumination adaptation of the TMOBrightness analysis of each frame
TOI (object)Adaptation of the TMOBrightness analysis per zone of frames
THI (hue)Saturation of color channel (due to clipping)Hue and brightness analysis per zone of frames
TCA (contrast)Global temporal filteringBrightness analysis per zone of frames
GhostingPixelwise temporal filteringConfidence weighting of pixelwise temporal filtering

References

Aydin T.O., Stefanoski N., Croci S., Gross M., Smolic A. Temporally coherent local tone mapping of HDR video. ACM Trans. Graph. 2014;33(6):196:1–196:13.

Banterle F., Artusi A., Debattista K., Chalmers A. Advanced High Dynamic Range Imaging: Theory and Practice. Natick, MA: AK Peters (CRC Press); 2011.

Boitard, R., 2014. Temporal Coherency in Video Tone Mapping. Ph.D. thesis, University of Rennes 1.

Boitard R., Bouatouch K., Cozot R., Thoreau D., Gruson A. Temporal coherency for video tone mapping. In: Proc. SPIE 8499, Applications of Digital Image Processing XXXV. 2012:84990D-1 84990D-10.

Boitard R., Cozot R., Thoreau D., Bouatouch K. Zonal brightness coherency for video tone mapping. Signal Process. Image Commun. 2014;29(2):229–246.

Brailean J., Kleihorst R., Efstratiadis S., Katsaggelos A., Lagendijk R. Noise reduction filters for dynamic image sequences: a review. Proc. IEEE. 1995;83(9):1272–1292.

Čadík et al., 2008 Čadík M., Wimmer M., Neumann L., Artusi A. Evaluation of HDR tone mapping methods using essential perceptual attributes. Comput. Graph. 2008;32(3):330–349.

Digital Multimedia Laboratory. DML-HDR Public Database. 2014. http://dml.ece.ubc.ca/data/DML-HDR/.

Drago F., Martens W.L., Myszkowski K., Seidel H.P. Perceptual evaluation of tone mapping operators. In: Proceedings of the SIGGRAPH 2003 Sketches. New York, NY: ACM Press; 2003:1.

Durand F., Dorsey J. Interactive tone mapping. In: Proceedings of the Eurographics Workshop on Rendering. Berlin/Heidelberg: Springer Verlag; 2000:219–230.

Durand F., Dorsey J. Fast bilateral filtering for the display of high-dynamic-range images. ACM Trans. Graph. 2002;21(3):257–266.

Eilertsen G., Wanat R., Mantiuk R.K., Unger J. Evaluation of tone mapping operators for HDR-video. Comput. Graph. Forum. 2013;32(7):275–284.

Farbman Z., Lischinski D. Tonal stabilization of video. ACM Trans. Graph. 2011;30(4):89:1–89:10.

Farbman Z., Fattal R., Lischinski D., Szeliski R. Edge-preserving decompositions for multi-scale tone and detail manipulation. In: ACM SIGGRAPH 2008 papers on, SIGGRAPH ’08. New York, NY: ACM Press; 2008:67:1–67:10.

Fattal R., Lischinski D., Werman M. Gradient domain high dynamic range compression. ACM Trans. Graph. 2002;21(3):249–256.

Ferwerda J. Elements of early vision for computer graphics. IEEE Comput. Graph. Appl. 2001;21(4):22–33.

Gastal E.S.L., Oliveira M.M. Domain transform for edge-aware image and video processing. ACM Trans. Graph. 2011;30(4):69:1–69:12.

Guthier B., Kopf S., Eble M., Effelsberg W. Flicker reduction in tone mapped high dynamic range video. In: Proc. of IS&T/SPIE Electronic Imaging (EI) on Color Imaging XVI: Displaying, Processing, Hardcopy, and Applications. 78660C-1. 2011;7866 78660C-15.

Irawan P., Ferwerda J.A., Marschner S.R. Perceptually based tone mapping of high dynamic range image streams. In: Proceedings of the Sixteenth Eurographics Conference on Rendering Techniques, EGSR ’05. Aire-la-Ville: Eurographics Association; 2005:231–242.

IRISA. IRISA Rennes FRVSense Public HDR Database. 2015. http://people.irisa.fr/Ronan.Boitard/.

ITU. Recommendation ITU-R BT.709-3: Parameter values for the HDTV standards for production and international programme exchange. International Telecommunications Union; 1998.

ITU. Recommendation ITU-R BT.2020: Parameter values for ultra-high definition television systems for production and international programme exchange. International Telecommunications Union; 2012.

Kang S.B., Uyttendaele M., Winder S., Szeliski R. High dynamic range video. ACM Trans. Graph. 2003;22(3):319–325.

Kiser C., Reinhard E., Tocci M., Tocci N. Real-time automated tone mapping system for HDR video. In: Proceedings of the IEEE International Conference on Image Processing. 2012:2749–2752.

Krawczyk G. MPI Public HDR Database. 2006. http://www.mpi-inf.mpg.de/resources/hdr/video/.

Krawczyk G., Mantiuk R. Display adaptive pfstmo documentation. 2007. http://pfstools.sourceforge.net/man1/pfstmo_mantiuk08.1.html.

Kronander J., Gustavson S., Bonnet G., Unger J. Unified HDR reconstruction from raw CFA data. In: IEEE International Conference on Computational Photography (ICCP). 2013:1–9.

Kronander J., Gustavson S., Bonnet G., Ynnerman A., Unger J. A unified framework for multi-sensor HDR video reconstruction. Signal Process. Image Commun. 2014;29(2):203–215.

Kuang J., Yamaguchi H., Liu C., Johnson G.M., Fairchild M.D. Evaluating HDR rendering algorithms. ACM Trans. Appl. Percept. 2007;4(2):9:1–9:27.

Lasserre S., Le Léannec F., Francois E. Description of HDR sequences proposed by Technicolor. In: ISO/IEC JTC1/SC29/WG11 JCTVC-P0228. San Jose, USA: IEEE; 2013.

Ledda P., Santos L.P., Chalmers A. A local model of eye adaptation for high dynamic range images. In: Proceedings of the 3rd International Conference on Computer Graphics, Virtual Reality, Visualisation and Interaction in Africa, AFRIGRAPH ’04. New York, NY: ACM Press; 2004:151–160.

Lee C., Kim C.S. Gradient domain tone mapping of high dynamic range videos. In: 2007 IEEE International Conference on Image Processing. III-461–III-464. 2007;3.

Mantiuk R., Daly S., Kerofsky L. Display adaptive tone mapping. ACM Trans. Graph. 2008;27(3):68:1–68:10.

Myszkowski K., Mantiuk R., Krawczyk G. High Dynamic Range Video. San Rafael, CA: Morgan & Claypool; 1–158. 2008;2.

Pattanaik S.N., Tumblin J., Yee H., Greenberg D.P. Time-dependent visual adaptation for fast realistic image display. In: Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH ’00. New York, NY: ACM Press/Addison-Wesley Publishing Co. 2000:47–54.

Pouli T., Artusi A., Banterle F. Color correction for tone reproduction. In: Color and Imaging Conference Final Program and Proceedings. 2013:215–220.

Ramsey S.D., Johnson III J.T., Hansen C. Adaptive temporal tone mapping. In: Computer Graphics and Imaging, 2004. 2004:3–7.

Reinhard E., Heidrich W., Debevec P., Pattanaik S., Ward G., Myszkowski K. High Dynamic Range Imaging, 2nd Edition: Acquisition, Display, and Image-Based Lighting. Los Altos, CA: Morgan Kaufmann; 2010.

Reinhard E., Stark M., Shirley P., Ferwerda J. Photographic tone reproduction for digital images. ACM Trans. Graph. 2002;21(3):267–276.

Stevens J.C., Stevens S.S. Brightness function: effects of adaptation. J. Opt. Soc. Am. 1963;53(3):375–385.

Tocci M.D., Kiser C., Tocci N., Sen P. A versatile HDR video production system. ACM Trans. Graph. 2011;30(4):41:1–41:10.

Tumblin J., Rushmeier H. Tone reproduction for realistic images. IEEE Comput. Graph. Appl. 1993;13(6):42–48.

Unger J. HDRv Public Database. 2013. http://www.hdrv.org/Resources.php.

Van Hateren J.H. Encoding of high dynamic range video with a model of human cones. ACM Trans. Graph. 2006;25(4):1380–1399.

Ward G. A contrast-based scale factor for luminance display. In: Heckbert P.S., ed. Graphics Gems IV. San Diego, CA: Academic Press Professional, Inc. 1994:415–421.

Xu D., Doutre C., Nasiopoulos P. Correction of clipped pixels in color images. IEEE Trans. Vis. Comput. Graph. 2011;17(3):333–344.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.141.31.125