[ieee 2014 27th sibgrapi conference on graphics, patterns and images (sibgrapi) - brazil...

8
High-Quality Reverse Tone Mapping for a Wide Range of Exposures Rafael P. Kovaleski, Manuel M. Oliveira Instituto de Inform´ atica, UFRGS Porto Alegre, Brazil Email: {rpkovaleski,oliveira}@inf.ufrgs.br Abstract—Reverse-tone-mapping operators (rTMOs) enhance low-dynamic-range images and videos for display on high- dynamic-range monitors. A common problem faced by previous rTMOs is the handling of under or overexposed content. Under such conditions, they may not be effective, and even cause loss and reversal of visible contrast. We present an rTMO based on cross-bilateral filtering that generates high-quality HDR images and videos for a wide range of exposures. Experiments performed using an objective image quality metric show that our approach is the only technique available that can gracefully enhance perceived details across a large range of image exposures. Keywords- reverse tone mapping; high dynamic range; cross- bilateral filter; I. I NTRODUCTION Recent developments in high-dynamic-range (HDR) hard- ware [2] indicate that HDR displays should become widely available in the near future. In this scenario, reverse tone map- ping operators (rTMOs) [3], [4], [5], [6] seek to enhance the brightness and contrast of low-dynamic-range (LDR) images and videos for HDR displays. The goal of these techniques is to improve the overall viewer’s experience, and their benefits have been verified by recent studies [7], [8], [9], [10]. A common limitation of previous rTMOs is their inability to handle under or overexposed content. In such situations, these operators will not improve the perceived contrast, and may even cause loss and reversal of visible contrast. We present an rTMO based on cross-bilateral filtering that generates high- quality HDR images and videos while supporting a wide range of exposure conditions. We evaluate the performance of our solution against previous rTMOs using DRIM [11], an objective image quality metric. These experiments show that ours is the only technique capable of gracefully enhancing perceived details across a large range of image exposures. A key component of several recent rTMOs is a brightness enhancement function (BEF) [4], [5], also known as an ex- pand map [3], [12]. BEFs are generated by determining areas where image information may have been lost (primarilly due to sensor saturation) and filling these regions using a smooth function. High-quality BEFs significantly improve the result of an rTM operation [5]. Our solution is designed around the automatic and efficient creation of high-quality BEFs, and builds upon the technique proposed by Kovaleski and Oliveira [5]. While their technique properly handles under- and well-exposed images, it does not perform as well for over-exposed ones. We analyze the reasons for such behavior, and show that it results from the fact that these BEFs are computed considering, for each pixel, the maximum intensity value among the R, G, and B channels. While the authors’ original motivation was to conservatively detect saturated pixels, this treats all channels equally, regardless of their perceptual differences. As a side effect, it prevents the technique from properly handling over- exposed areas. The key observation of our work follows from this analysis: the replacement of the maximum-intensity channel by the pixel’s luminance (which properly weights the perceptual contributions of the various channels) lends to a more robust solution, capable of handling a wider range of exposures. Although this difference might at first appear subtle, it is indeed quite significant: it results from a better understanding of the underlying problem, and extends the range of appli- cability of the technique. As a result, our solution is the only one that automatically performs reverse tone mapping of under-, well-, and over-exposed images and videos. Moreover, it lends to simpler equations that provide greater flexibility. Thus, while Kovaleski and Oliveira’s original technique only supports acceleration through the use of a bilateral grid, our solution allows one to directly use any of the existing bilateral- filtering acceleration strategies and implementations. Figures 1 and 2 show examples of BEFs generated using our approach. Figure 1 (b) depicts a BEF created for the LDR image shown on its left. The image in (c) is a tone-mapped version of the resulting HDR image. Note the many details that were not immediately visible in the original image, which contains both under- and over-exposed areas. The image in (d) shows the result produced by DRIM, where the blue pixels represent contrast enhancement. Figure 2 shows examples of BEFs computed with our technique using various bilateral- filtering implementations. In all examples, note the ability of our solution to produce smooth functions while preserving even fine details, such as small leaves and twigs. The contributions of our work include: An automatic, high-quality reverse-tone-mapping opera- tor for images and videos that supports a wide range of exposures (Section IV). Our solution improves Kovaleski and Oliveira’s technique [5] to robustly support overex- posed regions; An efficient approach for creating smooth high-quality 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images 1530-1834/14 $31.00 © 2014 IEEE DOI 10.1109/SIBGRAPI.2014.29 49

Upload: manuel-m

Post on 07-Apr-2017

215 views

Category:

Documents


3 download

TRANSCRIPT

Page 1: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

High-Quality Reverse Tone Mapping for a WideRange of Exposures

Rafael P. Kovaleski, Manuel M. OliveiraInstituto de Informatica, UFRGS

Porto Alegre, BrazilEmail: {rpkovaleski,oliveira}@inf.ufrgs.br

Abstract—Reverse-tone-mapping operators (rTMOs) enhancelow-dynamic-range images and videos for display on high-dynamic-range monitors. A common problem faced by previousrTMOs is the handling of under or overexposed content. Undersuch conditions, they may not be effective, and even cause lossand reversal of visible contrast. We present an rTMO based oncross-bilateral filtering that generates high-quality HDR imagesand videos for a wide range of exposures. Experiments performedusing an objective image quality metric show that our approachis the only technique available that can gracefully enhanceperceived details across a large range of image exposures.

Keywords- reverse tone mapping; high dynamic range; cross-bilateral filter;

I. INTRODUCTION

Recent developments in high-dynamic-range (HDR) hard-ware [2] indicate that HDR displays should become widelyavailable in the near future. In this scenario, reverse tone map-ping operators (rTMOs) [3], [4], [5], [6] seek to enhance thebrightness and contrast of low-dynamic-range (LDR) imagesand videos for HDR displays. The goal of these techniques isto improve the overall viewer’s experience, and their benefitshave been verified by recent studies [7], [8], [9], [10].

A common limitation of previous rTMOs is their inability tohandle under or overexposed content. In such situations, theseoperators will not improve the perceived contrast, and mayeven cause loss and reversal of visible contrast. We present anrTMO based on cross-bilateral filtering that generates high-quality HDR images and videos while supporting a widerange of exposure conditions. We evaluate the performanceof our solution against previous rTMOs using DRIM [11], anobjective image quality metric. These experiments show thatours is the only technique capable of gracefully enhancingperceived details across a large range of image exposures.

A key component of several recent rTMOs is a brightnessenhancement function (BEF) [4], [5], also known as an ex-pand map [3], [12]. BEFs are generated by determining areaswhere image information may have been lost (primarilly dueto sensor saturation) and filling these regions using a smoothfunction. High-quality BEFs significantly improve the resultof an rTM operation [5].

Our solution is designed around the automatic and efficientcreation of high-quality BEFs, and builds upon the techniqueproposed by Kovaleski and Oliveira [5]. While their techniqueproperly handles under- and well-exposed images, it does

not perform as well for over-exposed ones. We analyze thereasons for such behavior, and show that it results fromthe fact that these BEFs are computed considering, for eachpixel, the maximum intensity value among the R, G, andB channels. While the authors’ original motivation was toconservatively detect saturated pixels, this treats all channelsequally, regardless of their perceptual differences. As a sideeffect, it prevents the technique from properly handling over-exposed areas.

The key observation of our work follows from this analysis:the replacement of the maximum-intensity channel by thepixel’s luminance (which properly weights the perceptualcontributions of the various channels) lends to a more robustsolution, capable of handling a wider range of exposures.Although this difference might at first appear subtle, it isindeed quite significant: it results from a better understandingof the underlying problem, and extends the range of appli-cability of the technique. As a result, our solution is theonly one that automatically performs reverse tone mapping ofunder-, well-, and over-exposed images and videos. Moreover,it lends to simpler equations that provide greater flexibility.Thus, while Kovaleski and Oliveira’s original technique onlysupports acceleration through the use of a bilateral grid, oursolution allows one to directly use any of the existing bilateral-filtering acceleration strategies and implementations.

Figures 1 and 2 show examples of BEFs generated usingour approach. Figure 1 (b) depicts a BEF created for the LDRimage shown on its left. The image in (c) is a tone-mappedversion of the resulting HDR image. Note the many detailsthat were not immediately visible in the original image, whichcontains both under- and over-exposed areas. The image in (d)shows the result produced by DRIM, where the blue pixelsrepresent contrast enhancement. Figure 2 shows examples ofBEFs computed with our technique using various bilateral-filtering implementations. In all examples, note the ability ofour solution to produce smooth functions while preservingeven fine details, such as small leaves and twigs.

The contributions of our work include:∙ An automatic, high-quality reverse-tone-mapping opera-

tor for images and videos that supports a wide range ofexposures (Section IV). Our solution improves Kovaleskiand Oliveira’s technique [5] to robustly support overex-posed regions;

∙ An efficient approach for creating smooth high-quality

2014 27th SIBGRAPI Conference on Graphics, Patterns and Images

1530-1834/14 $31.00 © 2014 IEEE

DOI 10.1109/SIBGRAPI.2014.29

49

Page 2: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

(a) (b) (c) (d)

Fig. 1: Reverse tone mapping using our technique. (a) Input LDR image. (b) Brightness-enhancement function computedautomatically from the image in (a). (c) Tone-mapped version of the resulting HDR image, obtained using Reinhard et al.’s [1]local operator. (d) Output of the DRIM metric, blue represents contrast enhancement.

brightness enhancement functions for rTMOs based oncross-bilateral filtering (Section IV). Our method supportsany bilateral-filter acceleration technique, freeing the userfrom the limitations of any specific implementation.

II. RELATED WORK

Banterle et al. [3] proposed one of the first reverse-tone-mapping operators. They apply the inverse of Reinhard etal.’s tone-mapping operator [1] to an LDR image and use amedian-cut algorithm to cluster light sources in areas of highluminance. A density estimation of high-luminance regions,called expand map, is used to interpolate between the inputimage and the initial inverse-tone-mapped image, generatingthe HDR output. By designing a temporally coherent expandmap, this work was extended for videos [12].

The LDR2HDR framework of Rempel et al. [4] is basedon the use of brightness enhancement functions, which aresomehow similar to Banterle et al.’s expand maps [3]. First, theimage values are linearized; then, a binary mask (containingonly the pixels whose linearized values are above a certainthreshold) is blurred by a 2-D Gaussian kernel and then com-bined with an edge stopping function to maintain the contraston sharp edges. The resulting BEF is then used to scalethe linearized image values. This strategy is computationallyefficient, but sacrifices BEF quality.

Kovaleski and Oliveira [5] automatically generate high-quality edge-aware BEFs in real time using a bilateralgrid [13]. Their formulation, however, does not properlyhandles overexposures, a condition for which HDR images arehighly useful. Our work generalizes and improves Kovaleskiand Oliveira’s technique, allowing it to also handle overex-posures, as well as freeing it from the restriction to be usedwith a bilateral grid. Our solution supports all bilateral-filteringimplementations and acceleration strategies.

Meylan et al. [14] split the input image into diffuse andspecular components using some threshold. Specular areas areexpanded and then slightly blurred, to avoid visual artifactsand unnatural contours, while diffuse areas are mildly scaled.The segmentation of bright areas was improved by Didyket al. [15], who presented a semi-automatic technique forenhancing images and videos by labeling regions as diffusesurfaces, light sources, specular highlights, and reflections.Each area is then enhanced using different expansion func-tions. This system requires user intervention and is meant asa professional post-production tool.

Wang et al. [16] detail an interactive method to fill overex-posed and underexposed areas with texture from other, betterexposed regions, which are manually selected by the user. Incontrast with this and with Meylan et al.’s technique [14], oursis completely automatic.

Masia et al. [6] performed a detailed perceptual studythat considered viewers’ preferences comparing the resultsof three rTMOs and the original LDR image for variousexposure levels. The considered rTMOs were LDR2HDR [4],Banterle et al.’s [3], and linear contrast scaling [9]. Masiaand collaborators found that the performance of these oper-ators (measured as user ratings) decrease as the number ofoverexposed pixels increases. The authors also found that, inthe case of overexposure, scaling the images using a gammaexpansion (calculated from the estimated image key [17]) isa simple yet effective technique. As the gamma-expansionapproach was designed for overexposed images, it does notperform well for underexposed ones, achieving similar ratingsas the other tested operators [6], and being prone to loss andreversal of visible contrast (Figures 6 and 7). More recently,Masia and Gutierrez [18] extended the work of Masia et al. [6]by proposing more robust methods to calculate the gammavalue used for image expansion. These methods are still fairly

50

Page 3: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

simple to compute, but show a higher correlation with thedata gathered in their user studies. However, the fundamentalrestrictions of the original technique persist and, likewise, theyare not applicable to videos. In contrast, our rTMO supportsa wide range of exposures, producing high-quality results allacross it, and supports reverse tone mapping of videos.

Masia et al. [21] have also proposed a higher-level approachto reverse tone mapping that considers saliency informationalong with pixels intensity values. It, however, depends onuser input to assign area importance.

III. BILATERAL AND CROSS-BILATERAL FILTERING

A bilateral filter [22], [23], [24] is a nonlinear filter thatsimultaneously performs domain and range filtering. Usinga Gaussian function 𝐺 as decreasing function, the bilateral-filtered version 𝐼𝑏 of a grayscale image 𝐼 is defined as:

𝐼𝑏𝑝 =1

𝑊 𝑏𝑝

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐼𝑝 − 𝐼𝑞∣)𝐼𝑞

𝑊 𝑏𝑝 =

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐼𝑝 − 𝐼𝑞∣),(1)

where 𝐼𝑝 and 𝐼𝑞 correspond to the pixel values at positions 𝑝and 𝑞 in image 𝐼 , and 𝜎𝑠 and 𝜎𝑟 are the standard deviationsof Gaussian functions for the space and range domains,respectively. 𝑊 𝑏

𝑝 is a normalization factor.By replacing the term 𝐺𝜎𝑟

(∣𝐼𝑝 − 𝐼𝑞∣) in Equation 1 with𝐺𝜎𝑟

(∣𝐸𝑝 − 𝐸𝑞∣), for a second image 𝐸, one obtains a crossbilateral filter [25], [26] (Equation 2). Intuitively, the Gaussianfiltering weights are based on the spatial distances betweenpixel locations 𝑝 and 𝑞 in image 𝐼 , and on the differencesbetween pixel values 𝐸𝑝 and 𝐸𝑞 from image 𝐸.

𝐼𝑏𝑝 =1

𝑊 𝑏𝑝

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐸𝑝 − 𝐸𝑞∣)𝐼𝑞

𝑊 𝑏𝑝 =

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐸𝑝 − 𝐸𝑞∣).(2)

Due to the nonlinear nature of this filter, its use in realtime applications is limited. Some techniques allow it tobe computed at interactive rates by approximating the resultusing different methods. Durand and Dorsey [27] computethe filter response by linearly interpolating several discretizedintensity values, which are filtered with a Gaussian kernel inthe frequency domain. This work was extended by Paris andDurand [28] by interpreting their computation as a higher-dimensional linear convolution, which is followed by a trilin-ear interpolation and a division. A generalization of this idea isthe bilateral grid [13]. A bilateral grid Γ is a regularly-sampled3-D array, where the first two dimensions define the image’sspace domain, while the third one represents the image’s range.Filtering using a bilateral grid is performed in three steps:(i) grid creation, (ii) processing, and (iii) slicing. To processa bilateral grid, a 3-D function 𝑓 is applied to its contents,producing a new grid Γ′ = 𝑓(Γ). For the bilateral filter, 𝑓is a convolution by a Gaussian kernel, using 𝜎𝑠 and 𝜎𝑟 asvariances for the domain and range dimensions, respectively.

The slicing operation is used to extract a piecewise-smooth2-D map from the bilateral grid.

Using a bilateral grid, the bilateral-filter algorithm of Parisand Durand [28], which approximates the bilateral filter inEquation 1, can be expressed as:

𝑏𝑓(𝐼) = 𝑠𝐼(𝐺𝜎𝑠,𝜎𝑟⊗ 𝑔(𝐼)), (3)

where 𝑠𝐼 is the slicing operation using 𝐼 as reference image,𝑔(𝐼) is the bilateral grid created using image 𝐼 , ⊗ is theconvolution operator, and 𝐺 is a 3-D Gaussian kernel withspatial and range standard deviations given by 𝜎𝑠 and 𝜎𝑟,respectively. With a bilateral grid, cross-bilateral filtering isperformed using image 𝐸 to determine the grid positions,while storing the values from the data image 𝐼 . The slicingoperation then uses the edge image to recover the result.

Adams et al. [29] efficiently implement color bilateralfilters in 5D using uniform simplices. The use of simplicesresults in cheaper interpolation operations when comparedto Paris and Durand’s work [28]. More recently, Gastal andOliveira [20] compute manifolds that adapt themselves tothe signal. Weighted interpolation between these manifoldsgenerates the output with increased accuracy and speed, usinga smaller number of sampling points than previous techniques.

Fig. 3: BEF generation for a synthetic image. (left) Input imagecontaining high R, G, and B channel values. (center) BEFgenerated by the technique of [5]. It treats R, G, B regionsequally, despite their perceptual differences. (right) BEF gen-erated with our solution. Note how it properly represents thesubtle gradients found in (left).

IV. AN RTMO FOR A WIDE RANGE OF EXPOSURES

Kovaleski and Oliveira [5] generate brightness enhancementfunctions using a bilateral grid. The grid is created usingtwo different images, much like in cross-bilateral filtering.But the slicing operation is performed with a third image.Following the same notation of Equation 3, their techniquecan be expressed as

𝑏𝑓(𝐼) = 𝑠𝐵(𝐺𝜎𝑠,𝜎𝑟⊗ 𝑔𝐴(𝐶)), (4)

where 𝐴, 𝐵, and 𝐶 are single-channel images. Each pixelin 𝐴 contains the maximum intensity value among the RGBchannels of the corresponding pixel of 𝐼 . Image 𝐵 contains𝐼’s luminance information, and image 𝐶 contains all pixelsin 𝐴 whose values are above some threshold 𝑡 (230 forvideo, 254 for images). In their formulation, the maximum-intensity value among the RGB channels (image 𝐴) was usedas a conservative way of detecting saturated pixels. As we

51

Page 4: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

Input image Brute Force Bilateral Grid Real-Time Bilateral Filter Adaptive Manifolds

Fig. 2: Brightness-enhancement functions generated with our technique using different bilateral-filter algorithms. Input image,and BEFs created using brute force, real-time bilateral filter [19], bilateral grid [13], and adaptive manifolds [20].

building01 building02 building03 building04

Bui

ldin

gse

ries

Kov

ales

kian

dO

livei

raM

asia

and

Gut

ierr

ezO

urs

Fig. 4: Results of the DRIM metric [11] for the building series of bright/overexposed images. First row: input LDRimages. Other rows: DRIM output for the results produced by the rTM operators of Kovaleski and Oliveira [5], Masia andGutierrez’s [18], and ours. The colors green, blue, and red indicate pixels affected by loss of visible contrast, amplification ofinvisible contrast, and contrast reversal, respectively.

will show, this choice compromises the technique’s abilityto properly handle overexposures. Image 𝐼 has its valueslinearized using a gamma curve of 2.2 prior to the generationof images 𝐴, 𝐵 and 𝐶. 𝑔𝐴(𝐶) is a bilateral grid filtered usingthe edge information from image 𝐴, while storing values fromimage 𝐶. 𝜎𝑠 and 𝜎𝑟 are set to 100 and 30, respectively.

We rewrite Equation 4 using the structure of the bilateral-

filter equation (Equation 1):

𝐼𝑏𝑝 =1

𝑊 𝑏𝑝

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐵𝑝 −𝐴𝑞∣)𝐶𝑞

𝑊 𝑏𝑝 =

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐵𝑝 −𝐴𝑞∣),(5)

where 𝐴, 𝐵 and 𝐶 are the same as in Equation 4.A key observation is that the use of the maximum-intensity-

52

Page 5: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

graffiti01 graffiti02 graffiti03 graffiti04

Gra

ffiti

seri

esK

oval

eski

and

Oliv

eira

Mas

iaan

dG

utie

rrez

Our

s

Fig. 5: Results of the DRIM metric for the graffiti series of bright/overexposed images. First row: input LDR images. Otherrows: DRIM output for the results produced by the rTM operators of Kovaleski and Oliveira [5], Masia and Gutierrez’s [18],and ours. The colors green, blue, and red indicate pixels affected by loss of visible contrast, amplification of invisible contrast,and contrast reversal, respectively.

value image 𝐴, in Equation 5, tends to cause Kovaleski andOliveira‘s technique to be unable to enhance details in brightor overexposed LDR images. This is illustrated in Figure 3with a synthetic image containing high R, G and B values andsharp edges. By using the maximum-intensity channel values,their technique treats the R, G, and B channels equally, despitetheir perceptual differences. Such inconsistencies are reflectedin the resulting BEFs (Figure 3(center)). For comparison, theBEF generated with our solution is shown in Figure 3(right).Note how it properly represents the subtle gradients found inthe original image.

This observation also leads to a solution to the problem:replace the maximum-intensity-channel image 𝐴 with theluminance image 𝐵. This modification not only solves theissue with overexposed images, but also greatly simplifiesthe BEF generation process, turning Equation 5 into a cross-bilateral filtering operation:

𝐼𝑏𝑝 =1

𝑊 𝑏𝑝

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐵𝑝 −𝐵𝑞∣)𝐶𝑞

𝑊 𝑏𝑝 =

𝑞∈𝑆

𝐺𝜎𝑠(∥𝑝− 𝑞∥)𝐺𝜎𝑟

(∣𝐵𝑝 −𝐵𝑞∣).(6)

The use of Equation 6 instead of Equation 5 allows us toimmediately use any bilateral-filtering acceleration technique.This is illustrated in Figure 2, which shows BEFs computedwith our solution for two LDR images using four different

approaches: brute-force bilateral filter (no acceleration), bi-lateral grid [13], real-time bilateral filter [19], and adaptivemanifolds [20]. In contrast, Equation 5 would require acomplete rewrite of other bilateral-filtering techniques to beadapted to the algorithm in [5]. Furthermore, we use 𝜎𝑠 = 150,which corresponds to an angle of 1.2∘ at a viewing distanceof 3m for a 37” HDR display with a resolution of 1920x1080.This results in a blur filter spectrum containing low angularfrequencies of 0.5 cycles per degree or less, to which thehuman visual system is not very sensitive [4]. We use 𝜎𝑟 = 25,obtained through extensive testing. In our experience, thesevalues are better suited for BEF generation.

Given the computed BEFs, we generate the resulting HDRimages using similar steps as Rempel et al. [4] and Kovaleskiand Oliveira [5]: the input image 𝐼 is linearized using a 2.2gamma curve. From the linearized image, we generate a BEFusing Equation 6. Then, the BEF has its values scaled to the[1..𝛼] range. For our experiments we use 𝛼 = 4. Finally,the linearized input image has its values scaled accordingto the target-display capabilities and chosen 𝛼 value: we use0.3𝑐𝑑/𝑚2 for the black point and 1, 200𝑐𝑑/𝑚2 for the whitepoint. The resulting HDR image is then obtained by pointwisemultiplication of this scaled image by our rescaled BEF.

53

Page 6: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

car01 car02 car03 car04

Car

seri

esK

oval

eski

and

Oliv

eira

Mas

iaan

dG

utie

rrez

Our

s

Fig. 6: Results of the DRIM metric for the car series of dark/underexposed images. First row: input LDR images. Otherrows: DRIM output for the results produced by the rTM operators of Kovaleski and Oliveira [5], Masia and Gutierrez’s [18],and ours. The colors green, blue, and red indicate pixels affected by loss of visible contrast, amplification of invisible contrast,and contrast reversal, respectively.

V. RESULTS AND DISCUSSION

Masia et al. [6] compared the results of their gamma-expansion rTMO against three reverse tone mapping operators,namely the LDR2HDR [4], Banterle et al.’s operator [3],and linear contrast scaling [9]. For these comparisons, theyused images taken with different exposures. They have shownthat the gamma-expansion rTMO outperforms the comparedtechniques for overexposed images, while performing similarlyfor under and properly exposed images. This study, however,did not consider Kovaleski and Oliveira’s operator [5].

Masia et al. [6] calculate the gamma value using a linearregression of the form 𝛾 = 10.44𝑘 − 6.282, where 𝑘 is theimage key [17]. In a subsequent tech report, Masia andGutierrez [18] state that this equation was incorrectly reportedand present new equations for calculating the gamma value.

To achieve a comprehensive evaluation, we compare theresults of our technique against Kovaleski and Oliveira’s andthe revised version of Masia et al.’s (based on the robustregression equation presented by Masia and Gutierrez [18]).

To obtain an objective evaluation of these three techniques,we use the Dynamic Range Independent Image Quality Metric(DRIM) [11]. Such metric can detect three different types ofevents in the resulting HDR images: loss of visible contrast(shown in green), amplification of invisible contrast (shownin blue) and reversal of visible contrast (shown in red). Fora reverse-tone-mapping operator, loss and reversal of visible

contrast are undesirable, while amplification of invisible con-trast tends to increase perceived image quality [4].

For the experiments, we use the same images used inMasia et al.’s work [6], which are available as part of theirsupplemental materials. They correspond to series of underand overexposed images, some of which are shown on the toprows of Figures 4 to 7. Due to space constraints, not all imagessequences are shown in the paper, but the presented resultsare representative of all under and overexposed sequences.The building (Figure 4) and graffiti (Figure 5) seriescorrespond to bright or overexposed images, while the car(Figure 6) and flowers (Figure 7) series correspond to darkor underexposed images.

For the DRIM comparisons, our rTMO used a bilateralgrid [13]. The results produced by other bilateral-filter im-plementations performed equally well with DRIM. Using thebilateral grid, our BEFs are generated in under 1ms for imagesup to 1920 × 1080 pixels on an Intel Mobile Core 2 QuadQ9000 2.0 GHz CPU, using an nVidia 9400M GPU.

Figures 4 and 5 show that for overexposed images ourtechnique produces results that are at least as good as theones produced by the revised version of Masia et al.’s ap-proach [18]. Note that their technique has been designed toachieve its best performance for overexposed images. For bothtechniques there is an overall increase of contrast, even inareas where there is an apparent loss of detail (such as in

54

Page 7: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

flowers01 flowers02 flowers03 flowers04Fl

ower

sse

ries

Kov

ales

kian

dO

livei

raM

asia

and

Gut

ierr

ezO

urs

Fig. 7: Results of the DRIM metric for the flowers series of dark/underexposed images. First row: input LDR images. Otherrows: DRIM output for the results produced by the rTM operators of Kovaleski and Oliveira [5], Masia and Gutierrez’s [18],and ours. The colors green, blue, and red indicate pixels affected by loss of visible contrast, amplification of invisible contrast,and contrast reversal, respectively.

the lightpost in image building04 and in the graffiti inimages graffiti03-04).

For overexposed images, our solution outperforms Ko-valeski and Oliveira’s approach (see the larger clusters ofblue pixels in the DRIM results for our technique). Theseresults are explained by the differences between the equationsthat define the two methods, and by our choice of parametervalues that work better for a larger range of exposures. Asalready discussed in Section IV, the use of the maximum-intensity-value image (image 𝐴) in Equation 5 tends to causetheir technique to be unable to enhance details in bright oroverexposed images. Our solution avoids these problems byproperly weighting the three image channels using a luminanceimage (image 𝐵) in Equation 6.

Figures 6 and 7 show the results of the three rTMOapplied to a series of darker/underexposed images. For thefirst image in both sequences, the results produced by thegamma-expansion operator [18] suffer from significant lossand reversal of visible contrast, shown in the DRIM resultsin green and red, respectively. For the entire series, bestresults are obtained with our solution, followed by Kovaleskiand Oliveira’s. This should not be surprising, as the gamma-expansion technique was developed for overexposed images.

Figures 4 to 7 demonstrate the ability of our solution tohandle a wide range of exposures, enhancing perceived detailsall across it. Thus, it outperforms the gamma-expansion oper-

ator [18], which is not appropriate for underexposed images,as well as the Kovaleski and Oliveira’s original operator [5],which is not as effective for overexposed images.

Figure 8 illustrates the use of our solution to also cre-ate HDR videos. The algorithm is independently applied tothe individual frames. The top row shows frames from fireand smoke simulation videos. Their corresponding BEFs areshown in the middle row, while the result of the DRIMcomparison of the original LDR and resulting HDR videoframes are shown in the bottom row. Note the amplificationof invisible contrast in these frames, which contain both darkand bright regions. The cost of our rTM operator is the sameas the cost of the selected cross-bilateral filter implementation.

VI. CONCLUSION

We presented a high-quality reverse-tone-mapping operatorfor images and videos that supports a wide range of exposures.Our solution is based on the use of cross-bilateral filtering tocompute smooth BEFs that preserve sharp edges. It generalizesand improves Kovaleski and Oliveira’s technique [5], and canbe directly used with any bilateral-filtering strategy, freeingthe user from the limitations of any specific implementation.

We performed objective comparisons of our results againstthe ones produced by the gamma-expansion operator of Masiaand Gutierrrez [18] and by Kovaleski and Oliveira’s opera-tor [5] for a large range of image exposures using DRIM [11].

55

Page 8: [IEEE 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) - Brazil (2014.8.26-2014.8.30)] 2014 27th SIBGRAPI Conference on Graphics, Patterns and Images - High-Quality

Ori

gina

lFr

ame

Com

pute

dB

EF

DR

IMR

esul

t

Fig. 8: Frames from two HDR videos created using ourapproach. Top: frames from fire and smoke simulation videos.Middle: Corresponding BEFs. Bottom: Result of the DRIMcomparison of the original LDR and resulting HDR videoframes. Blue means amplification of invisible contrast.

These experiments show that our solution is the only one thatcan gracefully enhance perceived details across a large rangeof exposures. As such, it outperforms Masia and Gutierrez’soperator, which was designed for overexposed images, butdoes not work well for underexposed ones. It also outperformsKovaleski and Oliveira’s operator, which does a good jobfor well and underexposed images, but is not effective foroverexposed ones. Our solution can also be used to createHDR videos.

By expanding the range of image and video exposures, andby providing the user with the flexibility to choose amongany existing (or future) bilateral-filtering implementation, oursolution significantly enhances the repertoire of HDR imagingtools. It has the potential to enable new and creative HDRapplications that have not been previously possible due tothe lack of a fast reverse-tone-mapping operator capable ofhandling a large exposure range.

REFERENCES

[1] E. Reinhard, M. Stark, P. Shirley, and J. Ferwerda, “Photographic tonereproduction for digital images,” ACM TOG, vol. 21, pp. 267–276, 2002.

[2] H. Seetzen, W. Heidrich, W. Stuerzlinger, G. Ward, L. Whitehead,M. Trentacoste, A. Ghosh, and A. Vorozcovs, “High dynamic rangedisplay systems,” ACM TOG, vol. 23, no. 3, pp. 760–768, 2004.

[3] F. Banterle, P. Ledda, K. Debattista, and A. Chalmers, “Inverse tonemapping,” in GRAPHITE, 2006, pp. 349–356.

[4] A. G. Rempel, M. Trentacoste, H. Seetzen, H. D. Young, W. Heidrich,L. Whitehead, and G. Ward, “Ldr2hdr: on-the-fly reverse tone mapping

of legacy video and photographs,” ACM TOG, vol. 26, no. 3, p. 39,2007.

[5] R. P. Kovaleski and M. M. Oliveira, “High-quality brightness en-hancement functions for real-time reverse tone mapping,” The VisualComputer, vol. 25, no. 5-7, pp. 539–547, 2009.

[6] B. Masia, S. Agustin, R. W. Fleming, O. Sorkine, and D. Gutierrez,“Evaluation of reverse tone mapping through varying exposure condi-tions,” ACM Transactions on Graphics, vol. 28, no. 5, 2009.

[7] A. Yoshida, R. Mantiuk, K. Myszkowski, and H.-P. Seidel, “Analysisof reproducing real-world appearance on displays of varying dynamicrange,” Computer Graphics Forum, vol. 25, no. 3, pp. 415–426, 2006.

[8] H. Seetzen, H. Li, L. Ye, G. Ward, L. Whitehead, and W. Heidrich,“Guidelines for contrast, brightness, and amplitude resolution of dis-plays.” Soc. for Information Display (SID) Digest, pp. 1229–1233, 2006.

[9] A. O. Akyuz, R. Fleming, B. E. Riecke, E. Reinhard, and H. H. Bulthoff,“Do hdr displays support ldr content?: a psychophysical evaluation,”ACM Trans. Graph., vol. 26, no. 3, Jul. 2007.

[10] F. Banterle, P. Ledda, K. Debattista, M. Bloj, A. Artusi, and A. Chalmers,“A psychophysical evaluation of inverse tone mapping techniques,”Comput. Graph. Forum, vol. 28, no. 1, pp. 13–25, 2009.

[11] T. O. Aydin, R. Mantiuk, K. Myszkowski, and H.-P. Seidel, “Dynamicrange independent image quality assessment,” ACM Trans. Graph.,vol. 27, no. 3, pp. 69:1–69:10, Aug. 2008.

[12] F. Banterle, P. Ledda, K. Debattista, and A. Chalmers, “Expanding lowdynamic range videos for high dynamic range applications,” in Proc.24th Spring Conference on Computer Graphics, 2008, pp. 33–41.

[13] J. Chen, S. Paris, and F. Durand, “Real-time edge-aware image process-ing with the bilateral grid,” ACM TOG, vol. 26, no. 3, p. 103, 2007.

[14] L. Meylan, S. Daly, and S. Susstrunk, “The reproduction of specularhighlights on high dynamic range displays,” in In Proc. of the 14thColor Imaging Conference, 2006.

[15] P. Didyk, R. Mantiuk, M. Hein, and H.-P. Seidel, “Enhancement of brightvideo features for hdr displays,” Comput. Graph. Forum, vol. 27, no. 4,pp. 1265–1274, 2008.

[16] L. Wang, L.-Y. Wei, K. Zhou, B. Guo, and H.-Y. Shum, “High dynamicrange image hallucination,” in Render. Techniques, 2007, pp. 321–326.

[17] A. O. Akyuz and E. Reinhard, “Color appearance in high-dynamic-rangeimaging,” J. Electronic Imaging, vol. 15, no. 3, p. 033001, 2006.

[18] B. Masia and D. Gutierrez, “Multilinear regression for gamma expansionof overexposed content,” Universidad de Zaragoza (Tech Report RR-03-11), Tech. Rep., July 2011.

[19] Q. Yang, K.-H. Tan, and N. Ahuja, “Real-time O(1) bilateral filtering,”in CVPR, 2009, pp. 557–564.

[20] E. S. L. Gastal and M. M. Oliveira, “Adaptive manifolds for real-timehigh-dimensional filtering,” ACM TOG, vol. 31, no. 4, pp. 33:1–33:13,2012.

[21] B. Masia, R. W. Fleming, O. Sorkine, and D. Gutierrez, “Selectivereverse tone mapping,” in Cong. Espanol de Informatica Grafica, 2010.

[22] V. Aurich and J. Weule, “Non-linear gaussian filters performing edgepreserving diffusion,” in DAGM-Symposium, 1995, pp. 538–545.

[23] S. M. Smith and J. M. Brady, “Susan - a new approach to low levelimage processing,” IJCV, vol. 23, no. 1, pp. 45–78, 1997.

[24] C. Tomasi and R. Manduchi, “Bilateral filtering for gray and colorimages,” in ICCV, 1998, pp. 839–846.

[25] E. Eisemann and F. Durand, “Flash photography enhancement viaintrinsic relighting.” in ACM TOG, 2004, pp. 673–678.

[26] G. Petschnigg, R. Szeliski, M. Agrawala, M. F. Cohen, H. Hoppe,K. Toyama, and K. Toyama, “Digital photography with flash and no-flashimage pairs.” in ACM Transactions on Graphics, 2004, pp. 664–672.

[27] F. Durand and J. Dorsey, “Fast bilateral filtering for the display of high-dynamic-range images,” in SIGGRAPH, 2002, pp. 257–266.

[28] S. Paris and F. Durand, “A fast approximation of the bilateral filter usinga signal processing approach,” in ECCV (4), 2006, pp. 568–580.

[29] A. Adams, J. Baek, and M. A. Davis, “Fast high-dimensional filteringusing the permutohedral lattice,” CGF, vol. 29, no. 2, pp. 753–762, 2010.

56