visual sensitivity correlated tone reproduction for low dynamic …€¦ · to predict the...

14
Visual sensitivity correlated tone reproduction for low dynamic range images in the compression field Geun-Young Lee Sung-Hak Lee Hyuk-Ju Kwon Kyu-Ik Sohng Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021 Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

Upload: others

Post on 20-Oct-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

  • Visual sensitivity correlated tonereproduction for low dynamic rangeimages in the compression field

    Geun-Young LeeSung-Hak LeeHyuk-Ju KwonKyu-Ik Sohng

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • Visual sensitivity correlated tone reproduction forlow dynamic range images in the compression field

    Geun-Young Lee, Sung-Hak Lee,* Hyuk-Ju Kwon, and Kyu-Ik SohngKyungpook National University, School of Electronics Engineering, 80 Deahakro, Bukgu, Daegu 702-701, Republic of Korea

    Abstract. An image toning method for low dynamic range image compression is presented. The proposedmethod inserts tone mapping into JPEG baseline instead of postprocessing. First, an image is decomposedinto detail, base, and surrounding components in terms of the discrete cosine transform coefficients.Subsequently, a luminance-adaptive tone mapping based on the human visual sensitivity properties is applied.In addition, compensation modules are added to enhance the visually sensitive factors, such as saturation,sharpness, and gamma. A comparative study confirms that the transmitted compression images have goodimage quality. © The Authors. Published by SPIE under a Creative Commons Attribution 3.0 Unported License. Distribution or reproductionof this work in whole or in part requires full attribution of the original publication, including its DOI. [DOI: 10.1117/1.OE.53.11.113111]

    Keywords: tone mapping; low dynamic range; discrete cosine transform; human vision factor; contrast enhancement.

    Paper 141252 received Aug. 7, 2014; accepted for publication Oct. 29, 2014; published online Nov. 25, 2014.

    1 IntroductionThe luminance dynamic range in the real world is signifi-cantly large and the dynamic range of the eyes can shiftin response to the change or intensity of scenes.1 In contrast,image sensors that capture the luminance dynamic range arelimited to a certain intensity range, which is relatively verylow. Thus, there is a substantial difference between an imagecaptured by a sensor and the perceived scene.2 To reduce thisdiscrepancy, researchers have proposed a number of globaland local tone mapping methods. The global tone mappingmethod using only one mapping function has a relativelysimple computation, though it is insufficient to addresswide dynamic range, whereas the local tone mapping methodhas an adaptive function that may vary depending on spa-tially adjacent pixels. Furthermore, certain local methodsadopt human visual properties for local contrast enhance-ment, such as image color appearance model (iCAM)-basedmethods,3,4 logarithmic mapping,5 local eye adaptation,6 andhistogram adjustment.7 These make images similar to realscenes that an observer would perceive. It is widelyknown that human vision responds to luminance in such amanner that individual visual cells adjust each gain accord-ing to locally adapted luminance. Moreover, various experi-ments that help us understand the instinctive nature of humanvision have been conducted by psychophysicists. The resultsof the experiments are usually statistical data and need to becreated as the functions so that they are easy to use. Recently,tone mapping operators are extended into video streams.8,9

    These methods use temporally close frames to smooth outabrupt changes of luminance.10 In addition, for surveillancesystem, content-based tone mapping has been proposed.11 Itpresents inter and interframe object based tone mapping forthe enhancement of regions of interest (ROIs) in videostreams. Essentially, this content-based method has piece-wise global tone mapping based on features from detectedROIs. Generally, local tone mapping methods have better

    performance because the human visual system is a spatialcorrelation system sensitive to regional relative brightness,rather than a system described globally single tone curve.4

    Local tone mapping methods usually use image decom-position for edge preservation. Textures and detail informa-tion can be removed when a dynamic range is largelycompressed.3,12,13 The procedure for local tone mappingusing image decomposition is shown in Fig. 1. The detaillayer is preserved, whereas the base layer is compressedby tone mapping. The base layer has large features and isextracted by filtering the input image. The detail layer isa subtraction of the base layer from the input. After com-pressing the base layer, it is recomposed with the detaillayer. The details are not suppressed through tone mapping.Therefore, image decomposition is a necessary procedure forlocal tone mapping methods.

    This paper proposes a luminance-adaptive local tone map-ping method in the compression field for the contrastenhancement of low dynamic range images. Tone mappingis composed of simple local functions related to humanvision properties that respond to luminance change. Inorder to achieve this, we investigate human visual sensitivityproperties using two luminance adaptation functions for sim-ple or complex stimuli and contrast sensitivity functions(CSFs). Our image enhancement is based on these lumi-nance-adaptive human factors. In addition, we propose anovel image decomposition method in the compressiondomain using discrete cosine transform (DCT) band split-ting. The image decomposition is not only a necessarystep for local tone mapping, but also an initial step requiredfor merging the proposed tone mapping with JPEG baseline.The previous spatial domain based methods require severalGaussian kernels for multiscale tone mapping and detail-base separation. Moreover, edge stopped blurring techniquesto prevent halo artifact are computationally intensive.13 Theproposed method does not use any Gaussian kernel for edgepreserving and reduces the complexity of the process foradjusting sharpness and colors by cooperating with DCTcoefficients. Consequently, it performs well in terms of*Address all correspondence to: Sung-Hak Lee, E-mail: [email protected]

    Optical Engineering 113111-1 November 2014 • Vol. 53(11)

    Optical Engineering 53(11), 113111 (November 2014)

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

    http://dx.doi.org/10.1117/1.OE.53.11.113111http://dx.doi.org/10.1117/1.OE.53.11.113111http://dx.doi.org/10.1117/1.OE.53.11.113111http://dx.doi.org/10.1117/1.OE.53.11.113111http://dx.doi.org/10.1117/1.OE.53.11.113111http://dx.doi.org/10.1117/1.OE.53.11.113111mailto:[email protected]:[email protected]:[email protected]:[email protected]

  • simplicity of the process, detail preserving, tonal rendition,and halo artifact elimination. Further, for postcomplemen-tary processing, mask-based sharpness enhancement, visualgamma correction, and color compensation are accomplished.

    The remainder of this paper is organized as follows. InSec. 2, we discuss the luminance-adaptive human factors rel-evant to our research. In Sec. 3, we present the proposedalgorithm for image contrast enhancement using mask-based image decomposition and luminance-adaptive localtone mapping. In Sec. 4, we describe complementary proc-esses adopted for further image enhancement. In Sec. 5, wepresent simulations and comparative results. Finally, inSec. 6, we provide concluding remarks.

    2 Luminance-Adaptive Human Vision FactorsHuman vision accommodates variations in luminance througha process called light adaptation. In this section, we focus onthree human factors in light adaptation: two types of bright-ness functions, proposed by Stevens and Stevens14 andBartleson and Breneman,15 and contrast sensitivity func-tions.16 The brightness function represents the nonlinearitybetween perceived brightness and measured luminance ofthe same patch under various intensities of adapting lumi-nance. According to the brightness function proposed byStevens, brightness is increased sharply when human visionperceives the luminance of a patch to be increasing fromdarkness. It changes linearly over the threshold as shownin Fig. 2 on a logarithmic scale. Moreover, the slope andthreshold of a linear area increase with adapting luminance.In other words, in order for the perceived contrast ratio withbrightness to be preserved, the physical contrast ratio withluminance is decreased with an increase of adapting lumi-nance. In contrast to the simple patch experiments conducted

    by Stevens, Bartleson and Breneman conducted experimentsto predict the brightness for a complex stimulus. Accordingto the results by Bartleson and Breneman, brightness percep-tions of complex scenes, such as images, can be described byboth a power term and an exponential decay term.

    Human vision is more sensitive to change or differencethan the absolute value of luminance. Generally, image hav-ing a high contrast ratio is more distinct at lower levels ofadaptation.17,18 However, because of nonlinearity betweenperceived brightness and measured intensity under differentadaptations, it is impossible to fix the physical contrast ratiothat is suitable for an image with various intensity ranges. Toaddress this problem, Lee et al. obtained a curve representingthe relation between threshold luminance and adapting lumi-nance for constant brightness perception using the Stevens’results and Bartlenson–Breneman’s functions.19 As shown inFig. 3, the curve represents the highest and lowest luminanceperceived by human vision at each adapting luminance.Based on these extreme luminance values, the necessary con-trast ratio at each adapting luminance is shown in Fig. 4. Foran identical perception of a certain contrast ratio, humanvision requires a high luminance ratio at a low adapting lumi-nance, and vice versa; it requires a relatively low luminance

    Fig. 1 Procedure for image decomposition.

    Fig. 2 Brightness function proposed by Stevens.

    Fig. 3 Perceived extreme values for adapting luminance.

    Fig. 4 Contrast ratio in luminance for adapting luminance.

    Optical Engineering 113111-2 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • ratio at a high adapting luminance. In addition, to apply thisnonlinearity to Bartleson–Breneman’s functions, Lee et al.proposed visual gamma estimation for varying adaptationshown in Fig. 5. This shows that the exponent of the intensityfunction increases with increasing adapting luminance.Photographic images require gamma correction based onthe estimated visual gamma.

    Additionally, we examine the properties of the CSF ofhuman vision. The CSF specifically refers to the relationbetween contrast sensitivity and spatial frequency. In gen-eral, the CSF is measured by grating patterns that havechangeable contrast and spatial frequencies. Contrast sensi-tivity is an inverse of the detection threshold where the con-trast of a grating pattern cannot be perceived.20 As a relatedstudy, there is the experiment of van Meeteren and Vos.16

    According to their experiment, human vision is more sensi-tive to the contrast of grating patterns in high adapting lumi-nance. Furthermore, for higher adapting luminance, contrastsensitivity is saturated. Figure 6 shows the results of vanMeeteren and Vos. The CSF has band-pass shape and themaximum value of the CSFs increases for higher adaptingluminance.

    3 Proposed AlgorithmThe proposed tone mapping method is integrated with theprocedure for baseline encoding in JPEG to ensure thatthe input for tone mapping is not degraded. The methodis located between DCT and quantization in JPEG encoding.The overview of the proposed method is shown in Fig. 7. Aninput image has RGB color channels and color conversionfrom RGB to YCbCr.21 In the compression field, the com-ponent Ydct is decomposed into Surrounddct, Basedct, andDetaildct. Surrounddct is necessary to calculate the localadapting luminance La. Basedct and Detaildct represent thebase layer and detail layer, respectively. After mask-basedimage decomposition, Base is developed into tmBase by

    Fig. 5 Visual gamma in complex fields for adapting luminance.

    Fig. 6 Contrast sensitivity function proposed by van Meeteren and Vos.

    Fig. 7 Framework of the proposed algorithm.

    Optical Engineering 113111-3 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • applying luminance-adaptive tone mapping functionsaccording to the Surround value; then, Detaildct and tmBaseare enhanced with respect to sharpness, gamma, and color.The enhanced Detail 0dct and tmBase

    0dct are combined into

    Y 0dct, which can continue with the JPEG baseline. The chro-minance components CbCr are simply compensated by colorgain Gc.

    3.1 Mask-Based Image Decomposition

    Adaptive spatial filtering, such as the bilateral filter and sub-band coding by Laplacian pyramid, have been introducedfor image decomposition.13,22–24 These methods usuallyhave computational complexity and calculation burdens.4

    In JPEG baseline, DCT coefficients include informationfor extracting sub-band images, which represent detail,base, and surround.25 In this study, a simple method of imple-mentation for image decomposition in the compressed fieldis proposed. For detail preservation, an image is separatedinto two layers: detail layer and base layer, which representthe local texture and large features, respectively. The detaillayer represents local high-frequency components and thebase layer represents low-frequency components locally.This is shown in Fig. 8. The input image is decomposedusing the bilateral filter. It shows the characteristics of thebase layer and detail layer, which represent locally blurredimages and local textures, respectively.

    In JPEG baseline, DCT coefficients are computed withinan 8 × 8 pixel block. An image is converted from the spatialdomain to the frequency domain with an 8 × 8 pixel blocksize. Thus, it is possible to separate frequency componentslocally by splitting DCT coefficients in the block. Figure 9(a)shows the DCT block and location of coefficients for bandsplitting. The top-left coefficient is a direct current (DC)component of the block image. We assign a DC componentto the surround layer, set DC and the low-frequency compo-nents into the base layer and high-frequency componentsinto the detail layer. This image decomposition is imple-mented with a masking method. Figures 9(b) and 9(c) aremacro masks for extracting the base layer and detail layer.The use of DCT allows the integration of local tone mappingin JPEG baseline.

    For analysis of the proposed DCT mask, we compare aDCT mask and bilateral filtering to separate base and detaillayers. Detail layer images through bilateral filtering andDCT mask splitting are shown in Fig. 10. Blurring is distrib-uted throughout strong edges across foreground regions(trees) and background regions (sky and grass). As a result,a halo artifact occurs around strong edges. Blurred white out-lines near edges in the detail layer from bilateral filtering

    causes the halo artifact. In contrast, because a proposed detailseparation is conducted in an 8 × 8 DCT block, the regionwhere halo artifacts appear could not be larger than an 8 ×8 block. For pair comparison, tone mapped images are pro-duced by the same tone mapping operator (TMO) for eachseparated base layer. As shown in Fig. 11, the halo artifactsdifferentially appear in result images. A DCT maskingmethod, consequently, leads to diminished halo effects.

    3.2 Luminance-Adaptive Tone Mapping forthe Base Layer

    Some tone mapping functions are based on an electrophysio-logical model that predicts the response of photoreceptors(rods and cones) at any adaptation level.6 This has usuallybeen adopted by other authors to model perceived bright-ness; also, our tone mapping function is based on thismodel. The shapes of functions are similar to an S-shapedcurve in the logarithm domain.26,27 The basis function isgiven by

    Iout ¼ðIinÞn

    ðIinÞn þ C; (1)

    where Iin and Iout are the intensities of input and outputimages, respectively, and n and C are the parameters forthe S-shaped formulation.

    This basis function has been inspired by a power-functionresponse in CIECAM02 (Ref. 28), which presents the post-adaptation nonlinearities of cone responses. Figure 12 shows

    Fig. 8 Images decomposed using the bilateral filter: (a) input image, (b) base layer, and (c) detail layer.

    Fig. 9 Mask-based image decomposition in 8 × 8 discrete cosinetransform (DCT) block: (a) 8 × 8 DCT block, (b) mask for baselayer, and (c) mask for detail layer.

    Optical Engineering 113111-4 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • a relation between luminance intensity and cone responsesfor different adapting luminances. If the adapting luminanceis high, cone responses are right-shifted. Cones change theirsensitive area to a higher-intensity region for a higher adapt-ing luminance. This processing, called as luminance adapta-tion, is enacted in local cones on the retina. This is the reasonthat cone responses are applicable in local tone mapping.

    A proposed tone mapping function follows the basis func-tion, Eq. (1), with luminance adaptation processing. FromEqs. (2) to (6), analyzed brightness sensitivity propertiesare applied to local tone mapping. The proposed parametersare based on Stevens’ brightness functions and the analysisconducted by Lee et al. Tone mapping functions are given by

    tmBase ¼ 100ðαβBaseÞp

    ðαβBaseÞp þ δ; (2)

    αβ ¼ βðα − αmÞ þ αm; (3)

    α ¼ 0.003Lw

    �La

    maxðLaÞ�0.1

    ; (4)

    β ¼�Lm for Lm ≤ 0.51 − Lm for Lm > 0.5

    ; (5)

    δ ¼�0.39 expð5.80LmÞ − 0.37 for Lm ≤ 0.56.93 lnðLm − 0.41Þ þ 23.4 for Lm > 0.5 ; (6)

    where αβ is a compression level factor, δ is a luminance levelfactor, α is a contrast sensitivity factor, β is a weightingfactor, and αm is a mid-point value of α. Lw is a localwhite luminance map, La is a local adapting luminance map,and Lm is relative global luminance, which is an averagevalue of normalized La. p accounts for the slope of the func-tion, which is user-controllable and experimentally defined

    Fig. 10 Detail layer images (a) by a bilateral filter and (b) by a proposed DCT mask splitting.

    Fig. 11 Decomposed intensity images with tonemapped base and preserved detail (a) by a bilateral filterand (b) by a proposed DCT mask splitting.

    Fig. 12 Cone responses according to adapting luminance.

    Optical Engineering 113111-5 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • as 0.6 here. It is similar to that of CIECAM02 but modifiedfor higher overall contrast.

    In Eq. (3), the compression level factor, αβ, is defined asthe degree of local compression. It has been formulated onlyfor controlling the compression level without overall tonechanging. In the threshold luminance analysis shown inFig. 3, for identical contrast ratio perception under separateadapting luminance, the physical contrast ratio must increasefor a lower adapting luminance. This is because the humanvisual system is more sensitive to luminance change whenthe adapting luminance is lower, so a higher physical contrastratio is needed to keep local detail consistent in a dim sur-round viewing. In other words, to perceive a consistentbrightness contrast regardless of variations in the adaptingluminance, the image contrast should change in an exponen-tial decay toward the higher adapting luminance. First, a con-trast sensitivity factor, α, which is a basic factor of αβ,determines the contrast range of the image. It is derivedfrom the physical contrast ratio according to the relativeadaptation luminance at each white luminance in Fig. 4. Itdepends on the property that a high contrast ratio is requiredat a low adapting luminance, whereas at a high adaptingluminance, a relatively low contrast ratio is sufficient. Lais set to 0.2 Lw. Lw is obtained from a Gaussian-blurredintensity image in which the max luminance is set to2000 cd∕m2 for outdoor scenes. In order to calculate Lw,DC coefficients in DCT blocks are used, which are repre-sented as Surrounddct in Fig. 6. Then, α is weighted by factorβ of Eq. (5). A weighting factor, β, is designed to meetthe compression balance and prevent intensity saturation athigher La or gray out at lower La. β restricts a compressionrange at higher and lower adapting luminances. Here, toreduce the effect on the average luminance of an imageby β, a mid-point value αm of the overall α is fixed.

    In Eq. (6), the luminance level factor, δ, is designed toproperly adjust an average luminance level of a resultingimage, based on the analysis of average luminance for con-sistent brightness perception in Fig. 13. In viewing sceneswith uniform luminance distribution from dark to bright,the average luminance values will have a linear relationshipwith the adapting luminance, which is defined as ∼20% ofthe white luminance value of each scene. However, the

    human visual system has a nonlinearity property to perceiveaverage luminance for adapting luminance. Figure 13 showsthe difference between the physical average luminance andthe perceived average luminance. A bold line representsmedian luminance values from visual threshold luminancevalues of Fig. 3 and a dashed line shows the physical averageluminance for a uniform luminance distribution. From theanalysis, although adapting luminance linearly changes,the perceived average luminance is not proportional to thechanging ratio of the adapting luminance. This means thatas the adapting luminance is lowered, human visions needa relatively higher average luminance than physical lumi-nance to preserve average brightness; on the other hand, arelatively lower average luminance is needed for higheradapting luminance. The larger δ generates a lower averageluminance level in the output image. On the contrary, if animage is exposed for a short period, the small δ makes theoutput image brighter. The parameter, δ, is derived based onthe ratios between the values of bold and dashed lines forvarious adapting luminances; then it is adjusted using imageswith broad adapting luminance ranges.

    Figure 14 provides the resulting images with different val-ues of αβ and δ. First, the compression level factor αβ con-trols the overall dynamic range of the image. For a higher αβ(weighting factor β: 2.5), the dynamic range of a representedimage is more compressed (the bright portions are dimmedand the dark portions are lightened), whereas for lower val-ues of αβ (weighting factor β: 0.5), the compression is lower.Figure 14(c) has a smaller αβ value than Fig. 14(b), and thedynamic range of Fig. 14(c) is larger than Fig. 14(b). αβ isformulated for applying the visual contrast characteristic toa tone mapping function according to Lee’s analysis forStevens’ brightness function, which is shown in Fig. 4.Human vision requires a higher contrast ratio at a relativelylower adapting luminance,. Second, a luminance level factorδ effectively corrects an underexposed or overexposedimage. The represented image is toned down for a highmean value of the input image based on the experimentalresults of Fig. 3. As shown in Figs. 14(d) and 14(e), thechange of δ affects the average luminance of the outputimage. Based on this analysis, the factor δ is formulatedfor cooperating subjective experiments. These two fittedfunctions in Eqs. (5) and (6) are shown in Fig. 15.

    4 Additional Processing for Image Enhancement:Gamma, Sharpness, and Color

    4.1 Visual Gamma Correction

    Overall tone reproduction through TMOs changes brightnesscontrast in images and the perceived lightness (or relativebrightness) also changes as a function of different surroundluminance.1,19 In the experimental results of Bartleson andBreneman for complex stimuli, the exponent of the lightnessfunction increases with increasing adapting luminance, sophotographic images require gamma correction based onthe estimated visual gamma. Photographic images typicallyviewed in dim surroundings are reproduced using a powerfunction with a lower exponent value. Based on this, Leeet al. proposed the visual gamma given by

    γ ¼ 0.173 expð0.31 log LaÞ þ 0.329: (7)Fig. 13 Average luminance analysis for consistent brightnessperception.

    Optical Engineering 113111-6 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • The visual gamma as a function of the adapting luminancemeans that gamma correction should be conducted adaptivelywith local luminance as for human vision. Therefore, we adoptthe visual gamma as postprocessing after the proposed tonemapping. The output of tone mapping, tmBase, is gammacorrected according to the following equation:

    tmBase 0 ¼ maxðBaseÞ�

    tmBasemaxðBaseÞ

    �γ

    : (8)

    4.2 Sharpness Enhancement

    In order to compensate sharpness loss by the procedure forJPEG baseline, we apply CSF-based sharpening gain, Rcsf , toan existing mask-based unsharpening method. CSF refers tothe reciprocal of the minimum contrast ratio that humanvision can perceive at each spatial frequency. In JPEG base-line, the sharpness enhancement is applied adaptively byluminance adapting the CSF properties based on the mask-based sharpness filter. We consider the contrast sensitivity of

    Fig. 14 Toning results: (a) intensity image, (b) β ¼ 2.5, δ ¼ 0.1, (c) β ¼ 0.5, δ ¼ 0.1, (d) δ ¼ 0.1, β ¼ 0.5,and (e) δ ¼ 4, β ¼ 0.5.

    Fig. 15 (a) Contrast sensitivity factor, α, for adapting luminance La. (b) Luminance level factor, δ, forrelative Lm.

    Optical Engineering 113111-7 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • human vision for which a high contrast sensitivity means thatobjects are clearly visible. In order to design Rcsf , we com-pute the relative contrast sensitivity as a function of adaptingluminance using certain maximum values at each adaptingluminance: 5, 50, 100, 500, 1000, and 2000 cd∕m2. Themaximum contrast sensitivity at 5 cd∕m2 is set as a referencepoint. Figure 16 shows the ratio of maximum values to thereference at each adapting luminance. The proposed gain,Rcsf , is fitted with a rational function using these points,which is given by

    Rcsf ¼2.4La þ 26.9La þ 33.7

    : (9)

    The basis of the sharpness mask, Hðu; vÞ, is shown inFig. 17.29 The final sharpness enhancement using the CSFproperties is given as follows:

    Hcsfðu; vÞ ¼ Rcsf ½Hðu; vÞ − 1� þ 1; (10)

    Detail 0dct ¼ Detaildct �Hcsf ; (11)

    where Detaildct is a detail layer that is decomposed usinga mask-based image decomposition.

    4.3 Color Compensation

    Generally, during the tone mapping with a simplifieds-curve, the ratio of RGB signals changes and color

    saturation would be reduced.30 Although local tone mappingis applied to only the luminance channel, dynamic rangecompression generally results in an alteration of the ratioof chromatic channels and a reduction of color saturation.To correct this chronic defect of tone mapping, we adopta simple method for color compensation which restoresa ratio of color to luminance before tone mapping.30 Thismethod is given by

    Cb 0 ¼ μGcðCb − 128Þ þ 128; (12)

    Cr 0 ¼ μGcðCr − 128Þ þ 128; (13)

    Gc ¼tmBase 0

    Base; (14)

    where the color gain Gc is designed to preserve the ratio anda user-controlled factor μ, which prevents oversaturation, isexperimentally determined as 0.45. In our experiment, theuser-controlled factor of 0.45 is set for minimizing modifiedCIEDE2000 between reference images and proposedimages. Modified CIEDE2000 considers only the hue andchroma differences between these images.31

    5 Simulations and Results

    5.1 Objective and Subjective Assessment

    To conduct quantitative comparisons of the proposed methodwith existing tone mapping methods, several image assess-ment tools were employed, including the universal imagequality index32 (UIQI), the no-reference perceptual qualityassessment33 (NRPQA), the colorfulness metric ratio34,35

    (CMR), and structural fidelity of tone mapped images36

    (S). According to the mathematical definition of UIQI, thecloser the UIQI value is to one, the better is the image quality.Unlike UIQI, NRPQA does not need a reference image, as itis aimed specifically at no-reference quality assessment ofJPEG compressed images considering blurring and blockingas the most significant artifacts. As such, it is suitable forDCT-based image evaluation. A higher NRPQA value indi-cates a better image quality. CMR indicates the extent ofcolor in the resulting image relative to the reference image.S indicates local structural fidelity measure based on struc-tural similarity (SSIM),37 which contains three comparisoncomponents: luminance, contrast, and structure. Comparedwith SSIM, the luminance comparison component is miss-ing in S since TMOs locally change original intensity.Using all four of these numerical assessments, we com-pared the proposed method with previous approaches,such as iCAM06,3 a photographic tone reproduction basedon dodging and burning with a zone system33 (PTR), inte-grated surround retinex model38 (ISRM), and retinex-basedadaptive filter method39 (RAFM).

    Resulting images for these methods are shown inFigs. 18–21. Evaluation results for the full set of imagesare shown in Fig. 22. Note that according to UIQI results,the proposed method trails PTR by a slight margin, but itis competitive with other methods. NRPQA, which is a per-ceptual assessment for JPEG compressed images, presents amore objective evaluation than UIQI. Note that the proposed

    Fig. 16 Relative unsharp gain Rcsf for adapting luminance.

    Fig. 17 Coefficients of the basis sharpness mask, Hðu; vÞ.

    Optical Engineering 113111-8 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • Fig. 18 Result images for img3.

    Fig. 19 Result images for img4.

    Fig. 20 Result images for img5.

    Optical Engineering 113111-9 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • Fig. 21 Result images for img10.

    Fig. 22 Evaluation results using three metrics: universal image quality index, no-reference perceptualquality assessment, colorfulness metric ratio, and structural fidelity of tone mapped images.

    Optical Engineering 113111-10 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • method has the best NRPQA scores among all methodstested. The best score is assigned to the most robust methodabout blurring and blocking artifacts. According to CMRscores, the proposed method shows comparatively good per-formance and no halo artifacts, as is apparent in images proc-essed using ISRM. Finally, from S scores, it is confirmed thatthe proposed method has structural fidelity equal to or higherthan those of the other methods.

    In addition to objective evaluation, we conducted the psy-chophysical experiment based on score rating. An originalimage is first presented; then reconstructed images by eachmethod on a gray background are simultaneously shown ona display device: LG 47LM6700. Participants in the experi-ment are instructed to rate a score with 0-to-10 range foreach attribute: global tone, local contrast (halo), sharpness,and color (naturalness). In the experiment, the totalnumber of collected scores is 7 ðimagesÞ × 6 ðmethodsÞ×4 ðattributesÞ × 6 ðparticipantsÞ ¼ 1008. The average scoresand standard deviations are presented as color bars anderror bars in Fig. 23, respectively. The result shows thatthe proposed method is highly rated on the psychophysicalexperiment.

    Our overall assessment, based on qualitative comparisonof the entire set of test images, confirms that the proposedtone mapping method produces colorful, high-contrastimages with strongly enhanced details. In addition, accord-ing to subjective comparison, the proposed method hasgood preference scores for four-view, such as global tone,local contrast, sharpness, and color. All resulting images bythe proposed method and the original images are shown inFig. 24.

    5.2 Computation Time

    We compute the computation time of the methods with thetest setup as shown in Fig. 25. Considering the novelty ofthe proposed method that is inserted in JPEG baseline,JPEG encoding and decoding are conducted after tonemapping for the previous methods. For different resolutions(853 × 480, 1280 × 720, 1920 × 1080), computation timesin MATLAB® are listed in Table 1 (CPU: Intel i7-2600K3.40 GHz, RAM: 4 GB). In Table 1, the computation

    Fig. 23 Subjective evaluation using preference assessment.

    Fig. 24 Result images: input images (top) and out images (bottom) bythe proposed method.

    Fig. 25 Test setup for checking runtime.

    Optical Engineering 113111-11 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

  • time of our method is faster than those of iCAM06 andRAFM, but similar to those of PTR and ISRM.Compared with iCAM06 and RAFM adopting time-consum-ing tasks for edge preserving and anti-halo, such as the bilat-eral filter and anisotropic Gaussian functions, our methodimproves the edge resolution and halo artifact while savinga lot of computation time.

    6 ConclusionsA novel approach to enhance images using tone mapping inthe compression domain was presented. In order to combinetone mapping with JPEG baseline, we decomposed an imageusing mask-based DCT band splitting and suggested theluminance-adaptive tone mapping function, which wasbased on the brightness and contrast adaptation of humanvision. For image application, we adopted the Stevens’and Bartleson and Breneman’s experimental results and cor-related analysis in order to mimic human vision properties. Inaddition, the procedure involved sharpness enhancementbased on contrast sensitivity functions and color compensa-tion. For the evaluation results, the performance of the pro-posed method was compared with previous approachesthrough several image assessment methods. It was discov-ered that the proposed method outperformed previousapproaches in most cases. For optimal tone rendering, weare certain that the proposed method can be useful in physi-cal still cameras in order to compress the dynamic range inJPEG baseline.

    AcknowledgmentsThis research was supported by Basic Science ResearchProgram through the National Research Foundation ofKorea (NRF) funded by the Ministry of Education (2012-R1A1A2008362).

    References

    1. M. D. Fairchild, Color Appearance Models, 2nd ed., John Wiley &Sons, Ltd., Chichester, United Kingdom (2005).

    2. E. Reinhard et al.,High Dynamic Range Imaging: Acquisition, Displayand Image-based Lighting, Morgan Kaufmann, San Francisco (2005).

    3. J. Kuang, G. Johnson, and M. Fairchild, “iCAM06: a refined imageappearance model for HDR image rendering,” J. Vis. Commun.Image Represent. 18(5), 406–414 (2007).

    4. J. Xiao et al., “Hierarchical tone mapping based on image colourappearance model,” IET Comput. Vis. 8(4), 358–364 (2014).

    5. F. Drago et al., “Adaptive logarithmic mapping for displaying highcontrast scenes,” Comput. Graph. Forum 22(3), 419–426 (2003).

    6. P. Ledda, L. P. Santos, and A. Chalmers, “A local model of eye adap-tation for high dynamic range images,” in Proc. ACM AFRIGRAPH‘04, pp. 151–160, ACM, New York (2004).

    7. G.W. Larson, H. Rushmeier, and C. Piatko, “Avisibility matching tonereproduction operator for high dynamic range scenes,” IEEE Trans.Vis. Comput. Graph. 3(4), 291–306 (1997).

    8. S. B. Kang et al., “High dynamic range video,” ACM Trans. Graph.22(3), 319–325 (2003).

    9. C. Lee and C.-S. Kim, “Gradient domain tone mapping of highdynamic range videos,” in IEEE Int. Conf. on Image Processing,Vol. 3, pp. III-461–III-464, IEEE, New York (2007).

    10. R. Boitard et al., “Temporal coherency in video tone mapping,” Proc.SPIE 8499, 84990D (2012).

    11. Y. Chen et al., “Intra-and-inter-constraint-based video enhancementbased on piecewise tone mapping,” IEEE Trans. Circuits Syst.Video Technol. 23(1), 74–82 (2013).

    12. P. Ledda et al., “Evaluation of tone mapping operators using a highdynamic range display,” ACM Trans. Graph. 24(3), 640–648 (2005).

    13. F. Durand and J. Dorsey, “Fast bilateral filtering for the display ofhigh-dynamic-range images,” ACM Trans. Graph. 21(3), 257–266(2002).

    14. J. C. Stevens and S. S. Stevens, “Brightness function: effects of adap-tation,” J. Opt. Soc. Am. 53(3), 375–385 (1963).

    15. C. J. Bartleson and E. J. Breneman, “Brightness perception in complexfields,” J. Opt. Soc. Am. 57(7), 953–956 (1967).

    16. A. van Meeteren and J. J. Vos, “Resolution and contrast sensitivity atlow luminance,” Vis. Res. 12(5), 825–833 (1972).

    17. C. J. Bartleson, “Predicting corresponding colors with changes inadaptation,” Color Res. Appl. 4(3), 143–155 (1979).

    18. E. G. Heinemann and S. Chase, “A quantitative model for simultane-ous brightness induction,” Vis. Res. 35(14), 2007–2020 (1995).

    19. S. H. Lee et al., “The quantitative model of optimal threshold andgamma of display using brightness function,” IEICE Trans. Fundam.Electron., Commun. Comput. Sci. E89-A(6), 1720–1723 (2006).

    20. S. Westland et al., “Model of luminance contrast-sensitivity functionfor application to image assessment,” Color Res. Appl. 31(4), 315–319(2006).

    21. G. K. Wallace, “The JPEG still picture compression standard,”Commun. ACM 34(4), 30–44 (1991).

    22. E. H. Adelson et al., “Pyramid methods in image processing,” RCAEng. 29(6), 33–41 (1984).

    23. J. Shen et al., “Exposure fusion using boosting Laplacian pyramid,”IEEE Trans. Cybern. 44(9), 1579–1590 (2014).

    24. J. Shen, Y. Zhao, and X. Li, “Detail-preserving exposure fusion usingsubband architecture,” Vis. Comput. 28(5), 463–473 (2012).

    25. G. Y. Lee, S. H. Lee, and K. I. Sohng, “DBS separation and tone repro-duction using DCT band splitting for LDR image transmission,” inIEEE 2nd Global Conf. on Consumer Electronics, pp. 280–281,IEEE, New York (2013).

    26. E. Reinhard et al., “Photographic tone reproduction for digital images,”ACM Trans. Graph. 21(3), 267–276 (2002).

    27. J. Kuang et al., “Evaluating HDR rendering algorithms,” ACM Trans.Appl. Percept. 4(2) (2007).

    28. N. Moroney et al., “The CIECAM02 color appearance model,” inProc. IS&T 10th Color Imaging Conf., pp. 23–27, IS&T,Springfield, Virginia (2002).

    29. Y. K. Lee et al., “Improvement of LCDmotion blur in MPEG domain,”in Proc. of IPCV, Vol. 2, pp. 798–801, World Congress in ComputerScience Computer Engineering and Applied Computing, San Diego,California (2009).

    30. R. Mantiuk et al., “Color correction for tone mapping,” Comput.Graph. Forum 28(2), 193–202 (2009).

    31. H. J. Kwon et al., “Luminance adaptation transform based on bright-ness functions for LDR image reproduction,” Digital Signal Process.30, 74–85 (2014).

    32. Z. Wang and A. C. Bovik, “A universal image quality index,” IEEESignal Process. Lett. 9(3), 81–84 (2002).

    33. Z. Wang, H. R. Sheikh, and A. C. Bovik, “No-reference perceptualquality assessment of jpeg compressed images,” in Proc. IEEE Int.Conf. on Image Processing, pp. 477–480, IEEE, New York (2002).

    34. S. E. Susstrunk and S. Winkler, “Color image quality on the internet,”Proc. SPIE 5304, 118–131 (2004).

    35. J. Mukherjee and S. K. Mitra, “Enhancement of color images byscaling the DCT coefficients,” IEEE Trans. Image Process. 17(10),1783–1794 (2008).

    36. H. Yeganeh and Z. Wang, “Objective quality assessment of tone-mapped images,” IEEE Trans. Image Process. 22(2), 657–667 (2013).

    37. Z. Wang et al., “Image quality assessment: from error visibility tostructural similarity,” IEEE Trans. Image Process. 13(4), 600–612(2004).

    38. L. Wang, T. Horiuchi, and H. Kotera, “High dynamic range imagecompression by fast integrated surround retinex model,” J. ImagingSci. Technol. 51(1), 34–43 (2007).

    39. L. Meylan and S. Susstrunk, “High dynamic range image renderingwith a retinex-based adaptive filter,” IEEE Trans. Image Process.15(9), 2820–2830 (2006).

    Table 1 Computation time of methods in MATLAB® (in seconds).

    Methods

    853 × 480(standarddefinition)

    1280 × 720(high

    definition)

    1920 × 1080(full highdefinition)

    iCAM06 7.927 16.890 31.924

    Photographic tonereproduction

    3.421 8.125 17.964

    Integrated surroundretinex model

    4.774 9.383 21.045

    Retinex-based adaptivefilter method

    61.162 63.060 69.273

    Proposed 3.873 9.054 20.264

    Optical Engineering 113111-12 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

    http://dx.doi.org/10.1016/j.jvcir.2007.06.003http://dx.doi.org/10.1016/j.jvcir.2007.06.003http://dx.doi.org/10.1049/iet-cvi.2013.0230http://dx.doi.org/10.1111/1467-8659.00689http://dx.doi.org/10.1109/2945.646233http://dx.doi.org/10.1109/2945.646233http://dx.doi.org/10.1145/882262http://dx.doi.org/10.1117/12.929600http://dx.doi.org/10.1117/12.929600http://dx.doi.org/10.1109/TCSVT.2012.2203198http://dx.doi.org/10.1109/TCSVT.2012.2203198http://dx.doi.org/10.1145/1073204http://dx.doi.org/10.1145/566654.566574http://dx.doi.org/10.1364/JOSA.53.000375http://dx.doi.org/10.1364/JOSA.57.000953http://dx.doi.org/10.1016/0042-6989(72)90008-9http://dx.doi.org/10.1002/(ISSN)1520-6378http://dx.doi.org/10.1016/0042-6989(94)00281-Phttp://dx.doi.org/10.1093/ietfec/e89-a.6.1720http://dx.doi.org/10.1093/ietfec/e89-a.6.1720http://dx.doi.org/10.1002/(ISSN)1520-6378http://dx.doi.org/10.1145/103085.103089http://dx.doi.org/10.1109/TCYB.2013.2290435http://dx.doi.org/10.1007/s00371-011-0642-3http://dx.doi.org/10.1145/566654.566575http://dx.doi.org/10.1145/1265957http://dx.doi.org/10.1145/1265957http://dx.doi.org/10.1111/cgf.2009.28.issue-2http://dx.doi.org/10.1111/cgf.2009.28.issue-2http://dx.doi.org/10.1016/j.dsp.2014.03.008http://dx.doi.org/10.1109/97.995823http://dx.doi.org/10.1109/97.995823http://dx.doi.org/10.1117/12.537804http://dx.doi.org/10.1109/TIP.2008.2002826http://dx.doi.org/10.1109/TIP.2012.2221725http://dx.doi.org/10.1109/TIP.2003.819861http://dx.doi.org/10.2352/J.ImagingSci.Technol.(2007)51:1(34)http://dx.doi.org/10.2352/J.ImagingSci.Technol.(2007)51:1(34)http://dx.doi.org/10.1109/TIP.2006.877312

  • Geun-Young Lee received his BS and MS degrees in electronicsengineering from Kyungpook University, Daegu, Republic of Korea,in 2011 and 2013, respectively. He is currently pursuing his PhDdegree from Kyungpook National University, Daegu. His researchinterests include image and signal processing.

    Sung-Hak Lee received his BS, MS, and PhD degrees inelectronics engineering from Kyungpook National University in1997, 1999, and 2008, respectively. He worked at LG Elec-tronics from 1999 to 2004 as a senior research engineer. He hasworked at the School of Electronics Engineering of KyungpookNational University as a research professor. His research fieldhas been in color management, color appearance model, colorimage processing, and display applications for human visualsystem.

    Hyuk-Ju Kwon received his BS and MS degrees in electronics engi-neering from Kyungpook University, Daegu, Republic of Korea, in2010 and 2012. He is currently pursuing his PhD degree fromKyungpook National University, Daegu. His research interests includeimage and signal processing.

    Kyu-Ik Sohng is a professor at the School of Electronics Engineeringof Kyungpook National University, Daegu, Republic of Korea. Hereceived his BS and MS degrees in electronics engineering fromKyungpook National University, Daegu, in 1973 and 1975, respec-tively, and his PhD degree in electronics engineering from TohokuUniversity, Sendai, Japan, in 1990. His current research interestsinclude audio and video signal processing, color reproduction engi-neering, digital television, display and health, and automotive electron-ics engineering.

    Optical Engineering 113111-13 November 2014 • Vol. 53(11)

    Lee et al.: Visual sensitivity correlated tone reproduction for low dynamic range images. . .

    Downloaded From: https://www.spiedigitallibrary.org/journals/Optical-Engineering on 30 Mar 2021Terms of Use: https://www.spiedigitallibrary.org/terms-of-use