doctoral dissertation by marius pedersen

394
Image quality metrics for the evaluation of printing workflows Doctoral Dissertation by Marius Pedersen Submitted to the Faculty of Mathematics and Natural Sciences at the University of Oslo in partial fulfillment of the requirements for the degree Philosophiae Doctor (PhD) in Color Imaging

Upload: others

Post on 13-Jan-2022

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Doctoral Dissertation by Marius Pedersen

Image quality metricsfor the evaluation of printing workflows

Doctoral Dissertation by

Marius Pedersen

Submitted to the Faculty of Mathematics and Natural Sciences at theUniversity of Oslo in partial fulfillment of the requirements for the degree

Philosophiae Doctor (PhD) in Color Imaging

Page 2: Doctoral Dissertation by Marius Pedersen

© Marius Pedersen, 2011 Series of dissertations submitted to the Faculty of Mathematics and Natural Sciences, University of Oslo No. 1124 ISSN 1501-7710 All rights reserved. No part of this publication may be reproduced or transmitted, in any form or by any means, without permission. Cover: Inger Sandved Anfinsen. Printed in Norway: AIT Oslo AS. Produced in co-operation with Unipub. The thesis is produced by Unipub merely in connection with the thesis defence. Kindly direct all inquiries regarding the thesis to the copyright holder or the unit which grants the doctorate.

Page 3: Doctoral Dissertation by Marius Pedersen

Abstract

The aim of our research is to assess image quality of prints without the involvement of hu-man observers. The printing industry is continuously moving forward as new products andtechnologies are introduced to the market. The need to assess the quality is increased withthis growth, for example, to verify that these technology advancements lead to higher qualityprints.

In this thesis we describe the work carried out to use image quality metrics for the evalua-tion of printed images. The intended behavior of such metrics is to measure or predict imagequality as human observers would perceive it.

Following an introduction and background on quality assessment, we introduce the con-cept of image quality metrics. Existing image quality metrics are classified and a survey ofthem is given, to show how they are constructed and their differences. Following the survey,a new image quality metric, the Spatial Hue Angle MEtric (SHAME) is proposed, which ac-counts for two key aspects of image quality, namely region of interest and the human visualsystem. The evaluation of image quality metrics against the percept is a key aspect for en-suring that the metrics can substitute or assist human observers in the assessment of quality.Therefore, existing evaluation methods are presented and analyzed, revealing the need for anew method to assess the overall performance of image quality metrics. For that reason, anew method to evaluate overall performance is proposed, based on the rank order method.Using existing methods and the new evaluation method, a set of commonly used metrics areevaluated with a set of public databases. These databases contain digital images, with a rangeof different distortions and quality issues, and have quality ratings from human observers.

The knowledge gathered from the evaluation of image quality metrics on the existingdatabases was then applied to the assessment of printed images. Since the metrics require dig-ital images as input, a framework to digitize printed images is proposed. Using this frameworka set of metrics is evaluated against human observers, which shows that none of the existingmetrics predict overall image quality adequately. These findings lead us to break overall im-age quality into parts, more precisely quality attributes. Based on existing quality attributesa manageable set of six color printing quality attributes is proposed. The final set included:sharpness, color, lightness, contrast, artifacts, and physical. Through two experimental val-idation procedures these attributes are found to be a good foundation for the evaluation ofcolor prints. The image quality metrics are then used to describe each of the proposed qual-ity attributes separately to find the most appropriate metrics to measure the quality of eachattribute. Two experiments with human observers were carried out, which acted as the basisfor the evaluation and selection of metrics. The results show that for some attributes, such assharpness, suitable metrics can be found, but additional work is needed to find metrics thatcorrelate well with the percept for all of the attributes.

An area that may be improved with the use of image quality metrics is the reductionof quality values to a more manageable number (pooling), usually a single quality value. Wehave investigated the impact of pooling strategies on the performance of image quality metrics.This investigation shows that the performance is linked to the metric, and that the parametersfor the pooling strategies are very important. Even with the effort spent on pooling strategiesnone of the evaluated metrics performed well for the color quality attribute. This lead toa proposal for a new image quality metric designed for the color attribute, Total Variationof Difference (TVD) metric, which applies a spatial filtering to simulate the human visualsystem before quality is calculated. A comparison against the state of the art metrics showsan increased performance for the new metric.

III

Page 4: Doctoral Dissertation by Marius Pedersen

Lastly, we gather the work carried out in this thesis to develop a practical tool for theprinting industry to assess the quality of prints using image quality metrics, named the QualityAssistant. The Quality Assistant consists of all functions needed to evaluate quality, including:a test image suite, the framework for digitizing the prints, a set of image quality metrics, andvisualization tools. Through the work carried out in this thesis we have shown the applicabilityof image quality metrics for the evaluation of printing workflows.

IV

Page 5: Doctoral Dissertation by Marius Pedersen

Acknowledgements

There are many people who deserve to be acknowledged, without their support and help itwould not have been possible to complete this work. First of all, I would like to thank mymain supervisor, professor Jon Yngve Hardeberg from Gjøvik University College, who intro-duced me to the field of color imaging. Many thanks goes to my second supervisor ProfessorFritz Albregtsen from from the Department of Informatics at the University of Oslo, andNicolas Bonnier, my industry supervisor from Oce Print Logic Technologies. Their help andassistance has been invaluable. My deepest gratitude to all of you for your guidance andsupport during this work.

I would also like to thank the members of the Norwegian Color Research Laboratory atGjøvik University College: Peter Nussbaum, Aditya Sole, Ivar Farup, Gabriele Simone, ArneMagnus Bakke, Dibakar Raj Pant, Zhaohui Wang, Raju Shresta, and others.

Many of the employees at Oce have helped during the last years, and in particular I wouldlike to thank Christophe Leynadier, Jacques Perville, Arlette Del-Aguila, Kristyn Falken-stern, Medhi Felhi, Albrecht Lindner, Maria-Valezzka Ortiz-Segovia, Reneka Even, ChristineHurtret, and Marine Lauga. Many thanks to Oce for the pleasant work atmosphere in Creteil.Thanks to Frans Gaykema of Oce Venlo for his support and discussions.

Thanks also goes out to Zofia Baranczuk, Timothee Royer, Sebastien Akli Ajagamelle, AliAmirshahi, Eriko Bando, Faouzi Alaya Cheikh, Alessandro Rizzi, Guanqun Cao, and manyothers for their co-operation, help, and assistance.

My deepest appreciation is due to my family and friends, their support has been invalu-able. Finally, I would like to express my gratitude to Lene, who has borne my increasingabsence with good will.

Thank you all.

Gjøvik, NorwayAugust 2011Marius Pedersen

V

Page 6: Doctoral Dissertation by Marius Pedersen

VI

Page 7: Doctoral Dissertation by Marius Pedersen

CONTENTS

1 INTRODUCTION 31.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31.2 Aims . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41.3 Research methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41.4 Publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51.5 Thesis outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

PART I BACKGROUND 11

2 WHAT IS QUALITY? 132.1 Quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132.2 Image quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132.3 Print quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142.4 Definition used in this work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

3 EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY 153.1 Objective quality assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

3.1.1 Measuring quality using instruments. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153.1.1.1 Densitometer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153.1.1.2 Colorimeter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163.1.1.3 Spectroradiometer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183.1.1.4 Spectrophotometer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183.1.1.5 Glossmeter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18

3.1.2 Quality evaluation in the industry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183.1.2.1 Process-standard offset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

3.1.3 Standards for printing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203.1.3.1 ISO 12647 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203.1.3.2 USA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

Specifications for Web Offset Publications . . . . . . . . . . . . . . . . . . . . . . . 20General Requirements and Applications for Commercial Off-

set Lithography. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203.1.3.3 Germany . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

Der Bundesverband Druck und Medien e.V. . . . . . . . . . . . . . . . . . . . . . 203.1.3.4 Switzerland. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

System Brunner . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213.1.4 Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

3.1.4.1 Altona test suite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213.1.4.2 Roman16 bvdm reference images. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

VII

Page 8: Doctoral Dissertation by Marius Pedersen

CONTENTS

3.1.4.3 Visual print reference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223.1.4.4 Common test charts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

MacBeth ColorChecker color rendition chart . . . . . . . . . . . . . . . . . . . . 23IT8 test charts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23ECI 2002 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23Ugra/Fogra Media Wedge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23TC3.5 CMYK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25Contrast resolution test target . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25Addressability test target . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25Edge characteristics targets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

3.1.5 Measuring quality using algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273.2 Subjective quality assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

3.2.1 Psychophysical thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273.2.1.1 Method of adjustment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

Example method of adjustment: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283.2.1.2 Method of limits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

Example method of limits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31Improvements of the method of limits . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

3.2.1.3 Method of constant stimuli . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35Example method of constant stimuli . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

3.2.2 Psychophysical scaling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 363.2.2.1 Pair comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

Example pair comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 383.2.2.2 Category judgment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

Example category judgment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 433.2.2.3 Rank order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45

Example rank order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 453.2.3 Other methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

3.2.3.1 Mean opinion score . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 463.2.3.2 Double stimulus impairment scale . . . . . . . . . . . . . . . . . . . . . . . . . . . . 473.2.3.3 Double stimulus continuous quality scale . . . . . . . . . . . . . . . . . . . . . 473.2.3.4 Triplet comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 483.2.3.5 Quality ruler. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49

3.2.4 Experiment aspects. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 503.2.4.1 Number of observers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 503.2.4.2 Observer type . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513.2.4.3 Observer characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513.2.4.4 Experiment duration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513.2.4.5 Number of stimuli . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513.2.4.6 Type of stimuli . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 523.2.4.7 Standard test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

CIE recommendation for gamut mapping . . . . . . . . . . . . . . . . . . . . . . . . 52ISO 12640 test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52Canon Development Americas computer graphics images . . . . . . 53Sony sRGB standard images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53Kodak lossless true color image suite. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54DigiQ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

3.2.4.8 Stimuli labeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

VIII

Page 9: Doctoral Dissertation by Marius Pedersen

CONTENTS

3.2.4.9 Viewing conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56Controlled and uncontrolled environments . . . . . . . . . . . . . . . . . . . . . . . 56Viewing distance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56Viewing illumination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.2.4.10 Instructions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 573.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57

PART II IMAGE QUALITY METRICS 59

4 EXISTING IMAGE QUALITY METRICS 614.1 Classification of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63

4.1.1 Existing classification of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . 634.1.2 Proposal for classification of image quality metrics . . . . . . . . . . . . . . . . . . . . . . 64

4.2 Survey of existing image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 654.2.1 Mathematically based metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.2.1.1 Mean squared error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 664.2.1.2 ΔE∗

ab . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 664.2.1.3 ΔEE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 674.2.1.4 Difference between the metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

4.2.2 Low-level based metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 694.2.2.1 S-CIELAB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 694.2.2.2 S-DEE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 704.2.2.3 Adaptive image difference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 714.2.2.4 Comparison of selected metrics within the group . . . . . . . . . . . . . 72

4.2.3 High level based metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 724.2.3.1 SSIM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 724.2.3.2 Visual information fidelity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 744.2.3.3 Comparison of selected metrics within the group . . . . . . . . . . . . . 75

4.2.4 Other approaches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 764.2.4.1 Visual signal to noise ratio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 764.2.4.2 Color image similarity measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 764.2.4.3 Comparison of selected metrics within the group . . . . . . . . . . . . . 76

4.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

5 SHAME: A NEW IMAGE QUALITY METRIC 795.1 The hue angle algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 795.2 Spatial filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 805.3 Applying spatial filtering to the hue angle algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 825.4 Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 835.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83

6 HOW TO EVALUATE IMAGE QUALITY METRICS 856.1 Evaluation methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 856.2 Existing evaluation methods. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85

6.2.1 Correlation based methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 866.2.1.1 Pearson product-moment correlation coefficient . . . . . . . . . . . . . . 86

Significance test. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86Confidence intervals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

IX

Page 10: Doctoral Dissertation by Marius Pedersen

CONTENTS

6.2.1.2 Spearman’s rank-correlation coefficient . . . . . . . . . . . . . . . . . . . . . . . 876.2.1.3 Kendall tau rank correlation coefficient . . . . . . . . . . . . . . . . . . . . . . . 87

6.2.2 Error methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 896.2.2.1 Root mean squared error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 896.2.2.2 Mean absolute error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89

6.2.3 Outlier ratio. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 896.3 New method for the evaluation of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . 89

6.3.1 Example of the proposed method based on rank order . . . . . . . . . . . . . . . . . . . 916.3.1.1 Psychophysical experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 916.3.1.2 Z-scores . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 926.3.1.3 Selected image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 926.3.1.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93

6.3.2 Rank order based on metric order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 936.3.2.1 Overall observations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97

6.4 Which methods will we use for the evaluation of image quality metrics . . . . . . . . 986.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98

7 EVALUATION OF IMAGE QUALITY METRICS 997.1 TID2008 Database, Ponomarenko et al. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1017.2 Luminance changed images, Pedersen et al. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1077.3 JPEG and JPEG2000 compressed images, Simone et al. . . . . . . . . . . . . . . . . . . . . . . . . . 1077.4 IVC database, Le Callet et al. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1117.5 Contrast, lightness, and saturation alterations, Ajagamelle et al. . . . . . . . . . . . . . . . . . 1117.6 Gamut mapped images, Dugay et al. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1137.7 Overall performance of the metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1137.8 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116

PART III IMAGE QUALITY METRICS FOR THE EVALUATION OF PRINT-ING WORKFLOWS 119

8 A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS 1218.1 State of the art of frameworks to digitize printed images . . . . . . . . . . . . . . . . . . . . . . . . . 1218.2 A framework based on control points . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122

8.2.1 Transformation type. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1238.2.2 Interpolation method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1248.2.3 Evaluation of different transformation types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1248.2.4 Scanner . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126

8.2.4.1 Scanning resolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1268.2.4.2 Scanner ICC profile. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1288.2.4.3 Cropping of the scan. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1288.2.4.4 Scanner rotation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128

8.2.5 A comparison against another framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1308.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131

9 EVALUATION OF METRICS USING THE FRAMEWORK 1339.1 Experimental setup. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1339.2 Psychophysical results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1359.3 Selection of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136

X

Page 11: Doctoral Dissertation by Marius Pedersen

CONTENTS

9.3.1 Overall evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1379.3.2 Image-wise evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

9.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

10 WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT? 14110.1 State of the art . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141

10.1.1 Image quality attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14210.1.2 Image quality models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142

10.2 Investigation and selection of important quality attributes . . . . . . . . . . . . . . . . . . . . . . . . 14410.2.1 Color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14610.2.2 Lightness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14710.2.3 Contrast . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14710.2.4 Sharpness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14810.2.5 Artifacts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14810.2.6 Physical . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14810.2.7 Relations between quality attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149

10.3 Validation of the quality attributes - experiment I . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14910.3.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150

10.3.1.1 Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15010.3.1.2 Color workflow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15110.3.1.3 Viewing conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15110.3.1.4 Instructions given to the observers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153

10.3.2 Perceptual results. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15310.3.3 Fitting the quality attributes to the color printing quality attributes . . . . . 15410.3.4 Relations between the attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154

10.4 Validation of the quality attributes - experiment II . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15710.4.1 How to validate quality attributes? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15810.4.2 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159

10.4.2.1 Images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15910.4.2.2 Color workflow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16010.4.2.3 Viewing conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16010.4.2.4 Instructions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160

10.4.3 Fitting the quality attribute data to the color printing quality attributes . 16010.4.3.1 Discussion on the fitting of quality attributes . . . . . . . . . . . . . . . . . 163

Overlapping QAs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163Independence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163Global and local issues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164One child with several own children . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164Skewness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165Dimensionality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165

10.4.4 Observations on the color printing quality attributes . . . . . . . . . . . . . . . . . . . . . 16510.4.5 Dependence between the attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16710.4.6 Validation summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168

10.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 169

XI

Page 12: Doctoral Dissertation by Marius Pedersen

CONTENTS

11 IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES 17111.1 Selection of image quality metrics for the color printing quality attributes . . . . . . 171

11.1.1 Sharpness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17111.1.2 Color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17211.1.3 Lightness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17211.1.4 Contrast . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17211.1.5 Artifacts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172

11.2 Experiment I . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17311.2.1 Selected image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17311.2.2 Evaluation of the selected image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . 173

11.2.2.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17411.2.2.2 Evaluation of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . 174

Phase 1: naıve observers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174Phase 2: expert observers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177

11.2.3 Investigation of image characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18011.2.3.1 Selected image characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18011.2.3.2 Experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18111.2.3.3 Dominant color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18111.2.3.4 Colorfulness. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18311.2.3.5 Lightness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18311.2.3.6 Details . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18511.2.3.7 Overall observations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186

11.3 Experiment II. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18711.3.1 Selected image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18711.3.2 Evaluation of the selected image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . 187

11.3.2.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188Printing workflow. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188Observers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188Viewing conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188Experiment procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189

11.3.2.2 Experimental results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18911.3.2.3 Evaluation of image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . 189

Preparation of the printed images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189Evaluation method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190Evaluation results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190

11.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192

12 IMPROVING PERFORMANCE OF METRICS BY POOLING 19312.1 State of the art . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195

12.1.1 General formulation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19512.1.2 Quality based pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196

12.1.2.1 Minkowski pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19612.1.2.2 Local quality/distortion-weighted pooling . . . . . . . . . . . . . . . . . . . . 197

Monotonic function pooling. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197Percentile pooling. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197

12.1.3 Content based pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19812.1.3.1 Information content-weighted pooling . . . . . . . . . . . . . . . . . . . . . . . . 198

XII

Page 13: Doctoral Dissertation by Marius Pedersen

CONTENTS

12.1.3.2 Gaze based pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198Gaze-attentive fixation finding engine . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199Saliency map computed by fixation number and duration . . . . . . . 199Gaussian weighted sum model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199

12.1.3.3 Computational saliency model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199Itti’s saliency model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199Spectral residual model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199

12.2 Experimental setup. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20012.2.1 Test sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200

12.2.1.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20012.2.1.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200

12.2.2 Image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20012.2.3 Pooling methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20112.2.4 Evaluation methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201

12.3 Experimental results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20112.3.1 Sharpness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202

12.3.1.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20212.3.1.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203

12.3.2 Color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20312.3.2.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20312.3.2.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204

12.3.3 Lightness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20612.3.3.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20612.3.3.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206

12.3.4 Contrast . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20912.3.4.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20912.3.4.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209

12.3.5 Artifacts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21112.3.5.1 Test set 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21112.3.5.2 Test set 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 211

12.3.6 Overall observations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21212.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213

13 A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE 21513.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21613.2 The new color image quality metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21713.3 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 219

13.3.1 Test data sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21913.3.2 Evaluation procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22013.3.3 Results and discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 220

13.4 Investigation of image characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22413.4.1 Experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22413.4.2 Dominant color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22413.4.3 Colorfulness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224

13.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 227

XIII

Page 14: Doctoral Dissertation by Marius Pedersen

CONTENTS

14 QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY 22914.1 Pre-metric operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229

14.1.1 Padding the images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22914.1.2 Scanning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22914.1.3 Registration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23014.1.4 Registration validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 230

14.2 Measuring attributes using image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23114.2.1 Sharpness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231

14.2.1.1 Perceptual sharpness. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232

14.2.1.2 Computational sharpness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 234

14.2.2 Artifacts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23414.2.2.1 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 234

Perceptual noise. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235Computational noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235

14.2.2.2 Banding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236

14.2.3 Color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237Test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237

14.3 Visualization of results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23714.3.1 Spider and bar plots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23714.3.2 Quality maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 238

14.3.2.1 Box plots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23814.3.2.2 3 and 4 dimensional plots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24014.3.2.3 Color histograms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240

14.3.3 Scale differences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24214.4 Overview of the quality assistant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 243

14.4.1 Show test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24314.4.2 Pad images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24314.4.3 Scan profile evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24414.4.4 Image registration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24514.4.5 Image registration validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245

14.4.5.1 Visual rotation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24514.4.6 Calculate results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24614.4.7 Show results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246

14.4.7.1 Show test images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24714.4.7.2 XLS export . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24714.4.7.3 All images printer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24714.4.7.4 All printers image. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24814.4.7.5 4d plot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248

14.5 Evaluation of the Quality Assistant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25014.5.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250

XIV

Page 15: Doctoral Dissertation by Marius Pedersen

CONTENTS

14.5.1.1 Printers and print settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25014.5.1.2 Observers and instructions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25114.5.1.3 Viewing conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25114.5.1.4 Scanning and registration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251

14.5.2 Results and discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25214.5.2.1 Sharpness. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25214.5.2.2 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25314.5.2.3 Banding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25314.5.2.4 Color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255

14.5.3 Overall observations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25514.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256

15 CONCLUSION 257

16 FUTURE WORK 26116.1 Quality Assistant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26116.2 Quality attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26216.3 Image quality metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26216.4 Image characteristics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262

NOMENCLATURE 263

REFERENCES 280

PART IV APPENDICES 319

A OVERVIEW OF IMAGE QUALITY METRICS 321

B MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK 333B.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333

B.1.1 Tadmor and Tolhurst . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333B.1.2 Rizzi et al. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334B.1.3 Retinal-like subsampling contrast . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 335

B.2 The weighted-level framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336

C SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS 339C.1 Saliency map . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339C.2 Applicability of saliency maps to gamut mapping artifacts . . . . . . . . . . . . . . . . . . . . . . . 340C.3 Experimental framework: saliency models as artifact detectors . . . . . . . . . . . . . . . . . . 340

C.3.1 Global strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342C.3.2 Local strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 343

D DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING 347D.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347D.2 Proposed error diffusion worm measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349D.3 Evaluation of the error diffusion worm measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351

D.3.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351D.3.2 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352

D.3.2.1 Overall scores . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352

XV

Page 16: Doctoral Dissertation by Marius Pedersen

CONTENTS

D.3.3 Gradient image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 353D.3.4 Artistic image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355D.3.5 Highlight image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357D.3.6 Shadow image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358D.3.7 Comparison to other metrics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358

D.3.7.1 Overall performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 360D.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362

E FROM CONTRAST TO IMAGE DIFFERENCE: TWO NEW METRICS 363

F SPECIFICATIONS 365F.1 Monitors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365

F.1.1 Eizo CG241W . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365F.1.2 Eizo CG211 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367

F.2 Scanners . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 368F.2.1 Microtek ScanMaker 9800XL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 368F.2.2 Epson Expression 10000XL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369F.2.3 HP ScanJet G4050 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 370

F.3 Printers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 371F.3.1 Oce Colorwave 600 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 371F.3.2 HP Designjet 10ps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372F.3.3 HP Color LaserJet 4600dn. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 373

F.4 Paper . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374F.4.1 Oce LFM 050 Red Label . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374F.4.2 Oce LFM 054 Red Label . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374F.4.3 Stora Enso MultiCopy Original . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375F.4.4 HP office paper . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375

F.5 Viewing cabinet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376F.5.1 Verivide DTP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376

G LEGAL 377

XVI

Page 17: Doctoral Dissertation by Marius Pedersen
Page 18: Doctoral Dissertation by Marius Pedersen

CONTENTS

2

Page 19: Doctoral Dissertation by Marius Pedersen

1 INTRODUCTION

For centuries attempts to achieve accurate reproductions of pictorial images in an efficientway have been tried on various materials, and it is still an active field. Interest in image repro-duction spreads across several occupations, such as engineers, photographers, and scientists,and industries, such as printing, painting, photography, display, and film. Even in our digitalage most people’s preference is still towards paper rather than displays [184].

Many problems are encountered while working with the reproduction of images on paper,such as the limited color gamut that a reproduction device can reproduce, paper characteris-tics, and registration issues. Managing these problems has been a key issue for color repro-duction, and has been motivation for extensive research. Since reproduction devices are notable to create a reproduction that is identical to the original, differences between the originaland reproduction will occur. These differences impact the impression of the quality of the re-production. Due to immense research, technology advancements are rapid, and new and morerefined ways to deal with the limitations of a reproduction system are proposed continuouslyin order to achieve high quality images. One of the key factors customers consider when buy-ing or upgrading a color reproduction system is Image Quality (IQ) [118]. Quality assessmentis therefore needed to show if technology advances increase the quality of the produced image(Figure 1.1).

1.1 Motivation

There are essentially two types of IQ assessment: subjectively or objectively. Subjectiveevaluation is carried out by human observers, and is therefore influenced by the Human Visual

Quality?

Figure 1.1: When printing a digital image, the physical printed image might present differ-ences from the original. These differences contribute to the impression of quality. Extensiveresearch has been carried out to improve the quality, but in order to know if this researchimproves quality, some kind of quality assessment is needed.

3

Page 20: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

System (HVS). Objective evaluation of IQ can be carried out in many ways, a typical way isto use measurement devices gathering numerical values. Another way is to use algorithms,commonly known as IQ metrics, in an attempt to quantify quality.

Subjective evaluation is commonly used since it is a precise way to quantify IQ [116].However, it is both resource and time demanding, and in addition observers can be inconsis-tent. Because of these limitations, objective evaluation methods have been proposed. Theyhave lower cost, are less time and resource demanding, they produce consistent results, andthey require less competence of the user. One of these objective methods is IQ metrics, whichare based on different assumptions and incorporate different characteristics of the HVS. Animpressive number of IQ metrics have been proposed in the literature. Unfortunately, no IQmetric fully correlated with perceived quality has yet been proposed. Additionally, use of IQmetrics in the assessment of quality for printed images has not been extensively researched.Using IQ metrics in the printing process has a great potential, in terms of determining themost suitable processing of an image, to objectively assess the IQ during the printing processand of the final printed image, or to indicate where a loss of quality occurs in a print workflow.

1.2 Aims

IQ metrics have been proposed for different applications, such as video and images, and dis-tortions, for example compression and noise. As they are becoming more correlated withperceived IQ, application of IQ metrics are becoming more popular. They can be used tomonitor, evaluate, or improve quality.

The goal of this research is to adapt, use, and evaluate IQ metrics for measuring perceivedIQ. Adapt refers to modifying or developing methods to apply IQ metrics to printed images,it also refers to improvement of existing IQ metrics or development of new IQ metrics, whichare better correlated with subjective data than existing metrics. Use refers to the applicationof IQ metrics in a printing workflow, for example to measure the quality of a printed image.Evaluate refers to the evaluation of IQ metrics against perceptual data.

We will mainly focus on full-reference IQ metrics (where an original is known and it isused in the assessment of quality), and the applied field is color printing. We limit our workto halftoned images, but the methods are applicable to continuous tone images.

1.3 Research methods

Since we will investigate the relationship between IQ metrics and perceived quality, we willbe carrying out psychophysical experiments which require human observers. For these ex-periments we will use both qualitative and quantitative methods. Qualitative methods will beused to obtain a more complete and detailed description, while quantitative method will beused for numerical analysis of data and to get a more broaden understanding.

With the perceptual data obtained from experiments, analysis will be carried out anchoredin existing methods. This requires a good overview and understanding of the existing litera-ture, before commencing the task at hand. The work carried out in this thesis contains is boththeoretical, such as literature surveys, but also practical, such as experiments.

4

Page 21: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

1.4 Publications

The current study has led to the publication of several papers in international peer-reviewedconferences and international journals. The relations between the publications and the thesiscan be seen in Figure 1.2. Listed below are the main publications, and then the supportedpublications in reverse chronological order.

Main publications:

[355] M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality metrics: Clas-sification and evaluation. Foundations and Trends in Computer Graphics and Vision,2011. Submitted

[156] M. Gong and M. Pedersen. Pooling for color image quality. Journal of Visual Commu-nication and Image Representation, 2011. Submitted

[358] M. Pedersen, G. Simone, M. Gong, and I. Farup. A total variation based color imagequality metric with perceptual contrast filtering. In International conference on Perva-sive Computing, Signal Processing and Applications, Gjøvik, Norway, Sep 2011

[359] M. Pedersen, Y. Zheng, and J. Y. Hardeberg. Evaluation of image quality metrics forcolor prints. In A. Heyden and F. Kahl, editors, Scandinavian Conference on ImageAnalysis, volume 6688 of Lecture Notes in Computer Science, pages 317–326, YstadSaltsjobad, Sweden, May 2011. Springer-Verlag Berlin Heidelberg

[349] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Image quality metricsfor the evaluation of print quality. In F. Gaykema and S. Farnand, editors, Image Qual-ity and System Performance, volume 7867 of Proceedings of SPIE, pages 786702–1–786702–19, San Francisco, CA, Jan 2011

[347] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Estimating print qualityattributes by image quality metrics. In Color and Imaging Conference, pages 68–73,San Antonio, TX, USA, Nov 2010. IS&T/SID

[348] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Validation of qualityattributes for evaluation of color prints. In Color and Imaging Conference, pages 74–79, San Antonio, TX, USA, Nov 2010. IS&T/SID

[343] M. Pedersen and S. A. Amirshahi. Framework the evaluation of color prints usingimage quality metrics. In 5th European Conference on Colour in Graphics, Imaging,and Vision (CGIV), pages 75–82, Joensuu, Finland, Jun. 2010. IS&T

[346] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of image qualityfor color prints. Journal of Electronic Imaging, 19(1):011016–1–13, Jan 2010

[345] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of a new imagequality model for color prints. In Color Imaging Conference, pages 204–209, Albu-querque, NM, USA, Nov 2009. IS&T

[354] M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality metrics.Høgskolen i Gjøviks rapportserie 5, The Norwegian Color Research Laboratory (GjøvikUniversity College), Jun 2009. ISSN: 1890-520X

5

Page 22: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

[351] M. Pedersen and J. Y. Hardeberg. A new spatial hue angle metric for perceptual im-age difference. In Computational Color Imaging, volume 5646 of Lecture Notes inComputer Science, pages 81–90, Saint Etienne, France, Mar 2009. Springer Berlin /Heidelberg

[350] M. Pedersen and J. Y. Hardeberg. Rank order and image difference metrics. In 4thEuropean Conference on Colour in Graphics, Imaging, and Vision (CGIV), pages 120–125, Terrassa, Spain, Jun 2008. IS&T

Supporting publications:

[414] G. Simone, M. Pedersen, and J. Y. Hardeberg. Measuring perceptual contrast in digitalimages. Journal of Visual Communication and Image Representation, 2011. Accepted

[127] K. Falkenstern, N. Bonnier, H. Brettel, M. Pedersen, and F. Vienot. Weighing qualityattributes. In R. C. Baraas, editor, 21st symposium of the international colour visionsociety (ICVS), page 88, Kongsberg, Norway, Jul 2011. ISBN 978-82-8261-009-4

[128] K. Falkenstern, N. Bonnier, M. Pedersen, H. Brettel, and F. Vienot. Using metrics toassess the ICC perceptual rendering intent. In F. Gaykema and S. Farnand, editors,Image Quality and System Performance, volume 7867 of Proceedings for SPIE, pages786706–1–15, San Francisco, CA, Jan 2011

[410] G. Simone, V. Caracciolo, M. Pedersen, and F. A. Cheikh. Evaluation of a differenceof gaussians based image difference metric in relation to perceived compression arti-facts. In Advances in Visual Computing - 6th International Symposium, Lecture Notesin Computer Science, pages 491–500, Las Vegas, NV, Nov 2010. Springer

[126] K. Falkenstern, N. Bonnier, H. Brettel, M. Pedersen, and F. Vienot. Using image qualitymetrics to evaluate an ICC printer profile. In Color and Imaging Conference, pages244–249, San Antonio, TX, Nov 2010. IS&T and SID

[413] G. Simone, M. Pedersen, and J. Y. Hardeberg. Measuring perceptual contrast in uncon-trolled environments. In European Workshop on Visual Information Processing (EU-VIP), pages 102–107, Paris, France, Jul 2010. IEEE

[341] M. Pedersen. Objective image quality assessment of color prints. In G. Simone,J. Y. Hardeberg, and I. Farup, editors, The CREATE 2010 Conference, pages 146–150,Gjøvik, Norway, Jun 2010. ISBN: 978-82-91313-46-7

[61] G. Cao, M. Pedersen, and Z. Baranczuk. Saliency models as gamut-mapping artifactdetectors. In 5th European Conference on Colour in Graphics, Imaging, and Vision(CGIV), pages 437–443, Joensuu, Finland, Jun 2010. IS&T

[5] S. A. Ajagamelle, M. Pedersen, and G. Simone. Analysis of the difference of gaussiansmodel in image difference metrics. In 5th European Conference on Colour in Graphics,Imaging, and Vision (CGIV), pages 489–496, Joensuu, Finland, Jun 2010. IS&T

[6] S. A. Ajagamelle, G. Simone, and M. Pedersen. Performance of the difference of gaus-sian model in image difference metrics. In G. Simone, A. Rizzi, and J. Y. Hardeberg,editors, Gjøvik Color Imaging Symposium, number 4 in Høgskolen i Gjøviks rapport-serie, pages 27–30, Gjøvik, Norway, Jun 2009

6

Page 23: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

[340] M. Pedersen. 111 full-reference image quality metrics and still not good enough? InG. Simone, A. Rizzi, and J. Y. Hardeberg, editors, Proceedings from Gjøvik Color Imag-ing Symposium 2009, number 4 in Høgskolen i Gjøviks rapportserie, page 4, Gjøvik,Norway, Jun 2009

[415] G. Simone, M. Pedersen, J. Y. Hardeberg, and I. Farup. On the use of gaze informationand saliency maps for measuring perceptual contrast. In A-B Salberg, Hardeberg J.Y., and R. Jenssen, editors, 16th Scandinavian Conference on Image Analysis, volume5575 of Lecture Notes in Computer Science, pages 597–606, Oslo, Norway, Jun 2009

[352] M. Pedersen and J. Y. Hardeberg. Shame: A new spatial hue angle metric for perceptualimage difference. In Vision Sciences Society 9th Annual Meeting, Naples, FL, USA,May 2009

[353] M. Pedersen and J. Y. Hardeberg. SHAME: A new spatial hue angle metric for percep-tual image difference. Journal of Vision, 9(8):343, 8 2009. ISSN 1534-7362

[344] M. Pedersen, N. Bonnier, F. Albregtsen, and J. Y. Hardeberg. Towards a new imagequality model for color prints. In ICC Digital Print Day, Mar 2009

[342] M. Pedersen, F. Albregtsen, and J. Y. Hardeberg. Detection of worms in error diffu-sion halftoning. In S. P. Farnand and F. Gaykema, editors, Image Quality and SystemPerformance VI, volume 7242, page 72420L, San Jose, CA, USA, Jan 2009. SPIE

[417] G. Simone, M. Pedersen, J. Y. Hardeberg, and A. Rizzi. Measuring perceptual contrastin a multilevel framework. In B. E. Rogowitz and T. N. Pappas, editors, Human Visionand Electronic Imaging XIV, volume 7240 of Proceedings of SPIE, pages 72400Q–72400Q–9, San Jose, CA, Jan 2009

[416] G. Simone, M. Pedersen, J. Y. Hardeberg, and A. Rizzi. A multi-level framework formeasuring perceptual image contrast. Scandinavian Journal of Optometry and VisualScience, 1(1):15, Oct 2008

[169] J. Y. Hardeberg, E. Bando, and M. Pedersen. Evaluating colour image difference metricsfor gamut-mapped images. Coloration Technology, 124(4):243–253, Aug 2008

[357] M. Pedersen, A. Rizzi, J. Y. Hardeberg, and G. Simone. Evaluation of contrast measuresin relation to observers perceived contrast. In CGIV 2008 - Fourth European Conferenceon Color in Graphics, Imaging and Vision, pages 253–256, Terrassa, Spain, Jun 2008.IS&T

7

Page 24: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

1.5 Thesis outline

This thesis is intended to provide the reader with the understanding required to use IQ metricsto evaluate print quality. The thesis is structured into three parts, each with several chapters,to guide the reader through the different topics of the thesis.

• Part I is divided into two chapters that introduces important background knowledge tothe reader.

• Part II gives an in-depth introduction to IQ metrics in four different chapters, giving apresentation of IQ metrics, how to evaluate them, introduction of a new IQ metric, andevaluation of a set of IQ metrics.

• Part III investigates the use of IQ metrics in a printing workflow, and consists of sevenchapters. The chapters address how to use metrics with printed images, which qualityattributes to use, selection of IQ metrics for different attributes, improving metrics inthe pooling strategy, proposal of a new metric, and at last the Quality Assistant.

Then the conclusion is given, together with proposals of future work. Appendices are givenat last. An overview of the outline is shown in Figure 1.2 on the next page, linking chaptersand the publications above in Section 1.4.

8

Page 25: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

Chapter 1

Part I

Chapter 2

Chapter 3

Part II

Chapter 4

M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality metrics. Høgskolen i Gjøviksrapportserie 5, The Norwegian Color Research Laboratory (Gjøvik University College), Jun 2009.ISSN: 1890-520X. [354]

M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality metrics: Classification andevaluation. Foundations and Trends in Computer Graphics and Vision, 2011. Submitted. [355]

Chapter 6M. Pedersen and J. Y. Hardeberg. Rank order and image difference metrics. In 4th European Confer-ence on Colour in Graphics, Imaging, and Vision (CGIV), pages 120–125, Terrassa, Spain, Jun 2008.IS&T. [350]

Chapter 7

G. Simone, V. Caracciolo, M. Pedersen, and F. A. Cheikh. Evaluation of a difference of gaussiansbased image difference metric in relation to perceived compression artifacts. In Advances in VisualComputing - 6th International Symposium, Lecture Notes in Computer Science, pages 491–500, LasVegas, NV, Nov 2010. Springer. [410]

S. A. Ajagamelle, M. Pedersen, and G. Simone. Analysis of the difference of gaussians model inimage difference metrics. In 5th European Conference on Colour in Graphics, Imaging, and Vision(CGIV), pages 489–496, Joensuu, Finland, Jun 2010. IS&T. [5]

Part III

Chapter 8M. Pedersen and S. A. Amirshahi. Framework the evaluation of color prints using image qualitymetrics. In 5th European Conference on Colour in Graphics, Imaging, and Vision (CGIV), pages75–82, Joensuu, Finland, Jun. 2010. IS&T. [343]

Chapter 9

Chapter 10

M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of image quality for colorprints. Journal of Electronic Imaging, 19(1):011016–1–13, Jan 2010. [346]

M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Validation of quality attributes forevaluation of color prints. In Color and Imaging Conference, pages 74–79, San Antonio, TX, USA,Nov 2010. IS&T/SID. [348]

M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of a new image quality modelfor color prints. In Color Imaging Conference, pages 204–209, Albuquerque, NM, USA, Nov 2009.IS&T. [345]

Chapter 11

M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Estimating print quality attributes byimage quality metrics. In Color and Imaging Conference, pages 68–73, San Antonio, TX, USA, Nov2010. IS&T/SID. [347]

M. Pedersen, Y. Zheng, and J. Y. Hardeberg. Evaluation of image quality metrics for color prints. InA. Heyden and F. Kahl, editors, Scandinavian Conference on Image Analysis, volume 6688 of LectureNotes in Computer Science, pages 317–326, Ystad Saltsjobad, Sweden, May 2011. Springer-VerlagBerlin Heidelberg. [359]

M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Image quality metrics for the evaluationof print quality. In F. Gaykema and S. Farnand, editors, Image Quality and System Performance,volume 7867 of Proceedings of SPIE, pages 786702–1–786702–19, San Francisco, CA, Jan 2011.[349]

Chapter 12 M. Gong and M. Pedersen. Pooling for color image quality. Journal of Visual Communication andImage Representation, 2011. Submitted. [156]

Chapter 13M. Pedersen, G. Simone, M. Gong, and I. Farup. A total variation based color image quality met-ric with perceptual contrast filtering. In International conference on Pervasive Computing, SignalProcessing and Applications, Gjøvik, Norway, Sep 2011. [358]

Chapter 14

Chapter 15

Chapter 16

Appendix

Appendix A

Appendix B G. Simone, M. Pedersen, and J. Y. Hardeberg. Measuring perceptual contrast in digital images.Journal of Visual Communication and Image Representation, 2011. Accepted. [414]

Appendix CG. Cao, M. Pedersen, and Z. Baranczuk. Saliency models as gamut-mapping artifact detectors. In 5thEuropean Conference on Colour in Graphics, Imaging, and Vision (CGIV), pages 437–443, Joensuu,Finland, Jun 2010. IS&T. [61]

Appendix DM. Pedersen, F. Albregtsen, and J. Y. Hardeberg. Detection of worms in error diffusion halftoning.In S. P. Farnand and F. Gaykema, editors, Image Quality and System Performance VI, volume 7242,page 72420L, San Jose, CA, USA, Jan 2009. SPIE. [342]

Appendix E

Appendix F

Appendix G

Figure 1.2: Overview of thesis and how they are related to the publications.

9

Page 26: Doctoral Dissertation by Marius Pedersen

INTRODUCTION

10

Page 27: Doctoral Dissertation by Marius Pedersen

PART I

BACKGROUND

11

Page 28: Doctoral Dissertation by Marius Pedersen
Page 29: Doctoral Dissertation by Marius Pedersen

2 WHAT IS QUALITY?

Quality is a word we encounter in our daily life, and it refers to many different areas. To carryout research on the quality of prints a definition needs to be given.

2.1 Quality

A common definition of quality, regardless of field, is that quality is the conformance to re-quirements [90]. This definition is general, and has been adapted by many. Related definitionsare given by the International Organization for Standardization (ISO), who defines quality asthe totality of characteristics of an entity that bear on its ability to satisfy stated or impliedneeds [194] or as the ability of a set of inherent characteristics of a product, system or processto fulfill requirements of customers and other interested parties [196]. All of these definitionsrelate quality to some sort of requirements.

2.2 Image quality

The earliest known mentioning of IQ dates back to the invention of optical instruments, be-tween the years 1600 and 1620 [115]. The term became more and more common with theintroduction of photography and television. In the recent literature we find several definitions,such as:

• Janssen [217] defined quality in the context of visuo-cognitive systems to be the de-gree to which the image is both useful and natural. The usefulness of an image is theprecision of the internal representation of the image and the naturalness of an image tobe the degree of correspondence between the internal representation of the image andknowledge of reality as stored in memory.

• Yendrikhovskij [488] defined quality as a compromise between color realism (natural-ness constraint) and color discrimination (colorfulness constraint).

• Jacobson [216] defined IQ as the subjective impression found in the mind of the ob-server relating to the degree of excellence exhibited by an image.

• Engeldrum [116] defined IQ as the integrated set of perceptions of the overall degree ofexcellence of the image.

• ISO [202] and Keelan [230] defined IQ as the impression of the overall merit or ex-cellence of an image, as perceived by an observer neither associated with the act ofphotography, nor closely involved with the subject matter depicted.

13

Page 30: Doctoral Dissertation by Marius Pedersen

WHAT IS QUALITY?

• Fairchild [122] defined color IQ as the reductions in IQ corresponding to perceptiblevisual differences from some ideal and the magnitude of such differences.

• The International Imaging Industry Association (I3A) defined IQ as the perceptuallyweighted combination of all significant attributes of an image when considered in itsmarketplace or application [188].

2.3 Print quality

Print quality has also been defined, and the ISO 12647 standards [193] deal with qualityparameters in terms of process control for the graphic industry, and in these standards qualityis related to a set of quality parameters and tolerances. This follows the common definitionby Crosby [90] where quality is the conformance to requirements.

Southworth and Southworth [421] relate quality level of prints with an arbitrary or subjec-tive judgment as to the degree of ”goodness” or the absence of defects.

Field [143] adapted the framework proposed by Garvin [151] and applied it to color print-ing. The framework is divided into four quality facets:

• Conformance to specifications concerns the actual appearance of the image, and is theconformance to a set of tolerances.

• Excellence refers to how closely the reproduction approaches optimum or perfect ap-pearance. Judgment of excellence is usually carried out at the proofing stage.

• Aesthetics concerns creative aspects of the image.

• Permanence relates to durability, i.e. the ability to resist influence of light, chemicals,and moisture.

In addition to these, five non-appearance factors relating to the production and economy aregiven; delivery urgency, rapid change, creative interface, extras, and value.

2.4 Definition used in this work

The definitions above differ in terms of assumptions and restrictions. Because of this it isimportant to have a definition adapted to the setting in which it is to be used. The definitionof quality adopted in this work is the one by ISO [202] and Keelan [230]. This definition is anarrower definition than that of Jacobson [216] since it excludes personal attributes. However,it includes artifactual, preferential, and aesthetic attributes. This definition also allows theimage to be of better quality than an original and it accounts for image enhancements, incontrast to the definition by Fairchild [122]. In practice the definition chosen corresponds tothe one by Engeldrum [116], since in most relevant cases the user or customer has not takenthe photograph himself.

14

Page 31: Doctoral Dissertation by Marius Pedersen

3 EXISTING METHODS FOR THE

EVALUATION OF PRINT QUALITY

With the steady technology advancements, the need to verify if the new technology produceshigher quality prints than the current technology increases. There are two main classes ofmethods to assess IQ, objective and subjective. Objective assessment can involve the useof measurement devices to obtain numerical values, alternatively IQ metrics can be used.Subjective assessment is carried out by human observers, while objective assessment does notinvolve observers.

3.1 Objective quality assessment

Objective quality assessment has several advantages over subjective assessment: first it is lessresource demanding, in most cases more economic, faster, allows for real-time monitoring ofimaging systems, easy to use for optimization of image processing systems, requires little orno knowledge to use, and usually provides more consistent results than observers. This is whyit is the most common why to evaluate quality in the industry.

3.1.1 Measuring quality using instruments

Using physical instruments to measure quality has been and still is a popular way to assessquality. We will briefly mention a couple of different types of instruments useful to quantifyquality.

3.1.1.1 Densitometer

A densitometer measures the level of darkness (optical density) of a print. It measures theamount of light received from a stimulus and the result is calculated using a logarithmic out-put. A densitometer is commonly used to measure dot gain, the phenomenon causing printedmaterial to look darker than intended. This instrument is important in order to have correctdensity steps.

There are three main types of densitometers:

1. A reflection densitometer is used to measure the reflected light from a surface. A re-flection densitometer has a stable light source, optics to focus the light on a pre-definedspot on a surface, a set of filters for the spectral response, and a detector to monitor

15

Page 32: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

the reflected light as seen in Figure 3.1 [237, 438]. A sample is usually illuminatedfrom above, giving an illumination angle of 90◦ to the surface, and viewed at 45◦, fol-lowing the International Commission on Illumination (Commission Internationale del’eclairage, CIE) annotation this is written as (0◦:45◦x) where x indicates the azimuthaldirection of the reference plane [84]. Several other recommendations for geometry aregiven by CIE [84].

2. A transmission densitometer is used to measure the amount of light transmitted througha transparent material. Transmission densitometers follow the same principles as reflec-tion densitometers, but the sensor is placed beneath the material in order to measure thetransmittance (Figure 3.2).

3. A combination of the two above, measuring both reflected and transmitted light.

Polarizing filter

Aperture

Op�cs

Paper

Op�cs

Analyzer

Receiver

Display12345

Ink45

Light source

Color filter

Figure 3.1: Components of a reflection densitometer. A stable lightsource is used togetherwith an infrared filter, and an aperture to focus the light, which further passes through opticsbefore it reaches the sample surface. The reflected light from the surface passes through acolor filter before it reaches a sensor, which measures the reflected light. Results are convertedto a logarithmic scale before they are displayed. Figure inspired by Kipphan [237].

3.1.1.2 Colorimeter

The tristimulus colorimeter is the simplest form of measurement for color [281], which mea-sure tristimulus values. It has red, green, and blue photoreceptors, as the eye. A light illu-

16

Page 33: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Polarizing filter

Aperture

Op�cs

Film

Op�cs

Analyzer

Receiver

Display12345

Light source

Color filter

Figure 3.2: Components of a transmission densitometer. A stable light source illuminates thefilm, optics to focus the light on a pre-defined spot on a surface, a set of filters for the spectralresponse, and a detector to monitor the reflected light are the main components.

17

Page 34: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

minates the stimuli from 45◦ to the normal, light reflected along the normal is collected andpassed to a detector. The detector consists of three different filters with light sensitive diodes,which are combined to approximate the spectral response of the eye. A colorimeter is ef-fective for measuring color differences, but it has some limitations. The accuracy is limited,the colorimetric values are only valid for the light source (illuminant) in the instrument, andmetamerism cannot be indicated (the spectral response can be different but the tristimulusvalues are similar).

3.1.1.3 Spectroradiometer

A spectroradiometer measures the absolute spectral power distribution of a light source. Thelight is collimated by a lens into a dispersive element, and further decomposed into its spec-trum [395]. The spectrum is then sampled and recorded with a set of detectors. Spectrora-diometers can be used to measure both self-luminous and reflective objects. In the case ofreflective objects a spectrally smooth light source and a spectrally referenced sample is re-quired. Since these types of instruments are calibrated against a standard, the measurementuncertainty depends heavily on the reference standard [326].

3.1.1.4 Spectrophotometer

A spectrophotometer measures the ratio of reflected to incident light from a sample at manypoints across the visible spectrum [281]. In this instrument a light source illuminates a stim-ulus sample, and reflected light is passed to a spectral analyzer where the light is split intospectral components. The spectral analyzer has an advantage over the colorimeter. Unlikespectroradiometers, a spectrophotometer cannot measure self-luminous objects [395]. There-fore, spectrophotometers are useful for the calibration of printers and scanners, but not dis-plays. A spectrophotometer has a built in light source, and for color imaging purposes thelight source is usually at a 45◦ to the sensor. Geometry conditions are defined by CIE [84]. Aschematic cross section of a spectrophotometer is shown in Figure 3.3. The spectrophotome-ter is the most commonly used color measurement instrument in the graphic arts and printingindustry today [324].

3.1.1.5 Glossmeter

Gloss is an optical property of a surface, which is based on the interaction of light with physi-cal characteristics of a surface. A glossmeter gives the amount of reflected light from a sample.The measurements are dependent on the material and on the angle of the illumination. De-pending on the application different illumination and illumination angles are used. Gloss ismeasured on a scale from 0 to 100, so called Gloss Units (GU).

3.1.2 Quality evaluation in the industry

When considering printing presses there are two different approaches to ensure a stable print-ing process; optimized or standardized press behavior [324]. In optimized behavior one maxi-mizes the properties of the press without consideration to external specifications or standards,which can result in unique printing conditions and usually requires custom International ColorConsortium (ICC) profiles. The latter approach is to make the printing presses to conform to

18

Page 35: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Aperture

Op�cs

Sample

Op�cs

Light source

Gra�ng

Detector

Figure 3.3: Schematic cross section of a spectrophotometer. A light source illuminates thesample, the reflected (or transmitted) light is split into its spectral components by gratings(or by prisms or interference filters) before they reach the detector. Reproduction inspired bySharma [395].

a standard, for example ISO 12647-2 [201], or a specific reference. Process-Standard Offset(PSO) is one example of a method of standardized printing.

There are also commercial softwares available in the market. Quality Engineering Asso-ciates (QEA) provides test equipment for the printing industry. Their products include dif-ferent measurement instruments solutions, but also scanner or camera based solutions. TheirImage Quality Analysis R© (IAS) system follows the ISO-13660 [197] standard, and givesthe possibility to quantify print quality attributes as density, graininess, mottle, and banding.Imatest also provides software for automatic quality evaluation for digital imaging systemsand sensors. In a printing setting their GamutvisionTM software is the most relevant, whichvisualizes gamuts and rendering intents. It also evaluates ICC profiles and includes a perfor-mance test for printers. ImageXpert R© has products relevant for print evaluation as well, suchas scanner based solution for automatic quality evaluation attributes as mottle, banding, textquality, and line quality.

3.1.2.1 Process-standard offset

PSO is the description of an industrially orientated and standardized procedure for the creationof print products. PSO was developed by Fogra Graphic Technology Research Association(Fogra) in co-operation with the German Printing And Media Industries Federation. PSO isin conformance with ISO 12647 [193], and should therefore ensure the quality of the produc-tion of a print product, from data creation to the finished printing product. PSO certificationby Association for the Promotion of Research in the Graphic Arts Industry (UGRA)1 con-tains several functions; documentation, data reception, data creation, display, digital proofing,printing plate production, printing, and illumination. In some functions, as the digital proof-

1Visisted 02/07/11: http://www.ugra.ch/pso-certification.phtml

19

Page 36: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

ing, both subjective (visual inspection) and objective (measurements) evaluation is carriedout. In Norway PSO certification is carried out by the Norwegian Color Research Laboratory(Colorlab) following the same procedure as UGRA. In Germany PSO certification is carriedout by Fogra in collaboration with der Bundesverband Druck und Medien e.V. (bvdm).

3.1.3 Standards for printing

There are international standards, such as the ISO standards, and national standard for print-ing. We are giving a short introduction to the most relevant ones.

3.1.3.1 ISO 12647

ISO 12647 provides standard process control aims and tolerances for various printing methodsand processes. ISO 12647-1 [200] defines vocabulary, parameters, and measurement methods.ISO 12647-2 to 7 are targeted against different printing technologies, 12647-2 [201] for offsetprinting, 12647-3 [206] for newspaper printing, 12647-4 [207] for gravure printing, 12647-5[198] for silk screen printing, 12647-6 [208] for flexographic printing, and 12647-7 [210] fordigital proofing. These standards, such as the one for offset printing, describes control values,different paper types, tone values, solid tones, and tolerances.

3.1.3.2 USA

Specifications for Web Offset Publications Specifications for Web Offset Publications(SWOP) is an organization and the name of a set of specifications that it produces, and itaims at improving consistency and the quality of printed products. SWOP concerns only theUnited States, however the specifications are often referred to outside the United States [360].

General Requirements and Applications for Commercial Offset Lithography GeneralRequirements and Applications for Commercial Offset Lithography (GRACoL) is a documentcontaining general guidelines and recommendations that could be used as a reference sourceacross the industry for quality color printing developed by the Graphic Communications As-sociation (GCA), now called IDEAlliance. The different guidelines published are ”GRACoLx” (GRACoL, or G, followed by the version number), 7 being the last version published in2006.

In 2005 SWOP and IDEAlliance joined forces, and today SWOP and GRACoL are underthe same organization [360].

3.1.3.3 Germany

Der Bundesverband Druck und Medien e.V. Der Bundesverband Druck und Medien e.V.(bdvm) is the German Printing and Media Industry Federation. They have for a number ofyears published the standard ”ProzessStandard Offsetdruck” (PSO), which is a standard inco-operation with Fogra. It builds on the requirements from ISO 12647-2:2004 [201]. Thiswork thoroughly describes the implementation of the standard with all its parameters.

20

Page 37: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.1.3.4 Switzerland

System Brunner System Brunner is a private swiss company that since 1975 has publishedtheir own standard ”Eurostandard/Globalstandard”, which sets requirements for density, dotgain, and gray balance [478]. Thorough implementation of this standard achieves the sameresults as bdvm and it meets the requirements of the international standard ISO 12647-2 [387].

3.1.4 Test images

There are a range of test images, both pictorial and targets, in use in the industry. We willpresent the most commonly used images in the industry. We limit this overview to the imagesmainly used for measurement instruments, images for subjective assessment will be foundlater in Section 3.2.4.7.

3.1.4.1 Altona test suite

bdvm is responsible for the Altona test suite, which is the ”toolbox” in conformance withISO 12647-2:2004 [201]. The test suite contains a set of offset printed A3 sheets (referencesheets), a set of ICC profiles, a set of PDF/X-3 files, and a manual. In 2005 an updated versionwas published as a collaboration between bvdm, European Color Initiative (ECI) , EuropeanRotogravure Association (ERA) , UGRA, and Fogra. The Altona Test Suite files includes ameasure sheet (Altona measure as seen in Figure 3.4(a)), a visual control image (Altona visualas seen in Figure 3.4(b)) to evaluate overprints and spot colors, and a technical file (AltonaTechnical as seen in Figure 3.4(c)) to evaluate overprinting and fonts.

(a) Altona measure (b) Altona visual (c) Altona technical

Figure 3.4: Altona test suite. Figures reproduced from www.andigun.de, visited 28/06/11.

3.1.4.2 Roman16 bvdm reference images

The Roman16 bvdm reference images are published by the bvdm, and the suite contains 16reference images. The reference images are specially created test motifs for visual assessment,processing and output in premedia and printing. In the images all basic colors are covered,such as CMYK and RGB. One of the 16 reference images are shown in Figure 3.5.

21

Page 38: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.5: One of the Roman16 bvdm reference images. Image reproduced from http://www.cleverprinting.de/frame_newsletter0708_1.html, visited 30/06/11.

3.1.4.3 Visual print reference

Verband der Schweizer Druckindustrie (VSD, Association of the Swiss Printing Industry)established together with Ugra and others a set of test pages to perform quality control of theentire printing process. This set of reference prints are known as the visual print reference(Figure 3.6). The goal of the test pages is to control the printing production stage from start toend according to ISO 12467-2. The set contains eight pages with different images, and eachpage has an integrated chart for measurements (the Ugra/Fogra Media Wedge CMYK V3.0and the ECI TVI10 strip).

Figure 3.6: Visual print reference. Reproduced from http://www.ugra.ch/visual-print-reference-12.phtml, visited 30/06/11.

22

Page 39: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.1.4.4 Common test charts

A test chart is an arrangement of standardized color samples used for color comparisons andmeasurements. They are most commonly used to calibrate and to profile graphic devices, butalso for quality control.

MacBeth ColorChecker color rendition chart The MacBeth ColorChecker color rendi-tion chart was first produced in 1976. The test chart consists of 24 squares of painted samplesbased on Munsell colors. Color charts, such as the ColorChecker, are used to calibrate and toprofile devices. The ColorChecker was commercialized by X-Rite.

Figure 3.7: MacBeth ColorChecker color rendition chart. Reproduced from http://www.pbase.com/elliot/image/84099006/, visited 28/06/11.

IT8 test charts IT8 is a set of American National Standards Institute (ANSI) standardsfor color communications and control specifications. For printing the IT8.7/3 specificationare used for characterization of 4-color process printing. The IT8.7/3, as seen in Figure 3.8,defines two data sets; the basic ink value data set consists of 182 patches and the extendeddata set is 928 patches. The first 182 patches can be used as inputs to Neugebauer equations,masking equations, color mixing model calibration, black generation, and grey balance [161].The most common use of the IT8.7/3 test chart is for calibrating and profiling printers.

The IT8.7/2 is a color reflection target used for scanner calibration, while IT8.7/1 is atransmission target for scanners.

ECI 2002 The ECI 2002 target (Figure 3.9) has been developed by ECI [34]. The ECI 2002chart is a superset of the IT8.7/3 target, where all 928 patches of the IT8/7.3 are containedin the 1485 patches of the ECI 2002 chart. The chart is available in visual (structured) andrandom layout, the random layout is recommended for characterization of a press [109].

Ugra/Fogra Media Wedge The Ugra/Fogra Media Wedge CMYK is a standard tool forcontrolling the color transformation from data to digital proof or printing. The wedge consistsof 72 patches, which are defined with area coverages of the process colors C (Cyan), M(Magenta), Y (Yellow) and K (Black). There is also a correspondence between the wedge andthe most important patches of the ECI 2002 color chart. The wedge is also in correspondencewith ISO 12647.

23

Page 40: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.8: IT8.7/3 test chart for characterization of 4-color process printing.Reproduced from http://www.cgan.net/science/print/preprint/images/news43_06_0001.jpg, visited 30/06/11.

Figure 3.9: ECI 2002 test chart in visual layout. Downloaded from http://www.eci.org, visited 02/07/11.

Figure 3.10: Ugra/Fogra Media Wedge. Reproduced from http://www.ugra.ch/media-wedge-data.phtml, visited 30/06/11.

24

Page 41: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

TC3.5 CMYK The GretagMacbeth TC3.5 CMYK consists of 432 patches specified inCMYK (Figure 3.11). A variant of this target is the TC3.5 CMYK+Calibration target, whichhas 88 additional patches placed around the target. These targets are commonly used to createCMYK profiles.

Figure 3.11: TC3.5 CMYK test target.

Contrast resolution test target The contrast resolution test target shown in Figure 3.12consists of a two dimensional array of patches used to evaluate resolution. Each patch containsconcentric circles of varying size and contrast. These circles represent a sampling of imagedetail. This target is usually only used for the black printer due to easy evaluation, and sinceit is likely that the printing engine for black is no different from the the other colors [405].

Addressability test target Sigg [405] proposed an addressability test target (Figure 3.13),which will form the basis of the ISO 29112 standard for objective measurement of monochromeprinter resolution [491]. This test target determines the native addressability of a digital print-ing system, where addressability is defined by Sigg [405] as ”the maximum number of indi-vidually controllable spots per linear inch that a printer can lay down parallel or perpendicularto the direction of substrate motion through the printing system process”. The evaluation ofthe test chart is subjective. The chart consists of diverging ray patterns that will show a discon-tinuity at the native addressability of the printing system. The authors mention the possibilityof using scanner-based instruments instead of subjective evaluation.

Edge characteristics targets For edge raggedness and edge blurriness analysis a test target(Figure 3.14) is introduced in Zeise et al. [491]. It consists of three square printed solidregions orientated at 0, 8, and 24 degrees from the vertical. The evaluation of raggedness andblurriness is carried out within four regions on each side of the square.

25

Page 42: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.12: Contrast resolution test target. Reproduced from Sigg [405].

Figure 3.13: Addressability test target proposed by Sigg [405]. Figure reproduced from Zeiseet al. [491].

26

Page 43: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.14: Edge characteristics target. Figure reproduced from Zeise et al. [491].

3.1.5 Measuring quality using algorithms

The other objective method is to quantify IQ using algorithms, so called IQ metrics. Thesemetrics are models designed to automatically predict quality, either computational or percep-tual quality. Computational quality relates to the calculation of quality without modeling theHVS, while perceptual quality is dependent on the viewing conditions and the observers. Theadvantage of metrics compared to instruments is that they do not require special hardware andusually not special software. Most algorithms require little interaction from users, minimizingthe risk of human errors. These advantages over instruments and those of objective assess-ment methods over subjective methods give IQ metrics great potential to measure quality. Afull introduction to algorithms for measuring IQ is found in Part II.

3.2 Subjective quality assessment

Subjective assessment involves psychophysical experiments, where human observers are askedto grade a set of images according to a given criterion. There are several existing methods forcarrying out these types of experiments, which are introduced below. The use of subjectiveevaluation is important in this work, since we require subjective data for the evaluation of theobjective methods. Validation against subjective data is the only method to ensure that theobjective methods corresponds with perceived quality. Therefore a thorough introduction todifferent methods for subjective quality assessment is given.

3.2.1 Psychophysical thresholds

The first psychophysical experiments were carried out by Ernst Weber in the 1830’s. In theseexperiments it was found that the Just Noticeable Difference (JND) (difference threshold) isproportional to the stimulus magnitude, which today is referred to as Weber’s law. In the sametime period Gustav Fechner carried out similar experiments, and in the 1860’s he published”Elemente der Psychophysik” (Elements of psychophysics) [133], which was a study of therelationship between stimulus intensity and subjective experience. In psychophysics one canfind three traditional methods for measuring thresholds; method of adjustment, method oflimits, and method of constant stimuli.

27

Page 44: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.2.1.1 Method of adjustment

In the method of adjustment the observer would adjust a parameter until a visual match witha reference stimulus is obtained. This is carried out for a given set of stimuli with varyingcharacteristics, for example patches with varying intensity. The setting where a visual matchis found is recorded. The process is repeated for all stimuli in the test set. Further, statisticalanalysis can be carried out to find the Point of Subjective Equality (PSE) (the average settingthat gives a visual match), and the JND (the smallest difference between stimuli that can bereliably discriminated 50% of the time). Figure 3.15 shows an illustration of a set of statisticscommonly calculated.

0

0.25

0.5

0.75

1

30 40 50 60 70

Pro

ba

bil

ty o

f d

ete

c�o

n

Test intensity

LL UL

JND

PSE

IU

Figure 3.15: Example showing the Point of Subjective Equality (PSE), Just Noticeable Differ-ence (JND), Lower Limit (LL), Upper Limit (UL), and Interval of Uncertainty (IU). The PSEis located at the 50% probability of detection. The UI is the area between the LL and the UL.One JND equals the difference between the PSE and the upper or lower threshold.

Example method of adjustment: If a gray standard patch is given with an intensity of50, the observer would be given the standard patch and another patch with higher or lowerintensity, and then be instructed to adjust the intensity to match the standard (Figure 3.16).For the first trial the observer is shown the reference patch and one of the samples from thetest set, in this case 30. The observer adjusts the ”knob” ascending and the final intensity is 49,which is the intensity for which the observer perceives both patches as identical in intensity.Further, the same procedure is carried out for the other samples in the test set, in the secondtrial the observers are shown a test patch with an intensity of 70, and the final intensity is setto 52. For the third trial a patch of 40 is shown, and adjusted by the observer to 51. This isthen carried out until every patch in the test set is evaluated. Even though the example here iswith intensity patches, the setup can also be used to evaluate natural images.

The data gathered from the method of adjustment (Table 3.1) can be statistically analyzedto obtain information about the PSE and JND. Table 3.2 shows the results after a basic sta-tistical analysis of the recorded data. The PSE is the grand mean, in this case 50.6. Anothercommonly used statistics is the JND. This can also be calculated from the gathered data. JNDis usually defined as the point in the distribution that is equal to or greater than 75% of the

28

Page 45: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Reference Test set

50 30 7050

30

Trial 1

Ascending

Start intensity: 30

Final intensity: 49

70

Trial 2

Descending

Start intensity: 70

Final intensity: 52

40

Trial 1

Ascending

Start intensity: 40

Final intensity: 51

Figure 3.16: Example method of adjustment. The observer is given a pair of patches, thereference and one from the test set. The observer is instructed to adjust the intensity of the testpatch until it has the same intensity as the reference. The intensity that the observer indicatesas identical to the reference is recorded. Figure inspired by Ferwerda [141].

29

Page 46: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

values in the dataset. Given that the data is normally distributed we can find the point thatis equal to or greater than 75% of the data (75th percentile), this is where the z-scores (or astandard score indicates how many standard deviations an observation is above or below themean) are approximately 0.67. The JND can be found, with some assumptions, by taking theSTandard Deviation (STD) of the gathered data and multiplying it with 0.67. The JND is then0.67× ST D, and with the data from Table 3.2 we get a JND of: 0.67× 2.17 = 1.45. Basedon the PSE and JND the UL and LL thresholds can be derived as PSE ± JND, giving theUL threshold to be 52.05 and the LL threshold to be 49.15. By subtracting the Lower Limitthreshold from the Upper Limit threshold the Interval of Uncertainty (IU) can be found, inthis case 2.9. These results can be shown in a summary sheet (Table 3.3).

Table 3.1: Recorded data from an experiment carried out based on the method of adjust-ment, where A indicates ascending, and D descending trials. See Figure 3.16 for a visualrepresentation of the data for the first three trials.

Trial 1 2 3 4 5 6 7 8 9 10Series A D A D A D A D A D

Start intensity 30 70 40 60 35 65 45 55 42 62Final intensity 49 52 51 50 48 53 51 51 47 54

Table 3.2: Statistical analysis for the method of adjustment, where the mean and STD of theascending, descending, and grand mean of the series is calculated. The grand mean is equalto the PSE.

Series Mean STDAscending 49.20 1.79Descending 52.00 1.58

Grand 50.60 2.17

Table 3.3: Statistical analysis for the method of adjustment. A summary sheet shows the rele-vant statistical results, such as Point of Subjective Equality (PSE), Just Noticeable Difference(JND), Upper Limit (UL), Lower Limit (LL), and Interval of Uncertainty (IU).

Summary sheetPSE 50.60JND 1.45UL 52.05LL 49.15IU 2.90

3.2.1.2 Method of limits

The second method is the method of limits. Given the same setup as above, with a standardpatch of a given intensity, this patch is compared against a test set of patches where the ob-server is given one patch at a time in an ascending or descending order. For each patch theobserver is instructed to decide whether or not the standard patch is brighter or dimmer than

30

Page 47: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

the test patch. Usually each observer is shown both ascending and descending trials to mini-mize adaptation and expectation errors. Similar statistical analysis to the one for the methodof adjustments can be carried out to find the PSE, the JND, and other statistical measures. Thismethod provides more precise thresholds than the method of adjustment, but at the cost of aslight increase in complexity [395]. The method of limits only provides information aboutthe ”transitions”, where the observer indicates a change in one stimulus compared to another,all other stimuli do not provide any information. Because of this, the method of limits isoften used to get a rough idea of where the transition occurs, then further refinement of theexperiment is performed.

Example method of limits Following the same example as for the method of adjustmentwith a reference patch with an intensity of 50, and five test patches with a range of 30 to 70in intensity are used (Figure 3.17). In the first trial the observer is shown the reference patchand the patch with an intensity of 70. The observer perceives the 70 patch as brighter thanthe reference. Next the patch with a lower intensity (60) is shown and perceived as brighter.Next the patch with an intensity of 50 is shown and the observer perceives this as dimmer.We then record the midpoint between 50 and 60 as the cross point, in this case 55. The sameis done for an ascending trial, and then alternating order for the remaining trials. Table 3.4shows an example of data recorded in an experiment carried out based on the method of limits.This data can further be processed to obtain the same statistical measures as for the methodof adjustment; PSE, JND, STD, LL and UL thresholds, and IU. Table 3.5 shows the analysisrequired to obtain these measures. The PSE is the same as the mean, in this case 50. JNDequals to 3.53, the UL threshold 53.53, the LL threshold 46.47, the IU 7.06 (Table 3.6).

Table 3.4: Recorded data from an experiment carried out based on the method of limits.AS indicates ascending order, DS descending, Crosspoint is the point where the observerperceives a change compared to the reference, B indicates that the patch was brighter thanthe reference, and D indicates a dimmer patch with lower intensity.

Trial 1 2 3 4 5 6 7 8 9 10Series DS AS DS AS DS AS DS AS DS AS

Test intensity

70 B B B60 B B B B B B B50 D D B B B B D B D D40 D D D D D D D30 D D D

Crosspoint 55 55 45 45 45 45 55 45 55 55

Table 3.5: Statistical analysis method of limits, where the mean and STD of the ascending,descending, and grand mean of the series is calculated.

Series Mean STDAscending 49 5.48Descending 51 5.48

Grand 50 5.27

31

Page 48: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Reference Test set

50 30 7050

Trial 1

40 60

70

60

50 50

50

50 brighter

brighter

dimmer

------- 55

Trial 2

50

50

50 brighter

dimmer

dimmer

------- 55

40

50

60

Trial 3

70

60

50 50

50

50 brighter

brighter

dimmer

------- 45

4050

brighter

Figure 3.17: Example method of limits. The observers are given a pair, a reference patch anda patch from the test set. They are instructed to indicate whether the test patch is brighter ordimmer than the reference. For the first trial the reference and a test patch with an intensity of70 are shown, and the observer judges the test patch to be brighter than the reference. Thenthe next patch with lower intensity is shown, and this is continued until the answer changes todimmer. The recorded value from each trial is the average between the two values where theobserver changed from one answer to the other, for the first trial this would be ((60+50)/2) =55. Figure inspired by Ferwerda [141].

Table 3.6: Statistical analysis method of limits. The summary sheet shows the relevant sta-tistical results, such as Point of Subjective Equality (PSE), Just Noticeable Difference (JND),Upper Limit (UL), Lower Limit (LL), and Interval of Uncertainty (IU).

Summary sheetPSE 50.00JND 3.53UL 53.53LL 46.47IU 7.06

32

Page 49: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Improvements of the method of limits Improvements to the methods of limits have beensuggested as well, such as staircase methods, Parameter Estimation by Sequential Testing(PEST) [434], and QUEST [470]. The first improvement, staircase, usually starts with a highintensity stimulus, and reduces the intensity until the observer reaches the point where thestimulus is perceived as dimmer, then the staircase is reversed and the same is done in theopposite direction until the observer perceives the stimulus as brighter (Figure 3.18). Thisprocess continues until the results converge. This method is also known as ”1-down-1-up”.Statistical methods to analyze the results from the staircase method are available; see forexample Feeny et al. [136] who suggest different methods to calculate the PSE.

30

40

50

60

70

0 5 10 15

Test

in

ten

sity

Trial

Figure 3.18: Example of the staircase method. Blue diamond points indicate reduction inintensity, and red squares increasing intensity.

PEST [434] uses changes in step size to focus the adaptive track, stopping the track whenthe estimate is adequately defined (Figure 3.19). The starting level is usually significantlydifferent from the expected final value, and a number of samples are shown. After eachsample is shown at a fixed level a statistical test is applied to indicate if the performance atthat level is significantly better or poorer than the targeted performance (for example 75%correct detections). If above the targeted performance, the level is changed to the oppositeside of the expected final value, and the procedure is repeated. Then the step size is halved,and the process is repeated until the answers converge. The final estimate is simply the finalvalue determined by the trial placement procedure.

QUEST [470] can be considered as an improvement of PEST, where prior information isused along with the data to guide the placement of trials (Figure 3.20). It is a method where thetest intensity is chosen so the data gives the best fit to what is called a psychometric function.An advantage of QUEST is that it provides the maximum information gain per trial about thethreshold value, and doing so increased efficiency, allowing the threshold to be found in thesmallest number of trials.

Leek [264] gives a good introduction, comparison, and discussion of the staircase method,PEST, and QUEST.

33

Page 50: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

30

40

50

60

70

0 5 10 15 20 25

Test

in

ten

sity

Trials

Figure 3.19: PEST example. The first trials are usually far from an expected final value. Whenthe answers are higher than a targeted performance, the test intensity changes. This is doneuntil the answers converge. Blue diamond points are correct answers, while red squares arewrong answers.

30

40

50

60

70

0 5 10 15 20 25

Test

in

ten

sity

Trials

Figure 3.20: Example QUEST. The first trials are usually far from an expected final value.When the answers are higher than a targeted performance, the test intensity changes. This isdone until the answers converge. Blue diamond points are correct answers, while red squaresare wrong answers.

34

Page 51: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.2.1.3 Method of constant stimuli

The last method is the method of constant stimuli, where the property of a stimulus is notrelated from one trial to the next, but the stimulus is presented randomly. The advantage ofa random order is that the correct answer cannot be predicted by the observer, minimizingexpectation errors. However, the disadvantage is that it requires a large number of trials andis therefore time demanding. Statistical analysis can also be carried out on the data from thismethod to obtain the PSE and JND. This method has been shown to be inefficient comparedto adaptive methods [469], such as QUEST.

Given a stimulus with a reference value, a set of test stimuli (φ) are created surroundingthe intensity range of the reference. Each stimulus is shown to the observer repeatedly in arandom order, and the observer is asked to answer whether the stimulus is above or below thereference in intensity. The number of yes and no responses for the test stimuli are recorded,and the proportion (p) of yes responses is calculated. Based on the proportion of responsesa psychometric function can be constructed. Then a table with the proportion of detectionsand corresponding z-scores for the intensity values can be constructed using the normal dis-tribution table. Further, to obtain the PSE and other statistical measures additional analysis isrequired. The linear regression between the z-scores and test intensities is calculated, resultingin the slope and intercept of the fitting. These can further be used to obtain the PSE definedas −(intercept/slope) and the STD defined as 1/slope. UL, LL and the IU are similar to theones above.

Reference Test set

50 30 7050

Trial 1

40 60

7050 brighter

Trial 2

50 brighter60

Trial 3

50 dimmer40

Figure 3.21: Example of the method of constant stimuli, with a reference test intensity patchof 50 and 5 patches in the test set. The reference is compared against one of the patches fromthe test set. The observer is asked to judge whether the test patch is brighter or dimmer thanthe reference. Figure inspired by Ferwerda [141].

Example method of constant stimuli An experiment is set up with 20 trials per test in-tensity. For each test intensity the number of patches brighter and dimmer than the referenceis recorded (Table 3.7). Based on this, the percentage of patches judged as brighter than thereference can be calculated, and plotted as a psychometric function (Figure 3.22), and furtherthe z-scores can be derived (for example in Microsoft Excel using the normsinv function ornorminv in Matlab, which calculates the inverse of the cumulative distribution function). Thez-scores can then be plotted against the test intensities as seen in Figure 3.23. The linearregression between the z-scores and the test intensities will reveal the slope and intercept,which can be used to find the PSE. In this case the linear regression is 0.07x−3.79, where the

35

Page 52: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

slope is 0.07 and the intercept −3.79. The PSE is then the negative of the intercept divided bythe slope −(−3.79)/0.07 = 48.88. The STD is calculated as 1/0.07, the JND then becomes0.67×ST D = 0.67× (1/0.07) = 8.67. This can be visualized as seen in Figure 3.24, wherethe PSE is the point where 50% of the test patches are perceived as brighter and the remaining50% as dimmer. The IU is the distance between the UL and LL, surrounding the PSE. Thearea between the PSE and one of the thresholds is equal to one JND.

Table 3.7: Example constant stimuli. Each test has 20 trials, and based on the number of trialsjudged as dimmer the percentage of trials being dimmer can be calculated. This informationcan further be used to obtain z-scores.

Test intensity Brighter Dimmer Percentage brigther z30 2 18 0.1 -1.2840 4 16 0.2 -0.8450 9 11 0.45 -0.1360 17 3 0.85 1.0470 19 1 0.95 1.64

0

0.2

0.4

0.6

0.8

1

30 40 50 60 70

Pe

rce

nta

ge

bri

gth

er

Test intensity

Figure 3.22: Example constant stimuli. The test intensities are plotted against the percentageof trials perceived as being brighter.

3.2.2 Psychophysical scaling

Psychophysical threshold methods are useful to determine color tolerances, compression lim-its, and more. However, the goal is often to determine the scale of perception compared to asingle threshold. In order to achieve this, scaling methods have been proposed to obtain therelationship between physical changes and perceived changes. There are three common meth-ods for conducting psychophysical scaling experiments: category judgment, pair comparison,and rank order [116].

36

Page 53: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

y = 0.07x - 3.79

-2

-1.5

-1

-0.5

0

0.5

1

1.5

2

30 40 50 60 70z-sc

ore

Test intensity

Figure 3.23: Example constant stimuli. Linear regression is carried out between the test inten-sities and the z-scores, which reveals the slope (0.07) and intercept (-3.79). This informationis used to obtain the statistical measures.

0

0.25

0.5

0.75

1

30 40 50 60 70

PSELL UL

IU

JND

Test Intensity

Pro

ba

bil

ity

o

f d

ect

ec�

on

Figure 3.24: Example constant stimuli. The PSE is located at the 50% probability of detection.The interval of uncertainty (IU) is the area between the lower limit threshold (LL) and theupper limit threshold (UL). One JND equals the difference between the PSE and the upperlimit or lower limit thresholds.

37

Page 54: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.2.2.1 Pair comparison

In pair comparison experiments observers judge quality based on a comparison of image pairs,and the observer is asked which image in the pair is the best according to a given criterion, forexample which has the highest quality or is the least different from an original (Figure 3.25).These experiments can be either forced-choice, where the observer needs to give an answer, orthe observer is not forced to make a decision and may judge the two reproductions as equals(tie). In the case of pair comparison experiments no information on the distance betweenthe images is recorded, making it less precise than category judgment, but less complex. AMatlab script for the calculation of scale values from pair comparison data is found in Greenand MacDonald [162].

In these experiments the observer evaluates n reproductions for m reference images, re-sulting in (n(n− 1)×m)/2 comparisons. Usually in pair comparison experiments each pairof reproductions is shown twice, changing the position of the right and left reproductions toavoid bias, giving mn(n− 1) comparisons. With an increasing number of reproductions thenumber of trials increases very rapidly, which makes it unsuitable for experiments involvingmany reproductions.

Using Thurstone’s Law of Comparative Judgment [437], data collected from pair compar-ison experiments can be transformed into interval scale data. The results from this transforma-tion represent the distance of a given stimulus from the mean score of the set being evaluated.When calculating scaled values several assumptions must be satisfied [408]:

• Each sample has a single value that can describe its quality.

• Each observer estimates the quality of this sample with a value from a normal distribu-tion around this actual quality.

• Each sample has the same perceptual variance.

• Each comparison is independent.

However, all of these assumptions are not always valid in IQ experiments.Pair comparison is the most popular method to evaluate e.g. gamut mapping [80], and is

often preferred due to its simplicity, requiring little knowledge by the user.

Example pair comparison The observers are shown a pair of images, and instructed toselect one of them according to the instructions given. Given three different reproductions (A,B, and C), all combinations of them are shown. For a given pair the observer selects one, andthe result is recorded in an n× n raw data matrix, in this case 3× 3. In this matrix the valueone is given in column i and row j for reproduction i as compared with reproduction j. Anexample of these matrices for five different observers is shown in Table 3.8.

Then an n×n summed frequency matrix (Table 3.9) can be computed by summing all rawmatrices (Table 3.8). From the summed frequency matrix a percentage matrix (Table 3.10)can be found, by dividing the frequency matrix by the number of observations, in this case 5.This matrix shows the percentage of when a reproduction was preferred over another. Fromthe summed frequency matrix (Table 3.9) a Logistic Function Matrix (LFM) (Table 3.11) canbe found by using the formula proposed by Bartleson [29]:

LFM = ln

(f + c

N − f + c

), (3.1)

38

Page 55: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Reference Test set

50 30 7050

30

Trial 1

Le�

70

Trial 2

40

Trial 3

40 50

Le�

60

Right

Figure 3.25: Example pair comparison experiment. For the first trial the observer judged theleft patch to be closer to the reference, the same with the second trial, and in the third trialthe right. The observer judges all combinations of pairs.

Table 3.8: Frequency matrices for five observers each judging three different reproductions (A,B, and C). A 1 indicates that the reproduction was preferred over another, while 0.5 indicatesan equal preference.

(a) Observer 1

A B CA - 0 0B 1 - 0C 1 1 -

(b) Observer 2

A B CA - 0 0.5B 1 - 0C 0.5 1 -

(c) Observer 3

A B CA - 0.5 0.5B 0.5 - 0C 0.5 1 -

(d) Observer 4

A B CA - 0 0.5B 1 - 0C 0.5 1 -

(e) Observer 5

A B CA - 0 0B 1 - 1C 1 0 -

Table 3.9: Summed frequency tables based on the frequency tables for each observer (Table3.8).

A B CA - 0.5 1.5B 4.5 - 1C 3.5 4 -

Table 3.10: Percentage matrix based on the frequency table (Table 3.9).

A B CA - 0.1 0.3B 0.9 - 0.2C 0.7 0.8 -

39

Page 56: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

y = 0.7903x

-1.5

-1

-0.5

0

0.5

1

1.5

-2 -1.5 -1 -0.5 0 0.5 1 1.5 2

z-sc

ore

Logis�c func�on matrix values

Figure 3.26: LFM values from Table 3.11 plotted against z-scores computed using the inverseof the standard normal cumulative distribution based on the percentages from Table 3.10.The slope if the linear regression gives the coefficient to calculate the final z-scores, in thisexample 0.79.

where f is the value from the frequency matrix, N is the number of observations, and c is anarbitrary additive constant. Bartleson [29] proposed 0.5, which is the most commonly usedconstant [162, 302].

Table 3.11: Logistic Function Matrix (LFM).

A B CA - -1.61 -0.69B 1.61 - -1.10C 0.69 1.10 -

Values in the LFM can be transformed into z-scores using a simple scaling coefficient.This coefficient is found by looking at the relationship between the inverse of the standardnormal cumulative distribution for the percentage matrix (converts percentages into z-scores)and the LFM values, which is done using linear regression (Figure 3.26). The scaling coef-ficient for this example is the slope of the regression line. The z-scores are in the same wayas in the method of constant stimuli: in Microsoft Excel using the normsinv function or inMatlab with norminv. The coefficient (slope) for this example is 0.79, which is further usedto obtain the z-score matrix (Table 3.12) by multiplying the LFM matrix with the coefficient.

Table 3.12: Z-score matrix.

A B CA - -1.27 -0.55B 1.27 - -0.87C 0.55 0.87 -

The mean z-score for each reproduction is found by taking the average of each column

40

Page 57: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

(Table 3.13). The 95% Confidence Interval (CI) is found by using the number of observations:

CI = 1.96× σ√N

, (3.2)

where σ is the standard deviation and N the number of observations. Since the z-scores havea scale with units equal to σ

√2 the standard deviation can be set to 1, giving the confidence

interval of the mean to be 1.96×(1/√

N). For the given example: 1.96×(1/√

5) = 0.62. The95% confidence interval is then the mean z-scores ±CI. The most common way to visualizez-scores is by an error bar plot, as shown in Figure 3.27. The mean z-score value is indicatedby the center square, and the whiskers on each line show the 95% CIs. If two CIs overlap thetwo reproductions are not considered to be significantly different with a 95% confidence (asseen between reproduction B and C), if they do not overlap the difference between the twoCIs they are statistically significant with 95% confidence (as seen between reproduction Aand C).

It has been shown that each sample does not have the same perceptual variance for IQexperiments [107]. Because of this Montag [294] proposed a new way to calculate confidenceintervals:

CI = ±1.96σobs, (3.3)

whereσobs = b1(n−b2)

b3(N −b4)b5, (3.4)

where b1 = 1.76, b2 = 3.08, b3 = 0.613, b4 = 2.55, b5 = 0.491, n is the number of stimuli, andN the number of observations. This method takes into consideration the number of stimuli,while the method in Equation 3.2 does not. Equation 3.2 is the most commonly used method,and unless stated otherwise all CIs calculated in this work has been made using Equation 3.2.

Table 3.13: Mean z-scores. Higher values are better. CI = 1.96× (1/√

5) = 0.62.

A B C0.91 -0.20 -0.71

3.2.2.2 Category judgment

In category judgment the observer is instructed to judge an image according to a criterion, andthe image is assigned to a category (Figure 3.28). Five or seven categories are commonly used,with or without a description of the categories. This method is an extension of Thurstone’s lawof comparative judgment, and it is based on the law of categorical judgment [442]. One ad-vantage of category judgment is that information on the distance between images is recorded,but the task is more complex than pair comparison for the observers. Additionally, since thejudgments are more subjective the results are more observer dependent. Category judgmentexperiments are often faster than pair comparison, with fewer comparisons necessary. Be-cause of this it is better for large experiments compared to pair comparison. When designingthis type of experiment, one can consider showing all reproduction of the same scene at once,or they can be shown separately. The observers also have the ability to assign more than onereproduction to the same category. A Matlab script for the calculation of scale values fromcategory judgment data is found in Green and MacDonald [162].

41

Page 58: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

-1.5

-1

-0.5

0

0.5

1

1.5

2

A B

Z-s

core

s

Reproduc�on

C

Figure 3.27: Error bar plot of z-scores from a pair comparison experiment. The mean z-scoresfor three different reproductions are plotted with a 95% CI. Reproduction A can be consideredstatistically significant from reproduction C with 95% confidence, while A is not statisticallysignificantly different from B.

Reference Test set

50 30 7050

30

Trial 1

Categories: 1-7

7040 50 60

4 2 1 2 4

Figure 3.28: Category judgment experiment. The observer is shown the entire image set, andall stimuli are judged according to a criteria based on a seven step category scale with 1 beingthe best. In this example the stimuli with 50 in intensity has been judged as best and assignedto category 1.

42

Page 59: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Example category judgment As for pair comparison an n× n frequency matrix is calcu-lated from the raw experimental data (Table 3.14), where each column contains the frequencythat one reproduction is judged to belong to a category. These are summed to obtain a fre-quency matrix (Table 3.15). Based on the frequency matrix a cumulative frequency matrix,containing n×(n−1) values, is calculated (Table 3.16), and this is converted into a cumulativepercentage matrix (Table 3.17).

This matrix is then converted into a z-score matrix following the same steps as for paircomparison, first obtaining an LFM (Table 3.18), which is converted into z-scores (Table 3.19)based on a scaling coefficient found by linear regression between the LFM values and the z-scores of the cumulative percentage matrix. The coefficient for this example is 0.78.

Table 3.14: Category judgement example. Five observers judging each reproduction twice,results in 10 observations.

(a) Observer 1

7 6 5 4 3 2 1A 0 0 0 0 2 0 0B 0 0 1 1 0 0 0C 1 1 0 0 0 0 0

(b) Observer 2

7 6 5 4 3 2 1A 0 0 0 0 2 0 0B 0 0 0 1 1 0 0C 0 0 2 0 0 0 0

(c) Observer 3

7 6 5 4 3 2 1A 0 0 0 0 2 0 0B 0 0 1 1 0 0 0C 0 2 0 0 0 0 0

(d) Observer 4

7 6 5 4 3 2 1A 0 0 0 0 1 1 0B 0 0 0 1 1 0 0C 0 0 1 1 0 0 0

(e) Observer 5

7 6 5 4 3 2 1A 0 0 0 0 0 2 0B 0 0 0 2 0 0 0C 0 2 0 0 0 0 0

Table 3.15: Category judgment example: frequency table.

7 6 5 4 3 2 1A 0 0 0 0 7 3 0B 0 0 1 6 3 0 0C 1 5 3 1 0 0 0

Table 3.16: Category judgement example: cumulative frequency table.

7 6 5 4 3 2A 0 0 0 0 7 10B 0 0 1 7 10 10C 1 6 9 10 10 10

The next step consists of calculating an n× (n−2) difference matrix (Table 3.20), whichis found by taking the z-score values and subtracting the adjacent z-score value. The mean

43

Page 60: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Table 3.17: Category judgement example: Cumulative percentage matrix.

7 6 5 4 3 2A 0 0 0 0 0.7 1B 0 0 0.1 0.7 1 1C 0.1 0.6 0.9 1 1 1

Table 3.18: Category judgement example: LFM matrix.

7 6 5 4 3 2A -3.04 -3.04 -3.04 -3.04 0.76 3.04B -3.04 -3.04 -1.85 0.76 3.04 3.04C -1.85 0.37 1.85 3.04 3.04 3.04

Table 3.19: Category judgement example: z-score matrix.

7 6 5 4 3 2A -2.36 -2.36 -2.36 -2.36 0.59 2.36B -2.36 -2.36 -1.43 0.59 2.36 2.36C -1.43 0.29 1.43 2.36 2.36 2.36

Table 3.20: Category judgement example: difference matrix.

7 6 5 4 3A 0.00 0.00 0.00 2.95 1.77B 0.00 0.93 2.02 1.77 0.00C 1.72 1.15 0.93 0.00 0.00

Average 0.57 0.69 0.98 1.58 0.59

Table 3.21: Category judgement example: boundary estimates.

T1 T2 T3 T4 T5 T6Boundary 0 0.57 1.27 2.25 3.83 4.42

Table 3.22: Category judgement example: scale values.

Mean RankA 2.36 2.94 3.63 4.61 3.23 3.35 1B 2.36 2.94 2.70 1.66 1.46 2.22 2C 1.43 0.29 -0.17 -0.11 1.46 0.58 3

44

Page 61: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

of the column from this difference matrix is used to calculate the category boundaries (Table3.21), which is done by setting the origin (between category 6 and 7) to zero and adding theadjacent mean values.

Final z-scores, or scale values, are found by subtracting the z-scores from the boundaryestimates (Table 3.22). Confidence intervals are calculated similarly to pair comparison, basedon the number of observations.

3.2.2.3 Rank order

For rank order experiments the observer is presented with a number of images, who is asked torank them based on a given criterion. Rank order can be compared to doing a pair comparisonof all images simultaneously. If the number of images is high, the task quickly becomeschallenging to the observer. However, it is a fast way of judging many images and a simpletype of experiment to implement. A Matlab script for the calculation of scale values fromrank order data is found in Green and MacDonald [162].

Figure 3.29: Rank order example. The observer ranks the reproductions from best to worstaccording to a given criteria.

Example rank order For each stimulus the rank is collected. This data can be transformedinto pair comparison data based on comparative paired comparison modeling [91]. The sim-plest way of doing this is to compute the rank of each stimulus. Given that stimulus i is rankedj for k j times, its rank score (RS j) is found by:

RS j =∑n

j=1(n− j)k j

N(n−a), (3.5)

where n− j is a weighting factor, resulting in the lowest in the rank to have a factor of 0.These rank scores can further be converted into z-scores. To do so, one needs to examinethe difference between the rank order method and the pair comparison methods. For paircomparison a stimulus i gives a scaled value si:

si =1

n−1

n

∑j=1

z

(∑k=1 np jki

n

), (3.6)

45

Page 62: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

where z() indicates an operator to transform the proportion of choices to z-scores, p jki is abinary value representing whether a stimulus i is greater than stimulus j. In order to convertrank order data into a pair comparison matrix we have the following equation:

RSi =1

(n−1)n ∑k=1

n ∑j=1

npki j =1

n−1

n

∑j=1

[∑ j=1 npki j

n

]. (3.7)

Scaled values are obtained by applying the operator to Equation 3.7:

si = z

(1

n−1

n

∑j=1

[∑n

k=1 pki j

n

]). (3.8)

Given three reproductions (A, B, and C) and five observers we can obtain the ranking foreach observer (Table 3.23).

Table 3.23: Example rank order: the ranking of five different observers for three reproductions(A, B, and C)

Observers1 2 3 4 5

A 1 2 1 1 1B 2 1 3 2 3C 3 3 2 3 2

This raw data can be transformed into pair comparison data, and we can get a frequencymatrix (Table 3.24). This matrix can be processed with the same method as for pair compari-son to obtain z-scores.

Table 3.24: Example rank order: raw rank order data transformed into a pair comparisonfrequency matrix.

A B CA - 1 0B 4 - 2C 5 3 -

3.2.3 Other methods

In addition to the three methods introduced, there are also other methods for image qualityevaluation.

3.2.3.1 Mean opinion score

Mean Opinion Score (MOS) is defined by the International Telecommunication Union (ITU) [189]for audio quality, but has also been extensively used for IQ [322, 399, 482]. Observers are

46

Page 63: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

5

4

3

2

1

Excellent (Impercep�ble)

Good (Percep�ble but not annoying)

Fair (Slightly annoying)

Poor (Annoying)

Bad (Very annoying)

100

0

Figure 3.30: Quality scale and impairment scale.

told to judge the quality from one to five where five is the best quality (similar to categoryjudgment), it is also common to have a five point descriptive quality scale (bad, poor, fair,good, excellent) (Figure 3.30). The MOS is the arithmetic mean of the individual scores S(Equation 3.9), and is usually given with a 95% confidence interval.

MOS =1n

n

∑i=1

Si, (3.9)

where Si is the score from one observation, and n is the total number of observations. Thismethod assumes that the data is normally distributed. The MOS can also be used to calcu-late the Difference Mean Opinion Score (DMOS), where the MOS for the reference is thensubtracted from the MOS for the other images. This generates a DMOS for each image. TheDMOS value for a particular test represents the subjective quality of the test relative to thereference.

3.2.3.2 Double stimulus impairment scale

ITU [191] recommends a procedure similar to MOS and category judgment for the evaluationof television pictures. The method, Double Stimulus Impairment Scale (DSIS), consists ofshowing a reference picture first, then an impaired picture. The observer is instructed tojudge the quality of the impaired picture. A scale from one to five is used, and adjectivedescriptions are recommended. Two different descriptions are suggested, the impairment scale(very annoying, annoying, slightly annoying, perceptible but not annoying, imperceptible) andthe quality scale (excellent, good, fair, poor, bad) as seen in Figure 3.30. MOS can be derivedfrom the experimental data.

3.2.3.3 Double stimulus continuous quality scale

The Double Stimulus Continuous Quality Scale (DSCQS) is another method proposed byITU [191]. In this method the observers are simply asked to assess the overall quality ofeach presentation by inserting a mark on a vertical scale. The scales provide a continuousrating system to avoid quantizing errors, but they are divided into five equal lengths whichcorrespond to the normal ITU five-point quality scale. Each rating is converted from mea-

47

Page 64: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

surements of length on the score sheet to normalized scores in the range 0 to 100 (Figure3.30 left side), and then MOS can be calculated. Sheikh et al. [398] used a similar setup tothe DSCQS. However, the raw scores were converted into difference scores and further intoz-scores.

3.2.3.4 Triplet comparison

ISO 20462-2 (Photography: Psychophysical Experimental Methods for Estimating ImageQuality - Part 2: Triplet Comparison Method) [203] presents another way of doing psy-chophysical experiments, where three images are judged simultaneously [231, 235]. In thismethod the stimuli are presorted in three or more categories, so that only those stimuli fallingwithin certain ranges are subsequently rated against one another [231]. ISO proposes thethree following categories: ”favourable”, ”acceptable”, and ”unacceptable”. The three im-ages in each triplet are not ranked, but rated against a five-category scale (similar to ITU’srecommendation) or as the one ISO proposes; ”Favourable”, ”Acceptable”, ”Just acceptable”,”Unacceptable”, and ”Poor”. It is also possible to use a seven point category scale in the caseof many images. This method allows two or three stimuli to have the same rating within atriplet and a magnitude for the differences between stimuli can be quantified instead of a signas in pair comparison. The complete process is shown on Figure 3.31. The number of samplecombinations is fewer than for pair comparison (Figure 3.32), with the number of comparisons(N) equal to:

N =n(n−1)

6, (3.10)

where n is the number of samples. It has been shown that the triplet comparison is almost50% faster than pair comparison [203]. A function is also specified for the combinations ofsamples to be shown [203]:

f (i) = 1+modulo(i−1,n), (3.11)

where modulo indicates the remainder of the division of (i− 1) by n. For the case of n = 7,we get the following combinations (1, 2, 4), (2, 3, 5), (3, 4, 6), (4, 5, 7), (5, 6, 1), (6, 7, 2) and(7, 1, 3) without duplication. Scheffe’s method [295] is applied to obtain an interval scale,as described in Annex E of ISO 20462-2 [203]. This can further be converted into JNDs, asdescribed in Annex F of ISO 20462-2 [203].

Categorical step

Reduc�on of the number of samples

Triplet comparison method step

Interval scaling

Figure 3.31: Steps of the triplet comparison method. First the stimuli set is divided intocategories, then the number of samples are reduced, before triplet comparison is used for theevaluation according to a criteria. Finally interval scaling is used to obtain scaling.

48

Page 65: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Reference Test set

50 30 7050

30

Triplet comparison

7050 30

Pair comparison

70

50 70

30 50

Figure 3.32: Triplet comparison compared to pair comparison. One set is needed for tripletcomparison, while pair comparison needs three sets.

3.2.3.5 Quality ruler

The quality ruler uses a Standard Quality Scale (SQS) with a fixed numerical scale so that thestimuli are anchored to a physical standard, with one unit corresponding to one JND; it alsohas a zero point. A hardcopy quality ruler setup is shown in Figure 3.33, where the test stimuliis placed above the ruler. The ruler consists of several reference stimuli ordered from highestto lowest quality, spaced at approximately three JNDs, and labeled with an integer. The use ofreference stimuli results in more reliable results when assessing large sets of stimuli spanninga wide range of quality. The observer can slide the ruler back and forth in order to comparethe test stimuli with the reference stimuli. Since the reference stimuli are labeled 3, 6, 9, andso on, the observer can specify an integer between two reference stimuli, where one integerdifference corresponds to approximately one JND. A softcopy quality ruler is also defined byISO [204]. The quality ruler method [204, 230, 231] is more suitable for measuring largerquality differences than for example the triplet comparison method.

An average SQS score per stimulus can be easily obtained, and gives a quality value di-rectly expressed in JNDs. Another advantage of the quality ruler method is that scores fromdifferent experiments are easily compared. However, the process of obtaining SQS scoresis complex and delicate, and requires a lot of work in the implementation stage [373]. Thequality ruler is recommended to be used with artifactual attributes (for example noise or band-ing), rather than preference attributes (for example color, contrast, and tone), since observersgenerally agree that a higher level of artifacts are less desirable. In the case of preferentialattributes, the observers might not agree since there are large individual differences. Keelanand Urabe [231] recommends sharpness as a good reference attribute due to its strong effecton IQ and low variability among observers.

49

Page 66: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.33: Hardcopy quality ruler. Image reproduced from Keelan and Urabe [231].

3.2.4 Experiment aspects

In addition to the experimental method there are several other aspects that need to be takeninto account, such as the number of observers, their expertise, and the viewing conditions.

3.2.4.1 Number of observers

The number of observers in subjective experiments is important for the statistical analysisusually performed on the results, and the precision of the statistics [116, 117]. When a largenumber of observers are used, the average is more likely to be consistent with ”overall IQ”[274] and the precision of the estimated values increases [116]. It is complicated to answerprecisely how many observers are required, but guidelines can be found. In scale values (z-scores) the precision increases with the square root of the number of observers. A numberbetween 10 and 30 observers are recommended for typical scaling applications [116, 117].For the evaluation of gamut mapped images it is recommended by CIE to have at least 15observers carrying out pair comparison, category judgment, or ranking experiments [80]. 15observers are also recommended by the ITU [191] for the evaluation of television pictures.Keelan and Urabe [231] recommend a minimum of 10 observers for obtaining relative qualityvalues for JND, and a minimum of 20 observers for obtaining absolute quality scores on theSQS scale [204].

There is usually a trade-off between the number of stimuli and the number of observers.Due to time restrictions it is often more desirable to have a large number of observers than alarge stimuli material [395].

Recently, larger experiments have been carried out, such as for the Tampere Image Database(TID) [365] with 654 observers in different locations, or using the Internet, as done by Qiuand Kheiri [367]2, Simone et al. [413], or Fairchild [123], where an almost unlimited numberof observers can be recruited.

2Experiment accessible on http://hdri.cs.nott.ac.uk/siq/ (14/02/2011)

50

Page 67: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

If we collect numerous observations even trivial differences are declared statistically sig-nificant. Therefore, it is also important to control the amount of observations (number of datasamples). When case when two stimuli are judged to be different when they are not is com-monly referred to as type II errors in statistics [32], also known as a false negative since it failsto reject a false null hypothesis. Additionally, time and resources will be wasted by collectingtoo many observations, often for minimal gain.

3.2.4.2 Observer type

The expertise and background of the observers will influence the results of experiments [103,107, 108, 116, 175]. Usually observers are divided into two types (of groups): experts andnaıve. Those considered to be experts usually have experience in judging or evaluating im-ages [116]. The experts can, to a larger degree, distinguish among attributes and they oftenhave a more precise scale than non-experts. In experiments where small differences are neededto be quantified experts are usually more suitable than non-experts [116]. It has been shownthat experts have a stronger consensus in their response than non-experts [108, 357]. Expertsalso look at more regions of smaller and more precise size than non-experts [108], which alsohas been verified by eye-tracking experiments [339].

Cultural differences have also been found to influence IQ experiments [139, 140]. How-ever, the differences found by Fernandez and Fairchild [139], Fernandez et al. [140], and Gastand Tse [152] were small, and are therefore usually not taken into account.

3.2.4.3 Observer characteristics

A portion of the population has color vision deficiencies, and thus have a decreased abilityto perceive differences between some of the colors that others can distinguish. This willinfluence how they perceive images, and therefore they are usually not considered as optimalobservers for IQ assessment experiments. It is therefore considered good practice to test allobservers for color deficiency using pseudoisochromatic plates, for example with an Ishiharacolor test or the Dvorine test [31].

3.2.4.4 Experiment duration

A high number of stimuli will increase the time spent by observers, and recommendationsindicate that the duration of an experiment should be limited to avoid observers fatigue [191,252]. International Telecommunication Union [191] recommends not more than 30 minutes,while, Larabi [252] recommends that the median time over the observers should not be morethan 45 minutes. In research by Van Der Linde et al. [447] observers showed on average asustained level of concentration/effort during an eye tracking experiment lasting 1 hour.

3.2.4.5 Number of stimuli

Keelan and Urabe [231] state that a minimum of three scenes should be used in order to obtainrelative quality values of JND. For the SQS scale a minimum of six stimuli is required [204].ISO 20462-1 [202] reports that the number of test stimuli should be equal to or exceed threescenes, and preferably be equal to or exceed six scenes. CIE [80] recommends to use atleast one of the obligatory test images specified by CIE, together with at least three additional

51

Page 68: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.34: Ski image from the CIE guidelines, which is obligatory for evaluation of gamutmapping algorithms.

images, for the evaluation of gamut mapping algorithms. Field [143] indicates that betweenfive and ten images are required to evaluate color image quality issues.

The number of stimuli used is often depending on other aspects, such as the number ofobservers required, the experimental method, and the precision of the results.

3.2.4.6 Type of stimuli

There are also recommendations and guidelines for the selection of test stimuli. Two differenttypes of test stimuli are typically found in IQ experiments; pictorial images and research im-ages (i.e. test targets) [143]. Pictorial images are the most commonly used test stimuli, sinceobservers are confident in judging them. However, they must be chosen with care since thecontent might influence the results. Pictorial images are often difficult to use with instrumentssince consistent measurements are difficult to produce [143]. Research images are artificiallycreated test images, often made to test a specific problem. They have the advantage overpictorial images that they are content free and often have areas that can be read by measur-ing instruments. One of the best known research image is the Macbeth ColorChecker ColorRendition Chart (Figure 3.7).

There are several guidelines regarding the characteristics of test stimuli. They should forexample include several levels (low, medium, and high) of several different characteristics(such as tonal distribution, detail level, and saturation) [143]. It is recommended to test abroad range of images to reveal different quality issues [176]. For a complete overview ofcharacteristics of test images we refer the reader to Field [143] and CIE [80].

3.2.4.7 Standard test images

CIE recommendation for gamut mapping The CIE guidelines for gamut mapping [80]gives one image as obligatory for the evaluation of gamut mapping; the ski image (Figure3.34). In addition to this image ten graphics (Canon Development Americas computer graph-ics images) and three other pictorial images (Sony sRGB standard images) are recommended.

ISO 12640 test images ISO 12640 defines different sets of test images for evaluation ofdifferent processes.

52

Page 69: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.35: The eight natural images found in ISO 12640-1 [195].

Figure 3.36: The eight natural images found in ISO 12640-2 [199].

The image set in ISO 12640-1 [195] consists of 8 natural (Figure 3.35) and 10 syntheticimages. The natural images include skin tones, images with detail in the highlights or shad-ows, neutral colors, memory colors, and colors in areas that are difficult to reproduce. Thesynthetic images include resolution charts and uniform vignettes. The image set was devel-oped for comparison of color output systems such as printing, proofing, and color facsimile,and therefore they are in CMYK format.

ISO 12640-2 [199] defines XYZ/sRGB standard color image data. The image set consistsof 15 color images, encoded as both 16-bit CIEXYZ and 8-bit RGB digital data, for the eval-uation of quality changes. The set has eight natural images (Figure 3.36) and seven syntheticimages, spanning a range of different characteristics. This set is optimized for viewing ona reference sRGB display in the reference sRGB viewing environment, and relative to CIEstandard illuminant D65 (Description of viewing illumination is found in Section 3.2.4.9).

ISO 12640-3 [209] provides a test image data set with a large color gamut related toilluminant D50. The 18 test images, 8 natural (Figure 3.37) and 10 synthetic, are encoded as16-bit CIELAB digital data. The natural images are 16 bits per channel, while the syntheticimages are 8 bits per channel.

ISO 12640-4 [212] specifies a standard set of wide gamut display-referred color images.These are encoded as 16-bit Adobe RGB digital data. These images compliment the existingimages of ISO 12640-2 which are based on the sRGB display gamut. These test images havea larger color gamut than sRGB, and these images will require much less aggressive colorre-rendering going to print than sRGB encoded images.

Canon Development Americas computer graphics images Canon Development Ameri-cas proposed ten computer graphics (Figure 3.38), which are recommended by CIE for theevaluation of gamut mapping algorithms.

Sony sRGB standard images The Sony standard images [420] are three photographs pro-vided by Sony for CIE [80] for the evaluation of gamut mapping (Figure 3.39), being a part

Figure 3.37: The eight natural images found in ISO 12640-3 [209].

53

Page 70: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.38: Canon development Americas computer graphics images

Figure 3.39: Sony sRGB standard images.

of the recommended images from CIE. The set consists of two studio scenes, representing aportrait and a party image, and an outdoor image with a picnic theme.

Kodak lossless true color image suite Kodak proposed the Kodak lossless true color im-age suite [241], which is a set of 24 images (22 outdoor and two studio images) as seen inFigure 3.40. The photographs originally appeared on a Kodak Photo CD. This image suitehas become a very popular suite for evaluating different aspects of imaging. However, theseimages were made when digital images was a new concept. Therefore, their quality is limited,and they are probably not comparable to current digital photos.

DigiQ Halonen et al. [167] proposed the DigiQ image suite, which consists of three testimages for print quality evaluation (Figure 3.41). In designing the images, aspects taken intoconsideration included a recognizable theme, memory colors and shapes, surface materials,detail information and salience of objects. The images in the suite are 16-bit TIFF images inAdobe RGB format, with a resolution of 360 DPI and a print size of 100 × 150 mm.

3.2.4.8 Stimuli labeling

Labeling the stimuli is important for registration of results. However, if the label is visible tothe observer they might ”solve” the code and respond according to the code and not to the task.Therefore, placing code visible to the observer is usually not considered good practice [116,117]. A common practice is to put the code on the back side of the stimuli. If the code needsto be visible to the observer, it should be obscure. One option is for example to use a fourdigit code of random numbers (1000 possible combinations), or a two letter code of randomcharacters (625 possible combinations).

54

Page 71: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

Figure 3.40: Kodak lossless true color image suite.

Figure 3.41: DigiQ image suite.

55

Page 72: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.2.4.9 Viewing conditions

Controlled and uncontrolled environments Regardless of the experimental type the exper-imental condition can be either controlled or uncontrolled. Controlled experiments are carriedout in a laboratory where the viewing conditions meet standards (such as described by CIE[80]), while uncontrolled experiments can be carried out in the field or on the web. Web-basedexperiments get more popular, since they are easy to carry out, free, time-saving, and the num-ber of observers is almost unlimited. The most important reason not to carry out uncontrolledexperiments has been that the environments are not standardized, however research haveshown small differences between controlled and uncontrolled experiments [300, 422, 503].

Viewing distance The perceived quality of an image is related to the distance at which theobservers view it [116]. A change in the viewing distance will change the information thatour HVS can perceive or detect, and will therefore also changethe quality. Thus the view-ing distance should be kept constant when conducting experiments, in order to have similarconditions for all observers.

In experiments where the viewing distance is not controlled, the variability of the resultsmight increase and the results spread. In specific experimental methods, such as the qualityruler, a headrest bar is used to fix the distance from the stimuli to the observer (Figure 3.33).

Viewing illumination The illumination under which the stimulus is viewed has an impacton the perceived quality. The intensity of the illumination (illuminance), which is measured inlumens per square meter or lux, influences the responses from human observers. For emmisivedisplays, for example a Liquid Crystal Display (LCD) monitor, the amount of light that thescreen produces is measured in candelas per square meter. Controlling the illuminance isimportant since an increase in illuminance will increase the colorfulness, better known as theHunt effect [395], it has also been shown that the perceived hue can change with a change inluminance.

Another important aspect is the spectral power distribution of the illumination, also knownas the Correlated Color Temperature (CCT). Since the color appearance changes with thespectral distribution of the illumination, it is important to include this aspect in the experi-mental design. This is also an important factor when comparing results from different exper-iments. The most common light sources are standardized by the CIE; the light sources aregiven a code, for example D5000 (D50 short) where the D stands for daylight and the fourdigit number indicates the CCT. When conducting experiments involving both monitors andprinted stimuli the CIE [80] recommends using a monitor with a D65 white point while theprints should be viewed under D50.

The geometry of the illumination could have an impact on the results, for example whenviewing glossy prints. It is recommended to illuminate the test stimuli from an angle of 45◦from the normal to the sample, and view the sample normal to the surface [116]. This iscommonly achieved by the use of a normalized viewing booth.

The viewing conditions are important to control, and the CIE [80] gives guidelines fordifferent conditions depending on the task at hand. For printed samples and transparencies werefer the reader to ISO 3664:2009 [211] and on experiment design to ISO 20462-1:2004 [202].

56

Page 73: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

3.2.4.10 Instructions

The instructions given to the observers are among of the most important aspects when design-ing an experiment. In order to obtain useful and meaningful results, the observers need to begiven clear instructions. There are several considerations to make when writing the instruc-tions, such as what are the observers going to judge, in what context, what are the criteria forthe judgment, and are there any definitions needed for the observers to carry out their task?First the aspect of what should be judged needs to be decided upon, this could be overall IQor a specific attribute (for example sharpness or a type of artifact). Very often it is importantto give a definition together with the question in order for all observers to judge the imagesbased on the same foundation. In some cases visual references are useful, they can be usedas anchors. It is important that the question asked is precise and clear, and does not leave anydoubt or inconsistencies. The context in which the experiment is carried out is also an im-portant aspect, for example if you are to judge IQ, in the context of office documents, officialletters, or a private context.

3.3 Summary

In this chapter we have presented the common methods to measure IQ, both objective andsubjective. First objective methods were presented focusing on instruments, such as spec-trophotometers and colorimeters. Additionally, quality assessment in the industry with rele-vant standards and test image were introduced. Further, the most common subjective assess-ment methods were explained in detail together with experimental aspects, such as number ofobservers, test images, and instructions.

57

Page 74: Doctoral Dissertation by Marius Pedersen

EXISTING METHODS FOR THE EVALUATION OF PRINT QUALITY

58

Page 75: Doctoral Dissertation by Marius Pedersen

PART II

IMAGE QUALITY METRICS

59

Page 76: Doctoral Dissertation by Marius Pedersen
Page 77: Doctoral Dissertation by Marius Pedersen

4 EXISTING IMAGE QUALITY METRICS

An IQ metric is an objective mathematical way to calculate quality without human observers.Most of the IQ metrics in the literature have the same structure, as shown in Figure 4.1. Theimages (original and reproduction, or for some metrics just the reproduction) go through colortransformations from a standard color space (such as sRGB) to a more suitable color space tosimulate the HVS (such as an opponent color space). In this color space characteristics of theHVS is incorporated. The next step consists of calculating the quality, usually pixel-wise thatresults in a quality value for every pixel in the image. Most metrics also do pooling, wherethe quality values from the previous step are reduced to a lower number of values (usually onequality value).

Color space

transforms

Human visual

system models

Quality

calcula�on Pooling

Fewer

Many

Quality

value

Original and

reproduc�on

Figure 4.1: A general flow-chart of objective IQ assessment methods. First, the color images,original and reproduction, are transformed to another color space. Then the IQ metrics usu-ally incorporate characteristics of HVS before calculating quality. The next step consists ofpooling, where the quality values from the previous step are reduced to fewer quality values,which results in a value representing quality.

There are fundamentally three different types of IQ metrics; no-reference (also referred toas blind-reference), reduced-reference, and full-reference (Figure 4.2). For the first type onlythe reproduction is available, and the calculation of quality is based only on the reproductionwithout use of the original (Figure 4.2(a)). In reduced-reference some information of thereproduction and the original is used in the calculation of quality (Figure 4.2(b)), such as asub-signal or the histogram. In the last group both the complete reference and the reproductionare available (Figure 4.2(c)). The two first types of IQ metrics are considered as more difficulttasks than full-reference [464].

Many IQ metrics have been proposed in the literature, a brief summary of more than 100metrics was given by Pedersen and Hardeberg [354]. These metrics stem from different ideas,and they have been made for different purposes, such as to quantify a distortion, to benchmark,to monitor quality, to optimize a process, or to indicate problem areas. Because of this it isimportant to categorize and identify their scopes of use, but also to evaluate their performance.Existing surveys, such as the one by Wang and Bovik [456] mostly focused on grayscale IQ

61

Page 78: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

Reproduc�on

Image Quality

Metric

Quality value

(Q = 2.5)

(a) No-reference IQ metric general framework.

Reproduc�on

Original

Image Quality

Metric

Quality value

(Q = 2.5)

(b) Reduced-Reference IQ metric general framework.

Reproduc�on

Original

Image Quality

Metric

Quality value

(Q = 2.5)

(c) Full-reference IQ metric general framework.

Figure 4.2: IQ metrics can be divided into three groups, depending on the information avail-able. In no-reference metrics only the reproduction is used in the calculation of quality, inreduced reference some information from the original and reproduction is used, and full-reference metrics use the complete reference and reproduction.

metrics, and the one by Avcibas et al. [14] covers simple statistical metrics. In this chapter wecarry out a survey of color and grayscale IQ metrics. Our goal is to classify IQ metrics intoseparate groups, and to evaluate their correspondence with human observers. Such a surveycan be used to select the most appropriate IQ metric for a given problem or distortion. It willalso provide a better understanding of the state of the art of IQ metrics, which can be used toimprove or develop new IQ metrics better correlated with perception.

We limit our survey to full-reference IQ metrics, where both the complete original andreproduction are used in the calculation of quality, since the original is often available. Addi-tionally, more work has been carried out on full-reference IQ metrics than reduced-referenceand no-reference.

In the literature many different terms have been used, such as IQ, image difference, imagefidelity, and image similarity. IQ metrics refer to metrics with the goal of predicting IQ, whileimage difference metrics aims at predicting the perceived magnitude of differences betweenan original and a reproduction [221] without considering how the differences contribute toperceived IQ, image fidelity metrics predict visibility of image reproduction errors [500], andimage similarity metrics measures how one image matches the other [378]. If possible wewill use the general term IQ, which in this work also covers the other terms since predictionof differences are seen as one of the steps towards predicting IQ [224]. However, it does notmean that there is a relationship between difference and quality [407], since a difference insome cases can increase quality. If the original image is considered to be ideal any differencesin the reproduction will lead to a loss of quality [221], making metrics based on distancessuitable to measure IQ as well.

In addition, we mainly use the term metric as a general term even though not all metrics

62

Page 79: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

fulfill the requirements to be a metric in mathematical terms. Other terms, such as index,measure, and criterion, have also been used in the literature.

4.1 Classification of image quality metrics

Classification of existing IQ metrics is a good starting point for selecting the best metricsfor a given setting, such as to evaluate print quality. Without a classification of metrics onedoes not a have an organized way to telling the difference between the different metrics.Such an organization illuminates the relationship between metrics, and thereby increasing theunderstanding of them. Classification helps in the decision making of what metric to use, butalso in the development of new and improved metrics.

4.1.1 Existing classification of image quality metrics

Since IQ metrics have been proposed based on different approaches they can be divided intodifferent groups. These groups usually reflect different aspects of the metrics, such as theirintended use or construction. Several different researchers have classified metrics into groups,even though it can be difficult to find sharp boundaries between the numerous IQ metrics inthe literature.

Avcibas et al. [14] divided IQ metrics into six groups based on the information they use:

• Pixel difference-based measures such as mean square distortion.

• Correlation-based measures, that is, correlation of pixel values, or of the vector angulardirections.

• Edge-based measures, that is, measure of the displacement of edge positions or theirconsistency across resolution levels.

• Spectral distance-based measures, that is, the Fourier magnitude and/or phase spectraldiscrepancy on a block basis.

• Context-based measures are based on various functionals of the multidimensional con-text probability.

• HVS-based measures are measures that are either based on the HVS-weighted spectraldistortion measures or (dis)similarity criteria used in image base browsing functions.

Callet and Barba [59] divided IQ metrics into two distinct groups:

• Those who use a HVS model for low level perception, such as sub-band decompositionand masking effect.

• Those who use little information about the HVS for error representation, and push theeffort on prior knowledge on introduced distortion.

The authors comment that IQ metrics belonging to the last group fail to be robust since theyare specialized.

Wang and Bovik [456] classified IQ metrics based on three criteria:

63

Page 80: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

• Full-reference, no-reference and reduced reference metrics.

• General-purpose and application specific metrics.

• Bottom-up and top-down metrics.

These categories are based on information about the original image, knowledge about thedistortion process, and knowledge about the HVS. The authors state that it is not possibleto draw sharp boundaries between the different groups, but that a crisp classification of IQmetrics can be helpful.

Chandler and Hemami [70] divided IQ metrics into three different groups:

• Mathematically convenient metrics, which only operate on the intensity of the distor-tion.

• Metrics based on near-threshold psychophysics, these metrics take into account the vi-sual detectability of the distortions.

• Metrics based on overarching principles such as structural or information extraction.

Thung and Raveendran [436] divided full-reference IQ metrics into three groups similarlyto Chandler and Hemami [70]:

• Mathematical metrics.

• HVS-based metrics.

• Others.

Recently, Seshadrinathan and Bovik [391] divided techniques for image and video qualityinto three main categories:

• HVS-based approaches.

• Structural approaches.

• Information theoretical approaches.

The authors also mention that IQ assessment can be broadly classified as bottom-up and top-down approaches, where bottom-up approaches model the HVS and top-down approachescharacterize specific features of the HVS.

4.1.2 Proposal for classification of image quality metrics

As seen above there are many different ways to group IQ metrics. In order to present thevarious approaches we have divided the IQ metrics into four groups:

• Mathematically based metrics, which operate only on the intensity of the distortions.These metrics are usually simple, such as the Mean Squared Error (MSE) and PeakSignal to Noise Ratio (PSNR).

• Low-level based metrics, which take into account the visibility of the distortions usingfor example Contrast Sensitivity Functions (CSFs), such as the Spatial-CIELAB (S-CIELAB) [499].

64

Page 81: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

• High-level based metrics, which quantify quality based on the idea that our HVS isadapted to extract information or structures from the image. The Structural SIMilar-ity (SSIM) [458], which is based on structural content, or the Visual Image Fidelity(VIF) [396], which is based on scene statistics, are examples of metrics in this group.

• Other metrics, which are either based on other strategies or combine two or more of theabove groups. One example is the Visual Signal-to-Noise Ratio (VSNR) [70], whichtakes into account both low- and mid-level visual properties, and the final stage incor-porates a mathematically based metric.

The three first groups are similar to the classification by Chandler and Hemami [70]. However,there are metrics that do not fit directly into one or more of three first groups, requiring a fourthclassification group.

4.2 Survey of existing image quality metrics

For each of these groups we will present some selected metrics. Since an impressive numberof IQ metrics have been proposed [340, 354] the selection of metrics will be based on stan-dards, existing literature, frequence of use, and novelty. For an extensive overview of existingmetrics we refer the reader to Appendix A.

In addition to introducing a selection of metrics, we will also identify differences betweenthe metrics. For this a test target has been adopted. The target (Figure 4.3) has been devel-oped by Halonen et al. [167] to evaluate print quality, and is a part of the image suite fromDigiQ (Section 3.2.4.7). In this test target ten specific color patches from the GretagMacbethColorChecker were incorporated as objects in the scene; blue, green, red, yellow, magenta,cyan, orange, and three neutral grays. This target should show differences between a widevariety of the metrics, since it contains different frequencies, colors and objects. Since we areapplying full-reference IQ metrics, a reproduction is needed. This has been obtained by tak-ing the original image, converting it to sRGB (which is the original used in the comparison),and further applying an ICC profile (EuroScale Uncoated V2) with the relative colorimetricrendering intent, then converting the image to sRGB, using Adobe Photoshop CS3 for all pro-cedures. This should result in many different quality changes, which will provoke differencesin the metrics.

4.2.1 Mathematically based metrics

The first group of metrics, mathematically based ones, has been very popular probably dueto their easy implementation, and they are convenient to use for optimization. These metricsusually only work on the intensity of the distortion E:

E(x,y) = IO(x,y)− IR(x,y), (4.1)

where IO is the original image, IR is the reproduction, x and y indicate the pixel position.

65

Page 82: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

(a) Original (b) Modified

Figure 4.3: Test target from Halonen et al. [167] to differentiate between IQ metrics. Theoriginal image has been processed by applying an ICC profile with the relative colorimetricrendering intent.

4.2.1.1 Mean squared error

MSE is a mathematically based metric; it calculates the cumulative squared error between theoriginal image and the distorted image. Most of the metrics in this group are strict metrics, thatis, where (ρ(IO(x), IR(y))) is essentially an abstract distance, with the following properties:ρ(IO(x), IR(y)) = 0 if IO(x) = IR(y), symmetry, triangle inequality, and non-negativity. MSEis given as:

MSE =1

MN

M−1

∑y=0

N−1

∑x=0

[E(x,y)]2 , (4.2)

where x and y indicate the pixel position, M and N are the image width and height. Severalothers of these metrics have been proposed as well, such as the PSNR and Root Mean Square(RMS). The MSE and other metrics have been used due to its easy calculation and analyticaltractability [435]. These simple mathematical models are usually not well correlated withperceived IQ [70, 456]. Still, they have been of influence to other and more advanced metrics,such as the PSNR-HVS-M [366] that takes into account contrast masking and CSFs to modelthe HVS.

4.2.1.2 ΔE∗ab

Metrics measuring color difference also belong to the group of mathematically based metrics.One of the most commonly used color difference formula stems from the CIELAB colorspace specification published by the CIE [2004], with the idea of a perceptually uniformcolor-space. In such a color space it is easy to calculate the distance between two colors, byusing the Euclidean distance. In engineering, this is known as the RMS value, which is thesquare root of the MSE. The ΔE∗

ab takes as input a sample color with CIELAB values L∗r , a∗r ,

b∗r and a reference color L∗o, a∗o, b∗o. The distance between the two color samples is given by

ΔE∗ab =

√(ΔL∗)2 +(Δa∗)2 +(Δb∗)2, (4.3)

66

Page 83: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

where ΔL∗ = L∗o −L∗

r , Δa∗ = a∗o −a∗r and Δb∗ = b∗o −b∗r .

The CIELAB formula has served as a satisfactory tool for measuring perceptual differencebetween uniform color patches. The HVS is not as sensitive to color differences in fine detailsas compared to large patches, yet the CIELAB color formula will predict the same visualdifference between the two cases since it does not have a spatial variable [500]. Because ofthe popularity of the ΔE∗

ab it has also been commonly used to measure natural images, wherethe color difference of each pixel of the image is calculated. The mean of these differences isusually the overall indicator of the difference between the original and the reproduction:

ΔE∗ab =

1MN

m−1

∑x=0

n−1

∑y=0

ΔE∗ab(x,y). (4.4)

Other measures of the ΔE∗ab can be the minimum value or the maximum value in the computed

difference. Because of the widespread use [21, 41, 78, 169, 233, 234, 339, 350, 356, 382,419, 462, 498, 500], its position in the industry, and because it is often used as a referencemetric we consider the ΔE∗

ab to be included in an overview of IQ metrics. Extensions of theΔE∗

ab have been proposed when it became apparent that it had problems, especially in theblue regions. First the CIE proposed the ΔE∗

94 [83] and later the ΔE∗00 [275]. These are of

increasing complexity compared to the ΔE∗ab, and therefore many still use the ΔE∗

ab.

4.2.1.3 ΔEE

Another very interesting color difference formula was proposed by Oleari et al. [329] forsmall and medium color differences in the log-compressed OSA-UCS (Optical Society ofAmerica’s Committee on Uniform Color Scales) space. Because of its different origin andpromising results [411, 412] we include it in this overview of IQ metrics. The formula isdefined as following:

ΔEE =√

(ΔLE)2 +(ΔGE)2 +(ΔJE)2, (4.5)

where

LE =1bL

ln

[1+

bL

aL(10LOSA)

], (4.6)

with aL = 2.890, bL = 0.015,

GE = −CEcos(h), (4.7)

JE = CEsin(h), (4.8)

67

Page 84: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

with

h = arctan

(− J

G

), (4.9)

CE =1

bCln

[1+

bC

aC(10COSA)

], (4.10)

with aC = 1.256, bC = 0.050,

COSA =√

G2 + J2. (4.11)

ΔEE is statistically equivalent to ΔE∗00 in the prediction of many available empirical datasets.

However, ΔEE is the simplest formula providing relationships with visual processing. Theseanalyses hold true for CIE 1964 Supplementary Standard Observer and D65 illuminant. Thisformula can also be used in the same way as ΔE∗

ab to measure the color difference of naturalimages by taking the average color differences over the entire image.

4.2.1.4 Difference between the metrics

The metrics introduced above are based on different ideas and they have different origins,because of this they will be different in the way they work. In order to understand the funda-mental differences between the metrics we will investigate their similarities and differences.

Since the MSE has been shown to correlate poorly with perceived IQ we do not consider itin this part, where we will look at comparison between the ΔE∗

ab and ΔEE . The difference be-tween these two is in the color space, since both are based on the Euclidean distance betweentwo color values. Both color spaces are opponent color spaces, where both have a lightness,red-green, and blue-yellow axis. The OSA-UCS color space has a better blue linearity thanthe CIELAB color space [299]. The OSA-UCS system also has the unique advantage of equalperceptual spacing among the color samples [286].

Both the ΔE∗ab and ΔEE have been computed for the test image (Figure 4.3), and the results

can be seen in Figure 4.4. Both results have been normalized with the maximum value, makingit easier to compare the results. The first apparent difference is the difference in the blue vase,which is due to known issues in the blue regions of the CIELAB color space [50]. It is alsointeresting to notice the differences in the dark regions, this could be because the OSA-UCScolor space, which ΔEE is based on compensate for the reference background lightness [337].There is also a difference in the skin tones, where ΔEE has almost no difference between theoriginal and the reproduction, while ΔE∗

ab indicates a larger difference.It is important to state that we do not evaluate whether the the differences are correlated

with perceived difference or quality, we merely show the difference between the metrics.Evaluation of the IQ metrics can be found in Chapter 7.

4.2.2 Low-level based metrics

Metrics classified as low-level based metrics simulates the low level features of the HVS, suchas CSFs or masking. However, most of these metrics use a mathematically based metric, forexample one of the metrics introduced above, in the final stage to calculate quality.

68

Page 85: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

200 400 600 800 1000 1200 1400 1600 1800 2000

200

400

600

800

1000

1200

1400 min

max

(a) ΔE∗ab color difference map.

200 400 600 800 1000 1200 1400 1600 1800 2000

200

400

600

800

1000

1200

1400 min

max

(b) ΔEE color difference map.

Figure 4.4: Comparison of color difference formulas, a) ΔE∗ab and b) ΔEE. The difference

is calculated between an original image (Figure 4.3(a)) and a reproduction (Figure 4.3(b)).White indicates highest difference between the original and the modified version of the testtarget, black a low difference.

4.2.2.1 S-CIELAB

When it became apparent that the ΔE∗ab was not correlated with perceived image difference

for natural images, Zhang and Wandell [499] proposed a spatial extension of the formula(Figure 4.5). This metric should fulfill two goals, a spatial filtering to simulate the blurring ofthe HVS and a consistency with the basic CIELAB calculation for large uniform areas.

Original andreproduction

Colorseparation

Spatialfiltering

CIELABcalculation

S-CIELABrepresentation

Figure 4.5: S-CIELAB workflow. The spatial filtering is done in the opponent color space,and the color difference is calculated using ΔE∗

ab.

The image goes through color space transformations, first the RGB image is transformedinto CIEXYZ and then further into the opponent color space (O1, O2, and O3):

O1 = 0.279X +0.72Y −0.107Z,

O2 = −0.449X +0.29Y −0.077Z,

O3 = 0.086X −0.59Y +0.501Z.

Now the image contains a channel with the luminance information (O1), one with the red-green information (O2), and one with blue-yellow information (O3). Then a spatial filter isapplied, where data in each channel is filtered by a 2-dimensional separable spatial kernel:

f = k∑i

wiEi, (4.12)

69

Page 86: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

whereEi = kie

[−(x2+y2)/σ2i ], (4.13)

and ki normalize Ei such that the filter sums to one. The parameters wi and σi are different forthe color planes as seen in Table 4.1. k is a scale factor, which normalizes each color plane sothat its two-dimensional kernel f sums to one.

Table 4.1: The parameters used for the spatial filtering in S-CIELAB, where wi is the weightof the plane and σi is the spread in degrees of visual angle.

Plane Weights wi Spreads σi

Luminance0.921 0.02830.105 0.133-0.108 4.336

Red-Green0.531 0.03920.330 0.494

Blue-Yellow0.488 0.05360.371 0.386

Finally the filtered image is transformed into CIE-XYZ, and this representation is furthertransformed into the CIELAB color space where the ΔE∗

ab is applied to calculate color dif-ferences. The final result, S-CIELAB1 representation, consists of a color difference for eachpixel, these values are usually averaged to achieve one value for overall image difference.

The S-CIELAB metric was originally designed for halftoned images, but it has also beenused to calculate overall IQ for a wide variety of distortions. It is often used as a referencemetric due to this simple implementation and since it formed a new direction of IQ metrics, inaddition it is often used by researchers [5, 6, 18, 21, 41, 45, 138, 169, 173, 220, 233, 234, 339,350, 351, 356, 412, 490, 498, 500]. It has also lead to many other IQ metrics, which use thesame framework as S-CIELAB, such as SDOG-CIELAB [5, 6] (Appendix E) where the spatialfiltering was performed with the difference-of-gaussians, and S-CIELABJ [224] where theCSF was modfiied in terms of adjusting the peak sensitivity to correlate better with complexstimuli. Because of this it is natural to include S-CIELAB in an overview of IQ metrics.

4.2.2.2 S-DEE

ΔEE also has a spatial extension, Spatial-DEE (S-DEE), proposed by Simone et al. [412].This metric follows the S-CIELAB framework, but the color difference is calculated withΔEE instead of ΔE∗

ab. The spatial filtering with CSFs is also a bit different from the originalS-CIELAB, since the S-DEE incorporates CSFs from Johnson and Fairchild [224], wherethe filters are implemented in the frequency domain rather than in the spatial domain as inS-CIELAB.

The luminance filter is a three parameter exponential function, based on research byMovshon and Kiorpes [308].

CSFlum(p) = a · pc · e−b·p, (4.14)

116/02/2011: The implementation of S-CIELAB as described here is available at http://white.standford.edu/˜brian/scielab/scielab.html

70

Page 87: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

where a = 75, b = 0.22, c = 0.78, and p is represented as Cycles Per Degree (CPD).The luminance CSF is normalized so that the DC modulation is set to 1.0, by doing this the

luminance shift is minimized. This will also enhance any image differences where the HVSis most sensitive to them [223]. For the chrominance CSF, a sum of two Gaussian functionsare used.

CSFchroma(p) = a1 · e−b1·pc1 +a2 · e−b2·pc2 , (4.15)

where different parameters for a1, a2, b1, b2, c1, and c2 have been used as seen in Table 4.2.

Table 4.2: The parameters used in S-DEE for the spatial filtering in the frequency domain ofthe chrominance channels.

Parameter Red-Green Blue-Yellowa1 109.14130 7.032845b1 -0.00038 -0.000004c1 3.42436 4.258205a2 93.59711 40.690950b2 -0.00367 -0.103909c2 2.16771 1.648658

The filters are applied in the 010203 color space as for S-CIELAB, but the final step con-sists of applying the ΔEE rather than the ΔE∗

ab to get the color differences. The overall imagedifference is achieved by averaging the results over the image.

Since the ΔEE has promising results, so has the S-DEE [412], and because it challengesthe standard calculation of the CIELAB color difference it should be included in this surveyof IQ metrics. A variant of the S-DEE, SDOG-DEE, was proposed by Ajagamelle et al. [5, 6](Appendix E), where the spatial filter is carried out by the difference-of-gaussians.

4.2.2.3 Adaptive image difference

Wang and Hardeberg [459] proposed an Adaptive Bilateral Filter (ABF) for image differencemetrics. The filter blurs the image, but preserves edges, which is not the case when usingCSFs. The bilateral filter is defined as:

h(x) = k−1(x)∫ ∞

−∞

∫ ∞

−∞f (ε)c(ε,x)s( f (ε), f (x))dε, (4.16)

wherek(x) =

∫ ∞

−∞

∫ ∞

−∞c(ε,x)s( f (ε), f (x))dε, (4.17)

and where the function c(ε,x) measures the geometric closeness between the neighborhoodcenter x and a nearby point ε:

c(ε,x) = exp(−(ε− x)2

2σ2d

). (4.18)

The function s(ε,x) measures the photometric similarity between the neighborhood center x

71

Page 88: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

and a nearby point ε:

s(ε,x) = exp

(−( f (ε)− f (x))2

2σ2r

). (4.19)

σd is the geometric spread determined by the viewing conditions in pixels per degree:

σd =n/2

180/π · tan−1(l/(2m)), (4.20)

where n is the width of the image, l is the physical length in meters, and m is the viewingdistance in meter. σr is the range spread and is determined with image entropy:

σr = K/E, (4.21)

where K is a constant to rescale the entropy into an optimized value and

E = −∑i

pi log(pi), (4.22)

where pi is the histogram of the pixel intensity values of an image.

The filtering is performed in the CIELAB color space, and the color difference formulaused was ΔE∗

ab, because of this the metric can be said to follow the same overlaying frameworkas S-CIELAB.

Because of the new type of filtering preserving edges, which has shown to produce goodresults for amongst others gamut mapping [40], this metric is interesting to compare with, andit shows promising results compared to other metrics [459].

4.2.2.4 Comparison of selected metrics within the group

The difference between the metrics in this group is mainly in the filtering stage and in thequality calculation. Since the quality calculation is carried out with a color difference formula(ΔE∗

ab or ΔEE), we will focus on the filtering. The S-CIELAB and S-DEE use CSFs, wherethey differ in terms of the normalization of the filters. On the other side, ABF uses a bilateralfilter preserving edges.

Figure 4.6 shows the difference of an image filtered with the three different metrics. Wecan see that S-CIELAB (the second image from the left) is slightly darker than the original(left image), which is caused by the CSF and has been reported earlier [224]. The imagefiltered with the method from S-DEE does show such a lightness shift since the DC componentof the CSF is normalized to one. We can also notice the blurring of the images from S-CIELAB and S-DEE, and the loss of details. The S-DEE has a CSF which enhances thefrequencies where the HVS is most sensitive, and therefore it will in some cases preservemore details than S-CIELAB. The ABF on the right has a softer ”look” compared to the othermetrics, most noticeable is the preservation of the hard edges, which is the advantage of thebilateral filter.

72

Page 89: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

Figure 4.6: Comparison between the different spatial filtering methods. From left to rightwe can see the original image, the image filtered with S-CIELAB, the image filtered with S-DEE, and the image filtered with ABF. All images have been computed for the same viewingconditions. In S-CIELAB we perceive a lightness shift due to the fact that the luminance filteris not normalized, in S-DEE the luminance filter is normalized which corrects this issue, theABF has a softer look since it smooths on each side of an edge using an edge preservingbilateral filter.

4.2.3 High level based metrics

High-level based metrics quantify quality based on the idea that our HVS is adapted to extractinformation or structures from the image.

4.2.3.1 SSIM

The SSIM index proposed by Wang et al. [458] attempts to quantify the visible differencebetween a distorted image and a reference image. This index is based on the Universal ImageQuality (UIQ) index [455]. The algorithm defines the structural information in an image asthose attributes that represent the structure of the objects in the scene, independent of theaverage luminance and contrast. The index is based on a combination of luminance, contrast,and structure comparison. The comparisons are done for local windows in the image, theoverall IQ is the mean of all these local windows. The SSIM is specified as:

SSIM(x,y) =(2μxμy +C1)(2σxy +C2)

(μ2x +μ2

y +C1)(σ2x +σ2

y +C2), (4.23)

where μ is the mean intensity for signals x and y, and σ is the standard deviation of the signalsx and y. C is a constant defined as

C1 = (K1L)2, (4.24)

where L is the dynamic range of the image, and K1 � 1. C2 is similar to C1 and is defined as:

C2 = (K2L)2, (4.25)

73

Page 90: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

where K2 � 1. These constants are used to stabilize the division of the denominator. SSIM isthen computed for the entire image as:

MSSIM(X ,Y ) =1

W

W

∑j=1

SSIM(x j,y j), (4.26)

where X and Y are the reference and the distorted images, x j and y j are image content in localwindow j, and W indicates the total number of local windows. Figure 4.7 shows the SSIMflowchart, where the input is the original signal (image) and the distorted signal (image).

Figure 4.7: SSIM flowchart, signal x and y goes through a luminance and contrast measure-ment before comparison of luminance, contrast and structure. A combination of these resultsin the final similarity measure. Picture from Wang et al. [458].

The SSIM has received a lot of attention since its introduction, gone through extensiveevaluation, and it has influenced a number of other metrics, such as the color version SSIM-IPT by Bonnier et al. [41] and the color version of UIQ by Toet and Lucassen [439]. Becauseof this and the interesting idea of using structural information to measure IQ we include SSIMin this survey.2

4.2.3.2 Visual information fidelity

Sheikh and Bovik [396] proposed the Visual Information Fidelity (VIF) criterion, which is anextension of the Information Fidelity Criterion (IFC) by the same authors [397]. It quantifiesthe Shannon information present in the reproduction relative to the information present in theoriginal. The natural scene model used is a Gaussian Scale Mixture model in the waveletdomain, and as a HVS model they use an additive white Gaussian noise model.

The reference image is modeled by a Gaussian Scale Mixture in the wavelet domain. c isa collection of M neighboring wavelet coefficients from a local patch in a subband. Then c ismodeled as c =

√zu, where u is a zero-mean Gaussian vector and

√z is an independent scalar

random variable. The VIF assumes that the distortion of the image can be described locally

216/02/2011: The implementation of SSIM can be found at http://www.ece.uwaterloo.ca/˜z70wang/research/ssim/

74

Page 91: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

as a combination of a uniform wavelet domain energy attenuation with added independentadditive noise. So that visual distortion is modeled as a stationary, zero-mean, additive whiteGaussian noise process in the wavelet domain: e = c + n and f = d + n, where e and f arethe random coefficient vectors for the same wavelet subband in the perceived original andperceived distorted image. c and d are random vectors from the same location in the samesubband for the original and distorted image. n denotes the independent white Gaussian noisewith the covariance matrix Cn = σ2

nI, where σ2n is a HVS model parameter (variance of the

internal neuron noise) and I denotes the set of spatial indices for the random fields. The VIFis calculated as the ratio of the summed mutual information in the subbands, which can bewritten as following:

V IF =I(C;F |z)I(C;E|z) =

∑Ni=1 I(ci; fi|zi)

∑Ni=1 I(ci;ei|zi)

,

where i is the index of local coefficients patches, including all subbands.

The VIF3 criterion has shown to perform well compared to other state of the art met-rics [253, 396, 398], and combined with the use of statistical information it is interesting tocompare it against more traditional metrics.

4.2.3.3 Comparison of selected metrics within the group

It is difficult to compare the metrics within this group since SSIM produces a map and VIFproduces a single value. To be able to compare the SSIM against the other metrics, such asΔEE , we show the resulting map from the target (Figure 4.8). Since SSIM is a grayscale metricthe images have been converted from color to grayscale4. SSIM show the biggest differencein the dark regions of the image (Figure 4.8). It is also interesting to notice the similaritybetween the ΔEE (Figure 4.4(b)) and SSIM (Figure 4.8). To verify the similarities betweenthese two difference maps a 2-D correlation coefficient has been used:

r =∑m ∑n(Amn −A)(Bmn −B)√

(∑m ∑n(Amn −A)2)(∑m ∑n(Bmn −B)2), (4.27)

where A is the mean of one map, and B is the mean for the other map, respectively. Thismethod has previously been used to compare results from eye tracking maps [17, 339], andto compare an IQ metric to the regions marked by observers [61]. The correlation coefficientbetween these maps gives a value of -0.75 (minus since the ΔEE has the highest differenceas max, while SSIM has the highest difference as min. m = 2126 and n = 1417, which givesN = 2126×1417 = 3012542), indicating similarities between the two maps.

316/02/2011: The implementation of VIF is available at http://live.ece.utexas.edu/research/quality/.

4For all grayscale metrics we have converted the image using the rgb2gray function in Matlab unlessstated otherwise, following the recommendation of Wang et al. 16/02/2011: https://ece.uwaterloo.ca/

˜z70wang/research/ssim/. rgb2gray is defined as: 0.2989∗R+0.5870∗G+0.1140∗B, for an RGB image.17/04/2011: http://www.mathworks.com/help/toolbox/images/ref/rgb2gray.html

75

Page 92: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

200 400 600 800 1000 1200 1400 1600 1800 2000

200

400

600

800

1000

1200

1400

max

min

Figure 4.8: SSIM difference map between an original image (Figure 4.3(a)) and a reproduc-tion (Figure 4.3(b)).

4.2.4 Other approaches

Others metrics considered in this group are based on other approaches or metrics combiningtwo or more of the above groups.

4.2.4.1 Visual signal to noise ratio

Chandler and Hemami [70] proposed a metric, Visual Signal to Noise Ratio (VSNR), basedon near-threshold and suprathreshold properties of the HVS, incorporating both low-levelfeatures and mid-level features. The metric consists of two stages; first contrast thresholdsare used to detect visible distortions in the image, which is done in the wavelet domain bycomputing the Contrast Signal to Noise Ratio (CSNR). Then the contrast detection thresholdis computed based on the CSNR, which is done for each octave band. The contrast is furthercompared to the detection threshold, and if above the distortion is supratreshold (visible). Inthis case a second stage is carried out, where a model of global precedence is proposed toaccount for the mid-level properties of the HVS. The global precedence takes into accountthat contrast of distortions should be proportioned across spatial frequency. The final metricis computed as the combination of perceived contrast of the distortion and disruption of globalprecedence.

The VSNR is an interesting metric since it is based on contrast thresholds, therefore itwill only take into account the visible difference, different from the CSF based metrics wherethe entire image is modulated. In addition the use of both low-level features and mid-levelfeatures is compelling, and therefore we include VSNR5 in this survey.

4.2.4.2 Color image similarity measure

Lee et al. [261] proposed the Color Image Similarity Measure (CISM) with the intention forevaluating a halftoning algorithm. The CISM can be divided into two parts; one dealing withthe HVS and one with structural similarity. The input images are converted to the CIELAB

516/02/2011: The code for VSNR can be downloaded at http://foulard.ece.cornell.edu/dmc27/vsnr/vsnr.html

76

Page 93: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

color space where they are filtered with CSFs from Sullivan et al. [428] and Nasanen [313] forthe luminance channel and from Mullen [309] for the chrominance channels. After filteringthe images are converted to RGB before each channel is used as input to the SSIM [458]framework (as introduced above). From each of the RGB channels a mean SSIM value isobtained, which is combined in order to get one value for IQ. The final value is a weightedsum of the three channel:

CISM = ∑i

wiSSIMi, (4.28)

where i indicates the channel, and w is the weight (1/3 suggested by Lee et al. [261]).

4.2.4.3 Comparison of selected metrics within the group

Since both the metrics in this group produce one quality value, and not a map it is difficultto show the differences between them. Nevertheless, they have one fundamental difference:VSNR is made for grayscale images while CISM for color images. Also the difference be-tween the spatial filtering should be noted, where VSNR is built on contrast visibility andCISM on CSFs.

4.3 Summary

In this chapter we have discussed the structure of existing IQ metrics, and proposed a classi-fication into four different groups. Further, a survey of metrics is given, where metrics fromeach of the classifications groups have been presented and differences and similarities betweenthem have been discussed.

77

Page 94: Doctoral Dissertation by Marius Pedersen

EXISTING IMAGE QUALITY METRICS

78

Page 95: Doctoral Dissertation by Marius Pedersen

5 SHAME: A NEW IMAGE QUALITY

METRIC

Many of the existing IQ metrics weight the content of the images equally, even though itis a well-known fact that observers put more emphasis on certain regions when they judgeIQ [114]. It has also been shown that using information about where observers gaze canincrease the performance of IQ metrics [339, 356] and increase compression without loosingquality [47]. Based on this IQ metrics can be improved by using a metric that takes intoaccount regions of interest (i.e. does not weight all pixels equally). However, IQ metricsshould also simulate the HVS, since our perception of quality is a function of the viewingconditions [98], such as the distance an image is viewed at. Combining these two key aspectsof IQ could potentially result in a metric giving results that are better correlated with theresults of human perception. Therefore, in this chapter, we present a new IQ metric takingthese two aspects into account. The proposed metric, Spatial Hue Angle MEtric (SHAME),is based on the hue angle algorithm [177, 178], and two different spatial filtering methods aretested. We will first give an overview of the hue angle algorithm, and then the two spatialfiltering methods.

5.1 The hue angle algorithm

Hong and Luo [177, 178] proposed a full-reference color image difference metric built on theCIELAB color difference formula [85]. This metric is based on the known fact that systematicerrors over the entire image are quite noticeable and unacceptable. The metric is based onsome conjectures; summarized from Hong and Luo [177, 178] these are:

• Pixels or areas of high significance can be identified, and suitable weights can be as-signed to these.

• Pixels in larger areas of the same color should be given a higher weight than those insmaller areas.

• Larger color difference between the pixels should get higher weights.

• Hue is an important color perception for discriminating colors within the context.

The first step is to transfer each pixel in the image from L∗, a∗, b∗ to L∗, C∗ab, hab. Based on

the hue angle (hab) a histogram from the 360 hue angles is computed, and sorted in ascendingorder based on the number of pixels with same hue angle to an array k. Then weights can be

79

Page 96: Doctoral Dissertation by Marius Pedersen

SHAME: A NEW IMAGE QUALITY METRIC

applied to four different parts (quartiles) of the histogram, and by doing this Hong and Luocorrected the drawback that the CIELAB formula weights the whole image equally. The firstquartile, containing n hue angles, is weighted with 1/4 (that is, the smallest areas with thesame hue angle) and saved to a new array hist. The second quartile , with m hue angles, isweighted with 1/2. The third quartile, containing l hue angles, is given 1 as a weight and thelast quartile with the remaining hue angles is weighted with 9/4.

hist(i) =

⎧⎪⎪⎪⎨⎪⎪⎪⎩

k(i)∗1/4, i ∈ {0, ...,n}k(i)∗1/2, i ∈ {n+1, ...,n+m}k(i)∗1, i ∈ {n+m+1, ...,n+m+ l}k(i)∗9/4, otherwise

(5.1)

The average color difference, computed using ΔE∗ab, is calculated for all pixels having

the same hue angle and stored in CD[hue]. Then the overall color difference for the image,CDimage, is calculated by multiplying the weights based on the quartiles for every pixel withthe average CIELAB color difference for the hue angle

CDimage =359

∑0

hist[hue]∗CD[hue]2/4. (5.2)

5.2 Spatial filtering

We propose two different spatial filtering methods for the hue angle algorithm. The firstspatial filtering is adopted from S-CIELAB [499] (Section 4.2.2.1). The advantage of usingthis spatial filtering is that has been extensively evaluated [18, 21, 41, 45, 138, 169, 173, 220,233, 234, 339, 350, 351, 356, 490, 498, 500], and it has shown to produce good results for awide variety of distortions.

The image goes through color space transformations, first the RGB image is transformedinto CIEXYZ and further into the opponent color space (O1,O2, and O3) [499].

O1 = 0.279X +0.72Y −0.107Z,

O2 = −0.449X +0.29Y −0.077Z,

O3 = 0.086X −0.59Y +0.501Z.

Now the image contains a channel with the luminance information (O1), one with the red-green information (O2), and one with blue-yellow information (O3). Then a spatial filter isapplied, where data in each channel is filtered by a 2-dimensional separable spatial kernel:

f = k∑i

wiEi, (5.3)

whereEi = kie

[−(x2+y2)/σ2i ], (5.4)

and ki normalize Ei such that the filter sums to one. The parameters wi and σi are different for

80

Page 97: Doctoral Dissertation by Marius Pedersen

SHAME: A NEW IMAGE QUALITY METRIC

the color planes as seen in Table 5.1. k is a scale factor, which normalize each color plane soits two-dimensional kernel f sums to one.

Table 5.1: The parameters used for the spatial filtering, where wi is the weight of the planeand σi is the spread in degrees of visual angle.

Plane Weights wi Spreads σi

Luminance 0.921 0.02830.105 0.133-0.108 4.336

Red-Green 0.531 0.03920.330 0.494

Blue-Yellow 0.488 0.05360.371 0.386

The second spatial filtering proposed is adopted from Johnson and Fairchild [223]. Byspecifying and implementing the spatial filters using CSFs in the frequency domain, ratherthan in the spatial domain as the first spatial filtering, more precise control of the filters isobtained [223], but usually at the cost of computational complexity. Results also show in-creased performance of metrics where the spatial filtering is performed in the frequency do-main [223, 351, 412]. The luminance filter is a three parameter exponential function, basedon research by Movshon and Kiorpes [308].

CSFlum(p) = a · pc · e−b·p, (5.5)

where a = 75, b = 0.22, c = 0.78, and p is represented as Cycles Per Degree (CPD).The luminance CSF is normalized so that the DC modulation is set to 1.0, by doing this

the luminance shift is minimized, compared to the first spatial filtering method. This will alsoenhance any image differences where the human visual system is most sensitive to them [223].For the chrominance CSF, a sum of two Gaussian functions are used.

CSFchroma(p) = a1 · e−b1·pc1 +a2 · e−b2·pc2 , (5.6)

where different parameters for a1, a2, b1, b2, c1 and c2 have been used as seen in Table 5.2. Agraph of the luminance and chrominance CSFs is shown in Figure 5.1.

Table 5.2: The parameters used for the spatial filtering in the frequency domain of the chromi-nance channels.

Parameter Red-Green Blue-Yellowa1 109.14130 7.032845b1 -0.00038 -0.000004c1 3.42436 4.258205a2 93.59711 40.690950b2 -0.00367 -0.103909c2 2.16771 1.648658

81

Page 98: Doctoral Dissertation by Marius Pedersen

SHAME: A NEW IMAGE QUALITY METRIC

0 10 20 30 40 50 600

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8Contrast Sensitivity Functions

Cycles per degree

Sen

sitiv

ity

Luminance CSFRed/Green CSFBlue/Yellow CSF

Figure 5.1: CSFs used in the second spatial filtering.

5.3 Applying spatial filtering to the hue angle algorithm

The images are spatially filtered with the previously introduced spatial filtering methods. Thisresults in a filtered original and a filtered modified version of the original, which are usedas input to the hue angle algorithm, as shown in Figure 5.2. This can be considered as anextension of the S-CIELAB (Section 4.2.2.1) flowchart on Figure 4.5.

The hue angle algorithm, filtered respectively with the first and second filter, is from nowon referred to as SHAME-I and SHAME-II. The new metric will theoretically have severalkey features from both the S-CIELAB and the hue angle measure:

• Weight allocation: pixels in larger areas of the same color should be weighted higher.

• Simulation of the spatial properties of the human visual system.

• Undetectable distortions are ignored.

• Suitable for different kind of distortions, not only color patches.

• Generates one value for easy interpretation.

Transformimage

toopponent

colorspace

Applyspatialfilters

Transformfiltered

image toCIELABthen to

CIELCH

Calculatehistogram

Sorthistogram

andapply

weights

Calculateaverage

colordifference

Sum thecombined

hueweights

and colordifferences

Figure 5.2: Workflow of the proposed metric.

82

Page 99: Doctoral Dissertation by Marius Pedersen

SHAME: A NEW IMAGE QUALITY METRIC

5.4 Classification

SHAME can be considered as an approach combining several different features, and wouldbe classified together with the metrics in the group ”other approaches” in Section 4.2.4.

5.5 Summary

In this chapter a new IQ metric, SHAME, was proposed. SHAME takes into account theHVS and incorporates information about region of interest. Two different spatial filteringmethods are suggested, the one from Zhang and Wandell [499] and the one from Johnson andFairchild [223]. Evaluation of the metric is found later in Chapter 7.

83

Page 100: Doctoral Dissertation by Marius Pedersen

SHAME: A NEW IMAGE QUALITY METRIC

84

Page 101: Doctoral Dissertation by Marius Pedersen

6 HOW TO EVALUATE IMAGE QUALITY

METRICS

There are a number of algorithms involved in the reproduction of an image, for instance gamutmapping, halftoning, and compression. To identify the best reproduction among a numberof variants of the same reproduction algorithm (e.g. JPEG compression), a psychophysicalexperiment can be carried out. This will result in a scale with the visual difference of the re-productions from the original. These psychophysical experiments are both time and resourcedemanding. Because of this, objective methods such as IQ metrics have been introduced toentirely or partially eliminate psychophysical experiments. However, in order to know if anIQ metric correlates with the human percept, some kind of evaluation of the metric is required.The most common evaluation of IQ metrics is by comparing the results of the metrics to theresults of human observers. In this chapter we investigate the different methods to evaluateIQ metrics.

6.1 Evaluation methodology

We will give a brief introduction to the methodology for evaluation of IQ metrics. An ex-periment is set up following the instructions given in Section 3.2, where a set of images isevaluated by a group of observers. The results from the observers can be statistically ana-lyzed, as suggested in Section 3.2. The mean values from the subjective evaluation are usedas the basis against which the metrics are evaluated. For the same set of images the metricsare calculated, resulting in one value describing the quality of the image. The results from theIQ metrics are then compared against the results of the observers, forming the basis for theevaluation.

6.2 Existing evaluation methods

A number of existing evaluation methods are found in the literature. We will go through themethods in which IQ metrics are evaluated against perceptual data.

85

Page 102: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

6.2.1 Correlation based methods

Correlation is a common method for the evaluation of IQ metrics. Correlation describes thestatistical relationship between two or more variables.

6.2.1.1 Pearson product-moment correlation coefficient

The most common measure of correlation is the Pearson product-moment correlation coef-ficient [232], which is a linear correlation between two variables. Pearson’s correlation r isdefined as:

r =∑N

i=1(Xi −X)(Yi −Y )√∑N

i=1(Xi −X)2√

∑Ni=1(Yi −Y )2

, (6.1)

where X and Y are two variables and N is the number of samples. The correlation value ris between −1 and +1. For evaluation of metrics, one of the variables is the results from apsychophysical experiment (for example z-scores) and the other variable is the results froman IQ metric. Depending on the scale of the metric, whether a low number is high qualityor a high number is high quality, the best correlation will be −1 or +1, for easy comparisonof results the correlation values of the metrics where a low number indicates high quality theresults are usually multiplied with −1. Unless stated otherwise this is done for all results inthis work. This method has been extensively used in the evaluation of metrics [113, 169, 279,459, 495].

Significance test Significance tests can be carried out for Pearson’s correlation by using thefollowing equation:

t =r√

N −2√1− r2

, (6.2)

where N is the number of pairs of scores, and with N−2 degrees of freedom (number of valuesin the final calculation of a statistic). A t table can further be used to find the probability value(sometimes called p, i.e. the probability of obtaining a statistic as different from or moredifferent from the parameter specified in the null hypothesis).

Confidence intervals Confidence intervals can also be calculated for Pearson’s correlation.One consideration is that the sampling distribution of Pearson’s r is not normally distributed.Therefore, Pearson’s r is converted to Fisher’s z’ and the confidence interval is computedusing Fisher’s z’. The values of Fisher’s z’ in the confidence interval are then converted backto Pearson’s r.

The first step is to use the Fisher’s Z-transform:

z =12

ln

(1+ r1− r

), (6.3)

where r is the correlation coefficient. The confidence intervals for r are calculated on thetransformed r values (z′). The general formulation of confidence intervals for z′ is

z′ ± zσz′, (6.4)

86

Page 103: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

where the criterion z is the desired confidence level (1.96 gives a 95% confidence interval),and σz′ is defined as

σz′ =1√

N −3, (6.5)

where N is the number of correlation samples. The upper and lower limits confidence intervallimits for z′ are found by using Equation 6.4. To translate from the z-space back in to ther-space, it is necessary to invert Equation 6.3. This results in the following equation:

r =e2z −1e2z +1

. (6.6)

The upper and lower confidence interval limits for r can be found by using the upper andlower z values into Equation 6.6.

Figure 6.1 shows an overview of the lowest and highest correlation limits for N values from16 to 1024 (95% confidence interval). Figure 6.1(a) show the lowest limit, and Figure 6.1(b)the highest. The higher the number of data points (N) the smaller the confidence interval, fora small number of data points the confidence interval is large and important to be aware of.

6.2.1.2 Spearman’s rank-correlation coefficient

Another correlation measure often used [253, 279, 311, 364, 366, 398, 458, 495] is the Spear-man rank-correlation coefficient [232], which is a non-parametric measure of associationbased on the ranks of the data values that describes the relationship between the variableswithout making any assumptions about the frequency distribution. In the case of no tied ranksSpearman correlation (r) is defined as:

r = 1− 6∑di2

N(N2 −1), (6.7)

where the raw scores Xi and Yi are converted into ranks xi and yi, and the difference betweenthe ranks is di = xi−yi. N is the number of samples. In the case where ranks exist the Pearsoncorrelation coefficient (Equation 6.1) should be used for the calculation of correlation [32].Many report the Spearman correlation since we cannot expect a linear relation between themetric and perceived distortion due to the non-linearities in the HVS [439].

Confidence intervals can be calculated for Spearman with the same procedure as for Pear-son [10], since the distribution of Spearman’s r is similar to that of Pearson’s r.

6.2.1.3 Kendall tau rank correlation coefficient

A third correlation measure is also found in the literature [311, 364, 366, 495] for the eval-uation of metrics. The Kendall tau rank correlation coefficient [232] is a measure of rankcorrelation as Spearman, but the Kendall rank correlation interpretation is easier since thecorrelation represents the difference between two probabilities. Kendall tau rank correlationcoefficient is defined as :

τ =(number of concordant pairs)− (number of discordant pairs)

12N(N −1)

, (6.8)

87

Page 104: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

0.5

0.7

0.9

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

1664

2561024

Correlation

Lo

we

r li

mit

co

rre

lati

on

Number of datapoints

0.0-0.1 0.1-0.2 0.2-0.3 0.3-0.4 0.4-0.5 0.5-0.6 0.6-0.7 0.7-0.8 0.8-0.9 0.9-1.0

(a) Low limit of correlation

(b) High limit of correlation

Figure 6.1: Low and high correlation limits for different N values. The low limit shows thebottom of a 95% confidence interval of a correlation value, while the high limit shows the topof the 95% confidence interval. The figure shows values for N between 16 and 1024, and onlycorrelation values above 0.5. Figure b) has been rotated for better representation.

88

Page 105: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

where two random variables X and Y and their pair (xi,yi) and (x j,y j) are said to be concordantif the rank of both elements agree, and they are discordant if xi > x j and yi < y j or if xi < x j

and yi > y j. N is the number of samples, and τ is the Kendall correlation coefficient.

6.2.2 Error methods

6.2.2.1 Root mean squared error

Root Mean Squared Error (RMSE) calculation is used to quantify the error between fittedobjective data and corresponding subjective data [95, 190]. RMSE is given by

RMSE =

√1N

N

∑i=1

(Oi −Si)2, (6.9)

where N is the number of samples, O is the objective data, and S the subjective data. Thismethod has been used by a number of researchers [95, 253, 311, 396, 398, 458, 495].

6.2.2.2 Mean absolute error

The Mean Absolute Error (MAE) is used to measure how close predictions are from the actualvalues. MAE is given by

MAE =1N

N

∑i=1

|ei|, (6.10)

where N is the number of samples, and ei is the absolute error of fi−yi between the predictionfi and the true value yi. This method has been used by some researchers [279, 458].

6.2.3 Outlier ratio

Outlier ratio [450, 461] is the ratio of ”false” scores by an IQ metric to the total number ofscores. The ”false” scores are the scores that lie outside the interval [MOS−2σ,MOS+2σ],where MOS is the mean opinion score and σ the standard deviation of the MOS. This is a lessused method, but some researchers [279, 458] calculate this in addition to other methods.

6.3 New method for the evaluation of image quality metrics

Existing methods, such as those based on correlation, are sensitive to scale differences be-tween images [296] (Figure 6.2) and extreme values. Especially problematic is the use ofcorrelation when it is calculated on data coming from differently scaled and shifted scenes.These are problems when one would like to have a value indicating the overall performance ofa metric over a larger set of images. Because of this we propose a new method for evaluatingthe overall performance of IQ metrics (Figure 6.3). We propose to process the results fromIQ metrics in the same way as we commonly do with the results from a perceptual rank orderevaluation, i.e. that a metric simulates an observers. For each scene in the dataset we obtain

89

Page 106: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

0

0.5

1

1.5

2

2.5

0 2 4 6 8 10

Ob

se

rv

er

Sc

or

e

Metric Score

Figure 6.2: Metric values plotted against observer values for two different scenes, and theregression line (black). The correlation within each scene (circles) is rather high, but corre-lation of all scenes is lower due to scale differences between the scenes. One should note thatcorrelation is invariant to scaling and shift, while the slope and intercept of the regressionline is not.

a ranking based on the metric results, this is then used as input to the rank order method (asdescribed in Section 3.2.2.3). The result is z-scores for the overall performance of the metric,which can be directly compared to the z-scores from observers. One disadvantage of usingrank order is the number of scenes needed, in order to provide useful results the number ofscenes must be high (following the same recommendations regarding the number of observersas described in Section 3.2.4.1). The more scenes used the more accurate results. The moredata provided the more accurate the results will become, and the confidence intervals in themetric’s z-score will become smaller. The confidence intervals can be used as a measureof when a reproduction method is significantly better than another, and this provides moreinformation than Pearson’s or Spearman’s correlation coefficients.

Compare

z-scores

Image quality

metrics

• Calculate metric results

• Compute rank for each

scene

• Calculate z-scores

Observers

• Experiment results

• Calculate z-scores

Figure 6.3: The proposed method for evaluation of IQ metrics. The rank for the IQ metrics iscalculated for each scene, these rankings are used as input to the rank order method, whichresults in an overall z-score for the metric. Z-scores are also calculated for the observers. Thetwo z-scores can now be compared, if the metric z-scores are similar to the observer z-scoresthe metric has the same overall performance as the observers, indicating that it is able topredict overall perceived IQ.

90

Page 107: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

6.3.1 Example of the proposed method based on rank order

An example is given to show how the method can be used in the evaluation of IQ metrics.

6.3.1.1 Psychophysical experiment

Figure 6.4: Images from the experiment carried out by Dugay et al. [108]. These images havebeen gamut mapped with five different algorithms.

20 different images (Figure 6.4) were chosen for a psychophysical experiment [107, 108].Considering the color gamut of an Oce TCS 500 printer, the amount of out-of-gamut colorsranged from 25% to 100% with a mean of 57%. The images were reproduced using fivedifferent Gamut Mapping Algorithms (GMAs).

• HPminDE (Hue preserving minimum ΔE∗ab clipping) which is a baseline GMA pro-

posed by the CIE [80]. The algorithm does not change in-gamut colors at all, whileout-of-gamut colors are mapped to the closest color on the destination gamut whilepreserving the hue.

• SGCK (chroma-dependent SiGmoidal lightness mapping and Cusp Knee scaling) [80]is an advanced spatially invariant sequential gamut compression algorithm. The light-ness is first compressed by a chroma dependent sigmoidal scaling, resulting in highchroma colors being compressed less than neutral ones. The resulting colors are thencompressed along lines toward the cusp [303] of the destination gamut using a 90%knee scaling function. For the final compression the image gamut is used as the sourcegamut.

• Zolliker and Simon [502] proposed a spatial GMA; its goal being to recover local con-trast while preserving lightness, saturation and global contrast. A simple clipping isperformed as a first step; then by using an edge-preserving high pass filter the differ-ence between the original and gamut clipped image is filtered. The filtered image isthen added to the gamut clipped image. As a last step the image is clipped in order tobe in-gamut.

• Kolas and Farup [242] recently proposed a hue- and edge-preserving spatial color GMA.The image is first gamut clipped along straight lines toward the center of the gamut. A

91

Page 108: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

relative compression map is then created from the original and clipped image. Usingthis compression map, a new image can be constructed as a linear convex combinationof the original image and neutral gray image. This image is in turn filtered by an edge-preserving smoothing minimum filter. As the final step the gamut mapped image isconstructed as a linear convex combination of the original image and neutral gray usingthe filtered map.

• Farup et al. [131] proposed a multiscale algorithm (Gatta) preserving hue and localrelationship between closely related pixel colors. First a scale-space representation ofthe image and then gamut clipping the lowest scale is constructed. The resulting gamutcompression is then applied to the image at the next smallest scale. Operators are usedto reduce the effect of haloing. The process is repeated until all scales are treated. TheFourier domain is used to speed up the process.

The 20 different images have been evaluated by 20 observers in a pair comparison experi-ment [107, 108]. All observers had normal or corrected to normal color vision. The observerswere presented with the original image in the middle of the screen, with two different repro-ductions on each side. The observers were asked to pick the image with the most accuratereproduction with respect to the original image. When the observer had picked one image, anew pair of reproductions was shown until all combinations were evaluated. All pairs werealso shown twice in opposite order for consistency. The monitor was a Dell 2407WFP LCDdisplay calibrated with a D65 white point and a 2.2 gamma. The viewing conditions were cho-sen as close to the ones described in the CIE guidelines [80] as possible. The level of ambientillumination was measured to approximately 20 lux. The observer was seated approximately50 cm from the screen.

6.3.1.2 Z-scores

The z-scores are based on Thurstone’s law of comparative judgment [116, 437]. Data col-lected are transformed into interval scale data where scores represent the distance of a givenimage from the mean score of a set of images in units of the standard deviation of thescenes [302], and therefore being relative. The 95% confidence intervals are calculated inthe same way as proposed by Morovic [302]. The error bars are then computed as

X ± σ√N

,

where X is the Z-score, σ is the standard deviation, and N is the size of the sample set. Forthese experiments this is the number of observers multiplied with two, because each imagepair was shown twice for consistency. With this confidence interval there is a 95% estimatethat the value will be within the interval, and if the confidence interval of another GMA isoutside this interval the difference is significant.

6.3.1.3 Selected image quality metrics

Five IQ metrics have been chosen, ΔE∗ab, S-CIELAB [499], iCAM [125], SSIM [458], and the

hue angle algorithm [177]. All metrics except SSIM have a scale where closer to 0 indicate areproduction closer to the original, while SSIM has a scale between 0 and 1, where 1 indicatean identical reproduction.

92

Page 109: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

Figure 6.5: Z-score from the pair comparison experiment. Gatta cannot be differentiatedfrom SGCK and Zolliker. Kolas is rated barely lower than the Gatta, but SGCK cannot bedifferentiated from Zolliker. HPminDE receives the lowest z-score from the observers.

6.3.1.4 Results

From the pair comparison experiment z-scores were calculated, indicating the performanceof the different GMAs. From Figure 6.5 we can see that the Gatta algorithm gets the highestscore from the observers, indicating the lowest visual difference from the original. This algo-rithm gives statistically the same visual difference from the original as the SGCK and Zollikeralgorithm. Kolas has the fourth best score, but has the same visual difference as Zolliker.HPminDE clearly gives the highest visual difference from the original, the low score here in-dicating a large consensus among the observers about the low performance of this algorithm.

6.3.2 Rank order based on metric order

Aiming to develop a universal IQ metric, this metric should work across multiple scenes andin different conditions. One way of evaluating the performance of IQ metrics is to check thecorrelation between the perceived IQ and the calculated IQ [22, 305] as seen in Figure 6.6.We can see from Figure 6.6 that the data points are very spread, and there is no correlation.Thus it is not possible to use the IQ results as a way of evaluating the performance of the bestGMA. This is the case for all the metrics, where the correlation is generally low for all scenesas seen in Table 6.1 for both Pearson’s correlation and Spearman’s rank order correlation. TheSpearman rank order correlation does not take into account the frequency distribution of thevariables, and should therefore be less sensitive to extreme values, but as seen in Table 6.1this is not a good measure for overall prediction of performance. The z-score for each scenedoes not say anything about the difference between scenes; it is based on how preferred eachimage is within each scene. We follow the above proposal to process the IQ metric results inthe same way as we commonly do with the results from a perceptual rank order evaluation.

93

Page 110: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

0 5 10 15 20 25 30−2

−1.5

−1

−0.5

0

0.5

1

Metric score

Z−

scor

eZ−score vs. metric values for all scenes. Correlation −0.079

Figure 6.6: Z-score from observers against ΔE∗ab values. The data points are very spread, and

we get a very low correlation between the z-scores and ΔE∗ab values. The HPminDE algorithm

is rated as the best by ΔE∗ab, opposite of the observers (Figure 6.5).

Table 6.1: Correlation between all z-scores and all algorithms. The correlation here is lowfor all metrics both for Pearson and Spearman. SSIM has the highest Pearson’s correlation,however, not being higher than 0.16 it indicates a low performance. For the Spearman cor-relation the hue angle algorithm has the highest correlation, but still very low. The plot forΔE∗

ab with a linear fitted line and calculated Pearson’s correlation is found in Figure 6.6.

CorrelationMetric Pearson SpearmanΔE∗

ab -0.08 0.11SSIM 0.16 0.05S-CIELAB -0.06 0.10iCAM 0.01 0.07Hue angle -0.11 0.13

94

Page 111: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

The rank for each metric in the 20 scenes have been used as a basis for the overall perfor-mance of the GMAs, this correspond to 20 observers in a ranking experiment. If the resultsfrom the metrics match the overall results from the psychophysical experiment (Figure 6.5),the metrics predict perceived IQ. This method will only provide information about the orderof the image samples, not information of the distance between the samples. The number ofimages used should follow the recommendations for the number of observers (Section 3.2.4.1)to produce reliable data. CIE [80] recommends at least 15 observers, in this case 15 images.

In principle the rank order and pair comparison approaches provide the same informa-tion [91]. The rank order data has been used to generate corresponding pair comparisondata [91, 160], and the z-scores were computed as for a pair comparison experiment [116,302]. Babcock [17] got similar score for pair comparison, rank order, and graphical rating.This implicates that scale values from one type of experiment can be directly compared toscale values from another type of experiment. The rank order z-scores in this experiment havebeen calculated by using the Colour Engineering Toolbox [162].

Gatta HPminDE Kolaas SGCK Zolliker−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

Δ E*ab

− mean values

(a) ΔE∗ab rank order score.

−2 −1.5 −1 −0.5 0 0.5 1 1.5 2−0.7

−0.6

−0.5

−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

Algorithm Z−score

Obs

erve

r Z

−sc

ore

Perceived Z−score versus Δ E*ab

Z−score. Correlation: −0.62

(b) Observer z-score against ΔE∗ab rank order score.

Figure 6.7: Rank order score for ΔE∗ab values with a 95% CI, and these values plotted against

z-scores from observers with a 95% CI with a linear regression line.

From Figure 6.7(a) we can see results from the rank order for ΔE∗ab. The HPminDE have

the highest z-score, but this has the lowest z-score from the observers. This GMA will clipthe color to the minimum ΔE∗

ab and will always be rated as the best by the ΔE∗ab formula. The

SGCK gets a very low score in the ranking, while the observers rated this as one of the bestGMAs. Figure 6.7(b) shows z-score from the observers plotted against the rank order z-scorefrom ΔE∗

ab. The Pearson correlation here is -0.62, indicating a low performance for the ΔE∗ab.

This is important to notice that the HPminDE strongly influence the results, since it is judgedclearly worse than the other GMAs (Figure 6.5).

Figure 6.8(a) shows the results for SSIM, as we can see the HPminDE gets the lowest scoreby SSIM. This is the same as the observers. The four other GMAs cannot be differentiatedwith a 95% confidence interval. The observers also have very small differences between thesealgorithms; the Kolas algorithm has a score just lower than the SGCK, Gatta, and Zolliker.The score from SSIM is very similar to the score from the observers, this is also verified witha correlation between the scores of 0.98 (Figure 6.8(b)). The correct ranking of the HPminDEGMA is the basis for the excellent correlation here, and this ranking (Figure 6.8(a)) also reflectthe observers ranking (Figure 6.5).

95

Page 112: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

The overall Pearson correlation is only 0.16 and Spearman correlation is 0.05 (Table 6.1)between the z-scores for each scene and SSIM scores for each scene. These are therefore notgood measured of overall performance, even though high correlation can be found within eachscene in both measures. The correlation within a scene can be average, but the ranking can becorrect as seen on Figure 6.9. Here the Spearman correlation will perform well, while Pear-son’s only perform average. In other cases the Spearman correlation will perform average,while the Pearson correlation will perform excellent. Spearman’s rank order correlation canprovide low correlation where clusters of data are found, but the ranking within the clusteris not necessarily correct. When the ranking is used in the rank order method, the normaldistribution is taken into account and will therefore better handle extreme values. Becauseof this the ranking of the results within each scene and using these as a basis for the rankorder z-scores, SSIM will reflect perceived IQ. The overall Pearson correlation is only 0.16and Spearman correlation is 0.05 (Table 6.1) between the z-scores for each scene and SSIMscores for each scene. These are therefore not good measured of overall performance, eventhough high correlation can be found within each scene in both measures. The correlationwithin a scene can be average, but the ranking can be correct as seen on Figure 6.9. Here theSpearman correlation will perform well, while Pearson’s only perform average. In other casesthe Spearman correlation will perform average, while the Pearson correlation will perform ex-cellent. Spearman’s rank order correlation can provide low correlation where clusters of dataare found, but the ranking within the cluster is not necessarily correct. When the ranking isused in the rank order method, the normal distribution is taken into account and will thereforebetter handle extreme values. Because of this the ranking of the results within each scene andusing these as a basis for the rank order z-scores, SSIM will reflect perceived IQ.

Gatta HPminDE Kolaas SGCK Zolliker−2

−1.5

−1

−0.5

0

0.5

1SSIM − mean values

(a) SSIM rank order score.

−2 −1.5 −1 −0.5 0 0.5 1−0.7

−0.6

−0.5

−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

Algorithm Z−score

Obs

erve

r Z

−sc

ore

Perceived Z−score versus SSIM Z−score. Correlation: 0.98

(b) Observer z-score against SSIM rank order score.

Figure 6.8: Rank order score for SSIM values with a 95% CI, and these values plotted againstz-scores from observers with a 95% CI. The results here indicate a very high performance bythe SSIM, the correct ranking of the HPminDE is the main reason for this, but also the similarranking of the remaining GMAs. Since we only have five data points (N = 5), the CI will be[0.72,1.00]. However, this depends strongly on the HPminDE observation.

There are large differences between the score from iCAM and the observers, mostly dueto the calculated values for HPminDE in iCAM. iCAM has a spreading of the four best algo-rithms from the observers, and the HPminDE is rated as the best by iCAM, opposite of theobserver rating. The Pearson correlation between the observer z-score and iCAM z-score is

96

Page 113: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

0.98 0.982 0.984 0.986 0.988 0.99 0.992−0.6

−0.4

−0.2

0

0.2

0.4

0.6

SSIM score

Z−

scor

e

Correlation SSIM for scene 14: 0.75

Figure 6.9: Correlation between z-score and calculated IQ for SSIM on scene 14. The visualquality differences between the images is not correctly predicted by SSIM, but the ranking is.Based on this the rank order method on the SSIM data will be a better measure.

-0.68, resulting in a low performance by iCAM.The HPminDE receives the highest score by S-CIELAB, both the SGCK and Kolas get

low scores. The results for S-CIELAB is very similar to the ones found with ΔE∗ab. This

results in a low Pearson’s correlation of -0.62.The hue angle algorithm has similar results to ΔE∗

ab and S-CIELAB, this is not surprisingdue to the familiarity between these metrics. The ranking of the GMAs are the same with onlyminor differences in the z-score values, this results in almost the same Pearson correlationbetween observer z-score and metric z-score as S-CIELAB, with -0.72.

Table 6.2: Rank indicates the Pearson correlation between rank order z-score and observerz-score. Mean Pearson indicate correlation between ranked metric score and observer z-scorecalculated as Pearson’s correlation, where the correlation for each scene has been averaged.Mean Spearman is similar to mean Pearson but for Spearman’s rank order correlation.

CorrelationMetric Rank Mean Mean

Pearson Pearson SpearmanΔE∗

ab -0.62 0.28 0.17SSIM 0.98 0.84 0.78S-CIELAB -0.62 0.28 0.20iCAM -0.68 -0.03 -0.07Hue angle -0.72 0.27 0.15

6.3.2.1 Overall observations

SSIM is the IQ metric with the best fit between the observers and the algorithm z-score (Ta-ble 6.2), indicating a good prediction of perceived IQ. All other metrics have a correlation

97

Page 114: Doctoral Dissertation by Marius Pedersen

HOW TO EVALUATE IMAGE QUALITY METRICS

below 0, indicating that these metrics do not predict perceived IQ well. In all metrics exceptSSIM the HPminDE GMA has been miscalculated, i.e. given a too high rank by the metrics.These metrics are based on ΔE∗

ab and therefore the HPminDE will be given a high rank.

6.4 Which methods will we use for the evaluation of image

quality metrics

We will use a set of the methods presented above to evaluate IQ metrics. In order to compareour results to other researchers, and to follow the most common strategies, we adapted Pear-son (Section 6.2.1.1) and Spearman (Section 6.2.1.2) correlation. Additionally, to evaluatethe overall performance of the metrics we will use the proposed method based on rank order(Section 6.3). The method used will depend on many factors, such as the metrics being eval-uated and the experimental data. These methods require evaluation against perceptual datagathered from a large user group, and in some cases other evaluation methods might be usedto obtain a thorough evaluation.

6.5 Summary

In this chapter we have presented methods for evaluating IQ metrics against human observers.Further, a new method for evaluating the overall performance of metrics was proposed basedon the rank order method. An example of using the proposed method is given, before weintroduced the methods used in this thesis.

98

Page 115: Doctoral Dissertation by Marius Pedersen

7 EVALUATION OF IMAGE QUALITY

METRICS

The IQ metrics presented in Chapter 4 take different approaches in their common goal to pre-dict perceived IQ. These metrics are created to be correlated with perceived quality. In orderto assure that the metrics are correlated with human observers they need to be evaluated. Thisevaluation is usually carried out by comparing the results of the metrics against the resultsof human observers. The better correlated the metrics are the higher their performance. Todescribe their performance, in terms of correlation with the percept, we carried out exten-sive evaluations of a set of metrics (Table 7.1). The metrics selected are from the differentclassified groups of metrics from Chapter 4, ranging from standard metrics as the PSNR andΔE∗

ab to more advanced metrics such as S-CIELAB and SSIM, but also recently publishedmetrics. They also differ in their simulation of the HVS and whether they incorporate colorinformation. To evaluate the performance of the these metrics we follow the methodology ofChapter 6. For this evaluation we need a set of test images with a range of distortions, whichhas been judged by human observers. There are a number of databases available online forevaluation of metrics (Table 7.2), which contain images with distortions and perceptual data.There is a trend that the number of images and number of observers in the databases increases,and recently experiments have also been carried out on the web [123, 367]. Web-based exper-iment and other large experiments have the advantage with many observers, but also the thedisadvantage that controlling the viewing conditions is difficult, and therefore they might notbe the best to evaluate metrics taking into account the viewing conditions. Our goal is to useIQ metrics metrics to evaluate printed images, however, to our knowledge there is no databaseavailable with printed images, therefore the evaluation in this chapter is carried out with theexisting databases.

Among the few public databases providing images for evaluation of IQ metrics, we usedthe Tampere Image Database 2008 (TID2008) [365] and the IVC image database [58]. Thesetwo databases cover the distortions found in the other databases in Table 7.2. Additionally, weselected four datasets containing respectively luminance changed images [339, 356], JPEGand JPEG2000 compressed images [63, 410], images with global variations of contrast, light-ness, and saturation [6], and gamut mapped images [107, 108]. The datasets from Pedersenet al. [339, 356] and Ajagamelle et al. [6] are related to color differences, which is an im-portant aspect for metrics to be able to evaluate. The dataset from Simone et al. [63, 410]contains small compression artifacts, unlike the TID2008 database that has larger differences,this dataset will be able to tell if the metrics are able to evaluate just noticeable differences interms of compression. The dataset from Dugay [107, 108] is based on gamut mapping wherea number of different attributes are changed simultaneously, making it a difficult task for themetrics. These databases include a wide range of distortion, from color changes to struc-

99

Page 116: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.1: Metrics evaluated in terms of correlation with the percept. Year gives the year ofthe publication where the metric was proposed, Name is the (abbreviated) name of the metric,Author is the authors name, Type indicates the type of metric as given by the authors, HVSindicates if the metric has any simulation of the HVS, and Image indicates if the metric is acolor or grayscale metric. For the grayscale metrics the images were converted to grayscaleimages using the rgb2gray function in Matlab 7.5.0.

Year Name Author Type HVS Image- MSE - Image difference No Gray- PSNR - Image difference No Gray1976 ΔE∗

ab CIE Color difference No Color1995 ΔE94 CIE Color difference No Color1996 S-CIELAB Zhang and Wandell [499] Image difference Yes Color2001 ΔE00 CIE Color difference No Color2001 SCIELABJ Johnson and Fairchild [223] Image quality Yes Color2002 UIQ Wang and Bovik [455] Image quality No Gray2002 Hue angle Hong and Luo [177] Image difference No Color2003 QCOLOR Toet and Lucassen [439] Image fidelity No Color2004 SSIM Wang et al. [458] Image quality No Gray2006 VIF Sheikh and Bovik [396] Image fidelity Yes Gray2006 PSNR-HVS-M Egiazarian et al. [113] Image quality Yes Gray2006 SSIMIPT Bonnier et al. [41] Image difference No Color2007 VSNR Chandler and Hemami [70] Image fidelity Yes Gray2009 ΔEE Oleari et al. [329] Color difference No Color2009 S-DEE Simone et al. [412] Image difference Yes Color2009 SHAME Pedersen and Hardeberg [351] Image quality Yes Color2009 SHAME-II Pedersen and Hardeberg [351] Image quality Yes Color2009 ABF Wang and Hardeberg [459] Image difference Yes Color2010 SDOG-CIELAB Ajagamelle et al. [5] Image quality Yes Color2010 SDOG-DEE Ajagamelle et al. [5] Image quality Yes Color

100

Page 117: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.2: Image quality databases available online. Observers show the total number of ob-servers on which the scores are computed.

Database Type of distortion Originalscenes

Numberof distor-tions

Total num-ber of im-ages

Observers

TID20081 [365] JPEG, JPEG2000, blur, noise,etc.

25 17 1700 838

LIVE2 [399] JPEG, JPEG2000, blur, noise,and bit error

29 5 982 161

A573 [69] Quantization, noise, JPEG,JPEG2000, blur

3 6 57 7

IVC4 [58] JPEG, JPEG2000, LAR coding,and blurring

10 4 235 15

Toyama5 [179] JPEG and JPEG200 14 2 168 16CSIQ6 [254, 255] JPEG, JPEG2000, noise, con-

trast, and blur.30 6 866 35

WIQ7 [119] Wireless artifacts 7 Varying 80 301 18/02/2011: http://www.ponomarenko.info/tid2008.htm2 18/02/2011: http://live.ece.utexas.edu/research/quality3 18/02/2011: http://foulard.ece.cornell.edu/dmc27/vsnr/vsnr.html4 18/02/2011: http://www2.irccyn.ec-nantes.fr/ivcdb/5 18/02/2011: http://mict.eng.u-toyama.ac.jp/mictdb.html6 18/02/2011: http://vision.okstate.edu/index.php?loc=csiq7 18/02/2011: http://www.bth.se/tek/rcg.nsf/pages/wiq-db

tural changes, over a large set of images, covering many different characteristics, to assure anextensive evaluation of the IQ metrics.

The performance of each metric is estimated by calculating the correlation between theperceptual quality scores from psychophysical experiments and the quality values calculatedby the metric. We opted for two standard types of correlation:

• The product-moment correlation coefficient or Pearson correlation coefficient, whichassumes a normal distribution in the uncertainty of the data values and that the variablesare ordinal.

• The Spearman rank-correlation coefficient, which is a non-parametric measure of asso-ciation, based on the ranks of the data values, which describes the relationship betweenthe variables without making any assumptions about the frequency distribution.

7.1 TID2008 Database, Ponomarenko et al.

The TID2008 database contains a total of 1700 images, with 25 reference images and 17 typesof distortions over 4 distortion levels (Figure 7.1 and Table 7.3) [365]. This database can beconsidered as an extension of the well-known LIVE database [398]. MOS are the results of654 observers attending the experiments. No viewing distance is stated in the TID database,therefore we have used a standard viewing distance (50 cm) for the metrics requiring thissetting. We have used version 1.0 of the TID2008 database, and we have taken the naturallogarithm of the results of the metrics when calculating the correlation.

101

Page 118: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.3: Overview of the distortions in the TID database and how they are related to thetested subsets. The database contains 17 types of distortions over four distortion levels. Thesign ”+” indicates that the distortion type was used to alter the images of the subset and thesign ”-” that it was not considered for this subset.

DatasetType of distortion N

oise

Noise2

Safe

Hard

Simple

Exotic

Exotic2

Full

1 Additive Gaussian noise + + + - + - - +2 Noise in color components - + - - - - - +3 Spatially correlated noise + + + + - - - +4 Masked noise - + - + - - - +5 High frequency noise + + + - - - - +6 Impulse noise + + + - - - - +7 Quantization noise + + - + - - - +8 Gaussian blur + + + + + - - +9 Image denoising + - - + - - - +10 JPEG compression - - + - + - - +11 JPEG2000 compression - - + - + - - +12 JPEG transmission errors - - - + - - + +13 JPEG2000 transmission errors - - - + - - + +14 Non eccentricity pattern noise - - - + - + + +15 Local block-wise distortion - - - - - + + +16 Mean shift - - - - - + + +17 Contrast change - - - - - + + +

102

Page 119: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

(a) Image 1 (b) Image 2 (c) Image 3 (d) Image 4 (e) Image 5 (f) Image 6 (g) Image 7

(h) Image 8 (i) Image 9 (j) Image 10 (k) Image 11 (l) Image 12 (m) Image 13 (n) Image 14

(o) Image 15 (p) Image 16 (q) Image 17 (r) Image 18 (s) Image 19 (t) Image 20 (u) Image 21

(v) Image 22 (w) Image 23 (x) Image 24 (y) Image 25

Figure 7.1: The TID2008 database contains 25 reference images with 17 types of distortionsover four levels.

The correlations for the overall performance of the IQ metrics are listed in Table 7.4. Herethe correlation between all values from the IQ metrics and the subjective scores are found.The best metric for the full database regarding the Pearson correlation is the VIF, followedby VSNR, and UIQ. These are metrics designed and optimized for the distortions found inthe TID2008 database. The SSIM is not as good as its predecessor UIQ in terms of Pearsoncorrelation, but it has a higher Spearman correlation indicating a more correct ranking. Itis worth noting that metrics based on color difference formulae do not perform well, withS-CIELAB being the best with a correlation of 0.43. The simple color difference formulaeare found on the bottom part of the list with a very low correlation for the entire database,both in Pearson and Spearman. The low correlation in many of the IQ metrics can be causedby the high number of different distortions in the TID2008 database, since there might bescale differences between the different distortions and scenes. This problem occurs when theobservers rate images with different distortions to have the same quality, but the IQ metricsrate them to be different. However, looking at the same distortion the metric might have ahigh correlation, but taking the correlation of two or more distortions might be low due to thescale differences (For explanation of scale differences see Figure 6.2).

When looking at specific distortions (Table 7.5), PSNR-HVS-M is the best for five of theseven datasets. However, in the two last datasets PSNR-HVS-M does not perform well, with acorrelation of 0.26 in both sets. The reason for this is that PSNR-HVS-M has a problem withthe mean shift distortion. The PSNR-HVS-M incorporates a mean shift removal and ratesimages where a mean shift has occurred as similar even though the observers rate them differ-ently. VSNR and VIF do not have this problem and therefore they are able to obtain a highercorrelation for both the exotic and exotic2 dataset. SSIM has a low correlation in the exoticdataset, and a much higher correlation in exotic2. The reason for this is a couple of outliersin the exotic dataset, and with the introduction of more images in exotic2 several points are

103

Page 120: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.4: Comparison of the IQ metrics over all the images of the TID2008 database. Theresults are sorted from high to low for both Pearson and Spearman correlation. VIF, UIQ,and VSNR correlate rather well with subjective evaluation with a Pearson and Spearmancorrelation above 0.6. SDOG-CIELAB is calculated with Rc = 1, Rs=2, pyramid = 1, 1

2 , 14 , 1

8 , ...,and equal weighting of the level. SDOG-DEE is calculated with Rc = 3, Rs=4, pyramid =1, 1

2 , 14 , 1

8 , ..., and the variance used as weighting of the level to give more importance to thehigh-resolution levels. For more information on SDOG-CIELAB and SDOG-DEE we refer toAppendix E. S-CIELABJ includes only the modified CSF. For other metrics standard settingshave been used, unless otherwise is stated. N = 1700. For confidence limits we refer toFigure 7.2.

Metric Pearson Metric SpearmanVIF 0.74 VIF 0.75VSNR 0.71 VSNR 0.72UIQ 0.62 SSIM 0.63PSNR-HVS-M 0.59 PSNR-HVS-M 0.61SSIM 0.55 UIQ 0.60MSE 0.54 SSIM-IPT 0.57QCOLOR 0.52 MSE 0.56PSNR 0.51 PSNR 0.56SSIM-IPT 0.48 QCOLOR 0.48S-CIELAB 0.43 S-CIELAB 0.45SHAME-II 0.41 SHAME-II 0.41SDOG-CIELAB 0.38 ΔEE 0.38S-CIELABJ 0.32 SDOG-CIELAB 0.37SHAME 0.30 SHAME 0.35S-DEE 0.29 S-CIELABJ 0.31ABF 0.28 Hue angle 0.29ΔEE 0.27 S-DEE 0.29Hue angle 0.26 ΔE∗

ab 0.28SDOG-DEE 0.26 ABF 0.26ΔE∗

ab 0.23 SDOG-DEE 0.26ΔE∗

94 -0.06 ΔE∗00 0.24

ΔE∗00 -0.06 ΔE∗

94 0.24

104

Page 121: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

placed between the outliers and the rest increasing the correlation. Some of the problems inthe exotic dataset for SSIM stem from the contrast change distortion, but mainly from scaledifferences between the distortions (i.e. one distortion has an overall smaller difference thanthe other but they are judge similarly by the observers). Within a specific distortion a fairlyhigh correlation is found. Amongst the color difference based metrics S-CIELAB is the moststable metric, with a high correlation in the noise, noise2, hard, and simple dataset. SinceS-CIELAB incorporates spatial filtering it will be able to modulate the frequencies that areless perceptible, and therefore S-CIELAB shows good results for the datasets with noise. S-CIELAB has problems in the exotic and exotic2 because of scale differences, as many of theother metrics. The 95% CI of the correlation coefficients for all r’s are illustrated in Figure 7.2,with a 95% CI it is a 95% probability that the true correlation coefficient will be within theupper and lower limit.

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

Pe

ars

on

's c

orr

ela

�o

n c

oe

ffici

en

t r

±C

I

Pearson's correla�on coefficient r

Correla�on

Lower limit

Upper limit

Figure 7.2: Pearson correlation coefficients r plotted against the Pearson correlation coeffi-cient r ± 95% CI. N = 1700. Values calculated using the method described in Section 6.2.1.1.The blue line indicate the Pearson correlation coefficient, red the lower limit (with a 95% CI),and green the upper limit (95% CI). The CI is rather small compared to the CI in Figure 6.1,this is due to a high N.

105

Page 122: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Tabl

e7.

5:C

orre

lati

onfo

rth

eIQ

met

rics

for

the

seve

ndi

ffere

ntda

tase

tsin

the

TID

2008

data

base

.E

ach

data

set

isso

rted

desc

endi

ngfr

omth

ehi

ghes

tPea

rson

corr

elat

ion

toth

elo

wes

tPea

rson

corr

elat

ion.

N=

700,

800,

700,

800,

400,

400,

600

for

the

nois

e,no

ise2

,saf

e,ha

rd,s

impl

e,ex

otic

and

exot

ic2

data

set,

resp

ecti

vely

.Fo

rco

nfide

nce

lim

its

we

refe

rto

Fig

ure

6.1.

Dat

aset

Noi

seN

oise

2Sa

feH

ard

Sim

ple

Exo

ticE

xotic

2PS

NR

-HV

S-M

0.92

PSN

R-H

VS-

M0.

87PS

NR

-HV

S-M

0.92

PSN

R-H

VS-

M0.

81PS

NR

-HV

S-M

0.94

VSN

R0.

57V

IF0.

66V

SNR

0.83

VIF

0.85

VSN

R0.

84V

IF0.

79V

SNR

0.89

VIF

0.43

VSN

R0.

59V

IF0.

78V

SNR

0.82

VIF

0.82

S-C

IEL

AB

0.78

VIF

0.86

PSN

R-H

VS-

M0.

26SS

IM0.

57S-

CIE

LA

B0.

77S-

CIE

LA

B0.

75S-

CIE

LA

B0.

76V

SNR

0.75

MSE

0.83

SDO

GC

IEL

AB

0.23

UIQ

0.56

MSE

0.74

MSE

0.73

MSE

0.74

SSIM

-IPT

0.74

PSN

R0.

81M

SE0.

19Q

CO

LO

R0.

50PS

NR

0.72

PSN

R0.

72PS

NR

0.71

SSIM

0.73

S-C

IEL

AB

0.78

UIQ

0.16

SSIM

-IPT

0.49

SHA

ME

II0.

63SH

AM

EII

0.59

UIQ

0.65

UIQ

0.71

UIQ

0.77

SSIM

0.16

PSN

R-H

VS-

M0.

26S-

CIE

LA

BJ

0.59

UIQ

0.57

QC

OL

OR

0.60

ΔE∗ ab

0.68

QC

OL

OR

0.70

PSN

R0.

15M

SE0.

25A

BF

0.57

S-C

IEL

AB

J0.

54SH

AM

EII

0.58

Hue

Ang

le0.

67SS

IM0.

67SH

AM

E0.

12PS

NR

0.22

UIQ

0.53

SHA

ME

0.51

S-C

IEL

AB

J0.

57PS

NR

0.67

ΔE∗ ab

0.65

S-D

EE

0.04

SHA

ME

II0.

13SH

AM

E0.

51SD

OG

DE

E0.

51A

BF

0.56

MSE

0.67

AB

F0.

64SH

AM

EII

0.03

SDO

GC

IEL

AB

0.10

QC

OL

OR

0.50

SSIM

0.50

SHA

ME

0.52

SHA

ME

II0.

66H

ueA

ngle

0.62

SSIM

-IPT

0.03

S-D

EE

0.09

SDO

GD

EE

0.48

SSIM

-IPT

0.47

SDO

GC

IEL

AB

0.50

QC

OL

OR

0.65

SSIM

-IPT

0.61

QC

OL

OR

-0.0

2SH

AM

E0.

08Δ

EE

0.47

QC

OL

OR

0.46

SSIM

0.48

S-C

IEL

AB

J0.

65S-

CIE

LA

BJ

0.56

ΔE

E-0

.09

ΔE

E0.

06Δ

E∗ ab

0.46

AB

F0.

45SS

IM-I

PT0.

44A

BF

0.56

SHA

ME

II0.

55SD

OG

DE

E-0

.14

S-C

IEL

AB

0.00

SDO

GC

IEL

AB

0.45

ΔE

E0.

45Δ

E∗ ab

0.44

SDO

GD

EE

0.49

ΔE∗ 94

0.49

AB

F-0

.20

Hue

Ang

le-0

.01

Hue

Ang

le0.

45S-

DE

E0.

42SD

OG

DE

E0.

43S-

DE

E0.

49SD

OG

CIE

LA

B0.

48S-

CIE

LA

B-0

.21

SDO

GD

EE

-0.0

E∗ 94

0.44

SDO

GC

IEL

AB

0.40

Hue

Ang

le0.

42Δ

EE

0.45

ΔE∗ 00

0.48

S-C

IEL

AB

J-0

.22

S-C

IEL

AB

J-0

.03

SSIM

0.43

ΔE∗ 94

0.38

ΔE∗ 94

0.42

SDO

GC

IEL

AB

0.43

SHA

ME

0.48

Hue

Ang

le-0

.25

ΔE∗ ab

-0.0

E∗ 00

0.43

ΔE∗ 00

0.37

ΔE∗ 00

0.41

SHA

ME

0.40

ΔE

E0.

39Δ

E∗ ab

-0.3

0A

BF

-0.1

3SS

IM-I

PT0.

42H

ueA

ngle

0.36

ΔE

E0.

40Δ

E∗ 94

0.25

SDO

GD

EE

0.36

ΔE∗ 00

-0.5

E∗ 00

-0.2

1S-

DE

E0.

41Δ

E∗ ab

0.33

S-D

EE

0.39

ΔE∗ 00

0.25

S-D

EE

0.34

ΔE∗ 94

-0.5

E∗ 94

-0.2

1

106

Page 123: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

(a) (b) (c) (d)

Figure 7.3: Images in the experiment from Pedersen et al. [356]. These images were changedin lightness both globally and locally.

7.2 Luminance changed images, Pedersen et al.

The database includes four original images (Figure 7.3) reproduced with different changes inlightness. Each scene has been altered in four ways globally and four ways locally [339, 356].The globally changed images had an increase in luminance of 3 and 5 ΔE∗

ab, and a decrease inluminance of 3 and 5 ΔE∗

ab. The local regions were the front and background in two images,a horizontal line in one, and two smaller regions in the last. These regions had changes inluminance of 3 ΔE∗

ab. 25 observers participated in the pair comparison experiment, whichwas carried out on a calibrated CRT monitor, LaCIE electron 22 blue II, in a gray room. Theviewing distance was set to 80 cm, and the ambient light measured to 17 lux. The images werejudged in a pair comparison setup, with the original in the middle and the two reproductionson each side.

For this database, a group of IQ metrics correlate reasonably well with subjective assess-ment (Table 7.6). The original hue angle algorithm, SHAME, SHAME-II, S-CIELAB, andABF exhibit the best Pearson correlation coefficients, all with a Pearson correlation above0.8. It is also interesting to notice that all of these are based on color differences, using theΔE∗

ab as a color difference formula. Because the changes in this database have been carriedout according to ΔE∗

ab it is not surprising that these metrics perform well. The metrics witha low correlation usually have a problem with the images where a small local change fromthe original is highly perceivable. It is also interesting to notice that the IQ metrics based onstructural similarity, such as SSIM and UIQ, are outperformed by many other metrics. Simplemetrics, such as MSE and PSNR, also perform quite well.

7.3 JPEG and JPEG2000 compressed images, Simone et al.

The 10 original images (Figure 7.4) of this database were corrupted by JPEG and JPEG2000distortions [63, 410]. The images in the database was been selected according to the distor-tions that are being evaluated following the recommendations of Field [143] and CIE [80].Four different levels of compression were used for JPEG and JPEG2000, resulting in eightcompressed images for each scene giving a total number of 80 test images. The compressionrate in Bit Per Pixel (BPP) was calculated as:

BPP = (file size in bytes×8/image size in pixels). (7.1)

107

Page 124: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.6: Correlation between subjective scores and metric scores for the luminance changedimages from Pedersen et al. [356]. We can see that the metrics based on color differenceperform well, compared to the metrics based on structural similarity. The results are sortedfrom high to low for both Pearson and Spearman correlation. N = 32, which gives a CI forinstance for SHAME-II for Pearson 0.82 equal to [0.66,0.91]. For confidence limits we referto Figure 6.1.

Metric Pearson Metric SpearmanSHAME-II 0.82 ABF 0.78SHAME 0.81 S-CIELAB 0.77Hue angle 0.81 S-CIELABJ 0.76ABF 0.81 ΔE∗

ab 0.74S-CIELAB 0.80 Hue angle 0.73S-CIELABJ 0.78 SHAME-II 0.73ΔE∗

ab 0.76 MSE 0.72MSE 0.67 PSNR 0.72PSNR 0.66 SHAME 0.71ΔE∗

94 0.65 PSNR-HVS-M 0.68ΔEE 0.63 ΔE∗

94 0.68PSNR-HVS-M 0.63 ΔE∗

00 0.68ΔE∗

00 0.63 ΔEE 0.64UIQ 0.45 SSIM-IPT 0.51S-DEE 0.39 S-DEE 0.51VIF 0.39 UIQ 0.49SSIM-IPT 0.30 SSIM 0.49QCOLOR 0.28 QCOLOR 0.48SDOG-DEE 0.24 VIF 0.35SSIM 0.22 SDOG-DEE 0.32SDOG-CIELAB 0.20 SDOG-CIELAB 0.32VSNR 0.02 VSNR 0.02

108

Page 125: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

For the experiment the compression rates were chosen such that JPEG and JPEG2000 havesimilar values of the BPP. Table 7.7 shows the BPP used to compress each scene.

The images were used in a category judgment experiment, where the original was pre-sented on one side and the compressed on the other side. Each pair of images (original andcompressed) was displayed on an Eizo ColorEdge CG241W digital LCD display. The mon-itor was calibrated and profiled using GretagMacbeth Eye-One Match 3. The settings on themonitor were sRGB with 40% of brightness and a resolution of 1600×1200 pixels. The whitepoint was set to the D65 white point and the gamma was set to a value of 2.2. The display wasplaced at a viewing distance of 70 cm. A total of 18 observers participated in the experiment.

(a) Balls (b) Barbara (c) Cafe (d) Flower (e) House

(f) Mandrian (g) Parrots (h) Picnic (i) Poster (j) Sails

Figure 7.4: The 10 images in the dataset from Simone et al. [63, 410]. These images havehave been compressed with JPEG and JPEG2000.

Table 7.7: Selected bit per pixels for each image for the dataset from Simone et al. [63, 410].

Image scene BPP JPEG BPP JPEG2000Image 1: Balls 0.8783 0.9331 1.0074 1.1237Image 2: Barbara 0.9614 1.0482 1.1003 1.2569Image 3: Cafe 0.8293 0.9080 0.9869 1.0977Image 4: Flower 0.6611 0.7226 0.7844 0.8726Image 5: House 0.9817 1.0500 1.1445 1.2920Image 6: Mandrian 1.4482 1.5910 1.7057 1.9366Image 7: Parrots 0.6230 0.6891 0.7372 0.8169Image 8: Picnic 1.0353 1.1338 1.2351 1.3765Image 9: Poster 0.4148 0.4474 0.4875 0.5420Image 10: Sails 0.7518 0.8198 0.8774 0.9927

We found that all the metrics have low performance for the images from this database(Table 7.8), spanning a range in correlation below 0.4. This is probably because these partic-ular images were initially selected in order to determine the Just Noticeable Distortion (JND).Only small distortions were applied to the original images making it arduous for the observersto assess IQ, and therefore also very difficult for the IQ metrics. PSNR-HVS-M and PSNRhave the best Pearson correlation, but they do not perform well with a Pearson correlation of

109

Page 126: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.8: Correlation between subjective scores and metric scores for the JPEG andJPEG2000 compressed images from Simone et al. [63, 410]. The results are sorted fromhigh to low for both Pearson and Spearman correlation. All IQ metrics have a low correla-tion, both for Pearson and Spearman correlation. However, it is interesting to notice the morecorrect ranking by the color difference based IQ metrics, such as S-CIELAB and SHAME.N = 80 for this database. For confidence limits we refer to Figure 6.1.

Metric Pearson Metric SpearmanPSNR 0.31 SSIM-IPT 0.37PSNR-HVS-M 0.31 SHAME 0.36SSIM 0.29 PSNR-HVS-M 0.36MSE 0.26 S-CIELABJ 0.36SSIM-IPT 0.25 SSIM 0.35SDOG-CIELAB 0.23 S-CIELAB 0.33S-CIELAB 0.22 ABF 0.33VIF 0.21 PSNR 0.32S-CIELABJ 0.17 MSE 0.30ABF 0.16 ΔE∗

ab 0.28ΔE∗

ab 0.16 VIF 0.24QCOLOR 0.15 ΔEE 0.24SHAME 0.15 SHAME-II 0.24ΔE∗

00 0.13 Hue angle 0.23S-DEE 0.13 QCOLOR 0.21ΔEE 0.13 ΔE∗

00 0.20ΔE∗

94 0.13 S-DEE 0.19SDOG-DEE 0.11 ΔE∗

94 0.19SHAME-II 0.09 SDOG-CIELAB 0.18VSNR 0.08 SDOG-DEE 0.14Hue angle 0.07 UIQ 0.11UIQ 0.05 VSNR 0.11

only 0.31. It is interesting to notice the more correct ranking by the color difference based IQmetrics, such as S-CIELAB and SHAME, but still they are not performing very well.

This dataset is similar to the simple dataset in the TID2008 dataset, that contains JPEGand JPEG2000 compressed images in addition to gaussian blur and additive gaussian noise(Table 7.3). However, there is a large difference in the performance of the metrics, for exam-ple PSNR-HVS-M had a correlation of 0.94 for the simple dataset and only a correlation 0.31for the images from Simone et al. [63, 410]. The reason for this difference is the visual differ-ences between the images in the dataset, in the TID2008 database the differences between thedifferent levels of a distortion are high while in the images from Simone et al. [63, 410] theyare just noticeable. This results in a difficult task for the observers and of course very difficultfor an IQ metric.

110

Page 127: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

(a) (b) (c) (d) (e)

(f) (g) (h) (i) (j)

Figure 7.5: The images in the IVC database from Callet and Autrusseau [58]. These imageshave been compressed with three different compression techniques.

7.4 IVC database, Le Callet et al.

The IVC database contains blurred images and images distorted by three types of lossy com-pression techniques - JPEG, JPEG2000, and Locally Adaptive Resolution (LAR) [58]. A totalof 10 different images are found in this database (Figure 7.5). The viewing distance was setto 87 cm. Only the color images in the database have been computed.

The most accurate IQ metrics are UIQ, VSNR, and VIF (Table 7.9). It is interesting tonote that these metrics are based on different approaches; the VIF is based on statistics, UIQon structural similarity, and VSNR on contrast thresholds. Many of the gray scale metricshave a higher performance than the color based metrics, this could be explained by the factthat the changes in this database are not directly related to color. The IVC database is verysimilar to the simple dataset from the TID2008 database, since both contain blurring andcompression. We can also notice that the IQ metrics with a high correlation for the simpledataset in TID2008 also have a high correlation for the IVC database, reflecting the similaritiesbetween the two datasets.

7.5 Contrast, lightness, and saturation alterations, Ajaga-

melle et al.

This database contains a total of 10 original images (Figure 7.6) covering a wide range ofcharacteristics and scenes [6]. The images were modified on a global scale with separate andsimultaneous variations of contrast, lightness, and saturation. The experiment was carried outas a category judgment experiment with 14 observers. The viewing distance was set to 70 cm,and the ambient illumination was 40 lux.

PSNR-HVS-M has the best Pearson correlation, while SHAME-II has the best Spearmancorrelation (Table 7.10). We can notice that the differences between the metrics are low, forthe Spearman correlation 13 metrics have a correlation between 0.71 and 0.79. The color dif-

111

Page 128: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.9: Correlation between subjective scores and metric scores for the images from theIVC database. The results are sorted from high to low for both Pearson and Spearman corre-lation. N = 185 for the IVC database. For confidence limits we refer to Figure 6.1.

Metric Pearson Metric SpearmanVIF 0.88 VIF 0.90UIQ 0.82 UIQ 0.83VSNR 0.77 VSNR 0.78PSNR-HVS-M 0.73 SSIM 0.78SSIM-IPT 0.71 SSIM-IPT 0.77SSIM 0.70 PSNR-HVS-M 0.77PSNR 0.67 MSE 0.69S-CIELAB 0.61 PSNR 0.69QCOLOR 0.61 SHAME 0.68ΔEE 0.56 ΔE∗

ab 0.66SHAME 0.55 QCOLOR 0.66SDOG-DEE 0.55 S-CIELAB 0.66ΔE∗

ab 0.54 ΔEE 0.62ABF 0.51 SDOG-DEE 0.58MSE 0.51 ABF 0.56S-DEE 0.46 Hue angle 0.52S-CIELABJ 0.42 ΔE∗

00 0.47ΔE∗

00 0.38 ΔE∗94 0.46

ΔE∗94 0.37 S-DEE 0.46

SHAME-II 0.37 S-CIELABJ 0.43Hue angle 0.35 SHAME-II 0.42SDOG-CIELAB 0.29 SDOG-CIELAB 0.32

112

Page 129: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

(a) (b) (c) (d) (e)

(f) (g) (h) (i) (j)

Figure 7.6: Images altered in contrast, lightness, and saturation from Ajagamelle et al. [6].

ference formulae from CIE (ΔE∗ab, ΔE94, and ΔE00) all perform very well, the same also with

PSNR. This is different from many of the other databases, where these simple pixelwise met-rics do not correlate with perceived IQ. The high correlation over some of the other databasesis probably due to the global variations made by Ajagamelle et al., since the pixelwise IQmetrics weight all pixels equally. The results from this database are similar to the results fromthe lightness changed images from Pedersen et al. [356].

7.6 Gamut mapped images, Dugay et al.

In this dataset, 20 original images (Figure 6.4) were gamut mapped with five different algo-rithms [107, 108]: HPminDE, SGCK, Zolliker, Kolas, and Gatta. For more information onthe gamut mapping algorithms we refer the reader to Section 6.3.1.1.

The 20 different images were evaluated by 20 observers in a pair comparison experiment.The monitor used was a Dell 2407WFP LCD display calibrated with a D65 white point anda gamma of 2.2. The level of ambient illumination on the monitor was around 20 lux. Theviewing distance for the observers was approximately 50 cm.

We see from the results in Table 7.11 that most of the metrics does not correlate well withperceived quality. In gamut mapping many different attributes are changed simultaneously,and because of this the objective assessment is very complex. Previous research has alsoshown that IQ metrics have problems when multiple distortions occur simultaneously, as ingamut mapping [41, 169]. This is not the case for TID2008 and some of the other databasesevaluated here, since usually only one attribute changes at the time.

7.7 Overall performance of the metrics

Figure 7.7 shows the performance for all IQ metrics for the six different databases. We cansee that none of the 22 evaluated metrics have a high correlation for all databases. If wediscard the JPEG and JPEG2000 compressed images from Carraciolo et al. and the gamutmapped images from Dugay et al. some metrics perform rather well, such as the PSNR-HVS-M and PSNR, which have a correlation of above 0.5 for the remaining databases. However,the performance seems to be database or distortion dependent, and the metrics perform better

113

Page 130: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.10: Correlation between subjective scores and metric scores for the images fromAjagamelle et al. [6]. The results are sorted from high to low for both Pearson and Spearmancorrelation. PSNR-HVS-M has the highest Pearson correlation, while SHAME-II has the bestSpearman correlation. VSNR has the lowest correlation, which is caused by scale differencesbetween the difference images. N = 80. For confidence limits we refer to Figure 6.1.

Metric Pearson Metric SpearmanPSNR-HVS-M 0.79 SHAME-II 0.79ΔE∗

ab 0.75 SSIM-IPT 0.76ΔE∗

94 0.74 ABF 0.76ΔE∗

00 0.74 PSNR-HVS-M 0.75ABF 0.73 QCOLOR 0.75PSNR 0.72 UIQ 0.73S-CIELAB 0.67 SHAME 0.73SSIM-IPT 0.66 S-CIELAB 0.72SSIM 0.64 Hue Angle 0.72UIQ 0.63 ΔE∗

ab 0.72Hue Angle 0.62 SSIM 0.72SHAME-II 0.62 MSE 0.71QCOLOR 0.62 PSNR 0.71MSE 0.61 ΔE∗

00 0.69DEE 0.60 ΔE∗

94 0.69SDOG-DEE 0.59 DEE 0.68S-CIELABJ 0.58 S-CIELABJ 0.65VIF 0.53 SDOG-DEE 0.61SHAME 0.50 VIF 0.59SDOG-CIELAB 0.41 SDOG-CIELAB 0.48S-DEE 0.40 S-DEE 0.41VSNR 0.16 VSNR 0.14

114

Page 131: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Table 7.11: Correlation between subjective scores and metric scores for the gamut mappedimages from Dugay et al. [107, 108]. The results are sorted from high to low for both Pearsonand Spearman correlation. N = 100. For confidence limits we refer to Figure 6.1.

Metric Pearson Metric SpearmanVIF 0.31 VIF 0.27UIQ 0.31 UIQ 0.19QCOLOR 0.19 QCOLOR 0.12SSIM 0.16 VSNR 0.09MSE 0.09 SSIM 0.05VSNR 0.09 S-CIELABJ 0.03PSNR-HVS-M 0.07 PSNR-HVS-M 0.03SDOG-CIELAB 0.05 ΔE∗

00 0.00PSNR 0.04 ΔE∗

94 0.00ΔE∗

00 0.03 SSIM-IPT -0.02SDOG-DEE 0.02 MSE -0.02S-CIELABJ 0.02 SDOG-DEE -0.02ΔE∗

94 0.02 PSNR -0.03SSIM-IPT 0.01 ΔEE -0.06Hue angle 0.00 ΔE∗

ab -0.08ΔEE -0.01 SDOG-CIELAB -0.08ΔE∗

ab -0.03 Hue angle -0.08S-DEE -0.03 S-DEE -0.09SHAME -0.03 SHAME -0.09S-CIELAB -0.06 S-CIELAB -0.11ABF -0.08 ABF -0.11SHAME-II -0.10 SHAME-II -0.13

115

Page 132: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

for specific distortions, such as VIF for the IVC database with compression, and ABF andSHAME-II for the luminance changed images from Pedersen et al.

7.8 Summary

In this chapter we have evaluated IQ metrics from the survey in Section 4.2. For the evaluationdifferent state of the art databases have been used. However, no database with printed imagesis available, resulting in that evaluation of metrics for printed images require new databases.The results of the evaluation of the different databases indicate that the observed performanceof the different metrics can depend on both the type of distortion and on the image databasethat is used.

116

Page 133: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

Fig

ure

7.7:

Perf

orm

ance

ofth

eIQ

met

rics

acro

ssth

eda

taba

ses.

The

figur

esh

ows

the

Pear

son

corr

elat

ion

coef

ficie

nts

for

six

diffe

rent

sets

ofte

stim

ages

and

22IQ

met

rics

.

117

Page 134: Doctoral Dissertation by Marius Pedersen

EVALUATION OF IMAGE QUALITY METRICS

118

Page 135: Doctoral Dissertation by Marius Pedersen

PART III

IMAGE QUALITY METRICS FOR THE

EVALUATION OF PRINTING WORKFLOWS

119

Page 136: Doctoral Dissertation by Marius Pedersen
Page 137: Doctoral Dissertation by Marius Pedersen

8 A FRAMEWORK FOR USING METRICS

WITH COLOR PRINTS

Subjective assessment of print quality can be carried out by asking a group of observers aboutthe quality of the printed image. However, assessment of a printed image by IQ metrics ismore complicated. The original image is of a digital format and the printed image is of ananalog format, and because of this the printed image must be digitized before IQ assessmentcan be carried out. In this chapter we discuss the transformation from a physical reproductionto a digital reproduction with the goal of proposing a framework for using IQ metrics toevaluate the quality of color prints.

8.1 State of the art of frameworks to digitize printed images

A couple of frameworks have been proposed for using IQ metrics on printed images. Allthese frameworks follow the same procedure. Firstly, the printed image is scanned. Thenimage registration is performed to match the scanned image with the original. This is usuallyfollowed by a descreening procedure to remove halftoning patterns. Finally, an IQ metric canbe applied.

Grice and Allebach [163] proposed a framework for using the metrics specified in ISO/IEC13660 [213]. The framework is based on a test target that is printed and scanned before a set ofmetrics are applied. Different scanning resolutions are investigated, and the authors concludedthat a scanning resolution of 400 dpi might be too low for some metrics. The metrics appliedwere computational metrics (i.e. not design to account for the HVS) defined by ISO [213],such as metrics for contrast, mottle, darkness, and raggedness.

A framework was proposed by Zhang et al. [498] in 1997. First, the image is scanned,and then three additional scans are performed, each with a different color filter. This resultsin enough information to transform the image correctly to CIEXYZ. The printed image wasscanned with an Agfa Horizon flatbed scanner, and the scanning resolution was set to 1200dpi. In this article no information about the image registration was given, or on the descreeningprocedure. The applied IQ metric was S-CIELAB [499], and the images were color patches.

Another framework was proposed by Yanfang et al. [483] in 2008. Before printing twocontrol points are applied to the image, one to the upper left corner and one to the upper center,to help in the registration. The images were scanned with a Kodak i700 at 300 DPI beforeregistration, where the two control points are used for matching the printed image with theoriginal (Figure 8.1). Descreening was performed by the scanner at 230 LPI. No informationis given regarding the scaling of the image. The applied IQ metric was S-CIELAB [499].

121

Page 138: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

Original

and

reproduction

RGB

images

LAB

images

XYZ

images

S-CIELAB

images

S-CIELAB

difference

Scanning Registration Converting Calculating

Calculating

Figure 8.1: Framework by Yanfang et al. [483]

Recently, Eerola et al. [110, 111] proposed a new framework (Figure 8.2), which followsthe same steps as the previous frameworks. The printed reproduction is scanned, and then boththe original and the reproduction go through a descreening procedure, which is performed us-ing a Gaussian low-pass filter. Further, image registration is carried out, where local pointsare used with a Scale-Invariant Feature Transform (SIFT). A RANd SAmple Consensus prin-ciple (RANSAC) was used to find the best homography. This is different from the previousframeworks, since it uses local points instead of control points. Scaling was performed us-ing bicubic interpolation. Scanner resolution was set to 1250 DPI, and in Eerola et al. [110]LABMSE was the applied IQ metric. Several other metrics, as SSIM [458], Noise QualityMeasure (NQM) [99], PSNR, and UIQ [455], were used in Eerola et al. [111].

Figure 8.2: Framework from Eerola et al. [110, 111], where local points are used rather thancontrol points. Figure reproduced from Eerola et al. [110, 111].

8.2 A framework based on control points

We modify and propose a framework similar to the framework by Yanfang et al. [483], whichperforms image registration based on control points. These control points act as the basis forthe image registration, where the control points are used to perform different transformationprocedures. The advantage of using control points is that uniform images can be registered,opposed to frameworks based on local points.

As a first step in our framework, the image is padded with a white border and equippedwith four control points prior to printing. These points are black squares placed just outsidethe four corners of the image. Then the image is printed with a given device, such as an inkjet printer. Further, the image is scanned with a scanner. The ICC profile of the scanner is

122

Page 139: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

then assigned to the scanned image in order to achieve a correct description of the colors. Thenext step in the framework is to find the coordinates of the center of the control points, in boththe original image and the scanned image. This is done by a simple automatic routine basedon detection of squares. The scanned image can be affected by several geometric distortions,such as translation, scaling and rotation. Because of this, image registration must be carriedout. The coordinates for the control points, in both images, are used to create a transformationfor the registration. There are several possible transformation types for doing the registration,which are discussed below. In addition the interpolation method for scaling also has severalpossible methods, which are compared below. After the scanned image has been registered tomatch the original, a simple procedure is applied to remove the white padding and the controlpoints. Finally, an IQ metric can be used to calculate the quality of the printed image. Anoverview of the framework is shown in Figure 8.3. This framework differs from the one fromEerola et al. [110], not only in the registration method, but also in the descreening. In ourmodified framework we do not perform any specific descreening, but we leave this to the IQmetrics in order to avoid a double filtering of the image. It is therefore recommended that theIQ metrics perform some kind of simulation of the HVS, for example spatial filtering basedon contrast sensitivity.

Pad theimagewith

controlpoints

Printpaddedimage

Scanprintedimage

ApplyICC

profile

Imageregi-

stration

Removepadding

andcontrolpoints

Imagequalitymetrics

Figure 8.3: Overview of the proposed framework for using image quality metrics with printedimages.

8.2.1 Transformation type

The transformation type to be used with the framework must use the same or fewer numberof control points added to the image. In order to find the most appropriate transformation, wehave tested five different transformation types:

• Nonreflective similarity uses two pairs of control points, and is suitable when the imageis distorted by some combination of translation, rotation, and scaling.

[ u v ] = [x y 1]T, (8.1)

where x and y are the dimensions to be transformed, T is a 3-by-2 matrix that dependson 4 parameters, and u and v are the transformed dimensions.

• Similarity resembles the previous, but uses reflection in addition.

• Affine uses three pairs of control points, and is useful when the image exhibit shearing.

[ u v ] = [x y 1]T, (8.2)

123

Page 140: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

where T is a 3-by-2 matrix where all six elements can be different.

• Projective uses four pair of control points, and is commonly used when the imagesappears tilted.

[upwp

vpwp w ] = [x y w]T, (8.3)

where T is a 3-by-3, where all nine elements can be different.

• Piecewise linear uses four pair of control points, and is used when parts of the imageappear distorted differently. This transformation method applies affine transformationsseparately to triangular regions of the image.

Theoretically, translation, rotation, and scaling are needed to register a scanned image.However, we wanted to test other transformation methods as well, such as to account forshearing and local distortions. In order to find the most appropriate transformation type severaltests have been performed, which are discussed below.

8.2.2 Interpolation method

The interpolation method used for scaling the images are important, there are three mainmethods for doing this:

• bicubic interpolation,

• bilinear interpolation,

• nearest-neighbor interpolation.

The first method considers a 4×4 neighborhood, and accounts for the distance to the unknownpixel in the calculation. The second method considers a 2×2 neighborhood, and it takes theweighted average of these four pixels to arrive at its final interpolated value. The two firstmethods also incorporate an anti-aliasing filter to suppress moire patterns [451], which is alowpass filter. The latter method simply selects the value of the nearest point, and does notconsider the values of other neighboring points. To find the most appropriate interpolationmethod, i.e. the method introducing the least error, several tests have been carried out.

8.2.3 Evaluation of different transformation types

To ensure the best possible registration we evaluate the different transformation types (fivedifferent types as mentioned above) together with the different interpolation types1 (threedifferent types), which results in totally 15 combinations. These combinations are applied tothree different images shown in Figure 8.4. Before applying these combinations, the image,which is going to be registered, has been rotated and resized so that all the most relevantproblems during the scanning procedure are faced. Finally, the results have been comparedwith the original image.

The test images are selected to have a good combination of what would be expected forprinted images, containing both uniform areas and high frequency areas. The first image (Fig-ure 8.4(a)) is a typical scenery, the second image (Figure 8.4(b)) is an image with a fairly

1Standard parameters in Matlab have been used for the interpolation methods.

124

Page 141: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

(a) Image 1 (b) Image 2 (c) Image 3

Figure 8.4: Images used for evaluating different transformation types and interpolation meth-ods. The images differ in detail content, from having large uniform areas to high frequencycontent. This will reveal differences between the methods using local points and controlpoints.

uniform background, and the third image (Figure 8.4(c)) is an image with fine details. In-sisting on the uniformity and nonuniformity of the background is because edges have a greatinfluence on the errors when comparing the original and the registered image.

Figures 8.5 and 8.6 show the MSE between the original image and the registered imageafter applying the combinations of different transformation and interpolation for Figure 8.4(a)and Figure 8.4(c). Results for Figure 8.4(b) are similar to the results for Figure 8.4(a), and arenot shown. It should be kept in mind that the MSE values in the figures are normalized withthe highest value for the combinations.

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Nonereflective bicubic

Nonereflective bilinear

Nonereflective nearest

Similarity bicubic

Similarity bilinear

Similarity nearest

Affine bicubic

Affine bilinear

Affine nearest

Projective bicubic

Projective bilinear

Projective nearest

Piecewise bicubic

Piecewise bilinear

Piecewise nearest

MS

E

Figure 8.5: MSE between the original and registered image calculated for Figure 8.4(a)

As it can be seen, using a ”similarity” transformation and a ”bilinear” interpolation has thelowest MSE value. Inspection of the difference between the original image and the registeredimage show that the errors using this combination of transformation and interpolation occuron the edges. For the image in Figure 8.4(c) containing a lot of edges and details, bilinearinterpolation gives the lowest error of the interpolation methods for all five transformationtypes as seen in Figure 8.6. In addition to our results, research carried out by Acharya and

125

Page 142: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Nonereflective bicubic

Nonereflective bilinear

Nonereflective nearest

Similarity bicubic

Similarity bilinear

Similarity nearest

Affine bicubic

Affine bilinear

Affine nearest

Projective bicubic

Projective bilinear

Projective nearest

Piecewise bicubic

Piecewise bilinear

Piecewise nearest

MS

E

Figure 8.6: MSE between the original and registered image calculated for Figure 8.4(c)

Tsai [2] indicates that bilinear interpolation has the least error when reducing the size of animage, supporting our findings.

8.2.4 Scanner

There are two different scanner types available; Charge Coupled Device (CCD) and ContactImage Sensor (CIS). The disadvantage with a CIS scanner is the lack of depth-of-view [338].However, this is not a problem with prints, since the print is usually flat on the scan surface.

8.2.4.1 Scanning resolution

A central issue while scanning is to have a resolution high enough to capture the perceiveddetails of the printed image. In order to find the appropriate resolution an image was scannedwith a Microtek ScanMaker 9800XL scanner at different resolutions. The scanner was cali-brated using an IT8.7 Target and the profile was created with ProfileMaker.

An image (Figure 8.7) was prepared with control points just outside the corners, and itwas printed with an Oce ColorWave 600 wide format printer on Oce Red Label paper at aresolution of 150 pixels per inch. The printed image was then scanned with the followingresolutions: 72, 100, 150, 300, and 600 DPI without any automatic corrections. The scannedimage was used as input to the framework, where the errors were investigated using MSEand ΔE∗

ab. From the results in Table 8.1 we see that the values fluctuate, and do not give anyindication of the resolution needed. Since the printed image will be different from the original,just finding the lowest objective value is not appropriate, because of this further investigationis needed.

The next step after the objective evaluation of the scanning resolution is a subjective eval-uation. Visual inspections of the images reveal a higher detail level in the 600 DPI scans, thanat lower resolutions. However, at a normal viewing distance (60 cm) these details are not ap-parent in the printed image. The scanner resolution is also dependent on the visual angle the

126

Page 143: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

Figure 8.7: Scanned test images with control points outside the four corners.

Table 8.1: Scanning resolution test. MSE is in the 104.

Resolution (dpi) MSE ΔE∗ab

72 7.5 14.5100 6.5 12.5150 5.7 12.1300 5.9 12.9600 6.1 14.6

127

Page 144: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

prints are evaluated at, and can also be dependent on the IQ metrics Grice and Allebach [163].Metrics simulating the HVS usually perform some sort of blurring (filtering) of the imageand they might require lower resolution than computational metrics that do not simulate theHVS. The edge raggedness and edge blur metrics as defined by ISO [197] requires a scanningresolution of at least 600 DPI. Lee et al.’s [262] implementation of the edge raggedness andedge blur metrics require a resolution of 1200 DPI. Lim and Mani [267] state that a resolutionof 600 DPI should be sufficient to capture the details of a print. Zeise et al. [492] commentedthat a resolution of 400 DPI would be sufficient, but in many cases a higher resolution is needto limit aliasing artifacts from high-frequency image fluctuations. ISO/IEC 13360 standard[197] suggests a resolution of minimum 600 DPI for evaluation of binary monochrome textand graphic images. Stanek and Burns [424] scanning parameters for print analysis, and theyconcluded that a resolution higher than 800 DPI is not necessary when looking at the noisepower spectrum of different resolutions. It should also be noted that the scanning resolutioncan be dependent on the evaluation, for situations where the viewing distance is large a lowerresolution might be adequate since the image will be heavily filtered, for situations where theviewing distance is short higher resolutions might be required.

8.2.4.2 Scanner ICC profile

The ICC profile for the scanner should be created with a test target printed on the same printeras the image set, since the profile is dependent on the equipment [259, 260, 334]. A profilecreated with another setup, or a standard profile, might not match your individual device. Inthe case where several printers (with different ink) and a variety of papers, a profile should begenerated for each combination of paper and ink/printer.

8.2.4.3 Cropping of the scan

When scanning the printed images the operator will crop the image, which might result indifferences in the registered image. In order to investigate if cropping influences the registra-tion in the framework, a printed image was scanned several times with different crop regions.14 differences scans at different angles were carried out with three different crop regions. Asa measure of variation in the registration process we use the Coefficient of Variation (CV),which is defined as:

CV =σ

|μ| , (8.4)

where σ is the standard deviation, and μ the mean.For each scan the rotation angle is registered, and for the fourteen different angles we ob-

tain an average CV as low as 0.01, indicating very small variations in the registration process.

8.2.4.4 Scanner rotation

The angle at which the printed image is scanned at could also possibly influence the registra-tion. We use the same scans as for the previous section (8.2.4.3) to investigate if the rotationinfluences the registration. The printed image has been scanned 14 times at different angles,and each angle is scanned three times with a different crop, resulting in 42 different regis-tered images. This constitutes the basis for the investigation. For every registered image we

128

Page 145: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

have calculated the IQ values for SSIM [458] (Section 4.2.3.1) and S-CIELAB [499] (Sec-tion 4.2.2.1). If the rotation does not influence the final result of the metrics we would expectthat the results from the metrics are stable of the different registered images. We use CV(Equation 8.4) as a measure of the variation in the metric results. CV for SSIM is 0.00003and for S-CIELAB 0.003, both being very small, indicating that the variation between thedifferent rotations are small. We have also plotted the average results of the metrics for thethree crops against the average rotation of the three crops to reveal any relationship betweenthem. SSIM shows a correlation of -0.31 (Figure 8.8(a)) and S-CIELAB -0.01 (Figure 8.8(b)).The results do not indicate a relationship between the rotation and the metrics.

0.99630

0.99632

0.99634

0.99636

0.99638

0.99640

0.99642

0.99644

0.00 0.50 1.00

Avera

ge S

SIM

Average rotation

(a) SSIM

20.84

20.89

20.94

20.99

21.04

0.00 0.50 1.00

Ave

rag

e S

-CIE

LA

B

Average rotation

(b) S-CIELAB

Figure 8.8: Metric values plotted against the rotation angle. There is no indication of arelationship between the rotation and the results of S-CIELAB and SSIM. The blue solid lineis the linear regression line (r=-0.01 for S-CIELAB and r=-0.31 for SSIM), the red dashedlines indicate the 95% CI of the regression line, and the blue dotted curves indicate the total95% CI of the linear regression estimate [418].

129

Page 146: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

To ensure that there is no difference between different scanning angles, we have alsocomputed the CV for 25 scans with a rotation less than 0.3 degrees (average rotation 0.17degrees). The SSIM CV is 0.00005 and for S-CIELAB 0.001, both being very small andsimilar to the results for the images with a larger rotation. Based on these findings we cannotsay that the rotation will result in variations in the final metric values.

8.2.5 A comparison against another framework

The most important aspect to consider when selecting a framework for color prints is that theerrors introduced should be as small as possible. From the state of the art two different typesof frameworks can be found; those who use local interest points (features), such as the oneproposed by Eerola et al. [110], and those who use control points, such as the framework byYanfang et al. [483] and the modified framework presented above. Both have disadvantagesand advantages. For the local features no extra control points need to be added in the process,but local features will not work well for uniform or near-uniform surfaces, since no points canbe found in these areas. Using control points can be a disadvantage, since these need to beadded prior to printing, however, uniform images can be correctly registered, including colorpatches. We compare these two different types of IQ frameworks, the one proposed by Eerolaet al. [110] and our proposed framework.

We use the images in Figure 8.4 to compare the frameworks. For the proposed modifiedframework we use bilinear interpolation and similarity as transformation, for the frameworkby Eerola et al. [110] we use the same settings as in their paper. The best framework shouldhave the least difference between the original image and the registered image. The results forthe images are shown in Figure 8.9, and we can clearly see that the proposed framework basedon Control Points (CP) introduces less error than the framework based on Local Features (LF),both in terms of absolute difference and mean square error. The biggest difference is foundin the image with uniform areas (Figure 8.4(b)), which is a problem for frameworks basedon local features since no registrations points can be found. In images with a lot of detailsthe difference in error is almost insignificant, but the proposed framework performs slightlybetter. Figure 8.9(b) shows the relative error (normalized by the maximum error), and we cansee that for three different images we have the same tendency, with the CP based frameworkintroducing much less error than the LF framework.

1 2 30

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Image

Nor

mal

ized

err

or

Absolute difference CPAbsolute difference LFMSE CPMSE LF

(a) Error

1 2 30

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Image

Nor

mal

ized

err

or

Absolute difference CPAbsolute difference LFMSE CPMSE LF

(b) Relative error

Figure 8.9: Error of the proposed framework compared to the error from the framework fromEerola et al. [110].

In addition, we also checked the computational time used for the registration. The com-

130

Page 147: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

putational time for the proposed framework is stable over the three images. The other frame-work, which is based on local features, has varying computational time since the content ofthe image affects the number of registration points. Comparing the time used by the two dif-ferent frameworks, the proposed framework is more than 20 times faster than the frameworkby Eerola et al. [110]. The time differences are mainly from the number of registration points.Furthermore, the complexity of a framework based on local features is higher than that of aframework based on control points.

Based on the results shown here, we will use the proposed modified framework based oncontrol points to evaluate IQ.

8.3 Summary

In this chapter we have proposed a new framework for digitizing printed images in order toapply IQ metrics. The proposed framework is based on control points for the registration.Investigation shows that our framework outperforms a state of the art framework.

131

Page 148: Doctoral Dissertation by Marius Pedersen

A FRAMEWORK FOR USING METRICS WITH COLOR PRINTS

132

Page 149: Doctoral Dissertation by Marius Pedersen

9 EVALUATION OF METRICS USING THE

FRAMEWORK

We have used the framework explained in the previous chapter to evaluate a set of IQ metricson images from a color workflow.

9.1 Experimental setup

15 images were obtained from Cardin [64] (Figure 9.1). These were processed with two dif-ferent source profiles, the sRGB v2 perceptual transform and the sRGB v4 perceptual trans-form [185, 186]. These were further processed with four different softwares for obtaining thedestination profile:

• Basic rendering (black point compensation and hue preserving minimum ΔE∗ab).

• LOGO colorful from ProfileMaker Pro.

• LOGO Chroma plus from ProfileMaker Pro.

• Onyx from Onyx Mediacreator.

For each image this results in eight different reproductions, as seen in Figure 9.2. Theseimages were printed, with the Oce ColorWave 600 wide format CMYK printer on Oce RedLabel paper, at a resolution of 150 pixels per inch and at a size resulting in a printed repro-duction at approximately 8 by 10 inches.

The printed images were evaluated by 30 observers in a category judgment experimentwith three categories. The three categories were described as:

1. the most pleasant,

2. neither more nor less pleasant,

3. less pleasant.

The observers were presented with a reference image on an EIZO ColorEdge CG221 displayat a color temperature of 6500 Kelvin and luminance level of 80 cd/m2, following the specifi-cations of the sRGB. The image set was rendered for a sRGB display, and therefore a monitorcapable of displaying the sRGB gamut was the most adapted reproduction device for this setof images. The printed images were presented randomly in a controlled viewing room at a

133

Page 150: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

(a) 1 (b) 2 (c) 3 (d) 4 (e) 5 (f) 6 (g) 7

(h) 8 (i) 9 (j) 10 (k) 11 (l) 12 (m) 13

(n) 14 (o) 15

Figure 9.1: Images in the experiment. They cover a wide range of characteristics to ensurequality differences in many different quality attributes.

sRGB version 2 sRGB version 4

ProfileMaker Pro

LOGO Colorful

ProfileMaker Pro

LOGO Chroma +Onyx

Basic Rendering

BPC+HPminDE

Figure 9.2: Overview of the workflows used in the experiment. The image is either processedusing the sRGB v2 or sRGB v4. Furthermore, each of these is used as input to one of thefour rendering methods. This finally results in eight different printed reproduction of the sameimage. Figure inspired by Cardin [64].

134

Page 151: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

color temperature of 5200 Kelvins, an illuminance level of 450 ±75 lux and a color renderingindex of 96. The observers viewed the reference image and the printed image simultaneouslyfrom a distance of approximately 60 cm. The experiment was set up to follow the CIE guide-lines [80] as closely as possible. More details on the experiment can be found in Cardin [64]or Bonnier et al. [39].

The 15 images printed with the eight different workflows were scanned with a MicrotekScanMaker 9800XL at 600 DPI without any automatic corrections. The scanner was charac-terized using a Kodak IT8.7 target, since a printed test target was not available, and the profilewas built using ProfileMaker 5.0.8. Analysis of the scans showed that the scanner exhib-ited a slight shearing, which occurs due to mechanical issues. Because of this a transforma-tion method which takes into account shearing must be applied. Even though these methods(affine) introduce more error than the methods that do not correct for shearing (similarity), asseen in Figures 8.5 and 8.6.

9.2 Psychophysical results

30 observers participated in the experiment, where they judged the images on a three stepscale. The answers from the observers where processed with the Colour Engeneering Tool-box [162] to obtain z-scores. The results are shown in Figure 9.3. Onyx V4 has the highestz-score, but cannot be differentiated from Onyx V2. It is worth noticing the small differencebetween the highest and lowest z-score. This indicates a low visual difference between thedifferent workflows, and that the task was difficult for the observers. We will focus on theobjective evaluation of IQ, but an in depth analysis of the subjective evaluation can be foundin Cardin [64].

AB V2 AB V4 Onyx V2 Onyx V4 Cful V2 Cful V4 CPlus V2 CPlus V4−0.5

−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

0.5

Workflow

Z−

scor

es

Z−scores based on 30 observers for 15 images

Figure 9.3: Z-scores based on 30 observers for 15 images. Solid horizontal lines indicatecategories. All images are within the middle category, being ”neither more nor less pleasant”.

135

Page 152: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

9.3 Selection of image quality metrics

A set of IQ metrics were applied to evaluate the IQ of the printed images. Since no directdescreening is applied in the registration process IQ metrics not incorporating aspects of theHVS are most likely inappropriate. A set of metrics has been selected, and a brief introductionof them is given. For further information we refer the reader to their respective references orto Chapter 4.

• S-CIELAB [499] is often used as a reference metric, and has wide acceptance.

• S-CIELABJ [223] is a modified version of the original S-CIELAB, in terms of thespatial filtering.

• S-DEE [412] extends the S-CIELABJ with a more refined color difference calculation.

• SHAME [351] combines two state of the art metrics (the hue angle measure [177] andS-CIELABJ). We have included both SHAME-I and SHAME-II, which differ in termsof the spatial filtering. We refer the reader to Chapter 5 for more information.

• Adaptive Bilateral Filter (ABF) [459] uses bilateral filtering to simulate the HVS, beforethe color difference is calculated using ΔE∗

ab.

Further, IQ metrics taking into account structural information might be suitable, since theseare potentially good at detecting artifacts.

• SSIM [458], since this is commonly used and has received great publicity since it wasproposed in 2004. This metric works on a local neighborhood, and is therefore consid-ered as appropriate for our use.

• Cao et al. [61] proposed a metric designed to detect artifacts, which is based on thedifference of saliency. Saliency usually refers to a representation of the idea that certainparts of a scene are pre-attentively distinctive and create some form of significant visualarousal [227]. This metric should be suitable since the dataset from Cardin [64] isreported to have different artifacts, such as loss of details and contouring [39]. Thepercentage of pixels with artifacts has been used as a measure of quality for this metric.For more information on this metric we refer the reader to Appendix C.

Evaluation of performance is done by calculating the correlation coefficient between thesubjective score and the objective score, as introduced in Section 6.2.1. Three different kindof correlation are computed; the Pearson product-moment correlation coefficient, the Spear-man’s rank correlation coefficient, and the Kendall tau rank correlation coefficient [232]. Thefirst assumes that the variables are ordinal, and finds the linear relationship between variables.The second, Spearman, is a non-parametric measure of correlation that uses the ranks as basisinstead of the actual values. It describes the relationship between variables without makingany assumptions about the frequency distribution of the variables. The third, Kendall, is anon-parametric test used to measure the degree of correspondence between two rankings, andassessing the significance of this.

Two types of evaluation are carried out, first the overall performance of the metrics overthe entire image set, and then evaluation of the metrics for each of the 15 different images.

136

Page 153: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

9.3.1 Overall evaluation

We compare the overall score from the observers to the overall score by the metrics in orderto evaluate the overall performance of the metrics. The most common method for this is bycomputing the Pearson correlation for all scores. The results from this show that all metricshave a very low Pearson correlation, approximately around zero. This indicates that the IQmetrics cannot predict perceived IQ. In addition, both the Spearman and Kendall correlationcoefficients are similar to the Pearson correlation.

The low correlation is because of scale differences between the images, and because ofthis problem a new method [350] for evaluating overall performance of metrics was proposedin Section 6.3. In this method the scores for the metric are used to simulate an observer usingthe rank-order method, and results in a z-score plot as from the experiment (Figure 9.3). Theresults from the metric should be similar to the results from the observers if the metric exhibita good performance.

Figure 9.4 shows the results for the S-CIELAB IQ metric with this method. A visualinspection of the relation between the results from the metric and the observers show differ-ences, which is evidence that the metric does not predict perceived overall IQ. The z-scoresfrom the observers are plotted against the S-CIELAB z-scores in Figure 9.5. The Pearsoncorrelation between these are only 0.34, indicating a low correlation. The others also have alow correlation, as seen in Table 9.1.

AB V2 AB V4 Onyx V2 Onyx V4 Cful V2 Cful V4 CPlus V2 CPlus V4−1.5

−1

−0.5

0

0.5

1

1.5

2Metric Z−score

Workflows

Z−

scor

e

Figure 9.4: S-CIELAB z-score calculated with the method proposed in Section 6.3 for thesame 15 images as in Figure 9.3.

The small visual quality differences between the images contribute to making this a dif-ficult task for the IQ metrics. For this experiment the IQ metrics cannot predict perceivedoverall IQ.

137

Page 154: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

−1.5 −1 −0.5 0 0.5 1 1.5 2−0.5

−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

0.5

Metric z−score

Obs

erve

r z−

scor

e

Figure 9.5: Z-scores from observers (Figure 9.3) plotted against the z-scores from S-CIELAB(Figure 9.4). This gives a correlation of 0.34. Corresponding correlation for all metrics aregiven in Table 9.1.

Table 9.1: Overall performance of the metrics using the method proposed in Section 6.3.

Metric CorrelationS-CIELAB 0.34S-CIELABJ 0.14S-DEE -0.27SHAME-I -0.29SHAME-II 0.19ABF -0.09SSIM 0.18Cao et al. -0.60

138

Page 155: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

9.3.2 Image-wise evaluation

We have also analyzed the performance of the IQ metrics for each image, since previousresearch by Hardeberg et al. [169] has shown that some IQ metrics can be linked with certaincharacteristics of the image. Image-wise evaluation will reveal what causes the low overallperformance of the metrics. Figure 9.6 shows the Pearson correlation for the different IQmetrics for the 15 different images. We can notice a great variation in the results for thedifferent images, and also some variance between the IQ metrics. The Spearman and Kendallcorrelation coefficients follow the same tendency as the Pearson correlation.

Figure 9.6 shows that metrics such as S-CIELAB, S-CIELABJ , SHAME-I, SHAME-II,and ABF perform similarly in many of the images. In images 5, 7, 8, 11, and 12 they have agood correlation. S-DEE differs a bit from the previous metrics, most likely since it is basedon another color difference formula. SSIM is different from the other metrics in terms ofperformance, and it has generally low correlation, except in images 8, 11, and 12. It should benoted that SSIM is a gray scale metric, and does not take into consideration color information,which can explain why it is worse than the other metrics. The metric by Cao et al. performscontrary to the other metrics. This is not surprising, since it is based on artifact detection.We can see a very high correlation in image 3, where the observers mostly judged the imagesbased on detail visibility.

In the images where there is a large difference from the original, but the difference isincreasing the quality, the IQ metrics do not perform well. In the second image some of thereproductions have lost shadow details, but they have a small difference from the original.In this image observers preferred reproductions where details were preserved at the cost oflarger color differences. This is also the reason why the metric from Cao et al. is very similarto the observers. The problem of when a difference contributes to increasing IQ is a difficultissue to handle for the IQ metrics, and the IQ metrics are still not good enough to predictthis. However, in the images where a small difference from the original is preferred, such asimage 5, 7, 8, 11, and 12, the IQ metrics perform reasonably well. In these cases IQ metricsworking solely on the difference from the original are giving good results. This indicates thatthe metrics could be chosen based on the characteristics of the image. The results also showthat one metric is not able to predict perceived IQ, and that several metrics evaluating differentaspects might be required.

9.4 Summary

We have evaluated a set of IQ metrics against the percept using the framework presentedin Chapter 8. The results show that none of the evaluated IQ metrics correlate well withperceived overall quality. For some of the images in the dataset certain metrics perform well,which could indicate that characteristics of the images may influence the performance of theimage quality metrics.

139

Page 156: Doctoral Dissertation by Marius Pedersen

EVALUATION OF METRICS USING THE FRAMEWORK

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Image

Pear

son

Cor

rela

tion

Correlation between subjective score and metric score

S−CIELABS−CIELABJ

SHAME−ISHAME−IIS−DEESSIMABFCao et al.

Figure 9.6: Pearson correlation image-wise. For some images some IQ metrics have a highcorrelation, but the results vary over the 15 images in the dataset. The Spearman and Kendallcorrelation follow the same tendency as the Pearson correlation.

140

Page 157: Doctoral Dissertation by Marius Pedersen

10 WHAT QUALITY ATTRIBUTES ARE THE

MOST IMPORTANT?

Both objective and subjective evaluation of IQ are dependent on a number of Quality Attri-butes (QAs), which are terms of perception [479] such as sharpness, contrast, and saturation.These QAs influence the overall IQ differently, and knowledge about their importance canbe used to achieve an optimal reproduction of an image [135]. The importance of differentQAs has been investigated and acknowledged by many researchers [28, 45, 93, 97, 135, 269,304, 323, 383, 384], but there is, so far, no overall agreement on which QAs that are mostimportant. In this chapter we focus on QAs for the evaluation of color prints.

It has been a goal in IQ evaluation for many years to develop objective measures correlatedwith subjective quality. The advancements in this field have been driven by the desire to reducethe dependence on human observers and minimize both time and resources needed to quantifyIQ. To achieve this, the QAs used in subjective evaluation of quality and their importance needto be identified. These objective measures, such as IQ metrics (Chapter 4), can be used to helpobservers detect quality issues, identify where loss of quality occurs in a printing workflow orcompare the quality of different printing systems.

IQ models have been created to establish a link between subjective and objective quality.These models are theories of perception that enable the prediction of IQ [115]. They areintended to describe the overall IQ of a system and to help researchers in the evaluation ofIQ. IQ models are composed of QAs, and they show how QAs relate to each other and theirinfluence on overall IQ. The goal for IQ models is to evaluate all QAs and their relationships,however this is a very difficult and complex task. Most IQ models reduce the complexity byusing a subset of QAs composed of the most important QAs. By doing this, strengths andweaknesses of a given system can be meaningfully represented with a relatively small numberof QAs [314]. Several IQ models have been proposed [28, 97, 230, 499], however the searchfor better and improved IQ models is still ongoing. The goal of this chapter is to identify anddiscuss the most important QAs for the evaluation of color prints. These QAs can further beused to create a link between subjective and objective IQ in an IQ model.

10.1 State of the art

A survey of QAs and IQ models is given in this section.

141

Page 158: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.1.1 Image quality attributes

Norberg et al. [323] evaluated overall quality, as well as color rendition, sharpness, contrast,detail rendition in highlight and shadow areas, color shift, gloss, mottle, and print homogene-ity in a comparison of digital and traditional print technologies. In a study by Lindberg [269]12 different QAs (overall quality, gamut, sharpness, contrast, tone quality, detail highlights,detail shadow, gloss level, gloss variation, color shift, patchiness, mottle, and ordered noise)were used for the evaluation of color prints. Based on the evaluation performed by observersthese 12 QAs were reduced to two orthogonal dimensions using factor analysis; one relatedto print mottle and one related to color gamut. These two dimensions accounted for almostall variation in the data set. Hardeberg and Skarsbø [170] compared the quality of digitalprinting presses based on colorimetric and spatial measurements, visual xpert evaluations,and panel tests. For the visual tests th observers evaluated color pleasantness, total imagequality, smoothness, and detail rendition. Gast and Tse [152] evaluated six different QAs,blur, noise, banding, color rendition, tone reproduction and printer type. These QAs wereevaluated in terms of preference. Nussbaum et al. [325] investigated print quality evaluationof governmental purchase decisions. In their work they evaluated text quality, color gamut,repeatability, and register for general quality, color match, visual resolution, surface texture,logo alignment, and artifacts for logo assessment, and color match, accuracy, and artifacts forcopy quality. Additionally, several researchers have investigated the importance of QAs likesharpness [45], contrast [41], artifacts (for example noise [383] and banding [93]), natural-ness [135], and color [41, 97, 169, 178, 304].

Research on the combined influence of QAs has been carried out as well. In 1980 Sawyer[384] investigated the influence of sharpness and graininess on perceived IQ, and their com-bined influence. Two years later Bartleson [28] investigated the combined influence of sharp-ness and graininess on color prints. Both Sawyer and Bartleson showed results where theworst QA tended to determine the quality, and a change in other QAs would not increasequality. Natale-Hoffman et al. [314] investigated the relationship between color rendition andmicro uniformity on preference in 1999. This was considered by the authors as a step towardspredicting preference, without dependence on human observers.

Identification of QAs has also been recognized as important for IQ metrics. Morovic andSun [304] based an IQ metric on perceptual QAs, where the QAs were determined based onanswers from observers. Lightness, hue, chromaticity, details, and contrast were found to beimportant. Only the three first, being the most important according to Morovic and Sun [304],were incorporated in an IQ metric (ΔIcm). Later Wang and Shang [463] showed that definedQAs were beneficial for training IQ metrics.

10.1.2 Image quality models

A framework for IQ models was proposed by Bartleson [28] in 1982. His approach wasdivided into three parts:

1. identification of important QAs,

2. determination of relationships between scale values and objective measures,

3. combination of QA scale values to predict overall IQ.

Bartleson used this framework to investigate the combined influence of sharpness and graini-ness on the quality of color prints. This framework has the advantage of representing strengths

142

Page 159: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

and weaknesses of a given system by a relatively small number of QAs. Because of this andits perceptual considerations, the framework has been adopted by several researchers [97, 115,230]. We also adopt this framework in this chapter, where we mainly discuss the first part,identification of important QAs.

Dalal et al. [97] followed the framework in the creation of the document appearance char-acterization system, which is a two-sided appearance based system composed of QAs: onepart for the printer and one for materials and stability. For most QAs in the system, evaluationis performed by experts. The basic IQ is given by 10 QAs, for both the printer and materialsand stability. These describe different aspects of the system, such as color rendition, unifor-mity, tone levels, and stability. The document appearance characterization system has severaladvantages. It uses high level descriptors, which cover a wide range of IQ issues, like defectsand sharpness. The printer is also separated from materials and stability, allowing for separateanalysis. It is also a clear advantage that it is technology independent. In addition, the QAs inthe system are somewhat orthogonal (i.e. they do not influence each other).

This system has some drawbacks as well, since the evaluation is mostly carried out byexperts the results are influenced by the subjectivity of the expert. It might be unsuitablefor non-experts due to its complexity. This is because the QAs are associated with knownprinting problems and technological issues. This approach is different from the approachtaken by Morovic and Sun [304], where QAs were chosen based on answers from observers,resulting in more general QAs. The QAs in the model are also not adapted to IQ metrics,making it difficult to obtain a completely objective evaluation of IQ. It should be taken intoaccount that the document appearance characterization system was not intended to use onlyIQ metrics, since it is made for subjective evaluation by experts. In addition, the system doesnot directly account for the contrast QA, which has been regarded as an important QA byother researchers [41, 269, 304, 323].

Keelan [230] also adopted the framework proposed by Bartleson [28]. He first identi-fied important QAs, then the relationship between a subjective scale (based on just noticeabledifferences) and an objective metric is found. In the case where multiple QAs influence thequality of an image, the influence of each QA to overall IQ is found. Keelan adopted multi-variate formalism as a tool to combine the influence of each QA, in order to obtain a value foroverall IQ. QAs used in the model by Keelan are assumed to be independent, which is differ-ent from the QAs used by others [269, 304, 323]. The advantage is that QAs do not influenceeach other, and can be easily combined in order to achieve a value of overall IQ, which is notstraightforward for dependent QAs. However, the disadvantage is that it might be very diffi-cult to identify independent QAs. The model by Keelan also assumes that objective metricscan be readily designed. Nonetheless, a method to deal with this was proposed; rather thandealing with several QAs, the problem was approached by considering the non-independentQAs as a single QA with several facets.

Engeldrum [115] focused on the building of an IQ model, and partially adopted Bartle-son’s framework. He proposed the IQ circle, which is based on four elements; customerquality preference, technology variables, physical image parameters, and customer percep-tions. The last element, customer perceptions, contains the perceptual QAs (or ”nesses”),being the topic for this chapter. The quality circle shows the relationship between objectiveand subjective quality, but does not include which ”nesses” are important, or how they shouldbe quantified. Engeldrum also states that observers most likely will not be able to perceivemore than five QAs simultaneously. This statement is contradictory with the other IQ modelsand use of QAs, such as the document appearance characterization system, where in total 20QAs (10 for each side of the system) are evaluated. Norberg et al. [323] evaluated overall

143

Page 160: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

IQ and nine QAs. Lindberg [269] evaluated overall IQ in addition to 12 QAs, but analysisshowed that these QAs could be reduced to only two QAs.

Eerola et al. [112] proposed a model based on Bayesian network which connects the objec-tive measurements to the subjective scores. The model discovers links between the objectivemeasures and the QAs used by human observers, and the nature of the QAs and their influenceon each other.

Many other IQ models have been proposed as well [304, 458, 499]. Some of these areIQ metrics, which calculates one or several values representing IQ. One of these metrics isthe S-CIELAB [499] (Section 4.2.2.1), where a spatial pre-processing of the image is carriedout before the CIELAB color difference formula [82] is applied to calculate IQ. Metrics, likethe S-CIELAB and others, are most often constructed to quantify either overall quality or thequality of specific QAs. These metrics usually incorporate several stages of processing. Eachstage is linked to a specific aspect of IQ, where characteristics of different QAs are takeninto account. There are several different approaches to measure IQ. S-CIELAB [499] and theΔIcm [304] are built on the idea that color differences are responsible for a large proportionof differences between an original and its reproduction. Another IQ metric, SSIM [458](Section 4.2.3.1), is based on the degradation of structural information. Many of these metricshave been proposed for different purposes, such as image difference and image fidelity. Fora complete overview of full-reference IQ metrics we refer the reader to Chapter 4, Pedersenand Hardeberg [354], or Appendix A.

10.2 Investigation and selection of important quality attri-

butes

As a first step towards an IQ model relevant and important QAs must be identified. In orderto do this we have taken the approach of doing a survey of the existing literature. NumerousQAs have been considered as important and evaluated by researchers to quantify IQ. In ordernot to exclude QAs in this part of the investigation, we have included QAs based on bothtechnology and perception, and QAs used with different intentions. These QAs include, forexample, lightness [230, 304], sharpness [45, 239, 269, 320, 323], blur [152], contrast [41,269, 304, 323], contouring [161], noise/graininess [28, 152, 292, 320, 327, 383, 384], banding[23, 24, 93, 152], details [41, 169, 304, 320, 323], naturalness [135] , color [41, 97, 169], hue[178, 304], chroma [304], saturation [41], color rendition [97, 152], process color gamut [97],artifacts [41], mottle [239, 269, 379], gloss [245, 269, 323], color reproduction [292], tone re-production [152, 292], color shift [239, 323], ordered noise [323], patchiness [323], line qual-ity [97, 445], text quality [97], gamut size [12], adjacency [97], printer type [152], effectiveresolution [97], effective tone levels [97], gloss uniformity [97], skin color [169], paper rough-ness [239, 480], paper flatness [97], paper whiteness [238, 239], perceived gray value [320],structure changes [320], micro uniformity [97], macro uniformity [97], structure properties[320], color gamut [239], correctness of hue [182], colorfulness proportional to the original[182], correctness of lightness [182], edge sharpness [445], and edge raggedness [445].

When reducing these QAs found in the literature, there are several important issues toconsider, as mentioned previously, such as the intention of how QAs should be used, and theirorigin. A long term goal of this research is to create a link between subjective and objectiveIQ of color prints. With this intention the QAs should be based on perception and account fortechnological printing issues. The QAs should be general enough to be evaluated by observers,

144

Page 161: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

and in order not to exclude novice observers the QAs should be somewhat straightforward toevaluate. In addition, the QAs should be suitable for IQ metrics, being the intended objectivemethod. The existing sets of QAs and models do not fulfill all of these requirements, andtherefore a new set of QAs is needed.

Many of the QAs listed above are similar and have common denominators, which en-ables them to be grouped within more general QAs in order to reduce the dimensionality andcreate a more manageable evaluation of IQ. There is usually a compromise between gener-ality and accuracy when it comes to dimensionality. A small set of general QAs results inlower accuracy, but low complexity, while a higher dimensionality offers accuracy, but highercomplexity. We have linked most of above QAs to six different dimensions, considered asimportant for the evaluation of IQ. This results in a reasonable compromise between accuracyand complexity. We are also close to the statement by Engeldrum [115] that observers will notperceive more than five QAs simultaneously. We have reduced the QAs found in the literatureto the following Color Printing Quality Attributes (CPQAs) six:

• Color contains aspects related to color, such as hue, saturation, and color rendition,except lightness.

• Lightness is considered so perceptually important that it is beneficial to separate it fromthe color CPQA [230]. Lightness will range from ”light” to ”dark” [479].

• Contrast can be described as the perceived magnitude of visually meaningful differ-ences, global and local, in lightness and chromaticity within the image.

• Sharpness is related to the clarity of details [45] and definition of edges [66, 134].

• In color printing some artifacts can be perceived in the resulting image. These artifacts,like noise, contouring, and banding, contribute to degrading the quality of an image ifdetectable [55, 441].

• The physical CPQA contains all physical parameters that affect quality, such as paperproperties and gloss.

The six dimensions are concise, yet comprehensive, general high-level descriptors, beingeither artifactural, i.e., those which degrade the quality if detectable [55, 441], or preferential,i.e., those which are always visible in an image and have preferred positions [441].

We have turned to Venn diagrams to create a simple and intuitive illustration of the CPQAsand their influence on overall IQ. Venn diagrams may be used to show possible logical rela-tions between a set of attributes. However, it is not possible to create a simple Venn diagramwith six fold symmetry [164]. Therefore we illustrate the CPQAs using only five folds, leav-ing the physical CPQA out. This does not mean that the physical CPQA is less important thanthe other CPQAs.

The Venn diagram of Figure 10.1 illustrates how the overall IQ is influenced by one, two,three, four, or five of the CPQAs. Many of the CPQAs are interdependent [336], making qual-ity a multidimensional issue [319], in this case five dimensions. These CPQAs can influencethe overall quality differently, and therefore the ellipses may not have equal sizes or the samepositions in all situations. It is difficult to reduce all of the CPQAs used in the literature tosix dimensions while preserving independence together with having perceptual CPQAs. It isvery difficult, if not impossible, to do this while accounting for most of the aspects of IQ forcolor prints. This is a disadvantage since it leads to a more complex analysis of the results.However, methods have been proposed to deal with this problem [230].

145

Page 162: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

ColorSharpness

Artifacts

ContrastLightness

Figure 10.1: Simple Venn ellipse diagram with five folds used for an abstract illustration ofthe CPQAs. Five different CPQAs and the interaction between these are shown. Overall IQcan be influenced by one (yellow), two (red), three (blue), four (green), or five (gray) of theQAs.

The six CPQAs are a good starting point for the quality evaluation of color prints, andthey can be adapted to different situations. Each of these CPQAs can be divided into sub-QAs for adaptation to specific issues, and thereby increase the accuracy of the QAs. Theartifact QA can, for example, be divided into three sub-QAs; noise, contouring, and banding.Separate analysis of these can be advantageous since it allows for specific analysis either byexperts or IQ metrics. Splitting into sub-QAs can also prove useful when there is skewnessin distribution of CPQAs, resulting in better balance between the CPQAs. Furthermore, somesub-QAs can be placed under several main CPQAs, such as uniformity. This sub-QA canbe placed under color, but also under artifacts since lack of uniformity can be thought ofas an artifact. The placement of these sub-QAs must be done where it is most appropriate.Additionally, all CPQAs might not be used in a given evaluation of IQ, in this case CPQAscan be excluded. By being able to exclude CPQAs and divide CPQAs into sub-QAs, QAsused by other researches can be thought of as special cases of the proposed CPQAs.

In the following parts we take a closer look at the six different CPQAs, where links toQAs from the literature are identified.

10.2.1 Color

Color is a sensation. It is the result of the perception of light by the HVS [149]. The colorCPQAs includes color related issues, like hue, saturation, and color rendition. Lightness isnot a part of our color CPQAs. Since our HVS process lightness and chromaticity informationdifferently, it is convenient to treat these as separate CPQAs [230].

Many of the QAs used in the literature can be linked with one of the six proposed CPQAs.Within our color CPQAs we can find several QAs used by other researchers, such as color [41,97, 169], hue [178, 304], chroma [304], saturation [41], color rendition [97, 152], processcolor gamut [97], color reproduction [292], color shift [239, 323], gamut size [12], skincolor [169], color gamut [239], correctness of hue [182], and colorfulness proportional tothe original [182]. Many of these can also be connected to other QAs, which are discussedbelow.

146

Page 163: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.2.2 Lightness

A common definition of lightness is the visual sensation according to which the area in whichthe visual stimulus is presented appears to emit more or less light in proportion to that emittedby similarly illuminated areas perceived as a ”white” stimulus [479]. Variations in lightnessrange from ”light” to ”dark” [479].

Many QAs used by other researchers can be included within our lightness QA, like tonereproduction [152, 292], perceived gray value [320], correctness of lightness [182], and light-ness [230, 304]. Other QAs are more difficult to link with just one QA, such as colorshift [239, 323], gamut size [12], color gamut [239], process color gamut [97], and colorrendition [97, 152]. All of these have ties to lightness, but also to the color CPQA. Other QAswill influence lightness, but cannot be accounted for within the lightness QA, for examplepaper flatness and gloss level.

10.2.3 Contrast

Contrast is a difficult CPQA since there are many different definitions of contrast [357, 362,375]. Michelson [287] defined contrast as:

Imax − Imin

Imax + Imin,

where Imax and Imin represent the highest and lowest luminance. The Weber contrast [473] isdefined as:

I − Ib

Ib,

where I represents luminance of features and Ib the background luminance. Root-Mean-Square (RMS) contrast is another common way to define contrast:

RMS =[ 1

n−1

n

∑i=1

(xi − x)2]1/2

,

where xi is a normalized gray level value and x is the mean normalized gray level. Contrastcan also be defined as the visual property that makes an object distinguishable. This definitionis useful to express readability of prints. Another used definition of contrast is the lightnessratio between two areas in an image. Fedorovskaya [134] defines contrast as an integrated im-pression of differences in lightness, or lightness variation observed within the whole picture.Keelan [230] defines contrast, in the context of color and tone reproduction, as the relationshipbetween original scene lightness perceived by a photographer and final image (reproduced)lightness perceived by an observer.

Contrast is clearly difficult to define, and it will change according to the application. Evenso, the literature distinctly shows some common characteristics of contrast. It is related tochromaticity [415, 443], and because of this a definition of contrast based solely on lightnesscannot describe perceived contrast in color prints. As well as being correlated with color, con-trast is also related to local and global impressions [415, 443]. Therefore definitions workingon a global aspect are unsuitable for defining contrast in color prints.

The proposed CPQA is a perceptual QA, so a definition should relate to the HVS. To

147

Page 164: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

account for these characteristics of contrast, we define contrast for color prints as the perceivedmagnitude of visually meaningful differences, global and local, in lightness and chromaticitywithin the image. This definition takes into account the characteristics of contrast, and isapplicable to color prints.

Our contrast CPQA can be linked with the use of contrast by several researchers [41, 269,304, 323]. Due to our definition of contrast, it will have ties with many different other QAs,such as chroma, saturation, and lightness.

10.2.4 Sharpness

Caviedes and Oberti [66] relate the perception of sharpness to the clarity of detail and edgedefinition of an image. Bouzit and MacDonald [45] follow a similar thinking, and relatesharpness to details and edges. Fedorovskaya [134] defines overall sharpness as the overallimpression of clarity of edges observed within the whole picture. Sharpness is related to bothdetails and edges, and because of this our sharpness CPQAs is defined as the clarity of detailsand edge definition of an image.

QAs that are suitable to group within the sharpness CPQAs are diverse and many, includ-ing sharpness [45, 239, 269, 320, 323], details [41, 169, 304, 320, 323], line quality [97, 445],adjacency [97], blur [152], effective resolution [97], edge sharpness [445], and edge ragged-ness [445].

10.2.5 Artifacts

Different printing artifacts are found in this CPQA, such as noise, contouring, and banding.All of these have in common that if they are detectable they contribute to degrade the qualityof an image [441]. In this QA, definitions of sub-QAs can be useful. For printing, noise, con-touring, and banding are often considered as important artifacts. Image noise can be definedas random variations in brightness or color in an image. Contouring can be characterized asthe perceived color change within a small region exceeding a threshold [246]. This resultsin perceived discontinuity. Banding is the presence of extraneous lines or bands in a printedpage [24, 93], these appear as nonuniform light or dark lines across the page. This CPQAdiffer from the other since it has a zero point (i.e. no artifacts are presents), while the othershave an optimum.

We can link the proposed artifacts CPQA with several of the QAs used in the literature, forexample contouring [161], noise/graininess [28, 152, 292, 320, 327, 383, 384], banding [23,24, 93, 152], artifacts [41], effective tone levels [97], mottle [239, 269], ordered noise [323],patchiness [323], structure changes [320], and structure properties [320]. All of these willdegrade quality if detectable.

10.2.6 Physical

This CPQA is important since the first five CPQAs cannot account for physical QAs, such aspaper roughness and gloss level. Investigation of the literature has shown these physical QAsto be very important for overall IQ, and should be accounted for in the evaluation of IQ. Thereare several QAs used by researchers that fit within this attribute, like paper roughness [239,480], paper flatness [97], gloss [269, 323], and printer type [152].

148

Page 165: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.2.7 Relations between quality attributes

The six proposed CPQAs are not necessarily independent, and in order to calculate overall IQit is important to know which QAs influence other QAs and the magnitude of their influence.Investigation of the literature shows many of the relationships between CPQAs. Color can belinked to a number of other attributes, one of these is contrast [357, 441]. Color differencescan also be linked to different artifacts, as contouring [246] and banding [79]. One way topreserve details in gamut mapping is to introduce a slight color difference [41], relating colorto the sharpness QA, where details are included. Other relations to sharpness can be foundas well; in the case of sharpening a color difference can be introduced [76, 368]. In extremecases, sharpening can result in halo artifacts caused by color differences [230].

The last two relations for color can also occur due to changes in lightness, creating rela-tions to sharpness and artifacts. Lightness can be linked to contrast [56, 134, 357], but also toartifacts, such as banding [79] and contouring [246].

Contrast has already been linked to color [357, 441] and lightness [56, 134, 357]. It canbe linked to sharpness [26, 56, 57, 217, 221, 222, 392] as well, since an increase in contrastgenerally increases sharpness [222]. In the literature we also find relations to contouring [448]and banding [52] artifacts.

Above, sharpness has been linked with color [41, 76, 368], artifacts [230], and con-trast [26, 56, 57, 217, 221, 222, 392]. Another artifact that is commonly mentioned regardingsharpness is noise, and the relationship between these CPQAs has been extensively examinedin the literature [55, 104, 221, 222, 395].

Artifacts can be related to a number of different CPQAs. For noise it has already beenmentioned the relations to sharpness [55, 104, 221, 222, 395], and for halos to lightness andcolor [230]. While contouring is linked to contrast [448], banding can be related to both colorand lightness, since the bands can be caused by lightness and (or) color variations [79], butalso to contrast [52]. The relations for artifacts will change according to the different artifactsevaluated.

For the physical CPQA many relations can be found. Paper characteristics can for exampleinfluence color [238] and artifacts (as lack of smoothness [239]), while paper coating canaffect artifacts (for example lack of uniformity [323]).

There might also be situations where an increase in the quality of a CPQA will reducethe quality of another CPQA, one example can be the trade-off between noise and sharpness,which has been investigated in the literature [28].

The relations mentioned here do not address the magnitude of influence between CPQAs.

10.3 Validation of the quality attributes - experiment I

In this section we will take a closer look at the three first CPQAs; color, lightness, and contrastas seen on Figure 10.2. In order to investigate quality issues in a color workflow, and toconfirm the proposed QAs in the previous section, an experiment was carried out. A set ofimages was reproduced and investigated by 15 observers to find the most important QAs, andto learn which QAs observers use in the quality evaluation of a color workflow. The observerswere both male and female, and ranged from experts to non-experts.

The images were reproduced using the ICC perceptual rendering intent, which adjustscolor appearance to achieve the most attractive result on a medium different from the origi-nal [187]. In the evaluation of this color workflow, observers evaluate different QAs, and the

149

Page 166: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

ColorSharpness

Artifacts

ContrastLightness

Figure 10.2: A psychophysical experiment was carried out to investigate CPQAs in a colorworkflow. In this experiment a subset of the main CPQAs is considered to affect overall IQ;color, lightness, and contrast.

influence that these QAs have on overall IQ affects the observer’s judgment of quality. Forsome QAs the quality decreases, for other QAs the quality might increase, while some QAsneither increase nor decrease quality. Investigating only the QAs that influence quality mightnot give a correct representation of which QAs that are important. All QAs being evaluatedby observers should be investigated, because of this the instructions given to the observers arecrucial for obtaining correct results.

10.3.1 Experimental setup

10.3.1.1 Test images

Several guidelines have been given in the literature for the selection of images (Section 3.2.4.6),in the context of investigating IQ issues. Holm et al. [176] recommend the use of a broadrange of natural images as well as test charts to reveal the quality issues. The CommissionInternationale de l’Eclairage [80] suggests to include images with the following characteris-tics: high-key, low-key, low lightness contrast, leaves and sky, no neutrals, no white point,no black point, heavy cast, few hues, business graphic, and flesh tones. Buring et al. [54]propose to use natural images, as well as saturated colors. To achieve this we followed therecommendations of Field [143] and CIE [80], where the test images were chosen based onthe following criteria:

• Low, medium, and high levels of lightness,

• Low, medium, and high levels of saturation,

• Hue primaries,

• Low, medium, and high contrast ,

• Larger areas of the same color,

• Fine details,

• Memory colors as skin tones, grass, and sky blue,

150

Page 167: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

• Color transitions,

• Neutral gray.

Most of the images were pictorial with a wide range of scenes like landscapes, portraits, andpersonal items (such as jewelry, books, and clothes). This helps to characterize the impactsfor QAs [230], and ensures that the observer examines a wide variety of QAs. In addition tothe pictorial images a set of test charts were included, since these are content-free and have aselection of ”interest area” colors suitable for evaluation of different aspects of IQ [143].

A total of 56 images, as seen in the lower part of Figure 10.4, were used in this experiment.7 images from ISO [199], 2 images from CIE [80], 3 test charts, and 44 other images capturedby the authors, were included. The 44 images captured by the authors were in RAW format,which were converted to sRGB using Camera Raw 5.0 in Adobe PhotoShop CS4 with aresolution of 150 DPI and a 16-bits encoding. The RAW images were manipulated to lookoptimal by a professional, and then they were verified by two other professionals.

The images were printed at a resolution of 150 pixels per inch and at a size resulting in aprinted reproduction at approximately 8 by 10 inches.

10.3.1.2 Color workflow

The first step in the color workflow was to re-render the set of images from sRGB to theperceptual reference medium gamut [209] using the sRGB v4 perceptual transform. Theoutput profile (Figure 10.3) was generated using the TC3.5 CMYK test target, measured usinga GretagMacbeth Eye-One Pro spectrophotometer and ProfileMaker Pro 5.0.8. Then, as asecond step, linear lightness scaling compensation in the CIE XYZ color space plus the huepreserving minimum ΔE clipping gamut mapping algorithm [80] was applied to the image tore-render from the perceptual reference medium gamut to the gamut of the printing system.The linear lightness scaling were made between the black point CIELAB coordinates of eachimages to the black point CIELAB coordinates of the printing system contained in the outputprofile. The third and last step was to convert the color data from the profile connectionspace values to CMYK values of the printing system by a relative colorimetric transform. Theimages were then printed with the Oce ColorWave 600 wide format CMYK printer on OceRed Label paper.

10.3.1.3 Viewing conditions

The observers were presented with a reference image on an EIZO ColorEdge CG224 (someobservers on an EIZO ColorEdge CG221 since the experiment was carried out in two loca-tions) display at a color temperature of 6500 K and luminance level of 80 cd/m2, followingthe specifications of the sRGB. This set was rendered for sRGB display, and therefore a mon-itor capable of displaying the sRGB gamut was the most adapted reproduction device for thisset of images. In addition, the display was fitted with a monitor hood to prevent glare. Theprinted images were presented randomly in a controlled viewing room at a color temperatureof 5200 K, an illuminance level of 450 ±75 lux and a color rendering index of 96. The ob-servers viewed the reference image and the printed image simultaneously from a distance ofapproximately 60 cm. The experiment was set up to follow the CIE guidelines [80] as closelyas possible.

151

Page 168: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Figure 10.3: Gamuts for the experiment shown with three different projections in CIELABspace using ICC3D [132] (19/07/11: http://www.colorlab.no/icc3d). sRGBgamut on the top, perceptual reference medium gamut in the middle, and the gamut of theprinter on the bottom. The small outer wireframe indicate the sRGB gamut boundary, and thecircle indicates 100 on the a and b axis.

152

Page 169: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.3.1.4 Instructions given to the observers

The instructions given to the observers focused on the overall quality rating of the reproduc-tion and which QAs the observer used in the evaluation. Instructions specified that all QAsused in the evaluation should be stated, even if they did not influence IQ. The two followingquestions were given to the observers:

• Is the printed image a pleasing reproduction?

• According to you, which quality attributes influence the quality of the reproduction?

For the first question a scale from 1 to 7, where 1 represented the most pleasing reproduction,was given to the observers. A description for each level was proposed to help the observers intheir evaluation:

1. Most pleasing you can imagine,

2. Highly pleasing,

3. Very pleasing,

4. Fairly pleasing,

5. Moderately pleasing,

6. Poorly pleasing,

7. Least pleasing reproduction possible.

QAs used during the experiment were noted on a form, where QAs decreasing, increasing ornot influencing quality were denoted with different symbols. Since we want to see how QAsused by observers fit within in the QAs proposed previously, no descriptions or proposals forQAs were given to the observer in order to prevent any influence on the QAs and vocabularyused by the observer.

10.3.2 Perceptual results

15 observers participated in the experiment. Five observers rated the whole data set, and10 observers rated parts of the data set. A total of 452 evaluations were carried out by theobservers, where a scale value was given to each image and the observers described the QAsthey used in their evaluation. The evaluation was carried out in several sessions to preventobserver fatigue.

The overall average pleasantness of all images, based on the seven level scale, has beenfound to be between fairly and very pleasing (Figure 10.4). The whole scale was used in theexperiment, 10 observers used at least one of the extremes on the seven step scale. Analysisof the ratings given by the observers indicates that images with a majority of large areas withthe same color (especially shadow areas) and images with color transitions (both test chartsand natural images) are rated as least pleasing. Analysis of the answers from the observersindicates that contouring is found in the images with color transitions, and since the contouringis highly perceivable in these images the pleasantness is low. In some images with shadowareas details were lost or they were less visible, mainly due to the difference between the input

153

Page 170: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

and output gamut as seen in Figure 10.3. In images with color transitions color breaks or lossof smoothness occur reducing the pleasantness of the image, which is a result of the gamutclipping algorithm. The images rated to be most pleasant have average saturation, lightness,and contrast. Analysis of the results from the observers also reveals that the most pleasantimages are equal or more colorful than the original and have equal or better contrast thanthe original. We will focus on the QAs used by the observers, however, an evaluation andcomparison of quality between the sRGB v4 perceptual transform and sRGB v2 transformhas been carried out elsewhere [39].

10.3.3 Fitting the quality attributes to the color printing quality attri-

butes

For all observers a total of more than 50 QAs have been used in the evaluation, with an averageof 10 different QAs for each observer. For each image an average of 2.95 QAs were used, witha minimum of one and a maximum of eight. This indicates that observers do not consider ahigh number of QAs in their evaluation of IQ. Many of the QAs used on the experimentoverlap, such as lightness, brightness, luminance, and darkness. All of these are connected tothe lightness of the image, and have been grouped within the QA lightness. Similar groupingis done for other QAs to fit within the previously proposed set of QAs (Figure 10.1).

Color is the most frequently used QA by the observers. As seen in Figure 10.5 it has beenused to describe the IQ of more than 70 percent of the images. This is not surprising sincea color workflow was investigated, and the color QA is fairly wide, containing sub-QAs likehue, saturation, and colorfulness. These three sub-QAs are commonly used by the observers,and often used together. The second most used QA, sharpness, mainly contains two sub-QAs;edges and details. Details, both in highlights and shadows, have been frequently used by theobservers. Some observers also commented that loss of contrast led to a loss of perceivedsharpness, since edges and details were less prominent. This phenomenon has also beenacknowledged in the literature [56, 57, 230]. Contrast, the third most frequently used QA, is anarrower QA than color and sharpness. Because of this it is interesting to notice the frequentuse of contrast in the evaluation of color prints, and it confirms the need for a contrast QA inthe evaluation of IQ.

In the images sub-QAs as noise, contouring, banding and so on could be perceived. Theterm artifacts, or sub-QAs of this, were used in approximately 40 percent of the images.Lightness is considered in more than 30 percent of the images. Even though this is the leastfrequently used QA, it should be noted that some observers used the more general term colorrather than separating lightness and chromaticity.

10.3.4 Relations between the attributes

Analysis of the relations between QAs has also been carried out, using cross-tabulation andchi-square tests. The null hypothesis, H0, is that there is no relationship between two QAs.The alternative hypothesis, H1, is that there is a relationship between two QAs. p-valuesfrom this analysis are shown in Table 10.1. For some combinations of two QAs given a 5%significance level, H0 is rejected in favor of H1. This indicates dependence between color andlightness, but also between lightness and sharpness, contrast and artifacts, and artifacts and

154

Page 171: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

1 2 3 4 5 6 7 8 9 10111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455567. Least pleasing

6. Poorly pleasing

5. Moderately pleasing

4. Fairly pleasing

3. Very pleasing

2. Highly pleasing

1. Most pleasing

Rank

Ple

asan

tnes

s va

lue

Ranking of average pleasantness for each of the 56 images

Average pleasantness→

Figure 10.4: Average pleasantness rating for the 56 images in the experiment sorted from themost pleasing image to the least pleasing image. Each rating is plotted with a 95% confidenceinterval. Thumbnails of the images below the graph are sorted in the same order, from left toright and top to bottom. The most pleasing image in the top left corner, and least pleasing inthe bottom right corner.

155

Page 172: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Sharpness Color Artifacts Contrast Lightness0

10

20

30

40

50

60

70

80

Quality Attributes

Per

cent

age

Figure 10.5: Frequency of QAs used by the observers in the experiment. All QAs used by theobservers have been fitted to one of the important QAs proposed in the previous section.

lightness. These results show that use of these QAs occur simultaneously, but not how theyaffect each other or the overall IQ.

Table 10.1: Results from cross-tabulation and chi-square analysis. With a significance levelat 5% there is dependence between color and lightness, lightness and sharpness, contrast andartifacts and artifacts and lightness. Significance at 5% shown in bold.

Sharpness Color Artifacts ContrastColor 0.998Artifacts 0.168 0.219Contrast 0.239 0.723 0.009Lightness 0.005 <0.001 0.008 0.141

In the experiment observers distinguished between QAs that decreased IQ, did not in-fluence IQ, or increased IQ. Observers have marked more QAs to decrease quality than thetwo other groups, and more QAs to increase quality than QAs that do not influence IQ. Fig-ure 10.6 shows the distribution of the three groups for each of the five CPQAs. For theartifacts attribute, some observers stated that lack of artifacts increased quality. Observershave not considered artifacts where it does not influence IQ, indicating that artifacts only areconsidered when they are perceivable, or not present in areas where observers expect to findartifacts. In the sharpness CPQAs lack of details in several of the reproductions contribute todecreased IQ. Therefore we do not have an equal distribution between increase and decreasein IQ as for color, contrast, and lightness.

We have investigated the dependence between image characteristics and the usage ofthe three levels (decreasing, not influencing, and increasing quality) for each CPQA. Cross-tabulation shows dependence between the use of artifacts that decrease quality and imagesclassified as test charts. This is not surprising since contouring was most perceptible in theseimages. Dependence between fine details and increase in quality due to contrast can be foundas well, this indicates that contrast is important for the perception of details. Low lightnessand increase in IQ due to lightness also shows dependence between each other.

156

Page 173: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Sharpness Color Artifacts Contrast Lightness0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Quality Attributes

Nor

mal

ized

Fre

quen

cy

Decreasing qualityNot influencing qualityIncreasing quality

Figure 10.6: Normalized frequency distribution of QAs that increase, decrease, and do notinfluence IQ.

In this experiment the physical CPQA was not considered, and was therefore not a part ofthe analysis. There are also some QAs used by the observers that are difficult to link with oneof the five important CPQAs, for example naturalness and warmness. These CPQAs can belinked with changes in other CPQAs, such as color and lightness.

10.4 Validation of the quality attributes - experiment II

In the previous experiment almost all the QAs used by a group of observers were groupedwithin the proposed CPQAs [345, 346]. This validation was done by subjectively groupingthe numerous QAs used by the observers to one of the six CPQAs. However, since it wascarried out subjectively by the authors, it does not fully validate the aspects on which theCPQAs were selected. Additional validation is therefore required to ensure that they satisfythe intended needs, and to make sure that the proposed CPQAs are suitable to assess IQ. Thegoal of this section is to validate the CPQAs experimentally.

10.4.1 How to validate quality attributes?

The validation should be adapted to the criteria on which the CPQAs were selected. Thevalidation can be achieved by comparing data to the CPQAs, and analyzing the correspon-dence between the data and the CPQAs. Requirements need to be set to validate the CPQAs.Using the aspects on which they were selected (Section 10.2) we can derive the importantrequirements that the CPQAs should fulfill. For the CPQAs to be useful for the evaluation ofIQ, to be perceptual, and account for technological issues, they should be able to cover theentire field of IQ. All issues encountered in the evaluation of color prints should be describedusing the CPQAs, making this one of the requirements to validate. As many as possible ofthe QAs used by the observers should be accounted for within one of the CPQAs, and notoverlap several CPQAs. Minimum overlapping is considered as one of the requirements theCPQAs should fulfill. The CPQAs were selected to keep the number of QAs to a minimum,

157

Page 174: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

this is important for usability of the QAs, and for the CPQAs to be straightforward to use.Therefore dimensionality should be one of the requirements. For the CPQAs to be suitablefor IQ metrics and straightforward to use, it is important to keep independence.

Summarized, we have four different requirements the CPQAs should fulfill in order to bevalidated:

• the CPQAs should cover the entire field of IQ,

• few QAs should overlap the CPQAs (i.e. most of the QAs can be assigned to only oneof the proposed CPQAs),

• dimensionality should be kept to a minimum,

• low or no dependence should occur between CPQAs.

There are several ways to carry out the validation for these requirements. The valida-tion can be carried out subjectively or objectively. The drawback of the previous validation[345, 346] in Section 10.3 of the CPQAs was subjectivity. In order to minimize the subjec-tive influence, and to have an independent validation of the QAs; objective validation methodshave been investigated. It is preferable to have a fully objective method, where data, for exam-ple from an experiment, can be compared to the CPQAs. This requires a database containingall QAs, categorization of them, and their relations. To our knowledge such a database doesnot exist, and this method is therefore inapplicable. Another possible method is to use existingdefinitions of QAs to create relations between the QAs, resulting in a data structure, whichcan be visualized as a tree or data structure. This method is not completely objective, but itkeeps the subjectivity to a minimum. Therefore, we will adopt this method for the validationof the CPQAs.

Since the CPQAs are based on human visual perception, subjective data is required for thevalidation. In order to validate if the CPQAs cover the entire field of IQ it is required that theobservers use a wide variety QAs. Expert observers have been shown to be more precise thannon-experts [108] and they have a wider vocabulary. Therefore expert observers should berecruited for such experiments. In addition, the color workflow on which the data is collectedshould guarantee many different quality issues. The image set should also include a widevariety of characteristics to ensure that many different IQ issues are encountered.

There are various ways to carry out such an experiment. One way is to provide the CPQAsand their definitions to the observers, and ask the observers to use them in their judgment ofIQ. If the observers only use the CPQAs, one could argue that they cover all aspects of IQ.However, this experimental setup can restrict the observers to the CPQAs, and prevent themfrom using other QAs they normally would use. Another option is to record the QAs used bythe observers during the experiment, where the observers do not have prior knowledge of theCPQAs. This last option does not restrict the observers to the CPQAs. Therefore as it has thebest potential to validate the CPQAs, we adopt this method.

10.4.2 Experimental setup

10.4.2.1 Images

We have followed the same guidelines as given in Section 10.3.1.1 to select test images forthe validation. 25 images have been selected (Figure 10.7), where six are the same as those inthe first validation (Figure 10.4).

158

Page 175: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

To address the customer segment of Oce, we have also included 3D models, maps, posters,presentations, and pdf-like documents. The images have been collected from different sources.One image from ISO [204], two from CIE [80], ten images provided by the authors, one im-age from MapTube [283], three images from ESA [121], four images from Google 3D Ware-house [157], one image reproduced with permission from Ole Jakob Skattum, and one imagefrom Halonen et al. [167]. The images were 150 dpi 16-bit sRGB, saved as tiff files withoutcompression.

Figure 10.7: The 25 images used in the experiment to validate the quality attributes.

10.4.2.2 Color workflow

The images were printed on an Oce Colorwave 600 CMYK wide format printer on Oce RedLabel (LFM054) plain uncoated paper. The profile of the printer was generated using a Gre-tagMacbeth TC3.5 CMYK + Calibration test chart in ProfileMaker Pro 5.0.8. A round trip testwas carried out to ensure a correct profile as suggested by Sharma [393], and we performeda visual inspection of color gradients to verify that no artifacts occurred. The images wereprinted with three different rendering intents: perceptual, relative colorimetric, and relativecolorimetric with black point compensation.

10.4.2.3 Viewing conditions

The viewing conditions were identical to those presented in Section 10.3.1.3. Except that thereference image was presented on an EIZO ColorEdge CG221.

159

Page 176: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.4.2.4 Instructions

The instructions given to the observers focused both on the overall quality rating and on theQAs used in the evaluation:Rank the reproductions according to quality.- Elaborate on the attributes you use and quality issues you observe, i.e. all attributes youconsider.- If possible try to give an indication of the importance of the issues and attributes, andimportant areas.

The entire experiment was filmed, and the observers were encouraged to describe and talkabout their observations. The video enabled the authors to better capture the attributes usedby the observers than if they were to write down the attributes, since observers are usuallymore articulate orally.

10.4.3 Fitting the quality attribute data to the color printing quality

attributes

Four observers, all considered to be experts, were recruited for the experiment. This resultedin a total of 100 observations by the four observers for the 25 images, and more than six hoursof video were recorded. The video was transcribed by the authors with focus on the QAsused by the observers. Numerous QAs, more than 750 in total and more than 350 differentQAs, were used by the expert observers. This data constitutes the basis for the validationof the CPQAs. Figure 10.8 shows a tag cloud of the top 25 words from the raw transcribeddata [432].

Figure 10.8: Tag cloud with the 25 top words from the raw transcribed data. The font sizeof a tag in the tag cloud is determined by the number of times the word has been used by theobservers. Similar words have been grouped, such that details and detail are counted togetheras detail.

Since many of the words and phrases from the observers are similar and some synonyms,the QAs from the experiment need to be categorized. Similar words and phrases should begrouped together, and relations between terms found. We have chosen to use existing defi-nitions to accomplish this, and two different approaches can be taken with this method; top-down or bottom-up. In the top-down approach the relations are built from the most generalQAs and downwards to the most specific QAs. This requires building a full tree structurewith all relations, and then comparing it to the QAs used by the observers. In the bottom-upapproach, the starting points are the QAs used by the observers. These QAs are grouped intomore general attributes till the most general QA is reached. The advantage is that it does notrequire building a full tree structure prior to the analysis. Therefore, the bottom-up approachwas chosen to validate the CPQAs.

160

Page 177: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

An example of how the analysis is carried out; the observer has used the QA hue shift,this QA belongs to the more general QA hue. Using the relations of Pedersen et al. [345, 346]and the definition by Wyszecki and Styles [479], hue is considered a part of the more generalcolor QA, which is one of the CPQAs (Figure 10.9).

Color

Hue

Hue shi�

Figure 10.9: Bottom-up procedure for the attribute hue shift, which belongs to the hue at-tribute, which in turn belongs to the color attribute.

In the experiment observers described quality issues in the reproductions, differences be-tween the original and the reproductions, and differences between the reproductions. Sincethe physical CPQA was not changed in the experiment, we limit the discussion to five of thesix CPQAs, excluding the physical CPQA.

The bottom-up approach described above has been used to generate a tree for all the im-ages and observers in the experiment (Figure 10.10). In the following, we will show howthe QAs from the expert observers have been grouped and fitted to the CPQAs, but will onlypresent a part.

The observers have used many specific terms regarding color, such as red hue shift, yellowhue shift, and blue hue shift. All of these terms indicate a hue shift, which is a child of thehue attribute. A drift in hue also indicated a hue shift. Color dominance and color cast wereused to indicate a general hue shift.

The observers specifically indicated which colors had an increase in saturation. They alsotended to indicate which image was more saturated than another, rather than the other wayaround. Also, saturation loss was used in a more general way to indicate a global loss, whilesaturation increase was often used for a color or a region. Increase and loss of saturation areconsidered by the authors to be a shift in saturation. Shift in saturation is a sub-QA of satura-tion. Observers also used the following terms to describe saturation: intensity, chroma, purity,colorfulness, and vividness. Chroma is used for saturation in the Munsell color system [33],in the ISCC-NBS lexicon vividness is a level of saturation [33], purity is also a synonym forsaturation [453], colorfulness is considered the same as chroma [154, 453], and intensity isused about saturation density [425]. Based on these definitions these terms are considered asequal to saturation.

Saturation and hue are considered as children of the color attribute. For the general colorattribute observers used terms as color shift, color reproduction, and color rendering. Inaddition, the observers used the term chromaticity. Since the definition of chromaticity byOleari [328] contains both hue and saturation we can set this attribute as equal to the colorattribute.

10.4.3.1 Discussion on the fitting of quality attributes

Several issues were encountered while fitting the QAs.

Overlapping QAs Some of the QAs used by the observers are difficult to group withinonly one of the CPQAs. Naturalness is one of these attributes. We have argued earlier

161

Page 178: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Image Q

uality

color

hue

hue shift

red hue shift

yellow hue shift

green hue shift

orange hue shift

magenta hue shift

blue hue shift

purple hue shift

saturation

saturation shift

saturation increase

blue saturation

red saturation

pink saturation

orange saturation

magenta saturation

yellow saturation

green saturation

saturation loss

blue saturation loss

orange saturation loss mem

ory color

vegetation

grass treeleafes sky

skin color

skin color shift

red skin shift

yellow skin shift

orange skin shift

snow

sharpness

detail

detail lossdetail enhancem

ent

edge

edge loss

lightness

lightness shift

darklighter

neutrality

white gray

dynamic range

contrast

contrast shift

contrast loss

contrast increase

visual differences

artifacts

smoothness

contouring

noise banding

gloss banding

halo

Figure

10.10:T

heQ

Atree

generatedfrom

theattributes

usedby

fourexpert

observers.E

achlevel

ofa

sub-treehas

beensorted

fromleft

toright

basedon

frequency(high

tolow

).

162

Page 179: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

(Section 10.3) that naturalness could be accounted for by using several of the main or sub-attributes [345, 346]. In this experiment the observers used several QAs together with natu-ralness, very often a change in one or several of the other QAs lead to the impression of anunnatural image. In the five observations, where naturalness was used, the observers used theterm color in all of them, contrast in three of the five, and memory colors in four of the fiveobservations. In addition, it has been shown that naturalness depends on chroma and color-fulness [217], contrast [217], and memory colors [488]. Because of this, naturalness is mostlikely accounted for if these QAs are of reasonable quality.

The word gamut was also used by the observers, which is defined as the range of a setof colors [395]. Gamut cannot be listed as a sub-QA under one of the CPQAs, since it isdependent on both the color CPQA and the lightness CPQA. In the three observations wheregamut was used, both the lightness and the color QAs were used. Therefore, gamut can beaccounted for using the color and lightness CPQAs.

Readability and legibility are two terms from the experiment, which have been found tobe related in the literature [504], and they are often used about textual information. Researchhas shown that contrast is important for text readability [240] and text legibility [265]. Theseterms will also be influenced by sharpness. In five of the eleven observations where legibilityand readability were used, the observers also used contrast and sharpness, in the remainingsix observations either sharpness or contrast was used. This indicates that legibility and read-ability most likely can be accounted for with the CPQAs.

Memory colors are placed under the color CPQA, as the observers only specified colorchanges (saturation and hue) for these, and not changes in lightness. However, there mightbe situations where lightness should be considered as well, and memory colors in terms oflightness will become a sub-QA of lightness.

Independence Dynamic range is considered as a sub-QA of lightness, but it has also beenshown to influence contrast [62]. In the two observations with dynamic range, the observersindicated a relation to visibility of details. This issue is linked to the use of the phrase ”toodark”, which was often used together with ”detail loss”. In these cases, the observers per-ceived the regions where shadow details were lost, as larger dark uniform areas compared tothe original, and used the term ”too dark” or ”darker”.

The experimental data indicates that contrast is influenced by saturation and lightness, butalso that contrast is linked to detail. Since the definition of contrast contains both color andlightness it is perhaps the least independent QA. Furthermore, the experimental data showsthat the observers often use the contrast attribute separated from the color and lightness attri-butes, making contrast a very complex QA. Contrast is also important to account for bothnaturalness and readability. Without the contrast CPQA we would not cover the whole fieldof quality, and it is therefore required in order to fulfill the criteria on which the CPQAs wereselected, even at the expense of independence.

We carried out a cross-tabulation and chi-square analysis to investigate the dependencebetween the CPQAs. The null hypothesis H0 was that there was no relationship betweentwo CPQAs. The alternative hypothesis H1 was that there was a relationship between twoCPQAs. The p-values from this analysis are shown in Table 10.2. For some combinations oftwo CPQAs given a 5% significance level, H0 was rejected in favor of H1. The input data tothe analysis was whether or not one of the five CPQAs was used by the observers for eachof the 25 images. The disadvantage of this analysis is that it does not give any informationon the nature of the relationship between the CPQAs, it only gives information about whentwo CPQAs are used together. However, from the results we see a dependence between ar-

163

Page 180: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Table 10.2: P-values from cross-tabulation and chi-square analysis. With a significance levelat 5%, there is a dependence between artifacts and lightness, contrast and lightness, andbarely between artifacts and sharpness.

Sharpness Lightness Artifacts ContrastSharpness 0.405Lightness 0.566 0.198Artifacts 0.781 0.048 0.014Contrast 0.423 0.230 0.047 0.764

tifacts and lightness, which was also found earlier (Section 10.3.4) [345, 346]. There is alsodependence between artifacts and sharpness, and contrast and lightness. The observers indi-cated a relation between contrast and dynamic range, one of the sub-QAs of lightness. Thedependence analysis between dynamic range and contrast did not reveal a relation, neither fordetail visibility and dynamic range, nor for detail loss and dark. The reason for this mightbe the amount of data, four observers is too low for these specific QAs (Figure 10.12). Any-how, since this analysis does not cover the nature of the relations, further experiments arerequired to investigate the dependence between the CPQAs, where information on the ratingof each CPQA is needed. This information could not be gathered in our experiment since theobservers did not have prior knowledge about the CPQAs.

Global and local issues During the experiment the observers looked at both global andlocal quality issues. The QAs above can be divided into global and local attributes, and thisdifferentiation can be important for the assessment of IQ, but also in the method used tocombine results from different QAs.

One child with several own children In the tree structure (Figure 10.10) some QAs haveonly one child, and this child has several own children. It could then be argued that these QAscould be discarded, and the QA below could be linked to the parent of the removed QA. Forexample, saturation could be removed and replaced with saturation shift. However, observershave used these terms, as saturation alone, without further specification, which indicates thatthese levels are important and should be kept. Furthermore, in other situations there might beseveral children, such as for the edge QA, where one could suggest having two children as forthe detail QA, one for edge loss and another for edge enhancement.

Skewness The experimental data identifies skewness in terms of the number of sub-QAsbetween the different CPQAs, which was also found in Section 10.3.3 or see Pedersen et al.[345, 346]. Our experimental data shows that the color CPQA has significantly more sub-QAsthan the other CPQAs. This can be used as an additional argument for separating lightnessfrom the color CPQA, in order to reduce skewness. Additionally, separating these enablethe CPQAs to be easily adapted for the evaluation of grayscale images. The disadvantage ofskewness between the CPQAs is that it is not straightforward to combine IQ values from thedifferent CPQAs to one overall IQ value, since the CPQAs might have unequal weights.

Dimensionality Since all of the CPQAs have been used by the observers in the experiment,none of the CPQA can be removed directly to reduce the number of CPQAs. However, thecolor and lightness CPQA could be merged, but at the cost of increased skewness. There were

164

Page 181: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

39 observations where the observers used both color and lightness, indicating that the ob-servers differentiate between these CPQA. There were also nine observations where lightnesswas addressed without color.

It is not unlikely that the dimensionality can be reduced for specific workflows with spe-cific documents. Therefore we have also looked at the usage of the CPQAs for the non-photographic documents. For these images all the CPQAs have been used, and for this work-flow none of the CPQAs can be removed. However, there might be situations where only apart of the CPQAs are used to evaluate IQ.

10.4.4 Observations on the color printing quality attributes

The experimental data also leads to different observations on the CPQAs, which can be valu-able in the assessment of IQ. Figure 10.11 shows the frequency of use of the CPQAs in theexperiment. Color is the CPQA used most frequently by the observers, closely followed bysharpness. Artifacts is the least used CPQA by the experts. The results here indicate that thecolor and sharpness CPQAs should be evaluated in all images for IQ assessment. The lownumber of observations regarding artifacts could indicate that this CPQA only needs to beaccounted for in specific images, since the artifacts might be dependent on the characteristicsof the image. One example is banding, which has been perceived by some observers in theimages with large uniform areas, but not in the other images.

Color Sharpness Lightness Contrast Artifacts0

20

40

60

80

100

Quality attributes

Obs

erva

tions

Figure 10.11: Frequency of use for the CPQAs for the 100 observations. Color is the CPQAused the most by the observers, closely followed by sharpness. Lightness is the third mostused CPQA, with contrast being the fourth most used. Artifacts is the least used CPQA.

It is also interesting to look at the distribution of sub-QAs within the CPQAs. Detail isthe most often used sub-QA (Figure 10.12), closely followed by hue. Since the observerspaid much attention to loss of detail in the shadow regions, and since the rendering intentsreproduced these regions differently, detail is not surprisingly the most used sub-QA. Theperceptual rendering intent gave a slight hue shift in some of the images, which was often no-ticed by the observers, resulting in the frequent use of this attribute. The sub-QAs of artifactsare the least used, most likely since these are very specific. The artifact CPQA will containmany sub-QAs since it will cover many different artifacts.

It has been suggested that the first quality issue noticed by the observer is likely to bethe most important. We have analyzed this aspect. Figure 10.13 shows that color is by farthe most frequent first attribute used by the observers, dominating more than in the frequencytable for the whole experiment (Figure 10.11)

165

Page 182: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

7046

30

77

10

55

24

10

2416

24

86

2

1

10

100

Hu

e

Sa

tura

�o

n

Me

mo

ry c

olo

r

De

tail

Ed

ge

Lig

htn

ess

sh

i�

Ne

utr

ali

ty

Dyn

am

ic r

an

ge

Co

ntr

ast

sh

i�

Vis

ua

l d

iffe

ren

ces

Sm

oo

thn

ess

No

ise

Ba

nd

ing

Ha

lo

Color Sharpness Lightness Contrast Ar�facts

Ob

serv

a�

on

s (l

og

)

A�ributes

Figure 10.12: Distribution of sub-QAs within the CPQAs. Detail is most used.

Color Sharpness Lightness Contrast Artifacts0

10

20

30

40

50

60

Quality attributes

Obs

erva

tions

Figure 10.13: Number of observations based on the first CPQA used by the observers.

166

Page 183: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.4.5 Dependence between the attributes

In order to assess dependence between the QAs a second experiment was carried out. Theobservers rated overall quality and five different quality attributes on a seven step scale, wherethe ratings were used as basis for the analysis. First the correlation values between the ratingsfor the attributes were calculated (Table 10.3). As seen in the table non of the attributes havea very high correlation between them, the highest correlation values are found for sharpnessand contrast (0.52) and lightness and color (0.51). The highest correlation values are foundbetween the attributes and overall quality, which is expected since the observers base theirdecision of quality on the attributes. Artifacts is the attribute with the lowest correlationagainst overall quality, but the experimental data showed little differences in this attributecompared to the other attributes.

Table 10.3: Correlation values between the attributes based on the ratings from the observers.

Quality Color Lightness Contrast SharpnessColor 0.64

Lightness 0.63 0.51Contrast 0.65 0.41 0.49

Sharpness 0.59 0.40 0.38 0.52Artifacts 0.45 0.37 0.36 0.37 0.40

Additionally, we have calculated the z-scores based on the category ratings from the ob-servers for each image for overall quality and for the five quality attributes. These z-scoreshave been used to create a contingency table based on the correlation between between thez-scores of the attributes. For each of the 25 images in the experiment z-scores have beenobtained, and the z-score of one attribute is compared against the z-scores of the other attri-butes, where correlation between them is used as a measure of similarity. The mean over the25 images for each combination has been computed and reported in Table 10.4. A high meancorrelation value would indicate a relationship between the attributes. Between the attributesthe highest correlation is 0.54, which is found between sharpness and color, and sharpness andcontrast. The highest correlation values are found between overall quality and the attributes,which reflects the results from Table 10.3.

Table 10.4: Correlation values between the attributes based on the z-scores from the ob-servers.

Quality Color Lightness Contrast SharpnessColor 0.57

Lightness 0.68 0.31Contrast 0.51 0.06 0.50

Sharpness 0.40 0.54 0.38 0.54Artifacts 0.26 0.14 0.42 0.14 -0.16

A PCA analysis of the ratings from the observers for each CPQAs show that all attributescontributes, and that none of the attributes can be removed (Table 10.5). By using only oneCPQA 54% of the variance is accounted for, while using four CPQAs 92% is accounted for,while the remaining 8% is accounted for by using all CPQAs.

167

Page 184: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

Table 10.5: Proportion each principal components (CPQAs) contributes the representation oftotal variance.

Total variance proportionOne CPQA 0.54Two CPQAs 0.68

Three CPQAs 0.81Four CPQAs 0.92Five CPQAs 1.00

We can also analyze the linear correlation between the CPQAs, as independent variables,to the overall quality ratings, dependent variable, by using multiple linear regression. How-ever, this is under the assumption that the relationship between the variables is linear. Thecorrelation coefficients from this analysis shows a low correlation between the single attri-butes and the overall image quality ratings. The R2 gives a value of 0.66, which indicates howwell the regression line fits the data points. All CPQAs contribute to a positive fit of the alinear combination of them, indicating that none can be removed. The analysis also indicatesthat color and lightness contributes most to the fitting.

10.4.6 Validation summary

Prior to the experiment, we specified four requirements for the CPQAs to be validated. Firstthe CPQAs should cover the entire field of IQ. This is fulfilled if all the QAs recorded in theexperiment can be fitted within one or several of the CPQAs. This requirement is satisfied,and all the recorded QAs are accounted for within the CPQAs, either directly as a CPQA oras a sub-QA, or by using two or more of the CPQAs.

The second requirement was to have as few overlapping QAs as possible. Some of therecorded QAs overlap, such as naturalness, gamut, readability, and legibility. These overlap-ping QAs have been used totally 15 times, only a small percentage of the total number ofQAs used. The overlapping QAs can be accounted for using two or more of the CPQAs. Weconsider the number of overlapping QAs to be acceptable, and the overlapping QAs are notfrequently used by the observers. Thus the CPQAs satisfy the second requirement.

The third requirement was to keep the dimensionality to a minimum. None of the CPQAscan be directly removed, and all CPQAs have been used by the observers. However, as dis-cussed above the division between color and lightness has advantages and disadvantages.They could possibly be combined into a single QA. Nonetheless, without merging color andlightness, and considering the use of only lightness by the observers, the third requirement isalso satisfied.

The last requirement regarded dependence. The experimental results show some depen-dence between CPQAs, but as stated previously [345, 346] (Section 10.2.7) the CPQAs arenot fully independent, because it is very difficult, if not impossible, to account for all qualityissues while maintaining a low dimensionality. The experimental results indicate that contrastis the least independent CPQA. However, contrast cannot be removed since it is often usedby the observer. For that reason we consider the dependence found to be acceptable, but caremust be taken if the quality values for each CPQA are combined into an overall IQ value.

168

Page 185: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

10.5 Summary

This chapter has investigated QAs for the evaluation of IQ. Existing QAs used in the literaturehave been identified, and based on these a subset of attributes have been proposed, named theColor Printing Quality Attributes (CPQAs). The CPQAs consist of six attributes covering thefield of IQ. These attributes have been validated in two different experiments, and have beendeemed as a good starting point for evaluating IQ for color prints.

169

Page 186: Doctoral Dissertation by Marius Pedersen

WHAT QUALITY ATTRIBUTES ARE THE MOST IMPORTANT?

170

Page 187: Doctoral Dissertation by Marius Pedersen

11 IMAGE QUALITY METRICS TO

MEASURE QUALITY ATTRIBUTES

With a framework for using IQ metrics with prints and a set of attributes, the next step is toselect suitable IQ metrics for the CPQAs. Then these metrics should be evaluated to investi-gate their correspondence with the percept for each CPQA. We will investigate five of the sixCPQAs, leaving out the physical CPQA since this was not evaluated by the observers.

11.1 Selection of image quality metrics for the color print-

ing quality attributes

Numerous IQ metrics have been proposed in the literature, as seen in Section 4. These takeinto account different aspects of IQ, and therefore a single metric might not be suitable tomeasure all CPQAs. Because of this we will first discuss the properties that the metricsshould have for each CPQA, before a subset can be selected for further evaluation.

11.1.1 Sharpness

Sharpness is related to edges and details, and IQ metrics for sharpness should account forthese two factors. Most of the metrics for sharpness found in the literature work by the prin-ciple to find edges in the image, calculating quality in a local region at the edges, and thencombining the values from the local regions into an overall value representing sharpness qual-ity [229]. Some of these metrics, such as those proposed by Zhang et al. [493], correlate wellwith perceived sharpness. However, these methods do not directly take into account details.Another approach, which better takes into account details, is based on structural similarity,where the metrics usually work on local neighborhoods, such as SSIM [458]. There are alsoother metrics that are based on the visibility of salient objects to assess detail preservation,see for example Cao et al. [61] or Appendix C. Previous analysis of sharpness indicates arelation to contrast [346], therefore metrics accounting for local contrast could be suitable toassess sharpness. Research has also shown that perceived sharpness (blur) is dominated byluminance, and more important than chromatic information [226]. This is also supported byWebster et al. [471], who found that observers are very sensitive to luminance blur, and sensi-tivity to chromatic blur is weak. Therefore, it is not required that the metrics take into accountchromatic information.

171

Page 188: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

11.1.2 Color

There are many potential IQ metrics for the color CPQA. First of all the metrics for this CPQAshould account for color information, making all metrics based on color differences possiblecandidates. However, applying color difference formulas directly, such as the ΔE∗

ab, will mostlikely not predict perceived quality since the impression of quality is influenced by the view-ing conditions. Therefore, the metrics should incorporate a simulation of the HVS, such as theS-CIELAB [499] that performs a spatial filtering before applying the CIELAB color differ-ence formula. These color difference formulas, which many IQ metrics use, usually consist ofthree parts, one for lightness and two for chromaticity. Since lightness and color are separateCPQAs, they should be evaluated separately. As a result the chromaticity part of these met-rics will most likely be more suitable than using all three parts. Additionally, it has also beenargued in the literature that some regions are more important than others. Therefore, met-rics such as SHAME [351], which use different weighting functions are potentially suitablemetrics. For more information on SHAME we refer to Chapter 5.

11.1.3 Lightness

Metrics for the lightness CPQA should mainly follow the same principles as for the colorCPQA. IQ metrics based on color differences commonly calculate lightness quality separatedfrom color, such as S-CIELAB [499]. Metrics working only on grayscale can also be suitableif they analyze lightness, such as SSIM [458] that performs a comparison of lightness betweenthe original and the reproduction. However, metrics analyzing specific aspects in the lightnesschannel are most likely not suitable.

11.1.4 Contrast

The definition of the contrast CPQA states that contrast is both global and local, as well asdependent on lightness and chromaticity. Therefore metrics for the contrast CPQA shouldaccount for these aspects. Metrics computing contrast over a local neighborhood for furthercombining the local values into a global value for contrast are potentially suitable metrics,such as ΔLocal Contrast (ΔLC) [25] or SSIM [458]. Nevertheless, most contrast metrics donot account for chromaticity, and therefore they might not be optimal to measure contrast.One of the metrics that uses chromaticity to calculate contrast is the Weighted multi-LevelFramework (WLF) [417], which also takes into account locality and globality. This metric isexplained more in details in Appendix B.

11.1.5 Artifacts

The artifacts CPQA contains many different and specific QAs. Because of this it might bedifficult to find an overall metric to evaluate all aspects of this CPQA. However, artifacts havesome common denominators, if they are not detectable, they do not influence image quality.Therefore, the metrics used to measure artifacts should account for the sensitivity of the humanvisual system, such as VSNR [70] that has a threshold for when artifacts are perceptible. Thecharacteristics of artifacts is also an important issue, for noise metrics based on local contrastmight be suitable, as ΔLC [25], since noise affects local contrast. Artifacts like banding can bedetected by metrics that are using edge-preserving filters, for example the ABF metric [459],

172

Page 189: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

opposite of metrics performing non-edge preserving filtering, as the S-CIELAB [499]. Forsome sub-artifacts specific metrics might be required, such as for error diffusion worms thathave specific characteristics specially designed metrics as the Error Diffusion Worm Measure(Appendix D) are the most suitable.

11.2 Experiment I

11.2.1 Selected image quality metrics

We cannot evaluate all of the IQ metrics in the literature for all the CPQAs, and becauseof this a sub-set of metrics were selected, as shown in Table 11.1. The selection is basedon the results from previous evaluation [169, 343] (Chapters 7 and 9), the criteria on whichthe metrics were created, the discussion above, and their popularity. Many of the metricshave been created to assesss some aspects of IQ, and therefore only the suitable metrics foreach CPQA will be evaluated. Furthermore, for specific CPQAs we also evaluate parts of themetrics. For example, S-CIELAB combines the lightness and color differences to obtain anoverall value. When suitable, we will evaluate these separately in addition to the full metric.

Table 11.1: Selected IQ metrics for the evaluation of CPQAs.

��

��

��

��

��

��

MetricCPQA

Sharpness Color Lightness Contrast Artifacts

ABF [459] X X XBusyness [331] X XblurMetric [89] XCao [61] X XCW-SSIM[466] X X X XΔLC [25] X X X XLinLab [243] X X XMS-SSIM [467] X X X XM-SVD [402] X X XRFSIM [495] X X X XS-CIELAB [497] X X XS-DEE [412] X X XSHAME [351] X X XSSIM [458] X X X XVSNR [70] X X XWLF [417] X XYCXCzLab [244] X X X

11.2.2 Evaluation of the selected image quality metrics

The IQ metrics proposed for the CPQAs should be evaluated in order to find the best metric foreach CPQA. For doing this we compare the results from the IQ metrics against the subjectivescores from an experiment. We will here only show the results for the best performing metrics.

173

Page 190: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

11.2.2.1 Experimental setup

Two experimental phases were carried out for the evaluation of the metrics. 15 naıve observersparticipated in the first experimental phase, where they judged overall quality and the differentCPQAs on a seven step scale for a set of images. In the second phase, four expert observersranked the quality of a set of images and elaborated on different quality issues. We will givea brief introduction of the experimental setup, for more information see Section 10.4.2 orPedersen et al. [347, 348].

We selected the same 25 images as in Figure 10.7, identically processed and printed.Two sets of images were printed at the same time, one set for the first phase and one set forthe second phase. For more details on the processing of the images we refer the reader toSection 10.4.2.

The observers were presented with a reference image on an EIZO ColorEdge CG224 dis-play for the first phase, and an EIZO ColorEdge CG221 for the second phase. The colortemperature was set to 6500 K and the white luminance level to 80 cd/m2, following thespecifications of the sRGB. The printed images were presented in random order to the ob-servers in a controlled viewing room at a color temperature of 5200 K, at an illuminance levelof 450 ±75 lux and a color rendering index of 96. The observers viewed the reference imageand the printed image simultaneously from a distance of approximately 60 cm.

11.2.2.2 Evaluation of image quality metrics

We will compare the results of the metrics to the results of the observers, and the metric thatcorrelates the best with the observers is the most suitable metric. However, the first step isto turn the printed images into a digital format. This is done with the framework presentedabove in Chapter 8. A HP ScanJet G4050 was used for scanning the images from the firstexperimental phase, while an Epson 10000XL for the second phase. The resolution was setto 300 DPI. The scanners were characterized with the same test target as used to generate theprinter profile.

Since both experimental phases were carried out under mixed illumination, the CIECAM02chromatic adaptation transform [81] was used to ensure consistency in the calculations for themetrics. The measured reference white point of the monitor and the media were used as inputto the adaptation transform, following the CIE guidelines [81].

The evaluation of the metrics has been divided into two phases, one for the naıve observersand one for the expert observers. Each phase contained different methods for the evaluationadapted to the task given to the observers.

Phase 1: naıve observers In the first experimental phase each observer judged overall qual-ity and the five CPQAs for each image, which enabled us to compute z-scores [116] for eachof these. For the first phase 24 of the 25 images (Figure 10.7) were used in the experiment(the bride and groom image left out).

To assess the performance of the evaluated metrics we have adopted several different meth-ods. In order to achieve extensive evaluation we will investigate the performance of the IQmetrics both image by image, and the overall performance over the entire set of images. ThePearson correlation [232] is used for the image-wise evaluation, comparing the calculatedquality and observed quality. The mean of the correlation for each image in the dataset andthe percentage of images with a correlation above 0.6 is used as a measure of performance.While for the overall performance, we will use the rank order method [350] presented in Sec-

174

Page 191: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

tion 6.3 , where the correlation between the z-scores from the observers and the z-scores of themetric is the indication of performance. However, for the rank order correlation one shouldconsider that we only have three data points, and therefore it is also important to perform avisual comparison of the z-scores.

Sharpness: The results of the selected metrics for the sharpness CPQA are shown inTable 11.2. SSIM has a mean correlation of 0.29, but it has a correlation above 0.6 in 50%of the images. The rank order method used to evaluate the overall performance calculates z-scores for the metric, which can be compared against the z-scores from the observers. A metriccapable of correctly measuring the CPQA will have z-scores similar to the z-scores from theobservers. The correlation between the z-scores is used as a performance measure, and SSIMshows an excellent correlation (1.00) with a low p-value (0.03). Visual investigation of thez-scores from SSIM and the observers shows a striking resemblance, therefore SSIM seems tobe a suitable metric for the sharpness CPQA. Other versions of the SSIM, as the Multi Scale-SSIM (MS-SSIM) and Complex Wavelet-SSIM (CW-SSIM) increase the performance. Sincethese account better for the viewing conditions they might be more robust than the single scaleSSIM. Other metrics as the ΔLC and the Riesz-transform based Feature SIMilarity metric(RFSIM) also perform very well for this CPQA. We can also see from Table 11.2 that themetrics perform similar in terms of rank order correlation, and that for an overall evaluationof sharpness these metrics produce similar results.

Table 11.2: Performance of the metrics for the sharpness CPQA. Mean correlation impliesthat the correlation has been calculated for each image in the dataset, and then averagedover the 24 images. Percentage above 0.6 is the percentage of images where the correlationis higher than 0.6. The rank order correlation indicates the correlation between the metric’sz-scores computed with the rank order method (Section 6.3) and the observer’s z-scores forthe CPQA, in addition the p-value for the correlation is found in the parenthesis.

Metric Mean correlation Percentage above 0.6 Rank order correlationCW-SSIM 0.36 63 1.00 (0.05)ΔLC 0.26 50 0.97 (0.14)MS-SSIM 0.29 58 0.97 (0.15)RFSIM 0.34 63 0.99 (0.09)SSIM 0.29 50 1.00 (0.03)

Color: None of the evaluated IQ metrics perform significantly better than the others, interms of the mean correlation, percentage above 0.6 and rank order correlation (Table 11.3).The results here indicate that none of the evaluated metrics can accurately measure the colorCPQA. It is interesting to notice that all of these metrics are based on color differences, whichmight indicate that to find a suitable metric for color one should look of metrics based onanother principle than color differences.

Lightness: MS-SSIM shows the highest mean correlation for the evaluated IQ metrics(Table 11.4), it also has the highest percentage of images above 0.6 in correlation, and ithas an excellent rank order correlation with a low p-value. However, the single scale SSIMalso performs well. The other metrics in Table 11.4 also have a high rank order correlation,indicating that many metrics have an overall similarity to the observers rating.

175

Page 192: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.3: Performance of the metrics for the color CPQA. For further explanation seeTable 11.2.

Metric Mean correlation Percentage above 0.6 Rank order correlationLINLAB -0.25 17 -0.93 (0.24)S-CIELAB -0.29 13 -0.95 (0.19)S-DEE -0.34 13 -0.92 (0.25)SHAME -0.04 21 -0.25 (0.84)

Table 11.4: Performance of the metrics for the lightness CPQA. For further explanation seeTable 11.2. The subscript Lightness indicates only the lightness part of the metric.

Metric Mean correlation Percentage above 0.6 Rank order correlationΔLC 0.31 50 0.94 (0.22)MS-SSIM 0.50 63 0.99 (0.08)S-CIELABLightness 0.14 46 1.00 (0.01)SSIM 0.32 58 1.00 (0.05)VIF 0.34 54 0.99 (0.09)

Contrast: SSIM also performs well for this CPQA as seen in Table 11.5. It has a meancorrelation of 0.32, 50% of the images have a correlation above 0.6, and the rank order cor-relation is fairly high. It is worth noticing that using just the contrast calculation in SSIMthe number of images with a correlation above 0.6 is increased. Also by using MS-SSIM thenumber of images with a correlation above 0.6 is slightly increased compared to the singlescale SSIM. ΔLC has the highest mean correlation, and in 58% of the images a correlationabove 0.6, and the rank order correlation is high. It should be noticed that all metrics havefairly high p-values for the rank order correlation. The differences between the metrics aresmall for all performance measures, and therefore the results do not give a clear indication ofwhich metric that is the best.

Table 11.5: Performance of the metrics for the contrast CPQA. The subscript Contrast indicatesonly the contrast part of the metric. For further explanation see Table 11.2.

Metric Mean correlation Percentage above 0.6 Rank order correlationΔLC 0.34 58 0.97 (0.17)MS-SSIM 0.30 58 0.74 (0.47)RFSIM 0.32 54 0.94 (0.22)SSIM 0.32 50 0.86 (0.34)SSIMContrast 0.32 54 0.84 (0.37)

Artifacts: The observers gave similar results for the different rendering intents in termsof artifacts, because of this it is a very demanding task for the IQ metrics. The evaluationshows that RFSIM has the most images with a correlation above 0.6 together with MS-SSIM,but MS-SSIM has the highest mean correlation (Table 11.6). MS-SSIM and WLF have thehighest rank order correlation, but they do also have high p-values. Therefore the results donot give a clear indication of a suitable metric. One should also consider that the artifactsCPQA contains many different sub-QAs, therefore it could be difficult to find just one IQmetric to measure overall artifact quality, and several metrics might be required.

176

Page 193: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.6: Performance of the metrics for the artifacts CPQA. For further explanation seeTable 11.2.

Metric Mean correlation Percentage above 0.6 Rank order correlationMS-SSIM 0.23 46 0.61 (0.59)RFSIM 0.14 46 0.26 (0.83)SSIM 0.09 29 0.44 (0.71)WLF 0.02 33 0.76 (0.45)

Phase 2: expert observers In the second experimental phase a group of expert observerscommented on quality issues in the reproductions. A video of the experiment was used bythe authors to extract regions were the observers perceived quality issues. This enabled us toperform an in-depth evaluation of the IQ metrics, which ensures that the metrics are capableof measuring the different CPQAs. We will only include the metrics that performed well inthe first evaluation phase, since these are the ones most likely to be suitable for the CPQAs.

We will use the picnic image (Figure 11.1) to evaluate the IQ metrics. The observers in-dicated that this image contained a wide variety of QAs and different quality issues. Thesequality issues are the important issues for the IQ metrics to detect. Based on the commentsfrom the observers important regions have been found, each containing different quality is-sues:

• Tree: mainly details, but also lightness and contrast issues.

• Shoe: loss of details perceived in one of the reproductions.

• White shirt: a hue shift in one of the reproductions.

• Hair: a hue shift in the hair of the Asian girl in the middle.

• Pink shirt: one reproduction was too saturated.

• Grass: detail and saturation issues.

• Skin: a hue shift found in some reproductions.

• Cloth: one reproduction had a lighter red cloth than the others.

• Blanket: lightness issues.

• Sky: saturation and detail issues.

To evaluate the IQ metrics we compare the rank of the metrics, based on the mean value ofeach region, to the rank of the observers for each region. A mask for each region was createdbased on the comments from the observers (example shown in Figure 11.1(b)), and this maskwas used to obtain the ranking from the metrics. The observers did not rank all reproductionsfor all regions or quality issues, but instead they indicated which one was the best or the worst.We consider it to be important for the IQ metrics to predict which reproduction that is clearlybetter or worse. In addition to the ranking of the metrics, a visual inspection of the qualitymaps from each IQ metric has been carried out by the authors. This visual inspection willreveal more information about the performance of the metrics than the mean value.

177

Page 194: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.7: Ranking for the different regions in the image where observers commented onquality issues. P = perceptual rendering intent, R = relative colorimetric rendering intent,and B = relative colorimetric rendering intent with BPC. If (R,P) > B, then B was ranked asthe worst, but the observers did not rank the two other reproductions. () for the metric sideindicates that the mean values are not significantly different with a 95% confidence level. Amask was created based on the comments from the observers, and the mean of the results fromthe IQ metric was used a basis for the ranking.

(a) SSIM

Region Observers SSIM Correct rankTree P > R (B) P > B > R YesShoe P > R (B) P > B > R Yes

White shirt P > B (R) P > B > R YesHair (P,B) > R P > B > R Yes

Pink shirt (P,B) > R P > B > R YesGrass P > (R,B) P > B > R YesSkin R > B > P P > B > R NoCloth (B,R) > P P > B > R No

Blanket (R,B) > P P > B > R NoSky P > (R,B) P > B > R Yes

(b) ΔLC

Region Observers ΔLC Correct rankTree P > R (B) P > B > R YesShoe P > R (B) B > R> P No

White shirt P > B (R) P > B > R YesHair (P,B) > R P > B (B,R) No

Pink shirt (P,B) > R (B,P) > R YesGrass P > (R,B) P >B > R YesSkin R > B > P P > B > R NoCloth (B,R) > P P > B > R No

Blanket (R,B) > P P > B > R NoSky P > (R,B) P > B > R Yes

178

Page 195: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

(a) Picnic image (b) Tree mask

Figure 11.1: The picnic image has been used to show the differences of the IQ metrics. Onthe right side one of the masks, where the mean value from the IQ metrics has been calculatedwithin the black region.

SSIM: For the first experimental phase SSIM had a high performance for the sharpness,lightness, and contrast CPQAs, and performed fairly well for the artifact CPQA. In the secondphase SSIM was able to detect the correct order regarding details, and gives results similarto the observers, as seen from the tree, grass, and shoe regions in Table 11.7(a). The visualinspection supported this, and revealed that SSIM is able to detect even small loss of details.These findings correspond well with the results from the first experimental phase where SSIMwas one of the best performing metrics for the sharpness CPQA. SSIM also correctly detectedan area with a hue shift (hair), since this area in addition had a lightness shift. In the clothregion, where lightness differences were perceived by the observers, SSIM gave the correctranking. SSIM also gave the same ranking as the observers in the tree region, where lightnessand contrast were used by the observers. This shows that SSIM can be suitable to measureboth lightness and contrast, but further analysis is required to ensure that SSIM is able tomeasure these CPQAs. We can also notice that SSIM gives similar ranking for all regions inthe image.

ΔLC: In the first experimental phase ΔLC had one of the best performances for thesharpness CPQA. It is able to detect the detail issues in the grass and tree regions as seen inTable 11.7(b). This is also the reason why it performs quite well for the rank order correlationfor the sharpness CPQA in the first experiment phase. ΔLC also gave good results for thelightness CPQA, but it does not predict the lightness issues commented on by the expert ob-servers. For the contrast CPQA ΔLC correlated with the observers from the first experimentalphase. The experts did not directly comment on contrast, but in the pink shirt one of the re-productions was too saturated, which is one of the aspects looked at when evaluating contrast.ΔLC correctly detects this issue, even though it does account for chromaticity. Additionally,it correctly detects the detail issues as discussed above, which is also one of the aspects ofcontrast. However, since ΔLC does not account for chromaticity, it might not be suitable toevaluate all aspects of the contrast CPQA.

MS-SSIM: MS-SSIM is one of the best performing metrics for the artifact CPQA. Theanalysis for the second experimental phase is difficult, since MS-SSIM is a multilevel ap-proach it is difficult to compute a value for each region. However, investigation of the SSIMmaps for the different levels reveals that MS-SSIM is able to detect banding in several of the

179

Page 196: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

levels. The reason for the higher performance for the artifact CPQA and other CPQAs com-pared to the single scale SSIM might be that the multi-scale version takes better into accountthe viewing conditions, due to the subsampling of the image. The analysis for SSIM above isalso partly valid for MS-SSIM, since the first level of MS-SSIM is the same as SSIM.

CW-SSIM: CW-SSIM does not compute a map like SSIM and some of the other met-rics. But it is based on some principles that makes it suitable for some aspects. Linear anduniform changes correspond to lightness and contrast differences to which CW-SSIM is notsensitive because the structure is not changed, this is the reason why CW-SSIM most likelydoes not work well for contrast and lightness. However, in the sharpness CPQA where detailsare lost in several regions, the structure is changed and therefore it is suitable for sharpness.

RFSIM: This metric does not produce a quality map either, and therefore it is difficultto assess its performance in the second experimental phase. However, RFSIM uses an edgemap to detect key locations in the image. A result of this is that changes that occur at edgesare the main component for the quality calculation, which explains why RFSIM gives goodresults for sharpness, since sharpness is related to the definition of edges [66]. This wouldalso make RFSIM suitable for certain artifacts, such as banding, and could also explain whyit is one of the best metrics for the artifacts CPQA. However, the artifacts CPQA containsmany different artifacts, and therefore one cannot conclude that RFSIM is the best metric tomeasure the artifacts CPQA.

11.2.3 Investigation of image characteristics

The input image to a printing system has different Image Characteristics (ICs), which can bedefined as a collection of properties, that can represent an aspect of an image [429], for ex-ample, image gamut, image sharpness, or image contrast. These ICs will influence the qualityof the image differently, and they are dependent on the processing of the image. Informationabout ICs can be used for deciding the optimal processing of the image, such as the methodand parameters, but also to decide which IQ metrics that should be used in the evaluation ofthe image after it has been processed. As indicated in Chapters 9 and 10 and by Hardeberget al. [169] the performance of metrics might be dependent on ICs. Therefore, in this sectionwe investigate the influence of a set of characteristics on the performance of IQ metrics.

11.2.3.1 Selected image characteristics

Many different ICs are considered to influence IQ, such as out of gamut pixels [108, 155,280, 475], contrast [20, 38, 155], edges [41, 166], colorfulness [92, 94, 171], saturation[426], chroma [20, 155], image resolution [27], gamut size [12, 306, 475], gradients [49, 75],memory colors [96, 147, 488], uniform areas [19], detail level [19, 41, 155], image light-ness [20, 50, 155, 306], and dominant color [50]. We will investigate some of these, moreprecisely lightness, colorfulness, dominant color, and details. Lightness is often consideredas having a major impact on the processing and the final quality [20, 50, 155, 306], and itis therefore one of the most important ICs. Colorfulness has also shown to greatly influenceIQ [94, 171], being one of the ICs considered. Many images have a dominant color, andpreservation of this color is important for the impression of quality [50]. Details is also one

180

Page 197: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.8: Dominant color distribution based on the classification by two expert observers.

Dominant colorYes No11 13

of the ICs that observers use to evaluate quality, and preservation of details is crucial for ob-taining high quality images [41, 284, 429]. These four different ICs constitutes the basis forour analysis.

11.2.3.2 Experiment

The experimental images (Figure 10.7) are the basis for our investigation, and as done inSection 11.2.2 24 of the 25 images have been used. These images have been classified by twoexpert observers into a set of classes according to the selected ICs.

For dominant color the classes were ”yes” and ”no”. For colorfulness and lightness theclasses were ”high”, ”medium”, ”low”, and ”mixed”. For details the classes were ”very de-tailed”, ”detailed”, ”smooth”, ”very smooth”, and ”both detailed and smooth”.

The two experts carried out the experiment together, and were allowed to discuss eachimage and to come up with a decision on what class the image belonged to. The imageswere viewed on a calibrated EIZO ColorEdge CG224 from a distance of 60 cm, with similarviewing conditions as in Section 11.3.2.

The results for the metrics are based on the data collected in Section 11.2.2.2 for the naıveobservers, where the images have been evaluated by a group of observers and metrics arecalculated on the images. The observer data has then been compared against the metrics. Wewill calculate the performance, in terms of correlation, for images classified according to agiven IC.

11.2.3.3 Dominant color

The two experts divided the images for dominant color into two more or less equally largegroups (Table 11.8), 11 images were classified as having a dominant color, while the remain-ing 13 did not have a dominant color.

By computing the Pearson correlation between the IQ metric value and the z-scores fromthe observers for each image, we can analyze the performance for the images that are classifiedas having a dominant color to those classified as not having a dominant color. Each image hasthree reproductions (Section 10.4.2.2), resulting in a correlation value for each image with auncertainty. However, the uncertainty is decreased by taking the mean correlation value of thevalues for one group, making the results more reliable results.

One could hypothesize that in images with a dominant color, one would prefer to have acolor as close as possible to the original color (smallest color difference), since larger regionsof the same color should be considered as more important than smaller regions [177].

Our starting point is the S-CIELAB metric [499] as introduced in Section 4.2.2.1, sincethis is one of the most commonly used metrics in the literature. The observer data is basedon the color attribute. Figure 11.2(a) shows the correlation values for the images in the twoclasses, when the metric values are compared to the color quality values. The mean correlationfor the 11 images classified as having a dominant color is 0.11 (standard deviation 0.66), while

181

Page 198: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Dominant ColorNo dominant color

(a) S-CIELAB.

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Dominant ColorNo dominant color

(b) SSIM.

Figure 11.2: Performance of the metrics for images classified as having a dominant color(red) and images classified as not having a dominant color (blue).

for the 13 images classified as not having a dominant color the correlation is 0.44 (standarddeviation 0.59). This could be an indication that the hypothesis is not valid.

We then keep the same hypothesis that if a dominant color is apparent in the image, it isprobably important to maintain that color, and for images without a dominant color this is notas important. For the SSIM metric that performs well for sharpness, one could argue that forimages without a dominant color SSIM would work better. We have computed the results forSSIM for the two classes (Figure 11.2(b)). The sharpness evaluation done by the observersare the basis of the correlation. SSIM performs quite well for the images classified as nothaving a dominant color, and for those having a dominant color we see larger within-classvariations. The mean correlation for images not having a dominant color is 0.53 (standarddeviation 0.73), while for the other class 0.01 (standard deviation 0.57).

182

Page 199: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.9: Colorfulness distribution based on the classification by two expert observers.

ColorfulnessLow Medium High Mixed

4 10 8 2

11.2.3.4 Colorfulness

The experts also classified the images according to colorfulness, which enables us to carry outa similar analysis as for dominant color (Table 11.9).

We investigate the performance of S-CIELAB for the different classes of colorfulness,using the results from the observers for the color CPQA. From Figure 11.3(a) we can see thatS-CIELAB performs well for the images classified as having medium colorfulness, while thecorrelation for the other groups are varying (Figure 11.3(a)). Even by combining the low andmedium groups we have a higher correlation (0.34) compared to the high group (0.24). Themean correlation values are found in Table 11.10.

Table 11.10: Mean correlation for S-CIELAB for images classified according to colorfulness.

Colorfulness Correlation Standard deviationLow 0.01 0.65

Medium 0.48 0.55High 0.24 0.65

Mixed 0.13 1.22

SSIM performed well for the sharpness CPQA, but in some images it does not have a highcorrelation. We investigate the performance of SSIM based on the colorfulness. Since SSIMis a grayscale metric, and it does not take into account color information, one can hypothesizethat it performs better for images with low colorfulness than those with high colorfulness.This is to a certain extent seen in Figure 11.3(b), where SSIM performs excellent for thefour images classified as having low colorfulness. It also performs well for those with mixedcolorfulness, while for those classified as medium and high we find some high correlationvalues, but also some poor ones (Table 11.11). Observer data used for this analysis are theresults for the sharpness CPQA.

Table 11.11: Mean correlation for the SSIM metric for images classified according to color-fulness.

Colorfulness Correlation Standard deviationLow 0.94 0.05

Medium -0.05 0.77High 0.27 0.53

Mixed 0.80 0.13

11.2.3.5 Lightness

Lightness was classified into low, medium, high, and mixed (Table 11.12). When we lookat the results for the SSIM values against the sharpness CPQA results from the observers

183

Page 200: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Low colorfulnessMedium colorfulnessHigh colorfulnessMixed colorfulness

(a) S-CIELAB.

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Low colorfulnessMedium colorfulnessHigh colorfulnessMixed colorfulness

(b) SSIM.

Figure 11.3: Colorfulness classified images and the performance of the metrics.

184

Page 201: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.12: Lightness distribution based on the classification by two expert observers.

LightnessLow Medium High Mixed

6 13 2 3

for these classes we see that SSIM performs well for the images with low lightness (Figure11.4 and Table 11.13). This is not surprising, mainly because images with pixels below theblack point of the printer can lose details in these regions, especially the images that havebeen processed with the relative colorimetric rending intent using a gamut clipping algorithm.These clipping algorithms usually preserve saturation, but clip image details [42].

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Low lightnessMedium lightnessHigh lightnessMixed lightness

Figure 11.4: SSIM performance for images classified with different lightness levels.

Table 11.13: Mean correlation for the SSIM metric for images classified according to light-ness.

Lightness Correlation Standard deviationLow 0.92 0.11

Medium -0.01 0.68High 0.35 0.73

Mixed 0.26 0.73

11.2.3.6 Details

For the details IC we have grouped the answers for ”very smooth” and ”smooth” together,and also for ”detailed” and ”very detailed”. For these three groups (”smooth”, ”both smoothand detailed”, and ”detailed” as seen in Table 11.14) we look at the performance of the SSIMmetric. Since SSIM is based on structural similarity it should be more suitable for imageswith a higher detail level than those being smooth. Figure 11.5 and Table 11.15 show theperformance of the SSIM metric compared against the observers on the sharpness CPQA.

185

Page 202: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.14: Detail distribution based on the classification by two expert observers.

DetailsSmooth Both smooth and detailed Detailed

8 4 12

For this dataset we do not see any clear indication of a dependence of detail level on theperformance of SSIM.

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

SmoothBoth smooth and detailedDetailed

Figure 11.5: SSIM for images with different detail level for the sharpness CPQA.

Table 11.15: Mean correlation for the SSIM metric for images classified according to details.

Colorfulness Correlation Standard deviationSmooth 0.39 0.80

Both smooth and detailed 0.33 0.60Detailed 0.21 0.68

11.2.3.7 Overall observations

The analysis carried out in this section indicates that there might be a relationship betweenICs and the performance of IQ metrics, such as for SSIM when it comes to images withlow lightness. However, the number of images in the dataset, only 24, is too low to draw aconclusion. Further work with more images is required to determine the influence of ICs onthe performance of IQ metrics, and other ICs than those considered here should be subject forfurther investigations.

186

Page 203: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

11.3 Experiment II

11.3.1 Selected image quality metrics

Many metrics have been proposed (Chapter 4), and as in the previous section a selectionneeds to be done. For this experiment we select the same metrics as for the first experiment(Table 11.1), but we supplement with some additional metrics to ensure a thorough evaluation.Since many of the metrics are designed to account for specific aspects, only the ones suitablefor a given CPQA are evaluated. An overview of the 23 metrics selected for the evaluationand the CPQAs they evaluate is found in Table 11.16.

Table 11.16: Selected IQ metrics for the evaluation of CPQAs.

��

��

��

��

��

��

MetricCPQA

Sharpness Color Lightness Contrast Artifacts

ABF [459] X X XBusyness [331] X XblurMetric [89] XCao [61] X XCW-SSIM[466] X X X XΔLC [25] X X X XIW-SSIM [460] X X X XLinLab [243] X X XMS-SSIM [467] X X X XM-SVD [402] X X XPSNR-HVS-M[366] X X XPSNR-HVS [366] X X XRFSIM [495] X X X XRRIQA [465] X X X XS-CIELAB [497] X X XS-DEE [412] X X XSHAME [351] X X XSHAME-II [351] X X XSSIM [458] X X X XVIF [396] X X X XVSNR [70] X X XWLF [417] X XYCXCzLab [244] X X X

11.3.2 Evaluation of the selected image quality metrics

11.3.2.1 Experimental setup

We want to investigate the relationship between the percept of the CPQAs and IQ metrics. Inorder to do this we have carried out an experiment where human observers judge the qualityof the CPQAs on a set of printed images.

187

Page 204: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Figure 11.6: The ten test images used in the experiment. Each reproduced with four differentsettings.

Test images Ten images (Figure 11.6) were selected from the ISO standards [199, 209].The number of images follow the recommendation by Field [143], who recommend betweenfive and ten images, and the CIE [80], who recommend at least four images. The images wereselected to cover a wide range of characteristics, such as lightness from low to high levels,saturation from low to high levels, contrast from low to high levels, hue primaries, fine details,memory colors as skin tones. These different characteristics will ensure evaluation of manydifferent aspects of IQ. Due to the low number of images (10) investigation of ICs will not becarried out, since the number of images are too low for classification.

Printing workflow Firstly, the color space of all the images was changed to sRGB to definethe reference images. Secondly, then the color space was changed to CMYK using the outputprofile that was generated using a TC3.5 CMYK test target (Section 3.1.4.4), measured with aGretagMacbeth Eye-One Pro spectrophotometer and generated with ProfileMaker Pro 5.0.8.Finally the CMYK images were printed by a HP DesignJet 10ps printer with the HP softwareRIP v2.1.1 using four different modes on Stora Enso Multicopy Original plain paper: thebest print mode, with the resolution of 1200x1200, and the perceptual intent, the best modeand relative colorimetric intent, normal print mode, with the resolution of 600x600 and theperceptual intent, and the last with normal print mode and relative colorimetric intent. Thisresulted in the ten images having four different reproductions, giving a total of 40 images forthe observers to judge.

Observers Ten observers participated in the experiment, all had normal vision without vi-sual deficits. There were 3 females and 7 males with an average age of 23 years.

Viewing conditions The observers were presented with a reference image on an EIZO Col-orEdge CG224 at a color temperature of 6500 K and luminance level of 80 cd/m2. The imageset was rendered for sRGB display, and therefore a monitor capable of displaying the sRGBgamut was the most adapted reproduction device for the set of images. A hood was fitted tothe monitor to prevent glare. The printed images were presented randomly in a controlledviewing room at a color temperature of 5200 K, an illuminance level of 450 ±75 lux and acolor rendering index of 96. The observers viewed the reference image and the printed imagesimultaneously from a distance of approximately 60 cm. The experiment followed the CIEguidelines [80] as closely as possible.

188

Page 205: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Experiment procedure The observers were asked to compare one image selected from theten images at random to its four prints. Sharpness quality, color quality, lightness quality,contrast quality, artifacts quality, and the quality of the main characteristics were evaluated ona five step scale, where 1 indicated best quality and 5 the worst quality. The physical CPQAwas not evaluated since no physical parameter was changed.

11.3.2.2 Experimental results

From the experiment z-scores were calculated using the color engineering toolbox [162],which indicated the perceived differences between the four reproductions. These z-scoreswere calculated for each CPQA and the main characteristics, both image-wise and for thecomplete dataset.

It has been suggested in the literature that some regions of the image are more importantthan others [253, 356, 457]. In order to investigate the relationship between the CPQAs anddifferent regions of the image, we have calculated the Pearson correlation coefficients [232]between the main characteristics and the CPQAs. This analysis would reveal if the qualityof the CPQAs are related to the quality of main characteristics (region-of-interest). FromTable 11.17 we can see that in the different reproductions the main characteristics have varyingcorrelation coefficients with the CPQAs. This indicates that the quality of the CPQAs are notdirectly linked with main characteristics, but that other characteristics are important for theimpression of quality of most CPQAs. However, for some CPQAs and printing modes we seea high correlation between the main characteristics and the CPQAs, this might indicate thatIQ metrics performing a weighting of regions, such as SHAME, could be more suitable thanthose assigning equal weight to the entire image.

Table 11.17: Pearson correlation between z-scores of the main characteristics and the z-scores of the CPQAs for each printing mode and for all modes. N is the number of imagesused in the calculation of the correlation values.

CPQAsMode N Color Lightness Sharpness Contrast Artifacts

Best perceptual 10 0.85 0.47 0.55 0.92 0.28Best Relative 10 0.72 0.45 0.48 0.78 0.55

Normal perceptual 10 -0.02 0.60 0.30 0.61 0.71Normal relative 10 0.31 0.29 0.31 0.88 0.60

All 40 0.79 0.77 0.71 0.89 0.77

11.3.2.3 Evaluation of image quality metrics

In this part we evaluate the selected set of IQ metrics (Table 11.1) for each CPQA against theperceptual data from the experiment.

Preparation of the printed images The printed images cannot be directly used with IQmetrics, since the metrics require a digital input. Because of this the images need to bedigitized. To perform this we have adopted the framework [343] presented in Chapter 8. Firstthe images were scanned at a resolution of 600 DPI using an HP ScanJet G4050. The scannerwas characterized with the same test target as used to generate the printer profile. Since the

189

Page 206: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

experiment was carried out under mixed illumination, the CIECAM02 chromatic adaptationtransform [81] was used to ensure consistency in the calculations for the metrics. The CIEguidelines were followed [81], using the measured reference white point of the monitor andthe media were used as input to the adaptation transform.

Evaluation method Three different methods were adopted for the evaluation of the IQ met-rics. In order to evaluate all aspects of the metrics we will investigate the performance of theIQ metrics both image by image, and the overall performance over the entire set of images.The Pearson correlation [232] is used for the image-wise evaluation, comparing the calculatedquality and observed quality. The mean of the correlation for each image in the dataset andthe percentage of images with a correlation above 0.6 is used as a measure of performance.Overall performance is also an important aspect, and for this evaluation we will use the rankorder method [350], where the correlation between the z-scores from the observers and thez-scores of the metric is the indication of performance. With only four data points it is im-portant to carry out visual inspections of the z-scores to validate the correlation values. Thisevaluation is similar to one carried out for experiment I in Section 11.2.2.

Evaluation results Due to many IQ metrics and several CPQAs we will only show theresults of the best performing metrics for each CPQA.

Sharpness For sharpness the Structural SIMilarity (SSIM) based metrics perform well(Table 11.18). The MS-SSIM has the highest mean correlation with 0.73 and the highestnumber of images with a correlation above 0.6. It also performs among the best for therank order correlation. The results show that metrics based on structural similarity are well-suited to measure perceived sharpness quality. However, other approaches as the ΔLC and theRFSIM have very good performance, indicating that these might be suitable as well.

Table 11.18: Evaluation of IQ metrics for the sharpness CPQA.

Rank order

MetricMean

correlationPercentageabove 0.6

Correlation p-value

CW-SSIM 0.66 70 0.94 0.06ΔLC 0.43 50 1.00 0.00IW-SSIM 0.56 70 0.89 0.11MS-SSIM 0.73 80 0.94 0.06RFSIM 0.61 70 0.97 0.03SSIM 0.66 80 0.96 0.04

Color For the color CPQA none of the evaluated metrics perform well (Table 11.19).It should be noted that all of these metrics are based on color differences, and this might bean indication that using only the color difference from the original is not enough to predictperceived color quality. The color CPQA had a fairly high correlation for all modes betweenthe main characteristic and perceived IQ (Table 11.17), which might indicate that metricsgiving more importance to certain regions, such as SHAME and SHAME-II, could performbetter than the metrics that equally weight the entire image. The experimental results inTable 11.19 show that these metrics do not outperform other metrics.

190

Page 207: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.19: Evaluation of metrics for the color CPQA. Color indicates the color part of themetric.

Rank order

MetricMean

correlationPercentageabove 0.6

Correlation p-value

ABF 0.07 0 0.23 0.77LinLab -0.09 0 0.04 0.96S-CIELAB -0.27 0 -0.24 0.76S-DEEColor -0.38 0 -0.35 0.65SHAME 0.01 10 0.10 0.90SHAMEColor 0.05 20 0.12 0.88SHAME-II 0.23 30 0.27 0.73YCxCzLab 0.24 30 0.33 0.67

Lightness The SSIM based metrics perform very well for the lightness attribute (Ta-ble 11.20), the CW-SSIM has a mean correlation 0.86 and all images have a correlation above0.6. However, other metrics also perform well, such as the RFSIM, ΔLC, S-DEE with onlythe lightness part (S-DEELightness) and ABF with only the lightness part (ABFLightness). Theresults indicate that any of these are appropriate to measure lightness quality. These metricstake different approaches to measure lightness quality, indicating that different strategies aresuitable.

Table 11.20: Evaluation of metrics for the lightness CPQA. Lightness indicates the lightnesspart of the metric.

Rank order

MetricMean

correlationPercentageabove 0.6

Correlation p-value

ABFLightness 0.69 80 0.87 0.13CW-SSIM 0.86 100 0.93 0.07ΔLC 0.69 80 0.99 0.01IW-SSIM 0.85 80 0.95 0.05MS-SSIM 0.82 90 0.93 0.07RFSIM 0.86 90 1.00 0.00S-DEELightness 0.80 90 0.89 0.11SSIM 0.63 70 0.98 0.02

Contrast Many metrics perform well for the contrast CPQA (Table 11.21). The SSIMbased metrics all have a correlation above 0.6 in more than 70% of the images, they also havea high mean correlation and excellent rank order correlation. The RFSIM has a similar per-formance to the SSIM based metrics. All of these metrics would be appropriate for measuringcontrast. However, one should notice that all of the well performing metrics for contrast arebased on lightness, and none of them take color information into account. This might makethem inappropriate to measure contrast in images where color strongly contributes to the im-pression of contrast.

191

Page 208: Doctoral Dissertation by Marius Pedersen

IMAGE QUALITY METRICS TO MEASURE QUALITY ATTRIBUTES

Table 11.21: Evaluation of metrics for the contrast CPQA.

Rank order

MetricMean

correlationPercentageabove 0.6

Correlation p-value

CW-SSIM 0.72 90 1.00 0.00IW-SSIM 0.59 70 0.94 0.06MS-SSIM 0.72 80 1.00 0.00RFSIM 0.67 80 0.96 0.04SSIM 0.65 70 0.99 0.01

Table 11.22: Evaluation of metrics for the artifacts CPQA.

Rank order

MetricMean

correlationPercentageabove 0.6

Correlation p-value

CW-SSIM 0.83 90 0.97 0.03ΔLC 0.72 70 0.94 0.06IW-SSIM 0.83 90 0.99 0.01MS-SSIM 0.77 90 0.97 0.03RFSIM 0.82 90 0.99 0.01SSIM 0.60 70 1.00 0.00

Artifacts The performance for the artifacts CPQA (Table 11.22) follow the results ofmany of the other CPQAs. The SSIM based metrics perform well together with ΔLC andRFSIM. There are only minor differences between these, and any of them seem to be suitableto measure artifacts. However, artifacts can vary significantly and to measure specific artifactsspecially designed metrics might be required.

11.4 Summary

A set of IQ metrics for each CPQA has been evaluated in this chapter. The results indicatethat metrics based on structural similarity are suited to evaluate the sharpness CPQA. For thecolor CPQA none of the evaluated metrics perform well, and more work is needed to finda suitable metric for this attribute. Structural similarity based metrics also perform well forthe lightness CPQA, but other metrics based on color differences also perform well. For thecontrast CPQA the results are similar to sharpness, with structural similarity based metricsbeing the best. Good performance can also be found for the artifact CPQA, but this attributecontains many different sub-attributes which might require metrics adapted to the propertiesof the sub-artifact.

192

Page 209: Doctoral Dissertation by Marius Pedersen

12 IMPROVING PERFORMANCE OF

METRICS BY POOLING

An IQ metric has several different stages as mentioned in Chapter 4 (Figure 12.1), whereeach stage plays a critical role in the assessment of quality. In this chapter we extend theevaluation carried out in Chapter 11, and we investigate how pooling methods can improvethe assessment of print quality through quality attributes.

Pooling is a term commonly used in resource management and refers to the groupingof resources to maximize advantages and minimize risk. It has also been used in vision re-search where it describes the combination of responses of feature detectors in nearby locationsinto some statistic that summarizes the distribution of the features over some region of inter-est [43]. In medical research pooling often refers to the statistical process of generating asummary of data [333]. For IQ metrics, pooling is the technique used to reduce the numberof quality values to a more manageable number, most often one value. The main motivationfor doing pooling is that IQ metrics usually calculate quality locally (either pixel-wise or ina local neighborhood) producing an IQ value for each pixel, resulting in an extensive amountof data, which is difficult to analyze and process. Therefore, pooling is needed to reduce thenumber of values, usually to a single IQ value, since it is easier to manage one value thanmany. Previous research has shown that pooling is very important for achieving an IQ metriccorrelated with the percept [458, 463]. To demonstrate the importance and impact of poolingwe have selected one of the images (Figure 12.2) from the Dugay et al. [107, 108] dataset(Section 7.6). This image has been gamut mapped using five different gamut mapping algo-rithms. In the reproduction shown in Figure 12.2(b) we can notice a loss of details in certainregions, while other parts of the image have contrast changes. These are local changes, thatobservers notice and use in the judgment of quality, and since locality is important, the pool-ing strategies should account for this. The ranking based on three different pooling strategiescan be seen in Table 12.1, and they all produce different ranking results, one close to the

Color space

transforms

Human visual

system models

Quality

calcula�on Pooling

Fewer

Many

Quality

value

Original and

reproduc�on

Original and

reproduc�on

Color space

transforms

Human visual

system models

Quality

calcula�on

Quality

value

Figure 12.1: Stages of IQ metrics. In this chapter we focus on pooling.

193

Page 210: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

(a) Original (b) Reproduction

Figure 12.2: Original image and one of the gamut mapped images used in the calculation ofthe results in Table 12.1.

observers ranking (strategy 1) and one far from it (strategy 3). This shows that the poolingstrategy chosen may have a great impact on the performance of the metric.

Table 12.1: Difference between various pooling strategies. The ranking of the five gamutmapped images using three pooling strategies are different, and none of them completelymatch the observers ranking.

Image A B C D EPooling strategy 1 5 3 2 1 4Pooling strategy 2 5 3 1 2 4Pooling strategy 3 4 1 2 5 3

Observers rank 5 3 2 4 1

Pooling can usually be applied in three different stages, which can be referred to as spatialpooling, channel pooling and quality attribute pooling [258, 463]. Many IQ metrics produce aquality value for each pixel in the image; all of these values are usually referred to as a qualitymap. To combine the values in the quality map in the image domain is called spatial pooling.Most metrics also calculates quality in different channels, for example in a color space suchas CIELAB. Pooling values from different channels is referred to as channel pooling, whichusually results in a quality map. The last type of pooling, quality attribute pooling, is used tocombine several quality maps generated from different quality attributes (i.e., color, lightness,contrast, etc.). Spatial pooling is always needed, while channel pooling and quality attributepooling are needed only when the image is decomposed into different channels or differentquality maps are calculated for each quality attribute. In this chapter, we focus on spatialpooling methods.

The simplest spatial pooling method is to compute the average, which assumes that eachquality value is of equally significance. However, since the image features are highly spatiallynon-stationary, it is intuitively obvious that each region in an image may not bear the sameimportance as others [298]. This is also shown in Figure 12.1 where local changes occur in thereproduction. As a result, giving appropriate weights to each quality value before computingthe average is reasonable. Thus, the main issue is to determine the weights.

194

Page 211: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.1 State of the art

A lot of work has been done concerning spatial pooling approaches. There are two rationalesbehind these approaches. The first is that some regions in the image are more perceptuallysignificant (region of interest) so that distortions in these areas might drastically decrease IQ.These methods differ in their modeling of region significance. The second is that humansubjects tend to pay more attention to the low quality regions. Consequently, the low qualityregions are more important and should be weighted more heavily. The first method refers tolocal content of the image, while the latter refers to the local value of a quality map that isderived as a feature image. We call these two kinds of spatial pooling methods; content-basedand quality-based.

Existing pooling methods can be divided into these two groups with different subgroups:

• Quality based pooling (Section 12.1.2)

1. Minkowski pooling [102, 463] (Section 12.1.2.1)

2. Local quality/distortion-weighted pooling [297, 298, 463] (Section 12.1.2.2)

(a) Monotonic function pooling [463]

(b) Percentile pooling [297, 298]

• Content based pooling (Section 12.1.3)

1. Information content-weighted pooling [463] (Section 12.1.3.1)

2. Gaze based pooling (Section 12.1.3.2)

(a) Gaze-attentive fixation finding engine [297, 298]

(b) Saliency map computed by fixation number and duration [321]

(c) Gaussian weighted sum model [270]

3. Computational saliency model based pooling (Section 12.1.3.3)

(a) Itti’s saliency model [137, 214]

(b) Spectral residual model [180, 277]

12.1.1 General formulation

A general form of a spatial pooling approach is given by

M =∑n

i=1 wimi

∑ni=1 wi

, (12.1)

where wi is the weight given to the ith location and mi is the quality measure of the ith location.M is the pooled quality value. Most spatial pooling methods could be formulated in thisway. In a simple average pooling method, wi is the same over the image space. Actually, wi

might be related to quality mi value at each position or some other factors which are currentlyunknown to us.

Obviously, spatial pooling methods mainly consist of two steps. The first step is to findout what factors might be related to wi. The second step is to determine the exact function thatmaps the assumed factors to wi. The quality based methods assume that the weights wi might

195

Page 212: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

be related to the quality value mi, and the content based methods assume that the weights wi

might be related to image content in the local region around the ith pixel.We will first review quality based pooling methods, before we go through the content

based methods.

12.1.2 Quality based pooling

Wang and Moorthy [297] and Wang and Shang [463] assume that wi is determined by the IQvalue at the ith location in the quality map, i.e.,

wi = f (mi). (12.2)

The methods belonging to quality based pooling follow the principle that low quality valuesshould be weighted more heavily compared to higher quality values.

Nielsen et al. [317] classified weighting functions for gray-level cooccurrence matricesinto two classes depending on how the information is extracted. The first class is based on theposition in the matrix, and the other on the value of the matrix elements. We can also adaptthis classification to pooling, where the weighting in Equation 12.1 can fall into these twoclasses:

• Weighting based on position.

– This is a weighting based on the quality feature value, which covers Minkowskiand Monotonic Funtion pooling.

• Weighting based on value.

– This is a weighting based on the frequency of occurrence of a given quality featurevalue, which covers percentile pooling in a general way.

With this grouping, different than that of dividing Minkowski from local quality/distortion-weighted pooling, show a difference between the quality based pooling methods when they areformulated in terms of a histogram. This grouping shows the difference between Percentilepooling on one hand and Minkowski and Monotonic Function pooling, which has only adifference in the factor m, on the other.

12.1.2.1 Minkowski pooling

Wang and Shang [463] use Minkowski pooling [102] as a spatial pooling method, which waspreviously used by de Ridder [102] as a combination rule for digital-image-coding impair-ments. The Minkowski average is defined as

M =1N

N

∑i=1

mpi , (12.3)

where mi is the quality measure at the ith position of the quality map and N is the total numberof values in the map. We can easily see that Equation 12.3 can be reformulated according to

196

Page 213: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Equation 12.1, which is defined as

M =1N

N

∑i=1

mp−1i mi, (12.4)

which could be considered as the quality value at the ith position is weighted by wi = mp−1i

without normalization. If p = 1, it reduces to the arithmetic mean. If p > 1, wi is a monoton-ically increasing function of mi. If mi is a distortion measure, regions of low quality will beweighted more heavily. More emphasis will be put at highly distorted regions as p increases.

Since only the quality value is involved in Minskowski pooling, and not the position, thesummation may either be performed as a summation over all positions of the (2D) image asimplied by Equations 12.3 and 12.4:

M =1N ∑

x∑y

(m(x,y))p, (12.5)

or as a summation of the the (1D) histogram h(m) of IQ values m:

M =1N ∑

ih(mi)m

pi . (12.6)

12.1.2.2 Local quality/distortion-weighted pooling

Monotonic function pooling Wang and Shang [463] define wi in Equation 12.1 as

wi = mpi , (12.7)

where p ranges from 1/8 to 8 when f (mi) is a distortion measure, which makes sure that wi isa monotonically increasing function of mi. When mi is a quality measure, p ranges from -8 to-1/8 so that f (mi) is a monotonically decreasing function of mi.

Percentile pooling Moorthy and Bovik [297, 298] consider the statistical principle of heav-ily weighting the lower-percentile scores instead of an arbitrary monotonic function of quality.They rank the quality values in a quality map first and then merely weight the lowest p% qual-ity values more heavily. In this method, f (mi) is a step function with an adaptive threshold Tcomputed in a statistical manner. Assuming that mi is the distortion, f (mi) could be formu-lated as

wi = f (mi) =

{r : mi > T1 : mi < T

, (12.8)

where T is determined by the quality map and p.

197

Page 214: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.1.3 Content based pooling

Many researchers [137, 180, 214, 256, 270, 277, 297, 298, 321, 463] hold the opinion that theweight wi is determined by the image content of the local region around the ith location, i.e.,

wi = ci, (12.9)

where ci is a measure of perceptual significance of image content in the local region aroundthe ith location. They assume that an error that appears on a perceptually significant region ismuch more annoying than a distortion appearing in an inconspicuous area. The main issue ishow to model perceptual significance based on the image content.

12.1.3.1 Information content-weighted pooling

Wang and Shang [463] compute information content from local image regions to measureperceptual significance. Regions with higher information content are considered as moreperceptually significant, which implies that people will pay more attention to the regions withhigher information content. Hence, regions of higher information content should have largerweights.

The authors quantify information content (I) as the number of bits that can be receivedfrom a Gaussian image information source model which passes through a visual channel withadditive Gaussian noise. The information received is defined as

I =12

log2

(1+

SC

), (12.10)

where S is the power of the source model and C is the power of noise channel. We can see thatmore information is obtained if the ratio between source power and noise power is larger. Thefinal information content is given by the information in a local region in the reference imageand the distorted image,

I =12

log2

[(1+

Sr

C

)(1+

Sd

C

)], (12.11)

where Sr is the source power of a local region in the reference image, which could be cal-culated as σ2

x . Sd is the source power of a local region in the distorted image, which couldbe calculated as σ2

y . σ2x and σ2

y are the signal variances of a local region in the reference anddistorted image, respectively.

12.1.3.2 Gaze based pooling

These pooling methods use saliency maps generated from data gathered in eye tracking ex-periments. Different ways of doing the experiment and different methods used to calculatethe saliency maps result in different results. However, it has been reported that visual fixationdata improves IQ assessment [18, 356].

198

Page 215: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Gaze-attentive fixation finding engine When shown an image, a human tends to fixate atcertain points on the image. It is likely that these points are more perceptually significantwhen human beings assess IQ. Taking this into account, Moorthy and Bovik use a Gaze-Attentive Fixation Finding Engine (GAFFE) developed by Rajashekar [369] to predict tenfixations points in each image and then weight local regions around these points by a factor k,which is set to 265 by Moorthy and Bovik [297, 298]. Other pixels in the image are weightedby 1.

Saliency map computed by fixation number and duration Ninassi et al. [321] computesaliency maps in two different ways for each observer and for each picture. The first way isthe fixation number for each spatial location. The second way is the fixation duration for eachspatial location. To determine the most visually important regions, all the saliency maps aremerged yielding an average saliency map.

Gaussian weighted sum model Liu et al. [270] assume that each fixation location givesrise to a patch whose activity is Gaussian distributed. The width σ of the Gaussian patchapproximates the size of the fovea. The (k, l)th value in the saliency map is defined as

S(k, l) =T

∑j=1

exp[−(x j − k)2 +(y j − l)2

σ2 ], (12.12)

where S(k, l) is the saliency value in the position (k, l). x j,y j is the position of jth fixation andT is the total number of fixations.

12.1.3.3 Computational saliency model

These methods do not rely on visual fixation data, but model saliency directly, which mightbe inspired by the behavior of the HVS. Compared to eye tracking based saliency maps,these methods are more demanding because they could be directly used in IQ assessmentmethods. However, none of these saliency models capture all the aspects that might contributeto saliency detection in the HVS.

Itti’s saliency model Feng et al. [137] proposed a saliency based method for assessingvideo sequences affected by packet losses. Itti’s bottom-up Saliency based Visual AttentionModel (SVAM) [214] was used to calculate the saliency maps. The model is inspired by thebehavior and the neuronal architecture of the early primate visual system, which demonstratesgood performance in many vision applications. Feng et al. hypothesize that an error appearingon a saliency region is more annoying and show the improvement in video quality assessmentusing SVAM.

Spectral residual model Ma et al. [277] employ the spectral residual model to detectsaliency, which is used as weights in the pooling strategy. To compute the spectral residual(R), an image should firstly be transformed to the Fourier domain. We denote the amplitudespectrum by A( f ) and phase spectrum by P( f ). The spectral residual is defined as

R( f ) = log(A( f ))−h∗ log(A( f )), (12.13)

199

Page 216: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

where h is an averaging filter. The saliency map (SM) is generated by Inverse Fourier Trans-form (IFT):

SM = |IFT (exp(R( f )+ jP( f )))|2 (12.14)

12.2 Experimental setup

In order to evaluate the existing pooling methods we will compare the results from perceptualexperiments against the results of a set of IQ metrics using different pooling strategies. Twodifferent test sets have been selected for this.

12.2.1 Test sets

12.2.1.1 Test set 1

The first test is the same as in Section 11.2.2. The database contains 24 of the 25 refer-ence images, leaving out the bride and groom image as in Section 11.2.2.2. These images(Figure 10.7) have been chosen according to several image characteristics, such as lightness,contrast, saturation, etc.

The images were printed on an Oce Colorwave 600 CMYK wide format printer on OceRed Label (LFM054) plain uncoated paper using three different rendering intents: perceptual,relative colorimetric, and relative colorimetric with black point compensation. The CPQAsfor each printed image were judged by 15 observers. The results are then quantified as z-scores for these printed images. Finally, each reference image has three reproductions whosesubjective quality is represented as z-scores [116] of the five CPQAs. We will evaluate spatialpooling methods for each CPQA, since pooling might be attribute dependent.

In order to apply objective image quality metrics to the printed images, these images arescanned into digital images and stored without compression using the framework proposed inChapter 8. For more details on this dataset we refer the reader to Section 11.2.2 or 10.4, orPedersen et al. [347].

12.2.1.2 Test set 2

The experiment is also conducted on the ten image dataset (Figure 11.6), being the same as inSection 11.3.2. The images were printed by a HP DesignJet 10ps printer using four differentmodes: the best print mode and the perceptual intent, the best mode and relative colorimetricintent, normal print mode and the perceptual intent, and the last with normal print modeand relative colorimetric intent. Ten observers participated in the evaluation of the differentCPQAs, from which the results were calculated into z-scores. For details about the dataset werefer the reader to Section 11.3.2 or Pedersen et al. [359].

12.2.2 Image quality metrics

Six IQ metrics have been chosen for the evaluation based on the selection made in Sec-tions 11.2.1 and 11.3.1 , namely, SSIM [458], S-CIELAB [499], S-DEE [412], WLF [417],ABF [459] and ΔLC [25]. All of these metrics produce a quality map, making them suitable

200

Page 217: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

for evaluation of spatial pooling strategies. We will go through each CPQA to show the per-formance of the pooling methods. Since all metrics are not designed to evaluate all CPQAs,we only evaluate the suitable metrics for each CPQA, where the selection is based on theguidelines found in Section 11.1 and Pedersen et al. [349] (Table 12.2).

Table 12.2: Selected IQ metrics for the evaluation of the CPQAs and pooling strategies.

��

��

��

��

��

��

MetricCPQA

Sharpness Color Lightness Contrast Artifacts

SSIM [458] X X X XABF [459] X X XΔLC [25] X X X XS-CIELAB [499] X X XS-DEE [412] X X XWLF [417] X X

12.2.3 Pooling methods

For each metric, we test Spatial Average Pooling, Minkowski Pooling [463], MonotonicFunction Weighted pooling [463], Information Content Based Pooling [463], IG SaliencyModel Based Pooling [1], and Nonparametric Bottom-up (NB) Saliency Model Based Pool-ing [389, 390]. In the two saliency based methods the saliency map is used as a weightingmap using the general formulation in Equation 12.1.

12.2.4 Evaluation methods

We use three evaluation methods to compare the performance of different spatial poolingmethods for each IQ metric.

• Pearson Linear Correlation Coefficient (PLCC) as explained in Section 6.2.1.1. Themean PLCC of the whole database is calculated as the average of PLCCs of each image.

• The Percentage of Images (POI) with PLCC higher than 60% is also examined.

• Rank Correlation (RC), as presented in Section 6.3 and Pedersen and Hardeberg [350],which is the PLCC correlation between objective rank order z-score and subjective z-score to measure the order consistency between objective and subjective assessment. Tocompute the objective rank order z-score, we firstly rank the original objective scores ofeach scene’s three reproductions and then compute objective rank order z-score usingthe rank order function in the Colour Imaging Toolbox [162].

12.3 Experimental results

We compare the results from the IQ metrics with the different pooling methods to the percep-tual results.

201

Page 218: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.3.1 Sharpness

12.3.1.1 Test set 1

For the ΔLC metric the NB saliency model has the highest PLCC value, improving the PLCCby 0.1 compared to average pooling as seen in Table 12.3. But POI values between theseremains the same because the NB pooling method is not able to improve the low correlationvalues to high ones, although it improves some correlation values slightly. On the other hand,Minkowski pooling with p = 1

2 and p = 2, and the Monotonic Function pooling with p = 14 and

p = 12 change one bad correlation value to a very good one, but they decrease the correlation

values of other images. This is why POI improves while PLCC slightly increases or evendecreases. The above pooling methods show similar performance in terms of RC comparedto average pooling. For ΔLC we also see that the correlation decreases when the p for bothMonotonic Function and Minkowski is high.

For the SSIM metric, Monotonic pooling with p = 1 improves both PLCC and POI whileit maintains a high RC. The RC for Minkowski pooling is very dependent on the p value,decreasing the performance with increasing p values.

For both metrics the selected parameter in the pooling methods is important, and a wrongselection can decrease the performance of the metric. However, SSIM seems to be moredependent on the pooling method and parameters than ΔLC.

Table 12.3: Performance of the spatial pooling methods for the sharpness CPQA on TestSet 1. The ΔLC and SSIM metrics are selected for the evaluation. The best results for eachperformance measure and metric are shown in bold.

PoolingMethods

p ΔLC SSIMPLCC POI RC PLCC POI RC

Average 0.26 50% 0.97 0.29 50% 1.00

Minkowski

18 0.19 46% 0.86 0.37 54% 0.9214 0.21 46% 0.90 0.33 50% 0.9412 0.24 54% 0.95 0.31 54% 0.942 0.23 54% 1.00 0.20 38% 0.814 0.22 50% 0.98 0.14 33% -0.238 0.21 50% 0.98 0.01 33% -0.68

MonotonicFunction

18 0.29 50% 0.93 0.27 46% 0.9914 0.28 54% 0.95 0.28 42% 1.0012 0.25 54% 1.00 0.32 46% 1.001 0.26 50% 0.99 0.42 54% 0.992 0.24 50% 0.99 0.30 42% 0.994 0.13 42% 0.99 0.17 38% 0.898 0.05 38% 0.91 0.25 50% 0.91

IC 0.29 50% 0.96 0.32 50% 0.97IG 0.25 46% 0.93 0.10 38% 0.90NB 0.36 50% 1.00 0.33 54% 1.00

202

Page 219: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.3.1.2 Test set 2

For ΔLC, the spatial average pooling gives an excellent rank order correlation of 1.0 (Ta-ble 12.4). NB Saliency Model Based Pooling improves the performance in terms of meancorrelation and percentage of images above 0.6, with a rank order correlation comparable toaverage pooling. We can also notice that using Minkowski or Monotonic Function poolingwith a correct choice of p will increase the POI up to 70%. For SSIM, Minkowski Poolingwith p = 8 gives the best results concerning mean correlation and percentage of images above0.6. However, the rank order correlation drops a from 0.96 for average pooling to 0.7. Thespatial average pooling seems the best for SSIM, since it has the highest RC while maintain-ing good results for PLCC and ROI. The result is to some extent consistent with that of thefirst dataset.

Table 12.4: Performance of the spatial pooling methods for the sharpness CPQA on Test Set 2.ΔLC and SSIM are selected for the evaluation. The best results for each performance measureand metric are shown in bold.

PoolingMethods

p ΔLC SSIMPLCC POI RC PLCC POI RC

Average 0.43 50% 1.00 0.66 80% 0.96

Minkowski

18 0.42 40% 1.00 0.64 80% 0.9114 0.42 40% 1.00 0.65 80% 0.9112 0.43 50% 1.00 0.65 80% 0.912 0.44 60% 0.99 0.67 80% 0.954 0.46 70% 0.95 0.68 80% 0.908 0.31 50% 0.77 0.73 90% 0.70

MonotonicFunction

18 0.42 50% 0.99 0.65 80% 0.9314 0.42 60% 0.99 0.64 80% 0.9112 0.42 60% 0.97 0.61 80% 0.911 0.43 70% 0.96 0.02 30% -0.282 0.47 70% 0.89 -0.16 10% -0.824 0.33 20% 0.31 -0.36 0% -0.998 -0.26 10% -0.73 -0.41 0% -0.96

IC 0.39 50% 1.00 0.65 80% 0.93IG 0.42 60% 0.89 0.65 80% 0.88NB 0.51 70% 0.99 0.60 70% 0.92

12.3.2 Color

12.3.2.1 Test set 1

For the color CPQA, the three selected metrics all give very poor results (Table 12.5). Thismight be a consequence of the fact that quality does not monotonically decrease with colordifference between the reference image and printed images. Human observers might preferimages that are more saturated, even though they might have a larger color difference thanless saturated images. Table 12.5 shows that the pooling methods contribute little to the results

203

Page 220: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

when the metrics are very poor. The Monotonic Function pooling strategy improves the resultsslightly, possibly because local regions with large color difference that are really annoying aregiven larger weights.

Table 12.5: Performance of the spatial pooling methods for the color CPQA on Test Set 1.ABF, S-CIELAB, and S-DEE are selected for the evaluation. The best results for each perfor-mance measure and metric are shown in bold.

PoolingMethods

p ABF S-CIELAB S-DEEPLCC POI RC PLCC POI RC PLCC POI RC

Average -0.39 8% -0.99 -0.29 13% -0.95 -0.34 13% -0.92

Minkowski

18 -0.40 4% -0.99 -0.25 17% -0.95 -0.25 17% -0.9514 -0.40 4% -0.99 -0.25 17% -0.95 -0.26 21% -0.9512 -0.40 4% -0.98 -0.26 13% -0.95 -0.27 21% -0.922 -0.36 13% -0.99 -0.32 13% -0.93 -0.39 8% -0.904 -0.18 8% -1.00 -0.34 13% -0.97 -0.36 13% -0.958 -0.07 21% -0.99 -0.20 17% -0.98 -0.24 17% -0.98

MonotonicFunction

18 -0.39 13% -0.99 -0.30 8% -0.95 -0.36 8% -0.9314 -0.38 13% -0.99 -0.31 8% -0.95 -0.38 8% -0.9012 -0.36 13% -0.98 -0.33 8% -0.93 -0.41 8% -0.931 -0.29 13% -1.00 -0.13 25% -0.97 -0.44 8% -0.972 0.01 21% -0.94 0.11 38% -0.10 -0.25 25% -0.994 0.27 29% -0.97 -0.05 21% -0.99 0.06 33% -0.418 0.27 29% -0.97 -0.06 21% -1.00 0.08 29% -0.35

IC -0.38 4% -0.99 -0.27 13% -0.93 -0.35 13% -0.93IG -0.22 13% -0.94 -0.23 21% -0.97 -0.18 21% -0.95NB -0.27 13% -0.99 -0.21 8% -0.94 -0.28 13% -0.95

12.3.2.2 Test set 2

None of evaluated metrics shows satisfactory results for the color CPQA for most of thepooling methods (Table 12.6), which is similar to the results on the first dataset. However,the Monotonic Function Based Pooling with p = 2 significantly improves the performance ontest set 2 for the ABF metric, giving a PLCC of 0.66 (Figure 12.3), 70% of the images above0.6, and a RC of 0.98. The improved performance could be caused by the fact that large colordifferences strongly contributes to the impression of quality, and with a relative large p valuethese are given higher importance. We also see a similar increase for S-CIELAB and S-DEEbut the maximum performance is reached at p = 4 (Figure 12.3). The results for the differentmetrics also indicate that the selection of the pooling method and parameters is very importantfor the performance of the metric.

204

Page 221: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table 12.6: Performance of the spatial pooling methods for the color CPQA on Test Set 2.ABF, S-CIELAB, and S-DEE are selected for the evaluation. The best results for each perfor-mance measure and metric are shown in bold.

PoolingMethods

p ABF S-CIELAB S-DEEPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.07 0% 0.23 -0.27 0% -0.24 -0.42 0% -0.42

Minkowski

18 -0.14 0% 0.08 -0.32 0% -0.24 -0.48 0% -0.4214 -0.12 0% 0.08 -0.31 0% -0.24 -0.47 0% -0.4212 -0.06 0% 0.08 -0.30 0% -0.24 -0.46 0% -0.422 0.30 20% 0.68 -0.21 0% -0.24 -0.36 0% -0.384 0.57 60% 0.93 -0.02 10% -0.06 -0.24 0% -0.178 0.32 20% 0.38 -0.14 10% 0.09 -0.14 0% 0.16

MonotonicFunction

18 0.15 10% 0.45 -0.26 0% -0.25 -0.40 0% -0.4214 0.22 10% 0.55 -0.24 0% -0.24 -0.39 0% -0.4212 0.35 40% 0.72 -0.20 0% -0.24 -0.35 0% -0.381 0.52 50% 0.89 -0.07 10% -0.05 -0.27 0% -0.202 0.66 70% 0.98 0.16 30% 0.40 -0.12 0% -0.024 0.29 20% 0.32 0.23 40% 0.61 0.08 0% 0.378 -0.16 0% -0.30 -0.22 0% -0.43 -0.01 10% 0.28

IC 0.07 0% 0.23 -0.28 0% -0.25 -0.43 0% -0.42IG -0.01 0% 0.27 -0.28 0% -0.26 -0.33 0% -0.32NB 0.06 10% 0.38 -0.34 0% -0.29 -0.41 0% -0.48

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1/8 1/4 1/2 1 2 4 8

PLC

C

p value

Color - Monotonic Func�on pooling

ABF S-CIELAB S-DEE

Figure 12.3: PLCC performance of Monotonic Function pooling for the color attribute.

205

Page 222: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.3.3 Lightness

12.3.3.1 Test set 1

SSIM and ΔLC give slightly better results than ABF, S-CIELAB, and S-DEE in terms ofPLCC and POI (Table 12.7). The RC of SSIM and ΔLC is very high, while for the othermetrics it is low. For ABF, S-CIELAB, and S-DEE, the IG pooling method shows the bestperformance regarding PLCC and ROI, which is comparable to the best performance of SSIMand ΔLC. However, none of the spatial pooling methods is able to raise the low RC coeffi-cients of ABF, S-CIELAB, and S-DEE to satisfactory values. For ΔLC, none of the poolingmethods stands out in any of the performance measures, but a slight increase can be foundfor specific methods. For SSIM, the Monotonic Function pooling with p = 1

4 improves theperformance to the best one among all the combinations of evaluated metrics and poolingmethods (see Table 12.7). However, the parameter for Minkowski and Monotonic functionpooling is important for the performance.

12.3.3.2 Test set 2

For the lightness CPQA for the second test set, ΔLC and SSIM performs quite well for aver-age pooling (Table 12.8), while ABF, S-CIELAB, and S-DEE show poor results with spatialaverage pooling (Table 12.8). For ABF, S-CIELAB and S-DEE, Monotonic Function poolingwith p = 2 or p = 4 gives better results, but the selection of the parameter is important. Thisincrease in performance was not found in the first data set, which might indicate that poolingis image dependent. From Table 12.8 and Figure 12.5 we can see that SSIM and ΔLC hasdecreasing performance in terms of PLCC for increasing p values for Monotonic Functionpooling, while for the color difference metrics (S-CIELAB, ABF, and S-DEE) the PLCC isincreasing up to a certain p value. For Minkowski pooling ABF peaks at p = 4 (Table 12.8),while ΔLC and SSIM are high and more robust to changes in p compared with MonotonicFunction pooling (Figure 12.4). For ΔLC and SSIM, spatial pooling methods other than av-erage gives only slight improvements in the performance. For SSIM, Monotonic Functionpooling with p = 2 shows the best performance.

-0.60

-0.40

-0.20

0.00

0.20

0.40

0.60

0.80

1.00

1/8 1/4 1/2 2 4 8

PLC

C

p value

Lightness - Minkowski pooling

ABF

S-CIELAB

S-DEE

ΔLC

SSIM

Figure 12.4: PLCC performance of Minkowski pooling for the lightness attribute.

206

Page 223: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Tabl

e12

.7:

Perf

orm

ance

ofth

esp

atia

lpoo

ling

met

hods

for

the

ligh

tnes

sC

PQ

Aon

Test

Set1

.AB

F,S-

CIE

LA

B,S

-DE

E,Δ

LC

,and

SSIM

are

sele

cted

for

the

eval

uati

on.

The

best

resu

lts

for

each

perf

orm

ance

mea

sure

and

met

ric

are

show

nin

bold

.

Pool

ing

Met

hods

pA

BF

S-C

IEL

AB

S-D

EE

ΔL

CSS

IMPL

CC

POI

RC

PLC

CPO

IR

CPL

CC

POI

RC

PLC

CPO

IR

CPL

CC

POI

RC

Ave

rage

0.12

38%

0.11

0.29

50%

0.24

0.19

46%

0.33

0.31

50%

0.94

0.32

58%

1.00

Min

kow

ski

1 80.

1438

%0.

050.

3150

%0.

250.

1742

%0.

260.

3754

%0.

790.

3754

%0.

871 4

0.14

38%

0.05

0.31

54%

0.25

0.17

42%

0.26

0.36

54%

0.84

0.37

54%

0.90

1 20.

1438

%0.

120.

3154

%0.

250.

1842

%0.

340.

3454

%0.

910.

2954

%0.

892

0.09

38%

0.11

0.25

54%

0.30

0.16

42%

0.38

0.15

42%

1.00

0.12

46%

0.74

40.

0633

%0.

010.

2550

%0.

180.

1150

%0.

260.

1242

%1.

000.

0733

%-0

.34

80.

2046

%-0

.18

0.17

42%

0.12

-0.0

438

%-0

.25

0.14

33%

1.00

-0.0

433

%-0

.76

Mon

oton

icFu

nctio

n

1 80.

1238

%0.

110.

2850

%0.

240.

1846

%0.

310.

2546

%0.

890.

3458

%1.

001 4

0.11

38%

0.11

0.26

50%

0.24

0.17

46%

0.38

0.21

46%

0.91

0.38

63%

1.00

1 20.

0838

%0.

120.

2350

%0.

300.

1642

%0.

310.

1446

%0.

990.

3758

%0.

981

0.05

38%

-0.1

20.

1442

%0.

180.

1346

%0.

180.

0942

%1.

000.

5054

%0.

972

0.12

33%

-0.3

9-0

.08

29%

-1.0

0-0

.04

38%

0.08

0.10

42%

1.00

0.29

46%

0.97

40.

0538

%0.

180.

0946

%-0

.22

-0.0

933

%-0

.93

0.05

33%

1.00

0.15

42%

0.84

80.

0538

%0.

180.

1746

%-0

.13

-0.2

225

%-0

.96

0.16

38%

0.95

0.20

38%

0.86

IC0.

1438

%0.

110.

3050

%0.

300.

2150

%0.

310.

2650

%0.

920.

3054

%0.

94IG

0.35

50%

0.27

0.39

54%

0.18

0.30

54%

0.24

0.25

54%

0.89

0.32

50%

0.94

NB

0.16

42%

0.04

0.35

50%

0.29

0.16

42%

0.24

0.35

50%

0.98

0.29

58%

0.99

207

Page 224: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table12.8:

Performance

ofthespatialpooling

methods

forthe

lightnessC

PQ

Aon

TestSet2.AB

F,S-CIE

LA

B,S-D

EE

,ΔL

C,and

SSIMare

selectedfor

theevaluation.

The

bestresultsfor

eachperform

ancem

easureand

metric

areshow

nin

bold.

PoolingM

ethodsp

AB

FS-C

IEL

AB

S-DE

LC

SSIMPL

CC

POI

RC

PLC

CPO

IR

CPL

CC

POI

RC

PLC

CPO

IR

CPL

CC

POI

RC

Average

-0.080%

0.11-0.47

0%-0.35

-0.4710%

-0.520.69

80%0.99

0.6370%

0.98

Minkow

ski

18-0.29

0%-0.04

-0.520%

-0.35-0.53

10%-0.52

0.6180%

0.970.61

70%0.97

14-0.27

0%-0.04

-0.520%

-0.35-0.52

10%-0.52

0.6280%

0.970.62

70%0.97

12-0.22

0%-0.04

-0.500%

-0.35-0.50

10%-0.52

0.6580%

0.990.63

70%0.97

20.19

20%0.59

-0.3810%

-0.35-0.41

10%-0.48

0.7380%

0.980.63

80%0.95

40.56

60%0.88

-0.1320%

-0.18-0.29

10%-0.29

0.7680%

0.980.61

70%0.86

80.31

30%0.48

-0.070%

0.06-0.11

10%0.04

0.7070%

0.860.54

60%0.59

Monotonic

Function

180.01

0%0.34

-0.450%

-0.36-0.45

10%-0.52

0.7180%

0.990.63

70%0.97

140.10

10%0.45

-0.430%

-0.35-0.44

10%-0.52

0.7180%

0.990.62

70%0.97

120.26

20%0.64

-0.3810%

-0.35-0.41

10%-0.48

0.7280%

0.970.58

70%0.97

10.51

50%0.84

-0.2110%

-0.17-0.35

10%-0.31

0.7280%

0.980.07

30%-0.10

20.77

80%0.98

0.1230%

0.30-0.19

10%-0.13

0.7270%

0.96-0.08

10%-0.70

40.33

30%0.42

0.3640%

0.620.14

10%0.26

0.3530%

0.46-0.19

10%-0.99

8-0.07

10%-0.20

-0.0520%

-0.410.14

40%0.17

-0.2210%

-0.65-0.26

0%-0.97

IC-0.07

0%0.11

-0.460%

-0.36-0.47

10%-0.52

0.6880%

0.980.63

70%0.97

IG0.03

10%0.16

-0.2810%

-0.37-0.36

10%-0.43

0.7870%

0.960.55

60%0.83

NB

-0.030%

0.27-0.41

10%-0.41

-0.4610%

-0.580.72

80%1.00

0.6370%

0.98

208

Page 225: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

1/8 1/4 1/2 1 2 4 8

PLC

C

p value

Lightness - Monotonic Func�on pooling

ABF

S-CIELAB

S-DEE

∆LC

SSIM

Figure 12.5: PLCC performance of Monotonic Function pooling for the lightness attribute.

12.3.4 Contrast

12.3.4.1 Test set 1

For SSIM, Minkowski pooling with p = 18 and Monotonic Function pooling with p = 1 per-

form better than the average in terms of all three performance measures (Table 12.9). Thisis possibly because regions of larger contrast difference are more perceivable by human ob-servers. For the ΔLC metric, Minkowski pooling with p = 1

8 and p = 14 slightly improve

PLCC and RC at the cost of a lower POI. This means that two or three images get lowercorrelation, but the overall correlation increases. The result for WLF is similar to that forΔLC, but WLF seems to be less dependent on the parameters and pooling method for PLCCand POI than ΔLC and SSIM. However, the differences for the three metrics between the bestpooling method and other methods are small, and therefore the results for the first data setindicate that there is no pooling method that is overall better than others.

12.3.4.2 Test set 2

For the contrast CPQA for the second test set, SSIM and ΔLC have a good performance foraverage pooling, while WLF shows poor results (Table 12.10). Some improvement for SSIMand ΔLC can be found by changing the pooling method and the parameters, where Minkowskipooling provides more stable results for different p values than Monotonic Function. For theWLF a large increase in performance can be found compared to the average when usingMinkowski pooling, and they are more or less stable over different p values as seen in firstdata set (Table 12.9). However, Monotonic Function pooling does not give acceptable resultsas in the first data set (Table 12.9), which might indicate that pooling is image/database de-pendent. In general Minkowski pooling gives the highest performance of the pooling methodsregardless of the IQ metric for the second dataset.

209

Page 226: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table 12.9: Performance of the spatial pooling methods for the contrast CPQA on Test Set 1.SSIM, ΔLC and WLF are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p SSIM ΔLC WLFPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.32 50% 0.86 0.34 58% 0.97 0.38 63% 0.57

Minkowski

18 0.40 58% 1.00 0.37 42% 1.00 0.43 58% 0.6914 0.36 50% 0.99 0.38 50% 1.00 0.44 58% 0.6912 0.33 46% 0.99 0.37 50% 0.98 0.44 58% 0.692 0.22 46% 0.99 0.20 50% 0.86 0.27 54% 0.534 0.16 42% 0.26 0.20 50% 0.78 0.33 54% 0.808 0.07 33% -0.26 0.16 42% 0.76 0.34 54% 0.94

MonotonicFunction

18 0.31 54% 0.81 0.23 54% 0.99 0.35 50% 0.4014 0.32 54% 0.86 0.19 50% 0.98 0.33 50% 0.4012 0.36 58% 0.91 0.12 46% 0.88 0.23 50% 0.571 0.40 58% 0.94 0.19 50% 0.81 0.20 50% 0.762 0.17 50% 0.93 0.18 46% 0.79 0.33 54% 0.894 0.05 38% 1.00 0.11 33% 0.81 0.36 50% 0.948 0.13 38% 1.00 0.06 29% 0.60 0.24 54% 1.00

IC 0.34 54% 0.97 0.24 54% 0.98 0.37 54% 0.51IG 0.21 46% 0.59 0.30 58% 0.99 0.17 46% 0.73NB 0.33 54% 0.90 0.33 58% 0.92 0.28 54% 0.69

Table 12.10: Performance of the spatial pooling methods for the contrast CPQA on Test Set 2.SSIM, ΔLC and WLF are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p SSIM ΔLC WLFPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.65 70% 0.99 0.57 60% 0.93 -0.32 0% -0.31

Minkowski

18 0.60 80% 0.95 0.56 70% 0.96 0.54 60% 0.8714 0.61 80% 0.95 0.56 70% 0.96 0.54 60% 0.8712 0.63 70% 0.95 0.57 70% 0.92 0.55 60% 0.812 0.69 70% 1.00 0.56 70% 0.88 0.55 60% 0.714 0.71 80% 0.97 0.54 70% 0.85 0.54 50% 0.708 0.77 70% 0.77 0.34 50% 0.68 0.28 50% 0.42

MonotonicFunction

18 0.64 70% 0.97 0.56 60% 0.97 -0.35 0% -0.4414 0.62 80% 0.95 0.55 60% 0.97 -0.38 0% -0.4412 0.57 80% 0.95 0.53 70% 0.86 -0.42 0% -0.621 0.03 40% -0.30 0.51 70% 0.85 -0.39 10% -0.852 -0.12 20% -0.73 0.48 60% 0.85 -0.36 10% -0.954 -0.27 10% -0.90 0.16 20% 0.19 -0.10 10% -0.978 -0.37 0% -1.00 -0.27 10% -0.85 0.19 30% 0.16

IC 0.64 80% 0.97 0.54 60% 0.92 -0.23 0% -0.19IG 0.68 90% 0.96 0.51 60% 0.86 -0.57 10% -0.92NB 0.70 80% 0.92 0.62 60% 0.95 -0.44 0% -0.28

210

Page 227: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

12.3.5 Artifacts

12.3.5.1 Test set 1

SSIM combined with IG pooling gives the best POI for artifacts with 50% of the imagesabove 0.6, but with a low PLCC of 0.28 (Table 12.11). ABF, S-DEE, and S-CIELAB givevery poor results regardless of what pooling methods we use, with a PLCC around 0 andalways a negative RC. For the WLF metric, Monotonic Function pooling with p = 1

8 performsthe best in terms of PLCC and RC, but the results are only marginally better than SSIM. ForΔLC, Monotonic Function pooling gives better results for PLCC and RC than average pooling,however the POI remains low, indicating that the pooling methods work well for only a fewimages in the dataset.

Table 12.11: Performance of the spatial pooling methods for the artifact CPQA on Test Set 1.All the six metrics are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p ABF S-CIELAB S-DEEPLCC POI RC PLCC POI RC PLCC POI RC

Average -0.04 21% -0.81 -0.01 33% -0.72 -0.03 25% -0.65

Minkowski

18 -0.02 25% -0.84 -0.02 33% -0.71 0.01 29% -0.7014 -0.02 25% -0.84 -0.02 33% -0.71 0.02 29% -0.7012 -0.02 25% -0.80 -0.02 29% -0.71 0.01 25% -0.642 -0.08 21% -0.81 0.01 33% -0.67 -0.03 21% -0.614 -0.09 17% -0.86 0.01 29% -0.76 -0.04 25% -0.718 0.05 38% -0.94 -0.07 25% -0.80 -0.11 17% -0.96

MonotonicFunction

18 -0.04 21% -0.81 -0.01 33% -0.72 -0.03 25% -0.6614 -0.06 21% -0.81 -0.00 33% -0.72 -0.03 25% -0.6112 -0.08 21% -0.80 0.02 33% -0.67 -0.03 21% -0.661 -0.09 21% -0.92 0.13 42% -0.76 -0.03 17% -0.762 -0.07 17% -0.99 -0.09 25% -0.54 0.02 25% -0.824 -0.00 29% -0.76 0.01 29% -0.96 -0.15 21% -0.788 -0.00 29% -0.76 0.04 38% -0.92 0.02 29% -0.73

IC -0.04 25% -0.81 -0.02 33% -0.67 -0.02 25% -0.66IG -0.11 29% -0.70 -0.05 29% -0.76 -0.11 21% -0.72NB -0.12 25% -0.84 -0.02 29% -0.69 -0.09 17% -0.72

12.3.5.2 Test set 2

For the artifact CPQA for the second data set, SSIM and ΔLC perform quite well with averagepooling. However, the other four metrics show poor performance with average pooling. ForABF, S-CIELAB, and S-DEE Monotonic Function pooling with a large p gives much betterresults, which could be because severe artifacts are weighted higher. ABF combined withMonotonic Function pooling (p = 2) greatly improves the performance of this metric, and withthis setting it shows comparable results to the best settings for SSIM. Minkowski pooling (p =18 , 1

4) makes the very poor results of WLF to very good ones in terms all three performancemeasures. However, increasing the p value for WLF decreases the performance, showing the

211

Page 228: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table 12.12: Performance of the spatial pooling methods for the artifact CPQA on Test Set 1.All the six metrics are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p SSIM WLF ΔLCPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.09 29% 0.44 0.28 29% 0.77 -0.02 25% 0.18

Minkowski

18 0.06 29% 0.00 0.23 33% 0.66 0.04 29% -0.1314 0.10 33% 0.07 0.23 38% 0.66 0.03 25% -0.0512 0.06 29% 0.06 0.23 38% 0.66 0.01 25% 0.102 0.08 38% -0.21 0.17 29% 0.80 -0.04 25% 0.444 0.06 38% -0.98 0.08 33% 0.53 0.02 29% 0.558 0.06 42% -0.94 0.03 29% 0.25 0.11 33% 0.58

MonotonicFunction

18 0.12 29% 0.52 0.29 33% 0.88 -0.06 21% 0.0414 0.13 25% 0.44 0.25 29% 0.88 -0.08 21% 0.1112 0.20 42% 0.33 0.12 29% 0.77 -0.07 25% 0.401 0.20 33% 0.26 -0.00 25% 0.58 -0.07 25% 0.522 0.04 33% 0.29 0.02 29% 0.38 0.04 38% 0.554 0.11 38% -0.06 0.01 25% 0.25 0.14 42% 0.528 -0.01 33% -0.01 -0.03 38% -0.00 0.14 25% 0.74

IC 0.06 29% 0.18 0.26 33% 0.81 -0.05 21% 0.11IG 0.28 50% 0.76 0.21 33% 0.62 0.02 21% 0.04NB 0.08 33% 0.36 0.10 25% 0.67 0.02 25% 0.31

importance of a correct parameter. ΔLC gave good results for average pooling, and by using alow p value with Minkowski pooling the POI can be increased from 70% with average poolingto 90%, but the PLCC and RC are still similar to high values for average pooling.

12.3.6 Overall observations

In general for the methods using parameters the selection of the parameter is important forthe performance, and should be set with care. The results also indicate that pooling is metricdependent, and also that the parameters for a pooling strategy is metric dependent.

For sharpness, SSIM and ΔLC perform quite well with Average pooling, but a slight in-crease in performance can be achieved by using more advanced pooling strategies. For thecolor attribute the performance of the metrics can also be improved by using either Minkowskior Monotonic Function pooling compared to average pooling. However, the selection of the pvalues are important for obtaining the best performance as seen for Monotonic Function pool-ing in Figure 12.3. For the lightness attribute the results differ between the databases but alsobetween the metrics, indicating that pooling can be both image and metric dependent. Forthe contrast attribute the second dataset show more variation between the pooling methodsand between parameters, and the method together with the parameter is important for highperformance. For artifact, the saliency based pooling improves the performance of SSIM onboth data sets. For other metrics, improved performance can be found, such as for the ABFmetric using Monotonic Function on the second test set.

212

Page 229: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table 12.13: Performance of the spatial pooling methods for the artifact CPQA on Test Set 2.All the six metrics are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p ABF S-CIELAB S-DEEPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.06 20% 0.12 -0.31 10% -0.26 -0.45 0% -0.45

Minkowski

18 -0.11 10% 0.01 -0.34 10% -0.26 -0.51 0% -0.4514 -0.09 10% 0.01 -0.34 10% -0.26 -0.51 0% -0.4512 -0.05 10% 0.01 -0.33 10% -0.26 -0.49 0% -0.452 0.32 20% 0.59 -0.26 10% -0.31 -0.37 0% -0.404 0.65 50% 0.82 -0.10 10% -0.18 -0.23 0% -0.228 0.22 20% 0.34 -0.02 0% -0.17 -0.06 10% 0.11

MonotonicFunction

18 0.14 20% 0.35 -0.30 10% -0.27 -0.42 0% -0.4514 0.23 20% 0.46 -0.29 10% -0.31 -0.41 0% -0.4512 0.38 20% 0.61 -0.26 10% -0.31 -0.36 0% -0.401 0.62 60% 0.81 -0.16 10% -0.14 -0.27 0% -0.222 0.74 60% 0.91 0.16 10% 0.21 -0.09 20% -0.044 0.22 20% 0.30 0.34 30% 0.41 0.20 20% 0.358 -0.18 10% -0.34 -0.10 0% -0.62 0.15 20% 0.29

IC 0.07 20% 0.12 -0.31 10% -0.27 -0.46 0% -0.45IG 0.09 20% 0.24 -0.23 0% -0.29 -0.32 0% -0.34NB 0.11 20% 0.34 -0.32 0% -0.34 -0.45 0% -0.50

12.4 Summary

In this chapter pooling methods have been investigated with the goal of improving the per-formance of IQ metrics. Many different pooling methods have been evaluated for differentCPQAs, and their correlation with the percept has been used as a performance measure. Theoverall results indicate that the parameters for the pooling methods are important, but alsothat pooling is metric dependent. Improved performance can be found for specific metrics forspecific attributes.

213

Page 230: Doctoral Dissertation by Marius Pedersen

IMPROVING PERFORMANCE OF METRICS BY POOLING

Table 12.14: Performance of the spatial pooling methods for the artifact CPQA on Test Set 2.All the six metrics are selected for the evaluation. The best results for each performancemeasure and metric are shown in bold.

PoolingMethods

p SSIM WLF ΔLCPLCC POI RC PLCC POI RC PLCC POI RC

Average 0.60 70% 1.00 -0.14 30% -0.15 0.72 70% 0.94

Minkowski

18 0.55 50% 0.99 0.79 90% 0.94 0.70 90% 0.9514 0.56 60% 0.99 0.80 90% 0.94 0.70 90% 0.9512 0.58 60% 0.99 0.80 90% 0.89 0.71 80% 0.942 0.63 70% 0.98 0.77 80% 0.82 0.71 50% 0.914 0.63 70% 0.92 0.62 50% 0.81 0.73 60% 0.908 0.60 70% 0.65 0.25 20% 0.56 0.58 70% 0.79

MonotonicFunction

18 0.58 60% 1.00 -0.19 20% -0.28 0.71 60% 0.9714 0.56 50% 0.99 -0.23 20% -0.28 0.70 50% 0.9712 0.51 50% 0.99 -0.30 10% -0.47 0.69 50% 0.881 0.03 20% -0.14 -0.46 10% -0.75 0.69 50% 0.902 -0.17 0% -0.64 -0.49 10% -0.89 0.69 70% 0.924 -0.13 10% -0.92 -0.29 0% -0.99 0.19 20% 0.358 -0.21 10% -0.99 0.15 10% 0.33 -0.36 0% -0.75

IC 0.59 60% 1.00 -0.05 30% -0.02 0.68 70% 0.93IG 0.63 60% 0.89 -0.75 0% -0.97 0.76 70% 0.94NB 0.71 70% 0.98 -0.18 0% -0.11 0.75 80% 0.96

214

Page 231: Doctoral Dissertation by Marius Pedersen

13 A NEW METRIC FOR THE COLOR

QUALITY ATTRIBUTE

The evaluation carried out in the previous chapters (11 and 12) has shown that none of thecurrent IQ metrics are suitable to measure the color CPQA. In this chapter we proposed a newmetric designed for the color CPQA, where the focus is on the human visual system models(spatial filtering) and the quality calculation (third and fourth stage of Figure 13.1).

Since the introduction of Total Variation (TV) in image processing in 1992 by Rudin etal. [376], TV has become increasingly popular. The use of variational image processing hasbeen extended to several areas of computer vision, such as inpainting, segmentation, anddeblurring. Originally developed for grayscale images, TV has been extended to color imagesby Blomgreen and Chan [36]. Furthermore, in the last decades several efforts have been donefor developing fast and robust TV solvers such as the one by Chen and Tai [73]. We refer thereader to Chan and Shen [68] for a detailed overview of variational image processing methods.In this chapter we introduce TV in the field of IQ metrics.

Human observers are sensitive to various frequencies of a visual stimuli; the ContrastSensitivity Function (CSF) tells us how sensitive. If the frequency of visual stimuli is too high,we will not be able to differentiate between stimuli patterns (Figure 13.2). The use of CSFshas been popular in image quality metrics, such as the S-CIELAB [499] (Section 4.2.2.1) andS-DEE [412] (Section 4.2.2.2). In these metrics, the CSFs are commonly used to modulatefrequencies that are less perceptible [224]. The common way to do this is to use convolutionkernels to ”blur” the spatial frequencies that observers cannot perceive [499]. This method isfast, but does not result in the most precise filtering of the image [223]. Recent studies haveshown that contrast is one of the most relevant perceptual and IQ factors [346]. The historyof contrast is one century long, and measuring perceived contrast is not a trivial task [410].An important milestone was given by Peli [362] in 1990, who defined a local band-limited

Color space

transforms

Human visual

system models

Quality

calcula�on Pooling

Fewer

Many

Quality

value

Original and

reproduc�on

Original and

reproduc�on

Color space

transforms Pooling

Fewer

Many

Quality

value

Figure 13.1: Workflow of image quality metrics. In this chapter we focus on the human visualsystem models and the quality calculation.

215

Page 232: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

Figure 13.2: Contrast sensitivity threshold indicated by the curved line, if the contrastat a given spatial frequency is above the line it is invisible, while if below the line itis visible. Figure reproduced from http://www.cns.nyu.edu/˜david/courses/perception/lecturenotes/channels/channels.html, visited 12/07/11.

contrast for complex images. This work will be explained in details later in the chapter as itwill be relevant for our proposal of a new IQ metric for the color attribute.

13.1 Background

Peli [362] introduced a method to simulate the human visual system, where contrast at eachpoint in an image is calculated separately to account for variations across the image, and sincecontrast sensitivity depends on frequency, contrast is also calculated for different frequencybands.

Peli [362] proposes the idea of a pyramidal image-contrast structure where for each fre-quency band, the contrast is defined as the ratio of the bandpass-filtered image at that fre-quency to the low-pass image filtered to an octave below the same frequency (local luminancemean).

To define local band-limited contrast for a complex image, he obtains a band-limited ver-sion of the image in the frequency domain A(u,v):

A(u,v) ≡ A(r,θ) ≡ F(r,θ)G(r), (13.1)

where u and v are the respective horizontal and vertical spatial frequency coordinates, G(r)is a band-pass filter, r and θ represent the respective polar spatial frequency coordinates:r =

√u2 + v2, θ = tan−1 (u/v), and F(r,θ) is the Fourier transform of the image I(x,y).

In the spatial domain the filtered image a(x,y) can be represented similarly, that is, as:

a(x,y) = I(x,y)∗g(x,y), (13.2)

where ∗ is the convolution, and g(x,y) is the inverse Fourier transform of G(r).

216

Page 233: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

Figure 13.3: Original on the left, simulated filtered image in the center, and CSF based filteredimage on the right. The images simulate a distance of 200 cm.

In Peli’s approach of measuring local contrast, the pyramid is obtained as follows:

Ai(u,v) ≡ Ai(r,θ) ≡ F(r,θ)Gi(r), (13.3)

where Gi(r) is a cosine log filter centered at frequency of 2icycles/picture, expressed as:

Gi(r) =12

(1+ cos(π log2 r−πi)) . (13.4)

The resulting contrast at the band of spatial frequencies can be represented as a two-dimensional array ci(x,y):

ci(x,y) =ai(x,y)li(x,y)

, (13.5)

where ai(x,y) is the corresponding local luminance mean image and li(x,y) is a low-pass-filtered version of the image containing all energy below the band.

This filtering differs from other types of filtering because suprathreshold features retaincontrast and are not washed out [362] as seen in Figure 13.3.

13.2 The new color image quality metric

We propose a new color IQ metric based on contrast filtering and TV. First the original IO andreproduction IR are converted into the CIEXY Z color space. For each channel independently,the contrast of each pixel is calculated as described in Equation 13.5 in Section 13.1. Thecontrast c of each pixel is then compared against the contrast sensitivity threshold (T ) forthe corresponding channel for each band. If the contrast is suprathreshold the informationis perceptible and should be kept, if the contrast is subthreshold the information is discarded(Contrast threshold axis of Figure 13.2). The contrast of each pixel is calculated for each bandLi(x,y):

Li(x,y) =

{c(x,y) if c(x,y) > T

0 else. (13.6)

217

Page 234: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

The final filtered image L f is then the sum over the n different bands:

L f (x,y) =n

∑i=1

Li(x,y). (13.7)

For the luminance contrast sensitivity thresholds we use the same as Peli [363] while forthe chrominance thresholds we use the ones from Johnson and Fairchild [223] (also the sameas used for SHAME in Chapter 5), which are defined as:

CSFchroma(p) = a1 · e−b1·pc1 +a2 · e−b2·pc2 , (13.8)

where different parameters for a1, a2, b1, b2, c1, and c2 have been used as seen in Table 13.1.Since the CIEXY Z color space is not orthogonal, i.e. the X and Z channels contain lumi-

nance information, we separate these channels into a color part and a luminance part, filteredwith their respective contrast sensitivity thresholds. To obtain the luminance bandpass in-formation in the color channel (XBL), the lowpass information in the color channel (XL) isdivided by the lowpass information in the luminance channel (YL), and further multiplied withthe bandpass information in the luminance channel (YB).

XBL =XL

YLYB. (13.9)

The color information in the color channel (XBC) is found by subtracting the luminance band-pass information in the color channel (CBL) from the bandpass information in the same colorchannel (XB).

XBC = XB −XBL. (13.10)

After the filtering, the original and reproduction are converted to the log-compressed OSA-UCS color space as proposed by Oleari et al. [330]. Euclidean color differences in the OSA-UCS color space are shown to correlate well with perceived differences [335].

The new Total Variation of Difference (TVD) metric, given the original contrast filteredimage LO and its filtered reproduction LR, is defined as follows:

TV D =

√√√√∑j

(∫Ω| ∇LO j −∇LR j | dA

)2

∫Ω

√∑

j(LO j −LR j)

2dA, (13.11)

where√

∑ j(∫

Ω | ∇LO j −∇LR j | dA)2 is the TV term, while λ∫

Ω

√∑ j(LO j −LR j)

2dA is the

Color Difference (CD) term. ∇ is the gradient of the color channel j, Ω represent the im-age domain, and λ is the weighting parameter for the CD term following the notation usedby Blomgren and Chan [36]. The TV term is similar to the Color TV defined by Blomgrenand Chan [36], except that we take the gradient of the difference between the original andreproduction (∇LO j −∇LR j) instead of just the gradient of the image (∇LO j). The CD termis the Euclidean color difference, and is identical to the ΔEE from Oleari et al. [329] (Sec-tion 4.2.1.3) since we calculate the difference in the OSA-UCS color space.

We will also investigate other methods to reduce the number of IQ values into a sin-

218

Page 235: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

Table 13.1: The parameters for the contrast sensitivity thresholds for the chrominance chan-nels.

Parameter X channel Z channela1 109.14130 7.032845b1 -0.00038 -0.000004c1 3.42436 4.258205a2 93.59711 40.690950b2 -0.00367 -0.103909c2 2.16771 1.648658

gle number representing quality, so called pooling strategies (as introduced in Chapter 12).For the TV term (Equation 13.11) we will replace the standard outer norm (L2) over thecolor channels with the L1 norm, minimum, median and maximum. For the CD term (Equa-tion 13.11) we will replace the standard outer norm (average) over the image space withMinkowski (M) [463] (Equation 12.3), Monotonic Function (MF) [463] (Equation 12.7), in-formation content [463], and two different information content based pooling methods basedon saliency. The saliency weighting maps are used with Equation 12.1, where the saliencymodels from Achanta et al. [1] and Seo and Milanfar [389] are tested. We will also test 100 λvalues from 0 to 5, with equal steps.

13.3 Evaluation

Evaluation of the proposed metric is important to quantify its performance and agreement withthe human observers. We will evaluate the performance of the TVD metric by a comparisonwith the results of observers. Two different data sets have been selected for the evaluation.Comparison against the existing metrics evaluated in Chapter 11 will show the potential ofthe new metric.

13.3.1 Test data sets

The first test set [347] is the same as presented in Section 10.4.2, and contains 24 of the 25reference images in Figure 10.7. The images were printed on an Oce Colorwave 600 CMYKwide format printer on Oce Red Label (LFM054) plain uncoated paper using three differentrendering intents: perceptual, relative colorimetric, and relative colorimetric with black pointcompensation. Each printed image were judged by 15 observers. For details we refer thereader to Section 10.4.2 or Pedersen et al. [347].

The second test consists of ten images (Figure 11.6) is the same as in Section 11.3.2.The images were printed by a HP DesignJet 10ps printer on Stora Enso Multicopy Originalplain paper. The images were printed using four different modes: the best print mode andthe perceptual intent, the best mode and relative colormetic intent, normal print mode and theperceptual intent, and the last with normal print mode and relative colorimetic intent. Tenobservers judged the images according to color quality, from which z-scores were calculated.For details we refer to Section 11.3.2 or Pedersen et al. [359].

In order to apply objective IQ metrics to these printed images, the images are scanned intodigital images and stored without compression using the framework proposed in Chapter 8.

219

Page 236: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

In both data sets the observers were asked to judge the color quality of the images, and theratings have been quantified as z-scores [116].

13.3.2 Evaluation procedure

Three state of the art IQ metrics have been chosen for comparison in the evaluation: S-CIELAB [499], S-DEE [412], and ABF [459]. We use three same evaluation methods tocompare the performance of different IQ metrics as in Section 12.2.4: Pearson Linear Corre-lation Coefficient (PLCC) is calculated for each image between the metric scores and subjec-tive z-scores, and the first measure of performance is the mean PLCC of the whole databasecalculated as the average of PLCCs of each image. The second measure is the Percentage OfImages (POI) with PLCC higher than 60%. The last measure is the Rank Correlation (RC),which is the PLCC correlation between objective rank order z-score and subjective z-score.For more information on this evaluation method we refer the reader to Section 6.3 or Pedersenand Hardeberg [350].

13.3.3 Results and discussion

We will show the results for the following configurations of the TVD metric; TV term with L1and L2 pooling, λ equal to 0, 1, and the best λ, and for the color term we will show the meanpooling together with the best pooling method. The results from the evaluation can be seenin Tables 13.2 and 13.3 for the first and second dataset. For the first test set a combinationof the TV and CD terms, where the L1 pooling for the TV term, λ = 4 and the MF poolingwith p = 4 gives the highest correlation with the perceived color quality. With a λ = 1 weobtain results similar to existing metrics. Average pooling has the highest PLCC for λ = 0,an increase in performance can be found with both Minkowski pooling (Figure 13.5) andMonotonic Function pooling (Figure 13.6. We can see that most pooling methods producebetter results than average pooling, and that in many cases using a λ = 0 will generate thea higher PLCC than using a λ > 0. However, there are p values together with λ values thatproduce higher PLCC values then λ = 0, such as p = 8 in Minkowski pooling (Figure 13.5)and Monotonic Function pooling with p = 8, p = 4, and p = 1. It is also noticeable that theTVD metric has less variable PLCC values for different λ values than Monotonic Functionpooling.

TV shows the results for the TV term (Equation 13.11), without the spatial filtering andthe CD term. We see that it has similar performance for PLCC and POI, but a higher RC.Nonetheless, the performance of the new metric is not great, most likely since the visualdifferences of the first test is small, making the task very difficult for IQ metrics.

For the second test set an equal weighting of the TV and CD term gives similar results tothe state of the art metrics. However, by reducing the importance or removing the CD term(λ = 0) the performance of the TVD greatly improves (Table 13.3). Average pooling showsthe highest PLCC for lambda = 0 (Figure 13.7). The same can also be seen for Minkowskipooling (Figure 13.8) and for Monotonic Function pooling (Figure 13.9). The most stablepooling method is Minkowski pooling with p = 1/8 over different λ values (Figure 13.8). TVwithout the spatial filtering and without the CD term (last line in Table 13.3), gives slightlylower performance, indicating that the spatial filtering adds value to the TVD metric.

220

Page 237: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

Table 13.2: Results for the first test set. The highest PLCC and POI is found with L1 poolingfor the TV term, λ = 4, and using the Monotonic Function (MF) pooling with a p = 4 for thecolor term.

Metric PLCC POI RCS-CIELAB −0.29 13% −0.95S-DEE −0.34 13% −0.92ABF −0.39 8% −0.99TVD (L1/λ = 1/mean) −0.31 8% −0.94TVD (L2/λ = 1/mean) −0.30 13% −0.44TVD (L1/λ = 0) −0.15 13% −1.00TVD (L2/λ = 0) −0.16 13% −0.95TVD (L1/λ = 4/MFp=4) 0.18 29% −0.93TV (L1) −0.26 21% −0.12

-0.40

-0.30

-0.20

-0.10

0.00

0.10

0.20

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Average

Average

Figure 13.4: Performance of the TVD metric with average pooling for varying λ values forthe first dataset.

Table 13.3: Results for the second test set. The best results are found with a λ = 0, and witha λ = 0.5 and using the Minkowski (M) pooling with a p = 1/8.

Metric PLCC POI RCS-CIELAB −0.27 0% −0.23S-DEE −0.42 0% −0.42ABF 0.07 0% 0.23TVD (L1/λ = 1/mean) −0.25 0% −0.15TVD (L2/λ = 1/mean) −0.31 0% −0.29TVD (L1/λ = 0) 0.59 70% 0.98TVD (L2/λ = 0) 0.56 60% 0.92TVD (L1/λ = 0.5/Mp=1/8) 0.59 70% 0.98TV (L1) 0.53 60% 0.92

221

Page 238: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

-0.40

-0.30

-0.20

-0.10

0.00

0.10

0.20

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Minkowski

1/2

1/4

1/8

2

4

8

Figure 13.5: Performance of the TVD metric with Minkowski pooling for varying λ values forthe first dataset.

-0.40

-0.30

-0.20

-0.10

0.00

0.10

0.20

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Monotonic Func�on

1

1/2

1/4

1/8

2

4

8

Figure 13.6: Performance of the TVD metric with Monotonic Function pooling for varying λvalues for the first dataset.

222

Page 239: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

-0.40

-0.20

0.00

0.20

0.40

0.60

0.80

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Average

Average

Figure 13.7: Performance of the TVD metric with average pooling for varying λ values forthe second dataset.

-0.40

-0.20

0.00

0.20

0.40

0.60

0.80

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Minkowski

1/8

1/4

1/2

2

4

8

Figure 13.8: Performance of the TVD metric with Minkowski pooling for varying λ values forthe second dataset.

-0.40

-0.20

0.00

0.20

0.40

0.60

0.80

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5

PLC

C

λ

Monotonic Func�on

1/8

1/4

1/2

1

2

4

8

Figure 13.9: Performance of the TVD metric with Monotonic Function pooling for varying λvalues for the second dataset.

223

Page 240: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

13.4 Investigation of image characteristics

We will carry out a similar analysis as that of Section 11.2.3 for the TVD metric. This will in-dicate if there is any relationship between the characteristics of the image and the performanceof the TVD metric.

13.4.1 Experiment

The experimental images (Figure 10.7) and results for TVD from the first test set are thebasis for our investigation. These images have been classified by two expert observers into aset of classes, being the same as for Section 11.2.3. We examine the dominant color IC, aswell as the colorfulness IC. For dominant color the classes were ”yes” and ”no”, while forcolorfulness and lightness the classes were ”high”, ”medium”, ”low”, and ”mixed”.

13.4.2 Dominant color

For the calculation we use the best parameters from Table 13.2; L1 norm for the TV term, andMF pooling for the CD term, and λ = 4. The observer data is based on the color attribute.Figure 13.10 shows the correlation values for the images in the two classes, when the metricvalues are compared to the color quality values. The mean correlation for the 11 imagesclassified as having a dominant color is −0.13 (standard deviation 0.64), while for the 13images classified as not having a dominant color the correlation is 0.43 (standard deviation0.59). This is the same tendency as for S-CIELAB in Section 11.2.3.3.

We also test the result for the same metric with only the CD term (only color differences),using a mean pooling (Figure 13.10(b)). For the images classified as having a dominant colorwe obtain a mean correlation of −0.12 (standard deviation 0.68), and for those classified asnot having a dominant color a correlation of −0.41 (standard deviation 0.77). The meancorrelation for all images is −0.28 (standard deviation 0.60). These results indicate that aseparation of images when evaluating color quality might be useful, but only in terms ofcolor differences. This is also in agreement with the hypothesis that when a dominant coloris present preservation of this color is important. However, based on the small number ofsamples further work is needed to test the hypothesis.

13.4.3 Colorfulness

We also investigate the performance of the TVD metric for the different classes of colorful-ness, using the results from the observers for the color CPQA. From Figure 13.11(a) we cansee that the TVD metric performs well for the four images classified as having low colorful-ness, and poor for those with mixed colorfulness. However, it is important to notice that thesetwo classes are also those with the fewest number of images. The mean correlation values arefound in Table 13.4.

We have also investigated the color difference part of the TVD metric, as for dominantcolor, and the results are shown in Figure 13.11(b). We have as for the TVD used the responsefrom the observers for the color CPQA. The metric performs well for the two images classifiedas having mixed colorfulness, opposite of the TVD metric. The mean correlation values arefound in Table 13.5.

224

Page 241: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Dominant ColorNo dominant color

(a) TVD metric.

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Dominant ColorNo dominant color

(b) TVD metric with only the CD term.

Figure 13.10: Performance of the metrics for images classified as having a dominant color(red) and images classified as not having a dominant color (blue).

Colorfulness Correlation Standard deviationLow 0.85 0.17

Medium 0.09 0.58High 0.15 0.72

Mixed -0.64 0.40

Table 13.4: Mean correlation for the TVD metric for images classified according to colorful-ness.

225

Page 242: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Low colorfulnessMedium colorfulnessHigh colorfulnessMixed colorfulness

(a) TVD.

1 24−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Images

Cor

rela

tion

Low colorfulnessMedium colorfulnessHigh colorfulnessMixed colorfulness

(b) TVD with only the CD term.

Figure 13.11: Colorfulness classified images and the performance of the metrics.

Colorfulness Correlation Standard deviationLow -0.80 0.08

Medium -0.38 0.53High -0.17 0.79

Mixed 0.85 0.22

Table 13.5: Mean correlation for the TVD metric with only the CD term for images classifiedaccording to colorfulness.

226

Page 243: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

13.5 Summary

A new IQ metric, Total Variation of Difference (TVD), designed for the color CPQA has beenproposed in this chapter. The new metric incorporates a spatial filtering method based on thework by Peli [362] and the quality calculation is based on total variation. By a careful animproved selection of the parameters of the metric an increase in performance is obtained, interms of correlation with human observations, as compared to state of the art metrics.

227

Page 244: Doctoral Dissertation by Marius Pedersen

A NEW METRIC FOR THE COLOR QUALITY ATTRIBUTE

228

Page 245: Doctoral Dissertation by Marius Pedersen

14 QUALITY ASSISTANT - TOOL FOR

EVALUATING PRINT QUALITY

In this chapter we combine the work carried out in the previous chapters to develop a qualitytoolbox for the evaluation of print quality. With the toolbox, named Quality Assistant, one isable to quantify the quality of different quality attributes for a given printer using IQ metrics,and compare them against other printers. Such tools have several potential uses as pointedout by Hardeberg and Skarsbø [170] and Hardeberg [168]: trade-off analysis of speed andimplementation cost versus IQ in algorithm development, benchmarking printers, and doc-umentation of IQ improvements resulting from efforts spent on optimization of technologyparameters. Tool also incorporates more advanced functions, allowing the user for exampleto see where in the color space quality issues occur or in which regions problems occur. Thisinformation can be used to optimize quality, or in the development of new algorithms (forexample halftoning or gamut mapping), or printing techniques.

14.1 Pre-metric operations

This section presents the operations needed before IQ metrics can be applied.

14.1.1 Padding the images

The first step to use the Quality Assistant is to prepare the images for printing. This is done byadding registration points outside the printed image. These points are black squares placed asmall distance from the corner of the image (Figure 14.1). These control points are used in theregistration of the scanned image in order to have a match to the digital original, as explainedin Chapter 8. After the images have been padded with control points, they are printed on agiven device.

14.1.2 Scanning

After the image has been printed it needs to be scanned in order to obtain a digital repre-sentation. To ensure a correct representation of colors a profile for the scanner needs to begenerated. A profile for each media and print mode should be created since the profile is de-pendent on the equipment [259, 260, 334]. The profile is built on a TC 3.5 CMYK test target(Figure 14.2 and Section 3.1.4.4) from GretagMachbeth. This contains a total of 432 patches

229

Page 246: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Figure 14.1: Illustration of control points. Outside each corner of the image a black square isadded to assist in the registration.

covering a wide range of CMYK combinations. This test target is printed on the same printeras the other images, and then measured using a measurement device (for example a X-riteGretagMacbeth Eye-One IO). Then the printed test target is scanned, and compared againstthe measured values of the test target.

A round trip test should be carried out to measure the accuracy of the profile as suggestedby Sharma [393], as done in Chapter 10.4.2.2. We have followed the same procedure asSharma [393]: the raw scan of the test target is opened in Adobe Photoshop. The scannerprofile is selected (using Image→Mode→Assign Profile). Further, the images are processedto LAB using Image→Mode→Convert to Profile where the destination space should be LABColor. Absolute colorimetric is used as the rendering intent, and the Color ManagementModule (CMM) is ACE. This results in a LAB image, and for each patch in the test target thecolor difference can be calculated. ΔE∗

ab (Section 4.2.1.2) is computed between the measuredvalues and the values in the processed image. For further details on this method we refer thereader to Sharma [393] and Sharma and Fleming [394]. Standard statistical values as meanand maximum ΔE∗

ab can be used as an indication of the accuracy of the profile.

14.1.3 Registration

When the images have been scanned the next step is to perform registration as described inChapter 8. Registration is needed since the IQ metrics require the reproductions to be of thesame size and orientation as the original. First the control points added prior to printing arefound by a simple algorithm in both the original and the reproduction. Then the transformationbetween the points of the original and the points of the reproduction is found using an affinetransformation. Further, the transformation is applied to the scanned image with bilinearinterpolation. For more information on the registration we refer the reader to Chapter 8.

14.1.4 Registration validation

To ensure that the error introduced in the registration is minimal, validation of the registrationis performed. Even though there was no clear indication that the angle at which the image is

230

Page 247: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Figure 14.2: TC3.5 CMYK test target used for scanner characterization.

scanned influences the metric values (Section 8.2.4.4 and Figure 8.8) it is recommended thatthe angle is kept to a minimum. The rotation performed in the registration is calculated, andimages placed more than one confidence interval (95%) from the mean rotation angle (of allregistered images) should be rescanned with better precision. This is not a strict validation,but gives an indication of outliers among the scanned images.

14.2 Measuring attributes using image quality metrics

We will measure the IQ of different CPQAs using IQ metrics. Based on interaction with Ocewe focus on specific attributes, namely sharpness, two artifacts (noise and banding), and color.

14.2.1 Sharpness

For the sharpness CPQA we measure both perceptual sharpness and computational sharpness.

14.2.1.1 Perceptual sharpness

Metric For the sharpness attribute the SSIM [458] metric is applied (Section 4.2.3.1). Thismetric has shown good correspondence with observers for the sharpness CPQA (Chapter 11).The original and registered image are both transformed to the CIEXYZ color space. The orig-inal by using standard color transformations from the sRGB original, and the reproduction by

231

Page 248: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

using the scanner profile (scanner RGB to CIELAB to CIEXYZ or scanner RGB to CIEXYZ,depending on the Profile Connection Space (PCS) of the profile). The SSIM is calculatedbased on the Y channel in the CIEXYZ color space, the chrominance channels are not includedsince it has been shown that sharpness is mainly influenced by luminance [89, 226, 471], andluminance accounts for almost all the variability. Spatial average pooling is used to obtain onevalue for each image from the quality map produced by SSIM (Section 12.3.1). This resultsin a map and a mean value for each image in the dataset. The overall value for all images isthe mean of the mean value of all images.

Test images A set of five images is proposed to be used with the SSIM metric to evaluateperceptual sharpness (Figure 14.3). These images contain the different characteristics neededfor a complete evaluation of sharpness, namely edges and details. All test images are 240 dpisaved as uncompressed tif in the sRGB color space unless stated otherwise.

Figure 14.3: Images for the evaluation of perceptual sharpness.

We also investigate sharpness for maps (Figure 14.4), since these usually contains finelines and sharp edges. These maps are 600 dpi. The evaluation is carried out for three dif-ferent regions in one map and two in the other maps, where the regions represent differentcharacteristics.

14.2.1.2 Computational sharpness

Metric Two specific metrics are used to measure computational sharpness: edge blurri-ness [197] and edge raggedness [197].

Edge blurriness Edge Blurriness (EB) is defined as the appearance of being hazy orindistinct in outline; a noticeable transition of blackness from background to character [197].For the blurriness metric the reflectance factors at the 30% (R30) and 90% (R90) boundariesare computed (Figure 14.5). The R30 and R90 positions are calculated and the average distancebetween the R30 and R90 contours (DIS30−90) is calculated. EB is then defined as [491]:

EB =DIS30−90√

LD, (14.1)

where LD is the average density of a solid printed area.

232

Page 249: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

(a) Full maps

(b) Regions

Figure 14.4: Maps for the evaluation of perceptual sharpness. Top figure shows the full maps,while the cropped regions used for the metrics are shown in the bottom figure. The three firstfrom the left are from the first map, while the remaining two regions on the right are from thesecond map.

Printed

RegionPaper

Region

The fitted R30 line

(red line)

The fitted R90 line

(blue line)

Figure 14.5: Illustration of edge blurriness. Figure reproduced from Lee et al. [262].

233

Page 250: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Edge raggedness Edge Raggedness (ER) is the appearance of geometric distortion ofan edge from its ideal position [197]. ER is calculated by computing the R60 and fitting a lineto this (R60) as seen in Figure 14.6. Then the standard deviation of the distance from the fittedline to the R60 boundary edge is calculated:

ER =

√∑N

i=1(Ri60 −R60)2

N −1, (14.2)

where Ri60 is the individual positions (i) for the N aperture scans from the fitted line [491].

Printed

Region

Paper

Region

R60 boundary

Figure 14.6: Illustration of edge raggedness. Cyan line indicates the R60 boundary. Figurereproduced from Lee et al. [262].

Test images For computational sharpness a specific test target (Figure 3.14 and Section 3.1.4.4)is used, which consists of three square printed solid regions oriented at 0, 8, and 24 degreesfrom vertical [491]. These targets are printed, then scanned before being evaluated by theraggedness and blurriness metrics [197].

14.2.2 Artifacts

The artifacts CPQA is divided into sub-attributes, since different artifacts have distinctiveproperties. We focus on noise and banding.

14.2.2.1 Noise

Noise is, as sharpness, divided into perceptual and computational aspects.

234

Page 251: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual noise

Metric The perceptual aspect is calculated using the Noise Quality Metric (NQM) [99].This metric has been shown to correlate well with perceived noise [35, 70, 99, 255, 266, 398].The NQM is built on Peli’s contrast pyramid, as explained in Chapter 13. However, somechanges are done in order to account for contrast masking and to incorporate informationabout viewing medium and ambient luminance. The SNR is taken for the two filtered images(original and reproduction). The metric is defined as

NQM(db) = 10log10

(∑x ∑y O2(x,y)

∑x ∑y(O(x,y)− I(x,y))2

), (14.3)

where O and I are the filtered original and reproduction, and x and y indicate the imagewidth and height. NQM is calculated for each color channel since noise in the color channelshave been shown to influence perceived noise [225]. We use the CIEXYZ color space as forsharpness, where the average of the three mean values for each color channel is used as ameasure of noise.

Test images For noise it is important to have images with low, medium, and high fre-quency content, since perceived noise is dependent on frequency [225]. Knowing that noisealso is dependent on color [225] having a range of the most important hues is important. Weuse the same images (Figure 14.3) as for the evaluation of perceptual sharpness to evaluateperceptual noise. These images fulfill our requirements for evaluation of noise.

Computational noise

Metric Standard deviation is often used to describe noise [174, 271], and we also applythis estimate here. Noise can be both in the luminance and in the chrominance channels, andboth types influence IQ [225]. To achieve a good indication of noise we use a test target tocover different locations in the color space. This results in calculating the standard deviationfor each patch for each channel of the color space of a test target. Since the CIELAB colorspace has been shown to give good results for noise measures [404] we also adopt this as theworking color space. We calculate Computational Noise (CN) as:

CNi =13

3

∑j=1

N

∑i=1

(√

(Xi j −μi j)2) (14.4)

where N is the number of patches, Xi is patch number i, μi is the reference value of patch i,and j indicate the color channel. This is then calculated for each patch, and the overall noiseis given as

CN =1N

N

∑i

CNi, (14.5)

where i is the patch number, N the number of patches, and CN the overall noise.

Test images For the computational noise calculation we use a TC3.5 CMYK test target(Figure 14.2), since the metric requires color patches. This test target contains a total of 432

235

Page 252: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

patches covering a wide range of CMYK combinations, enabling noise estimation in differentlocations of the color space.

14.2.2.2 Banding

Metric Banding is evaluated using the Total Variation of Difference (TVD) metric (Chap-ter 13). The image quality map from TVD (from the color difference term) is subject topost-processing to detect banding lines. Banding has been shown to be dependent on fre-quency [93], and this is taken into account due to the spatial filtering carried out in the TVDmetric. To measure banding the TVD color difference map is summed to obtain a 1 dimen-sional representation horizontally and vertically, based on the assumption that banding occursperiodically. On each of these we look for re-occurring differences by using autocorrelation,which is the cross-correlation of the signal with itself. Autocorrelation is a common statisticsto find repeating patterns, and has also been used for estimation of banding [105, 371]. Weuse the sum of the mean of the autocorrelation horizontally and vertically as the measure ofbanding, the higher correlation the more banding.

Test images Two different types of test images are used for banding: pictorial images andtest targets. For the pictorial images we select images with uniform regions, since banding ismost visible in these regions [371], but also images with some regions with higher frequenciessince banding is dependent on frequency [93]. The pictorial images for banding are shown inFigure 14.7.

Figure 14.7: Pictorial images for the evaluation of banding.

We have selected two different test targets, a grayscale gradient (Figure 14.8(a)) and theGranger rainbow (Figure 14.8(b)).

Figure 14.8: Test targets for the evaluation of banding.

236

Page 253: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

14.2.3 Color

Metric The color attribute is calculated using the TVD metric, as defined in Chapter 13.Given an original filtered imaged LO and its filtered reproduction LR, TVD is calculated as

TV D =

√√√√∑j

(∫Ω| ∇LO j −∇LR j | dA

)2

∫Ω

√∑

j(LO j −LR j)

2dA, (14.6)

where√

∑i(∫

Ω | ∇LO j −∇LR j | dA)2 is the TV term, while λ∫

Ω

√∑i(LO j −LR j)

2dA is the

Color Difference (CD) term. ∇ is the gradient of the color channel j, Ω represents the imagedomain, and λ is the weighting parameter for the CD term, and j indicates the color channel.OSA-UCS is the used color space. λ is set to 0.5, and we use the L1 norm for the TV term, andMinkowski pooling with a p of 1/8, since these parameters gave a TVD metric well correlatedwith the perceived color quality in Chapter 13. It is also important to notice that the lightnesschannel is used here in addition to the color channels. The results shown here are based onboth lightness and color, but in the toolbox these results are also shown separately as suggestedin Chapter 10.

Test images Images for the color attribute should contain the most important hues, includesaturated colors, and skin tones. The images should also have regions with varying frequen-cies, i.e. having both smooth regions and detail-rich regions. Three images having theseproperties are shown in Figure 14.9.

Figure 14.9: Images for the evaluation of the perceptual color attribute.

14.3 Visualization of results

The Quality Assistant visualizes results differently depending on the attribute and the results.The visualizations create a dynamic tool adapted to be used in the field, and to provide asimple and intuitive way to understand the results.

14.3.1 Spider and bar plots

To visualize the results we mainly use spider plots (also known as radar chars, web charts,polar charts, or kiviat diagrams). It is a graphical method to display multivariate data (i.e.the results from different IQ metrics) as a two-dimensional chart of three or more variables,where each variable is represented on the axes starting from the same point (center). Thesespider plots have been used by researchers to visualize IQ results [97, 126, 128]. An example

237

Page 254: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

of a spider plot is shown in Figure 14.10. We use spider plots to visualize results with threeor more values. For data with one or two values we will use a standard histogram (bar plot)as seen in Figure 14.11.

0.25 0.5 0.75 1Image 1

0.25

0.5

0.75

1

Image 2

0.25

0.5

0.75

1Image 3

0.25

0.5

0.75

1

Image 4

0.25

0.5

0.75

1Image 5

Printer APrinter BPrinter C

Figure 14.10: Example of a spider plot, each axes indicates an image, and the colored linesindicate different printer. The higher the value the higher the quality.

14.3.2 Quality maps

Some IQ metrics provide IQ maps (Figure 14.12), where each pixel of the image is representedby a quality value. These quality maps are scaled and displayed to the full range of a colormap. When quality maps for the same image and same metric is compared they are displayedon the same scale for easy comparison.

14.3.2.1 Box plots

For metrics generating a quality map different statistics can be computed. A common methodin descriptive statistics to depict groups of numerical data is to use box plots (also known asa box-and-whisker diagram). A box plot (Figure 14.13) shows lower quartile (Q1), median(Q2), and upper quartile (Q3). The whiskers are lines extending from each end of the boxes toshow the extent of the rest of the data (1.5 times the inter quartile range). Notches around themedian indicates the 95% confidence interval (i.e. if the notches of two groups do not overlapthe median of the groups differ at a 5% significance level).

238

Page 255: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Printer A Printer B Printer C0

0.005

0.01

0.015

0.02

0.025

0.03

Printers

Q

ualit

y

Figure 14.11: Standard bar plot.

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Figure 14.12: Example of a quality map. The colorbar indicates quality, where a higher valuerepresent higher quality.

Printer A Printer B Printer C Printer D

0.6

0.7

0.8

0.9

1

1.1

1.2

Qua

lity

valu

es

Printers

Figure 14.13: Boxplot example.

239

Page 256: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

14.3.2.2 3 and 4 dimensional plots

For some types of attributes plots with more than two dimensions can provide an intuitive wayto visualize data. Noise is one example, where one can visualize noise in a three dimensionalspace (for example in CIELAB), making it easy to see in what regions of the color spacenoise is most prominent. We will use three dimensional plots where it is suitable, and extendit to four dimensions by changing the size of the data points in the three dimensional space.Figure 14.14 shows an example of such a plot for noise evaluation. Higher values and largersymbols indicate more noise shown in the CIELAB color space.

−500

50100

−100−50

050

10010

20

30

40

50

60

70

80

90

100

a

Standard deviation shown in the CIELAB color space.The higher the value the more noise.

b

L

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

Figure 14.14: Four dimensional plot for noise evaluation. The noise estimate is plotted in 3Din the CIELAB color space, where the fourth dimension is the size of the points (larger sizeequals more noise).

14.3.2.3 Color histograms

Four dimensional plots gives an intuitive method to visualize data. However, it is up to theviewer to make a decision on which colors that have the highest amount of error (quality loss).This decision might be subjective and vary depending on the observer. Therefore, we adoptthe color histogram proposed by Sun and Zheng [430], where we separate the color spaceinto unique 27 blocks (Figure 14.15), and show the quality values for each block. We usethe standard blocks from Sun and Zheng [430], where the color space (JCh) is segmentedinto eight segments in hue (0, 45, 135, 180, 225, 270, and 315 degrees), three sections forchroma (C = 20 and 50 as the threshold), and three lightness sections (J = 40 and 70) for lowchroma, two for mid-chroma (J = 55), and none for high lightness. This results in 27 differentblocks, representing different colors. For each block a quality value can be calculated, andthen visualized as seen in Figure 14.16.

240

Page 257: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Figure 14.15: 3d color histogram. The color space is segmented into 27 blocks. Figurereproduced from Sun and Zheng [430].

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 270

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

Bin number

Mea

n st

anda

rd d

evia

tion

of e

ach

bin

Noise based on color

Figure 14.16: Example color histogram segmentation for noise evaluation. Each bin (seg-ment) corresponds to one of the regions in Figure 14.15. Each bin is also shown with acolored circle representing the average color of that segment.

241

Page 258: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

14.3.3 Scale differences

The metrics used in the calculation have different scales, and in some cases opposite scales.SSIM has a scale from 0 to 1 (ascending scale), where 1 indicates high quality, while a colordifference based metric (as TVD) has a scale where 0 is highest quality (descending scale). Inorder to plot the results from different metrics together they need to have the same direction.There are several options for achieving this. One possibility is the multiplicative inverse:

x′ =1x, (14.7)

where x is the metric value and x′ is the multiplicative inverse of x. Using the multiplicativeinverse has the advantage that it is both right and left inverse. This is a non-linear mapping.

Another possibility is to use a linear mapping

x′ = −x+max(x), (14.8)

where x is the metric value and x′ is the inverse of x.Figure 14.17 shows the difference between the non-linear mapping and the linear mapping.

A metric value of 0 will be mapped to infinity using the multiplicative inverse, and valuesfrom 1 to the highest value will be mapped between 1 and 0. The linear mapping avoids thisproblem, and we will therefore use a linear mapping.

0

1

2

3

4

5

6

7

8

9

10

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2

Tra

nsf

orm

ed

va

lue

s

Metric values

Mul�plica�ve inverse Linear transfrom

Figure 14.17: Comparison of multiplicative inverse and linear mapping.

The linear mapping will result in the lowest quality value being mapped to 0. Since weuse spider plots for visualization of the results, this is unfortunate since the point is hardlyperceptible. Therefore, we use an ε argument to solve this:

x′ = −x+max(x)+ ε, (14.9)

where ε is 1 percent of the maximum value, and thereby avoiding a value of 0.

242

Page 259: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

This is done for all spider plots resulting in a higher value being higher quality, but forquality maps or 3d plots where it is common to use a descending scale the linear mappinginverse is not applied.

14.4 Overview of the quality assistant

The Quality Assistant consists of several functions, as shown in Figure 14.18. These func-tions are found in the first window of the Quality Assistant (Figure 14.19). The user canalso see which printers that are found in the center column, with the possibility to open thefolder where the toolbox is located. We will below give an introduction to the main features(Figure 14.18) of the toolbox.

Quality

Assistant

Show test

imagesPad images

Scan profile

evalua�on

Image

registra�on

Registra�on

valida�on

Calculate

results

Show

Results

Figure 14.18: Overview of the quality assistant.

Figure 14.19: Screenshot of the front window of the Quality Assistant.

14.4.1 Show test images

The show test images function displays the test images in the toolbox. The entire image setcan be displayed or images for specific attributes. In the main window these are shown asthumbnails, with the option to click on any image to display it at full size. A screenshot of thefunction is shown in Figure 14.20.

14.4.2 Pad images

The pad images function pads the images with the control points, which are used for theregistration. A padded image is shown in Figure 14.1.

243

Page 260: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Figure 14.20: Screenshot of the ”show test images” function in the Quality Assistant. Shownhere are the images for the color attribute. The user can click on these images to display themat fullsize.

14.4.3 Scan profile evaluation

The scan profile evaluation function carries out evaluation of the scanner profile as describedin Section 14.1.2. The result of the evaluation is shown as a color difference map, in termsof ΔE∗

ab, for each color patch of the test target (Figure 14.21). Additional statistics from theevaluation is shown below the color difference map, stating the mean, maximum, minimum,median, and variance of the results. Together with the window shown in Figure 14.21 the testtarget is also shown, in order for the user to identify which patches that have the largest andsmallest differences.

For the example shown in Figure 14.21 a mean value of 2.7 ΔE∗ab is found, with a max-

imum of 7.2. Comparing against the results of Sharma [393], the mean and maximum arecomparable. Sharma [393] states that a mean ΔE∗

ab of less than 3 is very good and likely toproduce excellent results in printer-based workflows. However, it is also important to keep inmind that a large maximum value can potenially create problems in particular colors.

Difference between the measured values and scanner values (ΔE*ab

)

5 10 15 20

2

4

6

8

10

12

14

16

18

1

2

3

4

5

6

7

Figure 14.21: Screenshot of ”scan profile evaluation”.

244

Page 261: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

14.4.4 Image registration

Image registration carries out registration of the scanned images in order to have them matchedwith the original image. The process of registration is explained in details in Chapter 8.

14.4.5 Image registration validation

The image registration validation function performs a validation of the registration. It analy-ses the rotation at which the print is scanned, and if above a threshold the user is recommendedto re-scan the print. The function also allows the user to see which images that are above thethreshold. A screenshot of the image registration validation is shown in Figure 14.22. This isnot a strict validation, and it is recommended that the user compares the results of one imageto the overall trend. If an outlier is found this image should most likely be re-scanned.

1 2 3 4 5 6 7 8 9 10 11 120

0.05

0.1

0.15

0.2

0.25

0.3

0.35

Images

Deg

rees

rot

ated

Registration validation

Threshold

Printer APrinter BPrinter CPrinter D

Figure 14.22: Screenshot of the validation of the registration. The threshold of showing therecommendation if images should be re-scanned is shown as a red line.

14.4.5.1 Visual rotation

The visual rotation function displays a representation of the average rotation and maximumrotation against an ideal rotation (i.e. no rotation). This is a visual tool to show the differencesbetween the scans. A screenshot of this function is shown in Figure 14.23.

No rotation − 0 degrees

Average rotation − 0.1147 degrees

Max rotation − 0.32916 degrees

Figure 14.23: Screenshot of the ”visual rotation” function.

245

Page 262: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual Color image

0.14 0.28 0.42 0.55Image 1

0.15

0.3

0.45

0.6Image 2

0.32

0.63

0.95

1.3

Image 3

Metric: TVDQuality Axis: [worst,best]

Printer APrinter BPrinter CPrinter D

Overall average:Printer A:0.37Printer B:0.53Printer C:0.03Printer D:0.81

Figure 14.24: Example of how the results are shown and how options are presented to theuser in the Quality Assistant.

14.4.6 Calculate results

The calculate results function calculates the results for the IQ metrics, either for all printersor for a specific printer. The results are saved for further analysis.

14.4.7 Show results

The show results function shows the results from the IQ metrics. The user has several optionswhen visualizing the results, showing the overall results for all quality attributes or the resultsfor specific attributes.

The results are visualized using the methods as introduced in Section 14.3. In each viewthe user has several options depending on the attribute. The name of the metric used for thecalculation of the quality is given in the top left corner, together with the quality axis. InFigure 14.24 the TVD metric is used, together with a quality axis where 0 is the worst qualityof the printers evaluated and maximum is on the outskirts of the spider plot. Other metricsmight have other scales, depending on the construction of the metric. Legend of the printersin the quality evaluation is given in the top right corner. In the bottom right corner the averageof the quality values for each printer in the spider plot is given as an indication of the overallperformance of the printers, where 1 is the best printer. On the bottom of the figure severaloptions are given for further investigation of the quality of the different printers. These arepresented in the next subsections.

In the example in Figure 14.24 the metric TVD gives printer D the highest score in all threeimages. printer C is the worst printer, and due to the linear mapping done (Section 14.3.3) thisis placed in the center of the plot. This does not imply that this is the lowest quality possible,but that the printer has the lowest quality among the ones tested. For the visual evaluation of

246

Page 263: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

the plots it important to look at the differences between the printers, and for doing this havinga reference printer will most likely be helpful. The average over the images, as shown in thebottom right corner of Figure 14.24, can be used as an indication of the overall performance ofthe printers for the given attribute. For the example printer D has the highest average, whichis also confirmed by the spider plot. It is importance to notice that even though the valuesare different, they might not be visually different, since no information on just noticeabledifferences are given.

14.4.7.1 Show test images

Shows the test images for the displayed attribute in a similar manner to what is done for theshow test images, as described in Section 14.4.1. Thumbnails are displayed of the images,with the possibility to click on them for full-size view.

14.4.7.2 XLS export

This function exports the values from the spider plot to Microsoft Excel sheet. For the attributedisplayed the function creates a new sheet in the Excel file where the data is written. Anexample of the exported values into a Excel sheet is shown in Figure 14.25. This exportenables the user to carry out additional analysis in other statistical softwares.

Figure 14.25: Example of the exported values into a Excel sheet.

14.4.7.3 All images printer

This function displays the quality maps (Figure 14.12) of a given printer for all images. Eachquality map can be clicked on for further information, where the quality map is displayedat a larger size and with a frame with detailed statistics on the quality map. The displayedinformation is the mean, maximum, minimum, and median values, in addition the standarddeviation of the map is given. Three more statistics are given, the Delta J, Delta C, andDelta h, representing the lightness, chroma and hue difference between the original and thereproduction calculated in the CIECAM02 JCh color space [86, 301]. A screenshot from theQuality Assistant is shown in Figure 14.26.

We can analyze the data from the example in Figure 14.26. In this case since the metricis TVD a higher values indicates lower quality, and TVD maps are shown as color differencemaps. We have a fairly high mean value, with a maximum close to the mean. This indicatesthat most of the values in the map have the same quality, which is also indicated by the lowstandard deviation. The minimum value is very low, which indicates that there are areas in theimage having a very high quality. Using the Delta J, C, h we can also extract information about

247

Page 264: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

0

5

10

15

20

25

30

Figure 14.26: Quality map and additional statistics.

the reason for the quality loss. We have a lower Delta J than Delta C, which indicates thatthe quality loss is larger in the chroma channel (indicated by C) than in lightness (indicatedby L). Delta h can be used to indicate if hue shifts occur compared to the original, and Deltah is high in the given example. Similar statistics are found for other metrics as well, but thespecific statistics depend on the metric.

14.4.7.4 All printers image

This function is similar to that of the previous function, but it displays the quality maps for allprinters for a given image. The quality maps are also clickable, giving the same informationas in Section 14.4.7.3 and Figure 14.26. In addition to showing the quality maps a boxplot (asdescribed in Section 14.3.2.1) based on the quality values is displayed (Figure 14.27), withthe same statistics as in Section 14.4.7.3.

14.4.7.5 4d plot

The 4d plot function displays quality values in a 3d plot with the fourth dimension being thesize of the symbols. A screenshot of the 4d plot for the color attribute is shown in Figure 14.28.In this screenshot color differences are displayed in the CIELAB color space, where the userhas the possibility to rotate the plot. Such a plot can reveal where in the color space a printerhas the largest color differences or most severe quality issues. In the case of Figure 14.28 mostof the errors can be found on the lower L scale, indicating that most of the color difference islocated in darker colors. Note that only the worst 5% of the points are shown in the plot.

A similar plot can also be seen for the noise artifact for the test target (Figure 14.2),but with additional statistics giving the mean, maximum, minimum, median, and standarddeviation. A screenshot of this plot is shown in Figure 14.29. We can notice that the higheststandard deviation is found in a few specific colors close to the neutral axis and fairly high on

248

Page 265: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Printer A Printer B Printer C Printer D 36

38

40

42

44

46

48

50

52Q

ualit

y va

lues

Printers

Boxplots for Perceptual Color image 1

Figure 14.27: Screenshot of the boxplot view in the Quality Assistant.

100120

140160

180

50

100

150

2000

50

100

150

200

a

Color difference between the original and the reproduction (worst 5% unique points)Scale from low quality (blue) to higher quality (red)

b

L

0.022

0.022

0.022

0.022

0.022

0.022

0.022

0.022

0.022

0.022

Figure 14.28: Screenshot of the 4d plot view in the Quality Assistant.

249

Page 266: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

−60 −40 −20 0 20 40 60 80

−50

0

50

10030

40

50

60

70

80

90

a

Standard deviation shown in the CIELAB color space.The higher the value the more noise.

Printer A

b

L

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

Figure 14.29: Screenshot of the 4d plot based on the computational noise metric.

the L axis. However, the highest density of noise is found in the darker colors. Pure colors,being in the outskirts have less noise than those towards the neutral axis. Such an analysisis easier to carry out using the Quality Assistant since panning, zooming, and rotation isavailable.

14.5 Evaluation of the Quality Assistant

We perform a final evaluation of the perceptual metrics in the Quality Assistant. The eval-uation will be carried out for the perceptual metrics, namely SSIM for sharpness, NQM fornoise, and TVD for color and banding. Computional metrics are not evaluated, since thesemight require special evaluation techniques, additionally for EB and ER a compliance stan-dard has been published in ISO 13660 [197].

First we introduce the experimental setup, then we show the results of the evaluation.

14.5.1 Experimental setup

14.5.1.1 Printers and print settings

The images have been printed on different printers and with different printer settings (Ta-ble 14.1), resulting four different reproductions. The printers handled color management with

250

Page 267: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

the default profile, and without automatic enhancement. The rendering intent was set to rela-tive colorimetric. For each printer and setting a complete set of images was printed.

Table 14.1: Printers and details for the evaluation of the Quality Assistant.

Printer Description Paper Print modeprinter A Low-end CMYKcm inkjet printer Copy paper (80 g/m2) Normalprinter B Low-end CMYK inkjet printer Copy paper (80 g/m2) Bestprinter C CMYKcm inkjet plotter Coated paper (130 g/m2) Productivityprinter D 12-channel inkjet plotter Coated paper (130 g/m2) Standard

14.5.1.2 Observers and instructions

Three expert observers, having print quality experience, were recruited for the experiment,being both male and female. All observers passed the Isihara color deficiency test (Sec-tion 3.2.4.3).

The observers were told to judge the quality of the printed images given the referenceaccording to each attribute on the five point ITU quality scale (Excellent, Good, Fair, Poor,Bad), as introduced in Figure 3.30. Together with the instructions the following definitions ofthe quality attributes were given:

• Sharpness is related to the clarity of details and definition of edges.

• Color contains aspects related to color, such as hue, saturation, lightness, and colorrendition.

• Artifacts

– Noise is unwanted variations [205], i.e. fluctuations in color and luminance.

– Banding appears as quasi-periodic density fluctuations in bands that extend acrossthe printed image horizontally or vertically [371, 372].

14.5.1.3 Viewing conditions

The original image for each scene was shown on a calibrated Eizo CG241W. The displayhad a color temperature of 6500 Kelvins and luminance level of 80 cd/m2, following thespecifications of the sRGB. A monitor hood was fitted to the display to prevent glare. Theilluminance level of the viewing room was 450±75 lux, with a color temperature of 5200Kelvins and a color rendering index of 96. The observers viewed the reference image and theprinted image simultaneously from a distance of approximately 60 cm. The experiment wasset up to follow the CIE guidelines [80] as closely as possible.

14.5.1.4 Scanning and registration

In order to use IQ metrics to quantify quality, the printed images needed to be scanned. A HPScanJet G4050 was used for scanning the images at 1200 dpi. A scanner profile was made foreach printer and each setting based on the printed test target. After scanning the images havebeen registered using the framework presented in Chapter 8.

251

Page 268: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual Sharpness image

0.23 0.46 0.69 0.91

Image 1

0.23

0.46

0.69

0.91

Image 2

0.24

0.48

0.72

0.96Image 3

0.210.430.640.86

Image 4

0.18

0.35

0.53

0.71

Image 5

0.24

0.47

0.71

0.95Image 6

Printer APrinter BPrinter CPrinter D

(a) SSIM.

Sharpness images

0.11 0.21 0.32 0.43

Image 1

0.13

0.25

0.38

0.5

Image 2

0.13

0.25

0.38

0.5Image 3

0.190.380.560.75

Image 4

0.15

0.3

0.45

0.6

Image 5

0.094

0.19

0.28

0.37 Image 6

Printer APrinter BPrinter CPrinter D

(b) Observers.

Figure 14.30: Spider plots for SSIM and observers for sharpness images.

14.5.2 Results and discussion

The observers results were calculated as Mean Opinion Score (MOS), and not z-scores due tothe low number of observers. The observer scores have been inverted and scaled between 0and 1. This will give spider plots similar to the IQ metrics, which enables easy comparison.

14.5.2.1 Sharpness

For sharpness the observers have judged printer D to have the highest sharpness quality forall six images (Figure 14.30(b)). SSIM has also given printer D the highest score for all siximages (Figure 14.30(a)). We can also notice that printer B is by the observers judged to bebetter than printer A for all but one image, and the same tendency is also found for SSIM.However, the distinctive indication by SSIM that printer B is better than printer A and printerC is not shared by the observers. We also evaluate the overall results of the observers to theoverall results of the metrics. The overall ranking by the experts were printer D, printer C,printer B, and printer A. For SSIM the mean of the images are taken as the overall indicatorfor quality, and the following ranking is achieved: printer D, printer B, printer A, and printerC. Printer C is the printer that is the difference, being placed fourth by SSIM and second bythe observers. SSIM gives printer C printer a very low score in image 5 (Figure 14.30(a))being some of the reason why it is ranked last.

For the maps the observers judged the entire maps, while the metrics are calculated fordifferent regions. Image 1-3 are three regions of the first map, and images 4 and 5 are tworegions in the second map. For the visualization we have plotted the scores from the observerssimilarly to the metric scores (images 1-3 are the scores for the first map and images 4 and 5for the second map). Figure 14.31(b) shows that the observers judge printer D to have highestquality of the evaluated printers. SSIM agrees for some images with this judgment, such asfor images 3 and 5. Only in image 3 has SSIM the large difference between printer D and theother printers, as the observer have.

For the overall judgment the observers have ranked printer D first, then printer B, printerA, and printer C, being the same ranking as for the images. Despite the differences for the

252

Page 269: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual Sharpness map

0.13 0.27 0.4 0.54

Image 1

0.15

0.31

0.46

0.62

Image 2

0.16

0.32

0.48

0.64Image 3

0.17

0.35

0.52

0.7

Image 4

0.2

0.4

0.6

0.8Image 5

Printer APrinter BPrinter CPrinter D

(a) SSIM.

Sharpness maps

0.25 0.5 0.75 1

Image 1

0.25

0.5

0.75

1

Image 2

0.25

0.5

0.75

1Image 3

0.15

0.3

0.45

0.6

Image 4

0.15

0.3

0.45

0.6Image 5

Printer APrinter BPrinter CPrinter D

(b) Observers.

Figure 14.31: Spider plots for SSIM and observers for sharpness maps.

individual images SSIM ranks the printers identically to the observers for the overall evalua-tion.

14.5.2.2 Noise

Figure 14.32(b) shows that the observers prefer printer D, being the printer with the leastamount of noise for all images. SSIM is in agreement with the observers (Figure 14.32(a)),except for image 5 where printer C has a slightly better score. The observers judge printer Ato have the most noise (lowest quality) in five of the six images. SSIM has a similar score,giving printer A the lowest score in all six images.

The observers and the metric has the same overall ranking for noise, with printer D beingthe printer with the least amount of noise, followed by printer C, printer B, and printer A.

14.5.2.3 Banding

Figure 14.33 shows the results for the metric (TVD with post-processing) and the observers forthe images in the banding evaluation. For the first image the metric correctly ranks printer Cas better than printer D, followed by printer B and printer A. For the second image the metricestimates printer C to be slightly better than printer D, before a difference down to the twoothers, where printer A is ranked before printer B. The ranking of the two worst ones for thisimage is the same as the observers. For the third image the metric ranks printer C to be betterthan the three others, but the difference between the printers is very small. The observers alsoindicate that the banding in this image is similar between the printers, but with printer B asbetter. For the overall ranking the observers prefer printer D over printer C, followed by a tiebetween printer A and printer B. The metric ranks printer C before printer D, and printer Abefore printer B.

For the targets the differences between the printers are small for the observers, but forthe first image observers judge printer A as better than printer B with the two other printerstied last. TVD correctly ranks printer A and printer B over printer C and printer D, with

253

Page 270: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual Noise image

5.3 11 16 21

Image 1

4.4

8.7

13

17

Image 2

4.8

9.6

14

19Image 3

5101520

Image 4

4.9

9.8

15

20

Image 5

5.5

11

16

22Image 6

Printer APrinter BPrinter CPrinter D

(a) NQM.

Noise images

0.19 0.38 0.56 0.75

Image 1

0.13

0.25

0.38

0.5

Image 2

0.19

0.38

0.56

0.75Image 3

0.150.30.450.6

Image 4

0.094

0.19

0.28

0.37

Image 5

0.15

0.3

0.45

0.6Image 6

Printer APrinter BPrinter CPrinter D

(b) Observers.

Figure 14.32: Spider plots for NQM and observers for noise images.

Perceptual Banding image

0.01 0.021 0.031 0.041

Image 1

0.0065

0.013

0.02

0.026Image 2

0.0049

0.0097

0.015

0.019

Image 3

Printer APrinter BPrinter CPrinter D

(a) TVD.

Banding images

0.13 0.25 0.38 0.5

Image 1

0.15

0.3

0.45

0.6Image 2

0.15

0.3

0.45

0.6

Image 3

Printer APrinter BPrinter CPrinter D

(b) Observers.

Figure 14.33: Spider plots for TVD and observers for banding images.

254

Page 271: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Printer A Printer B Printer C Printer D0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04Perceptual Banding target

Printers

Qua

lity

Image 1Image 2

(a) TVD.

Printer A Printer B Printer C Printer D0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4Banding images

Printer

Qua

lity

Image1Image 2

(b) Observers.

Figure 14.34: Bar plots for TVD and observers for banding targets.

printer A ranked before printer B. For the second image observer rank printer D as the bestslightly before the three other printers that get the same score. TVD on the other hand ranksprinter D as the worst one, with printer C first. However, the differences in metric scores arevery small. Additionally, it should be noted that the visual differences in the scores from theobservers between the targets are small, and therefore making it very difficult for the metric.Overall the observers prefer printer A over printer B and printer D, being tied, and printerC at last. However, the scores are very similar and with only three observers they are notstatistically significantly different. The metric ranks printer A over printer B, with printer Cslightly before printer D.

14.5.2.4 Color

The observers prefer printer D in terms of color for two of the three images for the colorattribute (Figure 14.35(b)). TVD has printer D as the best printer for all three images. How-ever, in the second image the difference between printer D and printer B is small. TVD ranksprinter B as the second best in two of the three images. We can also notice that printer C andprinter A are rated as the two printers having the lowest quality both by the observers, and byTVD for the two last images. In general TVD produces stable results over the three images,while the observers have slight differences from one image to the other.

The observers have given the following overall rank: printer D first, then printer B, printerC, and printer A at last. TVD has almost the same ranking, but switching the two last printers:printer D first, then printer B, printer A, and printer C.

14.5.3 Overall observations

Overall the metrics in the Quality Assistant agrees with the observers. However, in some attri-butes the scores from the observers are similar resulting in a difficult task for the metrics. Themetrics in the Quality Assistant are most likely not suited to measure very small differences,but they can be a valuable tool to indicate the performance of printers with larger differences.

14.6 Summary

In this chapter we have introduced the Quality Assistant, a tool for the industry to evaluatequality of printed images. The functions of the Quality Assistant have been presented and

255

Page 272: Doctoral Dissertation by Marius Pedersen

QUALITY ASSISTANT - TOOL FOR EVALUATING PRINT QUALITY

Perceptual Color image

0.14 0.28 0.42 0.55

Image 1

0.15

0.3

0.45

0.6Image 2

0.32

0.63

0.95

1.3

Image 3

Printer APrinter BPrinter CPrinter D

(a) TVD.

Color images

0.11 0.21 0.32 0.43

Image 1

0.094

0.19

0.28

0.37Image 2

0.083

0.17

0.25

0.33

Image 3

Printer APrinter BPrinter CPrinter D

(b) Observers.

Figure 14.35: Spider plots for TVD and observers for color images.

explained. An evaluation of the perceptual IQ metrics in the Quality Assistant has also beencarried out.

256

Page 273: Doctoral Dissertation by Marius Pedersen

15 CONCLUSION

The goal of this work was to adapt, use, and evaluate image quality metrics for measuringperceived image quality, in the applied field of color printing. Estimating perceived qualityof prints is not a trivial task, since quality is inherently multidimensional and subjective.Nevertheless, our aim was to have image quality metrics correlated with perceived quality,which is a complex and difficult task.

First, we introduced the concept of image quality and the existing methods to measurequality in Part I. Thereafter, in Part II we investigated existing image quality metrics in theliterature. First we classified existing metrics, resulting in a better understanding of existingmetrics and how they are constructed. Based on the knowledge gathered from this classifi-cation we proposed a new image quality metric in Chapter 5 taking into account region ofinterest and known aspects of the human visual system. Chapter 6 consisted of an overviewof existing methods for the evaluation of image quality metrics, which revealed that a methodfor evaluating overall performance was missing. Therefore, we proposed a new method forthe overall evaluation of image quality metrics based on rank order. In the final chapter ofPart II, we selected image quality metrics for evaluations against human observers using themethods based on correlation for evaluating the performance of the image quality metrics.

Part III focused on the use of image quality metrics for color printing. Applying imagequality metrics to printed images is not straightforward, since the metrics require correspond-ing digital images. In Chapter 8 we proposed a framework for converting printed images intodigital images, including scanning and registration. Evaluation of state of the art metrics usingthe framework in Chapter 9 showed that none of the metrics correlated well with perceivedoverall image quality. Therefore, Chapter 10 dealt with breaking overall IQ into quality attri-butes, reducing the complexity and dimensionality of IQ. A set of six color printing qualityattributes were proposed as the most important for the evaluation of quality of printed images,which were experimentally validated. Further, in Chapter 11, IQ metrics were selected forthe different color printing quality attributes, and evaluated against experimental data fromhuman observers. This evaluation showed that there is indeed room for improvement, andin Chapter 12 we investigated how one of the stages in image quality metrics, pooling, influ-enced the performance of image quality metrics. Our findings showed that the pooling methodis metric dependent and dependent on parameters. Even with the effort carried out with thepooling strategies none of the metrics performed well for the color quality attribute, which ledto a new metric for the color quality attribute that was introduced in Chapter 13. Chapter 14combines the work carried out in this thesis into a tool for objective quality assessment ofcolor prints, the Quality Assistant.

257

Page 274: Doctoral Dissertation by Marius Pedersen

CONCLUSION

Main contributions

• Classification and survey of image quality metrics, which organizes them and identifiesdifferences between metrics. Metrics are classified into four groups: mathematicallybased metrics, low-level based metrics, high-level based metrics, and other metrics.The classification contributes to understanding the relationship between metrics, andcan be used in the development of new metrics.

• Proposed a new method for evaluating the overall performance of image quality metrics.The new method is based on the rank order method, commonly used to statisticallyanalyze subjective data. The new method takes the image quality metric scores andinputs them to the rank order method, which generates z-scores enabling convenientcomparison against z-scores from subjective experiments. The method is less sensitiveto scale differences and extreme values than most existing methods.

• Extensive evaluation of image quality metrics were carried out. 22 image quality met-rics were evaluated against subjective data from six state of the art databases. Theresults show that the performance of the image quality metrics depend on the type ofdistortion, but also on the database. The results also indicate that the image qualitymetrics perform worse for images with multiple distortions.

• Two new image quality metrics have been developed: the Spatial Hue Angle MEtric(SHAME) and the Total Variation of Difference (TVD) metric. SHAME takes into ac-count the importance of local regions and simulation of the HVS. TVD merges totalvariation and contrast filtering to create a unique IQ metric well-correlated with per-ceived image quality.

• A framework for applying image quality metrics to printed images has been introduced.The framework uses control points to register scanned prints to match the original im-ages. The framework outperformed a state of the art framework, and enables the use ofmetrics on printed images.

• Proposal of the Color Printing Quality Attributes (CPQAs), which is a set of qualityattributes for the evaluation of color prints. The attributes included are the following:sharpness, contrast, color, lightness, artifacts, and physical. These have been validatedin two experiments, and are considered to be a good foundation for the evaluation ofprint quality.

• Image quality metrics have been suggested and evaluated for the CPQA, which con-stituted the basis for selecting the most appropriate metric for each CPQA. The resultsshow that metrics based on structural similarity perform well for sharpness, lightness,and contrast. Many metrics perform well for artifacts, depending on the artifact. Forcolor none of the evaluated metrics perform well.

• Evaluation of pooling strategies for image quality metrics identified that pooling is met-ric, parameter, and attribute dependent. The pooling strategy and its parameters shouldtherefore be chosen carefully depending on the metric and attribute.

• The Quality Assistant has been developed to assist researchers and developers in theprinting industry to evaluate quality. The Quality Assistant consists of all functionsneeded to evaluate quality, including: a test image suite, a framework for digitizing theprints, a set of image quality metrics, and visualization tools.

258

Page 275: Doctoral Dissertation by Marius Pedersen

CONCLUSION

In conclusion, our work has investigated the use of image quality metrics to measure thequality of color prints. New frameworks, metrics, and tools for measuring quality have beenproposed, and the work carried out in this project will be useful for the industry and for futureresearch and development in the image quality area.

259

Page 276: Doctoral Dissertation by Marius Pedersen

CONCLUSION

260

Page 277: Doctoral Dissertation by Marius Pedersen

16 FUTURE WORK

In the course of the research carried out in this thesis several possibilities for further researchhave been identified. The thoughts on future work can be divided into several sections, relatedto different parts of the thesis.

16.1 Quality Assistant

We have proposed IQ metrics for some of the quality attributes in the Quality Assistant, butthere are still more attributes that should be included to have a complete evaluation of quality.There is also room for improvement for the metrics in the Quality Assistant, they are still notperfectly correlated with perceived quality.

We have focused on inkjet printing and halftoned images, but the Quality Assistant shouldbe applicable for other printing technologies, so future work should include evaluation forother technologies, for example laser printing. However, it might be that technology-specificattributes need to be included. For laser printing other sub-artifacts should be used, such asghosting (i..e a vestigial image repeated at regular intervals of a page and appearing as lightor dark areas [51]).

To make the Quality Assistant even more usable, investigation should be carried out on thepossibility of using the scanners in the multi function printers. This would make the QualityAssistant easier to use in the field, since a separate scanner would not be needed.

The Quality Assistant uses full-reference metrics, and the disadvantage of this is that thereference image is needed. Extension of the Quality Assistant to no-reference IQ metricswould make it independent of the reference image, making it applicable to any printed im-age. For some attributes incorporation of no-reference metrics would be easier than others,especially those attributes where specific distortions occurs, such as noise and banding.

Usability testing of the Quality Assistant is another factor that could be investigated, inorder to ensure its ease of use, efficiency, and learnability. This could include testing theQuality Assistant in use in the field, or more controlled usability testing in a lab.

The spider plots for the metrics do not indicate any uncertainty, and it can be difficult todraw a conclusion whether two different printers are perceptually different or not. Investiga-tion of Just Noticeable Difference (JND) for the different metrics could solve this problem,and give an indication of a visual difference between two quality values.

Parametric as well as non-parametric statistical decision tools could either be appended orlinked to the Quality Assistant in order to rank or present similarity measures of the results inthe Quality Assistant.

261

Page 278: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

16.2 Quality attributes

Even though the CPQAs have been made for color printing, they are general and they couldpossibly be used in other fields such as for video or display quality. We believe that the use ofthe CPQAs could be beneficial in these fields as they have been for printing, but they mightbe subject to adjustments to fit the field.

In this work the physical CPQA has not been investigated, and work is needed to findsuitable metrics to measure this attribute.

In Chapter 12 pooling was discussed to reduce the number of quality values from themetric. However, pooling is also an important issue when combining quality values fromdifferent quality attributes. In order to obtain a value representing overall quality, pooling ofquality attributes is required.

The dependence between attributes is important in order to pool quality attributes, andshould be one of the steps in further work.

16.3 Image quality metrics

The spatial filtering method introduced in Chapter 13 is interesting, and further work shouldbe carried out to inspect the differences from other filtering methods and to evaluate its perfor-mance. In this work the TVD metric proposed in Chapter 13 has been evaluated using printedimages, and another possible next step would be to evaluate it on other databases, such asTID2008 [365] and LIVE [399].

Pooling was a central part of this work, and our findings show that the pooling strategycan influence the performance of IQ metrics. We believe that further investigation of poolingstrategies would reveal more advanced and better pooling strategies, that could improve IQmetrics for prints but also in general for IQ.

With IQ metrics correlated with perceived IQ they can also be used to optimize quality.The gamut mapping algorithms proposed by Alsam and Farup [8, 9] minimizes according toa constraint, where the constraint possibly could be set by IQ metrics. Similar approaches canalso be investigated for other processes, for example halftoning.

16.4 Image characteristics

In this work we also investigated the relationship between image characteristics and the per-formance of IQ metrics. It was mentioned in Section 11.2.3 that a larger dataset is required todraw a conclusion, and therefore a natural next phase would be to investigate characteristicson a large number of test stimuli. Other characteristics than those used in this work would alsobe interesting to look at, both for printed images and for other areas, such as compression.

Image characteristics could possibly be used to select the most suitable metrics that shouldbe used to evaluate an image. Image characteristics has the potential to be used to adapt theprocessing of images. Some work has been done adapting processing steps to characteristicsof the image, for example to improve halftoning [166, 247, 448] or assist in the gamut map-ping [75, 155, 181, 430]. We believe that image characteristics can be used as a decision toolfor the different processing steps.

262

Page 279: Doctoral Dissertation by Marius Pedersen

NOMENCLATURE

ABF Adaptive Bilateral Filter

ANSI American National Standards Institute

bdvm Der Bundesverband Druck und Medien e.V.

BPP Bit Per Pixel

CCD Charge Coupled Device

CCFL Cold Cathode Fluorescent Lamp

CCT Correlated Color Temperature

CI Confidence Interval

CIE Commission Internationale de l’eclairage

CIS Contact Image Sensor

CISM Color Image Similarity Measure

CMM Color Management Module

CN Computational Noise

Colorlab Norwegian Color Research Laboratory

CP Control Points

CPD Cycles Per Degree

CPQA Color Printing Quality Attribute

CRT Cathode Ray Tube

CSF Contrast Sensitivity Function

CSNR Contrast Signal to Noise Ratio

CV Coefficient of Variation

CW-SSIM Complex Wavelet Structural SIMilarity

DMOS Difference Mean Opinion Score

263

Page 280: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

DOG Difference of Gaussians

DPI Dots pr inch

DSCQS Double Stimulus Continuous Quality Scale

DSIS Double Stimulus Impairment Scale

EB Edge blurriness

ECA European Color Initiative

EDWM Error Diffusion Worm Measure

ER Edge Raggedness

ERA European Rotogravure Association

Fogra Fogra Graphic Technology Research Association

GAFFE Gaze-Attentive Fixation Finding Engine

GCA Graphic Communications Association

GMA Gamut Mapping Algorithm

GRACoL General Requirements and Applications for Commercial Offset Lithography

GU Gloss Units

HP Hewlett-Packard

HPminDE Hue preserving minimum ΔE∗ab clipping

HVS Human Visual System

I3A International Imaging Industry Association

IAS Image Quality Analysis

IC Image Characteristic

IC Information Content Based Pooling

ICC International Color Consortium

IFC Information Fidelity Criterion

IPS In-Plane Switching

IQ Image Quality

ISO International Organization for Standardization

ITU International Telecommunication Union

IU Interval of Uncertainty

264

Page 281: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

JND Just Noticeable Difference

JPEG Joint Photographic Experts Group

LAR Locally Adaptive Resolution

LCD Liquid Crystal Display

LF Local Features

LFM Logistic Function Matrix

LL Lower Limit

LPI Lines pr inch

MAE Mean Absolute Error

MOS Mean Opinion Score

MS-SSIM Multi Scale Structural SIMilarity

MSE Mean Squared Error

NB Nonparametric Bottom-up Saliency Model Based Pooling

NQM Noise Quality Measure

NQM Noise Quality Metric

OSA-UCS Optical Society of America’s Committee on Uniform Color Scales

PCS Profile Connection Space

PDF Portable Document Format

PEST Parameter Estimation by Sequential Testing

PLCC Pearson Linear Correlation Coefficient

POI Percentage of Images

PSE Point of Subjective Equality

PSNR Peak Signal to Noise Ratio

PSO Process-Standard Offset

QA Quality Attribute

QEA Quality Engineering Associates

RANSAC RANd SAmple Consensus

RC Rank Correlation

RFSIM Riesz-transfrom based Feature SIMilarity

265

Page 282: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

RMS Root Mean Square

RMSE Root Mean Squared Error

RSC Retinal-like Subsampling Contrast

S-CIELAB Spatial CIELAB

S-DEE Spatial-DEE

SGCK chroma-dependent SiGmoidal lightness mapping and Cusp Knee scaling

SHAME Spatial Hue Angle MEtric

SIFT Scale-Invariant Feature Transform

SQS Standard Quality Scale

SSIM Structural SIMilarity

STD Standard Deviation

SWAM Saliency based Visual Attention Model

SWOP Specifications for Web Offset Publications

TID Tampere Image Database

TV Total Variation

TVD Total Variation of Difference

UIQ Universal Image Quality

UL Upper Limit

VA Vertical Alignment

VIF Visual Image Fidelity

VSD Verband der Schweizer Druckindustrie

VSNR Visual Signal to Noise Ratio

WLF Weighted multi-Level Framework

266

Page 283: Doctoral Dissertation by Marius Pedersen

AUTHOR INDEX

Achanta, R. 201, 219, 281, 339

Acharya, T. 126, 281

Adelson, E. H. 281, 334, 336

Agaian, S. S. 310, 329

Ajagamelle, S. A. 6, 9, 70, 71, 99, 100, 111,

113, 114, 281, 330, 364

Akenine-Moller, T. 303, 329

Albregtsen, F. 5, 7, 9, 157, 158, 161,

163–165, 168, 171, 174, 196, 200, 201,

215, 219, 303, 305, 331

Alers, H. 49, 307

Alford, J. R. 301, 323

Algazi, R. 301, 323, 324

Algazi, V. R. 282, 325

Algeri, T. 307, 334–336, 363

Allebach, J. P. 121, 128, 144, 145, 148, 149,

171, 180, 282, 283, 285, 291, 292, 312,

315, 316, 324–326

Allen, A. R. 89, 288

Allen, E. 173, 187, 281, 304, 329

Alsam, A. 262, 281

Altman, D. G. 87, 281

Amirshahi, S. A. 5, 9, 173, 189, 305

An, K. 282, 328

Anderson, C. H. 281, 334, 336

Andersson, M. 144, 146, 147, 180, 282

Aoki, N. 141, 142, 144, 148, 308

Arce, G. R. 299, 347, 348, 351

Arditi, A. 163, 298

Arndt, D. 51, 288

Asada, T. 79, 289

Asikainen, R. 54, 65, 159, 293

Astola, J. 50, 66, 86, 87, 99–101, 187, 262,

289, 307, 328, 329

Autrusseau, F. 99, 101, 111, 285

Avadhanam, N. 282, 325

Avcibas, I. 62, 63, 282, 326

Axelson, P. 282, 351

Ayed, M. A. B. 282, 308, 326, 328

Babcock, J. S. 75, 95, 282

Baddeley, A. J. 315, 324

Bai, J. 70, 80, 198, 282

Bakke, A. M. 152, 180, 282, 290

Bala, R. 128, 229, 299

Balasubramanian, R. 180, 282

Bando, E. 7, 67, 70, 80, 86, 93, 113, 139,

142, 144, 146, 148, 173, 180, 282, 293

Bang, Y. 48, 128, 144, 148, 233, 234, 282,

283, 297, 299

Baranczuk, Z. 6, 9, 56, 75, 136, 139,

267

Page 284: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

171–173, 187, 200, 201, 283, 285, 311,

330, 345

Barba, D. 63, 194, 195, 198, 199, 285, 286,

299, 303, 327

Barten, P. G. J. 180, 283

Bartleson, C. J. 38, 40, 141–144, 148, 149,

283

Battisti, F. 50, 86, 87, 99–101, 262, 289, 307,

328, 329

Beatty, J. C. 180, 311

Bech, S. 51, 293

Beghdadi, A. 283, 284, 286, 294, 326, 327,

330

Belcher, S. J. 51, 283

Berche, S. 67, 70, 74, 80, 100, 113, 142–144,

146, 148, 149, 180, 181, 284, 329

Berenson, M. L. 51, 87, 283

Beretta, G. 163, 317

Bergen, J. R. 281, 334, 336

Berk, T. 161, 283

Bestmann, G. 23, 283

Bianco, S. 235, 283, 330

Biemond, J. 315, 323

Blin, J-L. 306, 325

Bliznyuk, V. 144, 148, 315

Blomgren, P. 215, 218, 283

Blommaert, F. J. J. 303, 324

Boi, J.M. 68, 301

Bolin, M. 283, 325

Bonnier, N. 5–7, 9, 67, 70, 72, 74, 80, 100,

113, 135, 136, 142–144, 146, 148, 149,

154, 157, 158, 161, 163–165, 168, 171,

174, 180, 181, 185, 200, 201, 215, 219,

237, 283, 284, 290, 305, 329, 331

Bosco, A. 292, 327

Bouman, C. A. 173, 180, 187, 291, 298, 315,

326, 351

Boureau, Y-L. 193, 284

Bouzerdoum, A. 284, 327

Bouzit, S. 70, 80, 141, 142, 144, 145, 148,

284

Bovik, A. C. 46, 48, 51, 61, 63, 64, 66, 72,

74–76, 87, 89, 92, 100, 101, 122, 129, 136,

144, 171–173, 187, 189, 193–195,

197–201, 231, 235, 262, 288, 302, 307,

309, 313, 314, 325–328

Bowen, J. 180, 262, 293

Bradley, A. P. 79, 284, 325

Brady, M. 136, 297

Brailean, J. C. 284, 322

Brambilla, C. 163, 317

Braun, G. J. 68, 180, 284

Braun, K. 51, 291

Brettel, H. 6, 67, 70, 74, 80, 100, 113,

142–144, 146, 148, 149, 180, 181, 237,

284, 290, 329

Briggs, J. C. 149, 261, 284, 285

Brooks, A. C. 285, 328

Brownston, L. 161, 283

Bryngdahl, O. 308, 324, 348, 351

Buchanan, J. W. 313, 325, 348, 350

Buckley, R. 70, 80, 316, 324

Buring, H. 150, 285

Burningham, N. 145, 149, 285

268

Page 285: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Burns, P. D. 128, 311, 316

Burt, P. J. 281, 334, 336

Calabria, A. J. 149, 154, 285

Caldera, M. 101, 289

Callet, P. Le 63, 99, 101, 111, 285

Canny, J. 285, 349

Cao, D. 141, 142, 144, 148, 236, 287

Cao, G. 6, 9, 75, 136, 139, 171, 173, 187,

285, 330, 345

Capra, A. 163, 285

Caracciolo, V. 6, 9, 99, 107, 109, 110, 215,

285, 310, 330

Cardin, N. 133–136, 285

Carli, M. 50, 66, 86, 87, 99–101, 187, 262,

289, 307, 328, 329

Carlson, C. F. 68, 180, 284

Carmel, S. 310, 331

Carnec, M. 286, 327

Castrorina, A. 163, 285

Caviedes, J. 145, 148, 180, 286

Chaddha, N. 286, 323

Chan, T. F. 215, 218, 283, 286

Chandler, D. M. 64–66, 75, 76, 87, 89, 100,

101, 172, 173, 187, 189, 198, 235, 286,

298, 299, 329

Cheikh, F. A. 6, 9, 99, 107, 109, 110, 215,

310, 312, 330

Chen, C. T. 284, 322

Chen, G. H. 286, 328

Chen, K. 215, 286

Chen, S. 286, 290, 330, 358

Chetouani, A. 286, 330

Cho, M-S. 67, 70, 80, 297

Cho, Y-H. 180, 262, 286

Choi, H-K. 48, 297

Choi, S. Y. 149, 286

Chou, C. 67, 286, 287, 323, 329

Chung, R. 149, 287

CIE 38, 50–53, 56, 79, 91, 92, 95, 107, 135,

150, 151, 159, 174, 188, 190, 247, 251,

287, 334

Ciocca, G. 235, 283, 330

Clarberg, P. 303, 329

Clarke, F. J. J. 287, 322

Colford, J. M. 193, 304

Comaniciu, D. 287, 343

Connah, D. 67, 70, 80, 93, 282

Cookingham, R. 145, 148, 149, 312

Corchs, S. 163, 285

Cordone, R. 147, 308, 335

Cormack, L. K. 46, 51, 101, 199, 262, 307,

309, 313

Couwenhoven, D. W. 311, 347

Cowan, W. B. 180, 311

Crean, P. A. 141–144, 146–148, 237, 288

Creelman, C. D. 33, 312

Crete, F. 173, 187, 232, 287

Crosby, P. B. 13, 14, 287

Cui, C. 45, 95, 141, 142, 144, 148, 180, 236,

287, 288

Cui, G. 67, 300, 325

Cui, L. 89, 288

Cusano, C. 180, 288

Dalal, E. N. 141–144, 146–148, 237, 288,

269

Page 286: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

303

Daly, S. 79, 288, 322

Damera-Venkata, N. 122, 235, 288, 301, 325,

347

Danielsen, H. E. 196, 303

de Freitas Zampolo, R. 288, 327

de Ridder, H. 141, 142, 144, 195, 196, 288,

291

de Veciana, G. 74, 309, 327

Deffner, G. 51, 288

Denecker, K. 124, 313

Deng, C. 317, 330

deQueiroz, R. 180, 282

Dijk, J. 149, 288

Dinet, E. 147, 312

Dolmiere, T. 173, 187, 232, 287

Donahue, K. D. 236, 288

Dong, W. 288, 329

Donohue, K. D. 236, 251, 307

Du, W. 282, 328

Dugay, F. 51, 91, 92, 99, 113, 115, 158, 180,

193, 289

Ebner, F. 68, 180, 284

ECI 23, 289

Eerola, T. 122, 123, 130, 131, 144, 289

Egiazarian, K. 50, 66, 86, 87, 99–101, 187,

262, 289, 307, 328, 329

Eidenvall, L. 141–144, 146–149, 304

El-Adawy, M. I. 87, 89, 303, 330

Elad, M. 297, 328

Elliot, S. L. 171, 232, 315

Enanoria, W. 193, 304

Endo, C. 79, 289

Engeldrum, P. G. 3, 4, 13, 14, 36, 50, 51, 54,

56, 92, 95, 141, 143, 145, 174, 200, 220,

289

Engelke, U. 101, 289

Eschbach, R. 56, 180, 282, 317

Eskicioglu, A. M. 173, 187, 309, 326

Estrada, F. 180, 201, 219, 281, 291, 339

European Space Agency 159, 290

Evans, B. L. 122, 235, 288, 301, 325, 347

Fairchild, M. D. 14, 50, 51, 62, 68, 70, 72,

75, 81, 83, 92, 95, 99, 100, 136, 149, 154,

180, 215, 218, 235, 247, 282, 284, 285,

290, 291, 296, 302, 325, 326

Falkenstern, K. 6, 237, 290

Fan, Z. 290, 309, 347, 351

Farrell, J. E. 38, 62, 67, 70, 80, 121, 173,

187, 298, 310, 313, 316, 324, 351

Farrugia, J. P. 290, 325

Farup, I. 5, 7, 9, 51, 67, 70, 71, 81, 91–93,

99, 100, 113, 115, 136, 147, 152, 158, 173,

180, 187, 193, 200, 201, 215, 218, 220,

262, 281, 282, 289, 290, 298, 304, 306,

310, 311, 330, 331, 336, 363

Fatemi, E. 215, 308

Fechner, G. T. 27, 290

Fedorovskaya, E. A. 141, 142, 144, 145,

147–149, 290, 291

Feeny, S. 33, 291

Feng, X. 195, 198, 199, 291

Feng, X-F. 70, 80, 291, 296, 324, 326

Fernandez, S. R. 51, 291

270

Page 287: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Ferwerda, J. A. 29, 32, 35, 291, 326

Field, G. G. 14, 52, 107, 150, 151, 188, 291

Field, S. 70, 80, 296

Fisher, P. S. 290, 358

Fitzhugh, A. 35, 314

Fleming, P. D. 230, 309

Flemming, P.D. 144, 148, 315

Floyd, R.W. 291, 347, 348, 351

Forrest, D. 261, 284

Fournier, A. 313, 325, 348, 350

Frankle, J. 291, 334, 336

Franti, P. 291, 325

Fredembach, C. 180, 291

Frese, T. 291, 326

Friedman, J.S. 146, 291

Gao, X. 292, 327

Gardner, M. J. 87, 281

Garvin, D. A. 14, 292

Gasparini, F. 163, 180, 285, 288

Gast, G. 51, 142, 144, 146–148, 292

Gatta, C. 92, 290

Gayle, D. 292, 327

Geisler, W. S. 122, 235, 288, 325

Giesen, J. 161, 172, 173, 187, 200, 201, 283,

313

Giger, M. L. 284, 322

Gill, D. 310, 331

Giorgianni, E. J. 161, 292

Golan, A. 180, 262, 292

Gomes, J. 149, 262, 313

Gong, M. 5, 9, 292, 306, 331

Google 3D Warehouse 159, 292

Gorley, P. 292, 329

Gorman, J. D. 193, 304

Granger, E. M. 292, 329

Green, P. 135, 136, 154, 284

Green, P. J. 23, 95, 144, 148, 292

Greenshields, K. W. 51, 283

Grice, J. 121, 128, 292

Grunbaum, B. 145, 292

Guarnera, M. 292, 327

Guarneri, I. 292, 327

Guo, J. 298, 324

Gupta, M. R. 180, 262, 293

Gusev, A. 173, 187, 309, 326

Gusev, D. A. 148, 149, 298

Ha, Y-H. 180, 262, 286

Hakkinen, J. 149, 307

Halonen, R. 54, 65, 122, 159, 289, 293

Haneishi, H. 79, 144, 146–148, 289, 301

Hardeberg, J. Y. 5–7, 9, 50, 51, 61, 65, 67,

70–72, 79–81, 86, 91–93, 99, 100, 107,

108, 113, 115, 136, 137, 139, 142, 144,

146–149, 152, 157, 158, 161, 163–165,

168, 171–174, 180, 187, 189, 190, 193,

198, 200, 201, 215, 219, 220, 229, 282,

289, 290, 293, 304–306, 310, 311, 314,

330, 331, 336, 348, 360, 364

Hasler, D. 180, 293

Hasselgren, J. 303, 329

Hatanaka, S. 303, 325

Havstad, A. 284, 327

Healey, G. 128, 229, 304

Heeger, D. J. 66, 293, 312, 323

271

Page 288: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Hel-Or, H. 180, 262, 292

Hemami, S. S. 64–66, 76, 100, 101, 172, 173,

187, 201, 219, 235, 281, 286, 329, 339

Hertel, D. W. 70, 80, 293

Herzog, P. G. 150, 285

Heverhagen, J. T. 235, 293

Heynderickx, I. E. 49, 195, 198, 199, 300,

307

Hinke, W. H. 296, 347, 351

Holliman, N. 292, 329

Holm, J. 52, 150, 293

Hong, E. 261, 284

Hong, G. 79, 92, 100, 136, 142, 144, 146,

181, 293, 326

Horita, Y. 101, 294

Horiuchi, T. 76, 299, 329, 348, 351

Hou, X. 195, 198, 294, 339

Huertas, R. 67, 100, 218, 304, 330, 363

Hughes, C. E. 315, 328

Hull, M. 185, 284

Hung-Shing, C. 262, 294

Hunt, R. W. G. 144, 146, 147, 294

Hunt, R. W.G. 247, 302

Hussein, A. E. 87, 89, 303, 330

Huttu, I. 144, 148, 304

Imai, F. H. 294, 326

Imai, J. 3, 294

International Color Consortium 133, 149,

294

International Imaging Industry Association

14, 294

International Telecommunication Union 46,

47, 50, 51, 89, 294

Iordache, R. 294, 326

ISO 13, 14, 19–21, 48–51, 53, 56, 128, 151,

159, 188, 232, 234, 250, 251, 295, 296

ISO/IEC 121, 296

Itti, L. 195, 198, 199, 296, 339

Ivkovic, G. 296, 327

Jacobson, R. E. 13, 14, 281, 296, 329

Janssen, T. J. W. M. 13, 149, 163, 296

Jarvis, J. F. 296, 347, 351

Jin, E. W. 70, 80, 296, 324

Johnson, G. M. 62, 70, 72, 81, 83, 92, 100,

136, 149, 215, 218, 235, 290, 296, 325,

326

Johnson, T. 52, 150, 293

Johnston, J. D. 308, 322

Judice, C. N. 296, 347, 351

Jung, E. 150, 285

Jung, J. 301, 329

Juricevic, I. 171, 232, 297

Kadir, T. 136, 297

Kaiser, P. K. 33, 291

Kalviainen, H. 122, 123, 130, 131, 144, 289,

308

Kamarainen, J-K. 122, 123, 130, 131, 144,

289

Karunasekera, S. A. 297, 323

Katajamaki, J. 171, 297

Kaufman, A. 161, 283

Kawayoke, Y. 101, 294

272

Page 289: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Keegan, B. F. 126, 305

Keelan, B. W. 13, 14, 48–51, 141, 143–149,

151, 154, 297, 312

Kendall, M. G. 86, 87, 136, 174, 189, 190,

297, 352

Kennedy, G. 193, 304

Kennedy, R. 126, 305

Keshk, H. A. 87, 89, 303, 330

Kheiri, A. 50, 99, 307

Khriji, L. 308, 328

Kim, J-S. 67, 70, 80, 297

Kim, S. H. 25–27, 128, 232–234, 299, 316

Kim, Y. J. 48, 297

Kim, Y-T. 180, 262, 286

Kimmel, R. 297, 328

Kingsbury, N. G. 297, 316, 323, 325

Kiorpes, L. 70, 81, 302

Kipphan, H. 16, 297

Kite, T. D. 122, 235, 288, 325

Klaman, M. 141–144, 146–149, 297, 304

Klassen, V. 70, 80, 316, 324

Klein, A. H. 144, 148, 313

Klein, S. A. 310, 322

Knoblauch, K. 163, 298

Kobayashi, H. 141, 142, 144, 148, 308

Koch, C. 195, 198, 199, 296, 303, 339

Kodak 54, 298

Kokaram, A. C. 316, 325

Kolas, Ø. 91, 298

Kolpatzik, B. 173, 187, 298, 351

Konik, H. 312, 330

Koo, B-K. 67, 70, 80, 297

Kopperud, S. 152, 290

Kostiuk, I. V. 262, 298

Kotani, K. 235, 301, 310, 323, 324

Kotera, H. 76, 262, 294, 299, 329, 348, 351

Kress, W. C. 128, 316

Kruse, B. 144, 147, 148, 298, 303

Kuo, C. 148, 149, 298

Kuo, C.-C. J. 298, 324

Kusuma, M. 101, 289

Kuznetsov, Y. V. 262, 298

Ladret, P. 173, 187, 232, 287

Lagendijk, R. L. 315, 323

Lai, Y.-K. 298, 324

Lam, E. P. 298, 329

Lambrecht, C. 298, 324

Land, L. 171, 232, 297

Larabi, C. 51, 298

Larabi, M-C. 86, 87, 89, 300

Larson, E. C. 75, 87, 89, 101, 189, 198, 235,

298, 299

Lau, D. L. 299, 347, 348, 351

Le Callet, P. 46, 286, 303, 327

Le Meur, O. 195, 198, 199, 303

LeCun, Y. 193, 284

Lee, B-S. 128, 229, 299

Lee, H-K. 180, 262, 286

Lee, J. 76, 299, 329, 348, 351

Lee, K-Y. 128, 233, 234, 299

Lee, L. 128, 229, 304

Lee, M-S. 299, 328

Leek, M. R. 33, 299

Legge, G. E. 163, 299

273

Page 290: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Leisti, T. 122, 144, 149, 289, 307

Lemahieu, I. 124, 313

Lensu, L. 122, 123, 130, 131, 144, 289, 308

Levine, D. M. 51, 87, 283

Leynadier, C. 72, 185, 284

Li, C. 247, 302

Li, F., S.and Zhang 235, 299

Li, H. 288, 329

Li, J. 298, 324

Li, Q. 187, 314

Li, S. 195, 198, 199, 300, 316, 330

Li, Y. 286, 323

Lim, W. 128, 300

Lin, F-S. 299, 328

Lin, Q. 300, 322

Lin, W. 315, 328, 329

Lin, X. 315, 328

Lindberg, S. 141–144, 146–149, 300, 304

Liu, H. 49, 195, 198, 199, 300, 307

Liu, K. 67, 287, 329

Liu, L-Y. 299, 328

Liu, T. 195, 198, 199, 291

Locke, M. H. 315, 329

Lodge, M. A. 235, 300

Loo, K. C. 298, 329

Loulou, M. 282, 326

Love, S. 141, 142, 144, 148, 236, 251, 287,

307

Lu, L. 89, 189, 313, 314

Lu, Z. K. 315, 328

Lubin, J. 300, 323, 324

Lucassen, M. P. 74, 87, 100, 312, 326

Luebker, A. 163, 299

Lukin, V. 50, 66, 86, 87, 99–101, 187, 262,

289, 307, 328, 329

Lundstrom, C. 50, 300

Luo, M. R. 67, 70, 79, 80, 91, 92, 100, 136,

142, 144, 146, 149, 181, 247, 286, 293,

297, 300, 302, 308, 311, 314, 322, 325,

326

Ma, L. 195, 198, 199, 235, 299, 300, 316,

330

Ma, Y-F. 300, 339

Maalouf, A. 86, 87, 89, 300

MacDonald, L. 70, 80, 141, 142, 144, 145,

148, 284

MacDonald, L.W. 180, 300

Madden, T. E. 161, 292

Mahlab, H. 292, 327

Mani, S. 128, 300

Mantiuk, R. 301, 327

MapTube 159, 301

Marini, D. 307, 334–336, 363

Marini, F. 235, 283, 330

Martinez, O. 236, 251, 307

Masmoudi, N. 282, 308, 326, 328

Matkovic, K. 303, 324

McCabe, G. P. 89, 302

McCann, J. 291, 334, 336

McCann, J. J. 181, 301

McCormick-Goodhart, M. 301, 327

McCulloch, M. 193, 304

McDonald, R. 287, 322

McDonnell, K. T. 161, 313

274

Page 291: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Medeghini, G. 307, 334–336, 363

Meer, P. 287, 343

Melgosa, M. 67, 100, 218, 304, 330, 363

Menegaz, G. 68, 301

Meng, T. H. Y. 286, 323

Mensonen, A. 144, 148, 304

Meyer, G. 283, 325

Michelson, A. 147, 301

Milanfar, P. 201, 219, 309

Miller, R. 76, 312

Miller, R.L. 311, 347

Mindru, F. 301, 329

Ming, Z. 121, 122, 130, 315

Mitsa, T. 301, 322, 323, 348, 351

Miyahara, M. 301, 323, 324

Miyake, Y. 70, 79, 80, 144, 146–148, 198,

282, 289, 294, 301, 326

Miyata, K. 144, 146–148, 301

Mizokami, Y. 171, 232, 315

Monga, V. 301, 347

Montag, E. D. 41, 301

Montgomery, D. C. 48, 301

Moore, D. S. 89, 302

Moorthy, A. K. 195–199, 302

Morimoto, A. 70, 80, 291, 326

Moroney, N. 56, 68, 247, 302

Morovic, J. 40, 91–93, 95, 141–144,

146–148, 180, 300, 302, 326

Mou, X. 86, 87, 89, 173, 187, 302, 316, 330,

331

Movshon, J. A. 70, 81, 302

Mueller, K. 161, 313

Mullen, K. T. 76, 302

Munkberg, J. 303, 329

Murphy, M. 149, 285

Myszkowski, K. 301, 327

Naguib, N. A. 87, 89, 303, 330

Nakaguchi, T. 70, 80, 198, 282

Nakauchi, S. 303, 325

Nakaya, F. 141–144, 146–148, 237, 288

Nasanen, R. 76, 303, 322

Natale-Hoffman, K. 141, 142, 303

Neuhoff, D. L. 145, 304, 325, 351

Neumann, L. 303, 324

Newell, J. 296, 324

Newman, T. 247, 302

Ng, Y. 148, 149, 298

Ngan, K. N. 195, 198, 199, 235, 299, 300,

316, 330

Nicolas, M. 173, 187, 232, 287

Nicolas, R. 147, 312

Niebur, E. 195, 198, 199, 296, 303, 339

Nielsen, B. 196, 303

Nijenhuis, M. R. M. 303, 324

Nilsson, F. 144, 145, 147, 148, 303, 325, 348

Ninassi, A. 195, 198, 199, 303

Norberg, O. 141–144, 146–149, 180, 282,

304

Noyes, Y. X. 128, 229, 304

Nussbaum, P. 18, 67, 70, 79, 80, 99, 107,

108, 113, 142, 189, 198, 304, 306

Nuutinen, M. 54, 65, 159, 293

Nyman, G. 122, 144, 149, 289, 307

Oberti, F. 145, 148, 180, 286

275

Page 292: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

O’Dell, S. 145, 148, 149, 312

Ogden, J. M. 281, 334, 336

Ohno, Y. 18, 304

Oittinen, P. 54, 65, 122, 144, 148, 159, 289,

293, 304

Ojanen, H. 149, 307

Oleari, C. 67, 70, 71, 81, 100, 136, 163, 173,

187, 200, 201, 215, 218, 220, 304, 310,

330, 363

Olives, J-L. 149, 307

Omadani, M. 3, 294

Omamiuda, M. 262, 294

Ong, E. P. 315, 328, 329

Ord, J. K. 86, 87, 136, 174, 189, 190, 297,

352

O’Reilly, R. 126, 305

Orfanidou, M. 173, 187, 304, 329

Osher, S. 215, 308

Otsu, N. 304, 342

Owens, R. A. 315, 324

Pai, M. 193, 304

Pai, N. 193, 304

Pan, Y. 149, 285

Pan, Z. 128, 229, 304

Panetta, K. 310, 329

Pant, D. R. 218, 304

Pappas, T. N. 285, 328

Parish, D. H. 163, 299

Parker, K. J. 316, 324

Pascale, D. 68, 305

Pattanaik, S. N. 315, 328

Patterson, A. 126, 305

Pedersen, M. 5–7, 9, 50, 51, 61, 65, 67, 70,

71, 75, 79–81, 86, 99, 100, 107–111, 113,

114, 136, 137, 139, 142, 144, 146–149,

157, 158, 161, 163–165, 168, 171–174,

180, 187, 189, 190, 198, 200, 201, 215,

219, 220, 237, 281, 285, 290, 292, 293,

305, 306, 310, 311, 330, 331, 336, 345,

348, 360, 364

Pedersen, M. A. 20, 306

Pefferkorn, S. 306, 325

Pekarovicova, A. 144, 148, 315

Peli, E. 147, 215–218, 227, 306, 307, 336

Pellacini, F. 291, 326

Pelli, D. G. 33, 314

Pelz, J. B. 75, 95, 282

Peroche, B. 290, 325

Pesquet-Popescu, B. 283, 326

Philips, W. 124, 313

Piatko, C. 308, 323

Pizlo, Z. 144, 145, 148, 149, 171, 282, 283,

285, 312, 315, 316, 324, 325

Po, L. M. 286, 328

Pointer, M. R. 149, 286

Ponce, J. 193, 284

Ponomarenko, N. 50, 66, 86, 87, 99–101,

187, 262, 289, 307, 328, 329

Purgathofer, W. 303, 324

Qiu, G. 50, 99, 307

Quiroga, M. 236, 251, 307

Radun, J. 149, 307

Rahardja, S. 315, 328

Rahmim, A. 235, 300

276

Page 293: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Rajashekar, U. 51, 199, 307, 313, 330

Rasmussen, D. R. 141–144, 146–148, 237,

288, 303

Raveendran, P. 64, 312

Rawashdeh, N. A. 236, 251, 307

Ray, L. 76, 312

Redi, J. 49, 307

Rees, M. 149, 287

Rigg, B. 67, 287, 300, 322, 325

Rindal, A. 152, 290

Ritala, R. 144, 289

Rizzi, A. 7, 51, 56, 92, 147, 149, 172, 173,

187, 200, 201, 290, 306–308, 311, 317,

334–336, 363

Rohlf, F.J. 129, 311

Rudin, L. I. 215, 308

Rushmeier, H. 308, 323

Russell, R. 62, 308

Rust, B. 308, 323

Saarelma, H. 171, 297

Sabir, M. F. 48, 75, 87, 89, 101, 235, 309

Sadovnikov, A. 144, 308

Safranek, R. J. 308, 322

Saito, R. 76, 299, 329, 348, 351

Samet, A. 282, 308, 326, 328

Sanders, P. 308, 323

Sankar, R. 296, 327

Sankur, B. 62, 63, 282, 326

Sano, C. 67, 308

Santoro, G. 292, 327

Sarlat, A. 135, 136, 154, 284

Sato, M. 141–144, 146–148, 237, 288, 303

Sato, T. 141, 142, 144, 148, 308

Sawyer, J. F. 141, 142, 144, 148, 308

Sayood, K. 62, 63, 282, 326

Scala, P. 163, 317

Schadenko, A. A. 262, 298

Scheermesser, T. 308, 324, 348, 351

Schettini, R. 163, 180, 235, 283, 285, 288,

330

Schmitt, F. 67, 70, 72, 74, 80, 100, 113,

142–144, 146, 148, 149, 180, 181, 185,

284, 329

Schulz, T. 21, 308

Seara, R. 288, 327

Seidel, H.-P. 301, 327

Seim, T. 308, 322

Seo, H. J. 201, 219, 309

Sequeira, J. 68, 301

Seshadrinathan, K. 64, 309

Shaked, D. 149, 297, 309, 328

Shang, X. 142, 193–198, 201, 219, 314

Sharma, A. 160, 230, 244, 309

Sharma, G. 18, 19, 31, 50, 56, 128, 149, 163,

229, 299, 309

Sheikh, H. R. 46, 48, 64, 72, 74–76, 87, 89,

92, 100, 101, 122, 129, 136, 144, 171–173,

187, 193, 200, 201, 231, 235, 262, 309,

314, 327, 328

Shen, J. 215, 286

Shiau, J. 309, 347, 351

Shibata, K. 101, 294

Shin, I-W. 251, 307

Shklyarov, D. 301, 327

277

Page 294: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Shnayderman, A. 173, 187, 309, 326

Shoham, T. 310, 331

Shohara, M. 235, 310

Shur, J. 313, 351

Sigg, F. 25–27, 232, 234, 310, 316

Silva, E. A. 310, 329

Silverstein, D. A. 310, 322

Silvestri, F. 66, 87, 187, 307, 329

Simon, K. 91, 317

Simoncelli, E. P. 61, 64, 72, 74, 76, 87, 89,

92, 100, 122, 129, 136, 144, 171–173, 187,

193, 200, 201, 231, 307, 314, 326, 327,

330

Simone, G. 5–7, 9, 50, 51, 67, 70, 71, 81, 99,

100, 107, 109–111, 113, 114, 136, 147,

149, 172, 173, 187, 200, 201, 215, 220,

281, 306, 308, 310, 311, 330, 331, 335,

336, 363, 364

Sinha, P. 62, 308

Skarsbø, S. E. 142, 229, 293, 304

Sobel, I. 297, 328

Sokal, R.R. 129, 311

Song, T. 67, 308, 311

Sony 53, 311

Southworth, D. 14, 311

Southworth, M. 14, 311

Sparow, I. 56, 311

Spaulding, K. E. 311, 347

Speigle, J. 70, 80, 291, 326

Stamm, T. 56, 311

Stanek, J. C. 128, 311

Steinberg, L. 291, 347, 348, 351

Stone, M. C. 163, 180, 311

Stuart, A. 86, 87, 136, 174, 189, 190, 297,

352

Stucki, P. 311, 347, 351

Styles, W.S. 141, 145, 147, 161, 315

Sullivan, B. J. 284, 322

Sullivan, J. 76, 312

Sun, J. 282, 328

Sun, P. 93, 141–144, 146–148, 302, 326

Sun, P-L. 180, 181, 240, 241, 262, 302, 312

Susstrunk, S. 201, 219, 281, 339

Svec, L. A. 171, 232, 315

Szlyk, J. 163, 298

Tadmor, Y. 312, 333, 335

TagCrowd 160, 312

Tai, X-C. 215, 286

Tastl, I. 52, 149, 150, 293, 309

Taylor, C. C. 312, 324

Taylor, M. M. 33, 312

Teo, P. C. 66, 293, 312, 323

Tharyan, P. 193, 304

Thomas, J. P. 33, 291

Thung, K-H. 64, 312

Thurstone, L. L. 38, 92, 312, 352

Tobias Associates 16, 312

Toet, A. 74, 87, 100, 312, 326

Tolhurst, D. J. 312, 333, 335

Tong, Y. 312, 330

Topfer, K. 145, 148, 149, 312

Torgerson, W. S. 41, 312

Tremeau, A. 147, 312, 330

Triantaphillidou, S. 173, 187, 281, 304, 329

278

Page 295: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Trontelj, H. 313, 351

Troter, A. Le 68, 301

Tsai, P-S. 126, 281

Tse, M-K. 51, 142, 144, 146–148, 292, 313

Tsumura, N. 70, 80, 144, 146–148, 198, 282,

294, 301, 326

Ucar, Y. 292, 327

Udomhunsakul, S. 46, 315

Ulichney, R. 313, 351

Urabe, H. 48–51, 297

Usui, S. 303, 325

Valberg, A. 308, 322

van der Linde, I. 199, 307

Varkur, K.L. 301, 322, 348, 351

Velho, L. 149, 262, 313

Venkatesh, M. V. 236, 288

Veryovka, O. 313, 325, 348, 350

Video Quality Experts Group 89, 313

Vienot, F. 6, 237, 290

Ville, D. Van De 124, 313

Vu, C. 198, 299

Vuori, T. 149, 307

Wah, R. L. 235, 300

Wan, X. 313, 329

Wandell, B. A. 62, 64, 66, 67, 69, 70, 80, 83,

92, 100, 121, 129, 136, 141, 144, 172, 173,

181, 187, 200, 201, 215, 220, 316, 317,

323, 325, 348, 358, 364

Wang, L. 161, 313

Wang, M. 313, 351

Wang, T. 292, 327

Wang, Y. 195, 198, 199, 291

Wang, Z. 46, 61, 63, 64, 66, 67, 71, 72, 74,

76, 86, 87, 89, 92, 100, 101, 122, 129, 136,

142, 144, 171–173, 187, 189, 193–198,

200, 201, 219, 220, 231, 262, 307, 309,

313, 314, 326, 327, 330

Ward, G. 308, 323

Watson, A. B. 33, 35, 314, 322

Webster, M. A. 171, 232, 297, 315

Weed, S. 180, 288

Westen, S. J. P. 315, 323

Westin, P. 141–144, 146–149, 304

Westland, S. 67, 70, 80, 297

Whittle, P. 147, 315

Wilhelm, H. 301, 327

Wilkins, A. 171, 232, 297

Williams, D. 128, 316

Wilson, D. L. 315, 324

Wiseman, J. 313, 351

Wolneykien, P. A. 262, 298

Wolski, M. 180, 315

Wong, P. W. 315, 324

Wright, W. D. 51, 283

Wu, S. Q. 315, 329

Wu, W. 180, 282, 315, 325

Wurgler, D. 21, 315

Wurm, L. H. 163, 299

Wyszecki, G. 141, 145, 147, 161, 315

Xie, D. 313, 329

Xie, S. L. 286, 328

Xu, J. 313, 329

Xu, R. 144, 148, 315, 328

279

Page 296: Doctoral Dissertation by Marius Pedersen

FUTURE WORK

Xue, W. 302, 331

Yamsang, N. 46, 315

Yanfang, X. 121, 122, 130, 315

Yang, C. L. 286, 328

Yang, D. 195, 198, 199, 291

Yang, X. K. 315, 328

Yao, S. 315, 328, 329

Yee, H. 316, 327

Yeh, E. M. 316, 325

Yendrikhovskij, S.N. 13, 163, 180, 316

Yu, Q. 316, 324

Yu, W. 121, 122, 130, 315

Yuasa, M. 51, 288

Zeise, E. K. 25–27, 128, 232, 234, 316

Zepernick, H-J. 101, 289

Zhang, B. 171, 316

Zhang, C. N. 288, 329

Zhang, D. 316, 331

Zhang, F. 316, 330

Zhang, H-J. 300, 339

Zhang, L. 86, 87, 89, 173, 187, 195, 198,

294, 302, 316, 330, 331, 339

Zhang, M. 302, 331

Zhang, X. 62, 64, 66, 67, 69, 70, 80, 83, 92,

100, 121, 129, 136, 141, 144, 172, 173,

181, 187, 200, 201, 215, 220, 316, 317,

323, 325, 348, 358, 364

Zhao, B. 317, 330

Zheng, Y. 5, 9, 200, 219, 306, 331

Zheng, Z-W. 240, 241, 262, 312

Zolliker, P. 56, 91, 161, 172, 173, 187, 200,

201, 283, 311, 313, 317

Zuffi, S. 163, 317

Zunino, R. 49, 307

280

Page 297: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[1] R. Achanta, S. S. Hemami, F. Estrada, and S. Susstrunk. Frequency-tuned salient re-gion detection. In IEEE International Conference on Computer Vision and PatternRecognition CVPR, pages 1597–1604, Miami Beach, Florida, Jun 2009.

[2] T. Acharya and P-S. Tsai. Computational foundations of image interpolation algo-rithms. Ubiquity, 8:1–17, Oct 2007.

[3] E. H. Adelson, C. H. Anderson, J. R. Bergen, P. J. Burt, and J. M. Ogden. Pyramidmethods in image processing. RCA Engineer, 29(6):33–41, Nov/Dec 1984.

[4] S. A. Ajagamelle. Analysis of the difference of gaussians model in perceptual imagedifference metrics. Master thesis, Gjøvik University College and Grenoble Institute ofTechnology, 2009.

[5] S. A. Ajagamelle, M. Pedersen, and G. Simone. Analysis of the difference of gaussiansmodel in image difference metrics. In 5th European Conference on Colour in Graphics,Imaging, and Vision (CGIV), pages 489–496, Joensuu, Finland, Jun 2010. IS&T.

[6] S. A. Ajagamelle, G. Simone, and M. Pedersen. Performance of the difference of gaus-sian model in image difference metrics. In G. Simone, A. Rizzi, and J. Y. Hardeberg,editors, Gjøvik Color Imaging Symposium, number 4 in Høgskolen i Gjøviks rapport-serie, pages 27–30, Gjøvik, Norway, Jun 2009.

[7] E. Allen, S. Triantaphillidou, and R. E. Jacobson. Image quality comparison betweenJPEG and JPEG2000. i. psychophysical investigation. The Journal of Imaging Scienceand Technology, 3(51):248–258, May/Jun 2007.

[8] A. Alsam and I. Farup. Colour gamut mapping as a constrained variational problem. InA-B. Salberg, J. Y. Hardeberg, and R. Jensen, editors, 16th Scandinavian Conferenceon Image Analysis, volume 5575 of Lecture Notes in Computer Science, pages 109–118, Oslo, Norway, Jun 2009. Springer-Verlag.

[9] A. Alsam and I. Farup. Spatial colour gamut mapping by means of anisotropic diffu-sion. In R. Schettini, S. Tominaga, and A. Tremeau, editors, Proceedings of the ThirdInternational Workshop on Computational Color Imaging (CCIW), volume 6626 ofLecture Notes in Computer Science, pages 113–124, Milano, Italy, Apr 2011. Springer.

[10] D. G. Altman and M. J. Gardner. Calculating confidence intervals for regression andcorrelation. British Medical Journal, 296:1238–1242, Apr 1988.

281

Page 298: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[11] K. An, J. Sun, and W. Du. Homogeneity based image objective quality metric. InProceedings of International symposium on Intelligent Signal Processing and Commu-nication Systems, pages 649–652, Hong Kong, Dec 2005.

[12] M. Andersson and O. Norberg. Color gamut - is size the only thing that matters? In2006 TAGA conference, pages 273–291, Vancouver, Canada, Mar 2006.

[13] N. Avadhanam and V. R. Algazi. Evaluation of a human-vision-system-based imagefidelity metric for image compression. In A. G. Tescher, editor, Applications of DigitalImage Processing XXII, volume 3808 of Proc. SPIE, pages 569–579, San Jose, CA,USA, 1999.

[14] I. Avcibas, B. Sankur, and K. Sayood. Statistical evaluation of image quality measures.Journal of Electronic Imaging, 11:206–223, 2002.

[15] P. Axelson. Quality measures of halftoned images. Master’s thesis, Linkoping Univer-sity, 2003.

[16] M. A. B. Ayed, A. Samet, M. Loulou, and N. Masmoudi. A new perceptual qualityassessment for image coding. In Proceedings of Visualization, Imaging, and ImageProcessing (VIIP2002), Malaga, Spain, Sep 2002.

[17] J. S. Babcock, J. B. Pelz, and M. D. Fairchild. Eye tracking observers during rank order,paired comparison, and graphical rating tasks. In Image Processing, Image Quality,Image Capture Systems Conference (PICS), pages 10–15, Rochester, NY, May 2003.

[18] J. Bai, T. Nakaguchi, N. Tsumura, and Y. Miyake. Evaluation of image corrected byretinex method based on S-CIELAB and gazing information. IEICE trans. on Fun-damentals of Electronics, Communications and Computer Sciences, Vol.E89-A(11):2955–2961, 2006.

[19] A. M. Bakke, I. Farup, and J. Y. Hardeberg. Predicting the performance of a spatialgamut mapping algorithm. In R. Eschbach, G. G. Marcu, S. Tominaga, and A. Rizzi,editors, Color Imaging XIV: Displaying, Hardcopy, Processing, and Applications, vol-ume 7241 of Proceedings of SPIE, page 724116, San Jose, CA, Jan 2009.

[20] R. Balasubramanian, R. deQueiroz, R. Eschbach, and W. Wu. Gamut mapping topreserve spatial luminance variations. In Color Imaging Conference, pages 122–128,Scottsdale, AZ, USA, 2000. IS&T.

[21] E. Bando, J. Y. Hardeberg, and D. Connah. Can gamut mapping quality be predicted bycolor image difference formulae? In B. Rogowitz, T. N. Pappas, and S. Daly, editors,Human Vision and Electronic Imaging X, volume 5666 of SPIE Proceedings, pages 180– 191, San Jose, CA, USA, Jan 2005.

[22] E. Bando, J. Y. Hardeberg, D. Connah, and I. Farup. Predicting visible image degra-dation by colour image difference fomulae. In The 5th International Conference onImaging Science and Hardcopy, pages 121–124, Xi’an, China, Aug 2004.

[23] Y. Bang, Z. Pizlo, and J. P. Allebach. Discrimination based banding assessment. InIS&T’s NIP19: International Conference on Digital Printing Technologies, pages 745–750, New Orleans, LA, Sep 2003.

282

Page 299: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[24] Y. Bang, Z. Pizlo, and J. P. Allebach. Perception based hardcopy banding metric. InIS&T’s NIP21: International Conference on Digital Printing Technologies, pages 78–83, Baltimore, MD, Sep 2005.

[25] Z. Baranczuk, P. Zolliker, and J. Giesen. Image quality measures for evaluating gamutmapping. In Color Imaging Conference, pages 21–26, Albuquerque, NM, USA, Nov2009. IS&T/SID.

[26] P. G. J. Barten. Evaluation of subjective image quality with the square-root integralmethod. Journal of the Optical Society of America A, 7:2024–2031, Oct 1990.

[27] P. G. J. Barten. Effects of quantization and pixel structure on the image quality of colormatrix displays. Journal of the Society for Information Display, 1:147 – 153, 1993.

[28] C. J. Bartleson. The combined influence of sharpness and graininess on the quality ofcolor prints. J. Photogr. Sci., 30:33–38, 1982.

[29] C. J. Bartleson. Optical Radiation Measurement. Vol. 5 - Visual Measurements. Aca-demic Press Inc., 1984.

[30] A. Beghdadi and B. Pesquet-Popescu. A new image distortion measure based onwavelet decomposition. In Seventh International Symposium on Signal Processing andIts Applications, volume 1, pages 485–488, Paris, France, Jul 2003.

[31] S. J. Belcher, K. W. Greenshields, and W. D. Wright. Colour vision survey usingthe ishihara, dvorine, bostrm and kugelberg, bostrm, and american-optical hardy-rand-rittler tests. British Journal of Ophthalmology, 42(6):355–359, Jun 1958.

[32] M. L. Berenson and D. M. Levine. Applied statistics: a first course. Prentice-Hall,Inc., Upper Saddle River, NJ, USA, 1988. ISBN 0-130-41476-X.

[33] T. Berk, L. Brownston, and A. Kaufman. A new color-naming system for graphicslanguages. IEEE Computer Graphics and Applications, 2:37–44, 1982.

[34] G. Bestmann. Technical recommendation: Color characterization target and data fileformat for 4-color printing. Technical report, Heidelberger Druckmaschinen AG, Apr2002. Version 1.2.

[35] S. Bianco, G. Ciocca, F. Marini, and R. Schettini. Image quality assessment by pre-processing and full reference model combination. In S. P. Farnand and F Gaykema,editors, Image Quality and System Performance VI, volume 7242, page 72420O, SanJose, CA, Jan 2009.

[36] P. Blomgren and T. F. Chan. Color TV: Total variation methods for restoration ofvector-valued images. IEEE Transactions on Image Processing, 7(3):304–309, Mar1998.

[37] M. Bolin and G. Meyer. A perceptually based adaptive sampling algorithm. In SIG-GRAPH ’98 Conference Proceedings, pages 409–418, Orlando, FL, Jul 1998. ACM.

[38] N. Bonnier. Contribution to Spatial Color Gamut Mapping. PhD thesis, TelecomParistech, 2008.

283

Page 300: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[39] N. Bonnier, P. Green, and A. Sarlat. Evaluating the use of the perceptual referencemedium gamut in ICC printing works. In Seventeenth Color Imaging Conference,pages 227–232, Albuquerque, NM, Nov. 2009. IS&T/SID.

[40] N. Bonnier, C. Leynadier, and F. Schmitt. Improvements in spatial and color adap-tive gamut mapping algorithms. In 4th European Conference on Colour in Graphics,Imaging, and Vision (CGIV), pages 341–346, Barcelona, Spain, Jun 2008. IS&T.

[41] N. Bonnier, F. Schmitt, H. Brettel, and S. Berche. Evaluation of spatial gamut mappingalgorithms. In Color Imaging Conference, pages 56–61, Scottsdale, AZ, Nov 2006.IS&T/SID.

[42] N. Bonnier, F. Schmitt, M. Hull, and C. Leynadier. Spatial and color adaptive gamutmapping: A mathematical framework and two new algorithms. In Color Imaging Con-ference, pages 267–272, Albuquerque, NM, Nov 2007. IS&T.

[43] Y-L. Boureau, J. Ponce, and Y. LeCun. A theoretical analysis of feature pooling invisual recognition. In J. Furnkranz and T. Joachi, editors, 27th International Conferenceon Machine Learning, pages 111–118, Haifa, Isreal, Jun 2010. Omnipress.

[44] A. Bouzerdoum, A. Havstad, and A. Beghdadi. Image quality assessment using aneural network approach. In Proceedings of the Fourth IEEE International Symposiumon Signal Processing and Information Technology, pages 330–333, Rome, Italy, Dec2004.

[45] S. Bouzit and L. MacDonald. Colour difference metrics and image sharpness. In ColorImaging Conference, pages 262–267, Scottsdale, AZ, Nov 2000. IS&T/SID.

[46] A. P. Bradley. A wavelet visible difference predictor. IEEE Transactions on ImageProcessing, 8:717–730, 1999.

[47] A. P. Bradley. Can region of interest coding improve overall perceived image quality?In B. Lovell and A. Maeder, editors, Proceedings of the APRS Workshop on DigitalImage Computing, pages 41–44, Brisbane, Australia, Feb 2003.

[48] J. C. Brailean, B. J. Sullivan, C. T. Chen, and M. L. Giger. Evaluating the EM al-gorithm for image processing using a human visual fidelity criterion. In InternationalConference on Acoustics, Speech and Signal Processing (ICASSP’91), volume 4, pages2957–2960, Toronto, Ontario, Canada, Apr 1991. IEEE.

[49] G. J. Braun and M. D. Fairchild. Gamut mapping for pictorial images. In TAGAProceedings: An Analysis of Sheetfed Lithographic Print Attributes, pages 645–660,Rochester, NY, Jan 1999.

[50] G. J. Braun, M. D. Fairchild, C. F. Carlson, and F. Ebner. Color gamut mapping in ahue-linearized CIELAB color space. In 6th Color Imaging Conference, pages 163–168,Scottsdale, AZ, USA, Nov 1998. IS&T/SID.

[51] J. C. Briggs, E. Hong, and D. Forrest. Analysis of ghosting in electrophotography. InInternational Conference on Digital Printing Technologies (NIP16), pages 403–407,Vancouver, British Columbia, Canad, Nov. 2000. IS&T.

284

Page 301: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[52] J. C. Briggs, M. Murphy, and Y. Pan. Banding characterization for inkjet printing. InImage Processing, Image Quality, Image Capture, Systems Conference, pages 84–88,Portland, OR, Mar 2000. IS&T.

[53] A. C. Brooks and T. N. Pappas. Structural similarity quality metrics in a coding con-text: exploring the space of realistic distortions. In B. E. Rogowitz, T. N. Pappas, andS. J. Daly, editors, Human Vision and Electronic Imaging XI, volume 6057 of SPIEProceedings, pages 299–310, San Jose, CA, USA, Feb 2006.

[54] H. Buring, P. G. Herzog, and E. Jung. Evaluation of current color management tools:Image quality assessment. In The European Conference on Colour Graphics, Imagingand Vision, pages 459–462, Aachen, Germany, Apr 2004. IS&T.

[55] N. Burningham, Z. Pizlo, and J. P. Allebach. Encyclopedia of Imaging Science andTechnology, chapter Image Quality Metrics, pages 598–616. Wiley, New York, 2002.

[56] A. J. Calabria and M. D. Fairchild. Perceived image contrast and observer preferencei. the effects of lightness, chroma, and sharpness manipulations on contrast perception.The Journal of Imaging Science and Technology, 47(6):479–493, 2003.

[57] A. J. Calabria and M. D. Fairchild. Perceived image contrast and observer preferenceII. empirical modeling of perceived image contrast and observer preference data. TheJournal of imaging science and technology, 47:494–508, 2003.

[58] P. Le Callet and F. Autrusseau. Subjective quality assessment IRCCyN/IVC database,2005. 18/07/11: http://www.irccyn.ec-nantes.fr/ivcdb/.

[59] P. Le Callet and D. Barba. A robust quality metric for color image quality assessment.In International Conference on Image Processing (ICIP), volume 1, pages 437–440,Barcelona, Spain, Sep 2003. IEEE.

[60] J. Canny. A computational approach to edge detection. IEEE Transactions on PatternAnalysis and Machine Intelligence, 8(6):679–698, 1986.

[61] G. Cao, M. Pedersen, and Z. Baranczuk. Saliency models as gamut-mapping artifactdetectors. In 5th European Conference on Colour in Graphics, Imaging, and Vision(CGIV), pages 437–443, Joensuu, Finland, Jun 2010. IS&T.

[62] A. Capra, A. Castrorina, S. Corchs, F. Gasparini, and R. Schettini. Dynamic range op-timization by local contrast correction and histogram image analysis. In InternationalConference on Consumer Electronics (ICCE ’06), pages 309 – 310, Las Vegas, Jan2006.

[63] V. Caracciolo. Just noticeable distortion evaluation in color images. Master’s thesis,Gjøvik University College and Roma Tre University, 2009. 18/07/11: http://colorlab.no/content/download/25901/274793/file/Caracciolo2009_Master_Thesis.pdf.

[64] N. Cardin. L’utilisation du perceptual reference medium gamut dans la gestion descoleours ameliore-t-elle la qualite d’images produites par impression jet d’encre? Mas-ter’s thesis, Ecole Nationale Superieure Louis-Lumiere Promotion Photographie, 2009.

285

Page 302: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[65] M. Carnec, P. Le Callet, and D. Barba. An image quality assessment method basedon perception of structural information. In IEEE International Conference on ImageProcessing (ICIP2003), pages 185–188, Barcelona, Spain, Sep 2003.

[66] J. Caviedes and F. Oberti. A new sharpness metric based on local kurtosis, edge andenergy information. Signal Processing: Image Communication, 19:147–161, 2004.

[67] N. Chaddha and T. H. Y. Meng. Psycho-visual based distortion measures formonochrome image and video compression. In Proc. of the Asilomar Conference onSignals, Systems and Computers, volume 2, pages 841–845, Pacific Grove, CA, Nov1993. IEEE.

[68] T. F. Chan and J. Shen. Image Processing And Analysis: Variational, PDE, Wavelet,And Stochastic Methods. Society for Industrial and Applied Mathematics, Philadelphia,PA, USA, 2005. ISBN 089871589X.

[69] D. M. Chandler and S. S. Hemami. Online supplement to ”VSNR: A visual signal-to-noise ratio for natural images based on near-threshold and suprathreshold vision”.18/02/11: http://foulard.ece.cornell.edu/dmc27/vsnr/vsnr_a57.pdf.

[70] D. M. Chandler and S. S. Hemami. VSNR: A wavelet-based visual signal-to-noiseratio for natural images. IEEE Transactions on Image Processing, 16(9):2284–2298,Sep 2007.

[71] G. H. Chen, C. L. Yang, L. M. Po, and S. L. Xie. Edge-based structural similarity forimage quality assessment. In Proceedings of International Conference in Acoustics,Speech and Signal Processing, volume 2, pages 933–936, Toulouse, France, May 2006.IEEE.

[72] G. H. Chen, C. L. Yang, and S. L. Xie. Gradient-based structural similarity for imagequality assessment. In IEEE International Conference on Image Processing, pages2929–2932, Atlanta, GA, USA, October 2006.

[73] K. Chen and X-C. Tai. A nonlinear multigrid method for total variation minimizationfrom image restoration. Journal of Scientific Computing, 33:115–138, 2007. ISSN0885-7474.

[74] S. Chen, A. Beghdadi, and A. Chetouani. Color image assessment using spatial ex-tension to CIE DE2000. In The International Conference on Consumer Electronics(ICCE), pages 1–2, Las Vegas, CA, USA, Jan 2008. ISBN: 978-1-4244-1458-1.

[75] Y-H. Cho, Y-T. Kim, H-K. Lee, and Y-H. Ha. Image dependent gamut mapping basedon color variations at each segmented region. In International conference on Digitalproduction printing and industrial applications, pages 26–27, Barcelona, Spain, May2003. IS&T.

[76] S. Y. Choi, M. R. Luo, and M. R. Pointer. Modelling image naturalness. In CGIV2008 - Fourth European Conference on Color in Graphics, Imaging and Vision, pages126–131, Terrassa, Spain, Jun 2008. IS&T.

[77] C. Chou and Y. Li. A perceptually tuned subband image coder based on the measure ofjust-noticeable-distortion profile. IEEE Tranactions on circuits and systems for videotechnology, 5:467–476, 1995.

286

Page 303: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[78] C. Chou and K. Liu. A fidelity metric for assessing visual quality of color images.In Proceedings of 16th International Conference on Computer Communications andNetworks (ICCCN), pages 1154–1159, Honolulu, HI, Aug 2007. IEEE.

[79] R. Chung and M. Rees. A survey of digital offset print quality issues. Technical report,Rochester Institute of Technology, 2007.

[80] CIE. Guidelines for the evaluation of gamut mapping algorithms. Technical ReportISBN: 3-901-906-26-6, CIE TC8-03, 156:2004.

[81] CIE. Chromatic adaptation under mixed illumination condition when comparing soft-copy and hardcopy images. Technical Report ISBN: 3-901-906-34-7, CIE TC8-04,162:2004.

[82] CIE. Colorimetry, 2nd ed. publication CIE 15.2, Bureau Central de la CIE, 1986.

[83] CIE. Industrial colour-difference evaluation, publication CIE 116-95, bureau central dela CIE, 1995.

[84] CIE. CIE 15:2004: Colorimetry. Technical report, 2004. 3rd edition.

[85] CIE. Colorimetry. Technical Report 15, 2004.

[86] CIE. A colour appearance model for colour management systems: CIECAM02. Tech-nical Report 159, CIE TC 8-01, Vienna, 2004. ISBN 978 3 901906 29 9.

[87] F. J. J. Clarke, R. McDonald, and B. Rigg. Modification to the JPC 79 colour differenceformula. Journal of the Society of Dyers and Colourists, 100:128–132, Apr 1984.

[88] D. Comaniciu and P. Meer. Mean shift: a robust approach toward feature space analy-sis. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 24(5):603–619,2002.

[89] F. Crete, T. Dolmiere, P. Ladret, and M. Nicolas. The blur effect: perception andestimation with a new no-reference perceptual blur metric. In B. E. Rogowitz, T. N.Pappas, and S. J. Daly, editors, Human Vision and Electronic Imaging XII, volume 6492of Proceedings of SPIE, page 64920I, San Jose, CA, Mar 2007.

[90] P. B. Crosby. The puzzlement of quality. Technical report, Philip Crosby Asso-ciates II, Sep 1982. 28/06/11: http://www.wppl.org/wphistory/PhilipCrosby/ThePuzzlementOfQuality.pdf.

[91] C. Cui. Comparison of two psychophysical methods for image color quality mea-surements: Paired comparision and rank order. In Color Imaging Conference, pages222–227, Scottsdale, AZ, Nov 2000. IS&T/SID.

[92] C. Cui. Gamut mapping with enhanced chromaticness. In Ninth Color Imaging Confer-ence: Color Science and Engineering: Systems, Technologies, and Applications, pages257–262, Scottsdale, AZ, USA, 2001.

[93] C. Cui, D. Cao, and S. Love. Measuring visual threshold of inkjet banding. In ImageProcessing, Image Quality, Image Capture, Systems Conference (PICS), pages 84–89,Montreal, Quebec, Canada, Apr 2001. IS&T.

287

Page 304: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[94] C. Cui and S. Weed. Colorfulness?-in search of a better metric for gamut limit colors. InImage Processing, Image Quality, Image Capture, Systems Conference (PICS), pages183–187, Portland, OR, Mar 2000. IS&T.

[95] L. Cui and A. R. Allen. An image quality metric based on a colour appearance model.In 10th International Conference on Advanced Concepts for Intelligent Vision Systems,volume 5259 of Lecture Notes In Computer Science, pages 696 – 707, Berlin, 2008.Springer-Verlag.

[96] C. Cusano, F. Gasparini, and R. Schettini. Image annotation for adaptive enhancementof uncalibrated color images. In S. Bres and R.Laurini, editors, Visual Informationand Information Systems, volume 3736 of Lecture Notes in Computer Science, pages216–225, Amsterdam, Netherlands, Jul 2006. Springerlink.

[97] E. N. Dalal, D. R. Rasmussen, F. Nakaya, P. A. Crean, and M. Sato. Evaluating theoverall image quality of hardcopy output. In Image Processing, Image Quality, ImageCapture, Systems Conference, pages 169–173, Portland, OR, May 1998. IS&T.

[98] S. Daly. The visible differences predictor: an algorithm for the assessment of imagefidelity. In Digital images and human vision, pages 179–206. MIT Press, Cambridge,MA, USA, 1993. ISBN 0-262-23171-9.

[99] N. Damera-Venkata, T. D. Kite, W. S. Geisler, B. L. Evans, and A. C. Bovik. Im-age quality assessment based on a degradation model. IEEE Transactions on ImageProcessing, 9:636–650, 2000.

[100] R. de Freitas Zampolo and R. Seara. A measure for perceptual image quality assess-ment. In International Conference on Image Processing (ICIP), volume 1, pages 433–436, Barcelona, Spain, Sep 2003.

[101] R. de Freitas Zampolo and R. Seara. Perceptual image quality assessment based onbayesian networks. In International Conference on Image processing (ICIP2004),pages 329–332, Singapore, Oct 2004.

[102] H. de Ridder. Minkowski-metric as a combination rule for digital-image-coding im-pariments. In B.E. Rogowitz, editor, Human Vision, Visual Processing, and DigitalDisplay III, volume 1666 of Proceedings of SPIE, pages 16–26, San Jose, CA, Feb1992.

[103] G. Deffner, M. Yuasa, and D. Arndt. Evaluation of display image quality: experts vs.non-experts. In Symp Soc Inf Disp Dig, volume 25, pages 475–478, 1994.

[104] J. Dijk. In search of an Objective Measure for the Perceptual Quality of Printed Images.PhD thesis, Technische Universiteit Delft, 2004.

[105] K. D. Donahue, C. Cui, and M. V. Venkatesh. Wavelet analysis of print defects. InImage Processing, Image Quality, Image Capture, Systems Conference (PICS), pages42–47, Portland, OE, Apr 2002. IS&T.

[106] W. Dong, Q. Yu, C. N. Zhang, and H. Li. Image quality assessment using rough fuzzyintegrals. In Proceedings of the 27th International Conference on Distributed Comput-ing Systems Workshops (ICDCSW’07), pages 1–5, Washington, DC, USA, 2007. ISBN0-7695-2881-3.

288

Page 305: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[107] F. Dugay. Perceptual evaluation of colour gamut mapping algorithms. Master thesis,Perceptual Evaluation of colour gamut mapping algorithms. Gjøvik University Collegeand Grenoble Institute of Technology, 2007.

[108] F. Dugay, I. Farup, and J. Y. Hardeberg. Perceptual evaluation of color gamut mappingalgorithms. Color Research & Application, 33(6):470–476, Dec 2008.

[109] ECI. New ”ECI 2002” target and conversion tools. 30/06/11: http://www.eci.org/doku.php?id=en:projects:eci2002.

[110] T. Eerola, J-K. Kamarainen, L. Lensu, and H. Kalviainen. Framework for applyingfull reference digital image quality measures to printed images. In A-B. Salberg, J. Y.Hardeberg, and R. Jenssen, editors, Scandinavian Conference on Image Analysis, vol-ume 5575 of Lecture Notes in Computer Science, pages 99–108, Oslo, Norway, June2009. Springer Berlin / Heidelberg.

[111] T. Eerola, L. Lensu, H. Kalviainen, J-K. Kamarainen, T. Leisti, G. Nyman, R. Halonen,and P. Oittinen. Full reference printed image quality: Measurement framework andstatistical evaluation. Journal of Imaging Science and Technology, 54(1):010201–(13),Jan/Feb 2010.

[112] T. Eerola, L. Lensu, J-K. Kamarainen, T. Leisti, R. Ritala, G. Nyman, andH. Kalviainen. Bayesian network model of overall print quality: Construction andstructural optimisation. Pattern Recogn. Lett., 32:1558–1566, August 2011. ISSN0167-8655. URL http://dx.doi.org/10.1016/j.patrec.2011.04.006.

[113] K. Egiazarian, J. Astola, N. Ponomarenko, V. Lukin, F. Battisti, and M. Carli. Twonew full-reference quality metrics based on HVS. In Proceedings of the Second Inter-national Workshop on Video Processing and Quality Metrics, pages 22–24, Scottsdale,USA, Jan 2006.

[114] C. Endo, T. Asada, H. Haneishi, and Y. Miyake. Analysis of the eye movements and itsapplications to image evaluation. In IS&T and SID’s 2nd Color Imaging Conference:Color Science, Systems and Applications, pages 153–155, Scottsdale, AZ, Nov 1994.

[115] P. G. Engeldrum. Image quality modeling: Where are we? In Image Processing, ImageQuality, Image Capture, Systems Conference (PICS), pages 251–255, Savannah, GA,Apr 1999. IS&T.

[116] P. G. Engeldrum. Psychometric Scaling, a toolkit for imaging systems development.Imcotek Press Winchester USA, 2000.

[117] P. G. Engeldrum. Psychometric scaling: Avoiding the pitfalls and hazards. In Im-age Processing, Image Quality, Image Capture Systems Conference, pages 101–107,Montreal, Quebec, Canada, Apr 2001. IS&T.

[118] P. G. Engeldrum. A theory of image quality: The image quality circle. Journal ofImaging Science and Technology, 48(5):446–456, sep/okt 2004.

[119] U. Engelke, M. Kusuma, H-J. Zepernick, and M. Caldera. Reduced-reference metricdesign for objective perceptual quality assessment in wireless imaging. Image Com-mun., 24:525–547, Aug 2009. ISSN 0923-5965.

289

Page 306: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[120] A. M. Eskicioglu, P. S. Fisher, and S. Chen. Image quality measures and their perfor-mance. In IEEE Transactions on Communications, volume 43, pages 2959–2965, Dec1995.

[121] European Space Agency. http://www.esa.int, 2010.

[122] M. D. Fairchild. Image quality measurement and modeling for digital photography. InInternational Congress on Imaging Science, pages 318–319, Tokyo, Japan, May 2002.

[123] M. D. Fairchild. Still photography throwdown: Silver halide vs. silicon. In Color andImaging Conference, pages 154–159, San Antonio, TX, Nov 2010. IS&T.

[124] M. D. Fairchild and G. M. Johnson. Meet iCAM: A next-generation color appear-ance model. In Color Imaging Conference, pages 33–38, Scottsdale, AZ, Nov 2002.IS&T/SID.

[125] M. D. Fairchild and G. M. Johnson. The iCAM framework for image appearance,image differences, and image quality. Journal of Electronic Imaging, 13:126–138,2004.

[126] K. Falkenstern, N. Bonnier, H. Brettel, M. Pedersen, and F. Vienot. Using image qualitymetrics to evaluate an ICC printer profile. In Color and Imaging Conference, pages244–249, San Antonio, TX, Nov 2010. IS&T and SID.

[127] K. Falkenstern, N. Bonnier, H. Brettel, M. Pedersen, and F. Vienot. Weighing qualityattributes. In R. C. Baraas, editor, 21st symposium of the international colour visionsociety (ICVS), page 88, Kongsberg, Norway, Jul 2011. ISBN 978-82-8261-009-4.

[128] K. Falkenstern, N. Bonnier, M. Pedersen, H. Brettel, and F. Vienot. Using metrics toassess the ICC perceptual rendering intent. In F. Gaykema and S. Farnand, editors,Image Quality and System Performance, volume 7867 of Proceedings for SPIE, pages786706–1–15, San Francisco, CA, Jan 2011.

[129] Z. Fan. A simple modification of error-diffusion weights. In Proceedings of SPIE,pages 113–114, Boston, MA, May 1992.

[130] J. P. Farrugia and B. Peroche. A perceptual image metric in computer graphics. In In-ternational conference on color in graphics and image processing (CGIP2000), pages13–17, Saint-Etienne, France, Oct 2000.

[131] I. Farup, C. Gatta, and A. Rizzi. A multiscale framework for spatial gamut mapping.IEEE Transactions on Image Processing, 16(10):2423–2435, 2007.

[132] I. Farup, J. Y. Hardeberg, A. M. Bakke, S. Kopperud, and A. Rindal. Visualizationand interactive manipulation of color gamuts. In Color Imaging Conference, pages250–255, Scottsdale, Arizona, Nov 2002. www.colorlab.no/icc3d.

[133] G. T. Fechner. Elemente der Psychophysik, volume 2. Breitkopf und Hartel, Leipzig,1860. (Reprinted, Bristol: Thoemmes Press, 1999).

[134] E. A. Fedorovskaya. Image quality as a problem of computational vision. In ImageProcessing, Image Quality, Image Capture, Systems Conference (PICS), pages 22–28,Rochester, NY, May 2003. IS&T.

290

Page 307: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[135] E. A. Fedorovskaya, F. J. J. Blommaert, and H. de Ridder. Perceptual quality of colorimages of natural scenes transformed in CIELUV color space. In Color Imaging Con-ference, pages 37–40, Scottsdale, AZ, Nov 1993. IS&T/SID.

[136] S. Feeny, P. K. Kaiser, and J. P. Thomas. An analysis of data gathered by the staircase-method. The American Journal of Psychology, 79(4):652–654, Dec 1966.

[137] X. Feng, T. Liu, D. Yang, and Y. Wang. Saliency based objective quality assessment ofdecoded video affected by packet losses. In IEEE International Conference on ImageProcessing, pages 2560–2563, San Diego, CA, Oct 2008. IEEE.

[138] X-F. Feng, J. Speigle, and A. Morimoto. Halftone quality evaluation using color visualmodels. In Image Processing, Image Quality, Image Capture, Systems Conference(PICS), pages 5–10, Portland, Oregon, USA, Apr 2002. IS&T.

[139] S. R. Fernandez and M. D. Fairchild. Observer preferences and cultural differencesin color reproduction of scenic images. In Color Imaging Conference, pages 66–72,Scottsdale, AZ, Nov 2002. IS&T.

[140] S. R. Fernandez, M. D. Fairchild, and K. Braun. Analysis of observer and cultural vari-ability while generating ”preferred” color reproductions of pictorial images. Journal ofImaging Science and Technology, 49(1):96–104, Jan/Feb 2005.

[141] J. A. Ferwerda. Psychophysics 101: How to design perception experiments. Shortcourse notes - Eighteenth Color and Imaging Conference., Nov 2010. San Antonio,TX.

[142] J. A. Ferwerda and F. Pellacini. Functional difference predictors (FDPs): measuringmeaningful image differences. In Conference Record of the Thirty-Seventh AsilomarConference on Signals, Systems and Computers, volume 2, pages 1388 – 1392, PacificGroove, CA, Nov 2003.

[143] G. G. Field. Test image design guidelines for color quality evaluation. In Color ImagingConference, pages 194–196, Scottsdale, AZ, Nov 1999. IS&T.

[144] R.W. Floyd and L. Steinberg. An adaptive algorithm for spatial grayscale. In Proceed-ings of the Society for Information Displays, volume 17, pages 75–77, 1976.

[145] J. Frankle and J. McCann. Method and apparatus for lightness imaging. United StatesPatent No. 4,384,336, 1983.

[146] P. Franti. Blockwise distortion measure for statistical and structural errors in digitalimages. Signal Processing: Image Communication, 13:89–98, 1998.

[147] C. Fredembach, F. Estrada, and S. Susstrunk. Segmenting memory colours. InIS&T/SID 16th Color Imaging Conference, pages 315–320, 2008.

[148] T. Frese, C. A. Bouman, and J. P. Allebach. A methodology for designing imagesimilarity metrics based on human visual system models. Technical Report TR-ECE97-2, Purdue University, West Lafayette, IN, USA., 1997.

[149] J.S. Friedman. History of Color Photography. Boston : American Photographic Pub-lishing, 1947.

291

Page 308: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[150] X. Gao, T. Wang, and J. Li. A content-based image quality metric. In Rough Sets,Fuzzy Sets, Data Mining, and Granular Computing, volume 3642 of Lecture notes incomputer science, pages 231–240. Springer Berlin / Heidelberg, 2005.

[151] D. A. Garvin. What does ”product quality” really mean? Sloan Management Review,Oct 1984.

[152] G. Gast and M-K. Tse. A report on a subjective print quality survey conducted atNIP16. In NIP17: International conference on Digital Printing Technologies, pages723–727, Fort Lauderdale, FL, Oct 2001.

[153] D. Gayle, H. Mahlab, Y. Ucar, and A. M. Eskicioglu. A full-reference color imagequality measure in the DWT domain. In 13th European Signal Processing Conference,EUSIPCO, page 4, Antalya, Turkey, Sep 2005.

[154] E. J. Giorgianni and T. E. Madden. Digital Color Management: Encoding Solutions.Addison-Wesley Longman Publishing Co., Inc., Boston, MA, USA, 2nd edition, 1998.

[155] A. Golan and H. Hel-Or. Novel workflow for image-guided gamut mapping. Journalof Electronic Imaging, 17(3):033004–1–11, 2008.

[156] M. Gong and M. Pedersen. Pooling for color image quality. Journal of Visual Commu-nication and Image Representation, 2011. Submitted.

[157] Google 3D Warehouse. http://sketchup.google.com/3dwarehouse/, 2010.

[158] P. Gorley and N. Holliman. Stereoscopic image quality metrics and compression. InAndrew J. Woods, Nicolas S. Holliman, and John O. Merritt, editors, StereoscopicDisplays and Applications XIX, volume 6803, page 680305, San Jose, CA, USA, Jan2008.

[159] E. M. Granger. A comparison of color difference data and formulas. In TAGA 60thAnnual Technical Conference, pages 191–202, San Francisco, CA, USA, March 2008.

[160] P. J. Green. Gamut mapping and appearance models in colour management. PhDthesis, University of Derby, UK., 2003.

[161] P. J. Green. Gamut mapping and appearance models in graphic arts colour manage-ment. PhD thesis, University of Derby, 2003.

[162] P. J. Green and L. MacDonald, editors. Colour Engineering: Achieving Device Inde-pendent Colour. John Wiley & Sons, 2002.

[163] J. Grice and J. P. Allebach. The print quality toolkit: An integrated print quality as-sessment tool. Journal of Imaging Science and Technology, 43(2):187–199, Mar/Apr1999.

[164] B. Grunbaum. The search for symmetric Venn diagrams. Geombinatorics, 8:104 –109, 1999.

[165] I. Guarneri, M. Guarnera, A. Bosco, and G. Santoro. A perceptual quality metric forcolor-interpolated images. In R. Rasmussen and Y. Miyake, editors, Image Quality andSystem Performance II, volume 5668 of SPIE Proceedings, pages 61–69, San Jose, CA,USA, Jan 2005.

292

Page 309: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[166] M. R. Gupta and J. Bowen. Ranked dither for robust color printing. In R. Eschbachand G. G. Marcu, editors, Color Imaging XII: Processing, Hardcopy, and Applications,Proceedings of SPIE, page 424, San Jose, CA, Jan 2007.

[167] R. Halonen, M. Nuutinen, R. Asikainen, and P. Oittinen. Development and measure-ment of the goodness of test images for visual print quality evaluation. In S. P. Farnandand F. Gaykema, editors, Image Quality and System Performance VII, volume 7529,pages 752909–1–10, San Jose, CA, USA, Jan 2010. SPIE.

[168] J. Y. Hardeberg. Color image quality for multi-function peripherals. In PICS 2002:Image Processing, Image Quality, Image Capture Systems Conference, pages 76–81,Portland, OE, Apr 2002.

[169] J. Y. Hardeberg, E. Bando, and M. Pedersen. Evaluating colour image difference met-rics for gamut-mapped images. Coloration Technology, 124(4):243–253, Aug 2008.

[170] J. Y. Hardeberg and S. E. Skarsbø. Comparing color image quality of 4 digital presses.In Proceedings of the 11th International Printing and Graphic Arts Conference, Bor-deaux, France, Oct 2002. 8 pages.

[171] D. Hasler and S. Susstrunk. Measuring colorfulness in natural images. In B. E. Ro-gowitz and T. N. Pappas, editors, Human Vision and Electronic Imaging VIII, volume5007 of Proceedings of SPIE, pages 87–95, San Jose, CA, Jan 2003.

[172] D. J. Heeger and P. C. Teo. A model of perceptual image fidelity. In Proceedingsof the International Conference on Image Processing (ICIP), volume 2, page 2343,Washington, DC, USA, Okt 1995. IEEE. ISBN 0-8186-7310-9.

[173] D. W. Hertel. Exploring s-CIELAB as a scanner metric for print uniformity. In R. Ras-mussen and Y. Miyake, editors, Image Quality and System Performance II, volume5668 of SPIE Proceedings, pages 51–60, San Jose, CA, Jan 2005.

[174] J. T. Heverhagen. Noise measurement and estimation in MR imaging experiments.Radiology, 245(3):638 639, Dec 2007.

[175] I. E. Heynderickx and S. Bech. Image quality assessment by expert and non-expertviewers. In B. E. Rogowitz and T. N. Pappas, editors, Human Vision and ElectronicImaging VII, volume 4662 of Proceedings of SPIE, pages 129–137, San Jose, CA, Jan2002.

[176] J. Holm, I. Tastl, and T. Johnson. Definition & use of the ISO 12640-3 referencecolor gamut. In Color Imaging Conference, pages 62–68, Scottsdale, AZ, Nov 2006.IS&T/SID.

[177] G. Hong and M. R. Luo. Perceptually based colour difference for complex images. InR. Chung and A. Rodrigues, editors, 9th Congress of the International Colour Asso-ciation, volume 4421 of SPIE proceedings, pages 618–621, Rochester, NY, USA, Jun2002.

[178] G. Hong and M. R. Luo. New algorithm for calculating perceived colour difference ofimages. Imaging Science Journal, 54(2):86–91, 2006.

293

Page 310: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[179] Y. Horita, K. Shibata, and Y. Kawayoke. Mict image quality evaluation database.18/02/2011: http://mict.eng.u-toyama.ac.jp/mictdb.html.

[180] X. Hou and L. Zhang. Saliency detection: A spectral residual approach. In In IEEEConference on Computer Vision and Pattern Recognition (CVPR07). IEEE ComputerSociety, pages 1–8, Minneapolis, MN, Jun 2007.

[181] C. Hung-Shing, M. Omamiuda, and H. Kotera. Adaptive gamut mapping method basedon image-to-device. In NIP15: International Conference on Digital Printing Technolo-gies, pages 346–349, Orlando, FL, USA, Oct 1999.

[182] R. W. G. Hunt. The reproduction of Color in Photography, Printing and Television.Fountain, Tolworth, UK, 1987.

[183] F. H. Imai, N. Tsumura, and Y. Miyake. Perceptual color difference metric for compleximages based on mahalanobis distance. Journal of Electronic Imaging, 10:385–393,Apr 2001.

[184] J. Imai and M. Omadani. Reasons why we prefer reading on paper rather than displays:Studies for seeking paper-like readability on electronic paper. The Journal of ImagingScience and Technology, 52(5):1–5, Sep-Oct 2008.

[185] International Color Consortium. White paper 26: Using thesRGB v4 ICC preference.icc profile. 28/3/11: http://color.org/ICC_White_Paper_26_Using_the_V4_sRGB_ICC_profile.pdf, .

[186] International Color Consortium. White paper 6: ICC version 2 & version 4 display pro-file differences. 28/3/11: http://color.org/ICC_White_Paper_26_Using_the_V4_sRGB_ICC_profile.pdf, .

[187] International Color Consortium. White paper number 2: Perceptual rendering intentuse case issues. 16/03/09: www.color.org, Jan 2005.

[188] International Imaging Industry Association. CPIQ initiative phase 1 white paper: Fun-damentals and review of considered test methods, Oct 2007. 15/02/2011: http://www.i3a.org/resources.

[189] International Telecommunication Union. Recommendation ITU-R P.800: Methodsfor subjective determination of transmission quality. Technical report, InternationalTelecommunication Union/ITU Radiocommunication Sector, 1996.

[190] International Telecommunication Union. Recommendation ITU-R BT.1676: Method-ological framework for specifying accuracy and cross-calibration of video quality met-rics. Technical report, 2004.

[191] International Telecommunication Union. Recommendation ITU-R BT.500-11:Methodology for the subjective assessment of the quality of television pictures. Techni-cal report, International Telecommunication Union/ITU Radiocommunication Sector,2009.

[192] R. Iordache and A. Beghdadi. A wigner-ville distribution-based image dissimilaritymeasure. In Proceedings of the 6th International Symposium on Signal Processingand Its Applications (ISSPA ’01), volume 2, pages 430–433, Kuala Lumpur, Malaysia,August 2001.

294

Page 311: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[193] ISO. ISO 12647: Graphic technology – process control for the manufacture of half-tonecolour separations, proof and production prints.

[194] ISO. ISO 5-4: Photography - density measurements - part 4: geomtric conditions forreflection density, 1995.

[195] ISO. ISO 12640-1: Graphic technology - prepress digital data exchange - CMYKstandard colour image data (cmyk/scid), 1997.

[196] ISO. ISO 11056 (E). sensory analysis - methodology - magnitude estimation method,1999.

[197] ISO. 13660: Information technology - office equipment - measurement of image qual-ity attributes for hardcopy output - binary monochrome text and graphic images, 2001.

[198] ISO. ISO 12647-5: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 5: Screen printing, 2001.

[199] ISO. ISO 12640-2: Graphic technology - prepress digital data exchange - part 2:XYZ/sRGB encoded standard colour image data (XYZ/SCID), 2004.

[200] ISO. ISO 12647-1: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 1: Parameters and measure-ment methods, 2004.

[201] ISO. ISO 12647-2: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 2: Offset lithographic pro-cesses, 2004.

[202] ISO. ISO 20462-1 photography - psychophysical experimental methods to estimateimage quality - part 1: Overview of psychophysical elements, jul 2004.

[203] ISO. ISO 20462-2 photography - psychophysical experimental methods to estimateimage quality - part 2: Triplet comparison method, jul 2004.

[204] ISO. ISO 20462-3 photography - psychophysical experimental methods to estimateimage quality - part 2: Quality ruler method, jul 2004.

[205] ISO. ISO 12231: Photography - electronic still picture imaging - vocabulary, 2005.

[206] ISO. ISO 12647-3: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 3: Coldset offset lithographyon newsprint, 2005.

[207] ISO. ISO 12647-4: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 4: Publication gravure print-ing, 2005.

[208] ISO. ISO 12647-6: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 6: Flexographic printing,2006.

[209] ISO. ISO 12640-3 graphic technology - prepress digital data exchange - part 3:CIELAB standard colour image data (CIELAB/SCID), 2007.

295

Page 312: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[210] ISO. ISO 12647-7: Graphic technology – process control for the production of half-tone colour separations, proof and production prints – part 7: Proofing processes work-ing directly from digital data, 2007.

[211] ISO. ISO 3664:2009 graphic technology and photography – viewing conditions, 2009.

[212] ISO. ISO 12640-4 graphic technology - prepress digital data exchange - part 4: Widegamut display-referred standard colour image data [Adobe RGB (1998)/SCID], 2011.

[213] ISO/IEC. 13660 - office equipment - measurment of image quality attributes for hard-copy output (binary monochrome text and graphic images), 1995. 7th working draft.

[214] L. Itti, C. Koch, and E. Niebur. A model of saliency-based visual attention for rapidscene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11):1254–1259, Nov 1998.

[215] G. Ivkovic and R. Sankar. An algorithm for image quality assessment. In InternationalConference on Acoustics, Speech, and Signal Processing, (ICASSP ’04), pages 713–716, Montreal, Canada, May 2004. IEEE.

[216] R. E. Jacobson. An evaluation of image quality metrics. J. Photograph. Sci., 43:7–16,1995.

[217] T. J. W. M. Janssen. Computational Image Quality. PhD thesis, Technische UniversiteitEindhoven, 1999.

[218] J. F. Jarvis, C. N. Judice, and W. H. Hinke. A survey of techniques for the display ofcontinuous tone pictures on bilevel display. Computer Graphics and Image Processing,5:13–40, 1976.

[219] E. W. Jin, X-F. Feng, and J. Newell. The development of a color visual difference model(CVDM). In Image Processing, Image Quality, Image Capture, Systems Conference,pages 154–158, Portland, Oregon, May 1998. IS&T.

[220] E. W. Jin and S. Field. A groundtruth database for testing objective metrics for imagedifference. In Image Processing, Image Quality, Image Capture, Systems Conference(PICS), pages 114–119, Rochester, NY, May 2003. IS&T.

[221] G. M. Johnson. Measuring images: differences, quality and appearance. PhD thesis,Rochester Institute of Technology, Mar 2003.

[222] G. M. Johnson and M. D. Fairchild. Sharpness rules. In 8th Color Imaging Conference,pages 24–30, Scottsdale, AZ, USA, Nov 2000. IS&T/SID.

[223] G. M. Johnson and M. D. Fairchild. Darwinism of color image difference models. InColor Imaging Conference, pages 108–112, Scottsdale, AZ, Nov 2001.

[224] G. M. Johnson and M. D. Fairchild. On contrast sensitivity in an image differencemodel. In Image Processing, Image Quality, Image Capture, Systems Conference(PICS), pages 18–23, Portland, OR, Apr 2002.

[225] G. M. Johnson and M. D. Fairchild. The effect of opponent noise on image quality.In R. Rasmussen and Y. Miyake, editors, Image Quality and System Performance II,volume 5668 of Proceedings of SPIE, pages 82–89, San Jose, CA, Jan 2004.

296

Page 313: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[226] I. Juricevic, L. Land, A. Wilkins, and M. A. Webster. Visual discomfort and naturalimage statistics. Perception, 39(7):884–899, 2010.

[227] T. Kadir and M. Brady. Saliency, scale and image description. International Journal ofComputer Vision, 45(2):83–105, 2001.

[228] S. A. Karunasekera and N. G. Kingsbury. A distortion measure for image artifacts basedon human visual sensitivity. In Proc. of the International conference on Acoustics,Speech and Signal Processing, volume 5, pages 117–120, Adelaide, SA, Australia,Apr 1994. IEEE.

[229] J. Katajamaki and H. Saarelma. Objective quality potential measures of natural colorimages. Journal of Imaging Science and Technology, 42(3):250–263, May/Jun 1998.

[230] B. W. Keelan. Handbook of Image Quality: Characterization and Prediction. MarcelDekker, New York, 2002.

[231] B. W. Keelan and H. Urabe. ISO 20462, a psychophysical image qualitt measurementstandard. In Y. Miyake and D. R. Rasmussen, editors, Image Quality and System Per-formance, volume 5294 of Proceedings of SPIE, pages 181–189, San Jose, CA, Jan2004.

[232] M. G. Kendall, A. Stuart, and J. K. Ord. Kendall’s Advanced Theory of Statistics: Clas-sical inference and relationship, volume 2. A Hodder Arnold Publication, 5 edition,1991.

[233] J-S. Kim, M-S. Cho, and B-K. Koo. Experimental approach for human perceptionbased image quality assessment. In R. Harper, M. Rauterberg, and M. Combetto, ed-itors, International Conference on entertainment computing, volume 4161 of Lecturenotes in computer science, pages 59–68, Cambridge, UK, Sep 2006. Springer Berlin /Heidelberg.

[234] J-S. Kim, M-S. Cho, S. Westland, and M. R. Luo. Image quality assessment for photo-graphic images. In AIC Colour 05 - 10th Congress of the International Colour Associ-ation, pages 1095–1098, Granada, Spain, May 2005.

[235] Y. J. Kim, Y. Bang, and H-K. Choi. Modelling vividness perception for colour laserprinter evaluation. In Color Imaging Conference, pages 299–303, Portland, OR, Nov2008. IS&T.

[236] R. Kimmel, D. Shaked, M. Elad, and I. Sobel. Space-dependent color gamut mapping:a variational approach. Image Processing, IEEE Transactions on, 14(6):796–803, 2005.

[237] H. Kipphan. Handbook of Print Media: Technologies and Production Methods.Springer-Verlag New York, Inc., Secaucus, NJ, USA, 2006. ISBN 3540335706.

[238] M. Klaman. The influence of paper whiteness and other parameters on the creatingof ICC-profiles for digital colour printers and conventional offset presses. In IARIGAI:Advances in Printing Science and Technology - Advances in Digital Printing, volume 6,Munich, Germany, Sep 1999. 13 pages.

[239] M. Klaman. Aspects on Colour Rendering, Colour Prediction and Colour Control inPrinted Media. PhD thesis, Stockholm University, 2002.

297

Page 314: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[240] K. Knoblauch, A. Arditi, and J. Szlyk. Effects of chromatic and luminance contrast onreading. Journal of the Optical Society of America A, 8:428–439, Feb 1991.

[241] Kodak. Kodak lossless true color image suite. http://r0k.us/graphics/kodak/.Visited 01/07/11.

[242] Ø. Kolas and I. Farup. Efficient hue-preserving and edge-preserving spatial gamutmapping. In Fifteenth Color Imaging Conference, pages 207–212, Albuquerque, NewMexico, Nov 2007.

[243] B. Kolpatzik and C. A. Bouman. Optimized error diffusion for high-quality imagedisplay. Journal of Electronic Imaging, 1(3):277–292, Jul 1992.

[244] B. Kolpatzik and C. A. Bouman. Optimal universal color palette design for error diffu-sion. Journal of Electronic Imaging, 4(2):131–143, Apr 1995.

[245] B. Kruse. The Characterization of Gloss from detailed Semi Goniometric Measure-ments. In TAGA, pages 307–313, San Diego, CA, Mar 2010.

[246] C. Kuo, Y. Ng, and D. A. Gusev. Perceptual color contouring detection and qualityevaluation using scanners. In International Conference on Digital Printing Technolo-gies (NIP), pages 581–585, New Orleans, LA, Sep 2003.

[247] Y. V. Kuznetsov, A. A. Schadenko, I. V. Kostiuk, and P. A. Wolneykien. The process-ing of fine detail for digital halftone printing. In IS&T’s NIP20: 2004 InternationalConference on Digital Printing Technologies, pages 632–637, Salt Lake City, UT, Oct2004.

[248] Y.-K. Lai, J. Guo, and C.-C. J. Kuo. Perceptual fidelity measure of digital color images.In B. E. Rogowitz and T. N. Pappas, editors, Human Vision and Electronic Imaging III,volume 3299 of Proc. SPIE, pages 221–231, San Jose, CA, Jul 1998.

[249] Y.-K. Lai, C.-C. J. Kuo, and J. Li. New image compression artifact measure usingwavelets. In J. Biemond and E. J. Delp, editors, Visual Communications and ImageProcessing, volume 3024 of Proc. SPIE, pages 897–908, San Jose, CA, Jan 1997.

[250] E. P. Lam and K. C. Loo. An image similarity measure using homogeneity regionsand structure. In S. P. Farnand and F Gaykema, editors, Image Quality and SystemPerformance V, volume 6808 of Proceedings of SPIE, pages 680811–680811–9, SanJose, CA, Jan 2008.

[251] C. Lambrecht and J. E. Farrell. Perceptual quality metric for digitally coded colorimages. In VIII European Signal Processing Conference EUSIPCO, IEEE proceedings,pages 1175–1178, Trieste, Italy, Sep 1996.

[252] C. Larabi. Subjective quality assessment of color images. In G. Simone, J. Y. Hard-eberg, and I. Farup, editors, The CREATE 2010 Conference, pages 373–377, Gjøvik,Norway, Jun 2010. ISBN: 978-82-91313-46-7.

[253] E. C. Larson and D. M. Chandler. Unveiling relationships between regions of interestand image fidelity metrics. In W. A. Pearlman, J. W. Woods, and L. Lu, editors, Vi-sual Communications and Image Processing, volume 6822 of SPIE Proceedings, pages68222A–68222A–16, San Jose, CA, Jan 2008.

298

Page 315: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[254] E. C. Larson and D. M. Chandler. Most apparent distortion: A dual strategy for full-reference image quality assessment. In S. Farnand and F. Gaykema, editors, Imagequality and system performance VI, volume 7242 of Proc. of SPIE, page 72420S, SanJose, CA, Jan 2009.

[255] E. C. Larson and D. M. Chandler. Most apparent distortion: full-reference image qual-ity assessment and the role of strategy. Journal of Electronic Imaging, 19(1):011006,Mar 2010.

[256] E. C. Larson, C. Vu, and D. M. Chandler. Can visual fixation patterns improve imagefidelity assessment? In International Conference on Image Processing, pages 2572–2575, San Diego, CA, Oct 2008. IEEE.

[257] D. L. Lau and G. R. Arce. Modern Digital Halftoning. Dekker, 2000.

[258] P. Le Callet and D. Barba. Robust approach for color image quality assessment. InT. Ebrahimi & T. Sikora, editor, Visual Communications and Image Processing, volume5150 of Proceedings of SPIE, pages 1573–1581, Lugano, Italy, Jul 2003.

[259] B.-S. Lee, R. Bala, and G. Sharma. Novel scanner characterization method for colormeasurement and diagnostics applications. In C. A. Bouman, E. L. Miller, and I. Pollak,editors, Computational Imaging IV, volume 6065 of Proc. of SPIE, pages 309–320, SanJose, CA, February 2006.

[260] B-S. Lee, R. Bala, and G. Sharma. Scanner characterization for color measurement anddiagnostics. Journal of Electronic Imaging, 16(4):043009, Oct-Dec 2007.

[261] J. Lee, T. Horiuchi, R. Saito, and H. Kotera. Digital color image halftone: Hybrid errordiffusion using the mask perturbation and quality verification. The Journal of ImagingScience and Technology, 51(5):391–401, Sep/Oct 2007.

[262] K-Y. Lee, Y. Bang, and S. H. Kim. ISO 29112 - edge blurriness & raggedness. Pre-sentation for the image quality assessment working group of ISO/IEC JTC1/SC28 onoffice equipment, Dec 2009.

[263] M-S. Lee, L-Y. Liu, and F-S. Lin. Image similarity comparison using dual-tree wavelettransform. In Pacific Rim Symposium Advances in Image and Video Technology, vol-ume 4319 of Lecture Notes in Computer Science, pages 189–197, Hsinchu, Taiwan,Dec 2006. Springer-Verlag Berlin Heidelberg.

[264] M. R. Leek. Adaptive procedures in psychophysical research. Perception & Psy-chophysics, 63(8):1279–1292, 2001.

[265] G. E. Legge, D. H. Parish, A. Luebker, and L. H. Wurm. Psychophysics of reading.XI. Comparing color contrast and luminance contrast. Journal of the Optical Societyof America A, 7:2002–2010, Oct 1990.

[266] F. Li, S.and Zhang, L. Ma, and K. N. Ngan. Image quality assessment by separatelyevaluating detail losses and additive impairment. IEEE Transactions on Multimedia,2011. Accepted for publication.

299

Page 316: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[267] W. Lim and S. Mani. Application of digital imaging to measure print quality. InInternational Conference on Digital Printing Technologies, pages 611–614, Toronti,Ontario, Canada, Oct 1998. IS&T.

[268] Q. Lin. Halftone image quality analysis based on a human vision model. In J. P.Allebach and B. E. Rogowitz, editors, Human Vision, Visual Processing, and DigitalDisplay IV, volume 1913 of Proceedings of SPIE, pages 378–389, San Jose, CA, Sep1993.

[269] S. Lindberg. Perceptual determinants of print quality. PhD thesis, Stockholm Univer-sity, 2004.

[270] H. Liu and I. E. Heynderickx. Studying the added value of visual attention in objectiveimage quality metrics based on eye movement data. In International Conference onImage Processing, pages 3097–3100, Cairo, Egypt, Nov 2009. IEEE.

[271] M. A. Lodge, A. Rahmim, and R. L. Wah. Simultaneous measurement of noise andspatial resolution in PET phantom images. Physics in Medicine and Biology, 55(4):1069–1081, Jan 2010.

[272] J. Lubin. Vision Models for Target Detection and Recognition, chapter A visual dis-crimination model for imaging systems design and evaluation, pages 245 – 283. WorldScientific Publishing Co, Inc., River Edge, NJ, USA, 1995. Singapore: World Scien-tific.

[273] J. Lubin. Sarnoff JND vision model: Algorithm description and testing. Technicalreport, Sarnoff Corporation, 1997. 5/10/2007: ftp://ftp.its.bldrdoc.gov/dist/ituvidq/old2/jrg003.rtf.

[274] C. Lundstrom. Measuring digital image quality. Technical report, Center for medicalimage science and visualization, Linkoping University, 2006. 04/06/09: http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-5586.

[275] M. R. Luo, G. Cui, and B. Rigg. The development of the CIE 2000 colour-differenceformula: CIEDE2000. Color Research and Application, 26(5):340–350, 2001.

[276] M. R. Luo and B. Rigg. BFD(l:c) colour-difference formula: Part 1 - development ofthe formula. Journal of the Society of Dyers and Colourists, 103:86–94, 1987.

[277] L. Ma, S. Li, and K. N. Ngan. Visual Horizontal Effect for Image Quality Assessment.IEEE Signal Processing Letters, 17(7):627–630, Jul 2010.

[278] Y-F. Ma and H-J. Zhang. Contrast-based image attention analysis by using fuzzy grow-ing. In 11th International Conference on Multimedia, pages 374–381, Berkeley, CA,Nov 2003. ACM. ISBN 1581137222.

[279] A. Maalouf and M-C. Larabi. A no reference objective color image sharpness metric.In 18th European Signal Processing Conference (EUSIPCO-2010), pages 1019–1022,Aalborg, Denmark, August 2010.

[280] L.W. MacDonald and J. Morovic. Assessing the effects of gamut compression in thereproduction of fine art paintings. In Proceedings of 3rd IS&T/SID Color ImagingConference, pages 194–200, Springfield, VA, 1995.

300

Page 317: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[281] R. MacDonald, editor. Colour Physics for Industry. Society if Dyers and Colourists, 2edition, 1997.

[282] R. Mantiuk, K. Myszkowski, and H.-P. Seidel. Visible difference predicator for highdynamic range images. In International Conference on Systems, Man and Cybernetics,volume 3, pages 2763–2769, Hague, Netherlands, Oct 2004. IEEE.

[283] MapTube. http://maptube.org/, 2010.

[284] J. J. McCann. Color gamut mapping using spatial comparisons. In R. Eschbach andG. G. Marcu, editors, Color Imaging: Device-Independent Color, Color Hardcopy, andGraphic Arts VI, volume 4300 of Proceedings of SPIE, page 126130, San Jose, CA,Jan 2001.

[285] M. McCormick-Goodhart, H. Wilhelm, and D. Shklyarov. A ”retained image appa-rance” metric for full tonal scale, colorimetric evaluation of photographic image stan-dard stability. In NIP20: International Conference on Digital Printing Technologies,pages 680–688, Salt Lake City, UT, Oct 2004. IS&T.

[286] G. Menegaz, A. Le Troter, J. Sequeira, and J.M. Boi. A discrete model for colornaming. EURASIP Journal on Advances in Signal Processing, 1, 2007. 10 pages.

[287] A. Michelson. Studies in Optics. University of Chicago Press, 1927.

[288] F. Mindru and J. Jung. Structure and color based comparison of halftone images. InInternational Congress of Imaging Science (ICIS06), pages 629–632, Rochester, NewYork, May 2006.

[289] T. Mitsa and J. R. Alford. Single-channel versus multiple-channel visual models forthe formulation of image quality measures in digital halftoning. In R. Eschbach, editor,Recent Progress in Digital Halftoning, volume 1, pages 14–16. IS&T, 1994.

[290] T. Mitsa and K.L. Varkur. Evaluation of contrast sensitivity functions for the formula-tion of quality measures incorporated in halftoning algorithms. In International Conf.Acostics, speech, signal processing: image and multidimensional signal processing,volume 5, pages 301–304, Minneapolis, MN, USA, Apr 1993. IEEE.

[291] M. Miyahara, K. Kotani, and R. Algazi. Objective picture quality scale (PQS) for imagecoding. IEEE Transactions on Communications, 46(9):1215–1226, Sep 1996.

[292] K. Miyata, N. Tsumura, H. Haneishi, and Y. Miyake. Subjective image quality formulti-level error diffusion and its objective evaluation method. Journal of ImagingScience and Technology, 43(2):170–177, Mar/Apr 1999.

[293] V. Monga, N. Damera-Venkata, and B. L. Evans. Image halftoning by error diffusion:A survey of methods for artifact reduction. Visisted 28/07/08: www.ece.utexas.edu,2003.

[294] E. D. Montag. Empirical formula for creating error bars for the method of pairedcomparison. Journal of Electronic Imaging, 15(1):222–230, Jan-Mar 2006.

[295] D. C. Montgomery. Design and Analysis of Experiments. John Wiley and Sons, NewYork, 2nd edition, 1984.

301

Page 318: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[296] D. S. Moore and G. P. McCabe. Introduction ot the practice of statistics. W. H. Freemanand Company, New York, 1989.

[297] A. K. Moorthy and A. C. Bovik. Perceptually significant spatial pooling techniques forimage quality assessment. In D. E. Rogowitz and T. N. Pappas, editors, Human Visionand Electronic Imaging XIV, volume 7240 of Proceedings of SPIE, page 724012, SanJose, CA, Jan 2009.

[298] A. K. Moorthy and A. C. Bovik. Visual importance pooling for image quality assess-ment. IEEE Journal of Selected Topics in Signal Processing, 3(2):193–201, Apr 2009.

[299] N. Moroney. A radial sampling of the OSA uniform color scales. In Color ImagingConference: Color Science and Engineering Systems, Technologies, and Applications,pages 175–180, Scottsdale, AZ, Nov 2003. IS&T.

[300] N. Moroney. Unconstrained web-based color naming experiment. In R. Eschbach andG. G. Marcu, editors, Color Imaging: Device-Independent Color, Color Hardcopy, andGraphic Arts VIII, volume 5008 of Proceedings of SPIE, pages 36–46, Santa Clara,CA, Jan 2003.

[301] N. Moroney, M. D. Fairchild, R. W.G. Hunt, C. Li, M. R. Luo, and T. Newman. TheCIECAM02 color appearance model. In Color Imaging Conference, pages 23–27,Scottsdale, AZ, Nov 2002. IS&T/SID.

[302] J. Morovic. To Develop a Universal Gamut Mapping Algorithm. PhD thesis, Universityof Derby, 1998.

[303] J. Morovic and M. R. Luo. The fundamentals of gamut mapping: A survey. Journal ofImaging Science and Technology, 45(3):283–290, 2001.

[304] J. Morovic and P. Sun. Visual differences in colour reproduction and their colorimet-ric correlates. In Color Imaging Conference, pages 292–297, Scottsdale, AZ, 2002.IS&T/SID.

[305] J. Morovic and P. Sun. Predicting image differences in color reproduction from theircolorimetric correlates. The Journal of imaging science and technology, 47:509–516,2003.

[306] J. Morovic and P-L. Sun. Methods for investigating the influence of image characteris-tics on gamut mapping. In Color Imaging Conference, pages 138–143, Scottsdale, AZ,Nov 1999. IS&T.

[307] X. Mou, M. Zhang, W. Xue, and L. Zhang. Image quality assessment based on edge.In F. H. Imai, F. Xiao, J. M. DiCarlo, N. Sampat, and S. Battiato, editors, DigitalPhotography VII, volume 7876 of Proceedings of SPIE, page 78760N, San Francisco,CA, Jan 2011.

[308] J. A. Movshon and L. Kiorpes. Analysis of the development of spatial sensitivity inmonkey and human infants. Journal of Optical Society of America A, 5:2166–2172,Dec 1988.

[309] K. T. Mullen. The contrast sensitivity of human colour vision to red-green and blue-yellow chromatic gratings. The Journal of Physiology, 359:381–400, 1985.

302

Page 319: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[310] J. Munkberg, P. Clarberg, J. Hasselgren, and T. Akenine-Moller. High dynamic rangetexture compression for graphics hardware. ACM Transactions on Graphics, 25(3):698–706, July 2006.

[311] N. A. Naguib, A. E. Hussein, H. A. Keshk, and M. I. El-Adawy. Contrast error dis-tribution measurement for full reference image quality assessment. In InternationalConference on Computer Theory and Applications, page 4, Alexandria, Egypt, Oct2008.

[312] S. Nakauchi, S. Hatanaka, and S. Usui. Color gamut mapping based on a perceptualimage difference measure. Color Research and Application, 24:280 –291, 1999.

[313] R. Nasanen. Visibility of halftone dot textures. IEEE Transactions on Systems, Man,and Cybernetics, 14:920–924, 1984.

[314] K. Natale-Hoffman, E. N. Dalal, D. R. Rasmussen, and M. Sato. Effect of selected im-age quality on overall preference. In Image Processing, Image Quality, Image Capture,Systems Conference (PICS), pages 266–269, Savannah, GA, Apr 1999.

[315] L. Neumann, K. Matkovic, and W. Purgathofer. Perception based color image differ-ence. Technical Report TR-186-2-97-21, Institute of Computer Graphics and Algo-rithms, Vienna University of Technology, Vienna, Austria, December 1997.

[316] E. Niebur and C. Koch. The Attentive Brain. MIT Press, Cambridge MA, Octorber1995.

[317] B. Nielsen, F. Albregtsen, and H. E. Danielsen. Statistical nuclear texture analysis incancer research: A review of methods and applications. Critical Reviews in Oncogen-esis, 14(2):89–164, 2008.

[318] M. R. M. Nijenhuis and F. J. J. Blommaert. Perceptual-error measure and its applicationto sampled and interpolated single-edged images. Journal of the Optical Society ofAmerica A, 14:2111–2127, September 1997.

[319] F. Nilsson. Objective quality measures for halftoned images. Journal of the OpticalSociety of America A, 16:2151–2162, 1999.

[320] F. Nilsson and B. Kruse. Objective quality measures of halftoned images. In Interna-tional Conference on Digital Printing Conference, pages 569–573, Seattle, WA, Nov1997. IS&T.

[321] A. Ninassi, O. Le Meur, P. Le Callet, and D. Barba. Does where you gaze on an imageaffect your perception of quality? applying visual attention to image quality metric. InInternational Conference on Image Processing, pages 169–172, San Antonio, TX, Sep2007. IEEE.

[322] A. Ninassi, O. Le Meur, P. Le Callet, and D. Barba. On the performance of humanvisual system based image quality assessment metric using wavelet domain. In B. E.Rogowitz and T. N. Pappas, editors, Human Vision and Electronic Imaging XIII, vol-ume 6806 of Proceedings of SPIE, pages 680610–680610–12, San Jose, CA, Jan 2008.

303

Page 320: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[323] O. Norberg, P. Westin, S. Lindberg, M. Klaman, and L. Eidenvall. A comparison ofprint quality between digital, offset and flexographic printing presses performed ondifferent paper qualities. In International Conference on Digital Production Printingand Industrial Applications, pages 380–385, Antwerp, Belgium, May 2001. IS&T.

[324] P. Nussbaum. Colour measurement and print quality assessment in a colour managedprinting workflow. PhD thesis, University of Oslo, 2011.

[325] P. Nussbaum, J. Y. Hardeberg, and S. E. Skarsbø. Print quality evaluation for govern-mental purchase decisions. In M. Lovrecek, editor, Advances in Printing Science andTechnology: proceedings of 31st international iarigai research conference, volume 31,pages 189–200. Acta Graphica publishers, 2004.

[326] Y. Ohno. CIE fundamentals for color measurements. In International Conference onDigital Printing Technologies, pages 540–545, Vancouver, Canada, Oct 2000. IS&T.

[327] P. Oittinen, I. Huttu, and A. Mensonen. End user preferences for quality of on-demandprinted products. In International Conference on Digital Printing Technologies, pages147 – 152, San Antonio, TX, Oct 1996. IS&T.

[328] C. Oleari. Misurare il colore. Hoepli, Milano, Italy, Oct 2008.

[329] C. Oleari, M. Melgosa, and R. Huertas. Euclidean color-difference formula for small-medium color differences in log-compressed OSA-UCS space. Journal of the OpticalSociety of America A, 26(1):121–134, 2009.

[330] C. Oleari, M. Melgosa, and R. Huertas. Euclidean color-difference formula for small-medium color differences in log-compressed OSA-UCS space. Journal of the OpticalSociety of America A, 26(1):121–134, 2009.

[331] M. Orfanidou, S. Triantaphillidou, and E. Allen. Predicting image quality using amodular image difference model. In S. P. Farnand and F Gaykema, editors, ImageQuality and System Performance V, volume 6808 of SPIE Proceedings, pages 68080F–68080F–12, San Jose, CA, Jan 2008.

[332] N. Otsu. A Threshold Selection Method from Gray-level Histograms. IEEE Transac-tions on Systems, Man and Cybernetics, 9(1):62–66, 1979.

[333] M. Pai, M. McCulloch, J. D. Gorman, N. Pai, W. Enanoria, G. Kennedy, P. Tharyan,and J. M. Colford. Systematic reviews and meta-analyses: an illustrated, step-by-stepguide. National Medical Journal of India, 17(2):86–95, 2004. ISSN 0970-258X.

[334] Z. Pan, Y. X. Noyes, J. Y. Hardeberg, L. Lee, and G. Healey. Color scanner char-acterization with scan targets of different media types and printing mechanisms. InR. Eschbach & G. G. Marcu, editor, Color Imaging: Device-Independent Color, ColorHardcopy, and Graphic Arts VI, volume 4300 of Proc. of SPIE, pages 58–63, San Jose,CA, December 2000.

[335] D. R. Pant and I. Farup. Riemannian formulation and comparison of color differenceformulas. Color Research and Application, 2011. (Accepted).

[336] T. N. Pappas and D. L. Neuhoff. Least-squares model-based halftoning. IEEE Trans-actions on Image Processing, 8(8):1102–1116, Aug 1999.

304

Page 321: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[337] D. Pascale. AN-7 the optical society of america uniform color scales (OSA UCS).Technical report, Babelcolor, Sep 2009. 21/01/10: http://www.babelcolor.com/download/AN-7%20The%20OSA%20UCS.pdf.

[338] A. Patterson, R. Kennedy, R. O’Reilly, and B. F. Keegan. Field test of a novel, low-cost, scanner-based sediment profile imaging camera. Limnology and OceanographyMethods, 4(3265):30–37, 2006.

[339] M. Pedersen. Importance of region-of-interest on image difference metrics. Master’sthesis, Gjøvik University College, 2007.

[340] M. Pedersen. 111 full-reference image quality metrics and still not good enough? InG. Simone, A. Rizzi, and J. Y. Hardeberg, editors, Proceedings from Gjøvik ColorImaging Symposium 2009, number 4 in Høgskolen i Gjøviks rapportserie, page 4,Gjøvik, Norway, Jun 2009.

[341] M. Pedersen. Objective image quality assessment of color prints. In G. Simone,J. Y. Hardeberg, and I. Farup, editors, The CREATE 2010 Conference, pages 146–150,Gjøvik, Norway, Jun 2010. ISBN: 978-82-91313-46-7.

[342] M. Pedersen, F. Albregtsen, and J. Y. Hardeberg. Detection of worms in error diffu-sion halftoning. In S. P. Farnand and F. Gaykema, editors, Image Quality and SystemPerformance VI, volume 7242, page 72420L, San Jose, CA, USA, Jan 2009. SPIE.

[343] M. Pedersen and S. A. Amirshahi. Framework the evaluation of color prints usingimage quality metrics. In 5th European Conference on Colour in Graphics, Imaging,and Vision (CGIV), pages 75–82, Joensuu, Finland, Jun. 2010. IS&T.

[344] M. Pedersen, N. Bonnier, F. Albregtsen, and J. Y. Hardeberg. Towards a new imagequality model for color prints. In ICC Digital Print Day, Mar 2009.

[345] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of a newimage quality model for color prints. In Color Imaging Conference, pages 204–209,Albuquerque, NM, USA, Nov 2009. IS&T.

[346] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Attributes of imagequality for color prints. Journal of Electronic Imaging, 19(1):011016–1–13, Jan 2010.

[347] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Estimating print qualityattributes by image quality metrics. In Color and Imaging Conference, pages 68–73,San Antonio, TX, USA, Nov 2010. IS&T/SID.

[348] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Validation of qualityattributes for evaluation of color prints. In Color and Imaging Conference, pages 74–79, San Antonio, TX, USA, Nov 2010. IS&T/SID.

[349] M. Pedersen, N. Bonnier, J. Y. Hardeberg, and F. Albregtsen. Image quality metricsfor the evaluation of print quality. In F. Gaykema and S. Farnand, editors, Image Qual-ity and System Performance, volume 7867 of Proceedings of SPIE, pages 786702–1–786702–19, San Francisco, CA, Jan 2011.

305

Page 322: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[350] M. Pedersen and J. Y. Hardeberg. Rank order and image difference metrics. In 4thEuropean Conference on Colour in Graphics, Imaging, and Vision (CGIV), pages 120–125, Terrassa, Spain, Jun 2008. IS&T.

[351] M. Pedersen and J. Y. Hardeberg. A new spatial hue angle metric for perceptual im-age difference. In Computational Color Imaging, volume 5646 of Lecture Notes inComputer Science, pages 81–90, Saint Etienne, France, Mar 2009. Springer Berlin /Heidelberg.

[352] M. Pedersen and J. Y. Hardeberg. Shame: A new spatial hue angle metric for perceptualimage difference. In Vision Sciences Society 9th Annual Meeting, Naples, FL, USA,May 2009.

[353] M. Pedersen and J. Y. Hardeberg. SHAME: A new spatial hue angle metric for percep-tual image difference. Journal of Vision, 9(8):343, 8 2009. ISSN 1534-7362.

[354] M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality met-rics. Høgskolen i Gjøviks rapportserie 5, The Norwegian Color Research Laboratory(Gjøvik University College), Jun 2009. ISSN: 1890-520X.

[355] M. Pedersen and J. Y. Hardeberg. Survey of full-reference image quality metrics: Clas-sification and evaluation. Foundations and Trends in Computer Graphics and Vision,2011. Submitted.

[356] M. Pedersen, J. Y. Hardeberg, and P. Nussbaum. Using gaze information to improveimage difference metrics. In B. Rogowitz and T. Pappas, editors, Human Vision andElectronic Imaging VIII, volume 6806 of SPIE proceedings, page 680611, San Jose,CA, USA, Jan 2008.

[357] M. Pedersen, A. Rizzi, J. Y. Hardeberg, and G. Simone. Evaluation of contrast measuresin relation to observers perceived contrast. In CGIV 2008 - Fourth European Confer-ence on Color in Graphics, Imaging and Vision, pages 253–256, Terrassa, Spain, Jun2008. IS&T.

[358] M. Pedersen, G. Simone, M. Gong, and I. Farup. A total variation based color im-age quality metric with perceptual contrast filtering. In International conference onPervasive Computing, Signal Processing and Applications, Gjøvik, Norway, Sep 2011.

[359] M. Pedersen, Y. Zheng, and J. Y. Hardeberg. Evaluation of image quality metrics forcolor prints. In A. Heyden and F. Kahl, editors, Scandinavian Conference on ImageAnalysis, volume 6688 of Lecture Notes in Computer Science, pages 317–326, YstadSaltsjobad, Sweden, May 2011. Springer-Verlag Berlin Heidelberg.

[360] M. A. Pedersen. Internationalt Standardiseret Grafisk Produktion ISO 12647. GrafiskArbejdsgiverforening, 2008.

[361] S. Pefferkorn and J-L. Blin. Perceptual quality metric of color quantization errors onstill images. In B. E. Rogowitz and T. N. Pappas, editors, Human Vision and ElectronicImaging III, volume 3299 of SPIE Proceedings, pages 210–220, San Jose, CA, Jan1998.

[362] E. Peli. Contrast in complex images. Journal of the Optical Society of America A, 7:2032–2040, 1990.

306

Page 323: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[363] E. Peli. Contrast sensitivity function and image discrimination. Journal of the OpticalSociety of America A, 18(2):283–193, 2001.

[364] N. Ponomarenko, F. Battisti, K. Egiazarian, J. Astola, and V. Lukin. Metrics perfor-mance comparison for color image database. In Fourth international workshop on videoprocessing and quality metrics for consumer electronics, Scottsdale, AZ, Jan 2009. 6pages.

[365] N. Ponomarenko, V. Lukin, K. Egiazarian, J. Astola, M. Carli, and F. Battisti. Colorimage database for evaluation of image quality metrics. In International Workshopon Multimedia Signal Processing, pages 403–408, Cairns, Queensland, Australia, Oct2008.

[366] N. Ponomarenko, F. Silvestri, K. Egiazarian, M. Carli, J. Astola, and V. Lukin. Onbetween-coefficient contrast masking of DCT basis functions. In Third InternationalWorkshop on Video Processing and Quality Metrics for Consumer Electronics VPQM-07, pages 1–4, Scottsdale, Arizona, USA, Jan 2007.

[367] G. Qiu and A. Kheiri. Social image quality. In S. P. Farnand and F. Gaykema, editors,Image Quality and System Performance VIII, volume 7867 of Proceedings of SPIE,page 78670S, San Francisco, CA, Jan 2011.

[368] J. Radun, T. Leisti, J. Hakkinen, H. Ojanen, J-L. Olives, T. Vuori, and G. Nyman.Content and quality: Interpretation-based estimation of image quality. ACM Trans.Appl. Percept., 4(4):1–15, 2008. ISSN 1544-3558.

[369] U. Rajashekar, I. van der Linde, A. C. Bovik, and L. K. Cormack. GAFFE: A gaze-attentive fixation finding engine. IEEE Transactions on Image Processing, 17(4):564–573, Apr 2008.

[370] U. Rajashekar, Z. Wang, and E. P. Simoncelli. Quantifying color image distortionsbased on adaptive spatio-chromatic signal decompositions. In Internatial Conferenceon Image Processing, pages 2213–2216, Cairo, Egypt, Nov 2009. IEEE.

[371] N. A. Rawashdeh, O. Martinez, M. Quiroga, and K. D. Donohue. Characterization ofprinter banding in regions of complex image content. In SoutheastCon, pages 433–438,Richmond, VA, Mar 2007. IEEE. ISBN: 1-4244-1029-0.

[372] N. A. Rawashdeh, I-W. Shin, K. D. Donohue, and S. Love. Printer banding estimationusing the generalized spectrum. In L. C. Cui and Y. Miyake, editors, Image Qualityand System Performance III, volume 6059 of Proceedings of Electronic Imaging, page60590F, San Jose, CA, Jan. 2006. SPIE.

[373] J. Redi, H. Liu, H. Alers, R. Zunino, and I. E. Heynderickx. Comparing subjective im-age quality measurement methods for the creation of public databases. In S. P. Farnandand F. Gaykema, editors, Image Quality and System Performance VII, volume 7529 ofSPIE, pages 752903–1 –752903–11, San Jose, CA, Jan 2010.

[374] A. Rizzi, T. Algeri, G. Medeghini, and D. Marini. A proposal for contrast measure indigital images. In CGIV 2004 – Second European Conference on Color in Graphics,Imaging and Vision, pages 187–192, Aachen, Germany, April 2004. IS&T.

307

Page 324: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[375] A. Rizzi, G. Simone, and R. Cordone. A modified algorithm for perceived contrast indigital images. In CGIV 2008 - Fourth European Conference on Color in Graphics,Imaging and Vision, pages 249–252, Terrassa, Spain, Jun 2008. IS&T.

[376] L. I. Rudin, S. Osher, and E. Fatemi. Nonlinear total variation based noise removalalgorithms. Physica D Nonlinear Phenomena, 60:259–268, Nov 1992.

[377] H. Rushmeier, G. Ward, C. Piatko, P. Sanders, and B. Rust. Comparing real and syn-thetic images: Some ideas about metrics. In Eurographics Workshop on RenderingTechniques, pages 82–91, Dublin, Ireland, Jun 1995. Springer-Verlag, London, UK.

[378] R. Russell and P. Sinha. Perceptually-based comparison of image similarity metrics.Technical report, Massachusetts institute of technology, Cambridge, MA, July 2001.AI Memo 2001-014/CBCL Memo 201.

[379] A. Sadovnikov, L. Lensu, and H. Kalviainen. On estimation of perceived mottlingprior to printing. In S. P. Farnand and F. Gaykema, editors, Image Quality and SystemPerformance V, Proceedings of SPIE, pages 68080B–68080B–12, San Jose, CA, Jan2008.

[380] R. J. Safranek and J. D. Johnston. A perceptually tuned sub-band image coder withimage dependent quantization and post-quantization data compression. In InternationalConference on Acoustics, Speech, and Signal Processing, volume 3, pages 1945–1948,Glasgow, UK, May 1989. IEEE.

[381] A. Samet, M. A. B. Ayed, N. Masmoudi, and L. Khriji. New perceptual image qualityassessment metric. Asian Journal of Information Technology, 4(11):996–1000, 2005.

[382] C. Sano, T. Song, and M. R. Luo. Colour differences for complex images. In ColorImaging Conference: Color Science and Engineering Systems, Technologies, Applica-tions, pages 121–126, Scottsdale, Arizona, Nov 2003. IS&T. ISBN / ISSN: 0-89208-248-8.

[383] T. Sato, N. Aoki, and H. Kobayashi. Improvement of digital print in preference byadding noise. In International Conference on Digital Printing Technologies, pages122–123, Baltimore, MD, Sep 2005. IS&T.

[384] J. F. Sawyer. Effect of graininess and sharpness on perceived print quality. In Pho-tographic Image Quality Symposium, pages 222–231, Oxford, UK, Sep 1980. RoyalPhotographic Society.

[385] T. Scheermesser and O. Bryngdahl. Texture metric of halftone images. Journal of theOptical Society of America A, 13:18–24, 1996.

[386] T. Scheermesser and O. Bryngdahl. Spatially dependent texture analysis and control indigital halftoning. Journal of the Optical Society of America A, 14:827–835, 1997.

[387] T. Schulz. Globalstandard system brunner and prozessstandard offsetdruck. Technicalreport, System Brunner Telegram, July 2010.

[388] T. Seim and A. Valberg. Towards a uniform color space: A better formula to describethe munsell and OSA color scales. Color research & application, 11(1):11–24, 1986.

308

Page 325: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[389] H. J. Seo and P. Milanfar. Nonparametric bottom-up saliency detection by self-resemblance. In Computer Vision and Pattern Recognition Workshops, pages 45–52,Miami, FL, Jun 2009. IEEE.

[390] H. J. Seo and P. Milanfar. Static and space-time visual saliency detection by self-resemblance. Journal of Vision, 9(12):1–27, 2009. ISSN 1534-7362. article 15.

[391] K. Seshadrinathan and A. C. Bovik. The Encyclopedia of Multimedia, chapter Imageand video quality assessment, pages 8–17. Springer-Verlag, 2009.

[392] D. Shaked and I. Tastl. Sharpness measure: Towards automatic image enhancement.In International Conference on Image Processing, volume 1, pages 937–940, Genoa,Italy, Sep 2005. IEEE.

[393] A. Sharma. Measuring the quality of ICC profiles and color management software. TheSeybold Report, 4(20):10–16, Jan 2005.

[394] A. Sharma and P. D. Fleming. Evaluating the quality of commercial ICC color manage-ment software. In TAGA Technical Conference, North Carolina, Apr 2002. 17 pages.

[395] G. Sharma. Digital Color Imaging Handbook. CRC Press, Inc., Boca Raton, FL, USA,2002.

[396] H. R. Sheikh and A. C. Bovik. Image information and visual quality. IEEE Transactionson Image Processing, 15(2):430–444, 2006.

[397] H. R. Sheikh, A. C. Bovik, and G. de Veciana. An information fidelity criterion forimage quality assessment using natural scene statistics. IEEE Transactions on ImageProcessing, 14(12):2117–2128, Dec 2004.

[398] H. R. Sheikh, M. F. Sabir, and A. C. Bovik. A statistical evaluation of recent full ref-erence image quality assessment algorithms. IEEE Transactions on Image Processing,15(11):3440–3451, 2006.

[399] H. R. Sheikh, Z. Wang, L. K. Cormack, and A. C. Bovik. LIVE image quality as-sessment database release 2. 18/07/11: http://live.ece.utexas.edu/research/quality, 2007.

[400] J. Shiau and Z. Fan. A set of easily implementable coefficients in error diffusion withreduced worm artfacts. In J. Bares, editor, Color imaging: device-independent color,color hardcopy, and graphic arts, volume 2658, pages 222–225. SPIE, 1996.

[401] A. Shnayderman, A. Gusev, and A. M. Eskicioglu. A multidimensional image qualitymeasure using singular value decomposition. In Y. Miyake and D. R. Rasmussen,editors, Image Quality and System Performance, volume 5294 of Proceedings of SPIE,pages 82–92, San Jose, CA, USA, Dec 2004.

[402] A. Shnayderman, A. Gusev, and A. M. Eskicioglu. An SVD-based grayscale imagequality measure for local and global assessment. IEEE Transactions On Image Pro-cessing, 15(2):422–429, 2006.

309

Page 326: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[403] T. Shoham, D. Gill, and S. Carmel. A novel perceptual image quality measure for blockbased image compression. In S. P. Farnand and F. Gaykema, editors, Image Qualityand System Performance VIII, volume 7867 of Proceedings of SPIE, page 786709, SanFrancisco, CA, Jan 2011.

[404] M. Shohara and K. Kotani. Measurement of color noise perception. In Image Pro-cessing (ICIP), 2010 17th IEEE International Conference on, pages 3225 –3228, HongKong, sep 2010. IEEE.

[405] F. Sigg. Test targets 6.0. In School of Print Media (CIAS)–Test Targets, chapter Testingfor Resolution and Contrast, pages 1–5. RIT, 2006.

[406] E. A. Silva, K. Panetta, and S. S. Agaian. Quantifying image similarity using measureof enhancement by entropy. In S. S. Agaian and S. A. Jassim, editors, Mobile Multi-media/Image Processing for Military and Security Applications, volume 6579, pages65790U.1–65790U.12, San Jose, CA, USA, Jan 2007. SPIE.

[407] D. A. Silverstein and J. E. Farrell. The relationship between image fidelity and imagequality. In International Conference on Image Processing, volume 1, pages 881 – 884,Lausanne , Switzerland, Sep 1996.

[408] D. A. Silverstein and J. E. Farrell. Quantifying perceptual image quality. In Image Pro-cessing, Image Quality, Image Capture, Systems Conference, pages 242–246, Portland,OE, May 1998. IS&T.

[409] D. A. Silverstein and S. A. Klein. DCT image fidelity metric and its application toa text-based scheme for image display. In J. P. Allebach and B. E. Rogowitz, editors,Human Vision, Visual Processing, and Digital Display IV, volume 1913 of Proceedingsof SPIE, pages 229–239, San Jose, CA, Sep 1993.

[410] G. Simone, V. Caracciolo, M. Pedersen, and F. A. Cheikh. Evaluation of a difference ofgaussians based image difference metric in relation to perceived compression artifacts.In Advances in Visual Computing - 6th International Symposium, Lecture Notes inComputer Science, pages 491–500, Las Vegas, NV, Nov 2010. Springer.

[411] G. Simone, C. Oleari, and I. Farup. An alternative color difference formula for comput-ing image difference. In G. Simone, A. Rizzi, and J. Y. Hardeberg, editors, Proceedingsfrom Gjøvik Color Imaging Symposium 2009, number 4 in Høgskolen i Gjøviks rapport-serie, pages 8–11, Gjøvik, Norway, Jun 2009. URL http://brage.bibsys.no/hig/bitstream/URN:NBN:no-bibsys_brage_9313/3/sammensatt_elektronisk.pdf.

[412] G. Simone, C. Oleari, and I. Farup. Performance of the euclidean color-difference for-mula in log-compressed OSA-UCS space applied to modified-image-difference met-rics. In 11th Congress of the International Colour Association (AIC), Sydney, Aus-tralia, Sep/Oct 2009.

[413] G. Simone, M. Pedersen, and J. Y. Hardeberg. Measuring perceptual contrast in un-controlled environments. In European Workshop on Visual Information Processing(EUVIP), pages 102–107, Paris, France, Jul 2010. IEEE.

[414] G. Simone, M. Pedersen, and J. Y. Hardeberg. Measuring perceptual contrast in digitalimages. Journal of Visual Communication and Image Representation, 2011. Accepted.

310

Page 327: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[415] G. Simone, M. Pedersen, J. Y. Hardeberg, and I. Farup. On the use of gaze informationand saliency maps for measuring perceptual contrast. In A-B Salberg, Hardeberg J.Y., and R. Jenssen, editors, 16th Scandinavian Conference on Image Analysis, volume5575 of Lecture Notes in Computer Science, pages 597–606, Oslo, Norway, Jun 2009.

[416] G. Simone, M. Pedersen, J. Y. Hardeberg, and A. Rizzi. A multi-level framework formeasuring perceptual image contrast. Scandinavian Journal of Optometry and VisualScience, 1(1):15, Oct 2008.

[417] G. Simone, M. Pedersen, J. Y. Hardeberg, and A. Rizzi. Measuring perceptual contrastin a multilevel framework. In B. E. Rogowitz and T. N. Pappas, editors, Human Visionand Electronic Imaging XIV, volume 7240 of Proceedings of SPIE, pages 72400Q–72400Q–9, San Jose, CA, Jan 2009.

[418] R.R. Sokal and F.J. Rohlf. Biometry: Principles and Practice of Statistics in BiologicalResearch. W.H. Freeman, San Francisco, CA, second edition, 1981.

[419] T. Song and M. R. Luo. Testing color-difference formulae on complex images using aCRT monitor. In Color Imaging Conference, pages 44–48, Scottsdale, AZ, Nov 2000.IS&T/SID.

[420] Sony. Sony sRGB standard images 1999. 01/07/11: http://www.colour.org/tc8-04/test_images/Sony/.

[421] M. Southworth and D. Southworth. Quality and productivity in the graphic arts.Graphic Arts Pub. Co., 1990. ISBN 0933600054.

[422] I. Sparow, Z. Baranczuk, T. Stamm, and P. Zolliker. Web-based psychometric evalu-ation of image quality. In S. P. Farnand and F. Gaykema, editors, Image Quality andSystem Performance VI, volume 7242 of Proceedings of SPIE, page 72420A, San Jose,CA, Jan 2009.

[423] K. E. Spaulding, D. W. Couwenhoven, and R.L. Miller. Improved error diffusion in-corporating a visual model. In B. E. Rogowitx and T. N. Pappas, editors, Human Visionand Electronic Imaging III, volume 3299 of Proceedings of SPIE, pages 452–460, SanJose, CA, jul 1988.

[424] J. C. Stanek and P. D. Burns. Scanning parameter selection for inkjet print analysis. InImage Processing, Image Quality, Image Capture, Systems Conference (PICS), pages135–139, Rochester, NY, May 2003.

[425] M. C. Stone. Course 20 notes: Color in information display principles, perception, andmodels. International Conference on Computer Graphics and Interactive Techniquesarchive, 2004.

[426] M. C. Stone, W. B. Cowan, and J. C. Beatty. Color gamut mapping and the printing ofdigital color images. ACM Trans. Graph., 7:249–292, October 1988. ISSN 0730-0301.

[427] P. Stucki. Mecca - a multiple error correcting computation algorithm for bi-level imagehard copy reproduction. Technical report, IBM research laboratory, Zurich, Switzer-land, 1981.

311

Page 328: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[428] J. Sullivan, L. Ray, and R. Miller. Design of minimum visual modulation halftonepatterns. IEEE Trans. Systems, Man, and Cybernetics, 21:33–38, 1991.

[429] P-L. Sun. The Influence of Image Characteristics on Colour Gamut Mapping. PhDthesis, University of Derby, Derby, UK, 2002.

[430] P-L. Sun and Z-W. Zheng. Selecting appropriate gamut mapping algorithms based ona combination of image statistics. In R. Eschbach and G. G. Marcu, editors, ColorImaging X: Processing, Hardcopy, and Applications, volume 5667 of Proceedings ofSPIE, pages 211–219, San Jose, CA, Jan 2005.

[431] Y. Tadmor and D. J. Tolhurst. Calculating the contrasts that retinal ganglion cells andlgn neurones encounter in natural scenes. Vision Research, 40:3145–3157, 2000.

[432] TagCrowd. 07/04/10: http://tagcrowd.com/.

[433] C. C. Taylor, J. P. Allebach, and Z. Pizlo. The image fidelity assessor. In ImageProcessing, Image Quality, Image Capture, Systems Conference (PICS), pages 237–241, Portland, Oregon, USA, May 1998. IS&T.

[434] M. M. Taylor and C. D. Creelman. PEST: Efficient estimates on probability functions.The Journal of the Acoustical Society of America, 41(4A):782–787, 1967.

[435] P. C. Teo and D. J. Heeger. Perceptual image distortion. In International ConferenceImage Processing, volume 2, pages 982 – 986, Austin, TX, Nov 1994. IEEE.

[436] K-H. Thung and P. Raveendran. A survey of image quality measures. In InternationalConference for Technical Postgraduates (TECHPOS), pages 1 – 4, Kuala Lumpur,Malaysia, Dec 2009. IEEE.

[437] L. L. Thurstone. A law of comparative judgment. Psychological Review, 34:237, 1927.

[438] Tobias Associates. Reflection densitometry. 09/02/11: http://www.densitometers.net/pdf/reflection_densitometry.pdf, 2002.

[439] A. Toet and M. P. Lucassen. A new universal colour image fidelity metric. Displays,24:197–204, 2003.

[440] Y. Tong, H. Konik, F. A. Cheikh, and A. Tremeau. Full reference image quality assess-ment based on saliency map analysis. Journal of Imaging Science and Technology, 54(3):030503, 2010.

[441] K. Topfer, B. W. Keelan, S. O’Dell, and R. Cookingham. Preference in image qualitymodelling. In Image Processing, Image Quality, Image Capture, Systems Conference(PICS), pages 60–64, Apr, Portland, OR 2002. IS&T.

[442] W. S. Torgerson. A Law of Categorical Judgment, chapter Consumer Behaviour, pages92–93. New York University Press, 1954.

[443] A. Tremeau, R. Nicolas, and E. Dinet. A measure of colour contrast correlated withhuman perception. In CGIV 2008 Fourth European Conference on Color in Graphics,Imaging and Vision, pages 120–125, Terrassa, Spain, Jun 2008. IS&T.

312

Page 329: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[444] H. Trontelj, J. E. Farrell, J. Wiseman, and J. Shur. Optimal halftoning algorithm de-pends on printing resolution. SID Digest, pages 749–752, 1992.

[445] M-K. Tse and A. H. Klein. Automated print quality analysis in inkjet printing: Casestudy using commercially available media. In Recent Progress in Ink Jet TechnologiesII, pages 515–519. IS&T, 1999.

[446] R. Ulichney. Digital halftoning. MIT Press, 1987.

[447] I. Van Der Linde, U. Rajashekar, A. C. Bovik, and L. K. Cormack. DOVES: a databaseof visual eye movements. Spatial vision, 22(2):161–177, 2009. ISSN 0169-1015.

[448] L. Velho and J. Gomes. Color halftoning with stochastic dithering and adaptive clus-tering. In First European Conference on Color in Graphics, Image and Vision (CGIV),pages 627–632, Poitiers, France, Apr 2002. IS&T.

[449] O. Veryovka, A. Fournier, and J. W. Buchanan. Multiscale edge analysis of halftonedimages. In B. E. Rogowitz and T. N. Pappas, editors, Human Vision and ElectronicImaging III, volume 3299 of SPIE Proceedings, pages 461–472, San Jose, CA, USA,Jan 1998.

[450] Video Quality Experts Group. Final report from the video quality experts group on thevalidation of objective models of video quality assessment, phase II. Technical report,August 2003.

[451] D. Van De Ville, K. Denecker, W. Philips, and I. Lemahieu. A quality measure forhalftones obtained by linear and nonlinear resampling. In J.-P. Veen, editor, Proceed-ings of the Workshop on Circuits, Systems and Signal Processing, pages 549–556, Veld-hoven, the Netherlands, Dec 2000. IEEE.

[452] X. Wan, D. Xie, and J. Xu. Quality evaluation of the halftone by halftoning algorithm-based methods and adaptive method. In L. C. Cui and Y. Miyake, editors, Image Qualityand System Performance IV, volume 6494 of SPIE Proceedings, page 64940U, SanJose, CA, USA, Jan 2007. SPIE.

[453] L. Wang, J. Giesen, K. T. McDonnell, P. Zolliker, and K. Mueller. Color design forillustrative visualization. IEEE Transactions on Visualization and Computer Graphics,14:1739–1754, 2008.

[454] M. Wang and K. J. Parker. Prediction of the texture visibility of color halftone patterns.Journal of Electronic Imaging, 11(2):195 – 205, April 2002.

[455] Z. Wang and A. C. Bovik. A universal image quality index. IEEE Signal ProcessingLetters, 9:81–84, 2002.

[456] Z. Wang and A. C. Bovik. Modern Image Quality Assessment. Morgan & ClaypoolPublishers, 2006.

[457] Z. Wang, A. C. Bovik, and L. Lu. Wavelet-based foveated image quality measurementfor region of interest image coding. In International Conference on Image Processing,pages 89–92, Thessaloniki , Greece, Oct 2001. IEEE.

313

Page 330: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[458] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli. Image quality assessment:from error visibility to structural similarity. IEEE Transactions on Image Processing,13(4):600–612, 2004.

[459] Z. Wang and J. Y. Hardeberg. An adaptive bilateral filter for predicting color imagedifference. In Color Imaging Conference, pages 27–31, Albuquerque, NM, USA, Nov2009. IS&T/SID.

[460] Z. Wang and Q. Li. Information content weighting for perceptual image quality assess-ment. IEEE Transactions on Image Processing, 20(5):1185 – 1198, Nov 2010.

[461] Z. Wang, L. Lu, and A. C. Bovik. Video quality assessment based on structural dis-tortion measurement. Signal Processing: Image Communication, 19(2):121132, Feb2004.

[462] Z. Wang and M. R. Luo. Experimental filters for estimating image differences. InCGIV 2008 - Fourth European Conference on Color in Graphics, Imaging and Vision,pages 112–115, Terrassa, Barcelona, Spain, Jun 2008. IS&T.

[463] Z. Wang and X. Shang. Spatial pooling strategies for perceptual image quality assess-ment. In International Conference on Image Processing, pages 2945–2948, Atlanta,GA, Oct 2006. IEEE.

[464] Z. Wang and E. P. Simoncelli. Reduced-reference image quality assessment using awavelet-domain natural image statistic model. In B. Rogowitz, T. N. Pappas, and S. J.Daly, editors, Human Vision and Electronic Imaging X, volume 5666 of Proc. SPIE,pages 149–159, San Jose, CA, USA, Jan 2005.

[465] Z. Wang and E. P. Simoncelli. Reduced-reference image quality assessment using awavelet-domain natural image statistic model. In B. Rogowitz, T. N. Pappas, and S. J.Daly, editors, Human Vision and Electronic Imaging X, volume 5666 of Proceedingsof SPIE, pages 149–159, San Jose, CA, Jan 2005.

[466] Z. Wang and E. P. Simoncelli. Translation insensitive image similarity in complexwavelet domain. In International conference on acoustics, speech and signal process-ing, volume 2, pages 573–576, Philadelphia, PA, Mar 2005. IEEE.

[467] Z. Wang, E. P. Simoncelli, and A. C. Bovik. Multi-scale structural similarity for im-age quality assessment. In Asilomar Conference on Signals, Systems and Computers,volume 2, pages 1398–1402, Pacific Grove, CA, Nov 2003. IEEE.

[468] A. B. Watson. DCT quantization matrices visually optimized for individual images.In J. P. Allebach and B. E. Rogowitz, editors, Human Vision, Visual Processing, andDigital Display IV, volume 1913 of SPIE Proceedings, pages 202–216, San Jose, CA,USA, February 1993.

[469] A. B. Watson and A. Fitzhugh. The method of constant stimuli is inefficient. PerceptionAnd Psychophysics, 47(1):87–91, 1988.

[470] A. B. Watson and D. G. Pelli. QUEST: A bayesian adaptive psychometric method.Perception & Psychophysics, 33(2):113–120, 1983.

314

Page 331: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[471] M. A. Webster, Y. Mizokami, L. A. Svec, and S. L. Elliot. Neural adjustments tochromatic blur. Spatial Vision, 19(2-4):111–132, 2006.

[472] S. J. P. Westen, R. L. Lagendijk, and J. Biemond. Perceptual image quality based ona multiple channel HVS model. In International Conference on Acoustics, Speech andSignal Processing, volume 4, pages 2351–2354, Detroit, MI, May 1995. IEEE.

[473] P. Whittle. Lightness, Brightness, and Transparency, chapter The psychophysics ofcontrast brightness, pages 35–110. Lawrence Erlbaum Associates, Hillsdale, NJ, 1994.

[474] D. L. Wilson, A. J. Baddeley, and R. A. Owens. A new metric for grey-scale imagecomparison. International Journal of Computer Vision, 24:5–17, 1997.

[475] M. Wolski, J. P. Allebach, and C. A. Bouman. Gamut mapping squeezing the most outyour color system. In Color Imaging Conference, pages 89–92, Scottsdale, AZ, Nov1994. IS&T.

[476] P. W. Wong. A mixture distortion criterion for halftones. Technical report, HP Labs,Mar 1997. 18/07/11: http://www.hpl.hp.com/techreports/97/HPL-97-49.html.

[477] W. Wu, Z. Pizlo, and J. P. Allebach. Color image fidelity assessor. In Image Processing,Image Quality, Image Capture, Systems Conference (PICS), pages 148–152, Monreal,Quebec, Canada, Apr 2001. IS&T.

[478] D. Wurgler. Globalstandard system brunner - standardization instead of chaos. Tech-nical report, System Brunner, 2004. 28/06/11: http://www.systembrunner.ch/homepage/files/english/eurostandard_offset.pdf.

[479] G. Wyszecki and W.S. Styles. Color Science, Concepts and Methods, QuantitativeData and Formulae. Wiley Interscience, Derby, UK, 2nd edition, Aug 2000.

[480] R. Xu, P.D. Flemming, A. Pekarovicova, and V. Bliznyuk. The effect of ink jet paperroughness on print gloss. Journal of Imaging Science and Technology, 49(6):660–666,Nov/Dec 2005.

[481] R. Xu, S. N. Pattanaik, and C. E. Hughes. High-dynamic-range still-image encoding inJPEG2000. IEEE Computer Graphics and Applications, 25(6):57–64, 2005.

[482] N. Yamsang and S. Udomhunsakul. Image quality scale ((IQS) for compressed imagesquality measurement. In International MultiConference of Engineers and ComputerScientists, pages 789– 794, Hong Kong, Mar 2009.

[483] X. Yanfang, W. Yu, and Z. Ming. Color reproduction quality metric on printing imagesbased on the s-cielab model. In International Conference on Computer Science andSoftware Engineering, volume 6, pages 294–297, Wuhan, Hubei, Dec 2008. IEEE.

[484] S. Yao, W. Lin, Z. K Lu, E. P. Ong, M. H. Locke, and S. Q. Wu. Image quality measureusing curvature similarity. In International Conference on Image Processing, volume 3,pages 437–440, San Antonio, TX, Sep 2007. IEEE.

[485] S. Yao, W. Lin, S. Rahardja, X. Lin, E. P. Ong, Z. K. Lu, and X. K. Yang. Perceivedvisual quality metric based on error spread and contrast. In International Symposiumon Circuits and Systems, volume 4, pages 3793 – 3796, Kobe, Japan, May 2005. IEEE.

315

Page 332: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[486] H. Yee. A perceptual metric for production testing. Journal of Graphics, GPU, andGame Tools, 9(4):33–40, 2004.

[487] E. M. Yeh, A. C. Kokaram, and N. G. Kingsbury. A perceptual distortion measure foredge-like artifacts in image sequences. In T. N. Pappas B. E. Rogowitz, editor, HumanVision and Electronic Imaging III, volume 3299 of SPIE proceedings, pages 160–172,San Jose, CA, Jan 1998.

[488] S.N. Yendrikhovskij. Color reproduction and the naturalness constraint. PhD thesis,Technische Universiteit Eindhoven, 1998.

[489] Q. Yu and K. J. Parker. Quality issues in blue noise halftoning. In G. B. Beretta andR. Eschbach, editors, Color Imaging: Device-Independent Color, Color Hardcopy, andGraphic Arts III, volume 3300 of SPIE proceedings, pages 376–385, San Jose, CA, Jan1998.

[490] Q. Yu, K. J. Parker, R. Buckley, and V. Klassen. A new metric for color halftonevisibility. In Image Processing, Image Quality, Image Capture, Systems Conference(PICS), pages 226–230, Portland, OR, USA, May 1998. IS&T.

[491] E. K. Zeise, S. H. Kim, and F. Sigg. Development of ISO/IEC 29112: test chartsand methods for measuring monochrome printer resolution. In S. P. Farnand andF. Gaykema, editors, Image Quality and System Performance VII, volume 7529 of Pro-ceedings of SPIE, pages 75290F–1–75290F–9, San Jose, Jan 2010.

[492] E. K. Zeise, D. Williams, P. D. Burns, and W. C. Kress. Scanners for analytic printmeasurement the devil in the details. In L. C. Cui and Y. Miyake, editors, Image Qual-ity and System Performance IV, volume 6494 of Proceedings of SPIE, page 64940L,San Jose, CA, Jan 2007.

[493] B. Zhang, J. P. Allebach, and Z. Pizlo. An investigation of perceived sharpness andsharpness metrics. In R. Rasmussen and Y. Miyake, editors, Image Quality and SystemPerformance II, volume 5668 of Proceedings of SPIE, San Jose, CA, USA, Jan 2005.

[494] F. Zhang, S. Li, L. Ma, and K. N. Ngan. Limitation and challenges of image qualitymeasurement. In P. Frossard, H. Li, F. Wu, B. Girod, S. Li, and G. Wei, editors, VisualCommunications and Image Processing conference, volume 7744 of Proceedings ofSPIE, pages 774402–774402–8, Huang Shan, China, Jul 2010.

[495] L. Zhang, L. Zhang, and X. Mou. RFSIM: A feature based image quality assessmentmetric using riesz transforms. In Internatonal Conference on Image Processing, pages321–324, Hong Kong, Sep 2010. IEEE.

[496] L. Zhang, L. Zhang, X. Mou, and D. Zhang. FSIM: A feature similarity index for imagequality assessment. IEEE Transactions on Image Processing, 2011.

[497] X. Zhang, J. E. Farrell, and B. A. Wandell. Application of a spatial extension toCIELAB. In Very high resolution and quality imaging II, volume 3025 of SPIE pro-ceedings, pages 154–157, San Jose, CA, USA, Feb 1997.

[498] X. Zhang, D. A. Silverstein, J. E. Farrell, and B. A. Wandell. Color image qualitymetric S-CIELAB and its application on halftone texture visibility. In COMPCON97Digest of Papers, pages 44–48, Washington, DC, USA, 1997. IEEE.

316

Page 333: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

[499] X. Zhang and B. A. Wandell. A spatial extension of CIELAB for digital color imagereproduction. In Soc. Inform. Display 96 Digest, pages 731–734, San Diego,CA, 1996.

[500] X. Zhang and B. A. Wandell. Color image fidelity metrics evaluated using image dis-tortion maps. Signal Processing - Special issue on image and video quality metrics, 70:201 – 214, 1998.

[501] B. Zhao and C. Deng. Image quality evaluation method based on human visual system.Chinese Journal of Electronics, 19(1):129–132, Jan. 2010.

[502] P. Zolliker and K. Simon. Adding local contrast to global gamut mapping algorithms. In3th European Conference on Colour in Graphics, Imaging, and Vision (CGIV), pages257–261, Leeds, UK, Jun 2006. IS&T.

[503] S. Zuffi, C. Brambilla, R. Eschbach, and A. Rizzi. Controlled and uncontrolled viewingconditions in the evaluation of prints. In R. Eschbach, G. G. Marcu, and S. Tominaga,editors, Color Imaging XIII: Processing, Hardcopy, and Applications, volume 6807,page 680714, San Jose, CA, Jan 2008. SPIE.

[504] S. Zuffi, P. Scala, C. Brambilla, and G. Beretta. Web-based versus controlled environ-ment psychophysics experiments. In L. C. Cui and Y. Miyake, editors, Image Qualityand System Performance IV, volume 6494 of SPIE proceedings, page 649407, San Jose,CA, Jan 2007.

317

Page 334: Doctoral Dissertation by Marius Pedersen

BIBLIOGRAPHY

318

Page 335: Doctoral Dissertation by Marius Pedersen

PART IV

APPENDICES

319

Page 336: Doctoral Dissertation by Marius Pedersen
Page 337: Doctoral Dissertation by Marius Pedersen

A OVERVIEW OF IMAGE QUALITY

METRICS

In Table A.1 we show an overview of image quality metrics. The metrics are sorted in chrono-logical order, then the name of the metric if given any by the authors, then the author namesin column three. For the type column the metrics are grouped by type. ID = image difference,IQ = image quality, IF = image fidelity, CD = color difference, HT = halftoning, DP = differ-ence predictor, VQ = video quality, IS = image similarity. For the column HVS, the metricmust have a HVS model or CSF filter that simulates the HVS. The metrics like SSIM, whoindirectly simulate the HVS is set to ”no”. MS indicates whether the metrics are multiscale.S/NS indicates whether the metric is spatial (FFS, FVS, VFS, VVS) or non-spatial (NS). Wehave divided the spatial metrics into 3 groups, FFS = fixed size of filter and fixed calculation,FVS = fixed size of filter and variable calculation, VFS = variable size of filter and fixed cal-culation, VVS = variable size of filter and variable calculation. Fixed size indicates that thefilter, block or similar is fixed for the whole image, variable size indicates that the filter orblock changes according to the image content. Fixed computation indicates the same calcula-tion within the filter or block, variable calculation indicates that the calculation is dependenton the image content. C/G indicates whether the metric are for color or grayscale images.The test column indicates what kind for evaluation that has been performed. This is eitherobjective or subjective, for the metrics where the authors has done the subjective test this ismarked with (A). Scenes indicates the number of scenes used in the original work where themetric was proposed, the same for modification and observers. For scenes the first numberof the number of scenes, while the second number in ( ) indicates the total number of images(originals × number of modifications). For observers the total number of observers is stated,and inside the ( ) number of experts are stated if this information is given by the author. -indicates that this information is not available or not stated by the authors. Figure A.1 showsrelations between selected metrics.

321

Page 338: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1:

Ove

rvie

wof

full

-ref

eren

ceIQ

met

rics

sort

edac

cord

ing

toye

ar.

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

1976

ΔE∗ ab

CIE

CD

No

No

NS

Col

or-

--

-

1984

CM

CC

lark

eet

al.[

87]

CD

No

No

NS

Col

or-

--

-

1984

Nas

anen

[313

]H

TY

esN

oFF

SG

ray

sub.

4(1

28)

Dot

profi

les

31

obse

rver

sin

the

mai

nex

peri

-

men

t,2

inco

ntro

lexp

erim

ent.

1986

SVF

Seim

and

Val

berg

[388

]C

DN

oN

oN

SC

olor

obj.

--

-

1987

BFD

Luo

and

Rig

g[2

76]

CD

No

No

NS

Col

or-

--

-

1989

Safr

anek

and

John

ston

[380

]

IQN

oN

oFF

SG

ray

obj.

30-

1990

SQR

IB

arte

n[2

6]IQ

No

No

FFS

Gra

ysu

b.5

(35)

Res

olut

ion

20su

b.da

tafr

omW

este

rink

and

Rou

fs.

1991

LIP

MSE

Bra

ilean

etal

.[48

]IF

Yes

No

FFS

Gra

ysu

b.1

(1)

Gau

ssia

nbl

ur1

(A)

Met

ric

used

tore

stor

ea

blur

red

imag

e.

1992

Mits

aan

dV

arku

r[2

90]

HT

Yes

No

VFS

Gra

ysu

b.11

(44)

Hal

fton

ing

123

met

rics

test

ed,

sam

epr

oced

ure

butd

iffe

rent

CSF

s.

1993

VD

PD

aly

[98]

DP

Yes

No

FFS

Gra

y-

--

-

1993

Wat

son

[468

]IQ

No

No

FFS

Gra

yob

j.-

2(8

)-

1993

Silv

erst

ein

and

Kle

in

[409

]

IQN

oN

oFF

SG

ray

obj.

--

-

1993

Lin

[268

]H

TY

esN

oFF

SG

ray

obj.

1(5

)H

alft

onin

g-

CSF

filte

rfr

omM

anno

set

al.

and

Nill

etal

.te

sted

.

Con

tinue

don

Nex

tPag

e...

322

Page 339: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

1993

Cha

ddha

and

Men

g[6

7]IQ

No

No

FFS

Gra

ysu

b.6

(48)

Com

pres

sion

artif

acts

60A

lso

exte

nded

tovi

deo.

Onl

yth

e

resu

ltsfr

omon

esc

ene

pres

ente

d.

1994

Mits

aan

dA

lfor

d[2

89]

HT

Yes

No

VFS

Gra

ysu

b.11

(44)

Hal

fton

ing

1994

Kar

unas

eker

aan

d

Kin

gsbu

ry[2

28]

IQY

esN

oFF

SG

ray

sub.

1(8

)L

appe

dO

rtho

gona

lTra

ns-

form

8O

bser

ver

expe

rtis

eno

tsta

ted.

1994

Teo

and

Hee

ger

[435

]IQ

Yes

Yes

VFS

Col

ors/

o1

(2)

-(A

)

1995

ΔE

94C

IEC

DN

oN

oN

SC

olor

--

--

Test

edin

vari

ous

pape

rson

vari

-

ous

type

ofsc

enes

.

1995

PDM

Hee

ger

and

Teo

[172

]IF

Yes

Yes

VFS

Gra

ysu

b.1

(3)

JPE

G(A

)E

xten

sion

ofTe

oan

dH

eege

r

[435

]

1995

PSPN

RC

hou

and

Li[

77]

IQY

esN

oFF

SG

ray

obj.

--

-

1995

Rus

hmei

eret

al.[

377]

ISY

esN

oFF

SG

ray

obj.

Var

ious

Synt

hetic

imag

es-

Com

pari

ngre

alan

dsy

nthe

ticim

-

ages

1995

PEM

Wes

ten

etal

.[47

2]IQ

Yes

No

VFS

Gra

ysu

b.6(

105)

PCM

,D

PCM

,D

CT

and

SBC

codi

ngat

diff

eren

t

bitr

ates

7(5

)

1995

VD

ML

ubin

[272

]IQ

Yes

Yes

VFS

Gra

ysu

b/ob

j-

--

Var

ious

test

ing

1996

S-C

IEL

AB

Zha

ngan

dW

ande

ll

[499

]

ID,

HT

Yes

No

FFS

Col

orob

j.-

JPE

G-D

CT,

half

toni

ng

and

patte

rns

-

1996

PQS

Miy

ahar

aet

al.[

291]

IQY

esN

oFF

SG

ray

sub.

5(25

)gl

obal

and

loca

ldis

tort

ion

9(9

)

Con

tinue

don

Nex

tPag

e...

323

Page 340: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

1996

CM

PSN

RL

ambr

echt

and

Farr

ell

[251

]

IQY

esN

oV

FSC

olor

sub.

1(4

00)

JPE

G2

1996

Sche

erm

esse

ran

dB

ryn-

gdah

l[38

5]

HT

No

No

VV

SG

ray

obj.

2(8

)H

alft

onin

g-

Als

ote

sted

on2

test

patte

rn.

1996

PQS

Miy

ahar

aet

al.[

291]

IQY

esN

oFF

SG

ray

sub.

5(2

5)-

9(9

)

1997

Sarn

off

JND

Vi-

sion

Mod

el

Lub

in[2

73]

IQY

esY

esFF

SC

olor

sub.

5(1

5)M

PEG

-2w

ithdi

ffer

ent

bit-

rate

s

20A

lso

test

edon

JPE

Gda

ta.

1997

Neu

man

net

al.[

315]

IDN

oN

oV

FSC

olor

obj.

--

--

1997

Won

g[4

76]

HT

No

No

FFS

Gra

y-

--

-

1997

Nije

nhui

san

dB

lom

-

mae

rt[3

18]

IQN

oN

oN

SG

ray

sub.

2(2

5)In

terp

olat

ion

6

1997

Lai

etal

.[24

9]IF

Yes

Yes

VFS

Col

orsu

b.1(

1)JP

EG

2000

(A)

1997

Sche

erm

esse

ran

dB

ryn-

gdah

l[38

6]

HT

No

No

VV

SG

ray

obj.

2(-

)H

alft

onin

g-

1997

Δg

Wils

onet

al.[

474]

ISN

oN

oV

FSG

ray

sub.

1(5

)JP

EG

,di

ffer

ent

dist

ortio

n

type

s

(A)

1998

IFA

Tayl

oret

al.[

433]

IFY

esY

esFF

SG

ray

obj.

--

--

1998

CV

DM

Jin

etal

.[21

9]ID

Yes

No

FFS

Col

orob

j.-

--

1998

Lai

etal

.[24

8]IF

Yes

No

FFS

Col

orsu

b.1(

1)JP

EG

2000

(A)

1998

Yu

etal

.[49

0]H

TY

esN

oFF

SC

olor

sub.

6(3

6)H

alft

onin

g8

(8)

1998

KL

KY

uan

dPa

rker

[489

]H

TN

oN

oFF

SC

olor

sub.

5(2

0)H

alft

onin

g10

Con

tinue

don

Nex

tPag

e...

324

Page 341: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

1998

Ver

yovk

aet

al.[

449]

HT

Yes

Yes

VFS

Gra

yob

j.1/

1(3

/3)

Hal

fton

ing

--

1998

EY

ehet

al.[

487]

VQ

Yes

No

FFS

Gra

ysu

b.1

(9)

Blo

ckar

tifac

ts8

Sequ

ence

of64

fram

es.

1998

CC

ET

Tvi

sual

met

ric

Peff

erko

rnan

dB

lin

[361

]

IQY

esN

oFF

SC

olor

sub.

6(3

0)M

PEG

-216

1998

Bol

inan

dM

eyer

[37]

DP

Yes

No

FFS

Col

orsu

b.-

-(A

)

1998

Fran

ti[1

46]

IQN

oN

oFF

SG

ray

sub.

3(4

2)C

ompr

essi

on15

-39

Poss

iblil

tyfo

rco

lor

imag

es

1999

WV

DP

Bra

dley

[46]

DP

Yes

No

FFS

Gra

yob

j.1

(3)

Noi

se-

1999

PDM

Ava

dhan

aman

dA

lgaz

i

[13]

IFY

esN

oFF

SC

olor

sub.

5/5(

50/7

5)C

ompr

essi

on5(

2)/5

(2)

1999

LSM

BPa

ppas

and

Neu

hoff

[336

]

HT

Yes

No

FFS

Gra

y-

--

-U

sed

for

half

toni

ngop

timiz

atio

n

1999

QM

aN

ilsso

n[3

19]

HT

Yes

No

VV

SG

ray

obj.

1(3)

Hal

fton

ing

-

1999

PDN

akau

chie

tal.

[312

]ID

Yes

No

FFS

Col

orsu

b.8

(48)

Gam

utm

appi

ng10

Use

dto

optim

ize

gam

utm

appi

ng

2000

DM

and

NQ

MD

amer

a-V

enka

taet

al.

[99]

IQN

oN

oFV

SG

ray

sub.

-V

ario

us(A

)

2000

Farr

ugia

and

Pero

che

[130

]

IDY

esN

oFF

SC

olor

obj.

4(8)

(A)

2001

ΔE

00C

IE[2

001]

CD

No

No

NS

Col

or-

--

-

2001

CIF

AW

uet

al.[

477]

IDY

esY

esFF

SC

olor

sub.

1(1)

Hue

(A)

2001

SCIE

LA

BJ

John

son

and

Fair

child

[223

]

IQ,

HT

Yes

No

FFS

Col

orsu

b.1(

72)

Shar

pnes

s-

Ext

ensi

onof

S-C

IEL

AB

Zha

ng

and

Wan

dell

[499

].

Con

tinue

don

Nex

tPag

e...

325

Page 342: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2001

Imai

etal

.[18

3]C

DN

oN

oN

SC

olor

sub.

6(12

)-

-

2001

SNR

WIo

rdac

hean

dB

eghd

adi

[192

]

ISN

oN

oFF

SG

ray

sub.

1(3)

Salt-

and-

pepp

erno

ise,

blur

ring

,JPE

G

5(0

)

2002

iCA

MFa

irch

ildan

dJo

hnso

n

[124

]

IDY

esN

oFF

SC

olor

--

--

2002

Hon

gan

dL

uo[1

77]

IDN

oN

oN

SC

olor

obj.

2L

ocal

colo

rch

ange

-

2002

UIQ

Wan

gan

dB

ovik

[455

]IQ

No

No

FFS

Gra

ysu

b.1

(8)

Dif

fere

ntdi

stor

tion

type

s22

2002

Feng

etal

.[13

8]IQ

Yes

No

FFS

Col

orsu

b.2

(14)

Hal

fton

ing

Ext

ensi

onof

CV

DM

2002

ΔI c

mM

orov

ican

dSu

n[3

04]

IDN

oN

oFF

SC

olor

sub.

7(3

2)G

amut

map

ping

-

2002

WM

SEA

yed

etal

.[16

]IQ

No

No

FFS

Gra

yob

j.2

(14)

Noi

se-

-

2002

HV

SR

EA

LA

vcib

aset

al.[

14]

IQN

oN

oFF

SG

ray

obj.

30(4

20)

JPE

G,

SPH

IT,

nois

e,an

d

blur

17M

etri

cbu

ildon

wor

kby

Fres

e

etal

.[14

8].

2003

Qco

lor

Toet

and

Luc

asse

n[4

39]

IFN

oN

oFF

SC

olor

sub.

2(2

1)Q

uant

izat

ion

4-16

2003

FDP

Ferw

erda

and

Pella

cini

[142

]

DP

Yes

No

NS

Gra

ysu

b.8

(24)

Com

pute

rge

nera

ted

im-

ages

18

2003

MSS

IMW

ang

etal

.[46

7]IQ

No

Yes

FFS

Gra

ysu

b.29

(344

)JP

EG

and

JPE

G2k

-Sc

enes

from

LIV

E

2003

M−

SVD

Shna

yder

man

etal

.[40

1]IQ

No

No

FFS

Gra

ysu

b.5

(30)

JPE

G,

JPE

G2k

,G

.Noi

se,

G.B

lur

,Sh

arpe

ning

,D

C

shif

t

10(5

/5)

Col

orex

tens

ion

poss

ible

2003

SNR

WAV

Beg

hdad

ian

dPe

sque

t-

Pope

scu

[30]

IQN

oN

oFF

SG

ray

sub.

1(3

)G

auss

ion

nois

e,JP

EG

and

grid

patte

rn

>25

Con

tinue

don

Nex

tPag

e...

326

Page 343: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2003

DQ

Man

dC

QM

deFr

eita

sZ

ampo

loan

d

Sear

a[1

00]

IQN

oN

oFF

SG

ray

sub.

4(4

5)Fr

eque

ncy

dist

ortio

n7

2003

Qua

lity

Ass

esse

rC

arne

cet

al.[

65]

IQY

esN

oV

FSC

olor

sub.

-(9

0)JP

EG

and

JPE

G2k

-A

lso

test

edon

LIV

E

2004

B-C

QM

deFr

eita

sZ

ampo

loan

d

Sear

a[1

01]

IQN

oN

oN

SG

ray

sub.

1(8

1)Fr

eque

ncy

dist

ortio

nan

d

nois

ein

ject

ion

2004

SSIM

Wan

get

al.[

458]

IQN

oN

oFF

SG

ray

sub.

29(3

44)

JPE

Gan

dJP

EG

2k-

Scen

esfr

omL

IVE

2004

Ivko

vic

and

Sank

ar

[215

]

IQY

esN

oFF

SG

ray

obj.

5(2

0)C

ontr

ast

stre

chin

g,w

hite

nois

e,bl

ur,J

PEG

2k

-

2004

NN

ET

Bou

zerd

oum

etal

.[44

]IQ

No

No

FFS

Gra

ysu

b.-

JPE

Gan

dJP

EG

2k-

Scen

esfr

omL

IVE

2004

I∗M

cCor

mic

k-G

oodh

art

etal

.[28

5]

IQN

oN

oFF

SC

olor

sub.

-2

prin

tsys

tem

s-

2004

HD

R-V

DP

Man

tiuk

etal

.[28

2]D

PN

oY

esV

FSG

rey

sub.

-Q

uant

izat

ion,

nois

e(A

)

2004

pdif

fY

ee[4

86]

IDY

esY

esFF

SC

olor

--

Film

prod

uctio

n-

2004

IFC

Shei

khet

al.[

397]

IFN

oN

oFF

SG

ray

sub.

29(3

44)

JPE

G,

JPE

G20

0,N

oise

,

Blu

r

20-

25Sc

enes

from

LIV

E

2005

CB

MG

aoet

al.[

150]

IQN

oN

oFF

SG

ray

sub.

29(3

44)

JPE

Gan

dJP

EG

2k-

Scen

esfr

omL

IVE

2005

CW

SSIM

Wan

gan

dSi

mon

celli

[466

]

IS,

IQ

No

No

FFS

Gra

yob

j.1

(12)

Dis

tort

ion

asJP

EG

,no

ise

etc.

-

2005

PQ

SIG

uarn

erie

tal.

[165

]IQ

No

No

-C

olor

sub.

-5

Inte

rpol

atio

nal

gori

thm

s-

2005

M-D

WT

Gay

leet

al.[

153]

IQN

oN

oN

SC

olor

sub.

5(3

0)JP

EG

,JP

EG

2k,

blur

,

nois

e,sh

arp

and

DC

-shi

ft

14St

ated

asa

colo

rm

etri

c,bu

ton

ly

oper

ates

onlu

min

ance

Con

tinue

don

Nex

tPag

e...

327

Page 344: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2005

VQ

ME

SCY

aoet

al.[

485]

IQN

oN

oFF

SG

ray

sub.

-(3

44)

JPE

Gan

dJP

EG

2k-

Scen

esfr

omL

IVE

2005

MH

IA

net

al.[

11]

IQN

oN

oFF

SG

ray

sub.

1(4

)JP

EG

,JP

EG

2k,

gaus

sian

nois

ean

dsp

eckl

edno

ise

(A)

2005

Kim

mel

etal

.[23

6]IS

No

Yes

FFS

Col

orsu

b.3

(3)

Gam

utm

appi

ng(A

)U

sed

for

gam

utm

appi

ngop

ti-

miz

atio

n

2005

Xu

etal

.[48

1]ID

No

No

NS

Col

orsu

b.-

Com

pres

sion

(A)

2005

Nw

MSE

Sam

etet

al.[

381]

IQY

esN

oFF

SG

ray

sub.

-JP

EG

2k,

JPE

G,

and

blur

-

ring

-Sc

enes

from

LIV

E

2006

VIF

Shei

khan

dB

ovik

[396

]IF

Yes

No

FFS

Gra

ysu

b.29

(344

)JP

EG

and

JPE

G2k

-Sc

enes

from

LIV

E

2006

WC

WSS

IMB

rook

san

dPa

ppas

[53]

VQ

Yes

Yes

FFS

Col

orsu

b.3

(5)

Vid

eoco

mpr

essi

onan

d

tran

smis

sion

dist

ortio

n

-V

ario

uste

stin

gof

the

met

ric.

2006

DT

WT-

SSIM

Lee

etal

.[26

3]IS

No

No

FFS

Gra

yob

j.10

(486

0)B

lurr

ing,

scal

ing

,rot

atio

n

and

shif

t

-Te

sted

onha

ndw

ritte

nda

taas

a

sim

ilari

tym

easu

re.

2006

ESS

IMC

hen

etal

.[71

]IQ

No

No

FFS

Gra

ysu

b.-

(489

)JP

EG

2k,

JPE

G,

and

blur

-

ring

-

2006

GSS

IMC

hen

etal

.[72

]IQ

No

No

FFS

Gra

ysu

b.-

(489

)JP

EG

2k,

JPE

G,

and

blur

-

ring

-

2006

UQ

I-H

VS

Egi

azar

ian

etal

.[11

3]IQ

Yes

No

FFS

Gra

ysu

b.2

(44)

Noi

se,

blur

,JP

EG

and

JPE

G20

00

56

2006

PSN

R-H

VS

Egi

azar

ian

etal

.[11

3]IQ

Yes

No

FFS

Gra

ysu

b.2

(44)

Noi

se,

blur

,JP

EG

and

JPE

G20

00

56

Con

tinue

don

Nex

tPag

e...

328

Page 345: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2006

PSN

R-H

VS-

ME

giaz

aria

net

al.[

113]

IQY

esN

oFF

SG

ray

sub.

2(4

4)N

oise

,bl

ur,

JPE

Gan

d

JPE

G20

00

56

2006

Min

dru

and

Jung

[288

]IQ

Yes

No

FFS

Col

orsu

b.1

(3)

Hal

fton

ing

(A)

2006

SSIM

IPT

Bon

nier

etal

.[41

]ID

No

No

FFS

Col

orsu

b.15

(90)

Gam

utm

appi

ng22

2006

mPS

NR

Mun

kber

get

al.[

310]

IQN

oN

oN

SC

olor

sub.

16(-

)H

DR

(A)

2007

RC

BM

Don

get

al.[

106]

IQN

oN

oFF

SG

ray

obj.

29(2

04)

JPE

GSc

enes

from

LIV

E

2007

SEM

ESi

lva

etal

.[40

6]IS

No

No

FFS

Gra

ysu

b.(2

33)

JPE

G-

Scen

esfr

omL

IVE

2007

CIS

ML

eeet

al.[

261]

HT

Yes

No

FFS

Col

orob

j.1

(28)

Hal

fton

ing

-

2007

VSN

RC

hand

ler

and

Hem

ami

[70]

IFY

esN

oFF

SG

ray

sub.

29(7

79)

--

Scen

esfr

omL

IVE

2007

P-C

IEL

AB

(ΔP

E)

Cho

uan

dL

iu[7

8]IF

No

No

VFS

Col

orsu

b.3

(6)

JND

profi

lean

dJP

EG

(A)

2007

QM

CS

Yao

etal

.[48

4]IQ

No

No

FFS

Gra

ysu

b.29

(344

)JP

EG

and

JPE

G2k

-Sc

enes

from

LIV

E

2007

DE

CO

R−

WSN

RW

anet

al.[

452]

HT

Yes

No

FFS

Gra

yob

j.1

(3)

Err

ordi

ffus

ion

-

2007

PSN

R-H

VS-

MPo

nom

aren

koet

al.

[366

]

IQY

ES

No

FFS

Gra

ySu

b.1

(18)

Noi

se

2008

DP

Gra

nger

[159

]C

DN

oN

oN

SC

olor

obj.

-

2008

Lam

and

Loo

[250

]IQ

No

No

NS

Gra

ysu

b.2

(6)

Noi

se

2008

SBL

CG

orle

yan

dH

ollim

an

[158

]

IQN

oN

oN

SG

ray

sub.

3(5

4)JP

EG

com

pres

sion

20

2008

”bus

ynes

s”O

rfan

idou

etal

.[33

1]IQ

No

No

FFS

Gra

ysu

b.10

(80)

JPE

Gan

dJP

EG

2kco

m-

pres

sion

10Ps

ycho

phys

ical

data

from

Alle

n

etal

.[7]

Con

tinue

don

Nex

tPag

e...

329

Page 346: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2008

Spat

ialΔ

E00

Che

net

al.[

74]

IQY

esN

oFF

SC

olor

obj.

1(1

)B

lurr

ing

-

2008

CE

DN

agui

bet

al.[

311]

IQN

oN

oFF

SG

ray

Sub

(227

)V

ario

us-

-

2009

ΔE

EO

lear

ieta

l.[3

29]

CD

No

No

FFS

Col

or-

--

-V

ario

uste

stin

g

2009

S-D

EE

Sim

one

etal

.[41

2]ID

Yes

No

FFS

Col

orSu

b25

(170

0)-

-Sc

enes

from

TID

2009

SHA

ME

Pede

rsen

and

Har

debe

rg

[351

]

IQY

esN

oFF

SC

olor

Sub

--

-V

ario

uste

stin

g

2009

SHA

ME

-II

Pede

rsen

and

Har

debe

rg

[351

]

IQY

esN

oFF

SC

olor

Sub

--

-V

ario

uste

stin

g

2009

CO

MB

Bia

nco

etal

.[35

]IQ

Yes

No

FFS

Col

orSu

b29

(779

)V

ario

us-

LIV

Eda

taba

se

2009

DR

ajas

heka

ret

al.[

370]

IQN

oN

oFF

SC

olor

Sub

1(10

)V

ario

us(A

)-

2009

AB

FW

ang

and

Har

debe

rg

[459

]

IDY

esN

oFF

SC

olor

Sub

10(4

20)

Var

ious

10-

2010

PSN

RV

Zha

oan

dD

eng

[501

]IQ

Yes

Yes

FFS

Gra

ySu

b(1

69)

Var

ious

-Sc

enes

from

LIV

E

2010

S DO

G-C

IEL

AB

Aja

gam

elle

etal

.[5]

IQY

esY

esFF

SC

olor

Sub

-V

ario

us-

Var

ious

eval

uatio

n

2010

S DO

G-D

EE

Aja

gam

elle

etal

.[5]

IQY

esY

esFF

SC

olor

Sub

-V

ario

us-

Var

ious

eval

uatio

n

2010

Cao

etal

.[61

]IQ

No

No

FFS

Col

orSu

b19

(38)

Gam

utm

appi

ng12

Exp

erto

bser

vers

2010

PSN

R-H

VS-

STo

nget

al.[

440]

IQY

esN

oFF

SC

olor

Sub

-V

ario

us-

TID

2008

and

LIV

Eda

taba

se

2010

PSN

R-H

VS-

M-S

Tong

etal

.[44

0]IQ

Yes

No

FFS

Col

orSu

b-

Var

ious

-T

ID20

08an

dL

IVE

data

base

2010

RFS

IMZ

hang

etal

.[49

5]IQ

Yes

No

FFS

Gra

ysu

b25

(170

0)-

838

TID

2008

2010

MD

OG

-DE

ESi

mon

eet

al.[

410]

IDY

esY

ES

FFS

Col

or-

--

-V

ario

us

2010

Sim

ple

Zha

nget

al.[

494]

IQN

oY

ES

FFS

Gra

y-

--

-Fi

veim

age

qual

ityda

taba

ses

Con

tinue

don

Nex

tPag

e...

330

Page 347: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

Tabl

eA

.1–

Con

tinue

d

Yea

rM

etri

cA

utho

r(s)

Type

HV

SM

SS/

NS

C/G

Test

Scen

esM

odifi

catio

nO

bser

vers

Com

men

t

2011

TV

DPe

ders

enet

al.[

358]

IQY

esN

oFF

SC

olor

Sub

-V

ario

uspr

inte

rs-

Prin

ted

imag

esfr

omPe

ders

en

etal

.[3

59]

and

Pede

rsen

etal

.

[347

]

2011

FSIM

Zha

nget

al.[

496]

IQY

esN

oFF

SG

ray

Sub

-V

ario

us-

Six

imag

equ

ality

data

base

s

2011

FSIM

CZ

hang

etal

.[49

6]IQ

Yes

No

FFS

Col

orSu

b-

Var

ious

-Si

xim

age

qual

ityda

taba

ses

2011

NSE

RM

ouet

al.[

307]

IQN

oY

esFF

SG

ray

Sub

-V

ario

us-

Six

imag

equ

ality

data

base

s

2011

BB

CQ

Shoh

amet

al.[

403]

IQN

oN

oFF

SG

ray

Sub

-V

ario

us-

Two

imag

ese

tsw

ithco

mpr

essi

on

331

Page 348: Doctoral Dissertation by Marius Pedersen

OVERVIEW OF IMAGE QUALITY METRICS

ΔE∗ ab

ΔE

94

Peff

erko

rnan

dB

linΔ

E00

Spat

ialΔ

E00

CM

CB

FDΔ

PE

PDS-

CIE

LA

B

Kim

mel

etal

.C

MPS

NR

CV

DM

CV

DM

exte

nsio

n

John

son

and

Fair

child

iCA

I cm

Neu

man

nFa

rrug

iaan

dPe

roch

eH

uean

gle

SHA

ME

SHA

ME

-II

I∗ colo

rV

DP

HD

R-V

DP

WV

DP

IFA

CIF

A

pdif

fPD

M

VD

M

Sarn

off

JND

Vis

ion

Mod

elSE

-VD

MB

olin

and

Mey

er

UIQ

SSIM

Min

dru

and

Jung

DT

WT-

SSIM

CW

SSIM

WC

WSS

IM

SSIM

IPT

MSS

IMC

ISM

CB

M

RC

BM

ESS

IM

GSS

IM

MH

IN

NE

TQ

MC

SL

aman

dL

oo

Qco

lor

UQ

I-H

VS

PSN

R-H

VS

NN

ET

Ivko

vic

and

Sank

ar

Fig

ure

A.1

:O

verv

iew

ofth

ere

lati

onsh

ips

betw

een

ase

lect

ion

ofth

eIQ

met

rics

.P

ink

oval

sar

eco

lor

met

rics

,w

hile

squa

red

gray

boxe

sar

egr

aysc

ale

met

rics

.A

nar

row

from

one

met

ric

toan

othe

rin

dica

tes

that

the

met

ric

inco

rpor

ates

feat

ures

from

the

pare

ntm

etri

c.

332

Page 349: Doctoral Dissertation by Marius Pedersen

B MEASURING PERCEPTUAL CONTRAST

IN A MULTI-LEVEL FRAMEWORK

In this appendix we present the Weighted multi-Level Framework (WLF) contrast metric.First a brief introduction to the background required for understanding the WLF metric, andthen the metric is defined.

B.1 Background

B.1.1 Tadmor and Tolhurst

Tadmor and Tolhurst [431] based their analysis of contrast on the Difference of Gaussians(DOG) model, which is modified and adapted to natural images.

In the conventional DOG model, the spatial sensitivity in the center component of thereceptive-fields is described by a bi-dimensional Gaussian with unit amplitude:

Center(x,y) = exp

[−(

xrc

)2

−(

yrc

)2]

, (B.1)

where the radius rc represents the distance at which the sensitivity decreases to 1/e with re-spect to the peak level and (x,y) are the spatial coordinates where the receptive-field is placed.The surround component is represented by another Gaussian curve, with a larger radius, rs:

Surround(x,y) = 0.85

(rc

rs

)2

exp

[−(

xrs

)2

−(

yrs

)2]

. (B.2)

When the central point of the receptive-field is placed at the location (x,y), the output of thecentral component is calculated as:

Rc(x,y) =i=x+3rc

∑i=x−3rc

j=y+3rc

∑j=y−3rc

Center(i− x, j− y)I(i, j), (B.3)

333

Page 350: Doctoral Dissertation by Marius Pedersen

MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK

while the output of the surround component is:

Rs(x,y) =i=x+3rs

∑i=x−3rs

j=y+3rs

∑j=y−3rs

Surround(i− x, j− y)I(i, j), (B.4)

where in both cases I(i,j) is the image pixel value at position (i,j).

The simplest case, with rc = 1 and rs = 2, results in a 7× 7 center mask and a 13× 13surround mask.

The result of the DOG model is obtained as:

DOG(x,y) =Rc(x,y)Rs(x,y)

. (B.5)

The conventional DOG model assumes that the response of a neuron depends uniquelyon the local luminance difference (ΔI) between the center and the surround. After the lightadaptation process, the gain of the ganglion cells of the retina and the Lateral GeniculateNucleus (LGN) neurons depends on the average local luminance I. Thus the model responsedepends on the contrast stimulus, , and the DOG model must be modified by a division by thelocal mean luminance. They propose the following criterium for the measure of contrast:

cT T (x,y) =Rc(x,y)−Rs(x,y)Rc(x,y)+Rs(x,y)

(B.6)

In their experiments, using 256 × 256 images, the overall image contrast is calculatedas the average local contrast of 1000 pixel locations taken randomly while assuring that thecenter and surround masks do not exceed the edges of the image:

CT T =1

1000

1000

∑n=1

cT Tn (B.7)

B.1.2 Rizzi et al.

Rizzi et al. [374] have developed a very simple and efficient measure, able to estimate globaland local components of contrast. It is based on two principles: to collect a simplified mea-sures of difference among neighboring pixel and to do it on various frequency levels, accord-ing to some evidences, which reports that the use of multilevel as an important implementationfeature to mimic HVS [3, 145].We will refer to this contrast measure as RAMM.

It performs a pyramid subsampling of the image to various levels in the CIELAB colorspace [85]. A pyramidal structure is created by halving the image at each iteration. For eachlevel, it calculates the local contrast in each pixel by taking the average difference between thelightness channel value of the pixel and the surrounding eight pixels, thus obtaining a contrastmap of each level. The final overall measure is a recombination of the average contrast foreach level:

CRAMM =1Nl

Nl

∑l=1

cl, (B.8)

334

Page 351: Doctoral Dissertation by Marius Pedersen

MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK

where Nl is the number of levels and, cl is the mean contrast in the level l:

cl =imax

∑i=1

jmax

∑j=1

ci, j, (B.9)

where imax and jmax indicate respectively the height and the width of the image, and ci, j is thecontrast of each pixel calculated as:

ci, j = ∑n∈N8

αn∣∣Pixeli, j −Pixeln

∣∣ .. (B.10)

The following values are used to define the weights of the neighboring pixels:

αn =1

4+2√

2

⎡⎢⎣

√2

2 1√

22

1 1√2

2 1√

22

⎤⎥⎦ (B.11)

This measure has a computational complexity of Θ(NlogN), where N is the number ofpixels, which is lower than alternative local methods, keeping a comparable level of correct-ness in the contrast estimate [374].

The steps of the measure are illustrated in Figure B.1(a).

B.1.3 Retinal-like subsampling contrast

Rizzi et al. [375] combined RAMM’s multilevel approach with Tadmor and Tolhursts evalua-tion of a color stimulus [375]. It works with the same pyramid subsampling as RAMM withthe following differences: it computes in each pixel of each level the DOG contrast instead ofthe 8-neighborhood local contrast and it computes the DOG contrast not only for the lightnessbut also for the two chromatic channels. The three independent measures of each channel arethen merged by a weighted linear combination. The final overall measure can be expressedby the formula:

CRSC = α ·CRSCL∗ +β ·CRSC

a∗ + γ ·CRSCb∗ , (B.12)

where α,β, and γ represent the weights of each color channel. The steps of the measure aredescribed in Figure B.1(b).

The computational complexity of the RSC measure is the same as for RAMM:

Θ(NlogN) , (B.13)

where N is the number of pixels, but with a slightly heavier multiplication constant due to theDOGs instead of the 8-neighbor difference computation.

As well as the previous presented measure, only one number of contrast is produced at theend, the averages of all the levels are averaged again among them with uniform weights.

This measure takes the name of Retinal-like Subsampling Contrast (RSC) and it derivesfrom the fact that the DOG model has been used successfully in many studies to describe thereceptive fields and responses of mammalian retinal ganglion cells and LGN neurones [431]

335

Page 352: Doctoral Dissertation by Marius Pedersen

MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK

and from the way of building the pyramid structure [374].

B.2 The weighted-level framework

In this section we introduce the Weighted-Level Framework (WLF) and we address mainlytwo aspects: the subsampling method and the weights in the level recombination [417].

An antialiasing filter is introduced in the subsampling in order to minimize distortionartifacts at low resolutions.

As demonstrated by Frankle and McCann [145], Adelson et al. [3] and [362], each levelhas a different contribution to contrast so we redefine Equation B.8 as follows:

Ci =1Nl

Nl

∑l=1

λl · cl, (B.14)

where Nl is the number of levels, cl is the mean contrast in the level l and i indicate the appliedcolor channel as before and the new parameter λl is the weight assigned to each level l.

The overall final measure is given by:

CWLF = α ·C1 +β ·C2 + γ ·C3, (B.15)

where α,β, and γ are the weights of each color channel. The measure can be extended todifferent color spaces such as XYZ and RGB and it is not limited to CIELAB.

The general structure of our proposed measure can be seen in Figure B.1(c) where the mostimportant novelties red are shown in red: the antialiasing filter in the pyramid and weightedrecombination of local contrast maps. In this framework the previously developed RAMMand RSC can be considered as special cases with uniform weighting levels in the CIELABcolor space.

For more details and evaluation of the WLF we refer the reader to Simone et al. [414, 415,417].

336

Page 353: Doctoral Dissertation by Marius Pedersen

MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK

(a) RAMM

(b) RSC (c) WLF

Figure B.1: Workflow for RAMM, RSC, and WLF. The difference between RAMM and RSC isfound in the neighborhood calculation of local contrast, where RAMM uses a 8-neighborhoodwhile RSC uses DOG-neighborhood. RAMM and RSC are just special cases with uniformweighting levels in CIELAB color space of WLF, which implements an antialiasing filter inthe subsampling, a weighted recombination of the local contrast maps and it is extended alsoto RGB and XYZ color space.

337

Page 354: Doctoral Dissertation by Marius Pedersen

MEASURING PERCEPTUAL CONTRAST IN A MULTI-LEVEL FRAMEWORK

338

Page 355: Doctoral Dissertation by Marius Pedersen

C SALIENCY MODELS AS

GAMUT-MAPPING ARTIFACT DETECTORS

C.1 Saliency map

Visual Saliency is the perceptual quality that makes an object, person, or pixel region stand outrelative to its neighbors and thus capture our attention. Visual attention results both from fast,pre-attentive, bottom-up visual saliency of the retinal input, as well as from slower, top-downmemory and volition based processing that is task-dependent [316].

Saliency estimation methods can broadly be classified as biologically based, purely com-putational, or a combination of both. In general, all methods employ a low-level approachby determining contrast of image regions relative to their surroundings, using one or morefeatures of intensity, color, and orientation [1]. Many models have been proposed based onthese low-level features and almost all have a low resolution and ill-defined object bound-ary [180, 214, 278]. A recent model [1] outperforms the rest in that it is able to efficientlyoutput full resolution saliency maps, establish well-defined boundaries of salient objects anddisregard high frequencies arising from texture, noise and blocking artifacts.

In brief, this saliency detection algorithm applies the Difference-of-Gaussian (DoG) filterfor band pass filtering. The DoG filter is widely used in edge detection and is suitable fordetecting the artifacts due to the contouring effect in our context. The DoG filter is given by:

DoG(x,y) = 12π [ 1

σ21e− x2+y2

2σ21 − 1

σ22e− x2+y2

2σ22 ] = G1(x,y,σ1)−G2(x,y,σ2),

where σ1 and σ2 are the standard deviation of the gaussian distribution. In order to ensure thesalient regions will be fully covered and not just highlighted on edges or in the center of theregions, we drive σ1 to infinity. This results in a notch in frequency at DC component whileretaining all other frequencies. To remove high frequency noise and textures, we use a smallGaussian kernel keeping in mind the need for computational simplicity. For small kernels, thebinomial filter approximates the Gaussian very well in the discrete case. We use 1/16[1, 4, 6,4, 1] giving ωhc = π/2.75.

The method of finding the saliency map S for an image I of width W and height H pixelscan thus be formulated as:

S(x,y) = ‖Iμ − Iωhc(x,y)‖, (C.1)

where Iμ is the mean image feature vector, Iωhc(x,y) is the corresponding image pixel vectorvalue in the Gaussian blurred version (using a 5 x 5 separable binomial kernel) of the originalimage, and ‖‖ is the L2 norm. Using the Lab color space, each pixel location is an [L,a,b]T

339

Page 356: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

vector, and the L2 norm is the Euclidean distance. We have the images transformed intoCIELAB space to take the effect of lightness on saliency into account. An illustration ofsaliency detection over the original can be found in Figure C.1

Figure C.1: The original image and its corresponding saliency map.

C.2 Applicability of saliency maps to gamut mapping arti-

facts

If an artifact is detectable in the image, it should be salient, and therefore saliency modelsshould be applicable for detecting artifacts. If salient regions are lost, such as important detailsin the images, there will be a difference in saliency between the original and reproduction.Also in this case saliency models should be able to detect artifacts.

The first step to ensure that saliency maps are applicable to detect artifacts is to investigatethe saliency maps from gamut mapped images. We have computed the saliency map foran original image and a gamut mapped image (Figure C.2). In the gamut mapped imagesignificant loss of details occur, being one of the artifacts we focus on. Investigation of thesaliency maps show that they are not similar, and that the saliency changes even though theimage is the same. In this case the saliency maps have been computed over the entire image.The results indicate that a difference of saliency can be suitable for detecting artifacts.

C.3 Experimental framework: saliency models as artifact

detectors

We propose a framework for using saliency models for the detection of gamut mapping arti-facts. Two different approaches, one global and one local, are explained in details below.

340

Page 357: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

100 200 300 400 500 600

50

100

150

200

250

300

350

400

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

100 200 300 400 500 600

50

100

150

200

250

300

350

400

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

100 200 300 400 500 600

50

100

150

200

250

300

350

400

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Figure C.2: The illustration of different gamut-mapped images and their correspondingsaliency maps. The first row shows the original image and the gamut mapped version withloss of details. In the next row the saliency map of the original is on the left, in the middlethe saliency map of the gamut mapped version of the original. The right image shows thenormalized absolute difference in saliency between the two normalized saliency maps. As wecan see the original and the gamut mapped version have difference salient regions, indicatingthat a change of saliency can be used to detect artifacts.

341

Page 358: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

C.3.1 Global strategy

The global approach is very straightforward as explained above. We detect saliency for theoriginal and gamut mapped image directly and then derive the difference between the twosaliency maps. An adaptive thresholding with the Otsu’s method, which chooses the thresholdto minimize the intraclass variance of the black and white pixels, is used to locate the regionswhere the artifacts lie in the difference image [332]. Figure C.3 show the steps of the globalapproach. The method can also be illustrated by the following formula,

Figure C.3: The experimental framework using the global method. Saliency maps are createdof both the original and reproduction, and the absolute different between these two maps aretaken to get an indication of where artifacts occur.

D(x,y) = S(x,y)−S(xg,yg), (C.2)

where S(x,y) is the saliency map derived from Equation C.1 and S(xg,yg) is the saliency mapupon the gamut-mapped image. D(x,y) is the difference image which shows the artifacts bythe global strategy.

We illustrate the global method in Figure C.4. For the images in the first row, the biggestdifference of the gamut mapped image from the original is the loss of details in the eagle’sbody. The saliency model is able to detect the attention shift due to the reproduction and giveus the right result.

342

Page 359: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

The contouring dominates the artifacts in the second row of images. However, since theheart of flower has a big color contrast from the background and attracts much of the attention,the saliency model will tell the difference in this region. It shows the inadequacy of theartifact detection using global approach and leads us to further investigate the image in specificregions. We build on the global strategy by using the saliency maps in local regions.

The artifact detection based on the global saliency gives a rough illustration of where theyare, but it can be misleading since it is always difficult to set the threshold to extract the artifactregions from the background. Some regions equally attract much attention from the originalto the reproduction and we cannot obtain the salient shift. Therefore we are required to pickout the errors from the specific regions.

Figure C.4: The original images are listed in the left column, the gamut mapped ones are inthe middle and the results from the global method are shown on the right. For the first row,the global saliency detection generally tells where the artifacts lie. However, in the next row,the heart of flower stands out of both original and reproduced image. The global thresholdingis no longer sufficient to detect the artifacts.

C.3.2 Local strategy

The local strategy is shown in Figure C.5, and is as follows: The original image is firstlytransformed into the CIELAB color space considering about its lightness, and then the mean-shift clustering [88] is applied on the images. Using the CIELAB color space, each pixelis represented in [L* a* b*] vectors, which means the lightness and chromatic values areseparated. In this way, we are able to investigate the influence of color in different lightness

343

Page 360: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

levels. The mean-shift algorithm is a nonparametric clustering technique which is commonlyused for edge preserving and image segmentation. It does not require prior knowledge of thenumber of clusters, the constrain of the shape of the clusters and it provides better boundary.The spatial band width is set as 3 and the range band width is 1.

Figure C.5: The experimental framework using the local method. By using the original imagea mask is created with the meanshift algorithm in the CIELAB colorspace. This mask is usedtogether with the saliency models to create segmented saliency maps of both the original andreproduction. Then the absolute difference of these two maps are taken to find artifacts.

With an adaptive threshold, the image could be separated into several parts allowing forthe analysis of specific artifacts in different lightness level. In the case of detail loss, whichis a problem in shadow and highlight areas, especially in gamut mapping, the image canbe divided into two parts. One part generally contains the high lightness while the other islow lightness. Due to the shadow effect, we tend to have more focus on the regions in lowlightness distribution. The adaptive thresholding is applied onto the mean-shift transformedimage of the original. The same mask partitioning the original image will also be used tofilter the reproduction, ensuring a correct comparison of regions. Then, the saliency detectionis applied and the difference is extracted upon the segmented regions to show the artifactregions.Compared with Equation C.2, the artifact detection based on local strategy is shown as,

M = Seg(Ilab(x,y)), (C.3)

andD(x,y) = S(M(x,y))−S(M(xg,yg)), (C.4)

344

Page 361: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

where Ilab(x,y) is the original image transformed into CIELAB color space. Seg is the mean-shift filtering taking the low lightness into account, and thereby M is the mask from the orig-inal image. Equation C.4 is similar to Equation C.2, except we apply the same mask on bothoriginal and gamut-mapped images.

Since the global model gives too high weight on the areas that capture the attention, localregions are now taking into account in depth. We can see that, the occurrence of artifacts ismuch higher in the regions with low lightness than the high lightness. Therefore, we focuson detecting the artifacts in the regions with low-lightness distribution. Figure C.6 show thatwe are able to obtain the regions where the artifacts occur from the model when comparedagainst the areas agreed upon by the authors to contain artifacts.

With the local method, we are able to obtain all the artifacts in the eagle’s body and itseyes for the first set. In the second set of images, the edges of the white blobs are detected byour method.

Figure C.6: The images in the left column illustrate the artifacts by our local saliency detec-tion model, and the columns are the artifacts agreed upon by the authors to contain artifacts.

For more details and evaluation of the metric we refer the reader to Cao et al. [61].

345

Page 362: Doctoral Dissertation by Marius Pedersen

SALIENCY MODELS AS GAMUT-MAPPING ARTIFACT DETECTORS

346

Page 363: Doctoral Dissertation by Marius Pedersen

D DETECTION OF WORMS IN ERROR

DIFFUSION HALFTONING

D.1 Introduction

Digital halftoning is the process of transforming a continuous tone image to a binary image,to allow printing or displaying on a bi-level display. Many different techniques have beenproposed for the transformation. One of these being error diffusion [144] halftoning, its sim-plicity and ability to produce high quality reproductions has contributed to its popularity. Thistechnique distributes the quantization error along the path of the image scan. This is done bya filter, determining the weight distribution of the error. The original four element filter byFloyd and Steinberg [144], as seen in Equation D.1, resulted in different artifacts.

One of the artifacts produced by error diffusion is worms, where black and white pixelsstring together are perceived as having a vermicular texture as seen on Figure D.1(a) [423].This artifact is found in the highlights and shadows of the halftoned image. The original filterby Floyd and Steinberg suffers from this artifact, and because of this numerous error diffusionalgorithms have been proposed.

116

[0 ∗ 73 5 1

](D.1)

Jarvis et al. [218] proposed a 12 element filter (Equation D.2), another 12 element filterwas proposed by Stucki [427] (Equation D.3), both of these break up worm patterns, butthey introduce artifacts in mid-tone areas [257]. Later Fan [129, 400] proposed a filter witha long tail (Equation D.4), this results in breaking up worm structures in the highlights andshadows while performing similar to the Floyd and Steinberg filter in the mid-tones [257].An overview of other filters, each with different advantages and disadvantages, is given byMonga et al. [293].

148

⎡⎣0 0 ∗ 7 5

3 5 7 5 31 3 5 3 1

⎤⎦ (D.2)

142

⎡⎣0 0 ∗ 8 4

2 4 8 4 21 2 4 2 1

⎤⎦ (D.3)

116

[0 0 ∗ 71 3 5 0

](D.4)

347

Page 364: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

(a) Left-to-right (b) Serpentine

Figure D.1: The result of using (a) Floyd and Steinberg left-to-right scan path and (b) Floydand Steinberg with serpentine scan path. The serpentine scan path breaks up worm patterns,and results in a visually more pleasing image as seen on Figure D.1(b).

Different approaches to scanning paths have been proposed to create visually more pleas-ing images. The original Floyd and Steinberg [144] algorithm used a left-to-right scan path,later a serpentine (boustrophedonic) processing was introduced where the scan is left-to-rightand then right-to-left (Figure D.1). Other more creative paths have been proposed based onspace filling curves, as the Peano and Hilbert paths [257] that break up worms, but introducesignificant amount of noise.

Quality assessment is needed to show if a halftone technique increase the quality of animage compared to other state of the art techniques. There are basically two different ways toassess the quality of a halftoned images; subjectively and objectively. Subjective evaluationis carried out by observers, and is therefore influenced by the Human Visual System (HVS).Objective evaluation can be carried out in many ways. One typical way is to use measurementdevices gathering numerical values. Another way is to use algorithms, commonly known asimage quality metrics, in an attempt to quantify quality. Image quality metrics are usuallydeveloped to take into account properties of the HVS, and thus with the goal of being wellcorrelated with subjective evaluations.

A number of these image quality metrics for halftone quality have been proposed. Leeet al. [261] proposed an image similarity measure to evaluate a hybrid error diffusion algo-rithm. Veryovka et al. [449] proposed a multiscale approach to analyze edges, and used it toevaluate different halftoning algorithms. Mitsa and Varkur [290] proposed an image qualitymeasure for halftoning based on a spatial filtering of Mean Square Error (MSE). Scheermesserand Bryngdahl [385] proposed a texture metric for halftone images, that find the occurrenceor absence of specific textures in quantized images. Nilsson [319] proposed an image qual-ity metric for halftone images that incorporated models of both the printer and observer. S-CIELAB was proposed by Zhang and Wandell [499] as a spatial pre-processor of the CIELABcolor difference formulae [82], and was designed to predict image difference for halftoning,distortion, etc. For a complete list of different metrics for halftoning and other areas we referthe reader to Pedersen and Hardeberg [355], Chapter 4, or Appendix A.

Even though many measures have been proposed for halftoning, many of them only workon a global scale without having a specific calculation for worms. A specific measure for

348

Page 365: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

worms could be used to preserve worms where they are wanted, as in edges, but removethem in areas where they are not wanted. A specific measure for worms could also be used toselect the most optimal halftoning algorithm for an image, resulting in the best possible outputimage. Such a measure for worms could also be included in a toolbox to detect halftoningartifacts.

D.2 Proposed error diffusion worm measure

This section introduces a measure for detection and quantification of worms, called ErrorDiffusion Worm Measure (EDWM). The measure is a multi-step measure as shown in Fig-ure D.2. It is a no-reference measure, where no information about the original is used in thecalculation.

Halftonedimage

Extracthighlights

andshadows

Cannyedgefilter

Removesinglepixels

Extract allpixels

within edgeboundary

Proximitydetection

Removesmall

worms

Finalmeasure

Figure D.2: Description of the workflow of the EDWM. Highlights and shadows are extractedfrom the halftoned image before a Canny edge filter is applied. Single pixels are removed,before all pixels within in the edge boundary are extracted. Then a proximity detection isperformed before small worms are removed. The final measure is calculated as the percentageof non-worm pixels divided by the number of pixels in the image.

Since worms are found in highlights and shadows of the image, the halftoned image isfiltered to remove parts without highlights and shadows. A sliding filter with width and heightof 10 pixels is used, where the mean of each location is found. A filter size of 10 × 10pixels was found to be sufficient to find the information needed to extract highlights andshadows. Thresholds for when a pixel is considered as highlight or shadow is set at 0.15 and0.85 respectively, on a scale where 0 is where white pixels (no dot of ink placed) are foundwithin the filter and 1 where all pixels are black (a drop of ink). These limits visually giveshighlights and shadows. The parts of the halftoned image within the limits (≤ 0.15 and ≥0.85 of a grayscale range from 0 to 1) are extracted for further processing.

A Canny edge filter is used to extract edges, where black pixels (a drop of ink) are foundnext to white pixel areas (no drop of ink), in the image [60]. This filter uses a low and ahigh threshold for hysteresis, and a sigma as the standard deviation for the Gaussian filter.Pixels placed close together will therefore be identified as forming an edge, and a worm canbe thought of as an edge going through the image. Figure D.3 shows an example where pixelsform a worm, and the Canny filter correctly detects this as an edge. In most cases a sigma of1, resulting in a 4 × 4 pixel Gaussian filter, yields good results together with a heuristicallychosen pair of thresholds depending on the data. Another advantage of using an edge filteris the fact that worms can be found in both highlights and shadow areas; in the highlightsdrops of ink make the worms, while in the shadow areas the ink-free areas make the worms.The edge filter does not differentiate between these, just the difference between dark and lightareas resulting in a robust method, and no special care must be taken if it is in a highlight orshadow area.

In the cases of a single pixel without any connecting pixels the Canny filter will result inan edge around the pixel. A simple filter is applied to remove these single pixels, along with

349

Page 366: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

(a) Halftoned image (b) Canny filtered image

Figure D.3: Error diffused image and image after Canny edge detection. The Canny edgedetection results in a ’shell’ around the black pixels.

other small pixels not connected to worms. This is done by checking the local neighbourhoodaround the pixel to identify whether it is connected to other pixels or not, if not connected thepixel is removed and not considered as a part of a worm.

In order to find the correct pixels that make up the worm the boundary of the object istraced, since this is a binary image a black and white boundary tracing is performed. Theexterior of the object is traced along with the boundaries of the holes.

The basis of the worm structure is now found, and the next step of the worm detection isa proximity calculation. The pixels inside a worm are extracted and for each pixel the closestnon-worm pixel is located, if this pixel is within a specified distance and angle it is consideredas part of the worm. The angle is based on the angle from the closest detected worm pixel(Figure D.4). This process is iterative, so when a pixel is found this pixel is then furtherchecked until all detected pixels are processed. The distance and angle can be set according tothe desired sensitivity. Testing has shown that a distance of 50 pixels and angle of 40 degreesgives good results.

Figure D.4: Proximity calculation. Square pixels indicate detected worm pixels. An anglelimit and distance limit is set. All non detected pixels (round pixels) inside the limit are found,in this case one pixel is inside both the angle and distance limit. The direction of the angleand distance limit is based on the closest detected pixel, this is done to detect pixels that arelinked. The process is iterative and is carried out until all pixels are processed.

All worms have now been detected regardless of size. It is well known that our humanvisual system filters out information that we cannot process. This also applies to worm, andbecause of this worms that or too small for observers to notice should not be taken into accountin the calculation [449]. Therefore the size of the worm is found by using dilation to increaseeach pixel. The worms will now ”grow” together, and the area of connected dilated pixels iscomputed. If the area is below a threshold, the pixels are removed from the detected worms.

The final value for the measure is calculated as the percentage of non-worm points dividedby the number of pixels in the image. This will result in a value between 0 and 1, where 1 is a

350

Page 367: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

Figure D.5: Experimental setup. Observers ranked four different images reproduced with fivedifferent error diffusion algorithms.

worm free image. This normalization will result in the possibility to compare EDWM valuesacross different images. Furthermore the map of worms, computed before the calculation ofthe final value, can be useful to detect areas where pixels can be rearranged to remove wormartifacts.

D.3 Evaluation of the error diffusion worm measure

Evaluation is an important aspect when developing a measure. An experiment was carriedout to evaluate the performance of the proposed measure. Correlation between the perceivedworms and predicted worms is adopted as the measure of performance. There are many waysto evaluate halftone quality, and our experiment is inspired by previous experiments carriedout by Mitsa and Varkur [290], Tontelj et al. [444] and Axelson [15].

D.3.1 Experimental setup

Four different images were chosen for a psychophysical experiment as seen on Figure D.6.One artistic image (Figure D.6(b)), with both highlights and shadow areas. One gradientramp (Figure D.6(a)), from white to black. Wang and Parker [454] stated that a gradientramp is a basic and important evaluation in most halftone evaluations. A gradient will re-veal worms, and is often used to demonstrate the performance of error diffusion algorithms[15, 243, 261, 336, 386, 400]. One highlight patch (Figure D.6(d), 2% black) and one shadowpatch (Figure D.6(c), 98% black). The highlight and shadow patch are areas where we nor-mally will have worms, and are therefore appropriate images for the evaluation error diffusionalgorithms.

The images are 600 pixels wide and 200 pixels high, except the artistic image wherethe height is 450 pixels. These images were reproduced with five different error diffusionalgorithms, Floyd and Steinberg (FS) [144], Stucki (S) [427], Fan (F) [129, 400], Jarvis,Judice and Ninke (JJN) [218] and Floyd and Steinberg Serpentine (FSS) [257, 446]. The

351

Page 368: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

images were printed on a HP ColorJet 4600DN on HP Office Paper White (80 g/m2) at 600DPI simulating 72 DPI.

The experiment was carried out in a dimmed room, with a GretagMacbeth Examoliteoverhead light source with D50 lighting as seen in Figure D.5. The observers were free tomove, and no restrictions on viewing distance were given. 12 observers participated in theexperiment, both males and females.

The experiment was divided into two parts. The first part was to rank the overall qualityof the five versions of the image, while the second part was to rank the images according tothe amount of worms found in the five versions. Before the second part the observers weretaught how to identify worms, and where they occur. This was done to secure reliable data.

(a) Gradient. (b) Artisticimage.

(c) Shadow image. (d) Highlight image.

Figure D.6: The four images used in the experiment, surrounded here by a black frame. Eachwas reproduced with five different error diffusion algorithms.

D.3.2 Results

This section contains results from the psychophysical experiment and the EDWM. Analysis ofthe results are also given. Two types of correlation will be computed for the results, the Pear-son product-moment correlation coefficient and the Spearman’s rank correlation coefficient[232]. The first assumes that the variables are ordinal, and looks for the linear relationshipbetween variables. The second, Spearman, is a non-parametric measure of correlation thatuses the ranks as basis instead of the actual values. This describes the relationship betweenvariables without making any assumptions about the frequency distribution of the variables.

D.3.2.1 Overall scores

The results from the 12 observers were combined and Z-scores were calculated using Thur-stones law of comparative judgement [437], the scores were obtained using the Colour Engi-neering Toolbox [162]. As seen in Figure D.7 Floyd and Steinberg Serpentine (FSS) has thehighest score both for quality and worms. For the worm artifact three algorithms cannot bedifferentiated by the observers, Stucki (S), Fan (F) and Jarvis, Judice and Ninke (JJN). TheFloyd and Steinberg algorithm (FS) gets the lowest score, and clearly has more worms thanthe other error diffusion algorithms. In the quality judgement the FS also has the lowest meanZ-score, but cannot be differentiated from S and F. The JJN was rated better than FS, butcannot be differentiated from S and F.

352

Page 369: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

There is also a high correlation between the quality score (Figure D.7(a)) and the wormscore (Figure D.7(b)), the Pearson correlation gives a value of 0.94 (p value: 0.0197). Thisindicates a strong relationship between the amount of worms and the quality of the image.The images evaluated in the experiment were chosen based on areas where worms in errordiffusion are present. Therefore the conclusion may not be valid in other images with lessworms, i.e. without large highlights and shadow areas.

FS S F JJN FSS−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1Quality All Images

Error Diffusion Algorithms

Z−

Sco

re

(a) Z-scores for all images based on quality judge-ment.

FS S F JJN FSS−1.5

−1

−0.5

0

0.5

1

1.5Worm All Images

Error Diffusion Algorithms

Z−

Sco

re

(b) Z-scores for all images based on worm judge-ment.

Figure D.7: Z-score for quality and worm judgement. From the figures we can see a similar-ity, indicating that worm artifacts are linked with the quality of error diffused images. ThePearson correlation gives a value of 0.94 (p value: 0.0197), supporting the similarities foundin the figures. We can also notice that in the quality judgement the observers use less of the Z-score scale then in the worm judgement, this indicate that observers agree more on the wormjudgement than in the quality judgement.

D.3.3 Gradient image

Figure D.8 and Table D.1 shows the results from the psychophysical experiment for the gra-dient ramp. For the quality judgement (Figure D.8(a)) JJN has the best quality, while FSScannot be differentiated from FS. In the worm judgement (Figure D.8(b)), the result is notas clear. FSS can be differentiated from FS, but not from any other algorithms. This alsoindicates that observers more easily pick one ”favorite” algorithm in the quality judgement(Figure D.8(a)), while in the worm judgement (Figure D.8(b)) the observers do not have ahigh consensus in ranking. Some observers also indicated that this scene was the hardest torank, both when it came to quality and worms.

Figure D.9 shows a scatter plot of the EDWM and observer scores. FS is correctly giventhe lowest score by EDWM as the observers did, and FSS is correctly given the highest EDWMscore. EDWM has ranked the S before F, the opposite of the observers, but the other error dif-fusion algorithms are ranked similarly to ranking of the observers. We also see that all pointsare within a 95% confidence interval. We get a high Pearson correlation, 0.93, indicating thatEDWM can predict the amount of perceived worms for this scene.

353

Page 370: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

FS S F JJN FSS

−1

−0.5

0

0.5

1

1.5

Quality Gradient Ramp Image

Error Diffusion Algorithms

Z−

Sco

re

(a) Z-score for gradient ramp based on quality judge-ment.

FS S F JJN FSS−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8Worm Gradient Ramp Image

Error Diffusion Algorithms

Z−

Sco

re

(b) Z-score for gradient ramp based on worm judge-ment.

Figure D.8: Z-score for quality and worm judgement for the gradient ramp. In the qualityjudgement observers agree more on the ranking than in worm judgement. This is a scenewhere most observers indicated to have problems with ranking according to worms.

Table D.1: Z-scores and EDWM scores for the gradient ramp. We can see that FSS has thehighest EDWM, i.e. the least worms. FS has the lowest value, indicating the most wormartifacts.

Algorithm EDWM Z-score wormsFloyd and Steinberg (FS) 0.98965 -0.4975Fan (F) 0.99133 0.0551Stucki (S) 0.99166 -0.2154Jarvis, Judice and Ninke (JJN) 0.99245 0.2763Floyd and Steinberg Serpentine (FSS) 0.99318 0.3815

354

Page 371: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

Figure D.9: Correlation for gradient ramp between Z-score and EDWM, plotted with 95%Z-score confidence intervals.

D.3.4 Artistic image

Figure D.10 shows the Z-scores for the artistic image, both for quality (Figure D.10(a)) andworm (Figure D.10(b)) judgement. For the quality judgement FSS cannot be differentiatedfrom F and S by the observers, but FSS was clearly better than JJN and FS. For the wormjudgement FS was judged to have more worms than the four other algorithms. These fouralgorithms gave almost the same amount of worms according to the observers.

FS S F JJN FSS−2

−1.5

−1

−0.5

0

0.5

1

Quality Artistic Image

Error Diffusion Algorithms

Z−

Sco

re

(a) Z-score for artistic image based on quality judge-ment.

FS S F JJN FSS−2

−1.5

−1

−0.5

0

0.5

1

Worm Artistic Image

Error Diffusion Algorithms

Z−

Sco

re

(b) Z-score for artistic image based on worm judge-ment.

Figure D.10: Z-score for quality and worm judgement for the artistic image.

Figure D.11 shows a scatter plot of the EDWM and observer score for the artistic image,while data can be found in Table D.2. We have a lower correlation than gradient ramp, butstill the correlation of 0.86 (p value: 0.0596) indicates a strong relation between the predicted

355

Page 372: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

Table D.2: Z-score and EDWM scores for the artistic image. S has been rated as the er-ror diffusion algorithm with the lowest amount of worms (i.e. best rating by the observers).FSS is given the highest EDWM value, but this also has a Z-score close to S and cannot bedifferentiated with a 95% confidence interval.

Algorithm EDWM Z-score wormsFloyd and Steinberg (FS) 0.98398 -1.3830Fan (F) 0.98768 0.2322Stucki (S) 0.99389 0.6196Jarvis, Judice and Ninke (JJN) 0.99187 0.0168Floyd-Steinberg Serpentine (FSS) 0.99503 0.5144

0.982 0.984 0.986 0.988 0.99 0.992 0.994 0.996 0.998−1.5

−1

−0.5

0

0.5

1

1.5Natural image, correlation:0.86(p value:0.0596)

EDWM

Z−

Sco

re

Figure D.11: Correlation for artistic image between Z-score and EDWM, plotted with 95%Z-score confidence intervals.

356

Page 373: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

FS S F JJN FSS−2

−1.5

−1

−0.5

0

0.5

1

1.5

2Quality Highlight Image

Error Diffusion Algorithms

Z−

Sco

re

(a) Z-score for highlight image based on qualityjudgement.

FS S F JJN FSS−2

−1.5

−1

−0.5

0

0.5

1

1.5

2Worm Highlight Image

Error Diffusion Algorithms

Z−

Sco

re

(b) Z-score for highlight image based on wormjudgement.

Figure D.12: Z-score for quality and worm judgement for the highlight image.

worms and perceived worms. Since the EDWM is a no-reference metric, it will not dis-criminate between worms made by edges in the image and worms in highlight/shadow areas.Because of this EDWM should be used with caution in natural images, because the worms inthese images could be wanted in order to reproduce edges. The good correlation found hereindicates that EDWM might be used on natural images.

D.3.5 Highlight image

Figure D.12 shows the results for the highlight image regarding quality (Figure D.12(a)) andworms (Figure D.12(b)), we can see that the ranking is almost similar for the quality andthe worm judgement. FFS has the same amount of worms and quality as JJN, but it can bedifferentiated from the rest.

Table D.3: Z-scores and EDWM scores for the highlight image. FSS is given the highest scoreby the observers, this image also receives the highest EDWM value. We can also see that theranking of EDWM values and observer scores is the same (Figure D.12).

Algorithm EDWM Z-score wormsFloyd and Steinberg (FS) 0.98873 -1.5002Fan (F) 0.99034 -0.7565Stucki (S) 0.99611 -0.3113Jarvis, Judice and Ninke (JJN) 0.99668 1.1630Floyd and Steinberg Serpentine (FSS) 0.99854 1.4050

Figure D.13 shows a scatter plot of the EDWM and observer score for the highlight image.We can see that the ranking of the EDWM values and observer scores is the same (TableD.3), indicating that the EDWM is able to correctly rank the images according to perceivedworms for this image. We also get a high correlation between the scores, indicating the highperformance of EDWM.

357

Page 374: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

0.988 0.99 0.992 0.994 0.996 0.998 1 1.002−2

−1.5

−1

−0.5

0

0.5

1

1.5

2Highlight image correlation:0.91(p value:0.0311)

EDWM

Z−

Sco

re

Figure D.13: Correlation for highlight image between Z-score and EDWM, plotted with 95%Z-score confidence intervals. We can see that the ranking of the EDWM values and observerscores is the same.

D.3.6 Shadow image

Results for the shadow image are found in Figure D.14 and Table D.4. JJN has the lowest qual-ity judgement (Figure D.14(a)) but cannot be differentiated from FS. For the worm judgement(Figure D.14(b)) FSS clearly gave less worms than the four other algorithms. Some observersindicated that this image was the easiest to judge compared to the three other images.

Table D.4: Z-scores and EDWM values for the shadow image. FSS is the best according tothe observers and EDWM.

Algorithm EDWM Z-score wormsFloyd and Steinberg (FS) 0.99069 -1.1752Fan (F) 0.99340 0.2116Stucki (S) 0.99223 -0.5446Jarvis, Judice and Ninke (JJN) 0.98933 -0.4714Floyd and Steinberg Serpentine (FSS) 0.99854 1.9797

Figure D.15 shows a scatter plot of the EDWM values and observer score for the shadowimage. A Pearson correlation of 0.92 (p value: 0.0226) indicates a high performance byEDWM.

D.3.7 Comparison to other metrics

EDWM’s performance is compared against the performance of other state of the art metrics.S-CIELAB [499], Structural Content [120], Average Distance [120] and MSE have been cho-

358

Page 375: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

FS S F JJN FSS

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5Quality Shadow Image

Error Diffusion Algorithms

Z−

Sco

re

(a) Z-score for shadow image based on qualityjudgement.

FS S F JJN FSS

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5Worm Shadow Image

Error Diffusion Algorithms

Z−

Sco

re

(b) Z-score for shadow image based on worm judge-ment.

Figure D.14: Z-score for quality and worm judgement for the shadow image.

0.988 0.99 0.992 0.994 0.996 0.998 1−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5Shadow image correlation:0.92(p value:0.0267)

EDWM

Z−

Sco

re

Figure D.15: Correlation for shadow image between Z-score and EDWM, plotted with 95%Z-score confidence intervals.

359

Page 376: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

sen for comparison. S-CIELAB was made with the intention to predict perceived differencebetween an original and for example a halftoned version of the original. Because of this S-CIELAB is a suitable metric to compare the EDWM against. Structural Content compareschanges in structure between two images, and a worms in the image will affect the struc-tural content. This makes the Structural Content metric suitable for comparison against theEDWM. Average Distance included to have a comparison against traditional measures as wellas new measures, such as the S-CIELAB, which simulate the HVS. The MSE is commonlyused by researchers to test the performance of new metrics, because of its widespread use inimage quality evaluation our results from the EDWM is compared against results from MSE.

All these metrics are full-reference metrics, i.e. they use information from the original im-age to calculate the difference or quality. The continuous tone image was used as a referencefor all metrics. Since EDWM is a no-reference metric, this must be taken into account whenanalyzing the results, and also the fact that none of the other metrics are specifically made forworm artifacts.

EDWM outperforms S-CIELAB, Structural Content, Average Distance and MSE. This isclearly seen on Table D.5, where EDWM is the only measure to have a correlation above 0.85.

Table D.5: Correlation between metric score and Z-score from the worm judgement, with pvalue in parentheses. Gray cells indicate the best correlation. We can see that EDWM out-performs the other metrics in all images. MSE and Average Distance has a good correlationfor the artistic image, but high p values (AD: 0.2985 and MSE: 0.1241) indicate that thecorrelation is by chance.

Metric Artistic Gradient Highlight Shadowimage ramp image image

EDWM 0.86 (0.0596) 0.93 (0.0220) 0.91 (0.0311) 0.92 (0.0226)S-CIELAB -0.59 (0.2906) 0.07 (0.9143) -0.48 (0.4106) 0.64 (0.2500)Structural Content 0.08 (0.8948) -0.09 (0.8836) -0.23 (0.7149) -0.15 (0.822)Average Distance -0.59 (0.2985) -0.19 (0.7550) -0.23 (0.7148) -0.15 (0.8206)MSE -0.77 (0.1241) -0.17 (0.7884) -0.23 (0.7148) 0.15 (0.8206)

D.3.7.1 Overall performance

Since the results from EDWM are normalized we can compare results obtained for differentimages. Figure D.16 shows the correlation between Z-scores from the worm judgement for allimages and EDWM values for all images (Tables D.1 - D.4). We can see a clear relationshipbetween the Z-scores and EDWM values, this is also confirmed by a Pearson correlation of0.81 (p value: 1e-005). Compared to the other metrics EDWM is significantly better (TableD.6). The other measures have a very low correlation, this is because they have big differencesin the values between images, resulting in a correlation close to zero.[350] Our measure alsohas the highest Spearman correlation, indicating that the ranking from it corresponds to theranking made by the observers.

360

Page 377: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

0.982 0.984 0.986 0.988 0.99 0.992 0.994 0.996 0.998 1 1.002−2.5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5Correlation:0.8(p value:1e−005)

ED

WM

Z−Score

Figure D.16: Correlation between all Z-score from worm judgement and all EDWM values,plotted with 95% confidence intervals. From the figure we can see a strong correlation (Pear-son correlation of 0.81, p value: 1e-005). The calculated Spearman correlation is 0.77 (pvalue 0.0001), indicating a correct ranking of predicted worms according to the observerranking.

Table D.6: Correlation between metric score and Z-scores from the worm judgement for allimages, with p values in parentheses. We can see that EDWM outperforms the other metricsin all images with a correlation of 0.81 (p value: 1e-005), and the Spearman correlationof EDWM is also higher than for the other metrics. The other measures have a very lowcorrelation due to big differences in the results between the images. Gray cells indicate thebest correlation.

Metric Pearson correlation Spearman correlationEDWM 0.81 (1e-005) 0.77 (0.0001)S-CIELAB 0.01 (0.9715) 0.06 (0.7967)Structural Content 0.00 (0.9876) -0.12 (0.6267)Average Distance 0.00 (0.9945) 0.00 (0.9849)MSE -0.02 (0.9451) 0.06 (0.7888)

361

Page 378: Doctoral Dissertation by Marius Pedersen

DETECTION OF WORMS IN ERROR DIFFUSION HALFTONING

D.4 Summary

A measure for worms, EDWM, has been proposed. The measure is simple to compute, anddoes not require any information about the original image. It allows for comparison of resultsacross different images, resulting in a robust measure. A psychophysical experiment was car-ried out to evaluate the performance of EDWM, and the results indicate a high performance.EDWM clearly outperforms other metrics.

362

Page 379: Doctoral Dissertation by Marius Pedersen

E FROM CONTRAST TO IMAGE

DIFFERENCE: TWO NEW METRICS

Two new metrics are proposed referred as SDOG-CIELAB and SDOG-DEE. They are in linewith the S-CIELAB approach (Section 4.2.2.1), but the spatial extension is based on thework initiated by Rizzi et al. [374] and the S-DEE metric (Section 4.2.2.2) by Simone etal.[412]. The improvements encompass a multi-level approach, the substitution of the orig-inal S-CIELAB spatial filtering with a DOG calculation, as introduced in Appendix B, andthe use of the ΔEE color-difference formula [329]. The general workflow of the metrics is asfollows (Figure E.1):

• The original image and its reproduction are first converted into the CIELAB color spacefor SDOG-CIELAB and into CIE XY Z for SDOG-DEE.

• The images are subsampled to various levels afterwards. The undersampling is simplesince the images are halfed, and the antialiasing filtering avoids artifacts at low resolu-tions.

• A pixelwise neighborhood contrast calculation is executed in each level using the DOGon the lightness and on the chromatic channels separately, thus providing local contrastmaps for each level and each channel.

• Local contrast errors are computed using ΔE∗ab for SDOG-CIELAB or ΔEE for SDOG-

DEE.

• A weighted recombination of the local contrast maps is finally computed, resulting inglobal ID metrics.

These metrics grew out of research on contrast measures, therefore one could expect the high-est correlations with images containing significant variations in contrast. The DOG modelwas chosen as a surrogate to the CSF filtering because it has revealed to be beneficial in theidentification of edges, while the CSFs are mainly used to modulate the less perceptible fre-quencies. The ΔEE formula was selected because it is statistically equivalent to CIEDE2000in the prediction of many available empirical datasets, but with greater simplicity and clearrelationships with visual processing.

Once that local contrast maps are generated for each level, how to reduce the concept ofcontrast from local values at each pixel location to a single number representing the global IDis an ongoing debate. The simplest strategy is taking the mean of each level and averaging alltogether.

363

Page 380: Doctoral Dissertation by Marius Pedersen

FROM CONTRAST TO IMAGE DIFFERENCE: TWO NEW METRICS

Original Reproduction

Color space transformationL*a*b*

XYZ

Subsampling with antialiasing filter

Pyramid levels

Calculation of the local contrast

Local contrast maps

Color difference calculationΔE*

ab ΔEE

Pyramid of local

contrast maps difference

Weighted recombination

Global image difference measure

Figure E.1: Workflow of the proposed metrics. The metrics are similar to the S-CIELAB(Section 4.2.2.1) described by Zhang and Wandell [499]. The improvements encompass amulti-level approach, the substitution of the S-CIELAB spatial filtering with a DOG calcula-tion, and the use of the ΔEE color-difference formula.

These two new metrics perform a weighted recombination of the levels, given by thefollowing equation:

GlobalID =1Nl

Nl

∑l=1

λl · cl,

where Nl is the number of levels, cl is the mean contrast in the level l, and λl is the weightassigned to each level l.

The metrics have several parameters for the the central component (Rc) and surroundcomponent (Rs) of the DOG filter, the nature of the subsampling, and the type of weightinglevels in the pyramid. In this work the subsampling is done using a pyramid where the pyramidis expressed by the series 1, 1

2 , 14 , 1

8 , 116 , .... For a detailed overview of the parameters, we refer

the reader to Simone et al. [413] and to Appendix B.For more information about these two metrics we refer the reader to Ajagamelle et al. [5],

Ajagamelle et al. [6], or Ajagamelle [4].

364

Page 381: Doctoral Dissertation by Marius Pedersen

F SPECIFICATIONS

In this appendix we reproduce the specification sheets of the devices and media used in thisstudy.

F.1 Monitors

Two different monitors have been used (Figures F.1 and F.2), and their specifications are foundin Tables F.1 and F.2.

F.1.1 Eizo CG241W

Figure F.1: Eizo CG241W. Reproduced from www.puremac.de.

365

Page 382: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

Table F.1: Eizo CG241W specifications.

Panel Size 24.1” / 61 cm (611 mm diagonal)Active Display Size (H V) 518.4 324 mmPanel Type VA (with overdrive circuit)Viewing Angles (H, V) 178◦, 178◦ (at contrast ratio of 10:1)Brightness 300 cd/m2 (maximum), 120 cd/m2 or less (recom-

mended)Contrast 850:1Native Resolution 1920 1200 (16:10 aspect ratio)Pixel Pitch 0.270 0.270 mmDisplay Colors 16.77 million from a palette of 68 billionWide Gamut Coverage Adobe RGB: 95%, NTSC:91%, sRGB/Rec.709:

98%, EBU 98%, SMPTE-C:100%, DCI:87%Look-Up Table 16 bits per color

366

Page 383: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.1.2 Eizo CG211

Figure F.2: Eizo CG211. Reproduced from www.eizo.pl.

Table F.2: Eizo CG211 Specifications

Panel Size 22.2” / 56.4 cm (563 mm diagonal)Active Display Size (H V) 478 299 mmPanel Type IPSViewing Angles (H, V) 170◦, 170◦ (at contrast ratio of 10:1)Brightness 200 cd/m2 (maximum), 100 cd/m2 or less (recom-

mended)Contrast 400:1Native Resolution 1920 1200 (16:10 aspect ratio)Pixel Pitch 0.249 0.249 mmDisplay Colors 16.77 million from a palette of 278 trillionnWide Gamut Coverage 98% of sRGB and Adobe RGBLook-Up Table 16 bits per color

367

Page 384: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.2 Scanners

Different scanners (Figures F.5,F.3, and F.4) have been used to digitize the printed images,and their specifications are found in Tables F.3, F.4, and F.5.

F.2.1 Microtek ScanMaker 9800XL

Figure F.3: Microtek ScanMaker 9800XL. Reproduced from www.burak.pl.

Table F.3: Microtek ScanMaker 9800XL specifications.

Type Flatbed color image scannerOptical Sensor CCD

Optical Resolution 1600dpi x 3200dpiHardware Resolution 2400 x 4800 dpi with Micro Step Drive technology

Color Bit Depth 48-bit internalOptical Density 3.7 Dmax

Maximum Scan Area 12” X 16.9”Light Source Cold Cathode Fluorescent Lamp (CCFL)

368

Page 385: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.2.2 Epson Expression 10000XL

Figure F.4: Epson Expression 10000XL. Reproduced from http://www.scanstation.co.uk.

Table F.4: Epson Expression 10000XL specifications

Type Flatbed color image scannerOptical Sensor Color Epson MatrixCCD line sensor

Optical Resolution 2400 dpiHardware Resolution 2400 x 4800 dpi with Micro Step Drive technologyMaximum Resolution 12,800 x 12,800 dpi

Effective Pixels 87,840 pixels / line (2400 dpi)Color Bit Depth 48-bit internal / external

Grayscale Bit Depth 16-bit internal / externalOptical Density 3.8 Dmax

Maximum Scan Area 12.2” x 17.2”Light Source Xenon gas cold cathode fluorescent lamp

Scanning Speed 2400 dpi in draft mode, Color 16.0 msec / line,Grayscale 5.3 msec / line, Line Art 5.3 msec / line

Focus Control AutoFocus optics system (CCD and lens unit)

369

Page 386: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.2.3 HP ScanJet G4050

Figure F.5: HP ScanJet G4050. Reproduced from http://www.officecritter.com.

Table F.5: HP ScanJet G4050 specifications.

Type Flatbed color image scannerOptical Sensor CCD

Optical Resolution 4800 dpiHardware Resolution 4800 x 9600 dpi

Bit Depth 96 bitGrayscale levels 256

Maximum Scan Area 21.6 x 31.1 cmLight Source Xenon gas cold cathode fluorescent lamp

370

Page 387: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.3 Printers

The specifications of the printers used in this work are found below (Figures F.6, F.7, F.8 andTables F.6, F.7, F.8).

F.3.1 Oce Colorwave 600

Figure F.6: Oce ColorWave 600. Reproduced from www.oce.com.

Table F.6: Oce Colorwave specifications.

Description Wide Format TonerPearl PrinterTechnology Oc CrystalPoint

Speed B&W: 31 sec/A0; Color: 34 sec/A0Warm-up Time 6 minutes (from cold start-up)Print Resolution Adressable raster: 1.200 dpiPaper Capacity Up to 200 m per roll, max 6 rolls

Output Sizes Width 279 mm - 1067 mmOutput Sizes Length 210 mm - 3000 mm (longer possible, might influ-

ence side margins)Paper Weight 60 g/m2 -160 g/m2Media Type Uncoated, recycled, colored, blue back, tyvek,

polyester films and more

371

Page 388: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.3.2 HP Designjet 10ps

Figure F.7: HP DesignJet 10ps. Reproduced from www.tiendatinta.com.

Table F.7: HP Designjet 10ps specifications.

Description Six-color printer (CMYKcm)Technology Inkjet

Speed up to 4.0 ppm - Color draft, up to 10.0 ppm - Colorbest, up to 1.4 ppm - Color normal

Print Resolution 2400 x 1200 dpiPaper Capacity 150.0 sheets

Max Media Size 13.0 in x 19.0 inPaper Weight 60 g/m2 -160 g/m2Media Type Heavy-weight coated paper, Glossy photo paper,

Opaque vinyl, High-gloss photo paper, Semi-glossphoto paper, Imaging film, Plain paper, Glossy pa-per, Coated paper, Glossy coated paper

372

Page 389: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.3.3 HP Color LaserJet 4600dn

Figure F.8: HP Color LaserJet 4600dn. Reproduced from www.hp.com.

Table F.8: HP Color LaserJet 4600dnspecifications.

Description Color laser printerTechnology Laser

Speed 17 ppm black and colorPrint Resolution 600 × 600 dpi

Media type Media type Paper (plain, glossy, film, recycled),envelopes, transparencies, labels, cardstock

373

Page 390: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.4 Paper

The specification of the different papers used is found in Tables F.9, F.10, F.11, and F.12.

F.4.1 Oce LFM 050 Red Label

Table F.9: Oce LFM 050 Red Label specifications.

Weight 75 g/m2

Whiteness CIE 159Thickness 99 μ m

Acidity 7.5 pHISO Brightness R457 + UV 108%ISO Brightness R457 - UV 88%

Opacity 92%Coating Not coated

F.4.2 Oce LFM 054 Red Label

Table F.10: Oce LFM 054 Red Label specifications.

Weight 75 g/m2

Whiteness CIE 159Thickness 103 μ m

Acidity 7.5 pHISO Brightness R457 + UV 109%ISO Brightness R457 - UV 88%

Opacity 92%Coating Not coated

374

Page 391: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.4.3 Stora Enso MultiCopy Original

Table F.11: Stora Enso MultiCopy Original specifications.

Weight 80 g/m2

Whiteness CIE 160Thickness 108 μ mOpacity 91,5%Coating Not coated

F.4.4 HP office paper

Table F.12: HP office paper specifications.

Weight 80 g/m2

Size A4

375

Page 392: Doctoral Dissertation by Marius Pedersen

SPECIFICATIONS

F.5 Viewing cabinet

F.5.1 Verivide DTP

The Verivide DTP viewing cabinet has been used in this work (Figure F.9 and Table F.13)

Figure F.9: Verivide DTP. Reproduced from www.verivide.com.

Table F.13: Verivide DTP specifications.

Light source D50Internal color Munsell grey N5

Viewing cavity 670×460 mmOverall size 695×610×390 mm

376

Page 393: Doctoral Dissertation by Marius Pedersen

G LEGAL

The author hereof has been enabled by Oce-Technologies B.V. to perform research activitieswhich underlies this document. This document has been written in a personal capacity. Oce-Technologies B.V. disclaims any liability for the correctness of the data, considerations andconclusions contained in this document.

377

Page 394: Doctoral Dissertation by Marius Pedersen