22.12.2013 Views

Additional Material, Journal of Imaging Science - Society for Imaging ...

Additional Material, Journal of Imaging Science - Society for Imaging ...

Additional Material, Journal of Imaging Science - Society for Imaging ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

JIST<br />

Vol. 51, No. 4<br />

July/August<br />

2007<br />

<strong>Journal</strong> <strong>of</strong><br />

<strong>Imaging</strong> <strong>Science</strong><br />

and Technology<br />

imaging.org<br />

<strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology


Editorial Staff<br />

Melville Sahyun, editor<br />

sahyun@infionline.net<br />

Donna Smith, production manager<br />

dsmith@imaging.org<br />

Editorial Board<br />

Philip Laplante, associate editor<br />

Michael Lee, associate editor<br />

Nathan Moroney, associate editor<br />

Mitchell Rosen, color science editor<br />

David S. Weiss, associate editor<br />

David R. Whitcomb, associate editor<br />

JIST papers are available <strong>for</strong> purchase<br />

at www.imaging.org and through<br />

ProQuest. They are indexed in<br />

INSPEC, Chemical Abstracts, <strong>Imaging</strong><br />

Abstracts, COMPENDEX, and ISI:<br />

<strong>Science</strong> Citation Index.<br />

Orders <strong>for</strong> subscriptions or single<br />

copies, claims <strong>for</strong> missing numbers,<br />

and notices <strong>of</strong> change <strong>of</strong> address<br />

should be sent to IS&T via one <strong>of</strong> the<br />

means listed below.<br />

IS&T is not responsible <strong>for</strong> the accuracy<br />

<strong>of</strong> statements made by authors and<br />

does not necessarily subscribe to their<br />

views.<br />

Copyright ©2007, <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong><br />

<strong>Science</strong> and Technology. Copying<br />

<strong>of</strong> materials in this journal <strong>for</strong> internal<br />

or personal use, or the internal or personal<br />

use <strong>of</strong> specific clients, beyond<br />

the fair use provisions granted by the<br />

US Copyright Law is authorized by<br />

IS&T subject to payment <strong>of</strong> copying<br />

fees. The Transactional Reporting Service<br />

base fee <strong>for</strong> this journal should be<br />

paid directly to the Copyright Clearance<br />

Center (CCC), Customer Service,<br />

508/750-8400, 222 Rosewood Dr.,<br />

Danvers, MA 01923 or online at<br />

www.copyright.com. Other copying<br />

<strong>for</strong> republication, resale, advertising or<br />

promotion, or any <strong>for</strong>m <strong>of</strong> systematic<br />

or multiple reproduction <strong>of</strong> any material<br />

in this journal is prohibited except with<br />

permission <strong>of</strong> the publisher.<br />

Library <strong>of</strong> Congress Catalog Card<br />

No. 59-52172<br />

Printed in the USA.<br />

<strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and<br />

Technology<br />

7003 Kilworth Lane<br />

Springfield, VA 22151<br />

www.imaging.org<br />

info@imaging.org<br />

703/642-9090<br />

703/642-9094 fax<br />

Manuscripts should be sent to the<br />

postal address above as describe at<br />

right. E-mail PDF and other files as requested<br />

to dsmith@imaging.org.<br />

Guide <strong>for</strong> Authors<br />

Scope: The <strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology (JIST) is dedicated to the advancement <strong>of</strong> imaging science knowledge, the<br />

practical applications <strong>of</strong> such knowledge, and how imaging science relates to other fields <strong>of</strong> study. The pages <strong>of</strong> this journal are<br />

open to reports <strong>of</strong> new theoretical or experimental results, and to comprehensive reviews. Only original manuscripts that have not<br />

been previously published nor currently submitted <strong>for</strong> publication elsewhere should be submitted. Prior publication does not refer<br />

to conference abstracts, paper summaries, or non-reviewed proceedings, but it is expected that <strong>Journal</strong> articles will expand in scope<br />

the presentation <strong>of</strong> such preliminary communication. Please include keywords on your title and abstract page.<br />

Editorial Process/Submission <strong>of</strong> Papers <strong>for</strong> Review: All submitted manuscripts are subject to peer review. (If a manuscript appears better<br />

suited to publication in the <strong>Journal</strong> <strong>of</strong> Electronic <strong>Imaging</strong>, published jointly by IS&T and SPIE, the editor will make this recommendation.)<br />

To expedite the peer review process, please recommend two or three competent, independent reviewers. The editorial staff, will<br />

take these under consideration, but is not obligated to use them.<br />

Manuscript Guidelines: Please follow these guidelines when preparing accepted manuscripts <strong>for</strong> submission.<br />

• Manuscripts should be double-spaced, single-column, and numbered. It is the responsibility <strong>of</strong> the author to prepare a succinct,<br />

well-written, paper composed in proper English. JIST generally follows the guidelines found in the AIP Style Manual, available<br />

from the American Institute <strong>of</strong> Physics.<br />

• Documents may be created in Micros<strong>of</strong>t Word, WordPerfect, or LaTeX/REVTeX.<br />

• Manuscripts must contain a title page that lists the paper title, full name(s) <strong>of</strong> the author(s), and complete affiliation/address <strong>for</strong><br />

each author. Include an abstract that summarizes objectives, methodology, results, and their significance; 150 words maximum.<br />

Provide at least four key words.<br />

• Figures should con<strong>for</strong>m to the standards set <strong>for</strong>th at www.aip.org/epub/submitgraph.html.<br />

• Equations should be numbered sequentially with Arabic numerals in parentheses at the right margin. Be sure to define symbols<br />

that might be confused (such as ell/one, nu/vee, omega/w).<br />

• For symbols, units, and abbreviations, use SI units (and their standard abbreviations) and metric numbers. Symbols, acronyms,<br />

etc., should be defined on their first occurrence.<br />

• Illustrations: Number all figures, graphs, etc. consecutively and provide captions. Figures should be created in such a way that<br />

they remain legible when reduced, usually to single column width (3.3 inches/8.4 cm); see also<br />

www.aip.org/epub/submitgraph.html <strong>for</strong> guidance. Illustrations must be submitted as .tif or .eps files at full size and 600 dpi;<br />

grayscale and color images should be at 300 dpi. JIST does not accept .gif or .jpeg files. Original hardcopy graphics may be sent<br />

<strong>for</strong> processing by AIP, the production house <strong>for</strong> JIST. (See note below on color and supplemental illustrations.)<br />

• References should be numbered sequentially as citations appear in the text, <strong>for</strong>mat as superscripts, and list at the end <strong>of</strong> the document<br />

using the following <strong>for</strong>mats:<br />

• <strong>Journal</strong> articles: Author(s) [first/middle name/initial(s), last name], “title <strong>of</strong> article (optional),” journal name (in italics), ISSN<br />

number (e.g. <strong>for</strong> JIST citation, ISSN: 1062-3701), volume (bold): first page number, year (in parentheses).<br />

• Books: Author(s) [first/ middle name/initial(s), last name], title (in italics), (publisher, city, and year in parentheses) page reference.<br />

Conference proceedings are normally cited in the Book <strong>for</strong>mat, including publisher and city <strong>of</strong> publication (Springfield, VA, <strong>for</strong> all<br />

IS&T conferences), which is <strong>of</strong>ten different from the conference venue.<br />

• Examples<br />

1. H. P. Le, Progress and trends in ink-jet printing technology, J. <strong>Imaging</strong> Sci. Technol. 42, 46 (1998).<br />

2. E. M. Williams, The Physics and Technology <strong>of</strong> Xerographic Processes (John Wiley and Sons, New York, 1984) p. 30.<br />

3. Gary K. Starkweather, “Printing technologies <strong>for</strong> images, gray scale, and color,” Proc. SPIE 1458: 120 (1991).<br />

4. Linda T. Creagh, “Applications in commercial printing <strong>for</strong> hot melt ink-jets,” Proc. IS&T’s 10th Int’l. Congress on Adv. In<br />

Non-Impact Printing Technologies (IS&T, Springfield, VA 1994) pp. 446-448.<br />

5. ISO 13655-1996 Graphic technology: Spectral measurement and colorimetric computation <strong>for</strong> graphic arts images (ISO,<br />

Geneva), www.iso.org.<br />

6. <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology website, www.imaging.org, accessed October 2003.<br />

Reproduction <strong>of</strong> Color: Authors who wish to have color figures published in the print journal will incur color printing charges.<br />

The cost <strong>for</strong> reproducing color illustrations is $490 per page; color is not available to those given page waivers, nor can color page<br />

charges be negotiated or waived. Authors may also choose to have their figures appear in color online and in grayscale in the printed<br />

journal. There is no additional charge <strong>for</strong> this, however those who choose this option are responsible <strong>for</strong> ensuring that the captions<br />

and descriptions in the text are readable in both color and black-and-white as the same file will be used in the online and<br />

print versions <strong>of</strong> the journal. Only figures saved as TIFF/TIF or EPS files will be accepted <strong>for</strong> posting. Color illustrations may be<br />

also submitted as supplemental material <strong>for</strong> posting on the IS&T website <strong>for</strong> a flat fee <strong>of</strong> $100 <strong>for</strong> up to five files.<br />

Website Posting <strong>of</strong> Supplemental <strong>Material</strong>s: Authors may also submit additional (supplemental) materials related to their articles<br />

<strong>for</strong> posting on the IS&T Website. Examples <strong>of</strong> such materials are charts, graphs, illustrations, or movies that further explain the<br />

science or technology discussed in the paper. Supplemental materials will be posted <strong>for</strong> a flat fee <strong>of</strong> $100 <strong>for</strong> up to five files. For<br />

each additional file, a $25 fee will be charged. Fees must be received be<strong>for</strong>e supplemental materials will be posted. As a matter <strong>of</strong><br />

editorial policy, appendices are normally treated as supplemental material.<br />

Submission <strong>of</strong> Accepted Manuscripts: Author(s) will receive notification <strong>of</strong> acceptance (or rejection) and reviewers’<br />

reports. Those whose manuscripts have been accepted <strong>for</strong> publication will receive correspondence in<strong>for</strong>ming them <strong>of</strong> the issue <strong>for</strong><br />

which the paper is tentatively scheduled, links to copyright and page charge <strong>for</strong>ms, and detailed instructions <strong>for</strong> submitting accepted<br />

manuscripts. A duly signed transfer <strong>of</strong> copyright agreement <strong>for</strong>m is required <strong>for</strong><br />

publication in this journal. No claim is made to original US Government works.<br />

Page charges: Page charges <strong>for</strong> the <strong>Journal</strong> is $80/printed page. Such payment is<br />

not a condition <strong>for</strong> publication, and in some circumstances page charges are<br />

waived. Requests <strong>for</strong> waivers must be made in writing to the managing editor prior<br />

to acceptance <strong>of</strong> the paper and at the time <strong>of</strong> submission.<br />

Manuscripts submissions: Manuscripts should be submitted both electronically<br />

and as hardcopy. To submit electronically, send a single PDF file attached to an e-<br />

mail message/cover letter to jist@imaging.org. To submit hardcopy, mail 2 singlespaced,<br />

single-sided copies <strong>of</strong> the manuscript to: IS&T. With both types <strong>of</strong> submission,<br />

include a cover letter that states the paper title; lists all authors, with complete<br />

contact in<strong>for</strong>mation <strong>for</strong> each (affiliation, full address, phone, fax, and e-mail); identifies<br />

the corresponding author; and notes any special requests. Unless otherwise<br />

stated, submission <strong>of</strong> a manuscript will be understood to mean that the paper has<br />

been neither copyrighted, classified, or published, nor is being considered <strong>for</strong><br />

publication elsewhere. Authors <strong>of</strong> papers published in the <strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong><br />

<strong>Science</strong> and Technology are jointly responsible <strong>for</strong> their content. Credit <strong>for</strong> the<br />

content and responsibility <strong>for</strong> errors or fraud are borne equally by all authors.<br />

JOURNAL OF IMAGING SCIENCE AND TECH-<br />

NOLOGY ( ISSN:1062-3701) is published bimonthly<br />

by The <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology,<br />

7003 Kilworth Lane, Springfield, VA 22151. Periodicals<br />

postage paid at Springfield, VA and at<br />

additional mailing <strong>of</strong>fices. Printed in Virginia,<br />

USA.<br />

<strong>Society</strong> members may receive this journal as part <strong>of</strong><br />

their membership. Forty-five dollars ($45.00) <strong>of</strong><br />

membership dues are allocated to this subscription.<br />

IS&T members may refuse this subscription by written<br />

request. Domestic institution and individual nonmember<br />

subscriptions are $195/year or $50/single<br />

copy. The <strong>for</strong>eign subscription rate is $205/year.<br />

For online version in<strong>for</strong>mation, contact IS&T.<br />

POSTMASTER: Send address changes to JOURNAL<br />

OF IMAGING SCIENCE AND TECHNOLOGY,<br />

7003 Kilworth Lane, Springfield, VA 22151.


JIST<br />

Vol. 51, No. 4<br />

July/August<br />

2007<br />

<strong>Journal</strong> <strong>of</strong><br />

<strong>Imaging</strong> <strong>Science</strong><br />

and Technology ®<br />

Feature Article<br />

283 Improved Calibration <strong>of</strong> Optical Characteristics <strong>of</strong> Paper by an Adapted<br />

Paper-MTF Model<br />

Safer Mourad<br />

General Papers<br />

293 Gloss Granularity <strong>of</strong> Electrophotographic Prints<br />

J. S. Arney, Ling Ye, Eric Maggard, and Brian Renstrom<br />

299 Forensic Examination <strong>of</strong> Laser Printers and Photocopiers Using Digital<br />

Image Analysis to Assess Print Characteristics<br />

J. S. Tchan<br />

310 Moiré Analysis <strong>for</strong> Assessment <strong>of</strong> Line Registration Quality<br />

Nathir A. Rawashdeh, Daniel L. Lau, Kevin D. Donohue,<br />

and Shaun T. Love<br />

317 Analysis <strong>of</strong> the Influence <strong>of</strong> Vertical Disparities Arising in Toed-in<br />

Stereoscopic Cameras<br />

Robert S. Allison<br />

328 Improved B-Spline Contour Fitting Using Genetic Algorithm <strong>for</strong> the<br />

Segmentation <strong>of</strong> Dental Computerized Tomography Image Sequences<br />

Xiaoling Wu, Hui Gao, Hoon Heo, Oksam Chae, Jinsung Cho, Sungyoung Lee,<br />

and Young-Koo Lee<br />

337 Colorimetric Characterization Model <strong>for</strong> Plasma Display Panel<br />

Seo Young Choi, Ming Ronnier Luo, Peter Andrew Rhodes, Eun Gi Heo,<br />

and Im Su Choi<br />

348 Real-Time Color Matching Between Camera and LCD Based on 16-bit<br />

Lookup Table Design in Mobile Phone<br />

Chang-Hwan Son, Cheol-Hee Lee, Kil-Houm Park, and Yeong-Ho Ha<br />

360 Solving Under-Determined Models in Linear Spectral Unmixing <strong>of</strong><br />

Satellite Images: Mix-Unmix Concept (Advance Report)<br />

Thomas G. Ngigi and Ryutaro Tateishi<br />

continued on next page<br />

imaging.org<br />

<strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology


IS&T BOARD OF DIRECTORS<br />

President<br />

Eric G. Hanson<br />

Department Manager<br />

Hewlett Packard Company<br />

Immediate Past President<br />

James R. Milch Jim<br />

Director Research & Innovation Labs.<br />

Carestream Health, Inc.<br />

Executive Vice President<br />

Rita H<strong>of</strong>mann<br />

Chemist, R&D Manager<br />

Il<strong>for</strong>d <strong>Imaging</strong> Switzerland GmbH<br />

continued from previous page<br />

368 Color Shift Model-Based Segmentation and Fusion <strong>for</strong> Digital Aut<strong>of</strong>ocusing<br />

Vivek Maik, Dohee Cho, Jeongho Shin, Donghwan Har, and Joonki Paik<br />

380 Error Spreading Control in Image Steganographic Embedding Schemes Using<br />

Unequal Error Protection<br />

Ching-Nung Yang, Guo-Jau Chen, Tse-Shih Chen, and Rastislav Lukac<br />

386 In Situ X-ray Investigation <strong>of</strong> the Formation <strong>of</strong> Metallic Silver Phases During the<br />

Thermal Decomposition <strong>of</strong> Silver Behenate and Thermal Development <strong>of</strong><br />

Photothermographic Films<br />

B. B. Bokhonov, M. R. Sharafutdinov, B. P. Tolochko, L. P. Burleva, and D. R. Whitcomb<br />

Conference Vice President<br />

Robert R. Buckley Rob<br />

Research Fellow<br />

Xerox Corporation<br />

Publication Vice President<br />

Franziska Frey<br />

Assist. Pr<strong>of</strong>., School <strong>of</strong> Print Media<br />

Rochester Institute <strong>of</strong> Technology<br />

Secretary<br />

Ramon Borrell<br />

Technology Strategy Director<br />

Hewlett Packard Company<br />

Treasurer<br />

Peter D. Burns<br />

Principal Scientist<br />

Carestream Health, Inc.<br />

Vice Presidents<br />

Choon-Woo Kim<br />

Inha University<br />

Laura Kitzmann<br />

Marketing Dev. & Comm. Manager<br />

Sensient <strong>Imaging</strong> Technologies, Inc.<br />

Michael A. Kriss<br />

MAK Consultants<br />

Ross N. Mills<br />

CTO & Chairman<br />

imaging Technology international<br />

IS&T Conference Calendar<br />

For details and a complete listing <strong>of</strong> conferences, visit www.imaging.org<br />

Digital Fabrication Processes Conference<br />

September 16–September 20, 2007<br />

Anchorage, Alaska<br />

General chair: Ross Mills<br />

NIP23: The 23rd International Congress on<br />

Digital Printing Technologies<br />

September 16–September 20, 2007<br />

Anchorage, Alaska<br />

General chair: Ramon Borrell<br />

IS&T/SID’s Fifteenth Color <strong>Imaging</strong><br />

Conference cosponsored by SID<br />

November 5–November 9, 2007<br />

Albuquerque, New Mexico<br />

General chairs: Jan Morovic<br />

and Charles Poynton<br />

Electronic <strong>Imaging</strong><br />

IS&T/SPIE 20th Annual Symposium<br />

January 26–January 31, 2008<br />

San Jose, Cali<strong>for</strong>nia<br />

General chairs: Nitin Sampat<br />

CGIV 2008: The Fourth European Conference<br />

on Color in Graphics, Image and Vision<br />

June 10–13, 2008<br />

Terrassa, Spain<br />

General chair: Jaume Pujol<br />

Archiving 2008<br />

June 24–27, 2008<br />

Bern, Switzerland<br />

General chair: Rudolf Gschwind<br />

Jin Mizuguchi<br />

Pr<strong>of</strong>essor, Yokohama National Univ.<br />

David Weiss<br />

Scientist Fellow, Eastman Kodak<br />

Company<br />

Chapter Director<br />

Franziska Frey – Rochester<br />

Patrick Herzog – Europe<br />

Takashi Kitamura – Japan<br />

Executive Director<br />

Suzanne E. Grinnan<br />

IS&T Executive Director<br />

ii J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 283–292, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Improved Calibration <strong>of</strong> Optical Characteristics <strong>of</strong> Paper<br />

by an Adapted Paper-MTF Model<br />

Safer Mourad <br />

Empa, Swiss Federal Laboratory <strong>for</strong> <strong>Material</strong>s Testing and Research, Laboratory <strong>for</strong> Media Technology,<br />

Dübendorf, Switzerland<br />

E-mail: safer.mourad@empa.ch<br />

Abstract. The calibration <strong>of</strong> color printers is highly influenced by<br />

optical scattering. Light scattered at microscopic level within printed<br />

papers induces a blurring phenomenon that affects the linearity <strong>of</strong><br />

the tone reproduction curve. The induced nonlinearity is known as<br />

optical dotgain. Engeldrum and Pridham analyzed its impact on<br />

printing, using Oittinen’s light scattering model. They determined the<br />

scattering and absorption coefficients based on spectral measurements<br />

<strong>of</strong> solid patches only. Their calibration achieves good independence<br />

<strong>of</strong> any printing irregularities. However, the microscopic<br />

knife-edge measurements <strong>of</strong> Arney et al. showed that the model<br />

overestimates the influence <strong>of</strong> the absorption coefficient. Unlike<br />

Oittinen’s model, we directly approach the laterally scattered light<br />

fluxes. This is achieved by an extended three-dimensional Kubelka-<br />

Munk model. We describe how to determine our coefficients using<br />

measurements <strong>of</strong> mere solid patches, which allows us to decouple<br />

the optical dot gain from other printing influences. Our improved<br />

model successfully corrects the observed overestimation and is able<br />

to predict Arney’s microscopic measurements. © 2007 <strong>Society</strong> <strong>for</strong><br />

<strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4283<br />

INTRODUCTION<br />

The appearance <strong>of</strong> halftone images is determined by optical<br />

scattering in paper, which greatly affects the calibration <strong>of</strong><br />

color printers. Optical scattering is the reason <strong>for</strong> the<br />

nonlinearity <strong>of</strong> the tone reproduction curve known as optical<br />

dot gain. This paper describes an improved and easy<br />

characterization method <strong>for</strong> optical dot gain.<br />

Engeldrum and Pridham 1 analyzed the effects <strong>of</strong> optical<br />

dot gain on printing using Oittinen’s scattering model. 2 This<br />

estimates the lateral extent <strong>of</strong> the scattering effect based on<br />

the Kubelka-Munk analysis <strong>of</strong> mere vertical radiant fluxes<br />

originally proposed <strong>for</strong> uni<strong>for</strong>m paints. 3 Engeldrum and<br />

Pridham estimated the values <strong>of</strong> the classical Kubelka-Munk<br />

fitting coefficients <strong>of</strong> light absorption and scattering by spectral<br />

measurements <strong>of</strong> printed solid patches, i.e., fulltone<br />

single-color patches printed with full area coverage. They<br />

applied the coefficients obtained to Oittinen’s model and<br />

predicted the optical lateral point spread function. Using<br />

simple solid patches <strong>of</strong>fers the significant advantage <strong>of</strong> attaining<br />

calibration independent <strong>of</strong> any halftone printing irregularities.<br />

However, Arney et al. compared microscopic<br />

<br />

IS&T Member<br />

Received Dec. 23, 2005; accepted <strong>for</strong> publication Jan. 28, 2007.<br />

1062-3701/2007/514/283/10/$20.00.<br />

knife-edge measurements <strong>of</strong> different paper substrates with<br />

the numerical results <strong>of</strong> Oittinen and Engeldrum’s model<br />

and observed overestimated influence <strong>of</strong> the absorption coefficient<br />

on the width <strong>of</strong> the predicted spread function. 4,5<br />

In contrast to Oittinen’s model, our approach intrinsically<br />

accounts <strong>for</strong> scattered lateral light fluxes. 6,7 Our concept<br />

extends the classical Kubelka-Munk model to threedimensional<br />

space and analyses the balances <strong>of</strong> diffuse light<br />

fluxes across the six faces <strong>of</strong> an elementary paper volume<br />

cube. Since the lateral fluxes are explicitly considered, the<br />

extended model improves the discrimination between the<br />

scattering and absorption coefficients. After a brief review <strong>of</strong><br />

the related background, we introduce the model used as a<br />

mathematical tool and demonstrate how to determine its<br />

fitting coefficients without using any microscopic device. We<br />

then present a few application results and compare them<br />

with the findings <strong>of</strong> Arney et al. 5<br />

BACKGROUND<br />

Nowadays, color printers are calibrated based on measurements<br />

<strong>of</strong> color patches that encompass the whole color<br />

gamut. Usually the patches are arranged within standardized<br />

color wedges and printed according to a known input set <strong>of</strong><br />

device-dependent sampling points. Together with the measured<br />

colors, the values <strong>of</strong> the sampling points constitute the<br />

printer’s tone transfer function. Common color management<br />

systems require these transfer functions in a tabulated<br />

<strong>for</strong>m called the calibration pr<strong>of</strong>ile. In practice, the tone<br />

transfer function <strong>of</strong> printers is highly nonlinear and needs to<br />

be measured on a large number <strong>of</strong> at least one thousand<br />

printed color sampling points. The measured color response<br />

deviations from linearity are referred to as dot gain and are<br />

induced by two distinct effects. The first effect is called mechanical<br />

or physical dot gain and arises due to the nonlinear<br />

response <strong>of</strong> the reproduction process. It leads, e.g., to differences<br />

between the intended dot size and the dot size actually<br />

printed. The second effect is the optical dot gain 8 , also called<br />

Yule-Nielsen effect 9 , and is generally caused by the phenomenon<br />

<strong>of</strong> scattered light within the paper substrate. It induces<br />

microscopically a blurring phenomenon, which is the reason<br />

<strong>for</strong> the optical dot gain. Both dot gains have similar influences<br />

on the printed halftone images and hence on the tone<br />

reproduction curve. There<strong>for</strong>e, it is hard to characterize their<br />

distinct impact by means <strong>of</strong> mere macroscopic reflectance<br />

283


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

measurements <strong>of</strong> reproduced color wedges. However, controlling<br />

and improving the image quality <strong>of</strong> color prints ultimately<br />

requires an analytical understanding <strong>of</strong> both effects<br />

in detail. In this publication we consider only the optical dot<br />

gain.<br />

In order to assess the optical effect, several authors have<br />

proposed and used direct microanalytical measurements.<br />

5,10–13 These approaches have a microscopic device in<br />

common, which projects a focused image <strong>of</strong> either a knife<br />

edge, 10 an isolated illumination point, 12 or a sinusoidal<br />

pattern 11 on top <strong>of</strong> the paper. Using a microspectrophotometer<br />

allows the visible optical effect <strong>of</strong> the scattered light to<br />

be characterized by spectral measurements <strong>of</strong> the spatial distribution<br />

<strong>of</strong> the reflected light. However, use <strong>of</strong> such a microscopic<br />

device is not always af<strong>for</strong>dable, especially from the<br />

perspective <strong>of</strong> printer manufacturers or graphic arts bureaux.<br />

We propose to simplify these ef<strong>for</strong>ts by introducing a<br />

calibration technique based on our light scattering and color<br />

halftone model. 6,7<br />

The analysis <strong>of</strong> the optical properties <strong>of</strong> paper can be<br />

considered as a special application <strong>of</strong> light scattering in turbid<br />

media. 14 From a fundamental point <strong>of</strong> view, light scattering<br />

can be derived from Maxwell’s equations; see, <strong>for</strong> instance,<br />

Ishimaru. 15 Concerning this approach, however, the<br />

same author also states: “... its drawback is the mathematical<br />

complexities involved, and its usefulness is limited.” [Ref.<br />

16, p. 2210]. On the other hand, transport theory directly<br />

models the transport <strong>of</strong> radiant power through turbid media.<br />

Because <strong>of</strong> its experimental con<strong>for</strong>mity, transport theory<br />

is preferred in a large number <strong>of</strong> applications. The pragmatic<br />

success is thereby emphasized and the approximating character<br />

<strong>of</strong> the solutions is accepted, which is especially the case<br />

<strong>for</strong> the simple and popular variants like the Kubelka-Munk<br />

two-flux theory, see, e.g., Ref. 17. In this tradition our color<br />

prediction model follows an engineering approach <strong>for</strong> printing<br />

applications, particularly in connection with the control<br />

and calibration <strong>of</strong> color halftone printers where “engineering”<br />

stands <strong>for</strong> a balance between simplicity <strong>of</strong> use and accuracy<br />

<strong>of</strong> prediction.<br />

Other halftone color prediction models are a continuing<br />

subject <strong>of</strong> past and current investigations. Emmel, 18,19 <strong>for</strong><br />

example, presents a recent survey and introduces a novel<br />

mathematical framework <strong>for</strong> spectral predictions <strong>of</strong> color<br />

halftone prints. The framework uses a global analytical approach<br />

based on matrix algebra that unifies most <strong>of</strong> the<br />

classical color prediction models. It is an efficient and intuitive<br />

model employing overall probabilities <strong>of</strong> photons entering<br />

and emerging through particular inking levels. However,<br />

the probabilistic description 20 used is “taken throughout the<br />

full sample area,” 19 which leads to a halftone-independent<br />

model. This makes it necessary to recalibrate the model’s<br />

coefficients <strong>for</strong> every different halftone technique used <strong>for</strong><br />

printing.<br />

As an alternative to Emmel’s approach, our proposed<br />

halftone prediction model 7 is particularly intended to meet<br />

the requirements <strong>of</strong> a halftone-dependent characterization <strong>of</strong><br />

digital printing devices. In order to be adaptable to arbitrary<br />

Figure 1. Diagram <strong>of</strong> an upper paper section with a light path scattered<br />

between the entry point x,y and the exit point x,y.<br />

halftone schemes we chose a numerical convolution approach<br />

using a separate optical modulation transfer function<br />

(MTF). The MTF model is founded on a three-dimensional<br />

extension <strong>of</strong> the Kubelka-Munk approach derived by analyzing<br />

multivariate partial differential equations with common<br />

computer algebra systems. 21 The derived extension approximates<br />

the scattered lateral light within semi-isotropic substrates.<br />

Although the one-dimensional Kubelka-Munk<br />

theory has methodological weaknesses, 17,22 the inaccuracy <strong>of</strong><br />

the predictions <strong>for</strong> the applications considered is limited.<br />

Moreover, it can be expected that the three-dimensional extension<br />

also increases the prediction accuracy. Like the underlying<br />

theory, the current approach relates the light propagation<br />

characteristics to a few substrate-dependent scattering<br />

and absorption coefficients. Specular and internal reflections<br />

at the interfaces and transmittances are considered as<br />

boundary conditions. Furthermore, the scattering concept<br />

can easily be extended to brightened fluorescent media as<br />

shown in Ref. 7.<br />

MTF-BASED SPATIAL-SPECTRAL HALFTONE<br />

PREDICTION MODEL<br />

In this section we describe the model used <strong>for</strong> light propagation<br />

in printed paper without elaborating the mathematical<br />

derivation. We are especially interested in an optical halftone<br />

model as a function <strong>of</strong> several easy predictable<br />

parameters such as scattering or partial reflection coefficients.<br />

The basis <strong>of</strong> the proposed approach 7 is to relate the<br />

reflected (reemitted) image and the local impact <strong>of</strong> the<br />

spread light to what is known as the point spread function<br />

(PSF) <strong>of</strong> paper. 10,23 The PSF models the scattered light intensity<br />

by the probability density h R x−x,y−y <strong>of</strong> a photon<br />

that enters the substrate at location x,y to exit at<br />

x,y; see Figure 1. Let x,y be the inner transmittance <strong>of</strong><br />

the print layer at location x,y, wherex,y= ink if<br />

x,y is covered by ink and x,y=1 otherwise, then the<br />

light reflected at point x,y <strong>of</strong> a halftone print is given by 23<br />

Rx,y = x,yh R x − x,y − yx,ydxdy.<br />

For simplicity, we ignore <strong>for</strong> the moment any specular or<br />

partial internal reflections at the paper interfaces. Usually,<br />

the computational ef<strong>for</strong>t <strong>of</strong> calculating the convolution integral<br />

is reduced by applying a two-dimensional Fourier trans<strong>for</strong>m<br />

to Eq. (1). Application <strong>of</strong> the commonly known convolution<br />

theorem replaces the integral operation by simple<br />

multiplication in the Fourier domain, yielding<br />

1<br />

284 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

Figure 2. Scheme <strong>of</strong> the paper bulk with the considered inner partial<br />

reflections and the illuminating flux i 0 .<br />

Rx,y = x,yF −1 †H R ,Fx,y‡,<br />

where F denotes the Fourier trans<strong>for</strong>m and F −1 its inverse<br />

[see the Appendix section and Bracewell 24 <strong>for</strong> details]. The<br />

variables (,) are the lateral frequencies in the Fourier domain<br />

and H R , is the MTF <strong>of</strong> paper, i.e., the Fourier<br />

trans<strong>for</strong>m <strong>of</strong> the PSF. Fx,y is the Fourier trans<strong>for</strong>m <strong>of</strong><br />

the inner transmittance <strong>of</strong> the print layer.<br />

We now extend Eq. (2) <strong>for</strong> practical purposes and incorporate<br />

additional surface refractive corrections in a similar<br />

way to Saunderson. 25 To begin with, we first supplement<br />

the specularly reflected fraction ap x,y at the upper side<br />

ink-air interface. (We use the following subscripts: ap: air to<br />

paper, pa: paper to air, and pb: paper to backing). Secondly,<br />

we distinguish between the incident transmitted fraction<br />

ap x,y and the emerging fraction pa x,y because <strong>of</strong> different<br />

illuminating and viewing geometries, yielding<br />

Rx,y = ap x,y + pa x,yF −1 †H R ,F ap x,y‡.<br />

In Eq. (3), the multiplication with the outward transmittance<br />

pa x,y represents the passage <strong>of</strong> the emerged light<br />

through the ink layer be<strong>for</strong>e being captured by a sensor. On<br />

the other hand, the multiplication <strong>of</strong> the MTF with the Fourier<br />

trans<strong>for</strong>med inward transmittance F ap x,y accounts<br />

<strong>for</strong> the spreading effect induced by the scattered light. Its<br />

propagation through the inner paper bulk is constrained at<br />

the bottom by the inner partial reflection pb and at the top<br />

interfaces by pa ; see Figure 2. We distinguish between both<br />

reflectances because the scattering bulk is usually faced by<br />

different media during measurements, as will be seen later.<br />

Note that Eq. (3) considers the halftone structures only<br />

through both inner transmittance factors ap x,y and<br />

pa x,y, i.e., we assume the MTF H R , to be independent<br />

from the local halftone structure. Consequently, we<br />

consider the situation depicted in Fig. 2, and the inner partial<br />

reflectances pa and pb are chosen halftone independently.<br />

Finally, with respect to multi-ink prints, an additional<br />

model is required <strong>for</strong> the calculation <strong>of</strong> the spectral inner<br />

transmittance <strong>of</strong> the overprinted ink layers. As such, Beer-<br />

Lambert’s multiplication <strong>of</strong> transmittances is a frequently<br />

chosen and simple approach. 26 More complex alternatives<br />

may be considered when dealing with fluorescent inks 19 or<br />

with ink penetration effects. 27<br />

In microscopic image analyses, Eq. (3) proves very useful.<br />

In this section we outline how to determine a mathematical<br />

expression <strong>of</strong> the MTF H R ,. The bulk <strong>of</strong> papers<br />

usually consists <strong>of</strong> a fiber network which deflects<br />

passing light rays in arbitrary directions. 28 This behavior is<br />

2<br />

3<br />

responsible <strong>for</strong> the scattering properties <strong>of</strong> the paper and,<br />

there<strong>for</strong>e, <strong>for</strong> its MTF. As mentioned earlier, the phenomenological<br />

Kubelka-Munk (KM) approach 3,29 is successfully<br />

used in application fields involving familiar light propagation<br />

problems. However, the KM approach considers only<br />

two diffuse radiant fluxes, one in the direction <strong>of</strong> incidence<br />

and the other in the opposite direction. In other words, the<br />

spatial distribution <strong>of</strong> the re-emitted light cannot be accounted<br />

<strong>for</strong> by using the results <strong>of</strong> the traditional KM<br />

theory. There<strong>for</strong>e, we approached the MTF by extending the<br />

KM theory to a three-dimensional space 7 in a similar way to<br />

the two-dimensional extension proposed by Berg. 30<br />

For an infinitesimal volume cube <strong>of</strong> the substrate, six<br />

light fluxes along and opposed to the coordinate axes x, y,<br />

and z are considered. The fluxes are non-negative by definition<br />

and are specified <strong>for</strong> −x,y and 0zD,<br />

where D is the thickness <strong>of</strong> the paper sheet. The study is<br />

confined to temporal steady state analyses <strong>of</strong> samples with a<br />

given absorption 0 that are illuminated with a light<br />

source <strong>of</strong> finite power. Hence the fluxes are expected to decay<br />

greatly in both lateral directions as x +y →. The<br />

correctness <strong>of</strong> this assumption was already shown by the<br />

measurements <strong>of</strong> Yule and others see, e.g., Ref. 10. Now, let<br />

us consider the intensity <strong>of</strong> the downward incident light flux<br />

along the vertical propagation direction z and call it i. The<br />

theory <strong>of</strong> KM is based on the assumption that the fractional<br />

amount <strong>of</strong> light lost by absorption between z and z+dz is<br />

given by dz, where the absorption density corresponds to<br />

the absorption coefficient K <strong>of</strong> the original KM theory. Also<br />

scattering decreases the considered intensity but, in our case,<br />

we consider different kinds <strong>of</strong> scattering densities: b and l ,<br />

where b is the back scattered intensity and l is the intensity<br />

scattered laterally to the initial direction. Accordingly,<br />

after passing dz the flux i is reduced by dz, where the<br />

coefficients obey<br />

= +4 l + b .<br />

We call the modeled scattering behavior semi-isotropic,<br />

where isotropic stands <strong>for</strong> the symmetry related to x,y,z and<br />

semi indicates l b . The absorbed light is lost to the system,<br />

but back-scattered light from the upward propagating<br />

flux is added to i. A quarter <strong>of</strong> the laterally scattered light<br />

from each lateral flux is also added to i. Applying the same<br />

reasoning to each <strong>of</strong> the remaining five fluxes we obtained a<br />

system <strong>of</strong> six coupled linear partial differential equations 7<br />

(PDE) that are familiar to the type <strong>of</strong> the seven-flux model<br />

<strong>of</strong> Yoon et al. 31 [Remark: According to the design <strong>of</strong> these<br />

PDEs, the incident light flux primarily propagates along the<br />

coordinate axes after an initial scattering event. However, in<br />

our applications, the resulting rotational asymmetry reaches<br />

numerically at most about one percent. 7 We disregard this<br />

asymmetry in favor <strong>of</strong> the obvious application advantages<br />

presented below].<br />

In order to derive the paper’s MTF, we used the generalized<br />

two-dimensional Fourier trans<strong>for</strong>m. This allows the<br />

PDE system to be reduced to a pair <strong>of</strong> equations similar to<br />

the original ordinary KM differential equations in z but with<br />

4<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 285


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

added dependencies on the spatial frequencies and . 7 For<br />

the PDE system at hand, the generalized two-dimensional<br />

Fourier trans<strong>for</strong>m is appropriate, since the fluxes are defined<br />

on the real plane R 2 and decay greatly as x +y →. 24,32,33<br />

Solving the system <strong>of</strong> trans<strong>for</strong>med PDEs at z=D yields the<br />

spectral reflectance MTF<br />

where<br />

H R , = Fh R x,y = A <br />

B <br />

,<br />

A = a 12 + a 21 − c pb e cD − a 12 + a 21 + c pb e −cD ,<br />

5.1<br />

5.2<br />

B =−a 21 + c + a 12 pa + pb + a 21 − c pa pb e cD + a 21<br />

− c + a 12 pa + pb + a 21 + c pa pb e −cD . 5.3<br />

The coefficients a 12 =a 12 , a 21 =a 21 , and c=c depend on<br />

the lateral frequencies , and are given by Eqs.<br />

(6.1)–(6.6):<br />

where<br />

2 2<br />

c =a 21 − a 12 , 6.1<br />

a 12 =<br />

a 21 =<br />

s 3<br />

s 1<br />

,<br />

s 2<br />

s 1<br />

,<br />

6.2<br />

6.3<br />

s 1 = − b 2 −2 l + b +2 l + b +4 2 2 − b 2 <br />

2 + 2 +16 4 2 2 ,<br />

s 2 = − b 2 −2 l + b +2 l + b −4 l 2 <br />

+4 2 − b + b −2 l 2 2 + 2 <br />

6.4<br />

+16 4 2 2 , 6.5<br />

s 3 = − b 2 −2 l + b b +2 l + b −4 l 2 <br />

+4 2 − b b + b −2 l 2 2 + 2 <br />

+16 4 b 2 2 . 6.6<br />

Likewise, the microspectral transmittance distribution is<br />

modeled by<br />

Tx,y = pa x,yF −1 †H T ,F bp x,y‡,<br />

with the spectral transmittance MTF H T ,<br />

7<br />

H T , = Fh T x,y =− 2c <br />

B <br />

.<br />

Here, bp x,y describes the fraction <strong>of</strong> light transmitted into<br />

the substrate through the bottom layer. Equation (7) implies<br />

that the optical spreading has no direct observable effect on<br />

transmittance measurements <strong>of</strong> single-side, upward-oriented<br />

printed paper sheets. This is consistent with microscopic<br />

transmittance images published by Koopipat et al. 13<br />

The two spatial <strong>for</strong>mulas, Eqs. (3) and (7), are the foundation<br />

used in predicting the spectral reflectance <strong>of</strong> arbitrary<br />

halftone prints presented and discussed in the Model Application<br />

and Model Discussion sections below. The following<br />

section considers the calibration <strong>of</strong> the optical parameters.<br />

MODEL CALIBRATION<br />

Our next objective is to determine the optical dot gain <strong>for</strong><br />

arbitrary halftones and dithering frequencies from a few<br />

macroscopic spectral measurements. More precisely, we need<br />

to calculate the isolated optical dot gain in order to distinguish<br />

between the different effects leading to printing<br />

nonlinearities. In order to meet this requirement, Eq. (3)<br />

describes the spectral reflectance Rx,y as a function <strong>of</strong> the<br />

bulk parameters D, , l , and b together with the surface<br />

refractive coefficients ap , ap , pa , pa , and pb . Un<strong>for</strong>tunately,<br />

only the paper thickness, D, is directly measurable<br />

with common instruments. There<strong>for</strong>e, we determine the remaining<br />

parameters in such a way as to best match the calculated<br />

results to the measured spectra <strong>of</strong> a small set <strong>of</strong> test<br />

patches. In order to avoid any printing irregularity and to<br />

increase the accuracy <strong>of</strong> the parameter estimation, the test<br />

patches are chosen to be as unambiguous as possible. In<br />

particular, we chose only solid patches <strong>of</strong> the primary<br />

colors—in our case cyan, magenta, yellow, and black prints,<br />

abbreviated to CMYK—plus a sample <strong>of</strong> paper-white (W).<br />

For this work, we consider only single side prints. As well as<br />

avoiding the printing irregularities, the uni<strong>for</strong>mity <strong>of</strong> the<br />

solid patches also reduces the matrix multiplication <strong>of</strong> Eq.<br />

(3) to a simple scalar multiplication. This is because<br />

Fx,y<strong>of</strong> a uni<strong>for</strong>m patch is only different from zero at<br />

zero frequencies ,0,0. Hence, <strong>for</strong> a solid patch,<br />

Eqs. (3) and (7) reduce to<br />

R solid = ap + pa H R 0,0 ap ,<br />

T solid = pa H T 0,0 bp .<br />

8<br />

9<br />

10<br />

In other words, the <strong>for</strong>m <strong>of</strong> the PSF h R x,y is not involved<br />

in the calibration process.<br />

Traditionally, the KM theory scattering and absorption<br />

coefficients K and S are determined using two distinct reflectance<br />

measurements: one measurement over a black<br />

backing and another over a white backing, both backings <strong>of</strong><br />

known reflectances. However, in our case, more measurements<br />

are required because we need to determine not only<br />

the scattering and absorption coefficients but also the inner<br />

transmittance <strong>of</strong> the primary inks used in addition to the<br />

286 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

Figure 3. Considered configurations <strong>of</strong> the calibration measurements <strong>of</strong> a paper sheet printed on one side<br />

with a solid patch. Upper left: Spectral reflectance on a known black backing. Upper right: Spectral reflectance<br />

on a known white backing. Lower left: Regular spectral transmittance. Lower right: Flipped spectral<br />

transmittance.<br />

refractive coefficients. There<strong>for</strong>e, we propose measuring the<br />

spectral reflectance and transmittance <strong>of</strong> each <strong>of</strong> the five<br />

solid test set patches (CMYK/W) in each <strong>of</strong> the four configurations<br />

depicted in Figure 3. These comprise the usual<br />

reflectance with a black and a white backing <strong>of</strong> known<br />

reflectances R b in addition to the transmittance <strong>of</strong> the<br />

samples in both a regular configuration and a flipped-over<br />

configuration yielding a total <strong>of</strong> nineteen different spectra.<br />

The spectral measurements were carried out using a<br />

standard R/T spectrophotometer (Gretag-Macbeth<br />

SpectroScan T). Its geometry <strong>of</strong> illumination (45°, collimated<br />

annular) and <strong>of</strong> viewing (0°) contradicts, strictly spoken,<br />

the general KM assumption <strong>of</strong> diffuse illumination and<br />

measurement. Nevertheless, according to Kubelka, 34 as we<br />

examine highly scattering paper substrates, we expect the<br />

geometric discrepancies at most to scale the determined values<br />

<strong>of</strong> the scattering and absorption coefficients by a constant<br />

factor. In our application we disregard its effect as long<br />

as the geometry <strong>of</strong> the measuring equipment is kept unchanged<br />

between calibration and prediction. The geometry<br />

<strong>of</strong> each measurement configuration also affects the assumed<br />

surface refractive corrections that are mutually connected by<br />

functional relations. We make these relations explicit by<br />

first-order approximations and introduce additional common<br />

measurement coefficients. 35 The considered surface refractive<br />

coefficients are illustrated in Figure 4. For simplicity,<br />

we assume the individual inks to have equal refractive indices,<br />

which may, however, deviate from the bulk refractive<br />

index <strong>of</strong> the unprinted paper. <strong>Additional</strong>ly, we disregard any<br />

wavelength dependency <strong>of</strong> the refractive indices. In order to<br />

list the refractive coefficients <strong>for</strong> each <strong>of</strong> the printed and<br />

unprinted cases, we hereafter identify each coefficient with<br />

one <strong>of</strong> the subscripts I or , respectively. (With regard to<br />

halftone patches, we vary these coefficients in accordance<br />

with the ink coverage).<br />

We begin with the situation <strong>of</strong> measuring the reflectance<br />

<strong>of</strong> the samples and consider the recorded fraction ap<br />

<strong>of</strong> the specular reflection s :<br />

apI = K s sI , ap = K s s , 11<br />

where K s amounts to the fraction captured by the instrument’s<br />

field <strong>of</strong> view. 36 Next, the incident transmission ap is<br />

determined by the transmitted fraction s =1− s in addition<br />

to what is known as the internal transmittance <strong>of</strong> the upper<br />

side layer [Ref. 26 p. 30] thus we approximate<br />

apI = 1− sI , ap =1− s . 12<br />

Similarly, we approximate the coefficient <strong>for</strong> the outward<br />

transmittance through the upper interface<br />

paI = 1− iI , pa =1− i . 13<br />

In Eq. (13), 1− i represents the fraction <strong>of</strong> the light transmitted<br />

diffusely from inside the scattering substrate. It thus<br />

differs from the fraction 1− s used in Eq. (12), which accounts<br />

<strong>for</strong> the collimated incidence. Accordingly, the amount<br />

<strong>of</strong> internal light reflected diffusely at the upper interface is<br />

Figure 4. Refraction and transmission coefficients at the interfaces <strong>of</strong> the<br />

paper sheet.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 287


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

Figure 5. Spectral calibration per<strong>for</strong>mance <strong>for</strong> the nonbrightened double-calandered APCO paper. The solid<br />

lines plot the spectral predictions and the crosses mark the corresponding spectral measurements. The paper<br />

has a thickness <strong>of</strong> about 99 m.<br />

paI = iI , pa = i . 14<br />

We continue at the bottom <strong>of</strong> the paper and consider the<br />

internal fraction <strong>of</strong> light reflected diffusely at the bottom<br />

interface. Similar to Eq. (14), it is determined by the internal<br />

reflection i plus the transmitted part through the bottom<br />

layer that is reflected back into the substrate at the backing<br />

pbI = iI + 1− iI 1− sI 2 R b ,<br />

pb = i + 1− i 1− s R b .<br />

15<br />

b = l = .<br />

17<br />

As already pointed out, we estimate the phenomenological<br />

model parameters as the set <strong>of</strong> values which minimizes<br />

the deviation between the nineteen measured spectra<br />

and their calculated predictions. We obtain the parameter<br />

estimates by using common least square minimization routines<br />

such as the lsqcurvefit <strong>of</strong> MATLAB. This routine<br />

solves nonlinear data-fitting problems in the least square<br />

sense. 37 In our case, the routine finds the parameters X<br />

sought that minimize the error to the measured spectra<br />

S meas <br />

Here, is squared because <strong>of</strong> the double light passage<br />

according to Beer-Lambert’s law. 26 Finally, <strong>for</strong> the case <strong>of</strong><br />

measuring the transmittance <strong>of</strong> the samples, the part transmitted<br />

from beneath the sample into the scattering substrate<br />

is approximated to<br />

1<br />

min<br />

X 2 S solidX − S meas 2 2 = 1 R solidi X<br />

2 i<br />

− R measi 2 + 1 T solidj X − T measj 2 ,<br />

2 j<br />

18<br />

bpI = 1− sI , bp =1− s . 16<br />

These relations reduce the multitude <strong>of</strong> refractive model parameters<br />

to K s , s , i <strong>for</strong> the printed and unprinted cases in<br />

addition to <strong>of</strong> each primary ink. Together with the<br />

scattering and absorption coefficients, six scalar coefficients<br />

and seven spectral coefficients are obtained that may be estimated<br />

using the nineteen available spectral measurements.<br />

However, in order to simplify the optimization process further,<br />

we limit the model to the isotropic case assuming both<br />

scattering coefficients to be equal, hence<br />

where R solid X and T solid X are the spectral reflectance<br />

and transmittance <strong>of</strong> the samples calculated with the<br />

parameters fixed to X according to Eqs. (9) and (10), respectively.<br />

The routine used finds the coefficients so that the<br />

solution is always bounded within an appropriately chosen<br />

range <strong>for</strong> each parameter <strong>of</strong> X.<br />

CALIBRATION RESULTS<br />

An example <strong>of</strong> the obtained calibration results is illustrated<br />

in Figure 5. The depicted spectra were obtained from a rep-<br />

288 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

Figure 6. Fitted estimates <strong>of</strong> the spectral absorption and scattering coefficients<br />

underlying the data <strong>of</strong> Fig. 5.<br />

Figure 8. From bottom to top reflects the dependency <strong>of</strong> the optical dot<br />

gain <strong>of</strong> the APCO paper on increasing screen frequencies. The simulated<br />

halftone technique is a conventional, circular dot screen illustrated below<br />

the chart.<br />

al. 40 <strong>for</strong> a critical review. Originally, the MD model was proposed<br />

in a densitometric <strong>for</strong>m due to limitations <strong>of</strong> the instruments.<br />

We use its spectral <strong>for</strong>m, which reads<br />

Figure 7. Fitted estimates <strong>of</strong> the spectral internal transmittances C ,<br />

M , Y , and K underlying the data <strong>of</strong> Fig. 5.<br />

resentative calibration <strong>of</strong> a non-brightened doublecalandered<br />

APCO paper 38 printed by a common, four-color<br />

xerographic desktop printer. The average calibration accuracy<br />

achieved 1E * 94 , which is in the order <strong>of</strong> the colorimetric<br />

discrimination per<strong>for</strong>mance <strong>of</strong> the human eye. Figure<br />

6 shows the fitted spectral model coefficients <strong>of</strong> absorption<br />

and scattering obtained. In this experiment, the estimates<br />

<strong>for</strong> the inner reflectance and transmittance coefficients<br />

<strong>of</strong> the printed interfaces are i =10%, s =1.25% and<br />

K s =35%. These scalar values, and particularly i differ from<br />

usual reports in related research. 36,39 In our opinion, this<br />

discrepency is mainly due to the different underlying approaches<br />

<strong>of</strong> accounting <strong>for</strong> the interactions between the partial<br />

reflections and the scattered differential light fluxes. Finally,<br />

Figure 7 plots the estimated internal transmittances<br />

<strong>of</strong> the separate CMYK inks.<br />

MODEL APPLICATIONS<br />

Given a calibrated parameter set <strong>of</strong> a particular paper-ink<br />

combination, such as those derived in the previous section,<br />

the spectral prediction Eqs. (3) and (7) allow the expected<br />

microimages <strong>of</strong> arbitrary halftone prints on that paper to be<br />

simulated. In this section, after recalling the basics <strong>of</strong> dot<br />

gain calculations, we apply the proposed approach and analyze<br />

optical dot gain predictions <strong>for</strong> idealized halftone prints.<br />

Common dot gain calculations <strong>of</strong> arbitrary halftones are<br />

based on the model <strong>of</strong> Murray-Davies (MD); see Wyble et<br />

R = 1−a t R g + a t R t ,<br />

19<br />

where R is the predicted reflectance spectrum, R g is<br />

the reflectance spectrum <strong>of</strong> the bare substrate and R t is<br />

the spectral reflectance spectrum <strong>of</strong> the color at full area<br />

coverage. The linear interpolation variable a t commonly refers<br />

to the theoretical area coverage <strong>of</strong> the predicted halftone<br />

print, i.e., the dot area <strong>of</strong> the binary image actually sent to<br />

the printer. Usually, the MD model overestimates the measured<br />

spectral reflectances R meas since it disregards the<br />

combined optical and mechanical dot gain effects as introducedintheBackground<br />

section. The overestimation gives<br />

rise to introduction <strong>of</strong> the effectiveareacoveragea eff , which<br />

refers to an estimated value that best fits the calculated reflectance<br />

R to the measured spectrum R meas . For a<br />

single wavelength, typically chosen at minimum reflectance,<br />

a eff is given by using Eq. (19)<br />

a eff = R meas − R g<br />

R t − Rg , 20<br />

with suppressed wavelength notation <strong>for</strong> simplicity. This allows<br />

the dot gain to be defined as<br />

= a eff − a t .<br />

21<br />

Use <strong>of</strong> Eq. (3) allows estimation <strong>of</strong> the optical dot-gain<br />

o <strong>of</strong> any given binary halftone image Ix,y with a dot area<br />

equal to a s . In this case o is given by<br />

o = a eff − a s = Rx,y − R g<br />

R t − R g<br />

− a s . 22<br />

In Eq. (22), Rx,y is the spatial average <strong>of</strong> the reflectance<br />

image predicted using Eq. (3) and replaces the measured<br />

reflectance R meas <strong>of</strong> Eq. (20). An illustrative example is seen<br />

in Figure 8 which depicts optical dot-gain predictions <strong>for</strong> the<br />

APCO paper 38 against the area coverage a s <strong>for</strong> various screen<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 289


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

Figure 9. Calculated optical dot gain curves at medium area coverage<br />

against the screen ruling with respect to two typically used halftone<br />

screens. The plots were predicted using the data <strong>of</strong> the APCO paper.<br />

Figure 11. Comparison <strong>of</strong> our LSF-predictions with normalized measurement<br />

points <strong>of</strong> a different but apparently similar paper type published by<br />

Arney et al. 5 The dots depict Arney’s measurements. The dashes and the<br />

solid line plot the obtained predictions <strong>for</strong> the ColorCopy and the APCO<br />

paper, respectively.<br />

particular the impact <strong>of</strong> the optical dot-gain’s variations<br />

around the commonly used ruling <strong>of</strong> 60 lpcm 150 lpi.<br />

In this region, any slight shift <strong>of</strong> the screen ruling obviously<br />

induces a non-negligible error arising solely from the optical<br />

dot gain.<br />

Figure 10. Calculated optical dot gain curves <strong>for</strong> two different types <strong>of</strong><br />

<strong>of</strong>fice paper plotted against screen ruling. The conventional circular dot<br />

screen was used. The dashes and the solid line plot the obtained predictions<br />

<strong>for</strong> the ColorCopy paper and the APCO paper, respectively.<br />

frequencies (so-called ruling). The type and magnitude <strong>of</strong><br />

the optical dot gain curves shown are close to analyses presented,<br />

e.g., by Gustavson 8 or by Yang [Ref. 41 p. 201], which<br />

apply a simple Gaussian-like function to approach the shape<br />

<strong>of</strong> the paper PSF.<br />

Clearly, <strong>for</strong> calculating the halftone reflectances, the<br />

simplifying assumption underlying Eq. (9) no longer holds<br />

true and the coefficients become spatially dependent. Hence<br />

we numerically evaluate the spatial <strong>for</strong>mula, Eq. (3), by using<br />

the standard inverse two-dimensional fast Fourier trans<strong>for</strong>m<br />

(iFFT). 24 Thereby we sample the coefficients <strong>of</strong> concern<br />

at spatial high resolution frequency grids , with a<br />

sampling frequency exceeding the Nyquist rate <strong>of</strong> the halftone<br />

structures, i.e., the coefficients are filled out at each<br />

x i ,y j according to the simulated binary image Ix i ,y j with<br />

the calibrated values <strong>of</strong> section Calibration Results.<br />

Another application example is given in Figure 9, which<br />

illustrates the influence <strong>of</strong> different types <strong>of</strong> halftone screens<br />

on the magnitude <strong>of</strong> the optical dot gain at medium area<br />

coverage <strong>of</strong> a s =50%. With the proposed simulation, it is<br />

also convenient to explore the effect <strong>of</strong> changing the type <strong>of</strong><br />

paper on the optical dot gain. For the case <strong>of</strong> the conventional<br />

circular dot halftone screen, Figure 10 shows the optical<br />

dot gain prediction <strong>for</strong> the APCO paper compared with<br />

results obtained with a similarly calibrated parameter set <strong>for</strong><br />

the ColorCopy <strong>of</strong>fice paper. 42 These graphs demonstrate in<br />

MODEL COMPARISON AND DISCUSSION<br />

To explore the per<strong>for</strong>mance <strong>of</strong> our microscopic predictions,<br />

we consider results <strong>of</strong> edge-trace measurements published by<br />

Arney et al. 5 For this purpose, we compare our prediction<br />

results with their published data <strong>of</strong> a measured line spread<br />

function (LSF) and the corresponding MTF <strong>of</strong> an <strong>of</strong>fice copy<br />

paper <strong>of</strong> a similar type to those analyzed in our laboratory.<br />

Figure 11 depicts a direct comparison <strong>of</strong> Arney’s measurements<br />

and our prediction <strong>of</strong> a microscopic line edge observation<br />

derived by using Eq. (3). The comparison <strong>of</strong> the<br />

MTF, the normalized modulus <strong>of</strong> the digital Fourier trans<strong>for</strong>m<br />

<strong>of</strong> the same data is illustrated in Figure 12. Both figures<br />

demonstrate good agreement between our predictions and<br />

the microdensitometric measurements. Note that the model<br />

parameters were calibrated using spectral measurements <strong>of</strong><br />

solid patches only.<br />

A further comparison arises from Arney’s discussion <strong>of</strong><br />

the influence <strong>of</strong> Kubelka-Munk’s absorption coefficient K on<br />

Oittinen and Engeldrum’s model. Oittinen, 2 as well as<br />

Engeldrum and Pridham, 1 suggested a simplified relationship<br />

between the Kubelka-Munk theory and the MTF <strong>of</strong><br />

paper. According to their approach, the MTF <strong>of</strong> paper approximates<br />

the vertical derivative <strong>of</strong> the classical KM<br />

function. 3 It depends on the thickness <strong>of</strong> the sample D, the<br />

scattering coefficient S, and the absorption coefficient K.<br />

However, with edge-trace measurements Arney et al. showed<br />

that K has only a small influence on the width <strong>of</strong> the MTF <strong>of</strong><br />

the paper compared to S and D. 4 Figure 13 illustrates this<br />

finding and compares the results <strong>of</strong> our model with those<br />

obtained by Oittinen and Engeldrum’s model. In particular,<br />

the figure plots the scalar measure k p proposed by Arney,<br />

which is equal to the inverse <strong>of</strong> the frequency at which the<br />

normalized MTF loses its half magnitude. The plot demon-<br />

290 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

our model shows good agreement. Moreover, the calibrated<br />

model agrees well with the experimental insight <strong>of</strong> Arney et<br />

al. 4 that the optical absorption coefficient <strong>of</strong> paper has an<br />

insignificant effect on the MTF width <strong>of</strong> paper.<br />

Figure 12. Comparison <strong>of</strong> the modulus <strong>of</strong> the digital Fourier trans<strong>for</strong>m <strong>of</strong><br />

the data presented in Fig. 11.<br />

ACKNOWLEDGMENTS<br />

The author would like to thank Pr<strong>of</strong>essor R. D. Hersch,<br />

École Polytechnique Fédérale de Lausanne (EPFL), Switzerland,<br />

<strong>for</strong> the continuing and supporting discussions. The<br />

constructive discussions with K. Simon, M. Vöge, and P.<br />

Zolliker are also gratefully acknowledged. Part <strong>of</strong> the investigation<br />

was financed by the Swiss Innovation Promotion<br />

Agency (grant KTI/CTI 6498.1 ENS-ET).<br />

Appendix<br />

We used the following <strong>for</strong>m <strong>of</strong> the two-dimensional Fourier<br />

trans<strong>for</strong>m F, <strong>of</strong> a two-dimensional function fx,y<br />

<br />

F, fx,ye =−<br />

−2ix+y dx dy,<br />

23<br />

and the inverse two-dimensional Fourier trans<strong>for</strong>m<br />

<br />

fx,y F,e =−<br />

2ix+y d d,<br />

24<br />

Figure 13. Comparing the involvement <strong>of</strong> the absorption coefficient K <strong>of</strong><br />

Oittinen and Engeldrum’s model with that <strong>of</strong> the absorption coefficient <br />

<strong>of</strong> our model obtained <strong>for</strong> the ColorCopy paper. The comparison is carried<br />

out using Arney’s data and his proposed k p scalar metric.<br />

strates the involvement <strong>of</strong> Oittinen and Engeldrum’s absorption<br />

coefficient K shown by Arney 5 and the more realistic<br />

influence <strong>of</strong> our absorption coefficient .<br />

CONCLUSIONS<br />

Our prediction model <strong>of</strong>fers a new way <strong>of</strong> minimizing the<br />

ef<strong>for</strong>t <strong>of</strong> characterizing and calibrating color halftone printers<br />

and extends the computational framework <strong>of</strong> controlling<br />

color printers online. The model used computes high resolution<br />

spectral color images <strong>of</strong> arbitrary halftone prints on<br />

common <strong>of</strong>fice paper and newsprint types <strong>of</strong> paper. Light<br />

scattering effects are accounted <strong>for</strong> by relating the appearance<br />

characteristics to a few substrate-dependent fitting coefficients.<br />

The main advantage <strong>of</strong> the approach is to uncouple<br />

the calibration <strong>of</strong> the scattering and absorption<br />

coefficients <strong>of</strong> the paper from printing irregularities without<br />

using a microscopic device. We base the calibration on reflectance<br />

measurements <strong>of</strong> solid patches because they are<br />

affected much less by the mechanical dot-gain than halftone<br />

patches are. Thereby, the calibration achieves a good independence<br />

from the printing process and requires merely a<br />

common spectrophotometer <strong>for</strong> reference measurements.<br />

Comparing published microspectral knife-edge measurements<br />

<strong>of</strong> Arney et al. 5 with corresponding simulations <strong>of</strong><br />

see Ref. 24 <strong>for</strong> details.<br />

REFERENCES<br />

1 P. G. Engeldrum and B. Pridham, “Application <strong>of</strong> turbid medium theory<br />

to paper spread function measurements”, Proc.-TAGA 1, 339–352<br />

(1995).<br />

2 P. Oittinen, “Limits <strong>of</strong> microscopic print quality”, in Advances in<br />

Printing <strong>Science</strong> and Technology (Pentech, London, 1982), Vol. 16, pp.<br />

121–138.<br />

3 P. Kubelka and F. Munk, “Ein Beitrag zur Optik der Farbanstriche”, Z.<br />

Tech. Phys. (Leipzig) 12, 593–601 (1931).<br />

4 J. S. Arney, E. Pray, and K. Ito, “Kubelka-Munk theory and the<br />

Yule-Nielsen effect on halftones”, J. <strong>Imaging</strong> Sci. Technol. 43, 365–370<br />

(1999).<br />

5 J. S. Arney, J. Chauvin, J. Nauman, and P. G. Anderson, “Kubelka-Munk<br />

theory and the MTF <strong>of</strong> paper”, J. <strong>Imaging</strong> Sci. Technol. 47, 339–345<br />

(2003).<br />

6 S. Mourad, P. Emmel, K. Simon, and R. D. Hersch, “Prediction <strong>of</strong><br />

monochrome reflectance spectra with an extended Kubelka-Munk<br />

model”, Proc. IS&T/SID Tenth Color <strong>Imaging</strong> Conference (IS&T,<br />

Springfield, VA, 2002) pp. 298–304.<br />

7 S. Mourad, Color Prediction <strong>for</strong> Electrophotographic Prints on Common<br />

Office Paper, Ph.D. Thesis, École Polytechnique Fédérale de Lausanne,<br />

Switzerland, 2003, http://diwww.epfl.ch/w3lsp/publications/colour.<br />

8 S. Gustavson, “Color gamut <strong>of</strong> halftone reproduction”, J. <strong>Imaging</strong> Sci.<br />

Technol. 41, 283–290 (1997).<br />

9 J. A. C. Yule and W. J. Nielsen, “The penetration <strong>of</strong> light into paper and<br />

its effect on halftone reproduction”, Proc.-TAGA, 65–76 (1951).<br />

10 J. A. C. Yule, D. J. Howe, and J. H. Altman, “The effect <strong>of</strong> the<br />

spread-function <strong>of</strong> paper on halftone reproduction”, Tappi J. 50,<br />

337–344 (1967).<br />

11 S. Inoue, N. Tsumura, and Y. Miyake, “Analyzing CTF <strong>of</strong> print by MTF<br />

<strong>of</strong> paper”, J. <strong>Imaging</strong> Sci. Technol. 42, 572–576 (1998).<br />

12 S. Gustavson, Dot Gain in Colour Halftones, Ph.D. Thesis, Dept. <strong>of</strong><br />

Electrical Engineering, Linköping University, Sweden, 1997.<br />

13 C. Koopipat, N. Tsumura, and Y. Miyake, “Effect <strong>of</strong> ink spread and<br />

optical dot gain on the MTF <strong>of</strong> ink jet image”, J. <strong>Imaging</strong> Sci. Technol.<br />

46, 321–325 (2002).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 291


Mourad: Improved calibration <strong>of</strong> optical characteristics <strong>of</strong> paper by an adapted paper-MTF model<br />

14 B. Philips-Invernizzi, D. Dupont, and C. Cazé, “Bibliographical review<br />

<strong>for</strong> reflectance <strong>of</strong> diffusing media”, Opt. Eng. (Bellingham) 40,<br />

1082–1092 (2001).<br />

15 A. Ishimaru, Wave Propagation and Scattering in Random Media<br />

(Academic Press, New York, 1978), Vols. I&II.<br />

16 A. Ishimaru, “Diffusion <strong>of</strong> light in turbid material”, Appl. Opt. 28,<br />

2210–2215 (1989).<br />

17 B. Hapke, “Kubelka-Munk theory: What’s wrong with it”, in Theory <strong>of</strong><br />

Reflectance and Emittance Spectroscopy (Cambridge University Press,<br />

Cambridge, UK, 1993), Chap. 11.<br />

18 P. Emmel, Modèles de prédiction couleur appliqués à l’impression jet<br />

d’encre, Ph.D. thesis, École Polytechnique Fédérale de Lausanne, 1998,<br />

http://diwww.epfl.ch/w3lsp/publications/colour.<br />

19 P. Emmel, “Physical models <strong>for</strong> color prediction”, in Digital Color<br />

<strong>Imaging</strong>, Handbook (The Electrical Engineering and Applied Signal<br />

Processing Series) (CRC Press, Boca Raton, FL, 2003), pp. 173–238.<br />

20 J. S. Arney, “A probability description <strong>of</strong> the Yule-Nielsen effect, I”, J.<br />

<strong>Imaging</strong> Sci. Technol. 41, 633–636 (1997).<br />

21 Wolfram Research, Inc. MATHEMATICA ® . http://www.wolfram.com.<br />

22 L. Yang and S. J. Miklavcic, “Revised Kubelka-Munk theory. III. A<br />

general theory <strong>of</strong> light propagation in scattering and absorptive media”,<br />

J. Opt. Soc. Am. A 22, 1866–1873 (2005).<br />

23 F. R. Ruckdeschel and O. G. Hauser, “Yule-Nielsen effect in printing: a<br />

physical analysis”, Appl. Opt. 17, 3376–3383 (1978).<br />

24 R. N. Bracewell, The Fourier Trans<strong>for</strong>m and its Applications, 3rd ed.<br />

(McGraw-Hill, New York, 2000).<br />

25 J. L. Saunderson, “Calculation <strong>of</strong> the color <strong>of</strong> pigmented plastics”, J. Opt.<br />

Soc. Am. 32, 727–736 (1942).<br />

26 G. Wyszecki and W. S. Stiles, Color <strong>Science</strong>, 2nd ed. (John Wiley & Sons,<br />

Inc., New York, 1982).<br />

27 L. Yang, R. Lenz, and B. Kruse, “Light scattering and ink penetration<br />

effectsontonereproduction”,J.Opt.Soc.Am.A18, 360–366 (2001).<br />

28 G. Kortüm, Reflectance Spectroscopy: Principles, Methods, Applications<br />

(Springer-Verlag, Berlin-Heidelberg, 1969).<br />

29 P. Kubelka, “New contributions to the optics <strong>of</strong> intensely light-scattering<br />

materials. Part II: Nonhomogenous layers”, J. Opt. Soc. Am. 44, 330–335<br />

(1954).<br />

30 F. Berg, Isotrope Lichtstreuung in Papier - Neue Überlegungen zur<br />

Kubelka-Munk-Theorie, Ph.D. Thesis, Technische Hochschule<br />

Darmstadt, 1997.<br />

31 G. Yoon, A. J. Welch, M. Motamedi, and M. C. J. Van Gemert,<br />

“Development and application <strong>of</strong> three-dimensional light distribution<br />

model <strong>for</strong> laser irradiated tissue”, IEEE J. Quantum Electron. 23,<br />

1721–1733 (1987).<br />

32 M. J. Lighthill, Introduction to Fourier Analysis and Generalised<br />

Functions (Cambridge University Press, Cambridge, UK, 1980).<br />

33 D. G. Duffy, Trans<strong>for</strong>m Methods <strong>for</strong> Solving Partial Differential<br />

Equations (CRC Press, Boca Raton, FL, 1994).<br />

34 P. Kubelka, “New contributions to the optics <strong>of</strong> intensely light-scattering<br />

materials Part. I”, J. Opt. Soc. Am. 38, 448–457 (1948).<br />

35 D. B. Judd and G. Wyszecki, Color in Business, <strong>Science</strong> and Industry, 3rd<br />

ed. (John Wiley & Sons, Inc., New York, 1975).<br />

36 F. R. Clapper and J. A. C. Yule, “The effect <strong>of</strong> multiple internal<br />

reflections on the densities <strong>of</strong> half-tone prints on paper”, J. Opt. Soc.<br />

Am. 43, 600–603 (1953).<br />

37 MathWorks. MATLAB Optimization Toolbox. Consider especially<br />

lsqcurvefit and fmincon, http://www.mathworks.com.<br />

38 ISO-2846-1. Graphic technology—Colour and transparency <strong>of</strong> ink sets <strong>for</strong><br />

four-colour-printing, 1st ed. (ISO, Geneva, 1997). The nonbrightened<br />

APCO II/II paper is specified in Annex A.<br />

39 D. B. Judd, “Fresnel reflection <strong>of</strong> diffusely incident light”, J. Res. Natl.<br />

Bur. Stand. 29, RP1504 (1942).<br />

40 D. R. Wyble and R. S. Berns, “A critical review <strong>of</strong> spectral models<br />

applied to binary color printing”, Color Res. Appl. 25, 4–19 (2000).<br />

41 L. Yang, S. Gooran, and B. Kruse, “Simulation <strong>of</strong> optical dot gain in<br />

multichromatic tone reproduction”, J. <strong>Imaging</strong> Sci. Technol. 45, 198–204<br />

(2001).<br />

42 Neusiedler, Color Copy paper. http://www.neusiedler.com(2000).<br />

Brightened, highly opaque <strong>of</strong>fice paper.<br />

292 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 293–298, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Gloss Granularity <strong>of</strong> Electrophotographic Prints<br />

J. S. Arney and Ling Ye<br />

Rochester Institute <strong>of</strong> Technology, Rochester, New York 14623<br />

E-mail: arney@cis.rit.edu<br />

Eric Maggard and Brian Renstrom<br />

Hewlett-Packard Co., Boise, Idaho 83714<br />

Abstract. The random variation in gloss <strong>of</strong>ten observed in images<br />

produced in electrophotographic printers has been examined by an<br />

analytical technique that combines the capabilities <strong>of</strong> a microdensitometer<br />

with a goniophotometer. The technique is called<br />

microgoniophotometry and measures both the spatial and the angular<br />

distribution <strong>of</strong> the specular component <strong>of</strong> reflected light. The<br />

analysis provides in<strong>for</strong>mation about the spatial variation <strong>of</strong><br />

specularly reflected light at all angles through which the specular<br />

light is reflected, not just at the equal/opposite angle at which gloss<br />

is traditionally measured. The results <strong>of</strong> this analysis have lead to an<br />

optical model <strong>of</strong> the random spatial variation in gloss. The results<br />

indicate that dry toner is typically not completely fused and can be<br />

described as a surface composed <strong>of</strong> two distinct regions. These two<br />

regions differ in the extent <strong>of</strong> fusing that has occurred, as manifested<br />

by their differences in specular reflectance characteristics.<br />

The difference in reflectance is manifested primarily in their different<br />

angular distributions <strong>of</strong> specular light and also in their spatial<br />

frequency. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4293<br />

INTRODUCTION<br />

A bidirectional reflectance distribution function (BRDF) is a<br />

useful way to characterize the angular distribution <strong>of</strong> specular<br />

light reflected from materials. 1–8 Moreover, one would<br />

expect the BRDF to be a necessary part <strong>of</strong> a complete instrumental<br />

characterization <strong>of</strong> visual attributes <strong>of</strong> gloss. 9 In<br />

addition to the angular distribution <strong>of</strong> the specular light, the<br />

spatial distribution <strong>of</strong> the specular light may also play a role<br />

in visual gloss. 10,11 As illustrated in Figure 1, gloss in electrophotographic<br />

prints is not always spatially uni<strong>for</strong>m. Indeed,<br />

spatial variations in gloss take many <strong>for</strong>ms. Artifacts<br />

such as streaking and banding are <strong>of</strong>ten observed in high<br />

gloss prints, and differential gloss involves differences in<br />

gloss between bordering regions <strong>of</strong> different color. The current<br />

report focuses on gloss granularity, which is the random<br />

gloss variation across a printed surface. Gloss granularity is<br />

illustrated in Fig. 1 with samples A and B showing different<br />

degrees <strong>of</strong> gloss granularity.<br />

Granularity analysis is an analytical technique that<br />

evolved during the 20th century to characterize silver halide<br />

photographic film. 12 The typical microdensitometer was an<br />

optical microscope with a fixed aperture and an electronic<br />

light detector. The film sample was scanned under the microscope<br />

and a trace <strong>of</strong> irradiance versus location was recorded.<br />

This technique is called microdensitometry. Currently,<br />

a microdensitometry scan may be per<strong>for</strong>med more<br />

easily by a s<strong>of</strong>tware routine applied to a digital image captured<br />

with a camera and appropriate microscope optics. 13,14<br />

Several reports have been published on the application <strong>of</strong><br />

microdensitometry techniques to the analysis <strong>of</strong> gloss<br />

granularity. 10,11,15 All <strong>of</strong> these techniques involve detection <strong>of</strong><br />

light at the specular angle (equal/opposite angle) while scanning<br />

across the surface <strong>of</strong> the sample. The current work<br />

extends this analytical technique to a measurement <strong>of</strong> the<br />

entire BRDF (goniophotometry) scanned spatially across the<br />

surface <strong>of</strong> a printed sample (microdensitometry). This analytical<br />

technique is called microgoniophotometry.<br />

THE MICROGONIOPHOTOMETER<br />

The microgoniophotometer has been described in detail in<br />

previous reports and is summarized in Figure 2. 1,2,16–18 The<br />

print sample is wrapped around a cylinder, and this presents<br />

all sample angles from −90° to +90° to the camera. The<br />

sample is illuminated with a linear light source placed at an<br />

angle <strong>of</strong> 20° from the camera. This places a bright specular<br />

line at the half angle, =10° between the camera and the<br />

source. Two images captured with this system are illustrated<br />

in Figure 3.<br />

As illustrated in Fig. 3, the specular component <strong>of</strong> the<br />

reflected light maintains its polarization and is observed only<br />

<br />

IS&T Member<br />

Received Jan. 3, 2007; accepted <strong>for</strong> publication Feb. 3, 2007.<br />

1062-3701/2007/514/293/6/$20.00.<br />

Figure 1. Examples <strong>of</strong> A rough and B smooth gloss granularity in<br />

electrophotographic prints produced by two different printers using different<br />

toners and fusing conditions.<br />

293


Arney et al.: Gloss granularity <strong>of</strong> electrophotographic prints<br />

Figure 5. BRDF <strong>of</strong> vs and BGDF <strong>of</strong> vs generated from Fig. 4.<br />

Curves are normalized to 1.00 at the peak value in order make a<br />

comparison.<br />

Figure 2. Schematic illustration <strong>of</strong> the microgoniophotometer. A linear<br />

polarizer is placed in front <strong>of</strong> the line light source, and another polarizer,<br />

called the analyzer, is in front <strong>of</strong> the camera.<br />

Figure 3. Images captured with the analyzer in front <strong>of</strong> the camera parallel<br />

to and perpendicular to the polarization direction <strong>of</strong> the light source<br />

polarizer.<br />

Figure 6. Micr<strong>of</strong>acets <strong>of</strong> the surface are randomly oriented at different tilt<br />

angles. If the facet tilt results in an equal/opposite angle between the<br />

camera and the light source, then light enters the camera. Otherwise the<br />

specular light misses the camera. A piece <strong>of</strong> shattered automobile window<br />

glass is a macroscopic illustration <strong>of</strong> bilevel gloss granularity.<br />

sample. A plot <strong>of</strong> the mean value versus tilt angle, vs , is<br />

a bidirectional reflectance distribution function, BRDF. A<br />

plot <strong>of</strong> the standard deviation versus tilt angle, vs , isa<br />

bidirectional granularity distribution function, BGDF. (See<br />

Figure 5.) It is the granularity <strong>of</strong> the specular light at each<br />

angle on the BRDF.<br />

Figure 4. The difference image A-B shows only the specularly reflected<br />

light. The mean, , and the standard deviation, , <strong>of</strong> the specular light is<br />

determined at each column in the image.<br />

in the image with parallel polarizers. Both the crossed and<br />

the parallel polarizers capture the same amount <strong>of</strong> diffuse,<br />

randomly polarized light. The difference image, (A-B) in<br />

Figure 4, shows only the specular light.<br />

The horizontal location <strong>of</strong> each column in the difference<br />

image (A-B) corresponds to a tilt angle, , on the print<br />

A FACET MODEL OF SPECULAR GRANULARITY<br />

Johansson, Béland, and MacGregor have introduced a model<br />

<strong>of</strong> specular reflection called the micr<strong>of</strong>acet model, 10,11 and<br />

the micr<strong>of</strong>acet model has been applied to the problem <strong>of</strong><br />

synthetic scene generation in computer graphics. 19 The<br />

micr<strong>of</strong>acet model assumes the surface that reflects the specular<br />

light can be described as a set <strong>of</strong> small facets, each at a<br />

randomly tilted angle, as illustrated schematically in<br />

Figure 6. The only facets that will deliver light to the camera<br />

are those facets tilted exactly to produces an equal/opposite<br />

294 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Arney et al.: Gloss granularity <strong>of</strong> electrophotographic prints<br />

Figure 7. Example <strong>for</strong> a sample <strong>of</strong> solid black toner printed by a typical<br />

electrophotographic printer. The solid line is Eq. 3, and the points are<br />

from experimental measurements <strong>of</strong> and 2 over the range −50° <br />

50°.<br />

angle between the source and the camera. Otherwise the<br />

light misses the camera. The result would be expected to be<br />

the bilevel image <strong>of</strong> specular glints, as illustrated in Fig. 6.<br />

The line light source used in the microgoniophotometer<br />

is assumed to be infinite in the direction colinear with the<br />

cylinder so that a facet tilt in the orthogonal direction, ,<br />

always directs light to the camera. There<strong>for</strong>e, the BRDF measured<br />

with the microgoniophotometer should be a direct<br />

measure <strong>of</strong> the random distribution <strong>of</strong> facet tilt angles in the<br />

direction. By normalizing the area under the BRDF, vs<br />

, to unity, the probability density function, P, <strong>for</strong> the<br />

random tilt angles, , can be <strong>for</strong>med as shown in Eqs. (1).<br />

The value <strong>of</strong> P at each angle, , is a measure <strong>of</strong> the fraction<br />

<strong>of</strong> the surface that contains facets at exactly angle :<br />

90<br />

K =−90<br />

d and P = <br />

K . 1<br />

Each facet that is at the correct specular angle delivers<br />

light at irradiance I to the camera. All other facets produce<br />

an irradiance <strong>of</strong> I=0. The result is irradiance I at the facet<br />

location projected onto the camera sensor plane. This bilevel<br />

set <strong>of</strong> facets should produce an average value and a standard<br />

deviation given by Eqs. (2) and (3). Note from Eq. (1) that<br />

the area under the BRDF ( vs ) is an experimental measure<br />

<strong>of</strong> the irradiance, I=K,<br />

= P · I, where I = K, 2<br />

2 = P · 1−P · I 2 .<br />

In order to test the facet model quantitatively, experimental<br />

measurements <strong>of</strong> 2 versus were carried out <strong>for</strong><br />

twenty samples <strong>of</strong> solid black (single toner) produced by<br />

different printers with different toners and different fusing<br />

conditions on different substrates. Values <strong>of</strong> P were calculated<br />

from with Eq. (1), and the data was plotted as 2<br />

versus P·1−P. Figure 7 is an example <strong>for</strong> a typical solid<br />

black toner printed by laser EP. The measured values <strong>of</strong> 2<br />

were much lower than predicted, and the data do not show<br />

the linearity <strong>of</strong> Eq. (3). Thus the facet model illustrated in<br />

Fig. 6 does not provide a complete, quantitative rationale <strong>for</strong><br />

the measured data.<br />

3<br />

Figure 8. The blurring effect <strong>of</strong> the camera pixels projected onto the<br />

surface facets.<br />

AN EXPANDED FACET MODEL<br />

It is not surprising that the experimentally measured values<br />

<strong>of</strong> 2 are lower than predicted. Equation (3) is based on the<br />

facets as if they were measured with infinite resolution.<br />

However, there is no reason to expect the surface facets to be<br />

large relative to the size <strong>of</strong> the camera pixels projected onto<br />

the surface. Indeed, if the camera pixels are larger than the<br />

facet size, the camera image will blur the image through a<br />

convolution with the effective aperture <strong>of</strong> the camera pixels.<br />

This is illustrated in Figure 8. The effect can be described<br />

quantitatively by modifying Eq. (3) with a blurring factor, k,<br />

as shown in Eq. (4):<br />

2 = P · 1−P · I 2 · k 2 .<br />

The nonlinearity observed in Fig. 7 requires additional<br />

modification <strong>of</strong> the facet model. Figure 9 suggests a modification<br />

based on the microstructure <strong>of</strong> the facets. Visual inspection<br />

<strong>of</strong> the printed samples in specular light indicates<br />

that the samples have a variety <strong>of</strong> different microstructures.<br />

Moreover, visual inspection <strong>of</strong> many samples suggests that<br />

the microstructures may be described as a population <strong>of</strong> two<br />

types <strong>of</strong> surfaces; one with well fused toner and the other<br />

with more poorly fused toner. This model is illustrated schematically<br />

in Figure 10.<br />

These two regions would be expected to contribute to<br />

the overall measured BRDF and granularity <strong>of</strong> the sample.<br />

This is described in Eqs. (5)–(7), where P a and P b are the<br />

probability density functions <strong>for</strong> the distribution <strong>of</strong> surface<br />

tilt angles in the two regions illustrated in Fig. 10, a and b<br />

are the rms granularity characteristic <strong>of</strong> the two regions, and<br />

F is the fraction <strong>of</strong> the surface that is region (a). Note that<br />

Eq. (7) reduces to Eq. (3) <strong>for</strong> P a =P b :<br />

P = F · P a + 1−F · P b ,<br />

2 = F · a2 · I 2 + 1−F · b2 · I 2 ,<br />

4<br />

5<br />

6<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 295


Arney et al.: Gloss granularity <strong>of</strong> electrophotographic prints<br />

Figure 11. P normalized BRDF versus angle <strong>for</strong> a typical solid black<br />

printed by laser EP. The solid line is experimental data. The dotted line is<br />

the model <strong>of</strong> Eqs. 5, 9, and10 with s a =5.1°, s b =12.7°, and<br />

F=0.3.<br />

Figure 9. Closeup <strong>of</strong> the specular band <strong>for</strong> experimental samples 1 and<br />

2.<br />

Figure 12. BRGF versus angle <strong>for</strong> a typical solid black printed by<br />

laser EP. The solid line is experimental data. The dotted line is the model<br />

<strong>of</strong> Eq. 8 with k a =0.95 and k b =0.20.<br />

Figure 10. Schematic illustration <strong>of</strong> partial fusing <strong>of</strong> toner.<br />

2 = F · P a · 1−P a · I 2 + 1−F · P b · 1−P b · I 2 .<br />

Equation (7) needs to be adjusted to account <strong>for</strong> the<br />

aperture effect <strong>of</strong> the camera pixels, as described above.<br />

However, one might expect the pixel aperture effect, the constant<br />

k in Eq. (4), not to be the same <strong>for</strong> the two regions.<br />

Thus we write Eq. (7). Equations (5)–(8) represent an expanded<br />

facet model <strong>of</strong> specular reflections:<br />

2 = F · P a · 1−P a · I 2 · k a<br />

2<br />

7<br />

Figure 13. Example <strong>for</strong> a sample <strong>of</strong> solid black toner printed by a typical<br />

laser EP printer. The solid line is Eq. 3, and the points are from experimental<br />

measurements <strong>of</strong> and 2 over the range −50° 50°.<br />

+ 1−F · P b · 1−P b · I 2 · k 2 b . 8<br />

By combining Eqs. (5), (9), and (10), the BRDF can be<br />

modeled by adjusting the parameters, s a , s b , and F to achieve<br />

APPLYING THE EXPANDED FACET MODEL<br />

the best fit with the experimental data. Figure 11 shows the<br />

In order to model the BRDF and BGDF, the two individual result <strong>for</strong> one <strong>of</strong> the printed samples. The model parameters<br />

PDF functions P a and P b are needed. These functions were s a , s b , and F were adjusted to achieve the minimum rms<br />

assumed to be normal distributions described by Eqs. (9) deviation from the experimental data.<br />

and (10):<br />

Equation (8) has two additional parameters, k a and k b ,<br />

that must be adjusted to model the BGDF, versus .<br />

1<br />

P a = e<br />

s a<br />

−2 2<br />

/2s a,<br />

9<br />

Figure 12 shows the minimum rms deviation between the<br />

2 model and the data, and Figure 13 shows the corresponding<br />

plot <strong>of</strong> versus P. The model provides a rationale <strong>for</strong> the<br />

1<br />

significant deviation from linearity predicted by Eq. (3).<br />

P b = e<br />

s b<br />

−2 2<br />

/2s b. 10<br />

2<br />

296 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Arney et al.: Gloss granularity <strong>of</strong> electrophotographic prints<br />

Figure 14. Examples <strong>of</strong> differences in behavior observed and modeled <strong>for</strong> other solid samples <strong>of</strong> black toner<br />

from different printers. Model parameters <strong>for</strong> s a , s b , F, k a , and k b are also shown.<br />

SPATIAL SIGNIFICANCE OF PARAMETERS k a AND k b<br />

Figure 14 illustrates the behavior <strong>of</strong> three additional samples<br />

<strong>of</strong> solid black toner printed by different electrophotographic<br />

printers. The differences in behavior are more easily observed<br />

by plotting versus P·1−P. The solid lines show<br />

the models that best fit the data, and the modeled values <strong>of</strong><br />

s a , s b , F, k a , and k b are also shown. From an analysis <strong>of</strong> 15<br />

samples <strong>of</strong> black toner produced in different printers, this<br />

behavior appears to be representative <strong>of</strong> typical electrophotographic<br />

samples.<br />

The physical meanings <strong>of</strong> parameters a , b , and F are<br />

indicated in the diagram <strong>of</strong> Fig. 10. In all cases s a s b , which<br />

suggests that the range <strong>of</strong> surface tilt angles in region (a) is<br />

less than the range <strong>of</strong> angles in region (b). This is reasonable<br />

if the toner in region (a) is more thoroughly fused than<br />

region (b). The fraction F in every case is less than 0.5,<br />

which suggests that there is less <strong>of</strong> the smooth region (a)<br />

than <strong>of</strong> the more rough region (b).<br />

The physical meaning <strong>of</strong> the parameters k a and k b is less<br />

obvious. In every case k a k b . This suggests the effect <strong>of</strong> the<br />

pixel aperture convolution with the facet size has more <strong>of</strong> a<br />

blurring effect in the rough region (b) than in the smooth<br />

region (a). A possible rationale <strong>for</strong> this observation may be<br />

that the rough region (b) is also a higher frequency region.<br />

The low pass filtering effect <strong>of</strong> the pixel aperture would indeed<br />

be expected to have a have a larger effect on the higher<br />

frequency region (b) than the lower frequency region (a).<br />

Thus k a and k b provide spatial in<strong>for</strong>mation about the gloss<br />

granularity in addition to the magnitude parameters s a and<br />

s b .<br />

As a check <strong>of</strong> the interpretation <strong>of</strong> k a and k b as indices<br />

<strong>of</strong> relative spatial frequency, the (A) image illustrated in<br />

Fig. 3 was low-pass filtered with a Gaussian kernel <strong>of</strong> radius<br />

R. Values <strong>of</strong> R were selected over the range R=0 (no filtering)<br />

to R=20 m. Each image was analyzed to extract experimental<br />

values <strong>of</strong> and as described above, and from<br />

fitting the model to each data set, values <strong>of</strong> the model parameters<br />

were determined as described above. The results<br />

are shown in Figure 15. As one would expect, the smoothing<br />

kernel had only a small effect on the width parameters, s a<br />

and s b . However, the values <strong>of</strong> k a and k b declined significantly,<br />

with k a decreasing much more than k b .<br />

Figure 15. Values <strong>of</strong> s a , s b , k a ,andk b <strong>for</strong> a printed sample <strong>of</strong> black toner<br />

analyzed through low pass filters <strong>of</strong> radius 0R20 m.<br />

DISCUSSION<br />

The behavior shown in Fig. 15 is consistent with the interpretation<br />

<strong>of</strong> k a and k b as noise attenuation factors related to<br />

the low pass filtering effect <strong>of</strong> the effective pixel aperture and<br />

the assumption that facets in the smooth region (a) are<br />

larger (lower frequency) than those in less well fused regions<br />

(b). The smaller facets in region (b) are low pass filtered to a<br />

larger extent than those in region (b) by the pixel aperture<br />

effect, so k b k a . Further filtering by the added Gaussian<br />

filters lowers both k a and k b , as expected, and they approach<br />

the same values <strong>for</strong> extreme low-pass filtering R=20 m.<br />

As discussed in a previous report, the width <strong>of</strong> the<br />

BRDF is an inverse index <strong>of</strong> traditional gloss. 13 A narrow<br />

curve correlates with a high gloss reading. In the current<br />

work, it appears that fused toner can be interpreted in terms<br />

<strong>of</strong> two spatial regions that differ in the degree <strong>of</strong> fusing. The<br />

well fused region has a narrow BRDF, indicated by the value<br />

<strong>of</strong> s a , and the poorly fused region has a broader BRDF indicated<br />

by s b . The magnitude <strong>of</strong> the rms deviation <strong>of</strong> gloss,<br />

called gloss granularity, is indicated by the values <strong>of</strong> k a and<br />

k b . As is typical <strong>of</strong> granularity indices, their magnitude is<br />

dependent on the effective spatial aperture <strong>of</strong> measurement.<br />

In this case that spatial aperture is the area <strong>of</strong> a camera pixel<br />

projected onto the surface. The range <strong>of</strong> behaviors <strong>of</strong> k a and<br />

k b observed in these experiments indicates that gloss granularity<br />

has a significant spatial frequency component that remains<br />

to be examined in future research.<br />

REFERENCES<br />

1 J. S. Arney and Hung Tran, “An inexpensive micro-goniophotometry<br />

you can build”, Proc. IS&T’s PICS Conference on Digital Image Capture,<br />

Reproduction, and Image Quality (IS&T, Springfield, VA, 2002) pp.<br />

179–182.<br />

2 J. S. Arney, H. Hoon, and P. G. Anderson, “A micro-goniophotometer<br />

and the measurement <strong>of</strong> print gloss”, J. <strong>Imaging</strong> Sci. Technol. 48, 458<br />

(2003).<br />

3 J. M. Bennett and L. Mattsson, Introduction to Surface Roughness and<br />

Scattering, 2nd ed. (Optical Soc. <strong>of</strong> America, Washington, DC, 1999),<br />

Chap. 3.<br />

4 J. C. Stover, Optical Scattering, Measurement and Analysis (McGraw Hill,<br />

NY, 1990).<br />

5 I. Nimer<strong>of</strong>f, “Two-parameter gloss methods”, J. Res. Natl. Bur. Stand.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 297


Arney et al.: Gloss granularity <strong>of</strong> electrophotographic prints<br />

58(3), 127 (1957).<br />

6 Standard Practice <strong>for</strong> Angle Resolved Optical Scatter Measurements on<br />

Specular or Diffuse Surfaces, Standard Procedure No. E 1392-96,<br />

American <strong>Society</strong> <strong>for</strong> Testing and <strong>Material</strong>s, 1996.<br />

7 I. Nimer<strong>of</strong>f, “Analysis <strong>of</strong> goniophotometric reflection curves”, J. Res.<br />

Natl. Bur. Stand. 48(6), 441 (1952).<br />

8 H. Rothe and D. Huester, “Application <strong>of</strong> circular and spherical statistics<br />

<strong>for</strong> the interpretation <strong>of</strong> BRDF measurements”, Proc. SPIE 3141(02), 13<br />

(1997).<br />

9 M. Colbert, S. Pattanaik, and J. Krivanek, “BRDF-shop: creating<br />

physically correct bidirectional reflectance distribution functions”, IEEE<br />

Comput. Graphics Appl. 26(1), 30 (2006).<br />

10 P.-Å. Johansson, “Optical homogeneity <strong>of</strong> prints”, doctoral thesis, KTH,<br />

Royal Institute <strong>of</strong> Technology, Stockholm, Sweden, 1999.<br />

11 M.-C. Béland, “Gloss variation <strong>of</strong> printed paper: relationship between<br />

topography and light scattering”, doctoral thesis, KTH, Royal Institute <strong>of</strong><br />

Technology, Stockholm, Sweden, 2001.<br />

12 R. E. Swing, An Introduction to Microdensitometry (SPIE Optical<br />

Engineering Press, Bellingham, WA, 1998).<br />

13 J. S. Arney, P. G. Engeldrum, and H. Zeng, “An Expanded Murray-<br />

Davies model <strong>of</strong> tone reproduction in halftone imaging”, J. <strong>Imaging</strong> Sci.<br />

Technol. 39, 502 (1995).<br />

14 J. S. Arney, C. Scigaj, and P. Mehta, “Linear color addition in halftones”,<br />

J. <strong>Imaging</strong> Sci. Technol. 45, 426 (2001).<br />

15 Y. Kipman, P. Mehta, K. Johnson, and D. Wolin, “A new method <strong>of</strong><br />

measuring gloss mottle and micro-gloss using a line-scan CCD camera<br />

based imaging system”, Proc. IS&T’s NIP17 (IS&T, Springfield, VA,<br />

2001) p. 714.<br />

16 J. S. Arney, L. Ye, and S. Banach, “Interpretation <strong>of</strong> gloss meter<br />

measurements”, J. <strong>Imaging</strong> Sci. Technol. 50(6), 567 (2006).<br />

17 J. S. Arney, P. G. Anderson, G. Franz, and W. Pfeister, “Color properties<br />

<strong>of</strong> specular reflections”, J. <strong>Imaging</strong> Sci. Technol. 50(3), 228 (2006).<br />

18 J. S. Arney, L. Ye, J. Wible, and T. Oswald, “Analysis <strong>of</strong> paper gloss”, J.<br />

Pulp Pap. Sci. 32(1), 19 (2006).<br />

19 M. Ashikhmin, S. Premoze, and P. Shirley, “A micr<strong>of</strong>aceted-based BRDF<br />

generator,” Proc. SIGGRAPH (ACM Press, NY, 2000) pp. 65–74.<br />

298 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 299–309, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Forensic Examination <strong>of</strong> Laser Printers and Photocopiers<br />

Using Digital Image Analysis to Assess Print<br />

Characteristics<br />

J. S. Tchan<br />

MATAR Research Group, London College <strong>of</strong> Communication, Elephant and Castle,<br />

London SE1 6SB, England<br />

E-mail: j.tchan@lcc.arts.ac.uk<br />

Abstract. The work in this paper describes a method that can assist<br />

the process <strong>of</strong> print identification with respect to the printing<br />

machine that produced it. The method used high spatial resolution<br />

and low-noise digital image analysis to measure the sharpness, intensity<br />

and size characteristics <strong>of</strong> individual text characters. The<br />

relative variations <strong>of</strong> these variables were used to identify the machine<br />

that produced the print under examination. The results<br />

showed that three machines could be distinguished and one <strong>of</strong><br />

these machines also showed differences in the print produced when<br />

the toner cartridge was changed. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong><br />

and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4299<br />

Received Oct. 12, 2006; accepted <strong>for</strong> publication Mar. 30, 2007.<br />

1062-3701/2007/514/299/11/$20.00.<br />

INTRODUCTION<br />

A reason why it is frequently not feasible to prosecute counterfeiters<br />

and document fraudsters is due to the difficulty in<br />

establishing links between the counterfeiters or fraudsters<br />

and their printing equipment. This is because <strong>of</strong> the extremely<br />

wide range <strong>of</strong> both cheap and expensive laser and<br />

ink jet printing machines available. This problem is exacerbated<br />

by the fact that new varieties <strong>of</strong> printing machine are<br />

being commercially produced on a continual and frequent<br />

basis.<br />

Also both laser and ink jet printers, which make up<br />

most <strong>of</strong> the cheap <strong>of</strong>fice printing market, have disposable<br />

ink and toner cartridges, or the cartridges can be refilled.<br />

This can prevent valid chemical analysis <strong>of</strong> any similarities or<br />

differences in chemical compositions from the various ink<br />

and toner cartridges. Also chemical analysis is a process that<br />

requires the destruction <strong>of</strong> part <strong>of</strong> the evidence.<br />

Methods which involve microscopy 1 are frequently used<br />

by <strong>for</strong>ensic scientists to determine the production source <strong>of</strong><br />

digital print. The linking <strong>of</strong> a document to a digital printer<br />

in these cases usually involves analyzing variables such as ink<br />

or toner overspray and assessing alignment, spacing and<br />

copy distortion.<br />

Investigations have been carried out by Oliver and<br />

Chen, 2 Tchan, Thompson, and Manning, 3 and Tchan 4,5 using<br />

digital image analysis to link documents to printing machines.<br />

Oliver and Chen have studied the relationship between<br />

the raggedness <strong>of</strong> print and text character distortion<br />

<strong>of</strong> different printers. Tchan, Thompson, and Manning have<br />

taken a similar approach but have also used neural networks<br />

to link the contrast, noise and edge characteristics <strong>of</strong> printed<br />

text to the printing machine that produced it. These attempts<br />

at using digital image analysis however could only<br />

provide a positive test <strong>for</strong> a small range <strong>of</strong> printing machines<br />

and do not account <strong>for</strong> the effect <strong>of</strong> replaceable toner and<br />

ink cartridges.<br />

The analysis <strong>of</strong> the actual shapes <strong>of</strong> the text characters<br />

produced by different print engines using digital image<br />

analysis is another possible way <strong>of</strong> fingerprinting printing<br />

machines, according to Tchan. 6 However, this method not<br />

only suffers from the influence <strong>of</strong> replaceable ink and toner<br />

components distorting the results <strong>of</strong> the analysis, but other<br />

drawbacks as well. First, processing huge amounts <strong>of</strong> image<br />

data is time-consuming due to the large number <strong>of</strong> fonts<br />

and their range <strong>of</strong> sizes from many different makes and<br />

models <strong>of</strong> printing machines. Secondly, the problem <strong>of</strong> ink<br />

spread on different kinds <strong>of</strong> paper or humidity conditions in<br />

ink jet printing processes distorts the shapes <strong>of</strong> text characters.<br />

An identification methodology <strong>for</strong> fingerprinting printing<br />

machines recently considered is to use a technique called<br />

ESDA 7 (Electrostatic Detection Apparatus). ESDA has been<br />

employed to detect and evaluate roller pressure marks on<br />

paper. 8 These pressure marks are due to the interaction between<br />

the paper feeder rollers and the paper substrate. The<br />

original application was <strong>for</strong> the detection <strong>of</strong> pen impressions<br />

several layers down in a notepad. It works by charging paper<br />

surfaces with a high voltage. If toner particles are applied to<br />

the charged paper surface, imperceptible pen marks due to<br />

writing on a piece <strong>of</strong> paper placed above this sheet may be<br />

revealed. This means <strong>for</strong> example, in a notepad, writing can<br />

be read from sheets many layers down from the top sheet<br />

that has the actual writing. As the ESDA technique has been<br />

shown to detect weak imperceptible pressure marks from<br />

pens, it might be able to detect pressure marks from printing<br />

rollers.<br />

If the pressure marks can be detected, then to link a<br />

document to the printer that produced it would require<br />

comparing the width and spacing characteristics <strong>of</strong> the roll-<br />

299


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Table I. List <strong>of</strong> the three printing machines and the printing machine with a change <strong>of</strong><br />

toner cartridge used in the investigation.<br />

Printing System<br />

Canon IR 3570 Photocopier<br />

HP 1200 Laser Jet Printer<br />

HP 4250 Laser Jet Printer Toner Cartridge 1<br />

HP 4250 Laser Jet Printer Toner Cartridge 2<br />

ers <strong>of</strong> the machine in question. However, detecting the pressure<br />

marks may not always be possible <strong>for</strong> the following<br />

reasons. First, the pressure exerted by the rollers is weak,<br />

generally much weaker than pen pressure. If the pressure<br />

marks are present, they will sometimes be difficult to detect,<br />

even with the most sophisticated digital image processing<br />

systems. Secondly, assuming that the pressure marks are detectable,<br />

heavy handling can destroy them.<br />

A similar technique that has been recently explored concerns<br />

ink jet printers and indentations on the paper after a<br />

printed sheet has been fed through the machine. 9 These indentations<br />

are caused by the spoke wheels that feed the paper<br />

through the ink jet printer and are most perceptible on<br />

moist parts <strong>of</strong> the paper caused by wet ink. Due to the heavy<br />

pressure imparted by the spoke wheels, the indentations can<br />

be seen using optical microscopy without the aid <strong>of</strong> ESDA.<br />

However not all ink jet printers have spoke wheels so this<br />

technique does not apply to all ink jet printing systems.<br />

Another type <strong>of</strong> method that has been considered <strong>for</strong><br />

the identification <strong>of</strong> laser printers exploits imperfections in<br />

the print known as banding. Banding imperfections are lines<br />

across the printed page when smooth print is required. 10<br />

The effect has been attributed to the following two causes.<br />

First, fine banding due to the imbalance <strong>of</strong> the rotor component<br />

<strong>of</strong> the polygon mirror or mechanical weaknesses <strong>of</strong><br />

the laser scanning unit. Secondly, rough banding caused by<br />

unsteady motion <strong>of</strong> the photoconductor drum or the fuser<br />

unit.<br />

Mikkilineni et al. 11 have devised a system that uses a<br />

scanner to analyze relative texture differences on a printed<br />

page caused by banding effects. This system has shown that<br />

9 out <strong>of</strong> a set <strong>of</strong> 10 laser printing machines were successfully<br />

identified.<br />

The method described in this paper is an alternative<br />

method <strong>of</strong> measuring banding effects in laser printers and<br />

photocopiers. Instead <strong>of</strong> scanning images and analyzing the<br />

relative texture <strong>of</strong> text characters, it uses a high resolution<br />

and low-noise digital image analysis system to measure the<br />

following variables in printed text. These variables are sharpness,<br />

intensity, and size. The following section describes the<br />

methodology and the experimental setup involved.<br />

EXPERIMENTAL PROCEDURE<br />

When a completely black page was printed out on a photocopier,<br />

two different laser printers and on one <strong>of</strong> the two<br />

laser printers with a different toner cartridge, Table I, the<br />

following effects in Figure 1 were seen. These are sketches <strong>of</strong><br />

Figure 1. Sketches <strong>of</strong> the lines produced by the four different printing<br />

samples used in the investigation.<br />

the banding lines seen with an indication <strong>of</strong> their dimensions<br />

and separations. It was observed that some <strong>of</strong> the horizontal<br />

lines from the HP 4250 <strong>for</strong> the same toner cartridge<br />

were not in fixed positions and the lines were not always<br />

equal in number <strong>for</strong> different printed sheets. It is unknown<br />

whether some <strong>of</strong> the lines were random or followed a complex<br />

pattern since further investigation <strong>of</strong> this effect has yet<br />

to be made. However, <strong>for</strong> this part <strong>of</strong> the investigation, only<br />

a confirmation <strong>of</strong> the existence <strong>of</strong> the banding effects that<br />

are a common feature <strong>of</strong> digital printers was required.<br />

The experimental procedure can also be separated into<br />

three distinct stages. In the first stage the banding effect was<br />

observed <strong>for</strong> a test page that was entirely covered in solid<br />

toner as stated above. In stage two, a test page <strong>of</strong> the same<br />

text character was produced and physical differences in the<br />

print were investigated <strong>for</strong> each printing machine used. This<br />

was completed using high-resolution digital image analysis.<br />

In stage three, a page <strong>of</strong> ordinary text was produced and<br />

patterns in the text were again investigated using highresolution<br />

digital image analysis. The process is illustrated in<br />

Figure 2, the flow chart below, and will be discussed in<br />

greater detail later.<br />

The effects <strong>of</strong> banding on printed text were investigated<br />

using a high-resolution digital image analysis system, which<br />

has been built specifically to analyze the print. Figure 3 illustrates<br />

how the camera was attached to the stand and how<br />

the lighting source was attached to the camera.<br />

300 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 2. The three different stages <strong>of</strong> the investigation.<br />

Figure 4. The three stages required to make the necessary measurements<br />

<strong>for</strong> the analysis.<br />

Table II. A summary <strong>of</strong> the measurements taken.<br />

Measurements<br />

Variable<br />

1 Location <strong>of</strong> the peak maximum <strong>for</strong> the print<br />

region, Fig. 5.<br />

2 The value <strong>of</strong> the peak maximum <strong>for</strong> the print<br />

region, Fig. 5.<br />

3 The text character area is calculated by counting<br />

the number <strong>of</strong> pixels below an arbitrarily selected<br />

threshold between the nonimage and image<br />

peaks, Fig. 5.<br />

4 Integration <strong>of</strong> the peak area <strong>for</strong> the print region<br />

divided by the text character area to determine<br />

the average image intensity, Fig. 5.<br />

Figure 3. The camera, lens, lighting system, and stand.<br />

The camera employed in the investigation was a<br />

Hamamatsu C4742-95 camera. This camera had a Peltier<br />

cooled CCD chip to increase the signal to noise ratio. The<br />

camera was attached firmly to a camera stand that weighs<br />

approximately 30 kg. The lighting unit was firmly screwed<br />

onto the lens <strong>of</strong> the camera. The lighting unit consisted <strong>of</strong> a<br />

circular array <strong>of</strong> red LEDS. The LEDS were connected to a<br />

laboratory power supply with low ripple.<br />

Figure 4 shows in block diagram <strong>for</strong>m the different<br />

hardware and s<strong>of</strong>tware components <strong>of</strong> the image analysis<br />

system. The image data from the camera was digitized to<br />

8 bit resolution using Matrox Mil s<strong>of</strong>tware. The data was<br />

subsequently analyzed using Visual Basic that was compatible<br />

with Matrox Mil. Algorithms were developed using Visual<br />

Basic Active X language that could per<strong>for</strong>m the following<br />

computations on individual text characters.<br />

Four measurements were taken from the text characters.<br />

They related to the image sharpness, intensity and size <strong>of</strong> the<br />

print under investigation. These are summarized in Table II.<br />

The image window size, Figure 5, was 480 by 483 and had a<br />

tonal resolution <strong>of</strong> 256.<br />

Measurement 1 is the position <strong>of</strong> the peak in standard<br />

8 bit gray scale <strong>for</strong> the tonal distribution <strong>of</strong> the printed region.<br />

The location <strong>of</strong> this peak has the lower gray scale<br />

value, in this case at about 55, Fig. 5. The other peak corresponds<br />

to the tonal distribution <strong>of</strong> the unprinted white paper.<br />

The position <strong>of</strong> the peak <strong>for</strong> the white paper used in the<br />

investigation is located at just over 100, Fig. 5. Such a low<br />

value is due to the arbitrarily low lighting exposure chosen<br />

<strong>for</strong> the CCD camera. Attempting to stretch the distance between<br />

the two peaks <strong>for</strong> the black print and white paper<br />

regions, by increasing the exposure too much, can sometimes<br />

reduce the precision <strong>of</strong> the system.<br />

Measurement 2 is the height <strong>of</strong> the peak from the tonal<br />

distribution <strong>of</strong> the printed region.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 301


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 6. The two sets <strong>of</strong> test targets that were used to produce the results<br />

in this investigation.<br />

Figure 5. An illustration <strong>of</strong> the data acquired from the image analysis<br />

system.<br />

Measurement 3 is the area <strong>of</strong> the printed text character,<br />

this is calculated by counting the number <strong>of</strong> pixels below a<br />

fixed arbitrary gray scale level.<br />

Measurement 4 averages the overall intensity <strong>of</strong> the<br />

printed region by integrating the total intensity from the<br />

printed region and dividing it by the number <strong>of</strong> pixels from<br />

the text character below the fixed arbitrary level chosen <strong>for</strong><br />

measurement 3. Figure 5 and Table II illustrate and summarize<br />

the measurements made. The measurements were taken<br />

individually and sequentially <strong>for</strong> each text character by<br />

manual alignment under the image window.<br />

Figure 6 shows the two print samples used in the investigation.<br />

They were both printed on the same batch <strong>of</strong> standard<br />

laser printer paper in all cases. The font was Times New<br />

Roman and the font size was 22 pts. 22 pts is a large-sized<br />

font; it was used because it facilitated relatively quick measurements<br />

to show that the method is viable.<br />

A test sheet was produced that consisted <strong>of</strong> a series <strong>of</strong><br />

the letter “W” in Times Roman font and 22 points in size.<br />

There were 17 W’s across the page and 30 down the page.<br />

The selection <strong>of</strong> the letter “W” was arbitrary. However, the<br />

size was important to facilitate ease <strong>of</strong> measurement when<br />

recording the data using the digital image analysis system.<br />

Figure 7. In the case <strong>of</strong> the normal text page a mask was required to<br />

eliminate the effect <strong>of</strong> adjacent text characters.<br />

In the case <strong>of</strong> the page <strong>of</strong> normal text, Fig. 6 on the<br />

right, a cardboard mask, Figure 7, was required to shield the<br />

effects <strong>of</strong> nearby letters influencing the readings since unwanted<br />

parts <strong>of</strong> letters appeared in the image window. The<br />

test set <strong>of</strong> “W’s” did not have this effect since the spacing <strong>of</strong><br />

the “W’s” was designed to eliminate the requirement <strong>for</strong> a<br />

mask. Figure 8 shows the “e’s” that were analyzed in the page<br />

<strong>of</strong> normal text.<br />

RESULTS<br />

First, the accuracy <strong>of</strong> the system was established by assessing<br />

variations in the intensity and area measurements <strong>for</strong> a<br />

302 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 8. The letter “e’s” that where selected from the test page <strong>for</strong> the<br />

classification <strong>of</strong> the two toner cartridges <strong>of</strong> the HP 4250.<br />

single text character sampled 50 times. A maximum value<br />

from the mean <strong>of</strong> ±0.05% was observed when the mask was<br />

not employed and ±0.03% when the mask was used <strong>for</strong> the<br />

intensity measurements. These error values were tripled in<br />

size <strong>for</strong> the area measurements. This text character print<br />

sample was also used to check <strong>for</strong> any substantial drift in the<br />

system at the beginning <strong>of</strong> each day over a period <strong>of</strong> about<br />

20 days that the measurements were taken. Substantial drift,<br />

which could be caused by small change in the LED voltage,<br />

lens settings or focus, did not occur over the period <strong>of</strong> data<br />

collection. This was probably due to the mechanical robustness<br />

<strong>of</strong> the optical system and the quality <strong>of</strong> the LED illumination<br />

system.<br />

Secondly, two W test pages from the HP 4250 printer,<br />

one from each <strong>of</strong> the two toner cartridges were analyzed.<br />

This, as in all <strong>of</strong> this investigation, required individual sequential<br />

manual alignment and subsequent measurement<br />

from each text character on a line <strong>of</strong> text. Figure 9 shows<br />

how the size <strong>of</strong> the letter “W” <strong>for</strong> line 1 and 10 <strong>of</strong> the grid<br />

changes down the page <strong>for</strong> the HP 4250 printer using the<br />

two different toner cartridges. The consistent patterns indicate<br />

that the digital image analysis system has recorded<br />

meaningful results.<br />

Thirdly, a demonstration was made <strong>of</strong> how the four<br />

printing samples could be distinguished using the W template.<br />

This was achieved by using three adjacent W test<br />

sheets from each <strong>of</strong> the four printing samples in a print run<br />

<strong>of</strong> 10. The reported results in this section used only the first<br />

horizontal and vertical lines <strong>of</strong> the W test sheets because <strong>of</strong><br />

Figure 9. A comparison <strong>of</strong> the text character size down line 1 and 10 <strong>for</strong> the two HP 4250 toner cartridges.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 303


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 10. A comparison <strong>of</strong> the text character area down line 1 <strong>of</strong> the W test pages. Left, <strong>for</strong> 3 adjacent<br />

pages; right, their average.<br />

the time-consuming nature <strong>of</strong> recording the measurements<br />

and associated time constraints <strong>of</strong> the researcher. The time<br />

problem was only discovered during the experimental phase<br />

and centered on alignment difficulties <strong>of</strong> the text characters<br />

in the image window. The left-hand graphs <strong>of</strong> Figures 10–13<br />

show the individual data from the three sheets and on the<br />

right-hand side their averages.<br />

It was shown that the peak size, the average intensity<br />

and the size measurements yielded useful in<strong>for</strong>mation <strong>for</strong><br />

the classification process. The position <strong>of</strong> the image peak<br />

304 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 11. A comparison <strong>of</strong> the text character area across line 1 <strong>of</strong> the W test pages. Left, <strong>for</strong> 3 adjacent<br />

pages; right, their average.<br />

remained constant at either 54 or 55 and provided no useful<br />

classification data. It does however provide a useful check on<br />

the stability <strong>of</strong> the illumination levels throughout the period<br />

when the measurements were taken. The experimental results<br />

using the W template show that all four printing<br />

samples could be differentiated by a combination <strong>of</strong> the<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 305


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 12. A comparison <strong>of</strong> the relative average intensity across the W test pages, left, <strong>for</strong> 3 adjacent pages;<br />

right, their average.<br />

analysis <strong>of</strong> the area variation <strong>of</strong> the text characters down the<br />

page and the average intensity, peak size and area variations<br />

<strong>of</strong> the text characters across the page, Figs. 10–13.<br />

Finally, the masking technique was employed, Fig. 7,<br />

with the printed page <strong>of</strong> normal text shown in Figs. 6 and 8<br />

to find differences from the different printing examples, in<br />

this case from the HP 4250 printer when the toner cartridge<br />

was changed. Figure 14 shows a shallow well or an inverted<br />

sharp spike at position 15 <strong>for</strong> toner cartridge 1 and an<br />

inverted sharp spike <strong>for</strong> toner cartridge 2 at position 14. This<br />

result was obtained from a print run <strong>of</strong> 201. The graphs,<br />

Fig. 14, show the results from sheets 1 <strong>of</strong> 101 and 201. The<br />

306 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 13. A comparison <strong>of</strong> the text character peak intensities across the W test pages. Left, <strong>for</strong> 3 adjacent<br />

pages; right, their average.<br />

measurements were also sampled <strong>for</strong> other sheets in the<br />

print run at 20 sheet intervals (numbers 1, 21, 41, 61, 81,<br />

101, 121, 141, 161, 181, and 201 in total) and were<br />

100% consistent in the fact that the inverted sharp spike at<br />

position 14 only appears <strong>for</strong> all print samples from toner<br />

cartridge 2 and not at all <strong>for</strong> toner cartridge 1.<br />

CONCLUSIONS<br />

The results <strong>of</strong> the investigation thus far demonstrate the potential<br />

<strong>of</strong> the method <strong>for</strong> the <strong>for</strong>ensic analysis <strong>of</strong> print, both<br />

in linking a machine to a particular document and to show<br />

whether a document has been tampered with.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 307


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

Figure 14. The area differences using the page <strong>of</strong> text <strong>for</strong> the HP 4250 laser printer <strong>for</strong> the two toner<br />

cartridges.<br />

It has been shown that a small number <strong>of</strong> toner based<br />

printing systems can be classified using high-resolution image<br />

analysis to measure the relative changes in the physical<br />

properties <strong>of</strong> individual text characters both across and<br />

down pages <strong>of</strong> printed text. Even at this stage <strong>of</strong> its development<br />

the system has potentially useful <strong>for</strong>ensic applications.<br />

Also, these results correlate with the work carried out by<br />

Mikkilineni et al. on the measurement <strong>of</strong> surface texture by<br />

scanning the text characters <strong>of</strong> laser printers.<br />

In this investigation more work is required on smaller<br />

text characters. In particular positive results are required on<br />

font sizes <strong>of</strong> 10 since this is typical <strong>for</strong> documents. If limitations<br />

in the hardware become apparent or the banding<br />

signatures become weaker when smaller font sizes are considered<br />

then a larger set <strong>of</strong> text characters could be analyzed<br />

statistically to try to overcome the limitations.<br />

However, the further work stated above requires assistance<br />

from better measurement and analysis techniques with<br />

greater precision. This is due to the labor intensive nature <strong>of</strong><br />

the experimental work. Statistical analysis techniques such as<br />

moving averages or autocorrelation analysis can enhance the<br />

data, thereby reducing the volume <strong>of</strong> data required from the<br />

308 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Tchan: Forensic examination <strong>of</strong> laser printers and photocopiers using digital image analysis…<br />

print samples <strong>for</strong> accurate classification. The current masking<br />

method is difficult to carry out in practice <strong>for</strong> a large<br />

number <strong>of</strong> small text characters and needs an alternative. A<br />

possible solution to this problem is to use or develop character<br />

recognition s<strong>of</strong>tware that can automatically isolate and<br />

classify individual text characters.<br />

The suggestions given <strong>for</strong> further work indicate, in the<br />

longer term, since data acquisition is time-consuming, an<br />

automated system that uses a high-resolution and fast scanning<br />

device is needed. This method is expensive to develop<br />

and implement but will greatly improve the volume <strong>of</strong> print<br />

that can be processed in a given time, and should enable<br />

machines to be examined more both quickly and more accurately<br />

from smaller font sized print.<br />

REFERENCES<br />

1 B. S. Lindbolm, and R. Gervais, Scientific Examination <strong>of</strong> Questioned<br />

Documents (Taylor and Francis, Boca Raton, FL, 2006).<br />

2 J. Oliver, and J. Chen, “Use <strong>of</strong> signature analysis to discriminate digital<br />

printing technologies”, Proc. IS&T’s NIP18 (IS&T, Springfield, VA, 2002)<br />

pp. 218–222.<br />

3 J. S. Tchan, R. C. Thompson and A. Manning, “The use <strong>of</strong> neural<br />

networks in an image analysis system to distinguish between laser prints<br />

and their photocopies”, J. <strong>Imaging</strong> Sci. Technol. 44(2), 132–144 (2000).<br />

4 J. S. Tchan, “Classifying digital prints according to their production<br />

process using image analysis and artificial neural networks”, Proc. SPIE<br />

3973, 105–116, (2000).<br />

5 J. S. Tchan, “The development <strong>of</strong> an image analysis system that can<br />

detect fraudulent alterations made to printed images”, Proc. SPIE 5310,<br />

151–159 (2004).<br />

6 J. S. Tchan, “Forensic analysis <strong>of</strong> print using digital image analysis”,<br />

Proc. SPIE 5007, 61–72 (2003).<br />

7 J. Levinson, Questioned Documents: A Lawyer’s Handbook (Academic<br />

Press, London, 2001).<br />

8 G. M. Laporte, “The use <strong>of</strong> an electrostatic detection device to identify<br />

individual and class characteristics on documents produced by printers<br />

and copiers-A preliminary study”, J. Forensic Sci. 49(3), 610–620 (2004).<br />

9 Y. Akao, K. Kobayashi, and Y. Seki, “Examination <strong>of</strong> spur marks found<br />

on inkjet printed documents”, J. Forensic Sci., 50(4), 915–923 (2005).<br />

10 J. You, “Banding reduction in an electrophotographic printer”, J.<br />

<strong>Imaging</strong> Sci. Technol. 49(6), 635–640 (2005).<br />

11 A. K. Mikkilineni, P. Chiang, G. N. Ali, G. T. C. Chiu, J. P. Allebach, and<br />

E. J. Delp, “Printer identification based on graylevel co-occurrence<br />

features <strong>for</strong> security and <strong>for</strong>ensic applications”, Proc. SPIE 5681,<br />

430–440 (2005).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 309


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 310–316, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Moiré Analysis <strong>for</strong> Assessment <strong>of</strong> Line<br />

Registration Quality<br />

Nathir A. Rawashdeh, Daniel L. Lau and Kevin D. Donohue <br />

University <strong>of</strong> Kentucky, ECE, 453 F. Paul Anderson Tower, Lexington, Kentucky 40506-0046<br />

E-mail: nathir@ieee.org<br />

Shaun T. Love<br />

Lexmark International, Inc., 740 W. New Circle Rd., Lexington, Kentucky 40550<br />

Abstract. This paper introduces objective macro and micro line<br />

registration quality metrics based on Moiré interference patterns<br />

generated by superposing a lenticular lens grating over a hardcopy<br />

test page consisting <strong>of</strong> high-frequency Ronchi rulings. Metrics <strong>for</strong><br />

macro and micro line registration are defined and a measurement<br />

procedure is described to enhance the robustness <strong>of</strong> the metric<br />

computation over reasonable variations in the measurement process.<br />

The method analyzes low frequency interference patterns,<br />

which can be scanned at low resolutions. Experimental measurements<br />

on several printers are presented to demonstrate a comparative<br />

quality analysis. The metrics demonstrate robustness to small<br />

changes in the lenticular lens and grating superposition angle. For<br />

superposition angles varying between 2° and 5°, the coefficients <strong>of</strong><br />

variance <strong>for</strong> the two metrics are less than 5%, which is small enough<br />

<strong>for</strong> delineating between test patterns <strong>of</strong> different print quality.<br />

© 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4310<br />

INTRODUCTION<br />

Image quality analysis is an important component in the<br />

development and operation <strong>of</strong> various digital imaging technologies,<br />

such as displays, scanners and printers. To produce<br />

visually pleasing images, devices must be designed to minimize<br />

defects, such as problems related to color registration<br />

and line quality. An efficient way <strong>of</strong> measuring imaging defects<br />

is through the use <strong>of</strong> special test targets, which are<br />

designed to test the limits <strong>of</strong> the respective imaging technology.<br />

Analysis based on test target results can be used to track<br />

and minimize image defects during the development phase.<br />

This paper presents a method <strong>for</strong> analyzing printed line<br />

quality by analyzing the Moiré patterns resulting from the<br />

superposition <strong>of</strong> a test pattern, consisting <strong>of</strong> finely spaced<br />

lines, and an array <strong>of</strong> cylindrical lenses <strong>of</strong> similar spacing.<br />

Other approaches to line quality attributes <strong>for</strong> hardcopy<br />

output include blurriness, raggedness, stroke width, darkness,<br />

contrast, fill, and registration. 1–3 The test targets <strong>for</strong><br />

these measures consist <strong>of</strong> a printed black line on a white<br />

background. The quality attributes are then quantified<br />

through measurements from the printed line. Blurriness<br />

measures the average transition length from light to dark,<br />

<br />

IS&T Member<br />

Received Jan. 5, 2007; accepted <strong>for</strong> publication Mar. 1, 2007.<br />

1062-3701/2007/514/310/7/$20.00.<br />

and raggedness measures the geometric distortion <strong>of</strong> the<br />

line’s edge from its ideal shape. Line width is the average<br />

stroke width measured from either edge along a direction<br />

normal to the line under analysis. Line darkness measures<br />

the mean line density, which can vary due to voids <strong>for</strong> example.<br />

The contrast attribute captures the relationship between<br />

the darkness <strong>of</strong> the line and that <strong>of</strong> its surrounding<br />

field by measuring the mean reflectance factors. Contrast<br />

can vary due to blurring, extraneous marks, haze, or substrate<br />

type. Fill refers to the appearance <strong>of</strong> darkness within<br />

the inner boundary <strong>of</strong> the line. One example <strong>of</strong> line registration<br />

is the color registration <strong>of</strong> the CMYK components in<br />

an inkjet printer. If the same line is printed once with each<br />

color, then ideally, all four color lines should collapse into<br />

one, and any consistent increase in line width would indicate<br />

position errors, or mis-registration, <strong>of</strong> one or more ink<br />

components. 3<br />

This paper introduces new metrics that differ from previous<br />

line quality attribute measures in that they are directly<br />

based on the printer’s ability to create fine detailed lines.<br />

While this metric may be influenced by measures such as<br />

raggedness and blur, its use <strong>of</strong> fine details makes it unique<br />

relative to previous measures. The measurement method involves<br />

the analysis <strong>of</strong> low frequency Moiré patterns that<br />

change according to small changes in the test patterns. The<br />

test pattern consists <strong>of</strong> finely spaced parallel lines, which an<br />

imperfect printer reproduces with some line placement (or<br />

registration) errors. The parallel lines are no longer uni<strong>for</strong>mly<br />

spaced in this case, and this is reflected in the resulting<br />

Moiré line shape. Moiré patterns are used as a nondestructive<br />

analysis tool in Moiré interferometry. For this<br />

method a photographic grid is printed on the surface <strong>of</strong> a<br />

material under investigation and is irradiated by coherent<br />

light. The interfering fringes (Moiré patterns) can indicate<br />

the presence <strong>of</strong> local stress and de<strong>for</strong>mation <strong>for</strong> in-plane<br />

displacement. 4,5 Moiré interferometry techniques have the<br />

advantage <strong>of</strong> being able to analyze a broad range <strong>of</strong> engineering<br />

materials in small analysis zones at high spatial resolution<br />

and sensitivity. This work extends the principles <strong>of</strong><br />

Moiré interferometry to assess line registration quality by<br />

analyzing the Moiré patterns produced by the superposition<br />

310


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

Figure 1. Moiré pattern <strong>of</strong> spacing P and direction <strong>for</strong>med by the<br />

angled superposition <strong>of</strong> a Ronchi ruling and a lenticular grating.<br />

<strong>of</strong> a lenticular grating on a printed Ronchi-ruling test pattern<br />

to characterize underlying printed line registration<br />

errors.<br />

The lenticular grating consists <strong>of</strong> a plastic sheet that is<br />

smooth on one side and holds an array <strong>of</strong> parallel cylindrical<br />

lenses or prisms on the other side. The printed test pattern is<br />

a Ronchi ruling (a rectangular spatial wave linear grating)<br />

with a similar line spacing as the lenticular grating. This<br />

quality assessment approach lends itself to automation, since<br />

the lenticular grating is thin enough to be fixed to the glass<br />

surface <strong>of</strong> a flatbed scanner and does not interfere with its<br />

automatic document feeder mechanism. Since only the<br />

shape <strong>of</strong> the Moiré lines is used <strong>for</strong> analysis, it is sufficient to<br />

use a relatively inexpensive scanner (or scan faster), because<br />

high-resolution detail and tone reproduction accuracy are<br />

not crucial. This paper presents the underlying equations<br />

affecting the critical details <strong>of</strong> the Moiré patterns, describes a<br />

procedure <strong>for</strong> robust measurement and computation <strong>of</strong><br />

macro and micro line quality metrics, and presents results<br />

<strong>for</strong> several printers. Measurements are analyzed and compared<br />

to a visual assessment <strong>of</strong> line quality based on a magnified<br />

view <strong>of</strong> the Ronchi pattern created with a high resolution<br />

scanner.<br />

The text is organized as follows. The Moiré Model section<br />

describes the Moiré line model and discusses normalization<br />

techniques and ranges <strong>of</strong> superposition angles <strong>for</strong><br />

robust measurements. The Line Registration Quality Measurements<br />

and Metrics section describes the measurement<br />

procedure and computation <strong>of</strong> the macro and micro line<br />

registration metrics. The Results and Analysis section presents<br />

measurement results from three different printers and<br />

analyzes measurement variability and quality assessment. Finally,<br />

the Conclusion section summarizes results and presents<br />

conclusions.<br />

MOIRÉ MODEL<br />

Figure 1 illustrates the Moiré fringe pattern produced by the<br />

superposition <strong>of</strong> a (printed) linear grid <strong>of</strong> spacing P 0 , and a<br />

lenticular grating <strong>of</strong> spacing P 1 at an angle . The Moiré<br />

lines are produced by the lenticular lenses intersecting with<br />

the individual lines <strong>of</strong> the Ronchi ruling. Only two lenses are<br />

illustrated in this figure; however, a sheet consisting <strong>of</strong> many<br />

lenticular lenses produces extended patterns <strong>of</strong> Moiré lines.<br />

Figure 2. Photograph <strong>of</strong> a printed horizontal Ronchi ruling test pattern at<br />

a small angle with a superimposed lenticular lens grating. Resulting Moiré<br />

patterns are dark vertical curved lines.<br />

The Moiré line spacing, as shown in Fig. 1, is related to the<br />

superposition parameters by 6,7<br />

P =<br />

P 0 P 1<br />

P 0 2 + P 1 2 −2P 0 P 1 cos .<br />

The angle <strong>of</strong> the Moiré lines with the base <strong>of</strong> the lenticular<br />

sheet is given by 6,7<br />

tan =<br />

P 1 sin<br />

P 0 − P 1 cos .<br />

An actual Moiré pattern from such a sheet is shown in Figure<br />

2. The Moiré lines deviate from straight lines due to<br />

printer imperfections. The Ronchi rule pattern was printed<br />

with a 0.4233 mm spacing, and the lenticular lens sheet consisted<br />

<strong>of</strong> lenses with a spacing <strong>of</strong> 0.630 mm (40 lenses per<br />

inch) and the lenses had a magnification factor <strong>of</strong> 1.505<br />

(making the effective Ronchi line spacing equal 0.637 mm).<br />

Fluctuations in the printed line spacing, P 0 , result from line<br />

registration errors and create deviations in the Moiré line<br />

angle , according to Eq. (2).<br />

The sensitivity <strong>of</strong> the resulting Moiré line direction<br />

angle to the superposition is shown in Figure 3, which<br />

plots Eqs. (1) and (2) as functions <strong>of</strong> . For the 10° interval<br />

shown, the Moiré line spacing decreases from around<br />

80 mm to 2.5 mm. Both P and exhibit relatively little<br />

change <strong>for</strong> greater than 4°. In practical implementations,<br />

the superposition angle cannot be precisely controlled. So<br />

ensuring that these changes do not significantly affect the<br />

metric is a critical issue to the usefulness <strong>of</strong> this method.<br />

There<strong>for</strong>e, selecting an around 4° reduces the impact <strong>of</strong><br />

small changes in the superposition angle. This results in<br />

multiple low-frequency Moiré lines over the test pattern <strong>for</strong><br />

robust analysis.<br />

1<br />

2<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 311


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

where x is the additive displacement term <strong>for</strong> underlying<br />

line spacing P 0 , and x is the deviation from Moiré angle<br />

. The angular deviations vary over the printed pattern<br />

based on changes in the underlying test pattern.Without loss<br />

<strong>of</strong> generality, this direction is denoted as a function <strong>of</strong> a<br />

single variable x. To separate the deviation terms, a Taylor<br />

series can be applied to the cotangent function and expanded<br />

about . After higher-order terms are dropped (assuming<br />

small deviations), Eq. (3) results in<br />

x− x sin2 <br />

P 1<br />

<br />

sin + cot + cot<br />

− P 0csc<br />

csc 2 P 1 csc 2 ,<br />

where terms in the parenthesis are constant over x and relate<br />

to a constant <strong>of</strong>fset on the Moiré angle . They are subtracted<br />

out in the estimation procedure. The effective gain<br />

term that scales the line position deviations to Moiré pattern<br />

angle deviations is given by<br />

4<br />

Figure 3. Plots <strong>of</strong> Moiré fringe spacing P and direction as a function <strong>of</strong><br />

the superposition angle between a Ronchi ruling and a lenticular grating.<br />

Circles indicate manual measurements; solid lines are plots <strong>of</strong> Eqs.<br />

1 and 2.<br />

Also included in the plots <strong>of</strong> Fig. 3, are actual measurements<br />

<strong>of</strong> the Moiré line spacing and angle <strong>for</strong> five values <strong>of</strong><br />

. The measurements were made by manually setting the<br />

superposition angle and using a ruler to visually measure<br />

the resulting Moiré spacing, and a protractor to measure the<br />

Moiré direction ø. The resulting measurements agreed well<br />

with Eqs. (1) and (2) as can be seen by the measurement<br />

marker on the graphs <strong>of</strong> Fig. 3. For the measurement system<br />

proposed in this work, the lenticular grating is <strong>of</strong> high precision,<br />

while the actual superposition angle may also be variable<br />

depending on the mechanics used to load the test sheet.<br />

The following equations show the critical parameters relating<br />

the underlying line registration to the Moiré pattern parameters<br />

used in the measurement. From this derivation, a<br />

normalization step is presented to reduce the sensitivity <strong>of</strong><br />

metrics computed from the Moiré pattern to variations in<br />

parameters <strong>of</strong> the measurement system.<br />

A relationship between changes in the underlying line<br />

spacing and change in the angle <strong>of</strong> the Moiré pattern can be<br />

seen from taking the reciprocal <strong>of</strong> Eq. (2) and adding deviation<br />

terms to the test pattern line spacing and Moiré pattern<br />

angles to obtain<br />

cot + x = P 0 + x − P 1 cos<br />

P 1 sin<br />

= P 0 + x<br />

P 1 sin − cot,<br />

3<br />

g m =− sin2 <br />

P 1 sin ,<br />

where the gain/sensitivity is determined by the lenticular<br />

grid spacing P 1 and superposition angle . An alternate<br />

derivation <strong>of</strong> g m can be obtained directly through the ratio<br />

<strong>of</strong> the root-mean-square (rms) deviations <strong>of</strong> and P 0 . This<br />

would eliminate the <strong>of</strong>fset (zero-order) term <strong>of</strong> Eq. (4) and<br />

allow the gain factor g m to be computed directly from the<br />

derivatives <strong>of</strong> Eq. (3) with respect to and P 0 . The gain<br />

factor g m in this case is simply /P 0 .<br />

Since the deviations, x, will be extracted from the<br />

Moiré patterns and used <strong>for</strong> characterization, the sensitivity<br />

to becomes an issue <strong>for</strong> consistent measurements (small<br />

changes in , <strong>for</strong> near zero, can result in large changes in<br />

the gain). This variability can be significantly reduced by<br />

dividing the measured angle deviation by the measured distance<br />

between the moiré lines, if the effective Ronchi pattern<br />

line spacing is close to that <strong>of</strong> the lenticular grid. With P 1<br />

equal to P 0 , Eq. (1) can be simplified using the half angle<br />

<strong>for</strong>mula to show the Moiré line spacing is related to by<br />

P 1<br />

5<br />

P =<br />

2 sin/2 . 6<br />

For small (as is the case here), sin approximately equals<br />

(in radians). Thus, by applying this approximation to Eqs.<br />

(5) and (6), the normalized gain becomes<br />

ḡ m = g m<br />

P − sin2 <br />

P 1<br />

2<br />

. 7<br />

This equation shows that the repeatability <strong>of</strong> the measurement<br />

is enhanced through this normalization. The dominant<br />

scale factor controlling the gain on the angular displacement<br />

is now primarily dependent on the lenticular gird spacing,<br />

which can be precisely controlled and does not change with<br />

the superposition angle. The next section describes the extraction<br />

<strong>of</strong> x and P <strong>for</strong>m the scanned Moiré patterns, and<br />

the development <strong>of</strong> the metrics based on the normalization<br />

described by Eq. (7).<br />

312 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

Figure 4. Line misregistration in 8.5 mm long vertical slices <strong>of</strong> the top twenty lines <strong>of</strong> the actual printed test<br />

patterns <strong>of</strong> printers LP1 top and LP2 bottom. Rulers show ideal line locations.<br />

LINE REGISTRATION QUALITY MEASUREMENT<br />

AND METRICS<br />

An example <strong>of</strong> a Moiré pattern used to extract parameters<br />

<strong>for</strong> quality metrics is shown in Fig. 2. The pattern was created<br />

with a five inch square printed test ruling and a lenticular<br />

grating at a superposition angle indicated by the line at<br />

the top <strong>of</strong> the figure. The Moiré patterns result in wavy lines<br />

along an angled path. Their deviation from a straight path<br />

indicates a faulty printed test pattern that is due to the x<br />

perturbations <strong>of</strong> Eq. (3). For example, the lines in the top<br />

half <strong>of</strong> the figure deviate to the left <strong>of</strong> the expected straight<br />

path and then back again. This corresponds to an increase,<br />

and then a decrease <strong>of</strong> angle x, which corresponds to<br />

changes in x according to Eq. (4). This section describes<br />

how these changes can be extracted, characterized, and used<br />

to <strong>for</strong>m line quality metrics. The underlying line imperfections<br />

<strong>for</strong> two laser printers are illustrated in Figure 4. This<br />

figure compares two portions <strong>of</strong> the printed Ronchi ruling<br />

test pattern. The slices are 8.5 mm long and contain the top<br />

20 printed lines. The figure contains regular tick marks to<br />

indicate line numbers and their expected locations. Observe<br />

at the junction <strong>of</strong> the line sets that the line spacing is not<br />

consistent between the two printers. Line 1 is aligned <strong>for</strong><br />

both prints; however, lines between 6 and 19 do not align,<br />

and lines from printer LP1 (top line set) deviate from the<br />

ideal locations from line 6 onward. The printed lines from<br />

LP2 (bottom line set) also deviate from the ideal locations,<br />

but the deviations are less pronounced and only start to<br />

become large from line 14 onward, indicating the line registration<br />

quality <strong>of</strong> printer LP2 is higher than that <strong>of</strong> printer<br />

LP1. The metrics described in this section will correctly assess<br />

this difference from values extracted over the whole<br />

printed line pattern.<br />

The printed test pattern used <strong>for</strong> the results presented in<br />

this paper is a 55 in. square Ronchi ruling. A lenticular<br />

lens sheet is superimposed at an angle <strong>of</strong> around 4° and the<br />

resulting pattern is scanned at 600 dpi on an HP ScanJet<br />

C7710A flatbed scanner. The scanned image results in a<br />

3000 by 3000 pixel image, which yields 10 pixels per blackwhite<br />

line pair (corresponds to a density <strong>of</strong> 60 line pairs per<br />

inch or a line spacing <strong>of</strong> 0.4233 mm). The targets are<br />

scanned in a monochrome setting and cropped to 2048 by<br />

2048 pixels to limit scan edge effects. The lenticular lens<br />

sheet is a Pacur LENSTAR large <strong>for</strong>mat polyester sheet with<br />

40 lenticules per inch. The sheet is 0.033 in. thick, which<br />

also corresponds to the lenticular focal length. The lenticules<br />

have a radius <strong>of</strong> 0.0146, and a width <strong>of</strong> 0.0251 inches. For<br />

the analysis, the scan is low-pass filtered to emphasize the<br />

lower frequency Moiré patterns <strong>of</strong> interest, using a rotationally<br />

symmetric two-dimensional Gaussian correlation kernel<br />

<strong>of</strong> size 8 and standard deviation parameter <strong>of</strong> 8. Luminance<br />

variability from the scanner, which <strong>of</strong>ten affects banding<br />

metrics, <strong>for</strong> example, is mitigated using this approach because<br />

only the shape <strong>of</strong> the Moiré lines are used in the<br />

metric and not their intensity. The angle between the test<br />

pattern and lenticular grating was determined (near 4°) by<br />

eye to produce Moiré patterns <strong>of</strong> good visibility and measurability<br />

after scanning <strong>for</strong> analysis purposes.<br />

The analysis program extracts the contiguous pixel locations<br />

<strong>of</strong> the local minima (or constant gray-level) <strong>for</strong>ming<br />

a pattern vertically oriented over the page. The groups <strong>of</strong><br />

pixel locations associated with the Moiré patterns are characterized<br />

by a best-fit (least squares) line to pixel minima to<br />

obtain an estimate <strong>of</strong> the Moiré line corresponding to a<br />

perfect line pattern. The groups <strong>of</strong> pixels near the line corresponding<br />

to the actual patterns are identified and<br />

smoothed using a higher-order polynomial (order 32). Since<br />

multiple lines exist over the page, a search <strong>for</strong> local minima<br />

is per<strong>for</strong>med with a best-fit line to identify each Moiré pattern.<br />

To describe this process, denote the scanned Moiré<br />

pattern image as Ix n ,y m ,wherex n and y m respectively represent<br />

the discrete row and column positions <strong>of</strong> the image<br />

matrix. As illustrated in Fig. 2, the origin <strong>of</strong> this coordinate<br />

system is located at the top left pixel. The algorithm searches<br />

<strong>for</strong> Moiré lines by assuming the <strong>for</strong>m<br />

Rx n ;m,b = mx n + b,<br />

where Rx is the y coordinate <strong>of</strong> Moiré line, and m and b<br />

are the slope and y intercept, respectively. The line parameters<br />

are found through an exhaustive search over a range <strong>of</strong><br />

b and m values in order to minimize the following cost<br />

function:<br />

8<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 313


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

N<br />

Cm,b = Ix n ,Rx n ;m,b,<br />

n=1<br />

where N is the total number <strong>of</strong> pixel rows in the scanned<br />

image. The parameters b and m associated with the best-fit<br />

line can be determined by<br />

m 0 ,b 0 = arg minCm,b.<br />

m,b<br />

9<br />

10<br />

Once the best fit is found over the image, the slope m 0 is<br />

fixed and b is incremented over the column pixels <strong>of</strong> Ix,y<br />

and local minima detected to find the other patterns. Since<br />

the images are relatively simple, the minima appear with<br />

good distinction, and a threshold can be set to ignore insignificant<br />

minimum peaks and collect the set <strong>of</strong> b values corresponding<br />

to local minima denoted by<br />

B = b 1 ,b 2 ,b 3 ,b 4 ,b 5 , ...,b Y ,<br />

11<br />

such that B is a vector, <strong>of</strong> length Y, containing the intercept<br />

values <strong>of</strong> the lines that are the best linear fits to the actual<br />

Moiré lines. The average distance between the minima is<br />

taken as an estimate <strong>of</strong> the Moiré line spacing given by<br />

Pˆ = 1 Y−1<br />

b i+1 − b i .<br />

Y −1i=1<br />

12<br />

The actual curved Moiré pattern can be found by locating<br />

the local minimum <strong>for</strong> each x coordinate in the neighborhood<br />

<strong>of</strong> each fitted line. For some lenticular grids; however,<br />

the locally dark image points appear at the lens intersections,<br />

creating a regular discontinuity over the pattern. To improve<br />

the detection <strong>of</strong> the Moiré pattern pixel, a midluminance<br />

gray level was used. There<strong>for</strong>e, the y coordinates <strong>of</strong> the actual<br />

Moiré patterns were determined by the pixels closest to<br />

the Moiré pattern gray level I m in the neighborhood <strong>of</strong> the<br />

fitted line. The collection <strong>of</strong> points <strong>for</strong> the ith Moiré pattern<br />

is denoted as<br />

Pˆ<br />

2<br />

S i x =arg minIx,y − I m Rx;m 0 ,b i −<br />

y<br />

y Rx;m 0 ,b i + Pˆ<br />

2,<br />

13<br />

where I m is the mean luminance <strong>of</strong> the Moiré patterns. Figure<br />

5 illustrates the results <strong>of</strong> this extraction process <strong>for</strong> two<br />

sample laser printer outputs. A 32-order polynomial was fitted<br />

to the locus <strong>of</strong> points from Eq. (13) in order to smooth<br />

and overlay the Moiré patterns, along with the best fit lines<br />

<strong>for</strong> visual inspection, on the actual scanned image. With the<br />

approach described above the need <strong>for</strong> smoothing is important<br />

because <strong>of</strong> the periodic dark bands <strong>of</strong> the lens intersections<br />

cause regular glitches in the points. While other methods<br />

can be used <strong>for</strong> smoothing, such as the median filter, this<br />

work uses the 32-order polynomial fitted to the points identified<br />

by Eq. (13). The results observed in Fig. 5 demonstrate<br />

that the extraction procedure is indeed capturing the basic<br />

Figure 5. Moiré analysis comparison between two laser printers. Straight<br />

lines indicate the ideal Moiré patterns, and curved lines are best-fit polynomials<br />

to actual Moiré patterns due to printer errors. a Laser printer<br />

LP1. b Laser printer LP2.<br />

elements <strong>of</strong> the Moiré patterns.<br />

The deviation from all lines is characterized by a mean<br />

deviateateachrow,givenby<br />

Y<br />

Lx n = 1 S¯ix n − Rx n ;m 0 ,b i ,<br />

Y i=1<br />

14<br />

where S¯i is the resulting polynomial fit to the points <strong>of</strong> S i in<br />

Eq. (13). The derivative <strong>of</strong> Lx is equal to the tangent <strong>of</strong><br />

angle x; and <strong>for</strong> small values <strong>of</strong> x, can be estimated<br />

with a numerical gradient as follows:<br />

314 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

Table II. Statistical measures <strong>of</strong> Moiré pattern deviation quality metrics <strong>for</strong> two laser<br />

Table I. Line registration quality metrics <strong>for</strong> LP1 as a function <strong>of</strong> varying in<br />

5 2.57 2.28 10 −9 13.7 10 −5<br />

degrees.<br />

and one inkjet printers.<br />

Pˆ NAV NMD<br />

Printer Pˆ NAV NMD<br />

2 6.14 2.11 10 −9 13.1 10 −5<br />

LP1 3 4.26 2.08 10 −9 12.7 10 −5<br />

3 4.26 2.08 10 −9 12.7 10 −5<br />

LP2 3.8 3.73 1.79 10 −9 11.7 10 −5<br />

4 3.12 2.29 10 −9 12.9 10 −5<br />

IP 3.7 3.66 2.37 10 −9 11.4 10 −5<br />

x n tanx n 1<br />

2 Lx n+1 − Lx n−1 .<br />

15<br />

The estimate <strong>of</strong> the Moiré angle can be used to compute the<br />

macro quality metric referred to as the normalized average<br />

variance (NAV), given by<br />

1<br />

2 L = x n 2 .<br />

Pˆ 2 N −1 n=1<br />

N<br />

16<br />

Note this metric reflects the average line registration error<br />

over the whole test pattern. A micro quality line metric can<br />

be taken over local portions <strong>of</strong> the test pattern and involve<br />

the row corresponding to the worst deviation. This metric is<br />

called the normalized maximum deviation (NMD), and is<br />

given by<br />

¯ L = 1 Pˆ max x n .<br />

n<br />

17<br />

RESULTS AND ANALYSIS<br />

To demonstrate the robustness <strong>of</strong> the metrics to superposition<br />

angle variation, a pattern from laser printer LP1 was<br />

scanned <strong>for</strong> four different values, and the resulting NAV<br />

and NMD quality metrics are presented in Table I. It can be<br />

seen that the spacing decreases with increasing angle, while<br />

the NAV and NMD measures stay relatively constant. Quantitatively,<br />

the coefficients <strong>of</strong> variances (CV) <strong>for</strong> the metrics<br />

over the variations are 4.9% and 3.3% <strong>for</strong> the NAV and<br />

NMD, respectively. The CV is the ratio <strong>of</strong> the standard deviation<br />

to the mean <strong>of</strong> a data set, and it provides a quantity<br />

related to the measurement resolution, which is affected by<br />

factors such as printer and scanner settings, as well as, properties<br />

<strong>of</strong> the lenticular lenses used, such as lens spacing and<br />

precision.<br />

As an example <strong>of</strong> how the quality metrics respond to<br />

different printers, the NAV and NMD metrics were computed<br />

using the outputs <strong>of</strong> two laser printers (LP1 and LP2,<br />

used in Fig. 5), and an inkjet printer denoted as IP. Table II<br />

shows a numerical comparison between these printer outputs,<br />

as well as the measurement parameters. The CV values<br />

computed from Table I can be used to examine the relative<br />

comparison <strong>of</strong> line registration quality between printers. For<br />

example, the difference between the NAV values as a percentage<br />

<strong>of</strong> their mean is 15% <strong>for</strong> printers LP1 and LP2 in<br />

Table II. This value is greater than the 4.9% variation expected<br />

from the measurement variability and thus indicates<br />

that the large scale (macro) line registration quality <strong>of</strong> LP2 is<br />

better than that <strong>of</strong> LP1 (consistent with observations in Figs.<br />

4 and 5). In addition, the NMD measurements differ by<br />

8.3%, which is greater than the 3.3% CV <strong>for</strong> the NMD measure.<br />

Comparing the inkjet printer IP with LP1, it is evident<br />

from the NAV values that IP has poorer quality (consistent<br />

with examinations <strong>of</strong> scaled-up observation <strong>of</strong> the line quality);<br />

however, the NMD values differ by 10.8% <strong>of</strong> their<br />

mean, which is grater than the 3.3% CV value. This indicates<br />

that even though LP1 has better line registration on a<br />

macro scale (on average across the page), it has greater isolated<br />

deviations than IP.<br />

These results suggest that the above measures can serve<br />

as a quality metric <strong>for</strong> printed line registration. The NAV<br />

measure reflects the average printed line spacing P 0 constancy<br />

over the length <strong>of</strong> the test page. A quasiperiodic pattern<br />

in Lx n reflects banding like intensity variations across<br />

the test page as observed in Fig. 4 <strong>for</strong> LP1. These shape<br />

variations reflected periodic fluctuations in the printed line<br />

spacing, which are likely due to the same problems causing<br />

banding, such as imperfect rollers in the print process direction<br />

or gear noise. Moreover, Lx n can isolate process<br />

motion-related banding causes from other ones that affect<br />

reflectance, such as toner or ink deposition inconsistencies.<br />

CONCLUSION<br />

This work outlines the use <strong>of</strong> Moiré analysis <strong>for</strong> the quantification<br />

<strong>of</strong> line registration. The line registration metrics developed<br />

here are based on modeling the interference between<br />

a lenticular lens sheet and a hardcopy test target containing a<br />

Ronchi ruling or linear grating, and they provide examples<br />

<strong>of</strong> how the resulting Moiré patterns can be used to measure<br />

line registration quality. There are clearly other metrics that<br />

can be derived from the extracted Moiré patterns that can<br />

emphasize other issues depending on the application. For<br />

instance, if the Moiré line deviations are quasiperiodic, it is<br />

likely that these deviations indicate the root cause <strong>of</strong> banding.<br />

There<strong>for</strong>e, metrics based on the periodicity <strong>of</strong> these deviations<br />

over macro regions can be used <strong>for</strong> banding characterization.<br />

The work derived general equations to help in<br />

designing <strong>of</strong> metrics that have good repeatability.<br />

The experimental setup presented in this work suggests<br />

methods <strong>for</strong> volume processing <strong>of</strong> hardcopy samples. This<br />

would require a scanner with an automatic document feeder.<br />

A lenticular lens sheet could then be embedded into the<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 315


Rawashdeh et al.: Moiré analysis <strong>for</strong> assessment <strong>of</strong> line registration quality<br />

scanner glass or fixed on top <strong>of</strong> the glass where the test sheet<br />

could slide over it. The resulting patterns would then be<br />

scanned, and s<strong>of</strong>tware applied to compute the per<strong>for</strong>mance<br />

analyses as described in this paper. The monochrome and<br />

low-resolution scans are relatively easy to produce and analyze<br />

on a computer. A potential problem resulting from an<br />

automatic document feeder is maintaining a constant superposition<br />

angle between test page and lens sheet; however, it<br />

has been shown that the proposed metrics are robust to<br />

small changes in the angle. A more significant problem<br />

would arise from variations in the distance between the test<br />

pattern and lenticular sheet, such as might result from<br />

trapped air or irregular pressure on the test pattern. In this<br />

case the Moiré line will be artificially skewed causing variations<br />

from the distance rather than line mis-registration. It<br />

would be important in such a system to ensure the automatic<br />

feed (or any other system) minimizes this variation <strong>for</strong><br />

accurate metrics.<br />

REFERENCES<br />

1 ISO/IEC 13660:2001 In<strong>for</strong>mation technology - Office equipment:<br />

Measurement <strong>of</strong> image quality attributes <strong>for</strong> hardcopy output, Binary<br />

monochrome text and graphic images (ISO, Geneva), www.iso.org.<br />

2 E. N. Dalal, A. Haley, M. Robb, D. Mashtare, J. Briggs, P. L. Jeran, T. F.<br />

Bouk, and J. Deubert, “INCITS W1.1 Standards <strong>for</strong> Perceptual<br />

Evaluation <strong>of</strong> Text and Line Quality”, Proc. IS&T PICS Conference<br />

(IS&T, Springfield, VA, 2003) pp. 102–103.<br />

3 Y. Kipman, “Image quality metrics <strong>for</strong> printers and media”, Proc. IS&T<br />

PICS Conference (IS&T, Springfield, VA, 1998) pp. 183–187.<br />

4 G. J. Indebetouw and R. Czarnek, Selected Papers on Optical Moiré and<br />

Applications (SPIE, Bellingham, WA, 1992).<br />

5 B. Han, D. Post, and P. Ifju, “Moiré interferometry <strong>for</strong> engineering<br />

mechanics: current practices and future developments”, J. Strain Anal.<br />

Eng. Des. 36, 101–117 (2001).<br />

6 F. Zandman, G. S. Holister, and V. Brcic, “The influence <strong>of</strong> grid<br />

geometry on moire fringe properties”, J. Strain Anal. 1, 1-10 (1965).<br />

7 A. Livnat and O. Kafri, “Moire pattern <strong>of</strong> a linear grid with a lenticular<br />

grating”, Opt. Lett. 7, 253 (1982).<br />

316 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 317–327, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Analysis <strong>of</strong> the Influence <strong>of</strong> Vertical Disparities Arising<br />

in Toed-in Stereoscopic Cameras<br />

Robert S. Allison<br />

Department <strong>of</strong> Computer <strong>Science</strong> and Centre <strong>for</strong> Vision Research, York University,<br />

4700 Keele St., Toronto, Ontario M3J 1P3, Canada<br />

E-mail: allison@cs.yorku.ca<br />

Abstract. A basic task in the construction and use <strong>of</strong> a stereoscopic<br />

camera and display system is the alignment <strong>of</strong> the left and<br />

right images appropriately—a task generally referred to as camera<br />

convergence. Convergence <strong>of</strong> the real or virtual stereoscopic cameras<br />

can shift the range <strong>of</strong> portrayed depth to improve visual com<strong>for</strong>t,<br />

can adjust the disparity <strong>of</strong> targets to bring them nearer to the<br />

screen and reduce accommodation-vergence conflict, or can bring<br />

objects <strong>of</strong> interest into the binocular field <strong>of</strong> view. Although camera<br />

convergence is acknowledged as a useful function, there has been<br />

considerable debate over the trans<strong>for</strong>mation required. It is well<br />

known that rotational camera convergence or “toe-in” distorts the<br />

images in the two cameras producing patterns <strong>of</strong> horizontal and<br />

vertical disparities that can cause problems with fusion <strong>of</strong> the stereoscopic<br />

imagery. Behaviorally, similar retinal vertical disparity patterns<br />

are known to correlate with viewing distance and strongly affect<br />

perception <strong>of</strong> stereoscopic shape and depth. There has been<br />

little analysis <strong>of</strong> the implications <strong>of</strong> recent findings on vertical disparity<br />

processing <strong>for</strong> the design <strong>of</strong> stereoscopic camera and display<br />

systems. I ask how such distortions caused by camera convergence<br />

affect the ability to fuse and perceive stereoscopic images. © 2007<br />

<strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4317<br />

INTRODUCTION<br />

In many stereoscopic viewing situations it is necessary to<br />

adjust the screen disparity <strong>of</strong> the displayed images <strong>for</strong> viewer<br />

com<strong>for</strong>t, to optimize depth perception or to otherwise enhance<br />

the stereoscopic experience. Convergence <strong>of</strong> the real<br />

or virtual cameras is an effective means <strong>of</strong> adjusting portrayed<br />

disparities. A long-standing question in the stereoscopic<br />

imaging and display literature is what is the best<br />

method to converge the cameras? Humans use rotational<br />

movements to binocularly align the visual axes <strong>of</strong> their eyes<br />

on targets <strong>of</strong> interest. Similarly, one <strong>of</strong> the easiest ways to<br />

converge the cameras is to pan them in opposite directions<br />

to “toe-in” the cameras. However, convergence through<br />

camera toe-in has side effects that can lead to undesirable<br />

distortions <strong>of</strong> stereoscopic depth. 1,2 In this paper we reanalyze<br />

these geometric distortions <strong>of</strong> stereoscopic space in the<br />

context <strong>of</strong> recent findings on the role <strong>of</strong> vertical disparities in<br />

stereoscopic space perception. We focus on a number <strong>of</strong> issues<br />

related to converged cameras and the mode <strong>of</strong> convergence:<br />

The effect <strong>of</strong> rectification; relation between the geometry<br />

<strong>of</strong> the imaging device and the display device; fused and<br />

Received Dec. 5, 2006; accepted <strong>for</strong> publication Mar. 7, 2007.<br />

1062-3701/2007/514/317/11/$20.00.<br />

augmented displays; orthostereoscopy; the relation between<br />

parallax distortions in the display and the resulting retinal<br />

disparity; and the effect <strong>of</strong> these toe-in induced retinal disparities<br />

on depth perception and binocular fusion.<br />

Our interests lie in augmented-reality applications and<br />

stereoscopic heads <strong>for</strong> tele-operation applications. In these<br />

systems a focus is on the match and registration between the<br />

stereoscopic imagery and the “real world” so we will concentrate<br />

on orthostereoscopic or near orthostereoscopic configurations.<br />

These configurations have well known limitations<br />

<strong>for</strong> applications such as visualization and cinema, and<br />

other configurations may result in displays that are more<br />

pleasing and easier to fuse. However, it is important to note<br />

that our basic analysis generalizes to other configurations,<br />

and we will discuss other viewing arrangements when<br />

appropriate. 3,4 In a projector-based display system with separate<br />

right and left projectors, or in binocular head mounted<br />

display (HMD) with independent left and right displays, the<br />

displays/projectors can also be converged mechanically or<br />

optically. In this paper we will also assume a single flat,<br />

fronto-parallel display (i.e., a monitor or projector display)<br />

so that the convergence <strong>of</strong> the projectors is not an issue.<br />

Since the left and right images are projected or displayed<br />

into the same plane we will refer to these configurations as a<br />

“parallel display.” In most cases similar considerations will<br />

apply <strong>for</strong> a HMD with parallel left and right displays.<br />

OPTIONS FOR CAMERA CONVERGENCE<br />

We use the term convergence here to refer to a variety <strong>of</strong><br />

means <strong>of</strong> realigning one stereoscopic half-image with respect<br />

to the other, including toe-in (or rotational) convergence<br />

and translational image shift.<br />

Convergence can shift the range <strong>of</strong> portrayed depth to<br />

improve visual com<strong>for</strong>t and composition. Looking at objects<br />

presented stereoscopically further or nearer than the screen<br />

causes a disruption <strong>of</strong> the normal synergy between vergence<br />

and accommodation in most displays. Normally accommodation<br />

and vergence covary but, in a stereoscopic display, the<br />

eyes should remain focused at the screen regardless <strong>of</strong> disparity.<br />

The accommodation-vergence conflict can cause visual<br />

stress and disrupt binocular vision. 5 Convergence <strong>of</strong> the<br />

cameras can be used to adjust the disparity <strong>of</strong> targets <strong>of</strong><br />

interest to bring them nearer to the screen and reduce this<br />

conflict.<br />

317


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Table I. Typical convergence <strong>for</strong> stereoscopic sensors and displays. “Natural” modes <strong>of</strong><br />

convergence are shown in bold.<br />

DISPLAY/SENSOR<br />

GEOMETRY<br />

Translation<br />

REAL OR VIRTUAL CAMERA CONVERGENCE<br />

Rotation<br />

Flat Horizontal Image Translation Toed-in camera, toed-in<br />

projector combination<br />

Spherical<br />

Differential translation <strong>of</strong><br />

computer graphics images<br />

Image sensor shift<br />

Variable baseline camera<br />

Human viewing <strong>of</strong> planar<br />

stereoscopic displays?<br />

Toed-in stereoscopic camera<br />

or robot head<br />

Haploscope<br />

Human physiological<br />

vergence<br />

Convergence can also be used to shift the range <strong>of</strong> portrayed<br />

depth. For example, it is <strong>of</strong>ten preferable to portray<br />

stereoscopic imagery in the space behind rather than in front<br />

<strong>of</strong> the display. With convergence a user can shift stereoscopic<br />

imagery to appear “inside” the display and reduce interposition<br />

errors between the stereoscopic imagery and the edges<br />

<strong>of</strong> the displays.<br />

Cameras used in stereoscopic imagers have limited field<br />

<strong>of</strong> view and convergence can be used to bring objects <strong>of</strong><br />

interest into the binocular field <strong>of</strong> view.<br />

Finally, convergence or more appropriately translation<br />

<strong>of</strong> the stereoscopic cameras can also be used to adjust <strong>for</strong><br />

differences in a user’s interpupillary distance. The latter<br />

trans<strong>for</strong>mation is not typically called convergence since the<br />

stereoscopic baseline is not maintained.<br />

In choosing a method <strong>of</strong> convergence there are several<br />

issues one needs to consider. What type <strong>of</strong> 2D image trans<strong>for</strong>mation<br />

is most natural <strong>for</strong> the imaging geometry? Can a<br />

3D movement <strong>of</strong> the imaging device accomplish this trans<strong>for</strong>mation?<br />

In a system consisting <strong>of</strong> separate acquisition and<br />

display systems is convergence best achieved by changing the<br />

imaging configuration and/or by trans<strong>for</strong>ming the images<br />

(or projector configuration) prior to display? If an unnatural<br />

convergence technique must be used, what is the impact on<br />

stereoscopic depth perception?<br />

Although camera convergence is acknowledged as a useful<br />

function, there has been considerable debate over the<br />

correct trans<strong>for</strong>mation required. Since the eyes (and the<br />

cameras in imaging applications) are separated laterally, convergence<br />

needs to be an opposite horizontal shift <strong>of</strong> left and<br />

right eyes images on the sensor surface or, equivalently, on<br />

the display. The most appropriate type <strong>of</strong> trans<strong>for</strong>mation to<br />

accomplish this 2D shift—rotation or translation—depends<br />

on the geometry <strong>of</strong> the imaging and display devices. We<br />

agree with the view that the trans<strong>for</strong>mation should reflect<br />

the geometry <strong>of</strong> the display and imaging devices in order to<br />

minimize distortion (see Table I). One could argue that a<br />

“pure” vergence movement should affect the disparity <strong>of</strong> all<br />

objects equally, resulting in a change in mean disparity over<br />

the entire image without any change in relative disparity<br />

between points.<br />

For example, consider a spherical imaging device such<br />

as the human eye where expressing disparity in terms <strong>of</strong><br />

visual angle is a natural coding scheme. A rotational movement<br />

about the optical centre <strong>of</strong> the eye would scan an<br />

image over the retina without distorting the angular relationships<br />

within the image. Thus the natural convergence<br />

movement with such an imaging device is a differential rotation<br />

<strong>of</strong> the two eyes, as occurs in physiological convergence<br />

(although freedom to choose various spherical coordinate<br />

systems complicates the definition <strong>of</strong> disparity 6 ).<br />

A flat sensor is the limiting <strong>for</strong>m <strong>of</strong> spherical sensor<br />

with an infinite radius <strong>of</strong> curvature, and thus the rotation <strong>of</strong><br />

the sensor becomes a translation parallel to the sensor plane.<br />

For displays that rely on projection onto a single flat, frontoparallel<br />

display surface (many stereoscopic displays with the<br />

notable exception <strong>of</strong> some head-mounted displays and haploscopic<br />

systems) depth differences should be represented as<br />

linear horizontal disparities in the image plane. The natural<br />

convergence movement is a differential horizontal shift <strong>of</strong><br />

the images in the plane <strong>of</strong> the display. Acquisition systems<br />

with parallel cameras are well-matched to such display geometry<br />

since a translation on the display corresponds to a<br />

translation in the sensor plane. This model <strong>of</strong> parallel cameras<br />

is typically used <strong>for</strong> the virtual cameras in stereoscopic<br />

computer graphics 7 and the real cameras in many stereoscopic<br />

camera setups.<br />

Thus horizontal image translation <strong>of</strong> the images on the<br />

display is the preferred minimal distortion method to shift<br />

convergence in a stereoscopic rig with parallel cameras when<br />

presented on a parallel display. This analysis corresponds to<br />

current conventional wisdom. If the stereo baseline is to be<br />

maintained then this vergence movement is a horizontal<br />

translation <strong>of</strong> the images obtained from the parallel cameras<br />

rather than a translation <strong>of</strong> the cameras themselves. For example,<br />

in computer-generated displays, the left and right half<br />

images can be shifted in opposite directions on the display<br />

surface to shift portrayed depth with respect to the screen.<br />

With real camera images, a problem with shifting the displayed<br />

images to accomplish convergence is that in doing so,<br />

part <strong>of</strong> each half-image is shifted <strong>of</strong>f <strong>of</strong> the display resulting<br />

in a smaller stereoscopic image.<br />

An alternative is to shift the imaging device (e.g., CCD<br />

array) behind the camera lens, with opposite sign <strong>of</strong> shift in<br />

the two cameras <strong>for</strong>ming the stereo rig. This avoids some <strong>of</strong><br />

the problems associated with rotational convergence discussed<br />

below. Implementing a large, variable range <strong>of</strong> convergence<br />

with mechanical movements or selection <strong>of</strong> subarrays<br />

from a large CCD can be complicated. Furthermore,<br />

many lenses have significant radial distortion and translating<br />

the center <strong>of</strong> the imaging device away from the optical axis<br />

increases the amount <strong>of</strong> radial distortion. Worse, <strong>for</strong><br />

matched lenses the distortions introduced in each sensor<br />

image will be opposite if the sensors are shifted in opposite<br />

directions. This leads to increased disparity distortion.<br />

Toed-in cameras can center the image on the optical axis<br />

and reduce this particular problem.<br />

If we converge nearer than infinity using horizontal im-<br />

318 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 1. A plan view <strong>of</strong> an array <strong>of</strong> points located in the X-Z plane at<br />

eye level. The solid dots show the true position <strong>of</strong> the points and also their<br />

reconstruction based on images from a parallel camera orthostereoscopic<br />

rig presented at a 0.7 m viewing distance. The open diamond shaped<br />

markers show the reconstructed position <strong>of</strong> the points in the array when<br />

the cameras are converged using horizontal image translation HIT. As<br />

predicted the points that are truly at 1.1 m move in to appear near the<br />

screen distance <strong>of</strong> 0.7 m. Also depth and size should appear scaled<br />

appropriately <strong>for</strong> the nearer distance. But notice that depth ordering and<br />

planarity are maintained. Circles at a distance <strong>of</strong> zero denote the positions<br />

<strong>of</strong> the eyes.<br />

age shift, then far objects should be brought toward the<br />

plane <strong>of</strong> the screen. With convergence via horizontal image<br />

shift, a frontal plane at the camera convergence distance<br />

should appear flat and at the screen distance. However,<br />

depth <strong>for</strong> a given retinal disparity increases approximately<br />

with the square <strong>of</strong> distance. Thus if the cameras are converged<br />

at a distance other than the screen distance to bring a<br />

farther (or nearer) target toward the screen, then the depth<br />

in the scene should be distorted nonlinearly but depth ordering<br />

and planarity are maintained (Figure 1). This apparent<br />

depth distortion is predicted <strong>for</strong> both the parallel and<br />

toed-in configurations. In the toed-in case it would be added<br />

to the curvature effects discussed below. Similar arguments<br />

can be made <strong>for</strong> size distortions in the image (or equivalently<br />

the apparent spacing <strong>of</strong> the dots in Fig. 1). See Woods 1<br />

and Diner and Fender 2 <strong>for</strong> an extended discussion <strong>of</strong> these<br />

distortions.<br />

It is important to note that these effects are predicted<br />

from the geometry and do not always correspond to human<br />

perception. Percepts <strong>of</strong> stereoscopic space tend to deviate<br />

from the geometric predictions based on the Keplerian projections<br />

and Euclidean geometry 6 ). Vergence on its own is<br />

not a strong cue to distance and other depth cues in the<br />

display besides horizontal disparity can affect the interpretation<br />

<strong>of</strong> stereoscopic displays. For example, it has been<br />

known <strong>for</strong> over 100 years that observers can use vertical<br />

disparities in the stereoscopic images to obtain more veritical<br />

estimates <strong>of</strong> stereoscopic <strong>for</strong>m. 8 In recent years, a role <strong>for</strong><br />

vertical disparities in human stereoscopic depth perception<br />

has been confirmed. 9,10<br />

Translation <strong>of</strong> the images on the display or <strong>of</strong> the sensors<br />

behind the lenses maintains the stereoscopic camera<br />

baseline and hence the relative disparities in the acquired or<br />

simulated image. Shifting <strong>of</strong> the images can be used to shift<br />

this disparity range to be centered on the display to ease<br />

viewing com<strong>for</strong>t. However, in many applications this disparity<br />

range is excessive and other techniques may be more<br />

suitable. Laterally shifting the cameras toward or away from<br />

each other increases or decreases the range <strong>of</strong> disparities<br />

corresponding to a given scene. Control <strong>of</strong> the stereo rig<br />

baseline serves a complementary function to convergence by<br />

adjusting the “gain” <strong>of</strong> stereopsis instead <strong>of</strong> simply the mean<br />

disparity. This function is <strong>of</strong>ten very useful <strong>for</strong> mapping a<br />

depth range to a useful or com<strong>for</strong>table disparity range in<br />

applications such as computer graphics, 4,11 photogrammetry,<br />

etc.<br />

In augmented reality or other enhanced vision systems<br />

that fuse stereoscopic imagery with direct views <strong>of</strong> the world<br />

(or with displays from other stereoscopic image sources),<br />

orthostereoscopic configurations (or at least consistent<br />

views) are important. In these systems, proper convergence<br />

<strong>of</strong> the camera systems and calibration <strong>of</strong> image geometry is<br />

required so that objects in the display have appropriate disparity<br />

relative to their real world counterparts. A parallel<br />

camera orthostereoscopic configuration presents true disparities<br />

to the user if presented on a parallel display. Thus,<br />

geometrically at least, we should expect to see true depth. In<br />

practice this seldom occurs because <strong>of</strong> the influence <strong>of</strong> other<br />

depth cues (accommodation-vergence conflict, changes in<br />

effective interpupillary distance with eye movements, flatness<br />

cues corresponding to viewing a flat display, etc.).<br />

In summary, an orthostereoscopic parallel-camera/<br />

parallel-display configuration can present accurate disparities<br />

to the user. 1,7 On parallel displays, convergence by horizontal<br />

shift <strong>of</strong> the images obtained from parallel cameras<br />

introduces no distortion <strong>of</strong> horizontal or vertical screen disparity<br />

(parallax). Essentially, convergence by this method<br />

brings the two half images into register with out changing<br />

relative disparity. This can reduce vergence-accommodation<br />

conflict and improve the ability to fuse the imagery. Geometrically,<br />

one would predict effects on perceived depth—<br />

the apparent depth <strong>of</strong> imagery with respect to the screen and<br />

the depth scaling in the image are affected by the simulated<br />

vergence. 1,13 However, this amounts to a relief trans<strong>for</strong>mation<br />

implying that depth ordering and coplanarity should be<br />

maintained. 2,10<br />

CAMERA TOE-IN<br />

While horizontal image translation is attractive theoretically,<br />

there are <strong>of</strong>ten practical considerations that limit use <strong>of</strong> the<br />

method and make rotational convergence attractive. For example,<br />

with a limited camera field <strong>of</strong> view and a nonzero<br />

stereo baseline there exists a region <strong>of</strong> space near to the<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 319


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 2. a The Toronto IRIS Stereoscopic Head 2 TRISH II, an example <strong>of</strong> a robot head built <strong>for</strong> a wide<br />

range <strong>of</strong> working distances. With such a system, a wide range <strong>of</strong> camera convergence is required to bring<br />

objects <strong>of</strong> interest into view <strong>of</strong> the cameras. With <strong>of</strong>f-the shelf cameras this can be most conveniently achieved<br />

with camera toe-in. b A hypothetical stereo rig with camera field <strong>of</strong> view . Objects in near working space<br />

are out <strong>of</strong> the binocular field <strong>of</strong> view which is indicated by the cross hatch pattern.<br />

cameras that cannot be seen by one or both cameras. In<br />

some applications such as landscape photography this region<br />

<strong>of</strong> space may be irrelevant; in other applications such as<br />

augmented reality or stereoscopic robot heads this may correspond<br />

to a crucial part <strong>of</strong> the normal working range (see<br />

Figure 2). Rotational convergence <strong>of</strong> the cameras can increase<br />

the near working space <strong>of</strong> the system and center the<br />

target in the camera images. 14 Other motivations <strong>for</strong> rotational<br />

convergence include the desire to center the target on<br />

the camera optics (e.g., to minimize camera distortion) and<br />

the relative simplicity and large range <strong>of</strong> motion possible<br />

with rotational mechanisms. Given that rotational convergence<br />

<strong>of</strong> stereo cameras is <strong>of</strong>ten implemented in practice, we<br />

ask what effects the distortions produced by these movements<br />

have on the perception <strong>of</strong> stereoscopic displays?<br />

It is well known that the toed-in configuration distorts<br />

the images in the two cameras producing patterns <strong>of</strong> horizontal<br />

and vertical screen disparities (parallax). Geometrically,<br />

deviations from the parallel-camera configuration may<br />

result in spatial distortion unless compensating trans<strong>for</strong>mations<br />

are introduced mechanically, optically or electronically<br />

in the displayed images, 2,12 <strong>for</strong> example unless a pair <strong>of</strong> projectors<br />

(or HMD with separate left and right displays) with<br />

matched convergence or a parallel display with special distortion<br />

correction techniques are used. 15,16 For the rest <strong>of</strong><br />

this paper we will assume a single projector or display system<br />

(parallel display) and a dual sensor system with parallel<br />

or toed-in cameras.<br />

The effects <strong>of</strong> the horizontal disparities have been well<br />

described in the literature and we review them be<strong>for</strong>e turning<br />

to the vertical disparities in the next section. The depth<br />

distortions due to the horizontal disparities introduced can<br />

be estimated geometrically. 1 The geometry <strong>of</strong> the situation is<br />

illustrated in Figure 3. The imaging space world coordinate<br />

system is centered between the cameras, a is the intercamera<br />

distance and the angle <strong>of</strong> convergence is (using the conventional<br />

stereoscopic camera measure <strong>of</strong> convergence rather<br />

than the physiological one).<br />

Let us assume the cameras converge symmetrically at<br />

point C located at distance F. A local coordinate system is<br />

attached to each camera and rotated ± about the y axis<br />

with respect to the imaging space world coordinate system.<br />

The coordinates <strong>of</strong> a point P=XYZ T in the left and right<br />

cameras is<br />

X<br />

l= X + a − Z sin<br />

l<br />

2cos<br />

Y l<br />

Y<br />

Z<br />

Z cos +X +<br />

2sin,<br />

a<br />

X<br />

r= X − a + Z sin<br />

r<br />

2cos<br />

Y r<br />

Y<br />

Z<br />

Z cos −X −<br />

2sin.<br />

a<br />

After perspective projection onto the converged CCD array<br />

(coordinate frame u-v centered on the optic axis and letting<br />

f=1.0) we get the following image coordinates <strong>for</strong> the left,<br />

u l ,v l T , and right, u r ,v r T ,arrays:<br />

= X + a − Z sin<br />

2cos<br />

u l<br />

v l= X l/Z l Z cos +X +<br />

Y l /Z l 2sin<br />

a<br />

Y<br />

Z cos +X +<br />

2sin,<br />

a<br />

1<br />

2<br />

320 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 3. <strong>Imaging</strong> and display geometry <strong>for</strong> symmetric toe-in convergence at point C and viewing at distance<br />

D plan view.<br />

Figure 4. Keystone distortion due to toe-in. a Left + and right images <strong>for</strong> a regularly spaced grid <strong>of</strong><br />

points with the stereo camera converged toed-in on the grid. b Corresponding disparity vectors comparing<br />

left eye with right eye views demonstrate both horizontal and vertical components <strong>of</strong> the keystone distortion.<br />

= X − a + Z sin<br />

2cos<br />

u r<br />

v r= X r/Z r Z cos −X −<br />

Y r /Z r 2sin<br />

a<br />

Y<br />

Z cos −X −<br />

2sin.<br />

a<br />

The CCD image is then reprojected onto the display screen.<br />

We assume a single display/projector model with central<br />

projection and a magnification <strong>of</strong> M with respect to the<br />

CCD sensor image resulting in the following screen coordinates<br />

<strong>for</strong> the point in the left, U l ,V l T , and right, U r ,V r T ,<br />

eye images:<br />

U l<br />

V l = M u l<br />

v l,<br />

U r<br />

V r = M u r<br />

v r.<br />

Toeing-in the stereoscopic rig to converge on a surface centers<br />

the images <strong>of</strong> the target in the two cameras but also<br />

introduces a keystone distortion due to the differential perspective<br />

(Figure 4). In contrast convergence by shifting the<br />

CCD sensor behind the camera lens (or shifting the half<br />

images on the display) changes the mean horizontal disparity<br />

but does not entail keystone distortion. For a given focal<br />

length and camera separation, the extent <strong>of</strong> the keystone<br />

distortion is a function <strong>of</strong> the convergence distance and not<br />

the distance <strong>of</strong> the target.<br />

To see how the keystoning affects depth perception, assume<br />

the images are projected onto a screen at distance D<br />

and viewed by a viewer with interocular distance <strong>of</strong> e. If the<br />

magnification from the CCD sensor array to screen image is<br />

3<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 321


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 5. Geometrically predicted perception curved grid <strong>of</strong> displayed<br />

images taken from a toed-in stereoscopic camera rig converged on a<br />

fronto-parallel grid made with 10 cm spacing asterisks based on horizontal<br />

disparities associated size distortion not shown. Camera convergence<br />

distance F and display viewing distance D are 0.70 cm<br />

e=a=62.5 mm; f=6.5 mm; see Fig. 3 and text <strong>for</strong> definitions. The<br />

icon at the bottom <strong>of</strong> the figure indicates the position <strong>of</strong> the world coordinate<br />

frame and the eyeballs.<br />

M and both images are centered on the display then geometrically<br />

predicted coordinates <strong>of</strong> the point in display space<br />

is (after Ref. 1)<br />

d<br />

P d =X d=<br />

eU l + U r <br />

2e − U r − U l <br />

eV l + V r<br />

<br />

<br />

Y d<br />

4<br />

2e − U r − U l <br />

Z<br />

eD<br />

e − U r − U l <br />

where U r −U l is the horizontal screen parallax <strong>of</strong> the point.<br />

If we ignore vertical disparities <strong>for</strong> the moment, converging<br />

the camera causes changes in the geometrically predicted<br />

depth. For instance, if the cameras toe-in to converge<br />

on a frontoparallel surface (parallel to the stereobaseline),<br />

then from geometric considerations the center <strong>of</strong> the object<br />

should appear at the screen distance but the surface should<br />

appear curved (Figure 5). This curvature should be especially<br />

apparent in the presence <strong>of</strong> undistorted stereoscopic<br />

reference imagery as would occur in augmented reality<br />

applications. 16 In contrast, if convergence is accomplished<br />

via horizontal image translation then a frontal plane at the<br />

camera convergence distance should appear flat and at the<br />

screen distance although depth and size will be scaled as<br />

discussed in the previous section.<br />

USE OF VERTICAL DISPARITY IN STEREOPSIS<br />

The pattern <strong>of</strong> vertical disparities in a stereoscopic image<br />

depends on the geometry <strong>of</strong> the stereoscopic rig. With our<br />

spherical retinas disparity is best defined in terms <strong>of</strong> visual<br />

angle. An object that is located eccentric to the median plane<br />

<strong>of</strong> the head is closer to one eye than the other (Figure 6).<br />

Hence, it subtends a larger angle at the nearer eye than at the<br />

further. The vertical size ratio (VSR) between the images <strong>of</strong><br />

an object in the two eyes varies as a function <strong>of</strong> the object’s<br />

eccentricity with respect to the head. Figure 6 also shows the<br />

variation <strong>of</strong> the vertical size ratio <strong>of</strong> the right eye image to<br />

the left eye image <strong>for</strong> a range <strong>of</strong> eccentricities and<br />

distances.<br />

It is evident that, <strong>for</strong> centrally located targets, the gradient<br />

<strong>of</strong> vertical size ratios varies with distance <strong>of</strong> the surface<br />

from the head. This is relatively independent <strong>of</strong> the vergence<br />

state <strong>of</strong> the eyes and the local depth structure. 17 Howard 18<br />

turned this relationship around and suggested that people<br />

could judge the distance <strong>of</strong> surfaces from the gradient <strong>of</strong> the<br />

VSR. Gillam and Lawergren 19 proposed a computational<br />

model <strong>for</strong> the recovery <strong>of</strong> surface distance and eccentricity<br />

based upon processing <strong>of</strong> VSR and VSR gradients. An alternative<br />

computational framework 10,20 uses vertical disparities<br />

to calculate the convergence posture and gaze eccentricity <strong>of</strong><br />

the eyes rather than the distance and eccentricity <strong>of</strong> a target<br />

surface. For our purposes, these models make the same predictions<br />

about the effects <strong>of</strong> camera toe-in. However, the<br />

latter model uses projections onto flat projection surfaces<br />

(hypothetical flat retinae) which is easier <strong>for</strong> visualization<br />

and matches well with our previous discussion <strong>of</strong> camera<br />

toe-in.<br />

With flat imaging planes, disparities are usually measured<br />

in terms <strong>of</strong> linear displacement in the image plane. If<br />

the cameras in a stereoscopic rig are toed in (or if eyes with<br />

flat retinae are converged), then the left and right camera<br />

images have opposite keystone distortion. It is interesting to<br />

note that in contrast to the angular disparity case the gradients<br />

<strong>of</strong> vertical disparities are a function <strong>of</strong> camera convergence<br />

but are affected little by the distance <strong>of</strong> the surface.<br />

These vertical disparity gradients on flat cameras/retinae<br />

provide an indication <strong>of</strong> the convergence angle <strong>of</strong> the cameras<br />

and hence the distance <strong>of</strong> the fixation point.<br />

For a pair <strong>of</strong> objects or <strong>for</strong> depth within an object, the<br />

relationship between relative depth and relative disparity is a<br />

function <strong>of</strong> distance from the observer. To an extent, the<br />

visual system is able to maintain an accurate perception <strong>of</strong><br />

depth <strong>of</strong> an object at various distances despite disparity<br />

varying inversely with the square <strong>of</strong> the distance between the<br />

object and the observer. This “depth constancy” demonstrates<br />

an ability to account <strong>for</strong> the effects <strong>of</strong> viewing distance<br />

on stereoscopic depth. The relationship between the<br />

retinal image size <strong>of</strong> an object and its linear size in the world<br />

is also a function <strong>of</strong> distance. To the degree that vertical<br />

disparity gradients are used as an indicator <strong>of</strong> the distance <strong>of</strong><br />

a fixated surface <strong>for</strong> three-dimensional reconstruction, toe-in<br />

produced vertical disparity gradients would be expected to<br />

indirectly affect depth and size perception. Psychophysical<br />

experiments have demonstrated that vertical disparity gradients<br />

strongly affect perception <strong>of</strong> stereoscopic shape, size and<br />

depth 9,10,21 and implicate vertical disparity processing in human<br />

size and depth constancy.<br />

322 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 6. a A vertical line located eccentric to the midline <strong>of</strong> the head is nearer to one eye than the other.<br />

Thus it subtends a larger angle in the nearer eye than the further adapted from Howard and Rogers 6 . b The<br />

gradient <strong>of</strong> vertical size ratio <strong>of</strong> the image <strong>of</strong> a surface element in the left eye to that in the right eye varies as<br />

a function <strong>of</strong> distance <strong>of</strong> the surface shown as a series <strong>of</strong> lines: distances <strong>of</strong> 70, 60, 50, 40, and 30 cm in<br />

order <strong>of</strong> steepness.<br />

VERTICAL DISPARITY IN TOED-IN STEREOSCOPIC<br />

CAMERAS<br />

First, consider a stereoscopic camera and parallel display system<br />

that intends to portray realistic depth and that has camera<br />

separation equal to the eye separation. If the camera is<br />

converged using the toe-in method at a fronto-parallel surface<br />

at the distance <strong>of</strong> the screen, then the center <strong>of</strong> the<br />

target will have zero horizontal screen disparity. However,<br />

the camera toe-in will introduce keystone distortion into the<br />

two images with the pattern <strong>of</strong> horizontal disparities predicting<br />

curvature as discussed above. What about the pattern <strong>of</strong><br />

vertical disparities? The pattern <strong>of</strong> vertical disparities produced<br />

by a toed-in camera configuration resembles the gradient<br />

<strong>of</strong> vertical size disparities on the retinae that can arise<br />

due to differential perspective <strong>of</strong> the two eyes. As discussed<br />

in the previous section, this differential perspective <strong>for</strong>ms a<br />

natural and rich source <strong>of</strong> in<strong>for</strong>mative parameters contributing<br />

to human stereoscopic depth perception.<br />

Given that camera toe-in generates such gradients <strong>of</strong><br />

vertical disparity in stereoscopic imagery, is it beneficial to<br />

use camera toe-in to provide distance in<strong>for</strong>mation in a stereoscopic<br />

display? In other words, should the toed-in configuration<br />

be used to converge the cameras and preserve the<br />

sense <strong>of</strong> absolute distance and size, shape and depth constancy?<br />

Perez-Bayas 22 argued that toed-in camera configurations<br />

are more natural since they present these vertical disparities.<br />

The principal problem with this claim is that it<br />

considers the screen parallax <strong>of</strong> stereoscopic images rather<br />

than their retinal disparities. These keystone distortions are<br />

in addition to the natural retinal vertical disparities present<br />

when viewing a scene at the distance <strong>of</strong> the screen.<br />

In order to estimate the effect on depth perception we<br />

need to consider the retinal disparities generated by the stereoscopic<br />

image. The keystone distortion occurs in addition<br />

to the retinal vertical disparity pattern inherent in the image<br />

because it is portrayed on the flat screen. Consider a frontoparallel<br />

surface located at the distance <strong>of</strong> the screen away<br />

from the camera and that we intend to display the surface at<br />

the screen. Projections onto spherical retinas are hard to<br />

visualize so let us consider flat retinae converged (toed-in) at<br />

the screen distance. Alternatively one could imagine another<br />

pair <strong>of</strong> converged cameras viewing the display, one centered<br />

at the center <strong>of</strong> each eye. The images on these converged flat<br />

retinae would <strong>of</strong> course have differential keystone distortion<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 323


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 7. a Simulation <strong>of</strong> the keystone distortion and gradient <strong>of</strong> VSR<br />

present in a stereo half image <strong>for</strong> a toed-in configuration. The plus symbols<br />

show the keystone distortion in the displayed image <strong>of</strong> a grid <strong>for</strong> a<br />

camera converged at 70 cm and the circle symbols indicated the exaggerated<br />

VSR distortion present in the retinal half image <strong>for</strong> an observer<br />

viewing the display at 70 cm flat retina. b Predicted distorted appearance<br />

circles in a set <strong>of</strong> frontal plane surfaces asterisks if depth from<br />

disparity is scaled according to the distance indicated by an exaggerated<br />

VSR. Typically the surface is not mislocalized in depth but curvature is<br />

induced. The predicted curvature based on the on the equations provided<br />

by Duke and Wilcox 28 is also shown diamonds. The simulated positions<br />

<strong>of</strong> the eyes are denoted by circles at zero distance and the screen by a<br />

line at 70 cm.<br />

when viewing a frontal surface such as the screen. When<br />

displaying images from the toed-in stereoscopic camera,<br />

which already have keystone distortion, the result is an exaggerated<br />

gradient <strong>of</strong> vertical disparity in the retinal images<br />

appropriate <strong>for</strong> a much nearer surface. For a spherical retina<br />

the important measure is the gradient <strong>of</strong> vertical size ratios<br />

in the image. The vertical size ratios in the displayed images<br />

imposed by the keystone distortion are in addition to the<br />

natural VSR <strong>for</strong> a frontal surface at the distance <strong>of</strong> the<br />

screen. Clearly, the additional keystone distortion indicates a<br />

nearer surface in this case as well [Figure 7(a)].<br />

From either the flat camera or spherical retina model we<br />

predict spatial distortion if disparities are scaled according to<br />

the vertical disparities, which indicate a closer target. Such a<br />

misjudgement <strong>of</strong> perceived distance would be predicted to<br />

have effects on perceived depth and size [open circles in Fig.<br />

7(b)]. There is little evidence that observers actually<br />

mislocalize surfaces at a nearer distance when a vertical disparity<br />

gradient is imposed. However, there is strong evidence<br />

<strong>for</strong> effects <strong>of</strong> VSR gradients on depth constancy processes.<br />

If a viewer fixates a point on a fronto-parallel screen,<br />

then at all screen distances nearer than infinity the images <strong>of</strong><br />

other points on the screen have horizontal disparity (retinal<br />

but not screen disparity). This is because the theoretical locus<br />

<strong>of</strong> points in three-dimensional space with zero retinal<br />

disparity, which is known as the horopter (the Vieth-Muller<br />

circle), curves inward toward the viewer and away from the<br />

frontal plane. The curvature <strong>of</strong> the horopter increases at<br />

nearer distances (Figure 8). 23 Thus a frontal plane presents a<br />

pattern <strong>of</strong> horizontal disparities that varies with distance. If<br />

depth constancy is to be maintained <strong>for</strong> fronto-parallel<br />

planes then the distance <strong>of</strong> the surface needs to be taken into<br />

account. Rogers and Bradshaw 21 showed that vertical disparity<br />

patterns can have a strong influence on frontal plane<br />

judgements, particularly <strong>for</strong> large field <strong>of</strong> view displays. Specifically,<br />

“flat”—or zero horizontal screen disparity—planes<br />

are perceived as curved if vertical disparity gradients indicate<br />

a distance other than the screen distance.<br />

In our case, the toe-in induced vertical disparity introduces<br />

a cue that the surface is nearer than specified by the<br />

horizontal screen disparity. Thus a zero horizontal screen<br />

disparity pattern <strong>for</strong> a frontal surface at the true distance<br />

would be interpreted as at nearer distance. The disparities<br />

would be less than expected from a frontal plane at the<br />

nearer distance. As a result, surfaces in a scene should appear<br />

curved more concavely than they are in the real scene. Notice<br />

that the distortion is in the opposite direction than the<br />

distortion created by horizontal disparities due to the<br />

keystoning.<br />

Thus the effect <strong>of</strong> vertical disparity introduced by the<br />

keystone distortion is complicated. The vertical disparity introduces<br />

a cue that the surface is nearer than specified by the<br />

horizontal screen disparity. Thus, from vertical disparities,<br />

we would expect a bias in depth perception and concave<br />

distortion <strong>of</strong> stereoscopic space. This may counter the convex<br />

distortions introduced by the horizontal disparities discussed<br />

above. So the surface may appear flatter than expected<br />

from the distorted horizontal disparities. But the<br />

percept is not more “natural” than the parallel configura-<br />

324 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

Figure 8. Disparity <strong>of</strong> a point on a fronto-parallel surface as a function <strong>of</strong> distance. Horizontal disparity <strong>for</strong> a<br />

given eccentricity increases with nearness due to the increasing curvature <strong>of</strong> the Veith-Muller circle see text.<br />

tion. Rather two distortions due to camera toe-in act to<br />

cancel each other out.<br />

Do toed-in configurations provide useful distance<br />

in<strong>for</strong>mation <strong>for</strong> objects at other distances or<br />

<strong>for</strong> nonorthostereoscopic configurations?<br />

Since the toe-in induced vertical disparity gradients are superimposed<br />

upon the natural vertical disparity at the retinae<br />

they do not provide natural distance cues <strong>for</strong> targets near the<br />

display under orthostereoscopic configurations.<br />

Nonorthostereoscopic configurations are more common<br />

than orthostereoscopic and we should consider the effects <strong>of</strong><br />

toe-in on these configurations. Magnification and minification<br />

<strong>of</strong> the images will scale the disparities in the images as<br />

well so that the vertical gradient <strong>of</strong> vertical size ratio will be<br />

relatively unchanged under uni<strong>for</strong>m magnification. Hence<br />

we expect a similar curvature distortion under magnification<br />

or minification.<br />

Hyperstereoscopic and hypostereoscopic configurations<br />

exaggerate and attenuate, respectively, the horizontal and<br />

vertical disparities due to camera toe-in and the magnitude<br />

<strong>of</strong> the stereoscopic distortions will be scaled. However, <strong>for</strong><br />

both configurations the sign <strong>of</strong> the distortion is the same<br />

and vertical disparities from camera toe-in predict concave<br />

curvature <strong>of</strong> stereoscopic space with increased distortion<br />

with an increased stereobaseline.<br />

For surfaces outside the plane <strong>of</strong> the screen, vertical<br />

keystone distortion from toe-in still introduces spatial distortion.<br />

A surface located at a distance beyond the screen in<br />

a parallel camera, orthostereoscopic configuration will have<br />

VSR gradients on spherical retinae appropriate to its distance<br />

due to the imaging geometry. For a toed-in camera<br />

system, all surfaces in the scene will have additional vertical<br />

disparity gradients due to the keystoning. These increased<br />

vertical disparity gradients would indicate a nearer convergence<br />

distance or a nearer surface thus the distance <strong>of</strong> the far<br />

surface should be underestimated and concave curvature introduced.<br />

The distance underestimation would be compounded<br />

by rescaling <strong>of</strong> disparity <strong>for</strong> the near distance<br />

which should compress the depth range in the scene.<br />

What about partial toe-in? For example, let us say we<br />

toed in on a target at 3mand displayed it at 1.0 m with the<br />

centers <strong>of</strong> the image aligned? Would the vertical disparities<br />

in the image indicate a more distant surface, perhaps even<br />

one at 3m (this would be the case if viewed in a haploscope)?<br />

A look at the pattern <strong>of</strong> vertical screen disparities in<br />

this case, however, shows that they are appropriate <strong>for</strong> a<br />

surface that is nearer than the 3msurface, and in fact nearer<br />

than the screen if the half images are aligned on the screen.<br />

Thus when the vertical screen disparities are compounded<br />

by the inherent vertical retinal disparities introduced by<br />

viewing the screen, the toe-in induced distortion actually<br />

indicates a nearer surface rather than the further surface<br />

desired. We will see below that vertical disparity manipulations<br />

can produce the impression <strong>of</strong> a further surface but the<br />

required trans<strong>for</strong>mation is opposite to the one introduced by<br />

camera toe-in.<br />

Do the toed-in configurations improve depth and size<br />

scaling?<br />

Vertical disparities have been shown to be effective in the<br />

scaling <strong>of</strong> depth, shape and size from disparity. 9,21 When the<br />

cameras are toed-in the vertical disparities indicate a nearer<br />

surface. There<strong>for</strong>e, camera toe-in should cause micropsia (or<br />

apparent shrinking <strong>of</strong> linear size) appropriate <strong>for</strong> the nearer<br />

distance. Similarly, depth from disparity should be scaled<br />

appropriate to a nearer surface and depth range should be<br />

compressed. Thus, if toe-in is used to converge an otherwise<br />

orthostereoscopic rig, then image size and depth should be<br />

compressed. Vertical disparity cues to distance are most effective<br />

in a large field <strong>of</strong> view display and the curvature, size<br />

and depth effects are most pronounced in these types <strong>of</strong><br />

displays. 9,21<br />

In the orthostereoscopic case with parallel cameras,<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 325


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

there are no vertical screen disparities and the vertical disparities<br />

in the retinal images are appropriate <strong>for</strong> the screen<br />

distance and no size or depth distortions due to vertical<br />

disparity are predicted. Vertical disparities in the retinal (but<br />

not display) images can thus help obtain veridical stereoscopic<br />

perception.<br />

I use computer graphics or image processing to render<br />

stereoscopic images. Can I use VSR to give an<br />

impression <strong>of</strong> different distances? If so how?<br />

Incorporating elements that carry vertical disparity in<strong>for</strong>mation<br />

(<strong>for</strong> example with horizontal edges) can lead to more<br />

veridical depth perception 8 and in this simple sense vertical<br />

disparity cues can assist in the development <strong>of</strong> effective stereoscopic<br />

displays. It is not certain that manipulating vertical<br />

disparity independent <strong>of</strong> vergence would be <strong>of</strong> use to content<br />

creators, but it is possible. In the lab we do this to look<br />

at the effects <strong>of</strong> vertical disparity gradients and to manipulate<br />

the effects <strong>of</strong> vertical disparities with vergence held constant.<br />

We have seen that toe-in convergence introduces a vertical<br />

disparity cue that indicates that a surface is nearer than<br />

other cues indicate. This will scale stereoscopic depth, shape<br />

and size appropriately, particularly <strong>for</strong> large displays. To<br />

make the surface appear further away the opposite trans<strong>for</strong>mation<br />

is required to reduce the vertical disparity gradients<br />

in the retinal image—this essentially entails “toe-out” <strong>of</strong> the<br />

cameras. VSR manipulations, intentional or due to camera<br />

toe-in, exacerbate cue conflict in the display as the distance<br />

estimate obtained from the vertical disparities will conflict<br />

with accommodation, vergence, and other cues to distance.<br />

FUSION OF VERTICAL DISPARITY<br />

In many treatments <strong>of</strong> the camera convergence problem it is<br />

noted that the vertical disparities introduced by toed-in<br />

camera convergence may interfere with the ability to fuse the<br />

images and cause visual discom<strong>for</strong>t. 24 Certainly, vertical fusional<br />

range is known to be less than horizontal fusional<br />

range 23 making it likely that vertical disparities could be<br />

problematic. Tolerance to vertical disparities depends on<br />

several factors including size <strong>of</strong> the display, and the presence<br />

<strong>of</strong> reference surfaces.<br />

When a stereoscopic image pair has an overall vertical<br />

misalignment, such as arises with vertical camera misalignment,<br />

viewers can compensate with vertical vergence and<br />

sensory fusional mechanisms. Vertical vergence is a disjunctive<br />

eye movement where the left and right eyes move in<br />

opposite directions vertically (vertical misalignment can also<br />

<strong>of</strong>ten be partially compensated by tilting the head with respect<br />

to the display). Vertical disparities are integrated over a<br />

fairly large region <strong>of</strong> space to <strong>for</strong>m the stimulus to vertical<br />

vergence. 25 Larger displays increase the vertical vergence response<br />

and the vertical fusional range. Thus we predict that<br />

vertical disparities will be better tolerated in large displays.<br />

In agreement with this Speranza and Wilcox 26 found up to<br />

30 minutes <strong>of</strong> arc <strong>of</strong> vertical disparity could be tolerated in a<br />

stereoscopic IMAX film without significant viewer discom<strong>for</strong>t.<br />

However, convergence via camera toe-in gives local<br />

variations in vertical disparity and thus images <strong>of</strong> objects in<br />

the display have spatially varying vertical disparities. Thus,<br />

averaging retinal vertical disparities over a region <strong>of</strong> space<br />

should be less effective in compensating <strong>for</strong> vertical disparity<br />

due to camera toe-in compared to overall vertical camera<br />

misalignment. Furthermore, any vertical vergence to fuse<br />

one portion <strong>of</strong> the display will increase vertical disparity in<br />

other parts <strong>of</strong> the display.<br />

The ability to fuse a vertically disparate image is reduced<br />

when nearby stimuli have different vertical disparities, particularly<br />

if the target and background are similar in depth. 27<br />

In many display applications the frame <strong>of</strong> the display is visible<br />

and serves as a frame <strong>of</strong> reference. In other applications<br />

such as augmented reality and enhanced vision displays the<br />

stereoscopic imagery may be imposed upon other imagery.<br />

Presence <strong>of</strong> these competing stereoscopic images will be expected<br />

to reduce the tolerance to vertical disparity due to<br />

camera convergence. 27 This indicates that vertical disparity<br />

distortions should be particularly disruptive in augmented<br />

reality displays where the stereoscopic image is superimposed<br />

on other real or synthetic imagery and parallel cameras<br />

or image rectification should be used.<br />

ADAPTATION AND SENSORY INTEGRATION OF<br />

TOE-IN INDUCED VERTICAL DISPARITY<br />

The human visual system relies on a variety <strong>of</strong> monocular<br />

and binocular cues to judge distance and relative depth in a<br />

scene. The effects <strong>of</strong> toe-in induced horizontal and vertical<br />

disparities on depth and distance perception discussed above<br />

will be reduced when viewing a scene rich in these cues. The<br />

extent <strong>of</strong> the perceptual distortion depends on perceptual<br />

biases and the relative effectiveness <strong>of</strong> the various cues. For<br />

example, Bradshaw and Rogers 21 per<strong>for</strong>med an experiment<br />

using dot displays to study size and depth scaling as a function<br />

<strong>of</strong> distance indicated by vertical disparities and vergence.<br />

They argued that use <strong>of</strong> vertical disparity in<strong>for</strong>mation<br />

to drive size and depth constancy requires measuring the<br />

relevant disparity gradients over a fairly large retinal area<br />

whereas vergence signals, correlated with egocentric distance,<br />

could be obtained during binocular viewing <strong>of</strong> a point<br />

source <strong>of</strong> light. Accordingly, when displays were small, subjects<br />

responded as if they were scaling the stimulus appropriate<br />

<strong>for</strong> the distance indicated by vergence; when displays<br />

were large subjects responded as if they were scaling the<br />

stimulus appropriate <strong>for</strong> the distance indicated by vertical<br />

disparity. When other cues reliably indicate a different distance<br />

than toe-in induced vertical disparities the effect <strong>of</strong> the<br />

latter on depth and size perception may be small. However,<br />

latent, even imperceptible, cue conflicts are believed to be a<br />

causal factor in simulator sickness symptoms such as eye<br />

strain and nausea. 5<br />

When sensory conflict is persistent, the visual system<br />

shows remarkable ability to adapt or recalibrate. Following<br />

prolonged viewing <strong>of</strong> a test stimulus that appears curved due<br />

to keystone-type vertical disparity trans<strong>for</strong>mations a nominally<br />

flat stimulus appears curved in the opposite direction.<br />

Duke and Wilcox 28 have claimed this adaptation is driven by<br />

326 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Allison: Analysis <strong>of</strong> the influence <strong>of</strong> vertical disparities arising in toed-in stereoscopic cameras<br />

the curvature in depth induced rather than by the vertical<br />

disparities directly. In general, such an aftereffect can reflect<br />

“habituation” or “fatigue” <strong>of</strong> mechanisms sensitive to the<br />

adapting pattern, or from a recalibration <strong>of</strong> the vertical disparity<br />

signal, or a change in the relative weighting <strong>of</strong> cues<br />

driving depth constancy. At the present time it is unclear<br />

which <strong>of</strong> these adaptive changes can be produced by prolonged<br />

exposure to keystone patterns <strong>of</strong> vertical disparity.<br />

The effects <strong>of</strong> vertical disparities induced by toe-in convergence<br />

also depends on context and may differ depending<br />

on the type <strong>of</strong> task being per<strong>for</strong>med by the subject and<br />

whether they involve size constancy, depth constancy, absolute<br />

distance judgements or other spatial judgements. For<br />

example, Wei et al. 29 reported that full-field vertical disparities<br />

are not used to derive the distance dependent gain term<br />

<strong>for</strong> the linear vestibulo-ocular reflex, a reflexive eye movement<br />

that compensates <strong>for</strong> head movements, under conditions<br />

where vertical disparities drive depth constancy.<br />

CONCLUSIONS<br />

In conclusion, we concur with conventional wisdom that<br />

horizontal image translation is theoretically preferred to<br />

toe-in convergence with parallel stereoscopic displays.<br />

Toed-in camera convergence is a convenient and <strong>of</strong>ten used<br />

technique that is <strong>of</strong>ten well-tolerated 24 despite the fact that it<br />

theoretically and empirically results in geometric distortion<br />

<strong>of</strong> stereoscopic space. The distortion <strong>of</strong> stereoscopic space<br />

should be more apparent in fused or augmented reality displays<br />

where the real world serves as a reference to judge the<br />

disparity distortion introduced by the toe-in technique. In<br />

these cases, and <strong>for</strong> near viewing when the distortions are<br />

large, the distortions may be ameliorated through camera<br />

rectification techniques 15,30 if resampling <strong>of</strong> the images is<br />

practical.<br />

It has been asserted by others that, since camera convergence<br />

through toe-in introduces vertical disparities into<br />

the stereoscopic imagery it should give rise to more natural<br />

or accurate distance perception than the parallel camera<br />

configuration. We have argued in this paper that these assertions<br />

are theoretically unfounded although vertical disparity<br />

gradients are an effective cue <strong>for</strong> depth and size constancy<br />

that could be used by creators <strong>of</strong> stereoscopic content. The<br />

geometrical distortions predicted from the artifactual horizontal<br />

disparities created by camera toe-in may be countered<br />

by opposite distortions created from the vertical disparities.<br />

However, when displayed on a single projector or monitor<br />

display the vertical disparity gradients introduced by<br />

unrectified, toed-in cameras do not correspond to the gradients<br />

experienced by a real user viewing a scene at the<br />

camera convergence distance. This is because the keystoning<br />

due to the camera toe-in is superimposed upon the natural<br />

vertical disparity pattern at the eyes.<br />

Our analysis and data 27 implies that stereoscopic<br />

display/camera systems that fuse or superimpose multiple<br />

stereoscopic images from a number <strong>of</strong> sensors should be<br />

more susceptible to toe-in induced fusion and depthdistortion<br />

problems than displays that present a single stereoscopic<br />

image stream. Analysis <strong>of</strong> toe-in induced vertical<br />

disparity rein<strong>for</strong>ces the recommendation that rectification <strong>of</strong><br />

the stereoscopic imagery should be considered <strong>for</strong> fused stereoscopic<br />

systems such as augmented reality displays or enhanced<br />

vision systems that require toed-in cameras to view<br />

targets at short distances.<br />

ACKNOWLEDGMENTS<br />

The support <strong>of</strong> the Ontario Centres <strong>of</strong> Excellence and<br />

NSERC Canada is gratefully acknowledged. An abbreviated<br />

version <strong>of</strong> this paper was presented at IST/SPIE Electronic<br />

<strong>Imaging</strong> 2004 [R. Allison, Proc. SPIE 5291, 167–178 (2004)].<br />

REFERENCES<br />

1 A. Woods, T. Docherty, and R. Koch, Proc. SPIE 1915, 36 (1993).<br />

2 D. B. Diner and D. H. Fender, Human Engineering in Stereoscopic<br />

Viewing Devices (Plenum Press, New York and London, 1993).<br />

3 L. Lipton, Foundations <strong>of</strong> the Stereoscopic Cinema (Van<br />

Nostrand–Reinhold, New York, 1982).<br />

4 Z. Wartell, L. F. Hodges, and W. Ribarsky, IEEE Trans. Vis. Comput.<br />

Graph. 8(2), 129 (2002).<br />

5 J. P. Wann, S. Rushton, and M. Monwilliams, Vision Res. 35(19), 2731<br />

(1995).<br />

6 I.P.HowardandB.J.Rogers,Depth Perception (I. Porteous, Toronto,<br />

2002).<br />

7 L. Lipton, The Stereographics Developers Handbook (Stereographics<br />

Corp., San Rafael, CA, 1997).<br />

8 H. von Helmholtz, Physiological Optics, English translation by J. P. C.<br />

Southall from the 3rd German edition <strong>of</strong> Handbuch der Physiologischen<br />

Optik, Vos, Hamburg (Dover, New York, 1962).<br />

9 B. J. Rogers and M. F. Bradshaw, Nature (London) 361, 253 (1993).<br />

10 J. Garding, J. Porrill, J. E. W. Mayhew, and J. P. Frisby, Vision Res. 35(5),<br />

703 (1995).<br />

11 M. Siegel and S. Nagata, IEEE Trans. Circuits Syst. Video Technol. 10(3),<br />

387 (2000).<br />

12 A. State, J. Ackerman, G. Hirota, J. Lee, and H. Fuchs, Proc. International<br />

Symposium on Augmented Reality (ISAR) 2001 (IEEE, Piscataway, NJ,<br />

2001) pp. 137–146.<br />

13 V. S. Grinberg, G. Podnar, and M. W. Siegel, Proc. SPIE 2177, 56 (1994).<br />

14 A. State, K. Keller, and H. Fuchs, Proc. International Symposium on<br />

Mixed and Augmented Reality (ISMAR) 2005 (IEEE, Piscataway, NJ,<br />

2005) pp. 28–31.<br />

15 N. Dodgson, Proc. SPIE 3295, 100 (1998).<br />

16 S. Takagi, S. Yamazaki, Y. Saito, and N. Taniguchi, Proc IEEE & ACM<br />

ISAR 2000 (IEEE, Piscataway, NJ, 2000) pp. 68–77.<br />

17 B. Gillam and B. Lawergren, Percept. Psychophys. 34(2), 121 (1983).<br />

18 I. P. Howard, Psychonomic Monograph Supplements 3, 201 (1970).<br />

19 B. Gillam, D. Chambers, and B. Lawergren, Percept. Psychophys. 44, 473<br />

(1988).<br />

20 J. E. W. Mayhew and H. C. Longuet-Higgins, Nature (London) 297, 376<br />

(1982).<br />

21 M. F. Bradshaw, A. Glennerster, and B. J. Rogers, Vision Res. 36(9), 1255<br />

(1996).<br />

22 L. Perez-Bayas, Proc. SPIE 4297, 251 (2001).<br />

23 K. N. Ogle, Researches in Binocular Vision (Hafner, New York, 1964).<br />

24 L. B. Stelmach, W. J. Tam, F. Speranza, R. Renaud, and T. Martin, Proc.<br />

SPIE 5006, 269 (2003).<br />

25 I. P. Howard, X. Fang, R. S. Allison, and J. E. Zacher, Exp. Brain Res.<br />

130(2), 124 (2000).<br />

26 F. Speranza and L. Wilcox, Proc. SPIE 4660, 18 (2002).<br />

27 R. S. Allison, I. P. Howard, and X. Fang, Vision Res. 40(21), 2985 (2000).<br />

28 P. A. Duke and L. M. Wilcox, Vision Res. 43(2), 135 (2003).<br />

29 M. Wei, G. C. DeAngelis, and D. E. Angelaki, J. Neurosci. 23, 8340<br />

(2003).<br />

30 O. Faugeras and Q. Luong, The Geometry <strong>of</strong> Multiple Images (MIT<br />

Press, Cambridge, MA, 2001).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 327


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 328–336, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Improved B-Spline Contour Fitting Using Genetic<br />

Algorithm <strong>for</strong> the Segmentation <strong>of</strong> Dental Computerized<br />

Tomography Image Sequences<br />

Xiaoling Wu, Hui Gao, Hoon Heo, Oksam Chae, Jinsung Cho, Sungyoung Lee and Young-Koo Lee<br />

Department <strong>of</strong> Computer Engineering, Kyunghee University, 1 Seochun-ri, Kiheung-eup, Yongin-si,<br />

Kyunggi-do, 449-701, South Korea<br />

E-mail: yklee@khu.ac.kr<br />

Abstract. In the dental field, 3D tooth modeling, in which each tooth<br />

can be manipulated individually, is an essential component <strong>of</strong> the<br />

simulation <strong>of</strong> orthodontic surgery and treatment. However, in dental<br />

computerized tomography slices teeth are located closely together<br />

or inside alveolar bone having an intensity similar to that <strong>of</strong> teeth.<br />

This makes it difficult to individually segment a tooth be<strong>for</strong>e building<br />

its 3D model. Conventional methods such as the global threshold<br />

and snake algorithms fail to accurately extract the boundary <strong>of</strong> each<br />

tooth. In this paper, we present an improved contour extraction algorithm<br />

based on B-spline contour fitting using genetic algorithm.<br />

We propose a new fitting function incorporating the gradient direction<br />

in<strong>for</strong>mation on the fitting contour to prevent it from invading the<br />

areas <strong>of</strong> other teeth or alveolar bone. Furthermore, to speed up the<br />

convergence to the best solution we use a novel adaptive probability<br />

<strong>for</strong> crossover and mutation in the evolutionary program <strong>of</strong> the genetic<br />

algorithm. Segmentation results <strong>for</strong> real dental images demonstrate<br />

that our method can accurately determine the boundary <strong>for</strong><br />

individual teeth as well as its 3D model while other methods fail.<br />

Independent manipulation <strong>of</strong> each tooth model demonstrates the<br />

practical usage <strong>of</strong> our method. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong><br />

and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4328<br />

INTRODUCTION<br />

The accurate 3D modeling <strong>of</strong> the mandible and the simulation<br />

<strong>of</strong> tooth movement play an important role in preoperative<br />

planning <strong>for</strong> dental and maxill<strong>of</strong>acial surgery. The 3D<br />

reconstruction <strong>of</strong> the teeth can be used in virtual reality<br />

based training <strong>for</strong> orthodontics students and <strong>for</strong> preoperatory<br />

assessment by dental surgeons. For 3D modeling tooth<br />

segmentation to extract the individual contour <strong>of</strong> a tooth is<br />

<strong>of</strong> critical importance. Automated tooth segmentation methods<br />

from 3D digitized images have been researched <strong>for</strong> the<br />

measurement and simulation <strong>of</strong> orthodontic procedures. 1<br />

These methods provide interstices along with their locations<br />

and orientations between the teeth <strong>for</strong> segmentation result.<br />

However, it does not give individual tooth contour in<strong>for</strong>mation<br />

which manifests more details that are helpful in dental<br />

study. A thresholding method, used in the existing segmentation<br />

and reconstruction systems, is known to be efficient<br />

<strong>for</strong> automatic hard tissue segmentation. 2,3 Some morphological<br />

filtering methods are used <strong>for</strong> creating intermediary<br />

Received Oct. 28, 2006; accepted <strong>for</strong> publication Mar. 30, 2007.<br />

1062-3701/2007/514/328/9/$20.00.<br />

slices by interpolation <strong>for</strong> modeling teeth in 3D. 4 The morphological<br />

operations are also combined with the thresholding<br />

method <strong>for</strong> dental segmentation in x-ray films. 2 However,<br />

neither the thresholding method nor the morphological<br />

filtering method is suitable <strong>for</strong> separating individual tooth<br />

regions using tooth computerized tomography (CT) slices,<br />

because some teeth touch each other and some are located<br />

inside <strong>of</strong> alveolar bone with a CT slice intensity pr<strong>of</strong>ile similar<br />

to teeth. 5 A modified watershed algorithm was suggested<br />

to create closed-loop contours <strong>of</strong> teeth while alleviating the<br />

over-segmentation problem <strong>of</strong> the watershed algorithm. 5 Although<br />

this reduces the number <strong>of</strong> regions significantly, it<br />

still produces many irrelevant basins that make it difficult to<br />

define an accurate tooth contour. A seed-growing segmentation<br />

algorithm 6 was suggested based on B-spline fitting <strong>for</strong><br />

arbitrary shape segmentation in sequential images. The best<br />

contour <strong>of</strong> an object is determined by fitting the initial contour<br />

passed by previous frame to the edges detected in the<br />

current frame. For the fitting operation, the objective function<br />

defined by the sum <strong>of</strong> distances between the initial contour<br />

and the object edges is used. For this algorithm to work<br />

properly, the complete object boundary should be extracted<br />

by global thresholding and the object should be located<br />

apart from other objects. If other objects are located nearby<br />

as in the case <strong>of</strong> the tooth CT image, the shape <strong>of</strong> the initial<br />

contour should be very close to the actual object contour to<br />

prevent being fitted to the boundaries <strong>of</strong> the nearby objects.<br />

Many snake algorithms have been proposed <strong>for</strong> medical<br />

image analysis applications. 7–10 However, in the CT image<br />

sequence where objects are closely located, the classical snake<br />

algorithms have not yet been successful due to difficulties in<br />

initialization and the existence <strong>of</strong> multiple extrema. It is only<br />

successful when it is initialized close to the structure <strong>of</strong> interest<br />

and there is no object which has similar intensity values<br />

to those <strong>of</strong> interest. 7 The snake models <strong>for</strong> object<br />

boundary detection search <strong>for</strong> an optimal contour that minimizes<br />

(or maximizes) an objective function. The objective<br />

function generally consists <strong>of</strong> the internal energy representing<br />

the properties <strong>of</strong> a contour shape and the external potential<br />

energy depending on the image <strong>for</strong>ce. The final shape<br />

<strong>of</strong> the contour is influenced by how these two energy terms<br />

are represented. However, many snakes tend to shrink when<br />

328


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

its external energy is relatively small due to the lack <strong>of</strong> image<br />

<strong>for</strong>ces. 7 Some snakes also suffer from the limited flexibility <strong>of</strong><br />

representing the contour shape and a large number <strong>of</strong> derivative<br />

terms in their internal energy representation. A<br />

B-spline based snake has been developed as a B-spline snake<br />

and B-snake to enhance the geometric flexibility and optimization<br />

speed by means <strong>of</strong> a small number <strong>of</strong> control<br />

points instead <strong>of</strong> snaxels. 11,12 B-spline snake controls contour<br />

shapes by a stiffening parameter as well as its control<br />

points, and detects object boundaries in noisy environments<br />

by using gradient magnitude in<strong>for</strong>mation instead <strong>of</strong> edge<br />

in<strong>for</strong>mation. This algorithm introduces a stiffening factor to<br />

the B-spline function 13 that varies the spacing between the<br />

spline knots and the number <strong>of</strong> sampled points used during<br />

the evaluation <strong>of</strong> the objective function. In addition, the<br />

factor controls the smoothness <strong>of</strong> curve and reduces the<br />

computation <strong>of</strong> the cost function. Although the algorithm<br />

was proposed to extract the contour <strong>of</strong> a de<strong>for</strong>mable object<br />

in a single image, it can be applied to the tooth segmentation<br />

in CT slices. However, in tooth CT data, the algorithm may<br />

cause the contour <strong>of</strong> a tooth to expand to include contours<br />

<strong>of</strong> nearby teeth and alveolar bone, or it may cause the contour<br />

to be contracted to a small region.<br />

A B-spline fitting algorithm employing a genetic algorithm<br />

(GA) was used to overcome local extrema indwelling<br />

in the vicinity <strong>of</strong> an object <strong>of</strong> interest. 14–17 In this case, it was<br />

shown that the GA does not require exhaustive search while<br />

avoiding high-order derivatives <strong>for</strong> curve fitting or matching<br />

problems. 18,19 However, the conventional GA-based B-spline<br />

fitting still suffers from the influence <strong>of</strong> other objects and<br />

<strong>of</strong>ten fails to extract the object boundary from the image<br />

sequences when similar objects are adjacent to each other.<br />

In this paper, we propose an improved B-spline contour<br />

fitting algorithm using a GA to generate a smooth and accurate<br />

tooth boundary <strong>for</strong> the 3D reconstruction <strong>of</strong> a tooth<br />

model. We devise a new B-spline fitting function by incorporating<br />

the gradient direction in<strong>for</strong>mation on the fitting<br />

contours to search the tooth boundary while preventing it<br />

from being fitted to neighboring spurious edges. We also<br />

present an evolution method to accelerate the search speed<br />

by means <strong>of</strong> automatic and dynamic determination <strong>of</strong> GA<br />

probabilities <strong>for</strong> crossover and mutation. Experimental results<br />

show that our method can successfully extract the individual<br />

tooth boundary, compared with other methods<br />

which fail to do so.<br />

BACKGROUND<br />

Dental CT images have the following two distinct characteristics:<br />

(1) An individual tooth <strong>of</strong>ten appears with neighboring<br />

hard tissues such as other teeth and alveolar bone, and<br />

(2) these neighboring hard tissues have the same or similar<br />

intensity values to the tooth <strong>of</strong> interest. Thus, the fixed<br />

threshold value <strong>for</strong> each tooth in each slice is not effective as<br />

shown in Figure 1. When we try to obtain a tooth region by<br />

thresholding method, the lower and upper limits <strong>of</strong> a threshold<br />

value can be displayed at each slice <strong>for</strong> a given tooth by<br />

the two curves in Fig. 1. Any threshold value within the limit<br />

Figure 1. Threshold values <strong>for</strong> a certain tooth computed at different slices<br />

by manual.<br />

produces the tooth region with the accuracy better than<br />

90%. It shows us that individual segmentation method is<br />

required <strong>for</strong> each tooth in each slice.<br />

There are many segmentation methods, each <strong>of</strong> which<br />

have their own limitations in separating individual tooth<br />

regions on CT images. 3–6 An optimal thresholding scheme 20<br />

can be attempted by taking advantage <strong>of</strong> the fact that the<br />

shape and intensity <strong>of</strong> each tooth changes gradually through<br />

the CT image slices.<br />

However, even if an optimal threshold is determined <strong>for</strong><br />

every slice, the result <strong>of</strong> the segmentation is found unsatisfactory<br />

because <strong>of</strong> neighboring hard tissue. For the 3D reconstruction<br />

<strong>of</strong> an individual tooth model, the tooth boundary<br />

needs to be defined more precisely.<br />

B-Spline Contour Fitting<br />

The B-spline curve has attractive properties <strong>for</strong> the representation<br />

<strong>of</strong> an object contour with arbitrary shape. They are<br />

also suitable <strong>for</strong> the curve fitting process and are summarized<br />

as follows.<br />

• An object <strong>of</strong> any shape, including those subsuming angular<br />

points, can be represented by a set <strong>of</strong> control<br />

points, a knot sequence, and a basis function. The shape<br />

<strong>of</strong> the contour can be adjusted by simply repositioning<br />

the control points in many fitting problems where the<br />

knot sequence and basis function can be fixed.<br />

• Little else remains to be different in the shape <strong>of</strong> the<br />

contour by deducting the number <strong>of</strong> control points<br />

within some tolerable limit <strong>for</strong> the purpose <strong>of</strong> reducing<br />

in<strong>for</strong>mation needed <strong>for</strong> fitting process. This allows the<br />

fitting process to be faster with fewer variables over<br />

which to optimize.<br />

We choose the uni<strong>for</strong>m cubic closed B-spline curve,<br />

shown as follows in Eqs. (1) and (2), to describe the object<br />

contours in the image.<br />

rs = r xs<br />

r y s =<br />

n−1<br />

x i B 0 s − i<br />

i=0<br />

n−1<br />

y i B 0 s − i, 1<br />

<br />

i=0<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 329


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

B 0 s =s 3 /2 − s 2 + 2/3 if t 0 s t 1 ,<br />

2−s 3 /6 if t 1 s t 2 ,<br />

0 otherwise<br />

In the equations, rs represents the coordinate <strong>of</strong> a contour<br />

pixel at a specific value <strong>of</strong> parameter s and x i ,y i represents<br />

coordinates <strong>of</strong> ith control point. The B-spline basis functions<br />

are translated copies <strong>of</strong> B 0 s. In the case <strong>of</strong> tooth<br />

segmentation we use a closed uni<strong>for</strong>m knot sequence, as<br />

t 0 ,t 1 ,...,t n =0,1,...,n and t 0 =t n where n is the total<br />

number <strong>of</strong> the control points.<br />

The B-spline fitting function f is represented in Eq. (3)<br />

(Ref. 11) as follows:<br />

M−1<br />

f = Irs k ,<br />

k=0<br />

where M is the total number <strong>of</strong> contour points. The fitting<br />

function is maximized when the contour con<strong>for</strong>ms to the<br />

object boundary. The B-spline fitting function makes use <strong>of</strong><br />

only external <strong>for</strong>ce computed based on the gradient magnitude<br />

on the contour. The smoothness constraint is implicitly<br />

represented by the B-spline itself.<br />

B-spline Contour Fitting using Genetic Algorithm<br />

The genetic algorithm is a probabilistic technique <strong>for</strong> searching<br />

<strong>for</strong> an optimal solution. The optimal solution is described<br />

by a vector, called a “chromosome,” which can be<br />

obtained by maximizing a fitting function. Hence the definition<br />

<strong>of</strong> the fitting function significantly affects the solution<br />

state. A sequence <strong>of</strong> evolutionary operations is repeated <strong>for</strong> a<br />

chromosome to evolve to its final state. The end <strong>of</strong> the evolutionary<br />

operation is determined by checking the fitness<br />

values, which represent the goodness <strong>of</strong> each chromosome in<br />

the population.<br />

A chromosome is a collection <strong>of</strong> genes, and a gene represents<br />

the control point <strong>of</strong> B-spline. Since the chromosome<br />

represents a complete contour and a gene uses the actual<br />

location <strong>of</strong> a control point, the search algorithm has neither<br />

ambiguity on the contour location nor potential bias to particular<br />

shapes. To reduce the size <strong>of</strong> a gene, we use the index<br />

value as a gene, instead <strong>of</strong> two coordinate values. 16,17 Composing<br />

a search area based on the indices provides a search<br />

area with arbitrary shape, where it is confined to search <strong>for</strong><br />

the final position <strong>of</strong> the control point to be found out. This<br />

scheme <strong>of</strong> chromosome guarantees that gene in<strong>for</strong>mation<br />

does not spread over the chromosome, which results in short<br />

length and order <strong>of</strong> schema. 16 Accordingly, there is a high<br />

probability to converge fast. A new generation is made<br />

through the sequence <strong>of</strong> evolutionary operations and, during<br />

the evolutionary processes, crossover and mutation steps affect<br />

the quality and speed <strong>of</strong> final solution significantly.<br />

2<br />

3<br />

IMPROVED B-SPLINE CONTOUR FITTING USING<br />

GENETIC ALGORITHM<br />

Fitting Function Based on Gradient Magnitude and<br />

Direction<br />

The fitting function measures the fitness <strong>of</strong> the possible contour<br />

to the object boundary in the current slice. The fitness<br />

value is the basis <strong>for</strong> determining the termination <strong>of</strong> the<br />

evolutionary process and selecting elite chromosomes <strong>for</strong><br />

mating pool generation. In the existing active contour models,<br />

the fitting function consists <strong>of</strong> the internal <strong>for</strong>ces controlling<br />

the smoothness <strong>of</strong> the contour and the external <strong>for</strong>ce<br />

used <strong>for</strong> representing the object boundary in<strong>for</strong>mation in<br />

the image. 7,12 One drawback <strong>of</strong> this representation is that it<br />

requires the determination <strong>of</strong> the weight values balancing<br />

these two components.<br />

B-spline snake makes use <strong>of</strong> a simple fitting function<br />

with only external <strong>for</strong>ce computed based on the gradient<br />

magnitude on the contour. The internal <strong>for</strong>ce terms are replaced<br />

by using a stiffening parameter and implicit smoothness<br />

constraint <strong>of</strong> the B-spline representation <strong>of</strong> a contour.<br />

However, in the image data such as the tooth CT image<br />

slices, those fitting functions <strong>of</strong>ten generate the contour fitted<br />

to the boundary <strong>of</strong> nearby object. They also generate the<br />

contour contracted to a small region unless the stiffening<br />

parameter is set properly.<br />

Note that the magnitude <strong>of</strong> the intensity difference may<br />

vary between the inside and outside <strong>of</strong> an object contour.<br />

However, if the relative intensity between two sides <strong>of</strong> a contour<br />

is maintained throughout the contour, the sign <strong>of</strong> the<br />

intensity difference made by two sides is inverted when the<br />

contour expands out to the boundary <strong>of</strong> another object.<br />

Hence, when fixing moving direction <strong>of</strong> parameter s along<br />

the curve, we are able to have knowledge <strong>of</strong> which side is<br />

inside (or outside) in advance. This enables us to know<br />

whether the contour is fitted to the object <strong>of</strong> interest or other<br />

adjacent objects. In this paper, the fitting function to be<br />

maximized is designed to take advantage <strong>of</strong> this property <strong>of</strong><br />

the data. This gradient direction in<strong>for</strong>mation allows the fitness<br />

function to penalize the portion <strong>of</strong> a contour fitted to<br />

the neighboring object.<br />

To compute the fitness value <strong>for</strong> a possible solution (or<br />

chromosome), we first generate the contour points from the<br />

B-spline representation <strong>of</strong> the solution and trace the contour<br />

as shown in Figure 2(a). At the kth contour point rs k ,a<br />

unit normal vector ns k is computed. Next, the inner region<br />

i<br />

and outer region pixel location p k and p o k , respectively, are<br />

identified by using ns k computed at the kth point rs k <br />

according to<br />

and<br />

p k o = rs k + ns k <br />

p k i = rs k − ns k .<br />

Then, the fitness value is determined based on gradient<br />

magnitude and direction in<strong>for</strong>mation, k ,ateachcontour<br />

point according to<br />

4<br />

5<br />

330 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

Figure 2. a Definition <strong>of</strong> inner and outer regions. b Illustration <strong>for</strong><br />

fitting function—right object is <strong>of</strong> interest, with adjacent left object, and<br />

thick black curve is a fitting curve. c Twisted contour.<br />

where<br />

and<br />

k =<br />

M−1<br />

f = k − k ,<br />

k=0<br />

k =Irs k if Ip k i − Ip k o 0,<br />

− Irs k if Ip k i − Ip k o 0,<br />

C, rs k = rs j <br />

, ∀ j 0,1, ... ,M −1 ∧ j k.<br />

0, rs k rs j <br />

Ip k i and Ip k o are intensity values <strong>of</strong> the inside and outside<br />

<strong>of</strong> the kth contour point, respectively. This equation is further<br />

illustrated by Fig. 2(b), where some portion <strong>of</strong> the contour<br />

attaches to another object and in this portion<br />

Ip k i Ip k o , so we assign the negative gradient magnitude<br />

to penalize the fitness value. The figure also shows that in<br />

other portions the contour correctly con<strong>for</strong>ms to the tooth<br />

boundary and in these portions Ip k i Ip k o , so we assign<br />

the positive gradient magnitude to the fitness value. Note<br />

that when there is no difference <strong>of</strong> gradient direction, which<br />

may happen if inner and outer pixel values are identical,<br />

then Ip k i =Ip k o . This aims at preventing the contour from<br />

being misfitted when the contour lies inside an object region<br />

having uni<strong>for</strong>m intensity values, such as the inside region <strong>of</strong><br />

a tooth.<br />

A constant-valued penalty C is deducted from the fitness<br />

value when the contour is twisted as shown in Fig. 2(c).<br />

Our experimental results showed that setting the penalty too<br />

high hindered searching the contour maximizing the sum <strong>of</strong><br />

gradient magnitudes. The proposed fitting method yields the<br />

best per<strong>for</strong>mance when C is set to around 0.1% <strong>of</strong> the sum<br />

<strong>of</strong> gradient magnitudes.<br />

6<br />

Improved Adaptive Evolutionary Operations<br />

The evolutionary process generates a new population <strong>of</strong> possible<br />

solutions through the following three genetic operators:<br />

reproduction (or selection), crossover, and mutation. The<br />

selection operation constructs the mating pool from the current<br />

population <strong>for</strong> the crossover operation. The results presented<br />

here use a tournament selection scheme. 16 The crossover<br />

operation generates two child chromosomes by<br />

swapping genes between the two parent chromosomes. In<br />

this paper we present one point cutting scheme by improved<br />

adaptive crossover probability. We also use an adaptive mutation<br />

probability scheme <strong>for</strong> our evolutionary process.<br />

The conventional GA generally uses fixed crossover and<br />

mutation probabilities. Adaptive genetic algorithm 21 (AGA)<br />

was proposed by Srinivas et al. that uses variable crossover<br />

and mutation probabilities that are determined automatically<br />

based on fitness values during fitting process <strong>for</strong> fast<br />

convergence rate. The probabilities <strong>for</strong> evolution are, there<strong>for</strong>e,<br />

no longer required to be set to constants. At the beginning<br />

stage <strong>of</strong> the fitting process, we consider all the possibilities<br />

<strong>of</strong> control point locations in the search area. As the<br />

process goes on, we obtain the evolutionary probabilities<br />

such that the possible solution near the optimal solution<br />

quickly converges to the actual solution. In AGA, 21 the crossover<br />

probability is adaptively determined depending on the<br />

fitness value f, according to<br />

f best − f<br />

1 , f f avg ,<br />

p c =k f best − f avg<br />

k 2 , f f avg ,<br />

where f best and f avg are the best and average fitness values in<br />

the mating pool, respectively, and k 1 and k 2 are constants<br />

and set to 1.0. Hence, if f=f best when ff avg , f is preserved,<br />

although the value <strong>of</strong> k 1 ensures high occurrence <strong>of</strong> crossover.<br />

If ff avg , crossover is operated without exceptions,<br />

since its corresponding chromosome has low fitness value.<br />

The mutation operation is also implemented by using<br />

the mutation probability p m as follows:<br />

f best − f<br />

3 , f f avg ,<br />

p m =k f best − f avg<br />

k 4 , f f avg ,<br />

where k 3 and k 4 are constants set to 0.5. As in the case <strong>of</strong><br />

crossover, the mutation operation does not affect the chromosome<br />

with the best fitness value. However if ff avg its<br />

mutation operation takes place with the most ambiguity<br />

since k 3 =0.5.<br />

In this paper we propose an improved adaptive crossover<br />

probability. To maintain the solution with high fitness<br />

value, we generate a random number p r and consider the<br />

relationship <strong>of</strong> p r with p c1 and p c2 ,wherep c1 and p c2 denote<br />

crossover probabilities generated from two parent chromosomes,<br />

father chromosome and mother chromosome respectively.<br />

When two parent chromosomes are selected, two children<br />

are generated as follows.<br />

7<br />

8<br />

(1) Generate a random number p r between0and1to<br />

determine the adaptive crossover probability, generate<br />

a random number p l between0and1todetermine<br />

the crossing site, and generate a random<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 331


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

number p s between 0 and 1 to determine which<br />

side <strong>of</strong> the crossing site the portion <strong>of</strong> the chromosome<br />

should exchange with the corresponding portion<br />

<strong>of</strong> its mate.<br />

(2) Replace f in Eq. (7) by the fitness value <strong>of</strong> each<br />

parent <strong>for</strong> computing the crossover probabilities,<br />

p c1 and p c2 .<br />

(3) If p r p c1 and p r p c2 , put the two parents to the<br />

next generation without change.<br />

(4) If p r is between p c1 and p c2 , thus p c1 p c2 and<br />

p s 0.5 then the left portion <strong>of</strong> the father chromosome<br />

should be exchanged with the corresponding<br />

portion <strong>of</strong> the mother chromosome to generate one<br />

child and put mother chromosome directly to the<br />

generation as another child. If p s 0.5 then the<br />

right portion from the father chromosome should<br />

be exchanged to generate one child and another<br />

child is a copy <strong>of</strong> the mother chromosome. Similarly<br />

if p c1 p c2 then the mother chromosome<br />

should be changed and put to the next generation<br />

while the father chromosome is put to the next<br />

generation without any change. In addition, the<br />

crossover scheme is determined by the value <strong>of</strong> p s .<br />

(5) If p r is less than both p c1 and p c2 , generate two child<br />

chromosomes as the normal crossover method<br />

does.<br />

In the proposed operation, the chromosomes with high<br />

fitness values can survive until a new chromosome with<br />

higher fitness is created. It supports rapid searching <strong>for</strong> an<br />

optimal solution by taking advantage <strong>of</strong> the crossover<br />

scheme swapping either side to the crossing site.<br />

EXPERIMENTAL EVALUATION<br />

We tested the proposed contour segmentation with two<br />

kinds <strong>of</strong> sets <strong>of</strong> data: synthetic images and two sets <strong>of</strong> real<br />

dental CT image sequences with a slice thickness <strong>of</strong> 0.67mm<br />

and 1mm and x-y resolution <strong>of</strong> 512512. Visual C++ with<br />

DICOM libraries 22 <strong>for</strong> reading 16-bit CT images and the 3D<br />

graphics library OpenGL were used as tools to implement<br />

the proposed algorithm. CT images are saved in DICOM<br />

<strong>for</strong>mat, an international standard <strong>for</strong> medical images, after<br />

acquisition through the commercially available Shimadzu<br />

Ltd. SCT-7800 CT scanner. The test data were prepared to<br />

reveal the capability <strong>of</strong> the proposed algorithm in finding an<br />

accurate boundary among many similar objects nearby. We<br />

compared the proposed algorithm with the existing B-spline<br />

snake algorithm that uses the gradient magnitude based external<br />

<strong>for</strong>ce in the fitting function. 11<br />

First, we applied these algorithms to a synthetic image<br />

similar to a tooth surrounded by alveolar bone. To generate<br />

the results, we constructed a B-spline contour with 8 control<br />

points and selected 20 initial chromosomes <strong>for</strong> each<br />

4040 window. For the following examples <strong>of</strong> B-spline<br />

snake the stiffening parameter is set to 2. As shown in Figure<br />

3, the proposed algorithm extracts an accurate object<br />

boundary while the existing B-spline snake fails.<br />

We also applied the two algorithms to real CT image<br />

Figure 3. Contours extracted from the synthetic data number <strong>of</strong> control<br />

points CP=8. a By B-spline snake method. b By the proposed<br />

method.<br />

sequences where an individual tooth <strong>of</strong>ten appears with<br />

neighboring hard tissues such as other teeth and alveolar<br />

bone. If too many control points are used <strong>for</strong> a contour, it<br />

reduces the smoothing effect on the curve and consequently<br />

generates twisted parts <strong>of</strong> contour as shown in Figure 4.<br />

Figure 5 shows part <strong>of</strong> test results using different set <strong>of</strong> slices,<br />

which have lower resolution. Since the test image is small, a<br />

1010 search area suffices <strong>for</strong> a control point.<br />

As shown in Fig. 5, an individual tooth <strong>of</strong>ten appears<br />

with neighboring hard tissues such as other teeth and alveolar<br />

bone, and the proposed algorithm produces better results<br />

than B-spline snake. The difference in the results stems from<br />

the fitting function.<br />

Part <strong>of</strong> the segmentation results <strong>of</strong> slice sequences is<br />

shown in Figure 6 and those <strong>of</strong> a molar having a more<br />

complicated shape are shown in Figure 7. In Fig. 6, the figures<br />

at the far left side show the results <strong>of</strong> teeth initialization<br />

<strong>for</strong> the first slice by applying a proper threshold to each<br />

tooth interactively. As the segmentation is per<strong>for</strong>med slice by<br />

slice, in contrast with the results <strong>of</strong> proposed method, malfitting<br />

error contained in the results <strong>of</strong> the existing method<br />

increases.<br />

332 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

Figure 4. Tooth contours extracted from CT image CP=16. a By the<br />

proposed method. b By B-spline snake.<br />

Table I lists part <strong>of</strong> the numerical results <strong>of</strong> the segmentation.<br />

N is the number <strong>of</strong> slices over which each tooth<br />

spans. FPE (false positive error) is the percent <strong>of</strong> area reported<br />

as a tooth by the algorithm, but not by manual segmentation.<br />

FNE (false negative error) is the percent <strong>of</strong> area<br />

reported by manual segmentation, but not by the algorithm.<br />

Similarity and dissimilarity indices, 23,10 which show the<br />

amount <strong>of</strong> agreement and disagreement, S agr and S dis ,respectively,<br />

between the algorithm area A alg and the manual<br />

segmentation area A man , are computed according to<br />

S agr =2 A man A alg<br />

A man + A alg<br />

,<br />

9<br />

S dis =2 A man A alg − A man A alg<br />

A man + A alg<br />

. 10<br />

Figure 5. Tooth contours extracted from CT image sequence CP=8.<br />

a By the proposed method. b By B-spline snake.<br />

These indices are calculated <strong>for</strong> validation on N slices <strong>of</strong><br />

each tooth. Averaged values <strong>of</strong> S agr as well as its minimum<br />

and maximum values are shown in Table I, and we conclude<br />

that the proposed method <strong>for</strong> segmentation isolates individual<br />

region <strong>of</strong> tooth successfully, in contrast with the results<br />

<strong>of</strong> B-spline snake shown in Table II.<br />

The proposed fitting method is designed <strong>for</strong> the fast<br />

contour extraction by the improved crossover method which<br />

uses a random number <strong>for</strong> copying genes <strong>of</strong> a superior chromosome<br />

to an inferior one when the random number falls<br />

into the range <strong>of</strong> crossover probabilities <strong>of</strong> its parents, p c1<br />

and p c2 . Furthermore, the proposed crossover method decides<br />

which part <strong>of</strong> crossing site will be exchanged between<br />

parent chromosomes. The decided part fosters chromo-<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 333


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

Table I. Segmentation results <strong>for</strong> 8 teeth <strong>of</strong> the proposed method from the same scans<br />

<strong>of</strong> CT set.<br />

Tooth N FPE% FNE% S agr S min S max S dis<br />

Figure 6. Tooth contours extracted from CT image sequence CP=16.<br />

a By the proposed method. b By B-spline snake.<br />

1 20 4.43 8.37 0.935 0.915 0.977 0.131<br />

2 22 7.88 3.45 0.945 0.916 0.973 0.111<br />

3 25 8.96 4.48 0.935 0.901 0.968 0.131<br />

4 24 8.46 6.47 0.926 0.905 0.970 0.148<br />

5 27 5.81 8.29 0.929 0.917 0.967 0.143<br />

6 26 2.07 7.05 0.953 0.923 0.971 0.094<br />

7 25 5.21 3.79 0.955 0.927 0.976 0.089<br />

8 23 5.69 1.42 0.965 0.932 0.983 0.069<br />

Table II. Segmentation results <strong>for</strong> 8 teeth <strong>of</strong> B-spline snake from the same scans <strong>of</strong> CT<br />

set.<br />

Tooth N FPE% FNE% S agr S min S max S dis<br />

1 20 6.12 27.21 0.814 0.574 0.952 0.373<br />

2 22 26.01 1.16 0.879 0.628 0.956 0.241<br />

3 25 45.86 11.28 0.756 0.316 0.897 0.487<br />

4 24 29.89 4.59 0.842 0.764 0.941 0.313<br />

5 27 28.06 8.06 0.836 0.726 0.933 0.328<br />

6 26 15.09 8.81 0.884 0.818 0.948 0.232<br />

7 25 27.98 5.03 0.852 0.755 0.936 0.296<br />

8 23 10.12 3.89 0.932 0.771 0.972 0.136<br />

Figure 8. Comparison <strong>of</strong> convergence rates.<br />

Figure 7. Extracted contours <strong>of</strong> molar CP=32. a By the proposed<br />

method. b By B-spline snake.<br />

somes to be competent with a high fitness value. We implement<br />

two genetic B-spline fittings with existing crossover<br />

methods to analyze the per<strong>for</strong>mance <strong>of</strong> the proposed crossover.<br />

Both existing methods generate the initial population<br />

randomly, with uni<strong>for</strong>m distribution, while using different<br />

crossover methods. “Method A” uses a fixed p c <strong>of</strong> 0.75 and<br />

“Method B” uses AGA, which determines p c adaptively. Figure<br />

8 compares the convergence rate <strong>of</strong> the proposed crossover<br />

method with those <strong>of</strong> the existing methods in terms <strong>of</strong><br />

the fitness value along chromosome generation. The figure<br />

shows that the proposed crossover method results in a better<br />

334 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

Figure 10. Manipulation <strong>of</strong> tooth. a Every tooth can be manipulated.<br />

b Simulation <strong>of</strong> having tooth out.<br />

Figure 9. Wireframe models <strong>of</strong> tooth and mandible. a 3D reconstruction<br />

<strong>of</strong> tooth. b 3D reconstruction <strong>of</strong> mandible.<br />

convergence rate than either method A or B. The proposed<br />

crossover method preserves the chromosomes with high fitness<br />

<strong>for</strong> fast convergence and the results shows it is effective<br />

to randomly select either side to crossing site <strong>for</strong> improved<br />

crossover operation.<br />

Individual segmentation <strong>of</strong> all teeth can be used to reconstruct<br />

a model <strong>of</strong> the mandible, as shown in Figures 9<br />

and 10. Every tooth can be separated from the jaw <strong>for</strong> simulation<br />

<strong>of</strong> dental treatments.<br />

CONCLUSIONS<br />

In this paper, we presented the improved genetic B-spline<br />

curve fitting algorithm <strong>for</strong> extracting individual smooth<br />

tooth contours from CT slices while preventing the contour<br />

from being twisted. This enables us to obtain individual accurate<br />

contours <strong>of</strong> teeth by overcoming the problem <strong>of</strong> the<br />

contour <strong>of</strong> a tooth expanding out to other teeth boundaries<br />

in the fitting process. Furthermore, we also devised the<br />

crossover method which accelerates convergence rate by<br />

means <strong>of</strong> both conserving chromosomes with high fitness<br />

value and allowing exchange <strong>of</strong> either side <strong>of</strong> cross site. The<br />

test results show that the proposed segmentation algorithm<br />

successfully extracts a smooth tooth contour under specific<br />

conditions such as the existence <strong>of</strong> objects in close vicinity.<br />

This paper also demonstrated the possibility <strong>of</strong> reconstruction<br />

<strong>of</strong> a 3D model in which each tooth was modeled<br />

and manipulated separately <strong>for</strong> the simulation <strong>of</strong> dental surgery.<br />

These anatomical 3D models can be used <strong>for</strong> facilitating<br />

diagnoses, pre-operative planning and prosthesis design.<br />

They will provide radiography <strong>of</strong> the mandible, an accurate<br />

mechanical model <strong>of</strong> the individual tooth and that <strong>of</strong> its root<br />

<strong>for</strong> endodontics and orthodontic operations. Hence the 3D<br />

reconstruction <strong>of</strong> the teeth can be used in virtual reality<br />

based training <strong>for</strong> orthodontics students and <strong>for</strong> preoperatory<br />

assessment by dental surgeons.<br />

ACKNOWLEDGMENTS<br />

This research was supported by the MIC (Ministry <strong>of</strong> In<strong>for</strong>mation<br />

and Communication), Korea, under the ITRC (In<strong>for</strong>mation<br />

Technology Research Center) support program<br />

supervised by the IITA (Institute <strong>of</strong> In<strong>for</strong>mation Technology<br />

Advancement) (IITA-2006–C1090–0602–0002). The authors<br />

are grateful to K. Blankenship and Y. Blankenship <strong>for</strong> their<br />

ef<strong>for</strong>t in pro<strong>of</strong>reading this paper.<br />

REFERENCES<br />

1 T. Kondo, S. H. Ong, and K. W. C. Foong, “Tooth segmentation <strong>of</strong><br />

dental study models using range images”, IEEE Trans. Med. <strong>Imaging</strong> 23,<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 335


Wu et al.: Improved B-spline contour fitting using genetic algorithm <strong>for</strong> the segmentation <strong>of</strong> dental...<br />

350–362 (2004).<br />

2 E. H. Said, D. E. M. Nassar, G. Fahmy, and H. H. Ammar, “Teeth<br />

segmentation in digitized dental X-ray films using mathematical<br />

morphology”, IEEE Trans. Inf. Forensics Security 1, 178–189 (2006).<br />

3 J. H. Ryu, H. S. Kim, and K. H. Lee, “Contour based algorithms <strong>for</strong><br />

generating 3D CAD models from medical images”, Int. J. Adv. Manuf.<br />

Technol. 24, 112–119 (2004).<br />

4 A. G. Bors, L. Kechagias, and I. Pitas, “Binary Morphological Shape-<br />

Based Interpolation Applied to 3-D Tooth Reconstruction”, IEEE Trans.<br />

Med. <strong>Imaging</strong> 21, 100–108 (2002).<br />

5 G. Bohm, C. Knoll, V. G. Colomer, M. Alcaniz-Raya, and S. Albalat,<br />

“Three-dimensional segmentation <strong>of</strong> bone structures in CT images”,<br />

Proc. SPIE 3661, 277–286 (1999).<br />

6 S. Liu and W. Ma, “Seed-growing segmentation <strong>of</strong> 3D surfaces from<br />

CT-contour data”, Computer-Aided Design 31, 517–536 (1999).<br />

7 M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: Active contour<br />

models”, Int. J. Comput. Vis. 1, 321–331 (1988).<br />

8 C. Han and W. S. Kerwin, “Detecting objects in image sequences using<br />

rule-based control in an active contour model”, IEEE Trans. Biomed.<br />

Eng. 50, 705–710 (2003).<br />

9 A. Klemencic, S. Kovacic, and F. Pernus, “Automated segmentation <strong>of</strong><br />

muscle fiber images using active contour models”, Cytometry 32,<br />

317–326 (1998).<br />

10 J. Klemencic, V. Valencic, and N. Pecaric, “De<strong>for</strong>mable contour based<br />

algorithm <strong>for</strong> segmentation <strong>of</strong> the hippocampus from MRI”, in Proc. 9th<br />

International Conference on Computer Analysis <strong>of</strong> Images and Patterns,<br />

Lect. Notes Comput. Sci. 2124, 298–308 (2001).<br />

11 P. Brigger, J. Hoeg, and M. Unser, “B-Spline snakes: A flexible tool <strong>for</strong><br />

parametric contour detection”, IEEE Trans. Image Process. 9, 1484–1496<br />

(2000).<br />

12 S. Menet, P. Saint-Marc, and G. Medioni, “Active contour models:<br />

overview, implementation and applications. Systems”, in Proc. Man and<br />

Cybernetics (IEEE Press, Piscataway, NJ, 1990), pp. 194–199.<br />

13 G. Farin, Curves and surfaces <strong>for</strong> CAGD, 4th ed. (Academic Press, New<br />

York, 1996), pp. 141–168.<br />

14 M. Cerrolaza, W. Annicchiarico, and M. Martinez, “Optimization <strong>of</strong> 2D<br />

boundary element models using B-splines and genetic algorithms”, in<br />

Engineering Analysis with Boundary Elements (Elsevier, Ox<strong>for</strong>d, 2000),<br />

Vol. 24, pp. 427–440.<br />

15 W. Annicchiarico and M. Cerrolaza, “Finite elements, genetic algorithms<br />

and B-splines: a combined technique <strong>for</strong> shape optimization”, Finite<br />

Elem. Anal. Design 33, 125–141 (1999).<br />

16 C. Ooi and P. Liatsis, “Co-evolutionary-based active contour models in<br />

tracking <strong>of</strong> moving obstacles”, Proc. International Conference on<br />

Advanced Driver Assistance Systems (IEEE, London, 2001), pp. 58–62.<br />

17 L. A. MacEachern and T. Manku, “Genetic algorithms <strong>for</strong> active contour<br />

optimization”, Proc. IEEE Int. Sym. <strong>for</strong> Circuits and Systems (IEEE Press,<br />

Piscataway, NJ, 1998), pp. 229–232.<br />

18 M.-S. Dao, F. G. B. De Natale, and A. Massa, “Edge Potential Functions<br />

(EPF) and Genetic Algorithms (GA) <strong>for</strong> Edge-Based Matching <strong>of</strong> Visual<br />

Objects”, IEEE Trans. Multimedia 9, 120–135 (2007).<br />

19 L. Ballerni and L. Bocchi, “Multiple Genetic Snakes <strong>for</strong> Bone<br />

Segmentation”, in Proc. EvoWorkshops, Lect. Notes Comput. Sci. 2611,<br />

346–356 (2003).<br />

20 R. C. Gonzalez and R. E. Woods, Digital Image Processing (Addison<br />

Wesley, Reading, MA, 1993), pp. 447–455.<br />

21 M. Srinivas and L. M. Patnaik, “Adaptive probabilities <strong>of</strong> crossover and<br />

mutation in genetic algorithms”, IEEE Trans. Syst. Man Cybern. 24,<br />

656–667 (1994).<br />

22 OFFIS Institute <strong>for</strong> In<strong>for</strong>mation Technology website, http://www.<strong>of</strong>fis.de/<br />

index-e.php/, accessed October 2006.<br />

23 A. P. Zijdenbos, B. M. Dawant, R. A. Margolin, and A. C. Palmer,<br />

“Morphometric analysis <strong>of</strong> white matter lesions in MR images: Method<br />

and validation”, IEEE Trans. Med. <strong>Imaging</strong> 13, 716–724 (1994).<br />

336 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 337–347, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Colorimetric Characterization Model <strong>for</strong> Plasma Display<br />

Panel<br />

Seo Young Choi, Ming Ronnier Luo and Peter Andrew Rhodes<br />

Department <strong>of</strong> Color & Polymer Chemistry, University <strong>of</strong> Leeds, Leeds, United Kingdom LS2 9JT<br />

E-mail: seoyoung228@googlemail.com<br />

EunGiHeoandImSuChoi<br />

PDP Division, Samsung SDI, 508 Sungsung-Dong, Chonan City, Chungchongnam-Do 330–300, South Korea<br />

Abstract. This paper describes a new device characterization<br />

model applicable to plasma display panels (PDP). PDPs are inherently<br />

dissimilar to cathode ray tube and liquid crystal display devices,<br />

and so new techniques are needed to model their color characteristics.<br />

The intrinsic properties and distinct colorimetric<br />

characteristics are first introduced followed by model development.<br />

It was found that there was a large deviation in colorimetric additivity<br />

and a variation in color due to differences in the number <strong>of</strong> pixels in<br />

a color patch (pattern size). Three colorimetric characterization<br />

models, which define the relationship between the number <strong>of</strong> sustain<br />

pulses and CIE XYZ values, were successfully derived <strong>for</strong> full<br />

pattern size: A three-dimensional lookup table (3D-LUT) model, a<br />

single-step polynomial model and a two-step polynomial model including<br />

three 1D LUTs with a trans<strong>for</strong>mation matrix. The single-step<br />

and two-step polynomial models having more than 8 terms and the<br />

3D LUT model produced the most accurate results. However, the<br />

single-step polynomial model was selected and extended to other<br />

pattern sizes because <strong>of</strong> its simplicity and good per<strong>for</strong>mance. Finally,<br />

a comprehensive model was proposed which can predict CIE<br />

XYZ at sizes different to that used <strong>for</strong> the training set. It was found<br />

that one combined training set <strong>for</strong>med using two different pattern<br />

sizes could give better results than a single-size training set.<br />

© 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4337<br />

INTRODUCTION<br />

Large size displays such as plasma display panel (PDP), liquid<br />

crystal display (LCD), and digital light processing (DLP)<br />

TV are promising candidates <strong>for</strong> replacing the cathode ray<br />

tube (CRT) displays that are currently in widespread use.<br />

Plasma display technology has the following characteristics:<br />

it is thin and light, has superior video image per<strong>for</strong>mance<br />

and uses a large screen size from 42 to 100 inches. These<br />

enable PDPs to be used inside stores <strong>for</strong> product promotion<br />

and, increasingly, <strong>for</strong> home theater. One <strong>of</strong> the desirable<br />

properties <strong>for</strong> large displays is to achieve a “lifelike” appearance<br />

under a range <strong>of</strong> practical viewing conditions as judged<br />

in terms <strong>of</strong> color and image quality. It is there<strong>for</strong>e important<br />

to investigate its colorimetric behavior and to make a characterization<br />

model based on the intrinsic physical properties<br />

<strong>of</strong> a PDP. Already, much research has been conducted to<br />

investigate the use <strong>of</strong> LCD and CRT technology, however<br />

relatively little work has been per<strong>for</strong>med <strong>for</strong> plasma<br />

displays. 1–5 Only one paper deals with PDP characterization<br />

based on the gain-<strong>of</strong>fset-gamma (GOG) model; (previously<br />

developed <strong>for</strong> CRT) at one pattern size, however the physical<br />

properties <strong>of</strong> PDPs and the pattern-size effect were not considered<br />

in this model. 6 In other words, the model could not<br />

be successfully extended to different pattern sizes. The International<br />

Electrotechnical Commission (IEC) has issued a<br />

standard IEC 61966-5 which includes methods and parameters<br />

<strong>for</strong> investigating the use <strong>of</strong> PDPs to display colored<br />

images in multimedia applications. 7 It does include the pattern<br />

size effect as a display area ratio characteristic, but<br />

changes in other characteristics such as color gamut due to<br />

pattern size were not considered. It assumes that a PDP’s<br />

RGB channels are independent. Un<strong>for</strong>tunately, PDPs typically<br />

exhibit significant additivity failure when compared to<br />

other display technologies. It is there<strong>for</strong>e essential that this<br />

additivity failure should be taken into account during the<br />

development <strong>of</strong> any device characterization model.<br />

The simplified structure <strong>of</strong> a PDP RGB cell is shown in<br />

Figure 1. A PDP is composed <strong>of</strong> two glass plates having a<br />

100 m gap and filled with a rare gas mixture which can<br />

<br />

IS&T member.<br />

Received Jul. 27, 2006; accepted <strong>for</strong> publication Mar. 1, 2007.<br />

1062-3701/2007/514/337/11/$20.00.<br />

Figure 1. The structure <strong>of</strong> a PDP’s RGB cells.<br />

337


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

and resultant CIE XYZ values at one particular pattern size.<br />

In addition, one <strong>of</strong> these models was extended to take into<br />

account other pattern sizes.<br />

Figure 2. A 16.7 ms frame includes 8 subfields. The black boxes are the<br />

durations <strong>of</strong> the sustain periods proportional to 1,2,4,…,128.<br />

include a 500 torr, Xe–Ne or Xe–Ne–He mixture which,<br />

when excited, results in the Xe atoms emitting vacuum ultraviolet<br />

(vuv) radiation at 147 and 173 nm, respectively.<br />

This vuv radiation then excites the red, green, and blue<br />

phosphors located on the rear glass plate. The discharge also<br />

emits red-orange visible light due to neon, which causes a<br />

subsequent reduction in color purity (see the Colorimetric<br />

characteristics <strong>of</strong> a PDP section). Each pixel has three individual<br />

RGB microdischarge cells. An alternating current<br />

(ac) is generated by dielectric barrier discharge operating in<br />

a glow regime to generate plasma in each cell. The ac voltage<br />

is approximately rectangular with a frequency in the order <strong>of</strong><br />

100 kHz and a rise time <strong>of</strong> about 200–300 ns. 8 Different<br />

intensity levels are obtained via the modulation <strong>of</strong> the number<br />

<strong>of</strong> ac pulses (sustain pulses) in a discharge cell. For CRT<br />

and LCD, intensity levels are controlled differently from a<br />

PDP, i.e., according to voltage level. In addition, the luminance<br />

output <strong>of</strong> a PDP is dependent on the pattern size<br />

displayed, even when the RGB input values remain the same.<br />

The average level <strong>of</strong> input video signal—a product <strong>of</strong> RGB<br />

input values and pattern size—increases in proportion to the<br />

increase in pattern size. This is also accompanied by an increase<br />

in power consumption. As a result, there is a need to<br />

regulate the power consumed <strong>for</strong> large area patterns by<br />

means <strong>of</strong> the automatic power control (APC) function. Specifically,<br />

this regulates power consumption to within a certain<br />

upper limit. Moreover, luminance output is affected by<br />

the APC function and generates different values dependent<br />

on pattern size.<br />

This paper describes an investigation into the colorimetric<br />

characteristics <strong>of</strong> a PDP and the development <strong>of</strong> three<br />

device characterization models which describe the relationship<br />

between the number <strong>of</strong> sustain pulses <strong>of</strong> RGB input<br />

PHYSICAL PROPERTIES OF A PDP<br />

Overall Transfer Process <strong>of</strong> the Input Video Signal<br />

As mentioned earlier, one unique feature <strong>of</strong> PDP is that, <strong>for</strong><br />

afixedRGB input, its luminance output varies according to<br />

the pattern size displayed. The average level <strong>of</strong> an input<br />

video signal increases not only in proportion to the RGB<br />

input but also to the increase in pattern size. Furthermore,<br />

power demand also grows, because a larger input video signal<br />

leads to a bigger discharge current in the PDP. To protect<br />

the electronic components from damage, it is necessary to<br />

impose a limit on power consumption. This is accomplished<br />

by controlling the discharge current. There are two methods<br />

<strong>for</strong> controlling this: adjusting the number <strong>of</strong> RGB sustain<br />

pulses and modifying the input level <strong>of</strong> the video signal. The<br />

PDP used in this study adopts the first method. The number<br />

<strong>of</strong> RGB sustain pulses is adjusted through the APC function,<br />

which is determined by the average intensity level <strong>of</strong> the<br />

input video signal. To explain the role <strong>of</strong> the APC function<br />

here, it is assumed that each discharge cell can display 256<br />

gray levels. Unlike a CRT, each cell is only capable <strong>of</strong> being<br />

turned on or <strong>of</strong>f (binary). Each gray level is obtained by<br />

modulating the number <strong>of</strong> sustain pulses during one frame<br />

(16.7 ms, 60 frames per second=60 Hz). A frame is divided<br />

into eight subfields, having weight ratios <strong>of</strong> 1, 2, 4…,128<br />

(Figure 2). The function <strong>of</strong> a subfield is to modulate light<br />

output over time. This is accomplished by dividing each<br />

video frame into shorter time periods where each cell is<br />

either turned on or <strong>of</strong>f. Each subfield has a sustain period<br />

(see black box in Fig. 2) whose duration is proportional to<br />

weight ratios, and an address period (see white box in Fig. 2)<br />

whose duration is the same <strong>for</strong> 8 subfields. The address periodisusedtoswitchonor<strong>of</strong>fagivencell.An8-bit<br />

binary<br />

coding is used to obtain 256 gray levels since there are 256<br />

possible levels that can be achieved by assigning on/<strong>of</strong>f to<br />

any combination <strong>of</strong> the eight subfields. In practice, the number<br />

<strong>of</strong> sustain pulses is determined by the sum <strong>of</strong> the product<br />

<strong>of</strong> the sustain pulse limit and the “weight ratios” which<br />

correspond to the proportion <strong>of</strong> subfields turned on. This<br />

calculation is shown in Table I. The sustain pulse limit, as<br />

mentioned previously, safeguards the display’s electronics. It<br />

Table I. One example <strong>of</strong> subfield configuration and the calculation process <strong>for</strong> the number <strong>of</strong> sustain pulses used <strong>for</strong> a color patch with input value <strong>of</strong> 5.<br />

Subfield 1 2 3 4 5 6 7 8<br />

Weight ratios 0.004<br />

1/255<br />

0.008 0.016 0.031 0.063 0.125 0.251 0.502<br />

128/255<br />

Configuration a 1 0 1 0 0 0 0 0 8 bits<br />

The sustain pulse limit, 2600, is given in the APC table defined by the manufacturer<br />

Calculation<br />

26000.004+26000.016=52<br />

the practical number <strong>of</strong> sustain pulses assigned to RGB cells<br />

a Binary coding: 0 is <strong>of</strong>f and 1 is on. This configuration corresponds to input value 5.<br />

Weight<br />

ratios’<br />

sum=1<br />

338 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Figure 3. Flowchart explaining the trans<strong>for</strong>mation <strong>of</strong> input video signal to<br />

the emission <strong>of</strong> light in a PDP.<br />

Figure 4. Plot <strong>of</strong> CIE X values versus the number <strong>of</strong> red sustain pulses at 4<br />

pattern sizes. Points 1, 2, and3 correspond to the maximum number<br />

<strong>of</strong> sustain pulses at 100%, 60%, and 30% pattern size, respectively.<br />

Table II. Maximum CIE XYZ <strong>for</strong> RGB and the range <strong>of</strong> the number <strong>of</strong> sustain pulses at 4, 30, 60, and 100% pattern size.<br />

Color CIE XYZ 4% 30% 60% 100%<br />

Red Max X 498.1 454.5 330.5 230.4<br />

Green Max Y 593.9 546.2 395.7 284.7<br />

Blue Max Z 1311.6 1175.6 809.9 551.7<br />

Sustain pulse range 0–2594 0–2594 0–1826 0–1260<br />

is determined from a manufacturer-defined APC table according<br />

to the average input video signal. Eight subfield<br />

combinations yield 256 gray levels corresponding to the supplied<br />

input values; however, the actual number <strong>of</strong> sustain<br />

pulses (and hence the brightness <strong>of</strong> each level) is controlled<br />

by the sustain pulse limit.<br />

The overall transfer process <strong>of</strong> input video signal to<br />

output stimulus can be expressed in terms <strong>of</strong> the steps<br />

shown in Figure 3, which includes an example <strong>for</strong> a full<br />

white pattern. RGB input values <strong>for</strong> the video signal are<br />

trans<strong>for</strong>med into the number <strong>of</strong> RGB sustain pulses via the<br />

PDP’s logic board. These are calculated from the sub-field<br />

configuration corresponding to the RGB input value and the<br />

sustain pulse limit assigned by the APC table. The number<br />

<strong>of</strong> sustain pulses are the same as the number <strong>of</strong> plasma<br />

discharges occurring in each cell. A succession <strong>of</strong> discharge<br />

pulses occurs between two sustain electrodes inside the front<br />

plate <strong>of</strong> the RGB cells according to the number <strong>of</strong> RGB<br />

sustain pulses assigned. The rare gas mixture (Xe–Ne or<br />

Xe–Ne–He) emits vacuum ultraviolet (vuv) photons at 147<br />

and 173 nm during discharge in RGB cells and those intensities<br />

are governed by the number <strong>of</strong> RGB sustain pulses.<br />

Xenon is used as a vuv emitter and neon acts as a buffer gas<br />

which lowers the breakdown voltage, i.e., the minimum voltage<br />

to initiate plasma. The vuv photons are converted into<br />

visible photons by the phosphor materials deposited on the<br />

inner walls <strong>of</strong> RGB discharge cells. Based on an understanding<br />

<strong>of</strong> this process, the final characterization model was developed<br />

between CIE XYZ values and the number <strong>of</strong> RGB<br />

sustain pulses (rather than simply RGB input values).<br />

Pattern Size Effect<br />

As mentioned in the Overall Transfer Process <strong>of</strong> the Input<br />

Video Signal section brightness varies according to pattern<br />

size. Figure 4 depicts CIE X values plotted against the normalized<br />

number <strong>of</strong> R sustain pulses. In the figure, there are<br />

four sets <strong>of</strong> data points corresponding to pattern sizes <strong>of</strong> 4%,<br />

30%, 60%, and 100% respectively. Each set includes 26 equal<br />

steps <strong>of</strong> the red channel. The decrease <strong>of</strong> maximum X value<br />

with increasing size shown in Fig. 4 can be explained due to<br />

an increased power demand by larger pattern sizes. To limit<br />

the total power consumption <strong>of</strong> a PDP, the number <strong>of</strong> sustain<br />

pulses must be lowered <strong>for</strong> larger color patches by the<br />

APC function. This results in the decrease in maximum X<br />

value with increasing size shown in Fig. 4. Table II illustrates<br />

also the size effect on color patches at 4% and 30% <strong>of</strong> the<br />

total screen area. It can be seen that they have a higher<br />

maximum range <strong>of</strong> the number <strong>of</strong> sustain pulses which result<br />

in larger CIE XYZ values than those <strong>for</strong> the 60% and<br />

100% pattern sizes. The highest number <strong>of</strong> sustain pulses <strong>for</strong><br />

the R primary color at 100% pattern size is 1260 while the<br />

4% pattern size is 2594, even though the input RGB values<br />

are the same (0,255,0). Hence, the resultant maximum Y<br />

value <strong>for</strong> the 4% pattern size is higher than <strong>for</strong> the 100%<br />

pattern size. This implies that the number <strong>of</strong> sustain pulses<br />

should be used as an input color specification, contrary to<br />

conventional approaches to display characterization which<br />

only consider the input digital RGB values.<br />

EXPERIMENTAL METHODS<br />

The colorimetric characterization <strong>of</strong> a 42-inch Samsung<br />

high-definition PDP (model PPM42H3) was evaluated. Its<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 339


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Table III. a Color patches <strong>for</strong> three characterization models at 100% pattern size.<br />

b Color patches used <strong>for</strong> developing single-step polynomial models at other pattern<br />

sizes.<br />

a<br />

Models<br />

at 100%<br />

pattern<br />

3D<br />

LUT<br />

Single-step<br />

polynomial<br />

model<br />

Three<br />

1D LUT<br />

Two-step<br />

polynomial model<br />

Trans<strong>for</strong>mation<br />

matrix<br />

Training<br />

sets<br />

Testing<br />

set<br />

6-<br />

level<br />

3D LUT<br />

6-, 5-, 4-<br />

& 3-level<br />

3D LUT<br />

52 steps<br />

<strong>of</strong> RGB<br />

115-color test set<br />

6-, 5-, 4-<br />

& 3-level<br />

3D LUT<br />

b<br />

4%, 30%, and 60%<br />

Pattern<br />

size<br />

Training<br />

sets<br />

Test sets<br />

5-, 4-, and 3-level 3D LUT<br />

at each pattern size<br />

Three 27-color test sets<br />

at three pattern sizes<br />

pixel resolution is 1024768 with an aspect ratio <strong>of</strong> 16:9<br />

and it is capable <strong>of</strong> addressing 512 intensity levels per channel,<br />

although only 256 were used here. Its pixel pitch is<br />

0.912 mm H0.693 mm V so that the total display area<br />

is 933.89532.22 mm 2 . Color patches were displayed in the<br />

middle <strong>of</strong> the screen and generated by a computercontrolled<br />

graphic card equipped with digital visual interface<br />

(DVI) output. This allows the PDP’s logic board to receive a<br />

digital signal directly from the computer.<br />

Contrary to typical display characterization, the number<br />

<strong>of</strong> sustain pulses was used as input color specification in this<br />

study, as explained previously in the Pattern Size Effect section.<br />

Color measurements were taken using a Minolta<br />

CS-1000 tele-spectroradiometer (TSR) in a dark room. The<br />

repeatability <strong>of</strong> both the TSR and PDP was evaluated using<br />

15 colors measured twice over a two-month period. The<br />

*<br />

median and maximum E ab were 0.38 and 1.18 during this<br />

time. This per<strong>for</strong>mance is considered satisfactory. Measurement<br />

patches consisted <strong>of</strong> rectangles which were 100%, 80%,<br />

60%, 45%, 30%, or 4% <strong>of</strong> the display area. The background<br />

was set to black (except <strong>for</strong> the 100% case).<br />

Three characterization models were first developed <strong>for</strong><br />

the 100% pattern size. Table III(a) describes the data set<br />

generated <strong>for</strong> the three types <strong>of</strong> characterization models: 3D<br />

LUT, single-step polynomial, and two-step polynomial<br />

model. For the 3D LUT model, 6 levels were used. For the<br />

other two models, 6-, 5-, 4-, and 3-level 3D LUTs were compared<br />

in order to determine which training set gave the optimum<br />

per<strong>for</strong>mance with the fewest number <strong>of</strong> measurements.<br />

The 6-level 3D LUT was first generated using 1, 15,<br />

43, 66, 105, and 255 digital input values <strong>for</strong> each <strong>of</strong> the RGB<br />

channels. These values were empirically determined to have<br />

approximately uni<strong>for</strong>m coverage <strong>of</strong> the CIE XYZ destination<br />

color space. The distribution <strong>of</strong> the 6-level training set is<br />

Figure 5. Plot <strong>of</strong> 216 colors <strong>of</strong> the 6-level 3D LUT in a XY, b YZ, and<br />

c a * b * plane, respectively.<br />

shown as an XY and YZ projection in Figures 5(a) and 5(b),<br />

respectively. In addition, an a * b * diagram depicting the<br />

whole <strong>of</strong> the 6-level training set can be seen in Fig. 5(c).<br />

Among the 6 digital input values, 1, 15, 43, 105, and 255<br />

were used to make the 5-level 3D LUT; 1, 15, 66, and 255<br />

340 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

were used <strong>for</strong> the 4-level 3D LUT; and 1, 43, and 255 <strong>for</strong> the<br />

3-level 3D LUT.<br />

Another model is called the “two-step polynomial”<br />

model which includes three 1D-LUTs between the normalized<br />

RGB luminance values and the number <strong>of</strong> RGB sustain<br />

pulses. This is followed by a trans<strong>for</strong>mation from normalized<br />

RGB luminance values to XYZ values via trans<strong>for</strong>mation<br />

matrix. Three 1D-LUTs were created <strong>for</strong> each <strong>of</strong> the<br />

RGB channels including 52 equal steps in RGB space. Linear<br />

interpolation was then used to predict the normalized luminance<br />

values between the data points. Six trans<strong>for</strong>mation<br />

matrices were used. One <strong>of</strong> them was the primary matrix<br />

obtained by measurements <strong>of</strong> RGB primary colors. This matrix<br />

can be used <strong>for</strong> the ideal case where there is little interaction<br />

among RGB channels and little unwanted emission.<br />

<strong>Additional</strong>ly, five trans<strong>for</strong>mation matrices <strong>for</strong> nonideal cases,<br />

were derived using polynomial regression between the measured<br />

XYZ values <strong>for</strong> 6-, 5-, 4-, and 3-level 3D LUTs and<br />

their corresponding normalized RGB luminance values. The<br />

cross terms RG, RB, GB, and RGB were included in the<br />

matrix to compensate <strong>for</strong> cross-channel interaction. The<br />

square terms R 2 , G 2 , and B 2 were also included, although<br />

these terms have no particular physical meaning. These<br />

terms were chosen because they were included in some previous<br />

display characterization studies. 1,9<br />

The per<strong>for</strong>mance <strong>of</strong> the three models was evaluated using<br />

three test sets. The first set includes 444 bright<br />

color patches that were chosen to correspond to L * values <strong>of</strong><br />

45, 85, 95, and 99 <strong>for</strong> each <strong>of</strong> the RGB channels. Two additional<br />

sets, including 24 colors L * 40 and 27 colors composed<br />

<strong>of</strong> three L * values (20, 60, and 90), were also added to<br />

verify model per<strong>for</strong>mance. The three sets were merged to<br />

<strong>for</strong>m a combined test set <strong>of</strong> 115 colors. The color difference<br />

E * ab between the predicted and measured values <strong>for</strong> these<br />

test colors was calculated to evaluate the accuracy <strong>of</strong> characterization<br />

models. All measured tristimulus values were corrected<br />

by subtracting those <strong>of</strong> the black level.<br />

A subsequent experiment was carried out to investigate<br />

model per<strong>for</strong>mance <strong>for</strong> different pattern sizes. Only the<br />

single-step polynomial model was further developed in this<br />

experiment. Three training sets at each pattern size were<br />

used to generate the 3D LUTs [see Table III(b)]. Per<strong>for</strong>mance<br />

was then evaluated by measuring 27 test colors consisting<br />

<strong>of</strong> combinations <strong>of</strong> three input levels producing L *<br />

values <strong>of</strong> 20, 60, and 90 <strong>for</strong> each channel.<br />

Table IV gives the terms used to develop the single-step<br />

model and the trans<strong>for</strong>mation matrices <strong>of</strong> the two-step<br />

model. The polynomial coefficients were computed from experimental<br />

data consisting <strong>of</strong> 216, 125, 64, or 27 colors measured<br />

from the 6-, 5-, 4-, and 3-level 3D LUTs, respectively.<br />

Each sample includes a set <strong>of</strong> RGB sustain numbers and<br />

their corresponding XYZ values. In the two-step model’s<br />

trans<strong>for</strong>mation matrix, a polynomial relationship was determined<br />

between the normalized RGB luminances and XYZ<br />

values. All calculations were executed using Matlab.<br />

Table IV. Detailed description <strong>of</strong> the terms used in the single- and two-step polynomial<br />

models.<br />

The matrices consisting <strong>of</strong><br />

various polynomial coefficients<br />

Independent variables<br />

33 R, G, B<br />

34 R, G, B,1<br />

35 R, G, B, RGB,1<br />

38 R, G, B, RG, RB, GB, RGB,1<br />

311 R, G, B, R 2 , G 2 , B 2 , RG, RB, GB, RGB,1<br />

320 311 plus R 3 , G 3 , B 3 , R 2 G, R 2 B, G 2 R, G 2 B,<br />

B 2 R, B 2 G<br />

335 320 plus R 3 G, R 3 B, G 3 R, G 3 B, B 3 R, B 3 G,<br />

R 2 GB, RG 2 B, RGB 2 , R 4 , G 4 , B 4 , R 2 G 2 , R 2 B 2 , G 2 B 2<br />

COLORIMETRIC CHARACTERISTICS OF A PDP<br />

Spectral Characteristics<br />

The spectral power distributions <strong>of</strong> the maximum intensity<br />

RGB primaries are shown in Figure 6(a). Two kinds <strong>of</strong> green<br />

phosphors were used <strong>for</strong> boosting luminance and stabilizing<br />

discharge: Zn 2 SiO 4 :Mn with a broad band at 526 nm and<br />

YBO 3 :Tb with a sharp peak at 544 nm, respectively. To improve<br />

red saturation, two types <strong>of</strong> red phosphors were<br />

mixed: Y,GdBO 3 :Eu with three main peaks at 593, 611,<br />

and 628 nm, together with YV,PO 4 :Eu which has a sharp<br />

peak at 620 nm. The <strong>for</strong>mer red phosphor appears redorange<br />

due to its main 593 nm emission peak, and it also<br />

possesses the highest conversion efficiency <strong>of</strong> vuv radiation<br />

into red visible light. On the other hand, YV,PO 4 :Eu,<br />

having a sharp main peak at 620 nm, <strong>of</strong>fers good red color<br />

purity. BaMaAl 10 O 17 :Eu was employed as the blue phosphor.<br />

It generates high luminance from a vuv excitation<br />

source but is weak under the harsh conditions <strong>of</strong> high energy<br />

vuv radiation. Consequently, it plays a key role in display<br />

longevity. For these reasons, the spectral properties <strong>of</strong> a<br />

PDP appear to be more complex than the other kinds <strong>of</strong><br />

displays. Fig. 6(b) is an enlargement <strong>of</strong> Fig. 6(a) and illustrates<br />

the red-orange emission <strong>of</strong> Ne gas. The intensity <strong>of</strong><br />

red-orange emission due to Ne gas was quite small compared<br />

with the other main peaks caused by phosphor materials.<br />

Hence the maximum radiance values used <strong>for</strong> Figs.<br />

6(a) and 6(b) are different. The fluctuations cause decrease<br />

in color purity as mentioned earlier. The characteristic redorange<br />

Ne gas emission at 585.2 nm results from atomic<br />

electronic transitions from the higher energy 2p quantum<br />

state to the lower lying 1s energy level. 10<br />

Temporal Stability<br />

A PDP needs time to reach a steady state <strong>for</strong> accurate measurement.<br />

As a result, temporal stability was evaluated over<br />

60 minutes using four pattern sizes, each consisting <strong>of</strong> a<br />

white color as shown in Figure 7. The white color at size 4%<br />

and having the highest number <strong>of</strong> sustain pulses shows the<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 341


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Figure 6. a Spectral power distributions <strong>of</strong> normalized RGB primaries. b Visible emission <strong>of</strong> neon gas<br />

between 580 and 680 nm.<br />

Figure 7. Plot <strong>of</strong> relative luminance values over time <strong>for</strong> white at 4, 30,<br />

60, and 100% pattern size.<br />

Figure 8. a White color patch at 30% pattern size with a black background.<br />

b White color patch at 30% pattern size with a white<br />

background.<br />

largest gap between its initial state and a steady state. For the<br />

4% and 30% color patches, the luminance decreases significantly<br />

at the beginning. Although the time to reach a steady<br />

state is similar <strong>for</strong> all four pattern sizes (around 40 min), the<br />

decrease in luminance is dependent on pattern size. A decrease<br />

in pattern size leads to an increase the number <strong>of</strong><br />

RGB sustain pulses accompanied by an increase in temperature<br />

<strong>of</strong> the RGB cells. If a small color patch is displayed on<br />

a PDP, the initial temperature is higher than that <strong>for</strong> a larger<br />

sized patch; however this is mitigated by a greater rate <strong>of</strong><br />

temperature change. As a result, the time to reach a steady<br />

state is not dependent on pattern size. Conversely, the decrease<br />

in luminance ratio is inversely proportional to pattern<br />

size due to the shorter time needed to reach a stable RGB<br />

cell temperature.<br />

Spatial Independence<br />

The two color patches shown in Figures 8(a) and 8(b) have<br />

the same center square color and different backgrounds.<br />

Spatial independence defines by how much the central white<br />

color is affected by changes in background color. 11 The central<br />

white colors <strong>of</strong> Figs. 8(a) and 8(b) have quite different<br />

CIE XYZ values (445, 444, 550) and (160, 166, 186), respectively,<br />

because the colorimetric characteristics <strong>of</strong> a PDP are<br />

dictated by the number <strong>of</strong> RGB sustain pulses as determined<br />

by the APC rather directly from the RGB input values. For<br />

example, in order to produce the same light output <strong>for</strong> the<br />

white in Figs. 8(a) and 8(b), there is a need to have different<br />

numbers <strong>of</strong> sustain pulses.<br />

Color Gamut<br />

The definition <strong>of</strong> color gamut is the range <strong>of</strong> colors that is<br />

achievable on a given color reproduction medium under a<br />

specified set <strong>of</strong> viewing conditions. Figure 9 shows the color<br />

gamut under dark viewing conditions defined by the primary<br />

and secondary colors <strong>of</strong> a PDP. The ranges <strong>of</strong><br />

tristimulus values displayed differ according to pattern size<br />

(as explained in Pattern size effect section). The color gamuts<br />

<strong>of</strong> four pattern sizes are plotted in a CIELAB a * b * diagram<br />

[Fig. 9(a)]. In addition, color gamuts <strong>of</strong> four pattern<br />

sizes at a hue angle <strong>of</strong> 137° are compared using a CIELAB<br />

C * L * diagram [Fig. 9(b)]. The CIE L * a * b * color coordinates<br />

were calculated based on the peak white at 4% pattern size<br />

representing an L * <strong>of</strong> 100 848.4 cd/m 2 . There<strong>for</strong>e, the<br />

C * L * coordinates <strong>of</strong> the white color patches <strong>for</strong> 30, 60, and<br />

100% pattern sizes show lower values than those at the 4%<br />

pattern size. As pattern size decreases, the range <strong>of</strong> colors<br />

achievable on a PDP becomes larger.<br />

Tone Reproduction Curve<br />

The tone reproduction curve depicts the relationship between<br />

RGB input values and their resultant luminance val-<br />

342 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Figure 9. a Color gamuts <strong>for</strong> the 4, 30, 60, and 100% pattern sizes plotted in an a * b * diagram. b Color<br />

gamuts at a hue angle <strong>of</strong> 137° <strong>for</strong> the 4, 30, 60, and 100% pattern sizes plotted in a C * L * diagram.<br />

ues. The RGB luminances <strong>of</strong> a CRT are controlled by cathode<br />

voltages. For PDPs, on the other hand, the number <strong>of</strong><br />

RGB sustain pulses controls luminance. Figures 10(a) and<br />

10(b) illustrate the intrinsic properties <strong>of</strong> the PDP studied.<br />

Figures 10(a) and 10(b) contain plots <strong>of</strong> normalized Y values<br />

<strong>for</strong> the 4% size (white luminance is 1) against the normalized<br />

number <strong>of</strong> sustain pulses and input values, respectively.<br />

It can be clearly seen in Fig. 10(a) that an increase in the<br />

number <strong>of</strong> RGB sustain pulses leads to an increase in RGB<br />

luminance. Furthermore, a larger patch size has a smaller<br />

range <strong>of</strong> sustain pulse numbers. This is because a larger<br />

pattern size is constrained to a lower maximum number <strong>of</strong><br />

sustain pulses in order avoid exceeding power consumption<br />

limitations. The points indicated by the arrows correspond<br />

to 95% relative luminance, with respect to the maximum<br />

white luminance value at each pattern size.<br />

The intrinsic TRC <strong>of</strong> a PDP [Figs. 10(a) and 10(b)]<br />

should be modified so that the slope <strong>of</strong> the low luminance<br />

range is much smaller than that at high luminances. Figure<br />

10(c) shows the result after gamma is modified by adjusting<br />

the number <strong>of</strong> sustain pulses. The shape <strong>of</strong> the TRC after<br />

gamma modification is the same regardless <strong>of</strong> pattern size.<br />

The usable range <strong>of</strong> the number <strong>of</strong> sustain pulses, however,<br />

depends on pattern size. For example, to produce a white<br />

patch on this PDP using an input value <strong>of</strong> 255, either 466,<br />

766, 1376, or 2594 RGB sustain pulses are assigned, respectively,<br />

<strong>for</strong> the 100%, 60%, 30%, and 4% pattern sizes. The<br />

number <strong>of</strong> sustain pulses available <strong>for</strong> white at 100% pattern<br />

size, 0 to 466, are quantized to 256 levels to make white<br />

luminance follow a power function <strong>of</strong> approximately 2.2 (the<br />

gamma value).<br />

Additivity<br />

The channel additivity was evaluated <strong>for</strong> white at four pattern<br />

sizes. The results are given in Table V in terms <strong>of</strong> percentage<br />

change in tristimulus values and color difference<br />

E * ab . The tristimulus values <strong>of</strong> the RGB patches having<br />

the same number <strong>of</strong> sustain pulses as the peak white patch<br />

were measured to evaluate additivity <strong>for</strong> each pattern size. A<br />

substantial difference between the tristimulus values <strong>for</strong><br />

white and the sum <strong>of</strong> the red, green and blue channels was<br />

found. The latter is larger than the <strong>for</strong>mer by about 15%<br />

*<br />

which corresponds to approximately 6 E ab units. The available<br />

power to a cell drops as other cells become active, leading<br />

to a reduction in brightness. This means that, <strong>for</strong> example,<br />

in terms <strong>of</strong> luminance, R+G+Bwhite. To<br />

counteract the problem <strong>of</strong> deviation from additivity, several<br />

matrix coefficients—including RGB channel cross terms<br />

(RG, RB, GB, and RGB)—were incorporated into the trans<strong>for</strong>mation<br />

matrices <strong>of</strong> the two-step model. In addition, a 3D<br />

LUT model was implemented in which many measured data<br />

points were included so as to compensate <strong>for</strong> the inherent<br />

additivity failure <strong>of</strong> a PDP.<br />

COLORIMETRIC CHARACTERIZATION MODEL FOR<br />

A PDP<br />

Testing the Models’ Per<strong>for</strong>mance at 100% Pattern Size<br />

As introduced in the third section, three types <strong>of</strong> characterization<br />

models were developed: 3D-LUT, single-step polynomial,<br />

and two-step polynomial. These were tested using the<br />

115-test color set. All <strong>of</strong> the models developed here are based<br />

on a 100% pattern size. The results are summarized in Table<br />

*<br />

VI in terms <strong>of</strong> mean and 95th percentile E ab units.<br />

It can be seen that the 3D-LUT model using tetrahedral<br />

interpolation 12 gave a reasonable prediction to the test data<br />

*<br />

with a mean and a 95th percentile <strong>of</strong> 1.3 and 2.5 E ab units.<br />

The two-step model using the primary matrix in which their<br />

coefficients were based on the measurement data gave quite<br />

poor per<strong>for</strong>mance with a mean difference <strong>of</strong> 4.3.<br />

Comparing different single-step polynomial models,<br />

there is a trend that the higher order polynomial models<br />

per<strong>for</strong>med better than the lower ones. However, this is only<br />

true <strong>for</strong> the models developed using more training samples.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 343


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

For those models developed using the 3-level and 4-level<br />

training samples, the higher term polynomial models did<br />

not exhibit more accurate prediction than the lower order<br />

models. This could be caused by over-fitting the measurement<br />

noise when using higher order polynomial models<br />

based on small number <strong>of</strong> training samples. Overall, the<br />

311 polynomial model developed using the 4-level training<br />

data set (which included 64 colors) was found to be<br />

acceptable <strong>for</strong> industrial applications. Using the 320<br />

model can lead to further improvements in the modeling<br />

per<strong>for</strong>mance.<br />

In comparing the single- and two-step models, the<br />

single-step model per<strong>for</strong>med slightly better than the twostep<br />

model except <strong>for</strong> the 33 model. This implies that<br />

single-step polynomial models with a higher order already<br />

consider the cross-talk between different channels in PDPs.<br />

There is needless to include a 1D-LUT normalization.<br />

Figure 11 shows different polynomial per<strong>for</strong>mances <strong>for</strong><br />

*<br />

a single-step model in terms <strong>of</strong> mean E ab using the training<br />

and testing data sets. It can be seen that the models<br />

predicted more accurately when the terms increase until<br />

reaching 311 and 320 polynomial models. For the<br />

335 model, this fits the training data set best, however it<br />

per<strong>for</strong>med poorly <strong>for</strong> the testing data set due to modeling<br />

the noise in the training data set.<br />

In real applications, both the <strong>for</strong>ward and reverse characterization<br />

models are used, i.e., from device signal to XYZ<br />

and vice versa. However, not all models are analytically invertible<br />

and so reverse models were developed having the<br />

same structure as the <strong>for</strong>ward model. The numerical reversibility<br />

<strong>of</strong> the single-step model was also tested. The testing<br />

procedure is shown in Figure 12 and does not require any<br />

color measurement. Here the 115-test color set, defined in<br />

terms <strong>of</strong> XYZ, was again used to first predict RGB sustain<br />

pluses using the reverse model and then further predict the<br />

corresponding XYZ via the <strong>for</strong>ward model. Finally, the color<br />

difference was calculated between the target XYZ and predicted<br />

XYZ values. The results <strong>for</strong> each combination <strong>of</strong> <strong>for</strong>ward<br />

and reverse polynomial model developed by the 4-level<br />

training data are given in Table VII. It can be seen that the<br />

311 polynomial model can give acceptable per<strong>for</strong>mance<br />

(its mean and 95th percentile are 0.3 and 0.8 E * ab , respectively).<br />

This can be further improved by using the 320<br />

model. Both models outper<strong>for</strong>med the other models<br />

studied.<br />

Figure 10. a The relationship between normalized number <strong>of</strong> sustain<br />

pulses and normalized white luminance <strong>for</strong> 4, 30, 60, and 100% pattern<br />

sizes be<strong>for</strong>e the modification <strong>of</strong> gamma. b The relationship between<br />

normalized input values and normalized white luminance <strong>for</strong> four pattern<br />

sizes be<strong>for</strong>e the modification <strong>of</strong> gamma. c The relationship between<br />

normalized input values and normalized white luminance <strong>for</strong> four pattern<br />

sizes after modifying gamma.<br />

Testing the Models’ Per<strong>for</strong>mance at Different Pattern<br />

Sizes<br />

Using the same approach as <strong>for</strong> the 100% pattern size (previous<br />

section), different single-step polynomial models developed<br />

using 3-, 4-, and 5-level 3D-LUT training data <strong>for</strong><br />

each <strong>of</strong> the 4%, 30%, and 60% pattern sizes. Very similar<br />

per<strong>for</strong>mances were found and so only the results from the<br />

30% pattern size are reported in Table VIII in terms <strong>of</strong> mean<br />

*<br />

and 95th percentile E ab values. The results showed that the<br />

311 and 320 polynomial models using 4- or 5-level<br />

training data gave a reasonable prediction. These results are<br />

very similar to those found at 100% pattern size (see Table<br />

VI).<br />

Developing a Single Characterization Model<br />

As mentioned in the Pattern size effect section, light output<br />

is proportional to the number <strong>of</strong> sustain pulses and the<br />

range these are regulated by the APC according to pattern<br />

size. Hence the characterization models developed earlier are<br />

only applicable to a single pattern size. A new method is<br />

developed here which aims to predict the colors displayed at<br />

different pattern sizes. In order to make a single model<br />

which can predict CIE XYZ at pattern sizes other than that<br />

used <strong>for</strong> the training set, it is necessary to select an appropriate<br />

training set covering the whole range <strong>of</strong> sustain pulses<br />

used <strong>for</strong> the test set. For example, it needs to predict CIE<br />

XYZ values <strong>of</strong> several colors in 80% and 45% sizes. There<br />

are two approaches to the selection <strong>of</strong> a training set <strong>for</strong> this<br />

purpose. First, a set having smaller size than 45% can be<br />

used because this can cover a higher range <strong>of</strong> sustain pulses<br />

than those available <strong>for</strong> the 80% and 45% sizes. Second, two<br />

344 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Table V. Tristimulus additivity failure and corresponding color difference <strong>for</strong> white at 4, 30, 60, and 100% pattern size.<br />

Pattern<br />

size<br />

RGB<br />

IVs<br />

Y<br />

cd/m 2 <br />

APC<br />

level<br />

No. <strong>of</strong><br />

sustain X Y Z<br />

*<br />

E ab<br />

100% 255 166.1 255 466 15.1% 13.9% 16.4% 5.6<br />

60% 255 263.2 219 766 16.7% 15.2% 19.2% 6.5<br />

30% 255 443.6 146 1376 17.6% 17.8% 19.9% 6.6<br />

4% 255 848.0 0 2594 16.7% 16.6% 21.1% 6.6<br />

Table VI. Testing the per<strong>for</strong>mance in terms <strong>of</strong> E * ab <strong>of</strong> the characterization models using the 115-color test set. The models were developed based on 6-, 5-, 4- and 3-level training sets.<br />

Training<br />

set<br />

3D<br />

LUT<br />

Single-step polynomial model<br />

Two-step polynomial model<br />

35 38 311 320 335 Primary matrix a 33 34 38 311 320 335<br />

6-level Mean 1.3 3.8 3.1 1.5 1.2 1.2 3.2 3.5 1.7 1.5 1.3 1.4<br />

95th 2.5 9.4 11.7 3.8 2.7 2.8 8.3 8.3 4.5 3.4 2.9 3.0<br />

5-level Mean 3.5 2.8 1.4 1.2 1.1 4.3 3.1 3.2 1.6 1.5 1.3 1.3<br />

95th 8.3 9.4 3.3 2.7 2.3 7.4 7.4 4.1 3.2 2.7 2.5<br />

4-level Mean 3.4 2.5 1.5 1.2 3.9 6.9 3.0 3.0 1.6 1.5 1.3 3.8<br />

95th 7.8 7.8 3.7 2.8 11.7 6.6 6.8 3.6 3.4 3.0 10.3<br />

3-level Mean 3.2 2.3 1.6 7.0 69.5 3.0 3.1 1.5 1.6 3.6 17.9<br />

95th 8.0 7.2 3.5 22.9 305.5 6.8 6.7 3.7 3.6 11.5 55.2<br />

a The primary matrix was obtained from RGB primary colors.<br />

Table VII. Reversibility result <strong>of</strong> polynomials <strong>for</strong> the 4-level training set in terms <strong>of</strong><br />

E ab .<br />

35 38 311 320 335<br />

Mean 2.3 1.6 0.3 0.2 1.7<br />

95th 5.9 4.9 0.8 0.5 4.8<br />

Figure 11. A comparison <strong>of</strong> average E * ab values against the terms used<br />

in the single-step polynomial model <strong>for</strong> the test and training set.<br />

training sets—one set having smaller pattern size than 45%<br />

and another having 100% size—can be combined to make a<br />

new training set. The reasoning behind the second method<br />

is to improve model accuracy. If two color patches at different<br />

pattern sizes but with same RGB sustain pulses are measured,<br />

a subtle difference in their XYZ values could be<br />

found. One practical example is that the luminance values<br />

<strong>for</strong> 30%, 60% and 100% pattern sizes using the same number<br />

<strong>of</strong> sustain pulses (408), are 111, 107, and 104, respectively.<br />

This is due to the available power to a cell being<br />

slightly different because <strong>of</strong> the different number <strong>of</strong> activated<br />

cells at different pattern sizes. Although the first training set<br />

may be sufficient, the polynomial coefficients computed by<br />

the second training set can be expected to take into account<br />

this small color difference due to pattern size. In real applications,<br />

the number <strong>of</strong> sustain pulses used to display complex<br />

images typically corresponds to those associated with<br />

40–50 % pattern sizes. There<strong>for</strong>e, we generated the first<br />

training set as 4-level 3D LUT <strong>for</strong> the 30% pattern size. In<br />

addition, a second training set was produced by combining<br />

two 4-level 3D LUTs <strong>of</strong> 100% and 30% pattern sizes. These<br />

two 4-level 3D LUTs were composed <strong>of</strong> different combinations<br />

<strong>of</strong> RGB sustain pulses. The test data included three<br />

27-color test sets at 80%, 60%, and 45% sizes. Tables IX(a)<br />

and IX(b) summarize the results from the first and second<br />

training sets. The results from the second training set show<br />

that the polynomials with 11, 20, and 35 terms per<strong>for</strong>med<br />

well and gave similar predictive accuracy <strong>for</strong> the three pattern<br />

sizes in Table IX(b). Table IX(a) summarizes the results<br />

<strong>for</strong> the models developed using only the 30% pattern size.<br />

The results in Table IX(b) are worse than those in Table<br />

IX(b) in all cases. This demonstrates that it is better to use<br />

the combined training set <strong>of</strong> 100% and 30% sizes <strong>for</strong> predicting<br />

midsized test colors.<br />

CONCLUSIONS<br />

The physical properties <strong>of</strong> a PDP which affect colorimetric<br />

characterization were examined. Also, colorimetric characteristics<br />

unique to PDP displays were investigated. Among<br />

those, a pattern-size influence and a substantial additivity<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 345


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

Figure 12. The process <strong>for</strong> testing reversibility.<br />

Table VIII. A comparison <strong>of</strong> the per<strong>for</strong>mances in terms <strong>of</strong> E * ab using the 27-color test set and 5-, 4-, and 3-level training sets at 30% pattern<br />

size.<br />

Training<br />

set<br />

Single-step polynomial model<br />

35 38 311 320 335<br />

5-level Mean 3.8 3.3 1.1 1.3 1.5<br />

95th 8.8 9.1 2.2 3.0 3.4<br />

4-level Mean 3.7 3.0 1.4 1.5 3.2<br />

95th 7.6 6.6 2.4 2.6 8.8<br />

3-level Mean 3.9 3.2 2.2 8.6 26.9<br />

95th 7.7 6.8 4.2 18.2 52.0<br />

Table IX. a Comparing models’ per<strong>for</strong>mance E * ab using each 27-color test set at 80%, 60%, and 45% pattern size. Each model was developed<br />

using the 4-level training set at 30% pattern size. b Comparing models’ per<strong>for</strong>mance E * ab using each 27-color test set at 80%, 60%, and 45%<br />

pattern size. Each model was developed using two 4-level training sets at 30% and 100% pattern sizes.<br />

a<br />

Training<br />

set<br />

Test<br />

set<br />

*<br />

E ab<br />

Single-step polynomial model<br />

35 38 311 320 335<br />

30%<br />

pattern<br />

size<br />

4-level<br />

80% Mean 5.1 5.3 2.1 2.5 2.9<br />

95th 9.5 10.6 4.2 5.2 7.6<br />

60% Mean 4.8 4.9 1.9 2.2 2.4<br />

95th 9.5 10.1 4.3 4.8 6.7<br />

45% Mean 3.9 3.7 1.8 1.7 2.4<br />

95th 7.7 7.6 3.2 3.1 6.3<br />

Training<br />

set<br />

Test<br />

set<br />

b<br />

*<br />

E ab<br />

Single-step polynomial model<br />

35 38 311 320 335<br />

Mixture<br />

<strong>of</strong><br />

100%<br />

4-level<br />

&<br />

30%<br />

4-level<br />

pattern<br />

size<br />

80% Mean 4.4 4.8 1.7 1.5 1.1<br />

95th 9.0 11.0 3.5 3.3 2.5<br />

60% Mean 4.2 4.4 1.7 1.5 1.3<br />

95th 8.8 10.3 3.7 2.8 2.3<br />

45% Mean 3.3 3.1 1.6 1.4 1.6<br />

95th 6.8 6.9 2.7 2.0 3.2<br />

failure were found. These must necessarily be considered<br />

when making an accurate colorimetric characterization<br />

model.<br />

Initially, three characterization methods were derived<br />

between the number <strong>of</strong> sustain pulses and CIE XYZ values<br />

at 100% pattern size in order to determine an appropriate<br />

model <strong>for</strong> a PDP. In the <strong>for</strong>ward direction, single- and twostep<br />

polynomial models, which each have more than 8<br />

terms, and a 3D LUT model showed the best results <strong>for</strong> the<br />

6-, 5-, and 4-level training set. However, the single-step<br />

model was eventually selected because <strong>of</strong> its simplicity. The<br />

required number <strong>of</strong> training set samples needed to obtain<br />

good model per<strong>for</strong>mance and requiring the least measurement,<br />

was 64 <strong>for</strong> the 4-level 3D LUT. Also the reversibility <strong>of</strong><br />

the single-step model was evaluated using a 4-level 3D LUT<br />

and this was shown to produce satisfactory results <strong>for</strong> 11-<br />

and 20-term polynomials. There<strong>for</strong>e, the single-step model<br />

was extended to various other pattern sizes. Their results<br />

346 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Choi et al.: Colorimetric characterization model <strong>for</strong> plasma display panel<br />

validated that the polynomial regression method using the<br />

4-level training set was a good characterization model <strong>for</strong><br />

this PDP display.<br />

Finally, one comprehensive training set consisting <strong>of</strong><br />

two 4-level 3D LUTs corresponding to 100% and 30% pattern<br />

sizes were produced to predict CIE XYZ at intermediate<br />

pattern sizes (i.e., sizes which were not present in the training<br />

set). These outcomes demonstrated that the single-step<br />

model could be successfully applied to estimate colors at<br />

different pattern sizes using just one combined training set.<br />

REFERENCES<br />

1 N. Katoh, T. Deguchi, and R. S. Berns, “An Accurate Characterization <strong>of</strong><br />

CRT Monitor (I) Verification <strong>of</strong> Past Studies and Clarification <strong>of</strong><br />

Gamma”, Opt. Rev. 8, 305 (2001).<br />

2 N. Katoh, T. Deguchi, and R. S. Berns, “An Accurate Characterization <strong>of</strong><br />

CRT Monitor (II) Proposal <strong>for</strong> an Extension to CIE Method and its<br />

Verification”, Opt. Rev. 8, 397 (2001).<br />

3 M. D. Fairchild and J. E. Gibson, “Colorimetric Characterization <strong>of</strong><br />

Three Computer Displays (LCD and CRT)”, Munsell Color <strong>Science</strong><br />

Laboratory Technical Report, http://www.cis.rit.edu/mcsl/research/PDFs/<br />

GibsonFairchild.pdf (2000).<br />

4 Y. S. Kwak and L. MacDonald, “Characterization <strong>of</strong> a Desktop LCD<br />

Projector”, Displays 21, 179 (2000).<br />

5 D. R. Wyble and H. Zhang, “Colorimetric Characterization Model <strong>for</strong><br />

DLP Projectors”, Proc. IS&T/SID 11th Color <strong>Imaging</strong> Conference (IS&T,<br />

Springfield, VA, 2003), pp. 346–350.<br />

6 G. Kutas and P. Bodrogi, “Colorimetric Characterization <strong>of</strong> HD-PDP<br />

Device”, in IS&T’s 2nd European Conference on Color Graphics, <strong>Imaging</strong><br />

and Vision, (IS&T, Springfield, VA, 2004,). pp. 65–69.<br />

7 Multimedia Systems and Equipment—Color Measurement and<br />

Management, Part 5: Equipment using Plasma Display Panels, IEC<br />

61966–5, 2001.<br />

8 J. P. Boeuf, “Plasma Display Panels: Physics, Recent Developments and<br />

Key Issues”, J. Phys. D 36, R53 (2003).<br />

9 P. Bodrogi and J. Schanda, “Testing a Calibration Method <strong>for</strong> Color CRT<br />

Monitors. A Method to Characterize the Extent <strong>of</strong> Spatial<br />

Interdependence and Channel Interdependence”, Displays 16(3), 123<br />

(1995).<br />

10 R. S. Van Dyck, C. E. Johnson, and H. A. Shugart, “Lifetime Lower<br />

Limits <strong>for</strong> the 3p 0 and 3p 2 Metastable States <strong>of</strong> Neon, Argon and<br />

Krypton”, Phys. Rev. A 5, 991 (1972).<br />

11 P. Green and L. MacDonald, Color Engineering: Achieving Device<br />

Independent Color (John Wiley and Sons Ltd, West Sussex, UK, 2002),<br />

p. 158.<br />

12 H. R. Kang, Color Technology <strong>for</strong> Electronic <strong>Imaging</strong> Devices (SPIE<br />

Optical Engineering Press, Bellingham, WA, 1997), p. 64.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 347


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 348–359, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Real-Time Color Matching Between Camera and LCD<br />

Based on 16-bit Lookup Table Design in Mobile Phone<br />

Chang-Hwan Son<br />

School <strong>of</strong> Electrical Engineering and Computer <strong>Science</strong>, Kyungpook National University, 1370,<br />

Sankyuk-dong, Buk-gu, Daegu 702-701, Korea<br />

Cheol-Hee Lee<br />

Major <strong>of</strong> Computer Engineering, Andong National University, 388, Seongcheon-dong, Andong,<br />

Gyeongsangbuk-Do 760-747, Korea<br />

Kil-Houm Park and Yeong-Ho Ha <br />

School <strong>of</strong> Electrical Engineering and Computer <strong>Science</strong>, Kyungpook National University, 1370,<br />

Sankyuk-dong, Buk-gu, Daegu 702-701, Korea<br />

E-mail: yha@ee.knu.ac.kr<br />

Abstract. Based on the concept <strong>of</strong> multimedia convergence, imaging<br />

devices, such as cameras, liquid crystal displays (LCDs), and<br />

beam projectors, are now built-in to mobile phones. As such, mobile<br />

cameras capture still images or moving pictures, then store them as<br />

digital files, making it possible <strong>for</strong> users to replay moving pictures<br />

and review captured still images. Increasingly, users want LCD in<br />

the mobile phone (we call it mobile LCD hereafter) to reproduce the<br />

same colors as the real scene. Accordingly, this paper proposes a<br />

method <strong>for</strong> color matching between mobile camera and mobile LCD<br />

that includes characterizing the mobile camera and mobile LCD,<br />

gamut mapping, camera noise reduction, and a 16-bit lookup table<br />

(LUT) design. First, to estimate the CIELAB values <strong>for</strong> the objects in<br />

the real scene, mobile camera characterization is achieved through<br />

polynomial regression <strong>of</strong> the optimal order determined by investigating<br />

the relation between captured RGB values and measured<br />

CIELAB values <strong>for</strong> a standard color chart. Thereafter, mobile LCD<br />

characterization is conducted based on 16-bit/pixel processing because<br />

<strong>of</strong> the reduced bit depth <strong>of</strong> the images displayed on a mobile<br />

LCD. In addition, a sigmoid model is used to find the luminance<br />

value corresponding to the RGB control signal, instead <strong>of</strong> using gain<br />

<strong>of</strong>fset gamma and S-curve models due to the adjustment <strong>of</strong> luminance<br />

curve made by a system designer <strong>for</strong> preference color reproduction.<br />

After completing the two types <strong>of</strong> characterization, gamut<br />

mapping is per<strong>for</strong>med to connect the source medium (mobile camera)<br />

with the target medium (mobile LCD), then a combination <strong>of</strong><br />

sigmoid functions with different parameters to control the shape is<br />

applied to the luminance component <strong>of</strong> the gamut-mapped CIELAB<br />

values to reduce camera noise. Finally, a three-dimensional RGB<br />

LUT is constructed using 16-bit/pixel-based data to enable color<br />

matching <strong>for</strong> moving pictures and inserted into the mobile phone.<br />

Experimental results show that moving pictures transmitted by a<br />

mobile camera can be realistically reproduced on a mobile LCD<br />

without any additional computation or memory burden. © 2007 <strong>Society</strong><br />

<strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4348<br />

<br />

IS&T Member.<br />

Received Dec. 1, 2006; accepted <strong>for</strong> publication Mar. 30, 2007.<br />

1062-3701/2007/514/348/12/$20.00.<br />

INTRODUCTION<br />

With the appearance <strong>of</strong> multimedia convergence in mobile<br />

phones that can now provide such functions as web browsing,<br />

3D games, television broadcasting, and image capturing,<br />

in addition to communication, manufacturers have invested<br />

heavily in super highway communication network, nextgeneration<br />

memory chips, and encryption technology <strong>for</strong><br />

reliable e-commerce operations. The use <strong>of</strong> color reproduction<br />

technology in mobile phones has also been recently<br />

introduced to support the development <strong>of</strong> mobile cameras,<br />

mobile beam projectors, and mobile liquid crystal displays<br />

(LCDs). In particular, with the rapid increase in mobile<br />

cameras, mobile phones can now capture and store still images<br />

or moving pictures as digital files, making it possible <strong>for</strong><br />

users to replay the moving pictures and review captured still<br />

images anytime and anywhere. However, mobile LCDs are<br />

currently unable to reproduce the original colors captured by<br />

a mobile camera due to a reduced bit-depth, lower backlight<br />

luminance, and weak resolution. 1 In addition, mobile cameras<br />

have a small lens, low dynamic range, and poor modulation<br />

transfer function (MTF), plus each device senses or<br />

displays in a different way, as they have unique<br />

characteristics. 2 As a result, there is a significant difference in<br />

the color appearance when captured images are displayed on<br />

a mobile LCD. There<strong>for</strong>e, real-time colormatching between<br />

mobile camera and mobile LCD in a mobile phone needs to<br />

be considered to ensure a better image quality.<br />

The aim <strong>of</strong> color matching is to achieve color consistency<br />

even when an image moves across various devices and<br />

undergoes many color trans<strong>for</strong>mations. 3 Several color<br />

matching approaches have already been suggested, <strong>for</strong> example,<br />

a simple method is to transmit the RGB digital values<br />

from the original device to the reproducing device, referred<br />

to as device-dependent color matching. Yet, since this<br />

method is no more than physical data transmission, accurate<br />

color matching cannot be achieved across various devices.<br />

Meanwhile, spectral-based approaches match the spectral reflectance<br />

curves <strong>of</strong> the original and reproduced colors, so the<br />

original and reproduction look the same under any il-<br />

348


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 1. The block diagram <strong>of</strong> the proposed method.<br />

luminant: i.e., there is no metamerism. However, the computation<br />

<strong>of</strong> reflectance is very complex and time consuming,<br />

making a spectral-based approach inappropriate <strong>for</strong> realtime<br />

color matching. Another method is colorimetric color<br />

matching to reproduce the same CIE chromaticity and relative<br />

luminance compared with the original color. This has<br />

already been widely applied to imaging devices, such as<br />

monitors, printers, and scanners based on the international<br />

color consortium (ICC) pr<strong>of</strong>ile, yet not to mobile phones,<br />

which have only been considered as a means <strong>of</strong> communication<br />

until quite recently. However, with multimedia convergence,<br />

mobile manufacturers have become aware <strong>of</strong> the<br />

importance <strong>of</strong> the ICC pr<strong>of</strong>ile <strong>for</strong> color matching between<br />

mobile cameras and mobile LCDs. Accordingly, this paper<br />

presents a real-time color matching system <strong>for</strong> mobile cameras<br />

and mobile LCDs based on the concept <strong>of</strong> the ICC<br />

pr<strong>of</strong>ile.<br />

The proposed color matching system is composed <strong>of</strong><br />

four steps: Characterization <strong>of</strong> the mobile LCD and mobile<br />

camera, gamut mapping, noise reduction, and a<br />

16-bit-based lookup table (LUT) design. The device characterization<br />

defines the relationship between the tristimulus<br />

values (CIEXYZ or CIELAB) and RGB digital values. In general,<br />

mobile camera characterization is modeled by a polynomial<br />

regression, and the more the polynomial order increases,<br />

the better the per<strong>for</strong>mance. However, <strong>for</strong> a higher<br />

polynomial order, most estimated tristimulus values exceed<br />

the boundary <strong>of</strong> the maximum lightness and chroma, making<br />

the implementation <strong>of</strong> mobile camera characterization<br />

difficult, as the relation between the tristimulus values and<br />

digital RGB values has not been analyzed. Thus a polynomial<br />

order is suggested based on investigating the relation<br />

between RGB digital values trans<strong>for</strong>med using the opponent<br />

color theory and CIELAB values. Meanwhile, <strong>for</strong> the mobile<br />

LCD characterization, a sigmoid function instead <strong>of</strong> a conventional<br />

method, such as the gain <strong>of</strong>fset gamma (GOG) or<br />

S-curve model, is used to estimate the luminance curve<br />

made by the system designer to achieve a preferable color<br />

reproduction or to improve the perceived contrast <strong>of</strong> the<br />

mobile LCD. Furthermore, the characterization is conducted<br />

based on 16-bit data processing, as a mobile LCD is controlled<br />

based on 16-bit data, in contrast to digital TVs or<br />

monitors with 24-bit data. After completing the two types <strong>of</strong><br />

characterization, a gamut-mapping algorithm is applied to<br />

connect the source medium (mobile camera) with the target<br />

medium (mobile LCD).<br />

Although the three processes mentioned above are sufficient<br />

to obtain colorimetric color matching <strong>for</strong> still images,<br />

noise reduction and an LUT design still need to be considered<br />

to achieve real-time color matching <strong>for</strong> moving pictures.<br />

In a mobile camera, various camera noises, such as<br />

CCD noise and thermal noise, are incorporated into moving<br />

pictures and further amplified after color matching, thereby<br />

degrading the image quality, especially in the dark region <strong>of</strong><br />

the achromatic axis. Thus, to solve this problem, a combination<br />

<strong>of</strong> two sigmoid functions with different parameters to<br />

control the shape is applied to the lightness component <strong>of</strong><br />

the gamut-mapped tristimulus values to change the contrast<br />

ratio. As a result, the lightness values <strong>for</strong> the camera noise<br />

are reduced in the dark region <strong>of</strong> the achromatic axis,<br />

thereby reducing the amplified camera noises. In addition, a<br />

three-dimensioinal (3D) RGB LUT is designed based on<br />

16-bit data to reduce the complex computation <strong>of</strong> serialbased<br />

processing and facilitate color matching <strong>for</strong> moving<br />

pictures.<br />

PROPOSED METHOD<br />

Figure 1 shows a block diagram <strong>of</strong> the proposed algorithm<br />

that can achieve real-time color matching between a mobile<br />

camera and a mobile LCD. First, to predict the CIELAB<br />

values <strong>of</strong> arbitrary objects in a real scene, the mobile camera<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 349


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

characterization is conducted by finding the relation between<br />

the RGB digital values <strong>of</strong> a standard color chart captured<br />

in a lighting booth and CIELAB values measured using<br />

a colorimeter. The CIELAB values estimated from the<br />

mobile camera characterization <strong>of</strong> the input RGB values are<br />

then trans<strong>for</strong>med into an achievable color range that can be<br />

reproduced by the mobile LCD, referred to as gamut mapping.<br />

Next, the lightness values <strong>of</strong> the gamut-mapped<br />

CIELAB values are changed using the parameters <strong>of</strong> a sigmoid<br />

function provided by a visual experiment to reduce the<br />

camera noise incorporated into a moving picture, then combined<br />

with two untouched color signals. Thereafter, the<br />

modified gamut-mapped CIELAB values are converted into<br />

color matched RGB values <strong>for</strong> display on the mobile LCD<br />

based on a sigmoid-based mobile-LCD characterization to<br />

consider the luminance curve adjusted <strong>for</strong> the preferred<br />

color reproduction, along with 16-bit data processing due to<br />

the reduced bit depth in the mobile LCD. Finally, a 3D-RGB<br />

LUT is constructed using the 16-bit/pixel-based data to enable<br />

color matching <strong>for</strong> moving pictures and inserted into<br />

the mobile phone, thereby allowing 24-bit moving pictures<br />

to be reproduced on a mobile LCD with a higher quality<br />

image.<br />

CHARACTERIZATION OF THE MOBILE LCD BASED<br />

ON 16-bit DATA PROCESSING<br />

The display characterization predicts the tristimulus value<br />

<strong>for</strong> the input digital value and may be conducted by a<br />

measurement-based approach or modeling-based<br />

approach. 4–6 A measurement-based approach measures a lot<br />

<strong>of</strong> patches made by combination <strong>of</strong> input digital values using<br />

a colorimeter and estimates the tristimulus value by the<br />

interpolation method or polynomial regression <strong>for</strong> an arbitrary<br />

digital value. There<strong>for</strong>e, this approach improves the<br />

characterization accuracy, yet requires a lot <strong>of</strong> measurement<br />

data and extensive memory and is relatively complex. Meanwhile,<br />

a modeling-based approach finds the relationship between<br />

the digital input data and tristimulus value based on a<br />

mathematical function with a smaller number <strong>of</strong> data measurements.<br />

The GOG and S-curve models have been used as<br />

typical mathematical functions and have been applied to different<br />

types <strong>of</strong> display. In general, the GOG model is appropriate<br />

<strong>for</strong> CRT display because its electro-optical transfer<br />

function, the relationship between the grid voltage and beam<br />

current, follows a power curve shape, while LCD display is a<br />

binary device that switches from an OFF state to an ON state<br />

and follows the S-shaped curve, thereby adapting the<br />

S-curve model <strong>for</strong> LCD characterization. The overall procedure<br />

<strong>of</strong> modeling-based approach is identical except that<br />

electro-optical transfer function is modeled with a different<br />

mathematical function. The first step <strong>of</strong> modeling-based<br />

characterization is to convert the digital value to luminance<br />

value <strong>for</strong> each RGB channel. This can be done by estimating<br />

the coefficient <strong>of</strong> mathematical function with optimization<br />

programming. In the case <strong>of</strong> the GOG model, mathematical<br />

function can be described as<br />

Y ch =k g,ch d ch<br />

2 N −1 + k o,ch,ch<br />

, 1<br />

where ch represents RGB channel, d ch is input digital value,<br />

and N is the bit number; k g,ch ,k o,ch , are the gain, <strong>of</strong>fset,<br />

and gamma parameters <strong>of</strong> the GOG model, respectively. Y ch<br />

is the normalized luminance value corresponding to the normalized<br />

input digital value <strong>for</strong> each channel. To get all parameters<br />

<strong>of</strong> the GOG model, the digital value <strong>of</strong> each channel<br />

is independently sampled by a uni<strong>for</strong>m M interval,<br />

which assumes no channel interaction that the light emitted<br />

from a pixel location is dependent only on R, G, B triplet <strong>for</strong><br />

that pixel and is independent <strong>of</strong> input digital value <strong>for</strong> other<br />

pixels. 7 Then, the CIEXYZ values <strong>for</strong> M-sample digital values<br />

are acquired by measuring the displayed patches created<br />

by M-sampled digital values with a colorimeter. At this time,<br />

even though displayed patches are made by 8-bit data <strong>for</strong><br />

each channel, the 8-bit based M-sampled RGB digital values<br />

corresponding to the measured CIEXYZ values must be<br />

practically converted to 5,6,5-bit data in the mobile LCD,<br />

and thus the digital values <strong>of</strong> the 8-bit based R-channel and<br />

B-channel are divided by 8, while that <strong>of</strong> the G-channel is<br />

dividedby4:<br />

d R = d R<br />

2 R, d G = d G<br />

2 G, d B = d B<br />

, 2<br />

B<br />

2<br />

where d R , d G , and d B is the digital value <strong>of</strong> displayed patch<br />

<strong>for</strong> each channel and R,G,B is the difference <strong>of</strong> bitnumber<br />

between 8-bit/channel <strong>of</strong> patches and bit-number/<br />

channel in mobile LCD. There<strong>for</strong>e, d ch in Eq. (1) is substituted<br />

with d R ,d G ,d B , which is used to find the luminance<br />

curve <strong>of</strong> mobile LCD.<br />

Of the measured CIEXYZ values, the Y values are selected<br />

as Y ch , assuming that the shapes <strong>of</strong> X, Y, and Z are<br />

identical after normalization, which is referred to the<br />

channel-chromaticity constancy that spectrum <strong>of</strong> light from<br />

a channel has the same basic shape and only undergoes a<br />

scaling in amplitude as the digital value <strong>for</strong> each channel is<br />

varied. 7 Finally, the pairs <strong>of</strong> M-sampled digital value<br />

d R ,d G ,d B and Y values are substituted in Eq. (1), yielding<br />

all parameters <strong>of</strong> the GOG model using optimization nonlinear<br />

programming. The second step is to trans<strong>for</strong>m the<br />

luminance value <strong>of</strong> each channel calculated by the GOG<br />

model to the CIEXYZ value. This stage can be simply<br />

achieved by a matrix operation:<br />

X r,max X g,max X R<br />

Y Y r,max Y g,max Y b,max Y G<br />

Z=X Z r,max Z g,max Z b,maxY Y B,<br />

where Y R , Y G , and Y B are the luminance values <strong>of</strong> each channel,<br />

Y ch=R,G,B . In each column, the matrix coefficients are the<br />

CIEXYZ value at the maximum digital value <strong>of</strong> each channel,<br />

and can be directly measured with a colorimeter.<br />

Through the above-mentioned two steps, display characterization<br />

can be accomplished. In the case <strong>of</strong> the S-curve<br />

3<br />

350 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 2. Electro-optical transfer function <strong>for</strong> mobile LCD; a GOG model, b GOG model except saturation<br />

region, c S-curve model, and d sigmoid model.<br />

model, only the power-curved function shown in Eq. (1) is<br />

replaced with the S-shaped mathematical function in the<br />

process <strong>of</strong> display characterization<br />

d ch /2 N −1 ch<br />

Y ch =A ch<br />

d ch /2 N −1 ch + C ch ,<br />

where A ch , ch , ch , and C ch are parameters, respectively.<br />

Equation (4) has various S-shaped curves according to the<br />

parameter values, and if both ch , and C ch is zero, Eq. (4)<br />

follows the gamma curve as Eq. (1). All parameters in Eq.<br />

(4) can be obtained by applying the same process, the first<br />

step explained in the GOG model. Using these parameters,<br />

the input digital value is converted into the luminance value<br />

and then trans<strong>for</strong>med into the CIEXYZ value through a matrix<br />

operation.<br />

To conduct the characterization <strong>of</strong> mobile LCD, we apply<br />

conventional methods to a cellular phone, a Samsung<br />

SCH-500. In a mobile phone, each RGB pixel value is represented<br />

by (5,6,5) bit and image size is fixed at 240320.<br />

Figures 2(a)–2(c) show the electro-optical transfer function<br />

4<br />

resulting from the GOG model, the GOG model without the<br />

saturation region, and the S-curve model. In Figure 2, three<br />

types <strong>of</strong> lines represent the estimated luminance values obtained<br />

by using conventional characterization <strong>for</strong> each channel,<br />

while three types <strong>of</strong> marks indicate the measured luminance<br />

values <strong>for</strong> each channel. In Fig. 2, the shape <strong>of</strong> the<br />

electro-optical transfer function <strong>for</strong> mobile display is different<br />

from a power-curved shape <strong>of</strong> CRT display or S-curved<br />

shape <strong>of</strong> LCD display. As the input digital value moves toward<br />

the middle point, the gradient <strong>of</strong> the luminance curve<br />

rapidly increases and immediately decreases, producing a<br />

saturation region. This is due to the adjustment <strong>of</strong> the luminance<br />

curve by the system designer intended to enhance<br />

the contrast ratio and overcome the low channel-bit number.<br />

As a result, a conventional GOG model or S-curve model<br />

does not follow the luminance curve <strong>of</strong> the saturation region<br />

and is not directly applied to mobile LCD. There<strong>for</strong>e, we<br />

used the sigmoid function to model the electro-optical<br />

transfer function <strong>of</strong> mobile LCD based on visual observation<br />

<strong>of</strong> the luminance curve. The sigmoid function is expressed<br />

as<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 351


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Table I. Estimated parameter <strong>of</strong> sigmoid function.<br />

a-parameter<br />

c-parameter<br />

R-channel 11.9149 0.4647<br />

G-channel 11.1892 0.4508<br />

B-channel 11.3273 0.4359<br />

Table II. Per<strong>for</strong>mance <strong>of</strong> mobile LCD characterization with various methods.<br />

*<br />

Average E ab<br />

*<br />

Maximum E ab<br />

GOG model 15.655 32.4424<br />

GOG model except<br />

8.7614 17.5898<br />

saturation region<br />

S-curve model 6.9801 15.2279<br />

Sigmoid model 3.9683 14.6831<br />

1<br />

sigmoidx,a,c =<br />

1 + exp− ax − c . 5<br />

The sigmoid function is a symmetrical function with respect<br />

to c and is a constant value if it is zero. The shape <strong>of</strong> a<br />

sigmoid function depends on the absolute value <strong>of</strong> a, and as<br />

the absolute value <strong>of</strong> a increases, the gradient <strong>of</strong> the sigmoid<br />

function rapidly increases with respect to c. Figure 2(d)<br />

shows the electro-optical transfer function resulting from<br />

the sigmoid model. In Fig. 2(d), the estimated luminance<br />

curve closely follows the measured luminance value and it is<br />

predicted that the estimation error will be reduced. The estimated<br />

coefficients <strong>of</strong> the sigmoid function are shown in<br />

Table I. The estimated curve is nearly symmetric with respect<br />

to 0.45 and the absolute value <strong>of</strong> a to determine the<br />

shape <strong>of</strong> the sigmoid function is independent <strong>of</strong> channel and<br />

is almost the same.<br />

To evaluate the per<strong>for</strong>mance <strong>of</strong> each method, the<br />

CIE1976 color difference E * ab was used to measure the<br />

characterization error, which is the Euclidian distance between<br />

estimated CIELAB value and measured CIELAB<br />

value. Sixty-four patches were tested and Table II shows the<br />

characterization error <strong>of</strong> various model-based methods. The<br />

GOG model had the largest color difference and the characterization<br />

error was still severe although the GOG model was<br />

used except in the saturation region. For the S-curve model,<br />

*<br />

the average E ab was approximately 6.9 and is normal color<br />

difference. However, in the middle region, the estimated luminance<br />

value shows a significant difference compared with<br />

measured luminance value. The sigmoid model has a good<br />

average color difference smaller than 6.0, which is indistinguishable<br />

in human vision.<br />

DECISION OF POLYNOMIAL ORDER FOR THE<br />

CHARACTERIZATION OF MOBILE CAMERA<br />

The camera characterization is to find the relationship between<br />

the tristimulus value and digital RGB value. Through<br />

the accurate camera characterization, we can get in<strong>for</strong>mation<br />

about an object color in real scene and reproduce the object<br />

color on mobile LCD. The general procedure <strong>of</strong> camera<br />

characterization is shown in Figure 3. First, a standard color<br />

chart such as a Macbeth or Gretag Color Chart is placed<br />

with 0/45° geometry in a lighting booth, where an illuminant<br />

is set at D65 to reflect the perceived color corresponding<br />

to a daylight condition. 8 The standard color chart<br />

is then captured by a mobile camera set with aut<strong>of</strong>ocusing to<br />

avoid color clipping. Captured RGB digital values <strong>of</strong> each<br />

patch in the standard color chart are averaged to reduce the<br />

camera noise and nonuni<strong>for</strong>mity <strong>of</strong> illumination. Next, the<br />

tristimulus value <strong>of</strong> the standard color chart is acquired by<br />

measuring each patch <strong>of</strong> the color chart or standard data<br />

provided by the manufacturer. Finally, polynomial regression<br />

with least square fitting is applied to find the relationship<br />

between captured RGB digital values and measured<br />

tristimulus values. 9,10<br />

In general, the per<strong>for</strong>mance <strong>of</strong> camera characterization<br />

becomes better as the polynomial order increases. Practically,<br />

Figure 3. The procedure <strong>for</strong> mobile camera characterization.<br />

352 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 4. The characteristics <strong>of</strong> cellular camera; a L * vs average RGB<br />

value <strong>of</strong> gray sample, b a * vs R−B, andb * vs G−B.<br />

<strong>for</strong> a higher polynomial order, most estimated tristimulus<br />

values exceed the boundary <strong>of</strong> the maximum lightness and<br />

chroma, and there is difficulty in implementing mobile camera<br />

characterization. This is because the characteristic curve<br />

<strong>of</strong> the mobile camera, i.e., the relationship between the<br />

tristimulus value and digital RGB value, is not analyzed to<br />

suggest an appropriate polynomial order. To determine the<br />

polynomial order, the RGB digital value is manipulated<br />

based on opponent color theory and is compared with the<br />

Figure 5. The characteristics <strong>of</strong> PDA camera; a L * vs average RGB<br />

value <strong>of</strong> gray sample, b a * vs R−B, andb * vs G−B.<br />

CIELAB value. The CIELAB space is an opponent color coordinate<br />

composed <strong>of</strong> a lightness signal and two types <strong>of</strong><br />

color signals obtained by the difference <strong>of</strong> three color signals.<br />

Thus the RGB digital value is trans<strong>for</strong>med into a lightness<br />

signal and two color signals, R+G+B/3, R−B, and<br />

G−B, just as in opponent color space. Figures 4 and 5<br />

show the relationship between the manipulated RGB values<br />

and CIELAB values <strong>for</strong> a cellular camera and PDA camera.<br />

From a visual evaluation, the distribution <strong>of</strong> the measure-<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 353


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

ment data was found to be slightly dispersed due to nonuni<strong>for</strong>m<br />

illumination intensities according to the spatial position<br />

on the color chart, where the lux-meter measurements<br />

<strong>for</strong> the four corners were 1990 lux, 2055 lux, 1955 lux, and<br />

1922 lux, respectively. Although ef<strong>for</strong>ts were made to correct<br />

the nonuni<strong>for</strong>mity <strong>of</strong> the illumination intensity, the modeling<br />

<strong>of</strong> lux-meter measurements according to their distance<br />

from the center <strong>of</strong> the color chart is not trivial work due to<br />

their random distribution, there<strong>for</strong>e, this issue has been carried<br />

over to future work. However, it was still clear that the<br />

manipulated RGB digital values were roughly linear to the<br />

CIELAB values:<br />

R + G + B L * , R − B a * , G − B b * . 6<br />

3<br />

There<strong>for</strong>e, a first-order polynomial is adopted, and mathematical<br />

modeling <strong>of</strong> mobile camera characterization is expressed<br />

as linear equations:<br />

L * =1+ L,R R + L,G G + L,B B,<br />

a * =1+ a,R R + a,G G + a,B B,<br />

b * =1+ b,R R + b,G G + b,B B.<br />

Equation (7) can be equally expressed in vector <strong>for</strong>m,<br />

P = V T ,<br />

7<br />

1 , ...,1 n<br />

L,1 , a,1 , b,1<br />

R 1 , ...R n L,2 , a,2 , b,2<br />

V =1 = 8<br />

G 1 , ...G n L,3 , a,3 , b,3<br />

B 1 , ...B n, L,4 , a,4 , b,4,<br />

1<br />

P =L * , a * *<br />

1 , b 1<br />

· · ·<br />

· · ·<br />

L * n , a * n , b<br />

*,<br />

n<br />

where n is the patch number <strong>of</strong> color chart. The ultimate<br />

goal <strong>of</strong> mobile camera characterization is deriving the coefficients<br />

<strong>of</strong> the linear equation, which can be obtained by<br />

pseudoinverse trans<strong>for</strong>mation <strong>of</strong> Eq. (8):<br />

= VV T −1 VP.<br />

Using the derived coefficients, an arbitrary captured digital<br />

value can be converted into the CIELAB value. However,<br />

some <strong>of</strong> the estimated CIELAB values may exceed the maximum<br />

value <strong>of</strong> CIELAB space caused by the error <strong>of</strong> linear<br />

regression. To solve this problem, the lightness value is subtracted<br />

from the amount <strong>of</strong> excessive lightness, and two<br />

color signals are linearly compressed while preserving their<br />

hue value:<br />

9<br />

Table III. Estimation errors <strong>of</strong> mobile camera characterization.<br />

Cellular camera<br />

Samsung SCH-100<br />

PDA camera<br />

Samsung SPH-M400<br />

*<br />

Average E ab<br />

L * = L * *<br />

− L max − 100,<br />

*<br />

Maximum E ab<br />

4.3605 12.2098<br />

6.2638 16.8828<br />

10<br />

a * = k a * *<br />

/a max , b * = b * /a * a * , 11<br />

where k is the constant value <strong>for</strong> color-signal compression.<br />

L max and a max are the estimated maximum lightness value<br />

and color signal value, respectively. Table III shows the per<strong>for</strong>mance<br />

<strong>of</strong> mobile camera characterization <strong>for</strong> the color<br />

chart; the PDA camera shows poorer per<strong>for</strong>mance than the<br />

cellular camera. When observing the moving picture transmitted<br />

from the mobile camera, the PDA camera is subject<br />

to more noise than the cellular camera, which will produce a<br />

large characterization error.<br />

REAL-TIME COLOR MATCHING BETWEEN MOBILE<br />

CAMERA AND MOBILE LCD BASED ON 16-bit<br />

LUT DESIGN INCLUDING NOISE REDUCTION<br />

The process <strong>of</strong> colorimetric color matching reproduces the<br />

same CIE chromaticity and relative luminance compared<br />

with the original color, and has been widely applied to imaging<br />

devices such as monitors, printers, and scanners based<br />

on the ICC pr<strong>of</strong>ile, but not to mobile phones because mobile<br />

phones have been considered to be primarily communication<br />

devices. However, with multimedia convergence, mobile<br />

manufacturers have become aware <strong>of</strong> the necessity <strong>of</strong> color<br />

matching between mobile camera and mobile LCD. With the<br />

characterization <strong>of</strong> mobile camera and mobile LCD, to<br />

implement the color matching system, achievable ranges <strong>of</strong><br />

colors (gamut) must be considered. Figure 6 shows the<br />

gamut difference between a mobile camera under D65 environment<br />

(point) and a mobile LCD (solid color). As shown<br />

in Fig. 6, the gamut <strong>of</strong> mobile camera is larger than that <strong>of</strong><br />

mobile LCD, and has a regular <strong>for</strong>m resulting from the use<br />

<strong>of</strong> linear equations. Thus, significant parts <strong>of</strong> the mobile<br />

camera gamut can be unachievable by the gamut <strong>of</strong> mobile<br />

LCD, and it is necessary to alter the original colors (mobile<br />

camera) to ones that a given output medium (mobile LCD)<br />

is capable <strong>of</strong> reproducing. This power is frequently referred<br />

to as gamut mapping. In this paper, gamut mapping with<br />

variable and multiple anchor points is used to reduce any<br />

sudden color changes on the gamut region boundary and<br />

increase the lightness range reduced in conventional gamut<br />

mapping toward an anchor point. 11<br />

In general, the per<strong>for</strong>mance <strong>of</strong> colorimetric color<br />

matching between cross media, such as a monitor and<br />

printer, depends on the gamut mapping and device characterization.<br />

However, in the case <strong>of</strong> a mobile camera, various<br />

camera noises, such as CCD noise and thermal noise, can be<br />

354 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 6. Gamut mismatches between mobile camera point and mobile<br />

LCD solid color; a projected to a * ,b * plane and b projected<br />

to L * ,b * plane.<br />

included in moving pictures and become amplified after<br />

color matching, especially in the dark regions <strong>of</strong> the achromatic<br />

axis, although not in the chromatic region due to the<br />

blending <strong>of</strong> the reproduced signal. To solve this problem, the<br />

combination <strong>of</strong> two sigmoid functions with different parameters<br />

is applied to the lightness component <strong>of</strong> gamutmapped<br />

tristimulus values to change the contrast ratio. The<br />

sigmoid function is expressed as<br />

n=i<br />

1<br />

S i = e<br />

n=0 −100x n /m − x 0 2 /2 2 ,<br />

2<br />

i =1,2,. .,m,<br />

S i − minS<br />

S LUT =<br />

maxS − minS L *<br />

*<br />

*<br />

max out − L min out + L min out .<br />

12<br />

13<br />

Equation (12) is a discrete cumulative normal function S,<br />

where x 0 and are the mean and standard deviation <strong>of</strong> the<br />

normal distribution, respectively, and m is the number <strong>of</strong><br />

points used in the discrete lookup table. X n is the gamutmapped<br />

lightness component <strong>of</strong> CIELAB values and this<br />

value is then scaled into dynamic range <strong>of</strong> the mobile LCD,<br />

*<br />

*<br />

as given in Eq. (13), where L min out and L max out are blackpoint<br />

and white-point lightness value <strong>of</strong> the mobile LCD. In<br />

Eq. (12), x 0 controls the centering <strong>of</strong> the sigmoid function,<br />

Figure 7. Modified sigmoid function <strong>for</strong> noise reduction; a sigmoid<br />

functions with different parameters and b the combination <strong>of</strong> two sigmoid<br />

functions.<br />

and controls the shape. To find the parameters to conceal<br />

the camera noise through the lightness remapping, visual<br />

experiments were repeated based on adjustment <strong>of</strong> two parameters,<br />

and thus we found that the combination <strong>of</strong> two<br />

sigmoid functions is needed. In Figure 7(a), the solid line<br />

with x 0 =30 and =11.025 is the optimal curve to reduce<br />

the camera noise in the dark region, yet remapped lightness<br />

values in the bright region are significantly increased, <strong>for</strong>ming<br />

a saturation region. Thus another sigmoid function with<br />

x 0 =40 and =27.35, represented by dotted line in Fig. 7(a),<br />

is applied to the gamut-mapped light values larger than the<br />

input lightness value <strong>of</strong> 20 in order to make the lightness<br />

value <strong>of</strong> the reproduced image similar to that <strong>of</strong> the original<br />

image. Ultimately, the combination <strong>of</strong> two sigmoid functions,<br />

as shown in Fig. 7(b) expressed by the solid line, decreases<br />

the lightness value <strong>of</strong> camera noise in the dark region<br />

<strong>of</strong> achromatic axis, and from this result, amplified camera<br />

noise is hardly observed by a human eye.<br />

This kind <strong>of</strong> serial processing mentioned above, including<br />

the characterization, gamut mapping, and noise reduc-<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 355


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Table IV. Example <strong>of</strong> bit quantization.<br />

Table V. The example <strong>of</strong> proposed 3D-RGB LUT.<br />

6-bit quantization<br />

8-bit input data && 8-bit masking data<br />

R G B R G B<br />

R-channel: 00110<br />

G-channel: 011001<br />

B-channel: 10011<br />

00110011 && 11111000<br />

01100110 && 11111100<br />

10011001 && 11111000<br />

tion, has computational complexity and is not appropriate<br />

<strong>for</strong> real-time processing. There<strong>for</strong>e, a 3D-RGB LUT is constructed<br />

based on N-grid points <strong>for</strong> each channel. The input<br />

RGB digital values are uni<strong>for</strong>mly sampled by nnn grid<br />

points, which are processed by serial color matching, resulting<br />

in new corresponding output RGB values. The input<br />

RGB digital value and output RGB digital values are stored<br />

in the 3D-LUT and arbitrary input RGB values are calculated<br />

by interpolation. This 3D-LUT can be inserted into the<br />

mobile LCD without any difficulties associated with memory<br />

and computation. In actuality, in a mobile phone, a moving<br />

picture has 8-bits per channel, while the displayed RGB image<br />

on the LCD screen is represented by 5,6,5 bits per<br />

channel. Thus, be<strong>for</strong>e displaying an image on the LCD<br />

screen, 24-bit moving picture data is quantized into 16-bit<br />

data through a bit operation used in program language. For<br />

26 64 0 24 64 17<br />

32 64 0 25 64 18<br />

0 0 6 3 3 4<br />

6 0 6 10 3 2<br />

13 0 6 14 0 1<br />

19 0 6 17 0 1<br />

26 0 6 21 0 3<br />

32 0 6 23 0 4<br />

0 13 6 0 19 4<br />

example, suppose that the moving picture data to be displayed<br />

is (51,102,153). The final data are calculated by applying<br />

the AND operation (&&) with 8-bit masking data.<br />

Table IV shows an example <strong>of</strong> the AND operation.<br />

EXPERIMENTS AND RESULTS<br />

To conduct a subjective experiment <strong>of</strong> colorimetric color<br />

matching, test images were captured using a mobile camera;<br />

these included both face image and color chart images cap-<br />

Figure 8. The experimental results with the cellular phone; a and b device-dependent color matching, c<br />

and d proposed color matching.<br />

356 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 9. The experimental results with the PDA camera; a and b device-dependent color matching, c<br />

and d proposed color matching.<br />

tured in a lighting booth with D65 illumination. Statistically,<br />

the face image is one <strong>of</strong> the most frequently captured images,<br />

and people are very sensitive to their skin color displayed<br />

on mobile LCD. For this reason, a face image representing<br />

the skin color was selected as a test image. Similarly,<br />

the reason why the color chart image captured under D65<br />

illumination was used as a test image was that the characterization<br />

<strong>of</strong> the mobile camera was conducted under D65<br />

illumination, and subjective per<strong>for</strong>mance <strong>of</strong> color matching<br />

can be easily evaluated by comparing the displayed image<br />

with the real object as seen in the lighting booth. In addition,<br />

device-dependent color matching was compared to<br />

evaluate the per<strong>for</strong>mance <strong>of</strong> colorimetric color matching.<br />

Device-dependent color matching directly transmits the captured<br />

image to mobile LCD, while colorimetric color matching<br />

sends the captured image through the 3D-RGB LUT,<br />

which is quantized and transmitted to the mobile LCD.<br />

Table V shows a part <strong>of</strong> the data set stored in the 3D-RGB<br />

LUT designed to the 16-bit system. In the R channel and B<br />

channel, the maximum digital value is 2 5 , whereas<br />

G-channel’s maximum digital value is 2 6 . Based on the<br />

16-bit LUT, colorimetric color matching between mobile<br />

camera and mobile LCD can be processed in real time.<br />

SUBJECTIVE EXPERIMENT OF DEVELOPED COLOR<br />

MATCHING BASED ON 16-bit LUT DESIGN<br />

Figure 8 shows the captured images that are displayed on the<br />

cellular phone. Figures 8(a) and 8(b) show the images resulting<br />

from device-dependent color matching, while Figs.<br />

8(c) and 8(d) show the images resulting from LUT-based<br />

colorimetric color matching. In Fig. 8(a), even though the<br />

picture is taken against the light, the face region is very<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 357


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

Figure 11. Quality evaluations <strong>of</strong> device-dependent color matching and<br />

proposed color matching.<br />

seen in the D65 daylight, especially the red and green hues.<br />

Figure 9 shows the results <strong>of</strong> color matching <strong>for</strong> a PDA<br />

phone; the same effect is shown. Figure 10 shows the resulting<br />

images <strong>of</strong> colorimetric color matching considering the<br />

noise reduction. Figure 10(a) is the resulting image obtained<br />

by conventional colorimetric color matching, and its image<br />

quality is significantly degraded by the camera noise. By applying<br />

the combination <strong>of</strong> two sigmoid functions in Fig.<br />

7(b) to conventional colorimetric matching, the contrast ratio<br />

<strong>of</strong> reproduced image is changed and from this result,<br />

camera noise is not observable to the human eye, as shown<br />

in Fig. 10(b). Consequently, color matching based on 3D-<br />

LUT accurately reproduces the object color seen in the real<br />

scene and thus improves the color fidelity <strong>of</strong> the mobile<br />

display. For moving pictures, the same results decreasing the<br />

camera noise can be achieved with no problems <strong>of</strong> computation<br />

and memory.<br />

Figure 10. The results <strong>of</strong> noise reduction; a be<strong>for</strong>e lightness remapping<br />

and b after lightness remapping.<br />

bright due to the tendency <strong>of</strong> the electro-optical transfer<br />

function <strong>of</strong> mobile LCD to saturate the bright region as<br />

shown. In addition, the colorfulness <strong>of</strong> “table” and “cloth”<br />

region is more decreased than the original color, and the<br />

image quality is degraded. As shown in Fig. 8(c), the skin<br />

color in the “face” region is more natural and realistic than<br />

in Fig. 8(a), and the object colors such as “cloth” and “table”<br />

arewellreproducedonLCD.FortheMacbethColorchart<br />

seen in Fig. 8(b), colors <strong>of</strong> each patch are washed out and<br />

exhibit major differences in appearance, compared with<br />

original color seen in the D65 lighting booth, because<br />

device-dependent color matching is only physical data transmission.<br />

On the other hand, the result shown in Fig. 8(d)<br />

adquately represents the colorfulness <strong>of</strong> the original color<br />

QUANTITATIVE EVALUATION OF THE DEVELOPED<br />

COLOR MATCHING<br />

To evaluate colorimetric color matching based on 16-bit<br />

RGB LUT, a Macbeth Color Chart composed <strong>of</strong> 24 patches<br />

was used as a test image. For quantitative evaluation <strong>of</strong> the<br />

device dependent color matching, the Macbeth Color Chart<br />

is previously captured in the D65 lighting booth, and is displayed<br />

on mobile LCD. Then, the CIELAB value <strong>of</strong> each<br />

patch is measured using a colorimeter and compared with<br />

the CIELAB data <strong>of</strong> the Macbeth Color Chart measured in<br />

the D65 lighting booth, thus calculating the CIE 1976 color<br />

difference. For the proposed color matching, the Macbeth<br />

Color Chart is again captured in the D65 lighting booth, and<br />

is displayed on mobile LCD through use <strong>of</strong> the 16-bit RGB<br />

LUT. Then, the CIELAB value <strong>of</strong> each patch is measured<br />

using a colorimeter and compared with the CIELAB data <strong>of</strong><br />

Macbeth Color Chart measured in the D65 lighting booth.<br />

Figure 11 shows the result <strong>of</strong> quantitative evaluation using<br />

1976 Color difference. In Fig. 11, several patches corresponding<br />

to the proposed color matching have a larger color<br />

difference than <strong>for</strong> the conventional method, due to the<br />

characterization errors <strong>of</strong> the mobile camera and mobile<br />

LCD. However, the proposed color matching has a lower<br />

average color difference <strong>of</strong> 15.56, whereas device-dependent<br />

color matching has the average color difference <strong>of</strong> 24.395.<br />

358 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Son et al.: Real-time color matching between camera and LCD based on 16-bit lookup table design in mobile phone<br />

There<strong>for</strong>e, the proposed color matching achieves better colorimetric<br />

reproduction than the conventional method, and it<br />

is concluded that object color transmitted by mobile camera<br />

in real time can be accurately and realistically reproduced on<br />

a mobile LCD.<br />

CONCLUSIONS<br />

This paper presented a method <strong>for</strong> real-time color matching<br />

between mobile camera and mobile LCD, involving characterization<br />

<strong>of</strong> the mobile camera and mobile LCD, gamut<br />

mapping, noise reduction, and a LUT design. The characterization<br />

<strong>of</strong> the mobile LCD is conducted based on 16-bit<br />

processing, plus a sigmoid function is used to estimate the<br />

electro-optical transfer function. Meanwhile, <strong>for</strong> the characterization<br />

<strong>of</strong> the mobile camera, the optimal polynomial order<br />

is determined by trans<strong>for</strong>ming the captured RGB data<br />

into opponent color space and finding the relationship between<br />

the trans<strong>for</strong>med RGB values and the measured<br />

CIELAB values. Following the two types <strong>of</strong> characterization,<br />

gamut mapping is executed to overcome the gamut difference<br />

between the mobile camera and the mobile LCD, then<br />

noise reduction processing is applied to the lightness component<br />

<strong>of</strong> the gamut-mapped CIELAB values. Finally, to reduce<br />

the complex computation <strong>of</strong> serial-based color matching,<br />

a 3D RGB LUT is designed based on 16-bit data and<br />

inserted into the mobile phone. Experiments demonstrated<br />

that the proposed color matching realistically reproduced<br />

object colors from a real scene on a mobile LCD and improved<br />

the fidelity color <strong>of</strong> the mobile display. The LUT was<br />

also designed without any further computation or memory<br />

burden, making real-time processing possible.<br />

Acknowledgments<br />

This work is financially supported by the Ministry <strong>of</strong> Education<br />

and Human Resources Development (MOE), the<br />

Ministry <strong>of</strong> Commerce, Industry and Energy (MOCIE), and<br />

the Ministry <strong>of</strong> Labor (MOLAB) through the fostering<br />

project <strong>of</strong> the Lab <strong>of</strong> Excellency.<br />

REFERENCES<br />

1 J. Luo, “Displaying images on mobile device: capabilities, issues, and<br />

solutions”, Wirel. Commun. Mob. Comput. 2, 585–594 (2002).<br />

2 J. Y. Hardeberg, Acquisition and reproduction <strong>of</strong> color images:<br />

colorimetric and multispectral approaches, Universal Publishers,<br />

Dissertation.com, 2001.<br />

3 H. R. Kang, Color Technology <strong>for</strong> Electronic Image Device (SPIE Optical<br />

Engineering Press, Bellingham, WA, 1996).<br />

4 R. S. Berns, “Methods <strong>for</strong> characterizing CRT displays”, Displays 16,<br />

173–182 (1996).<br />

5 Y. S. Kwak and L. W. MacDonald, “Characterisation <strong>of</strong> a desktop LCD<br />

projector”, Displays 21, 179–194 (2000).<br />

6 N. Tamura, N. Tusmura, and Y. Miyake, “Masking model <strong>for</strong> accurate<br />

colorimetric characterization <strong>of</strong> LCD”, Proc. IS&T/SID Tenth Color<br />

<strong>Imaging</strong> Conference (IS&T, Jmigtiel, VA, 2002), 312–316.<br />

7 G. Sharma, “LCD versus CRTs color-calibration and gamut<br />

consideration”, Proc. IEEE 90, 605–622 (2002).<br />

8 M. D. Fairchild, Color Appearance Models (Addison-Wesley, Reading,<br />

MA, 1998).<br />

9 G. Hong, M. R. Luo, and P. A. Ronnier, “A study <strong>of</strong> digital camera<br />

colorimetric characterization based on polynomial modeling”, Color<br />

Res. Appl. 26, 76–84 (2001).<br />

10 M. R. Pointer, G. G. Attridge, and R. E. Jacobson, “Practical camera<br />

characterization <strong>for</strong> colour measurement”, <strong>Imaging</strong> Sci. J., 49, 63–80<br />

(2001).<br />

11 C. S. Lee, Y. W. Park, S. J. Cho, and Y. H. Ha, “Gamut mapping<br />

algorithm using lightness mapping and multiple anchor points <strong>for</strong> linear<br />

tone and maximum chroma reproduction”, J. <strong>Imaging</strong> Sci. Technol. 45,<br />

209–223 (2001).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 359


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 360–367, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Solving Under-Determined Models in Linear Spectral<br />

Unmixing <strong>of</strong> Satellite Images: Mix-Unmix Concept<br />

(Advance Report)<br />

Thomas G. Ngigi and Ryutaro Tateishi<br />

Center <strong>for</strong> Environmental Remote Sensing, Chiba University, 1-33 Yayoi, Inage, Chiba, 263-8522, Japan<br />

E-mail: tgngigi@hotmail.com<br />

Abstract. This paper reports on a simple novel concept <strong>of</strong> addressing<br />

the problem <strong>of</strong> underdetermination in linear spectral unmixing.<br />

Most conventional unmixing techniques fix the number <strong>of</strong> endmembers<br />

on the dimensionality <strong>of</strong> the data, and none <strong>of</strong> them can<br />

derive multiple 2 + end-members from a single band. The concept<br />

overcomes the two limitations. Further, the concept creates a processing<br />

environment that allows any pixel to be unmixed without any<br />

sort <strong>of</strong> restrictions (e.g., minimum determinable fraction), impracticalities<br />

(e.g., negative fractions), or trade-<strong>of</strong>fs (e.g., either positivity<br />

or unity sum) that may be associated with conventional unmixing<br />

techniques. The proposed mix-unmix concept is used to generate<br />

fraction images <strong>of</strong> four spectral classes from Landsat 7 ETM+data<br />

(aggregately resampled to 240 m) first principal component only.<br />

The correlation coefficients <strong>of</strong> the mix-unmix image fractions versus<br />

reference image fractions <strong>of</strong> the four end-members are 0.88, 0.80,<br />

0.67, and 0.78. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and<br />

Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4360<br />

PROBLEM STATEMENT / INTRODUCTION, AND<br />

OBJECTIVE<br />

“… the number <strong>of</strong> bands must be more than the number <strong>of</strong><br />

end-members…” is perhaps the most ubiquitous statement<br />

in the field <strong>of</strong> linear spectral unmixing. This is simply because<br />

most <strong>of</strong> the conventional unmixing techniques are<br />

based on least squares, 1 convex geometry, 2 or combination<br />

<strong>of</strong> both and the number <strong>of</strong> end-members (unknowns) is<br />

dependent on the dimensionality (equations) <strong>of</strong> the data.<br />

Least squares can unmix as many end-members as up to the<br />

dimensionality <strong>of</strong> the data, and at the very best exceed by<br />

one when the unity constraint is en<strong>for</strong>ced. In convex geometry,<br />

the number <strong>of</strong> determinable end-members (at the<br />

unmixing stage) is equal to the number <strong>of</strong> vertices <strong>of</strong> the<br />

data simplex, and this number exceeds the dimensionality <strong>of</strong><br />

the data by one. After extracting the end-member spectra,<br />

most <strong>of</strong> the convex geometry-based techniques apply the<br />

least squares approach (combined case) in computing the<br />

fractions <strong>of</strong> the end-members.<br />

Some linear spectral unmixing techniques include Sequential<br />

Maximum Angle Convex Cone (SMACC) Spectral<br />

Tool, 3 (Generalized) Orthogonal Subspace Projection, 4,5<br />

Convex Cone Analysis, 6 N-FINDR, 7 Orasis, 7 and Iterative<br />

Error Analysis. 7 Keshava 8 gives a detailed account <strong>of</strong> spectral<br />

unmixing techniques. A number <strong>of</strong> commercially available<br />

s<strong>of</strong>tware, including ENVI, IDRISI Kilimanjaro, PCI, and<br />

ERDAS Imagine, have linear spectral unmixing modules.<br />

The greatest fundamental commonality <strong>of</strong> all conventional<br />

linear spectral unmixing techniques is that none <strong>of</strong> them can<br />

derive multiple end-members 2 + from a single band. The<br />

object <strong>of</strong> the mix-unmix concept is to overcome this problem<br />

and unmix as many end-members as can be deciphered<br />

from the reference data and without introducing any sort <strong>of</strong><br />

restrictions, impracticalities, or trade-<strong>of</strong>fs that may be associated<br />

with conventional unmixing techniques.<br />

DESCRIPTION OF THE MIX-UNMIX CONCEPT<br />

As the term implies, the model consists <strong>of</strong> two branches,<br />

namely, mixing and unmixing. The mixing branch entails<br />

development <strong>of</strong> hypothetical mixed pixels on the basis <strong>of</strong><br />

desired end-members’ actual digital numbers (DNs).<br />

Unmixing involves determination <strong>of</strong> each real image pixel’s<br />

DN’s contributory end-members and their fractions by<br />

back-propagating through the mixing branch using a pixel<br />

<strong>of</strong> the same DN in the hypothetical image as a proxy. This<br />

preliminary study demonstrates the concept on a single<br />

simulated band.<br />

Mixing Branch<br />

Nominally, the end-members are paired up hierarchically<br />

into a single hypothetical mixed class (Figure 1;<br />

EM1=end-member 1, EM1.2=combined end-members 1<br />

and 2). Essentially, in pairing up, each and every DN from a<br />

member <strong>of</strong> a pair is combined with each and every DN from<br />

the other member, at complementary percentages ranging<br />

from 0% to 100%, giving rise to various “mixture tables”<br />

(MTs) whose number depends on the ranges <strong>of</strong> training<br />

DNs <strong>of</strong> the two members.<br />

Theory <strong>of</strong> the mixing branch and <strong>for</strong>mation <strong>of</strong> mixture<br />

tables (MTs)<br />

The number <strong>of</strong> possible DN combinations, MTs, <strong>of</strong> two<br />

members, A and B, <strong>of</strong> a pair is equal to the product <strong>of</strong> their<br />

training DN ranges, i.e.,<br />

MTs = A max DN − A min DN + 1 B max DN<br />

Received Dec. 5, 2006; accepted <strong>for</strong> publication Mar. 22, 2007.<br />

1062-3701/2007/514/360/8/$20.00.<br />

where<br />

− B min DN + 1,<br />

360


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Figure 1. Bottom-up pairing up <strong>of</strong> end-members as well as the resultant super-end-members—pairing up<br />

involves mixing the two members-<strong>of</strong>-a-pair’s DN ranges at all complementary fractions. In case the number <strong>of</strong><br />

super-end-members is even but not a multiple <strong>of</strong> four, one level <strong>of</strong> mixing is skipped <strong>for</strong> one pair as indicated<br />

by EM5 and EM6 <strong>for</strong> six end-members. For an odd number, one end-member is simply carried <strong>for</strong>ward to the<br />

next level individually as indicated by EM7 <strong>for</strong> seven end-members. In this case, EM5.6 and EM7 have the<br />

same hierarchical status as EM1.2.3.4. At the base <strong>of</strong> the branch are training DN ranges and assumed<br />

fractions <strong>of</strong> the end-members—the DNs are known by in situ observation, from spectral libraries, or identification<br />

<strong>of</strong> pure pixels in the image to be unmixed, etc. At the top <strong>of</strong> the branch are hypothetical pixels’ DN values<br />

resulting from mixing all the end-members’ spectra at all possible complementary fractions.<br />

A max DN = maximum DN <strong>of</strong> A,<br />

A min DN = minimum DN <strong>of</strong> A,<br />

Subsequently, the total number <strong>of</strong> possible DNs and percentages<br />

combinations <strong>of</strong> the two is<br />

MTs N%s.<br />

The expression also gives the total number <strong>of</strong> possible mixture<br />

pixels <strong>of</strong> the two.<br />

Thus all pixels, in a hypothetical band, composed <strong>of</strong><br />

only two end-members, EM1 and EM2, would be defined by<br />

the following expression—discussed assuming: that the endmembers’<br />

training DNs range from, respectively, 10–89 and<br />

90–150 in the band; a mixture interval <strong>of</strong> 10%, and assuming<br />

linear mixing.<br />

B max DN = maximum DN <strong>of</strong> B,<br />

B min DN = minimum DN <strong>of</strong> B.<br />

The number <strong>of</strong> possible percentages combinations, N%s, <strong>of</strong><br />

the two members is given by<br />

1<br />

N%s=100 % ÷ MI +1,<br />

where<br />

MI = adopted mixture interval.<br />

where<br />

• f 1,i =percentage <strong>of</strong> EM1 in pixel i Table Ia 1st<br />

column,<br />

• f 2,i =percentage <strong>of</strong> EM2 in pixel i Table Ia 2nd<br />

column,<br />

• f 1 +f 2 =100%,<br />

• DN 1,i =DN <strong>of</strong> EM1 in pixel i Table Ia 2nd row,<br />

• DN 2,i =DN <strong>of</strong> EM2 in pixel i Table Ia 3rd row,<br />

• DN 1,2,i =mixture DN <strong>of</strong> DN 1,i and DN 2,i in pixel i<br />

Table Ia all cells excluding the first two columns<br />

and first three rows,<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 361


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Table I. a MTs <strong>of</strong> EM1 and EM2. The EM1.2 DNs are computed as: EM1.2 DN=EM1 % EM1 DN +EM2 % EM2 DN. Note that EM1% and<br />

EM2% are complementary.<br />

MT= 1 2 3−<br />

4 880 4 881<br />

4 879<br />

EM1 DN= 10 11 88 89<br />

EM2 DN= 90 90 150 150<br />

EM1% EM2% EM1.2 DN<br />

0 100 90 90 150 150<br />

10 90 82 82 143 143<br />

20 80 74 74 137 137<br />

30 70 66 66 131 131<br />

40 60 58 58 125 125<br />

50 50 50 51 119 119<br />

60 40 42 43 113 113<br />

70 30 34 35 107 107<br />

80 20 26 27 100 101<br />

90 10 18 19 94 95<br />

100 0 10 11 88 89<br />

b Min-max MTs <strong>of</strong> EM1 and EM2.<br />

c Min-max LUTs <strong>of</strong> EM1 and EM2.<br />

Min-MT<br />

EM1 DN= 10<br />

EM2 DN= 90<br />

Min-LUT<br />

Max-MT<br />

EM1 DN= 89<br />

EM2 DN= 150<br />

Max-LUT<br />

EM1 % EM2 %<br />

EM1.2 DN<br />

Amount <strong>of</strong><br />

overlap with<br />

EM1.2 vector<br />

90–150<br />

0 100 90 150 60<br />

10 90 82 143 53<br />

20 80 74 137 47<br />

30 70 66 131 41<br />

40 60 58 125 35<br />

50 50 50 119 29<br />

60 40 42 113 23<br />

70 30 34 107 17<br />

80 20 26 101 11<br />

90 10 18 95 5<br />

100 0 10 89 0<br />

• min EM1DN=minimum DN <strong>of</strong> EM1,<br />

• max EM1DN=maximum DN <strong>of</strong> EM1.<br />

Bounding mixture tables, and various numbers <strong>of</strong> endmembers<br />

From Table I(a), it is very clear that, given constant fractions<br />

<strong>of</strong> EM1 and EM2, the mixture class DNs (EM1.2 DNs) always<br />

fall between the values in the first and last MTs, thus<br />

the two MTs fully give the ranges <strong>of</strong> all possible DNs <strong>of</strong><br />

EM1.2. Hereinafter, the two are referred to as min-MT and<br />

max-MT, respectively, and min-max MTs collectively [Table<br />

I(b)].<br />

Similarly, min-max MTs <strong>of</strong> the other paired endmembers<br />

are generated: <strong>for</strong> EM3 and EM4; in Eq. (1) EM1,<br />

EM2, and EM1.2 are replaced with EM3, EM4, and EM3.4,<br />

respectively. Table II(a) shows min-max MTs <strong>of</strong> EM3 and<br />

EM4 shows the DN ranges are 151–180 and 181–210, respectively.<br />

Next, second level min-max MTs are developed from<br />

the above first level MTs: <strong>for</strong> EM1.2 and EM3.4; in Eq. (1)<br />

EM1, EM2, and EM1.2 are replaced with EM1.2, EM3.4, and<br />

EM1.2.3.4, respectively. Table III(a) shows min-max MTs <strong>of</strong><br />

EM1.2, and EM3.4. Since EM1.2 represents EM1 and EM2,<br />

and EM3.4 represents EM3 and EM4, subsequently, the second<br />

level min-max MTs inherently represent all the possible<br />

DN outcomes <strong>of</strong> mixing all the end-members EM1, EM2,<br />

EM3, and EM4 at all possible complementary fractions.<br />

362 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Table II. a Min-max MTs <strong>of</strong> EM3 and EM4. The EM3.4 DNs are computed as: EM3.4<br />

DN=EM3 % EM3 DN +EM4 % EM4 DN. Note that EM3% and EM4% are<br />

complementary. b Min-max LUTs <strong>of</strong> EM3 and EM4.<br />

Min-MT<br />

EM3 DN= 151<br />

EM4 DN= 181<br />

Min-LUT<br />

Max-MT<br />

EM3 DN= 180<br />

EM4 DN= 210<br />

Max-LUT<br />

Amount <strong>of</strong><br />

overlap with<br />

EM3.4 vector<br />

172–201<br />

EM3 % EM4 % EM3.4 DN<br />

0 100 181 210 20<br />

10 90 178 207 29<br />

20 80 175 204 29<br />

30 70 172 201 29<br />

40 60 169 198 26<br />

50 50 166 195 23<br />

60 40 163 192 20<br />

70 30 160 189 17<br />

80 20 157 186 14<br />

90 10 154 183 11<br />

100 0 151 180 8<br />

For more end-members, the process is successively repeated<br />

as shown in Fig. 1. For three end-members in Eq. (1)<br />

EM1, EM2, and EM1.2 are replaced with EM1.2, EM3, and<br />

EM1.2.3, respectively.<br />

Unmixing Branch<br />

This is similar to the mixing branch (Fig. 1) but with the<br />

arrows (processing) reversed and the MTs renamed look-uptables<br />

(LUTs)—Tables I(c), II(b), and IIIb. As discussed below,<br />

a real image pixel DN is fractionalized into two highest<br />

level super-end-members, each <strong>of</strong> which is then split into its<br />

two constituent end-members. The process continues until<br />

the finest level (end-members <strong>of</strong> interest) from which the<br />

mixing branch was constructed (see Figure 2).<br />

Fractionalization<br />

This discussion demonstrates the unmixing process on a<br />

single-band image composed <strong>of</strong> the four end-members outlined<br />

in the Mixing Branch section. For each DN in the<br />

band, all the vectors in which it lies are identified, e.g., Table<br />

III(b) EM.1.2.3.4 italicized DNs give all the possible vectors<br />

<strong>for</strong> DN 180, with the lower nodes located in Table III(b-1)<br />

and the upper nodes in Table III(b-2)—the first vector is<br />

172–204 (bold). Each one <strong>of</strong> these vectors is a combination<br />

<strong>of</strong> two minor vectors, one apiece from EM1.2 and EM3.4<br />

(italicized); e.g., <strong>for</strong> the vector 172–204, the constituent vectors<br />

are 90–150 (bold) from EM1.2 and 181–210 (bold)<br />

from EM3.4.<br />

The most probable vector (MPV) in which the DN 180<br />

lies is computed as<br />

Figure 2. Top-bottom fractionalization <strong>of</strong> a pixel; first into two highest level super-end-members, then effectively<br />

into second highest level four super-end-members by fractionalizing each <strong>of</strong> the highest level super-endmembers<br />

into two. The process is repeated successively until the lowest level end-members that were used to<br />

build up the mixing branch. At the top <strong>of</strong> the branch is a universe <strong>of</strong> values encompassing all the DNs in the<br />

image to be unmixed—all: assuming that the image is composed <strong>of</strong> only the end-members used in the mixing<br />

branch. At the base <strong>of</strong> the branch are estimated contributory percentages fractions <strong>of</strong> the end-members cf.<br />

Fig. 1.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 363


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Table III. a-1 Min-MT <strong>of</strong> EM1.2 and EM3.4. The EM1.2.3.4 DNs are computed as: EM1.2.3.4 DN=EM1.2% EM1.2 DN +EM3.4% EM3.4 DN. Note that EM1.2% and EM3.4% are complementary.<br />

b-1 Min-LUT <strong>of</strong> EM1.2 and EM3.4.<br />

EM3.4 DN b<br />

EM1.2<br />

DN a<br />

EM1.2<br />

%<br />

EM3.4<br />

%<br />

181 178 175 172 169 166 163 160 157 154 151<br />

EM1.2.3.4 DN<br />

90 0 100 181 178 175 172 169 166 163 160 157 154 151<br />

90 10 90 172 169 167 164 161 158 156 153 150 148 145<br />

90 20 80 163 160 158 156 153 151 148 146 144 141 139<br />

90 30 70 154 152 150 147 145 143 141 139 137 135 133<br />

90 40 60 145 143 141 139 137 136 134 132 130 128 127<br />

90 50 50 136 134 133 131 130 128 127 125 124 122 121<br />

90 60 40 126 125 124 123 122 120 119 118 117 116 114<br />

Rows 8 to 117<br />

10 70 30 61 60 60 59 58 57 56 55 54 53 52<br />

10 80 20 44 44 43 42 42 41 41 40 39 39 38<br />

10 90 10 27 27 27 26 26 26 25 25 25 24 24<br />

10 100 0 10 10 10 10 10 10 10 10 10 10 10<br />

a-2 Max-MT <strong>of</strong> EM1.2 and EM3.4.<br />

b-2 Max-LUT <strong>of</strong> EM1.2 and EM3.4.<br />

EM3.4 DN d<br />

EM1.2<br />

DN c<br />

EM1.2<br />

%<br />

EM3.4<br />

%<br />

210 207 204 201 198 195 192 189 186 183 180<br />

EM1.2.3.4 DN<br />

150 0 100 210 207 204 201 198 195 192 189 186 183 180<br />

150 10 90 204 201 199 196 193 191 188 185 182 180 177<br />

150 20 80 198 196 193 191 188 186 184 181 179 176 174<br />

150 30 70 192 190 188 186 184 182 179 177 175 173 171<br />

150 40 60 186 184 182 181 179 177 175 173 172 170 168<br />

150 50 50 180 179 177 176 174 173 171 170 168 167 165<br />

150 60 40 174 173 172 170 169 168 167 166 164 163 162<br />

Rows 8 to 117<br />

89 70 30 125 124 124 123 122 121 120 119 118 117 116<br />

89 80 20 113 113 112 111 111 110 110 109 108 108 107<br />

89 90 10 101 101 100 100 100 99 99 99 99 98 98<br />

89 100 0 89 89 89 89 89 89 89 89 89 89 89<br />

a Column 1 elements are from EM1 and EM2 min-MT Table Ib, column 3<br />

b Row 1 elements are from EM3 and EM4 min-MT Table IIa, column 3<br />

c Column 1 elements are from EM1 and EM2 max-MT Table Ib, column 4<br />

d Row 1 elements are from EM3 and EM4 max-MT Table IIa, column 4<br />

where<br />

n<br />

lower nodes<br />

i=1<br />

cMNxDN =<br />

n<br />

, 2<br />

cMNyDN =<br />

n<br />

upper nodes<br />

i=1<br />

n<br />

, 3<br />

• cMNxDN=lower node <strong>of</strong> DN 180 MPV from combined<br />

classes M and N (EM1.2 or EM3.4),<br />

• cMNyDN=upper node ditto,<br />

• lower nodes=all the EM1.2.3.4 italicized DN in Table<br />

III(b-1),<br />

• upper nodes=ditto Table III(b-2),<br />

• n=number <strong>of</strong> EM1.2.3.4 italicized DN vectors=count<br />

<strong>of</strong> EM1.2.3.4 italicized DN nodes in Table III(b-1) or<br />

Table III(b-2).<br />

From Eqs. (2) and (3), cMNxDN=156 and cMNyDN<br />

=190. From Table III(b), the pair <strong>of</strong> nodes most close to the<br />

pair 156/190 is 156/191 and it is adopted as the MPV <strong>for</strong><br />

the DN 180. This vector 156–191 [Table III(b) bold and<br />

underlined] lies at the intersection <strong>of</strong> EM1.2 vector 90–150<br />

364 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

given rise to the EM1.2 vector 90–150 is taken to be proportional<br />

to the amount <strong>of</strong> overlap with it. Each probability is<br />

also taken to be the probability <strong>of</strong> the corresponding paired<br />

percentages (PPs) having given rise to the EM1.2 vector 90–<br />

150 since the PV was developed from them (PPs). Similarly<br />

<strong>for</strong> Table II(b) PVs and PPs in the case <strong>of</strong> EM3.4 vector<br />

172–201. There are seven possible overlap scenarios as depicted<br />

by Figure 3. Table I(c) last column gives the weights<br />

<strong>of</strong> the EM1 and EM2 vectors to EM1.2 vector 90–150, and<br />

Table II(b) ditto EM3 and EM4 vectors to EM3.4 vector<br />

172–201.<br />

From Fig. 3 and Tables I(c) and II(b), the most probable<br />

percentage contribution (MPPC) <strong>of</strong> each daughter class to<br />

its mother class is computed as<br />

Figure 3. Possible universal overlap scenarios. A is EMx.y’s most probable<br />

vector, e.g., EM1.2 MPV 90-150 or EM3.4 MPV 172-201; all/<br />

some <strong>of</strong> the other non-arrowed lines B-H are vectors contained in<br />

EMx.y’s min-max LUTs e.g., Table Ic <strong>for</strong> EM1.2, or Table IIb <strong>for</strong><br />

EM3.4; arrowed lines are the respective overlaps.<br />

and EM3.4 vector 172–201. There<strong>for</strong>e, by extension, the DN<br />

180 most probably resulted from these EM1.2 and EM3.4<br />

vectors as the combination most probably gave rise to the<br />

DN 180 MPV 156–191.<br />

Further, percentages-wise, the DN 180 could have resulted<br />

from any <strong>of</strong> the paired percentages associated with the<br />

EM1.2.3.4 italicized DNs vectors. The most probable contributory<br />

paired percentages (MPPC) are computed as<br />

where<br />

MPPC x.y =<br />

n<br />

i% x.y p i <br />

i=1<br />

±<br />

n<br />

n p i<br />

i=1<br />

n<br />

i=1<br />

p i v i<br />

2<br />

n<br />

n 2 <br />

i=1<br />

p i<br />

,<br />

• i% x.y =ith paired percentages <strong>of</strong> x EM1.2<br />

and y EM3.4,<br />

• p i =weight <strong>of</strong> i% x.y =count <strong>of</strong> i% x.y ’s EM1.2.3.4<br />

italicized DNs,<br />

• n=count <strong>of</strong> probable contributory paired<br />

percentages,<br />

• v=i% x.y −MPPC x,y . The second term in Eq. 4<br />

is computed after the first one.<br />

From Eq. (4), EM1.2% =16.67% ±2.34% and EM3.4%<br />

=83.33% ±2.34%. Hence, the DN 180 most probably resulted<br />

from these EM1.2 and EM3.4 percentages combinations<br />

as the pair most probably gave rise to the DN 180<br />

MPV 156-191.<br />

Next, the EM1.2 90-150 and EM3.4 172-201 vectors are<br />

checked against the lower level min-max LUTs, Tables I(c)<br />

and II(b), respectively, and all the vectors with which they<br />

(EM1.2 and EM3.4 vectors) overlap <strong>for</strong>m the universe <strong>of</strong><br />

possible vectors (PVs) from which they (EM1.2 and EM3.4<br />

vectors) or, in other words, a fraction <strong>of</strong> the value 180, arose.<br />

The probability (weight) <strong>of</strong> each <strong>of</strong> the Table I(c) PVs having<br />

4<br />

where<br />

cM %=<br />

q<br />

cM% i p i<br />

i=1<br />

±<br />

q<br />

q p i<br />

i=1<br />

q<br />

i=1<br />

p i v i<br />

2<br />

q<br />

q 2 <br />

i=1<br />

p i<br />

,<br />

5<br />

• cM% =MPPC <strong>of</strong> daughter class cM (EM1 or EM2) to<br />

its mother class (EM1.2). EM3 or EM4 <strong>for</strong> EM3.4;<br />

• cM% i =percent <strong>of</strong> cM’s ith probable vector—Table I(c)<br />

columns 1 and 2 <strong>for</strong> EM1 and EM2, respectively; Table<br />

II(b) columns 1 and 2 <strong>for</strong> EM3 and EM4, respectively;<br />

• p i =overlap range <strong>of</strong> cM’s ith probable vector with its<br />

cM mother’s MPV. Table I(c) last column <strong>for</strong> EM1<br />

and EM2. Table II(b) last column <strong>for</strong> EM3 and EM4;<br />

• q=count <strong>of</strong> probable paired-percentages;<br />

• v=cM%-cM% i . The second term in Eq. (5) is computed<br />

after the first one.<br />

From Eq. (5) and Tables I(c) and II(b), the MPPCs <strong>of</strong><br />

EM1 and EM2 to EM1.2, and EM3 and EM4 to EM3.4 are;<br />

EM1=71% ±2.84%, EM2=29% ±2.84%, EM3<br />

=59% ±2.42%, and EM4=41% ±2.42%.<br />

The MPPC <strong>of</strong> an end-member to the original pixel DN<br />

is simply the product <strong>of</strong> all MPPCs along the path from the<br />

end-member itself to the pixel DN. Hence, <strong>for</strong> end-member:<br />

• 1=EM1% EM1.2% =71% 16.67%<br />

=11.84±1.73%,<br />

• 2=EM2% EM1.2% =29% 16.67%<br />

=04.83±0.83%,<br />

• 3=EM3% EM3.4% =59% 83.33%<br />

=49.16±2.44%,<br />

• 4=EM4% EM3.4% =41% 83.33%<br />

=34.17±2.23%.<br />

The standard deviation <strong>of</strong> product AB is computed as<br />

+ 2<br />

B<br />

, 6<br />

B<br />

AB = AB 2<br />

A<br />

A<br />

where k =standard deviation <strong>of</strong> k.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 365


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Figure 4. Left location <strong>of</strong> the study area; right 30 m resolution Landsat<br />

ETM+ data RGB=342.<br />

Figure 5. Left reference data: 30 m resolution spectral classes. The<br />

spectral classes correspond to broad in<strong>for</strong>mation classes dense vegetation<br />

C1, dense vegetation/bare land C2, bare land/dense vegetation<br />

C3, and bare land C4; right raw image: 240 m resolution first<br />

principal component. The rectangles show locations <strong>of</strong> reference black<br />

and mix-unmix red training sites.<br />

STUDY TEST<br />

Four spectral end-members are mapped from a single-band<br />

simulated raw image.<br />

Study Area and Data<br />

Landsat ETM+ data, <strong>of</strong> 21 February 2000, covering<br />

southern-central Kenya is utilized to generate both the reference<br />

and raw data. The land covers in the area transition<br />

from dense vegetation (<strong>for</strong>est) to bare land (Figure 4).<br />

Simulation <strong>of</strong> Reference and Raw Data<br />

K-mean classification is run on the ETM+ data bands 1, 2,<br />

3, 4, 5, and 7 to produce four spectral classes. The spectral<br />

classes are adopted as reference data. The six bands data is<br />

resampled to 240 m resolution (mimics moderate resolution<br />

data, e.g., MODIS bands 1 and 2–250 m resolution) and<br />

then principal components trans<strong>for</strong>mation (PCT) executed<br />

on the new data set. Each <strong>of</strong> the resampled bands and PCs is<br />

taken as a candidate raw image.<br />

Selection <strong>of</strong> Band to Unmix, and its Unmixing<br />

A section from the 30 m resolution reference spectral classes<br />

image, black rectangle in Figure 5, hereinafter referred to as<br />

Figure 6. a Comparison <strong>of</strong> DNs’ distribution curves <strong>of</strong> C1, C2, C3,<br />

and C4 in mix-unmix training site across original bands and principal<br />

components PCs. The least overlap between the curves occurs in PC1<br />

and, thus, it is adopted as the raw band to unmix. Y axes=frequencies,<br />

and X axes=DNs—but values not shown. b Training DNs <strong>of</strong> EM1,<br />

EM2, EM3, and EM4.<br />

reference training site, is geographically overlaid on each<br />

candidate raw image (240 m resolution) and pure pixels in<br />

the overlay section, red rectangle in Fig. 5, hereinafter referred<br />

to as mix-unmix training site, <strong>of</strong> the candidate raw<br />

image <strong>for</strong> each spectral class identified. A pixel in the mixunmix<br />

training site is pure if the geographically corresponding<br />

8-pixel8-pixel block in the reference training site is<br />

composed <strong>of</strong> a single class, 8 is the ratio <strong>of</strong> the two resolutions.<br />

Figure 6(a) compares the four spectral classes’ purepixels’<br />

DNs’ distribution curves in the mix-unmix training<br />

site. Since the four spectral classes exhibit the highest spectral<br />

dissimilarity between themselves in the first PC, it is<br />

adopted as the raw image to be unmixed. The training DN<br />

ranges <strong>of</strong> the four spectral classes (now denoted as endmembers,<br />

EMs) are as shown in Fig. 6(b). The raw image<br />

(first PC) is unmixed under the mix-unmix concept on the<br />

basis <strong>of</strong> the training DNs into the four end-members.<br />

Mix-Unmix Fraction Images versus Reference Fraction<br />

Images<br />

Reference fraction images <strong>of</strong> the four spectral classes (Fig. 5)<br />

are generated by computing the percentage coverage <strong>of</strong> each<br />

class in every 8-pixel8-pixel block (each block is 240 m<br />

366 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Ngigi and Tateishi: Solving under-determined models in linear spectral unmixing <strong>of</strong> satellite images: mix-unmix concept<br />

Figure 7. Data processing flowchart.<br />

240 m). The reference fraction images are compared with<br />

the mix-unmix fraction images. Figure 7 outlines the entire<br />

image processing flow, and Figure 8 compares the fraction<br />

images. The correlation coefficients <strong>of</strong> the mix-unmix image<br />

fractions versus the reference image fractions <strong>of</strong> the four<br />

end-members are 0.88 (EM1), 0.80 (EM2), 0.67 (EM3), and<br />

0.78 (EM4).<br />

DISCUSSION<br />

The mix-unmix fraction images show similar transition patterns<br />

(highest to lowest concentration levels) as the reference<br />

fraction images <strong>for</strong> all the end-members; though the correlation<br />

coefficients are not “very high.” Although a mixture<br />

interval <strong>of</strong> 10% is used throughout this study, any value that<br />

as a divisor <strong>of</strong> 100 gives a whole number can be used. We<br />

cannot use 100 itself as it would mean that each pixel contains<br />

just a single end-member.<br />

FUTURE<br />

As discussed in the Mixing Branch section, only the extreme<br />

DN values (i.e., bounding mixture tables) are used in this<br />

study. Also, the training DNs <strong>of</strong> end-members are assumed<br />

to be “frequency-less.” As the mix-unmix s<strong>of</strong>tware develops,<br />

all mixture tables and training DNs’ distribution curves will<br />

be incorporated.<br />

The effect <strong>of</strong> adopted mixture interval and overlap <strong>of</strong><br />

training DNs on accuracy <strong>of</strong> the concept will be addressed<br />

on implementation <strong>of</strong> the above. Also, per<strong>for</strong>mance <strong>of</strong> the<br />

concept across different numbers <strong>of</strong> end-members, different<br />

resolutions, and different geographical scales will be tested.<br />

Figure 8. Reference fraction images upper row and Mix-unmix fraction<br />

images lower row <strong>of</strong> four end-members first column=EM1, second<br />

=EM2, third=EM3, fourth=EM4. White and black are background,<br />

i.e., 0%.<br />

CONCLUSIONS<br />

This preliminary investigation shows that the mix-unmix<br />

concept is capable <strong>of</strong> addressing the problem <strong>of</strong> underdetermination<br />

in linear spectral unmixing—a very revolutionary<br />

dimension in data processing as the number <strong>of</strong> endmembers<br />

is not pegged on that <strong>of</strong> available bands. It is the<br />

only method that truly solves the problem <strong>of</strong> underdetermination.<br />

Sequential Maximum Angle Convex Cone<br />

(SMACC) Spectral Tool does not work on a single band, and<br />

Generalized Orthogonal Subspace Projection cannot generate<br />

additional bands from a single band. Further, the mixunmix<br />

concept creates a processing environment that allows<br />

any pixel to be unmixed without any sort <strong>of</strong> restrictions<br />

(e.g., minimum determinable fraction), impracticalities (e.g.,<br />

negative fractions), or trade-<strong>of</strong>fs (e.g., either positivity or<br />

unity sum) that may be associated with conventional unmixing<br />

techniques.<br />

REFERENCES<br />

1 Y. E. Shimabukuro and J. A. Smith, “The least squares mixing methods<br />

to generate fraction images derived from remote sensing multispectral<br />

data”, IEEE Trans. Geosci. Remote Sens. 29, 16 (1991).<br />

2 J. W. Boardman, “Geometric mixture analysis <strong>of</strong> imaging spectrometry<br />

data”, Proc. Int. Geosci Remote Sens Symposium 4, 2369 (1994).<br />

3 J. Gruninger, A. J. Ratkowski, and M. L. Hoke, “The sequential<br />

maximum angle convex cone (SMACC) endmember model”, Proc. SPIE<br />

5425 1 (2004).<br />

4 R. Hsuan and C. Yang-Lang, “Error Analysis <strong>for</strong> Band Generation in<br />

Generalized Process Orthogonal Subspace Projection”, IEEE Geoscience<br />

and Remote Sensing Symposium Proceedings, IGARSS (IEEE Press,<br />

Piscataway, NJ, 2005).<br />

5 I. Emmett, “Hyperspectral Image Classification Using Orthogonal<br />

Subspace Projections: Image Simulation and Noise Analysis”, http://<br />

www.cis.rit.edu/~ejipci/Reports/osp_paper.pdf (2001).<br />

6 A. Ifarraguerri and C. Chang, “Multispectral and Hyperspectral Image<br />

Analysis with Convex Cones”, IEEE Trans. Geosci. Remote Sens. 37, 756<br />

(1999).<br />

7 M. E. Winter and E. M. Winter, “Comparison <strong>of</strong> approaches <strong>for</strong><br />

determining end-members in hyperspectral data”, Proc. IEEE Aerospace<br />

Conference (IEEE Press, Piscataway, NJ, 2000).<br />

8 N. Keshava, “A Survey <strong>of</strong> Spectral Unmixing Techniques”, Lincoln Lab.<br />

J. 14, 55 (2003).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 367


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 368–379, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Color Shift Model-Based Segmentation and Fusion<br />

<strong>for</strong> Digital Aut<strong>of</strong>ocusing<br />

Vivek Maik<br />

Image Processing and Intelligent Systems Lab, Graduate School <strong>of</strong> Advanced <strong>Imaging</strong> <strong>Science</strong>, Multimedia<br />

and Film, Chung Ang University, Seoul 156-756, South Korea<br />

E-mail: vivek5681@wm.cau.ac.kr<br />

Dohee Cho<br />

Digital/Scientific <strong>Imaging</strong> Lab, Graduate School <strong>of</strong> Advanced <strong>Imaging</strong> <strong>Science</strong>, Multimedia and Film, Chung<br />

Ang University, Seoul 156-756, South Korea<br />

Jeongho Shin<br />

Department <strong>of</strong> Web In<strong>for</strong>mation Engineering, Hankyong National University, Anseong 456-749,<br />

South Korea<br />

Donghwan Har<br />

Digital/Scientific <strong>Imaging</strong> Lab, Graduate School <strong>of</strong> Advanced <strong>Imaging</strong> <strong>Science</strong>, Multimedia and Film, Chung<br />

Ang University, Seoul 156-756, South Korea<br />

Joonki Paik<br />

Image Processing and Intelligent Systems Lab, Graduate School <strong>of</strong> Advanced <strong>Imaging</strong> <strong>Science</strong>, Multimedia<br />

and Film, Chung Ang University, Seoul 156-756, South Korea<br />

Abstract. This paper proposes a novel color shift model-based<br />

segmentation and fusion algorithm <strong>for</strong> digital aut<strong>of</strong>ocusing <strong>of</strong> color<br />

images. The source images are obtained using new multiple filteraperture<br />

configurations. We shift color channels to change the focal<br />

point <strong>of</strong> the given image at different locations. For each respective<br />

location we then select the optimal focus in<strong>for</strong>mation and, finally,<br />

use s<strong>of</strong>t decision fusion and blending (SDFB) to obtain fully-focused<br />

images. The proposed aut<strong>of</strong>ocusing algorithm consists <strong>of</strong>: (i) color<br />

channel shifting and alignment <strong>for</strong> varying focal positions; (ii) optimal<br />

focus region selection and segmentation using sum modified Laplacian<br />

(SML); and (iii) SDFB, which enables smooth transition<br />

across region boundaries. By utilizing segmented images <strong>for</strong> different<br />

focal point locations, the SDFB algorithm can combine images<br />

with multiple, out-<strong>of</strong>-focus objects. Experimental results show per<strong>for</strong>mance<br />

and feasibility <strong>of</strong> the proposed algorithm <strong>for</strong> aut<strong>of</strong>ocusing<br />

images with one or more differently out-<strong>of</strong>-focus objects. © 2007<br />

<strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4368<br />

Received Sep. 25, 2006; accepted <strong>for</strong> publication Mar. 22, 2007.<br />

1062-3701/2007/514/368/12/$20.00.<br />

INTRODUCTION<br />

Demand <strong>for</strong> digital aut<strong>of</strong>ocusing techniques is rapidly increasing<br />

in many visual applications, such as camcorders,<br />

digital cameras, and video surveillance systems. Until now,<br />

most focusing ef<strong>for</strong>ts have been put on gray scale images.<br />

Even with specialized color processing techniques, each color<br />

channel is processed independently <strong>for</strong> aut<strong>of</strong>ocusing applications.<br />

In this paper, a novel aut<strong>of</strong>ocusing algorithm utilizing<br />

color shift property is proposed, which can restore an<br />

image with multiple, differently focused objects. We propose<br />

a new filter-aperture (FA) model <strong>for</strong> aut<strong>of</strong>ocusing color images.<br />

The proposed method overcomes the fusion with multiple<br />

source images as it uses a single input image. The FA<br />

model separates and distributes the out-<strong>of</strong>-focusing blur<br />

into different color channels. The multiple FA models also<br />

make it possible to generate as many source images as necessary<br />

<strong>for</strong> fusion-based aut<strong>of</strong>ocusing. Multiple focal points<br />

are spotted on the image and color channel shifting aligns<br />

each channel with the respective focal point. For each alignment<br />

the sum modified Laplacian (SML) operator is used to<br />

obtain a numerical measure indicating the degree <strong>of</strong> focus <strong>of</strong><br />

that image. The in-focus pixels are selected and combined at<br />

each process using s<strong>of</strong>t decision fusion and blending (SDFB)<br />

to produce the in-focus image with maximum focus metric.<br />

The SML operator can also be used to estimate a number <strong>of</strong><br />

focal points starting from the minimum degree <strong>of</strong> focus in<br />

the input image. The proposed algorithm does not use any<br />

restoration filter, which usually results in undesired artifacts,<br />

such as ringing, reblurring, and noise clustering.<br />

The rest <strong>of</strong> the paper is organized as follows. The following<br />

section summarizes existing techniques, and presents<br />

the major contribution <strong>of</strong> the proposed work. The section<br />

titled “Multiple FA model” gives a detailed description <strong>of</strong> the<br />

multiple FA method and “Digital Aut<strong>of</strong>ocusing Algorithm”<br />

describes the proposed aut<strong>of</strong>ocusing algorithm. “Experi-<br />

368


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

EXISTING STATE-OF-THE-ART AUTOFOCUSING<br />

METHODS<br />

FA Model<br />

The conventional photo sensor array uses micro lenses in<br />

front <strong>of</strong> every pixel to concentrate light onto the photosensitive<br />

region. 1 In this paper, we can interpret the optical<br />

design in a gradual step that we are able to make the multiple<br />

detectors beneath the each micro lens, instead <strong>of</strong> multiple<br />

arrays <strong>of</strong> detectors. The artificial compound eye sensor<br />

(insect eyes) is composed <strong>of</strong> a micro lens array and a photo<br />

sensor. 2 However, the imaging quality <strong>of</strong> these optical designs<br />

is fundamentally inferior to a camera system with a<br />

large single lens; the resolution <strong>of</strong> these small lens arrays is<br />

severely limited by diffraction. The “wave front coding”<br />

system 3 is similar to the proposed system (see Figure 1) in<br />

that it provides a way to decouple the trade-<strong>of</strong>f between<br />

aperture size and depth <strong>of</strong> field, but their design is very<br />

different. Rather than collecting and resorting rays <strong>of</strong> light,<br />

they use aspheric lenses that produce images with a depthindependent<br />

blur. Deconvolution <strong>of</strong> these images retrieves<br />

image details at all depths as shown in Figure 2.<br />

Figure 1. Block diagram <strong>of</strong> the proposed algorithm.<br />

mental Results” shows the simulation results and comparisons<br />

with existing methods. Finally, we have the concluding<br />

remarks.<br />

Aut<strong>of</strong>ocusing Methods<br />

The traditional aut<strong>of</strong>ocusing system in a camera usually<br />

consists <strong>of</strong> two different modules: analysis and control. The<br />

analysis module estimates a degree-<strong>of</strong>-focus <strong>of</strong> an image<br />

projected onto the image plane. The control module per<strong>for</strong>ms<br />

focusing functions by moving the lens assembly to the<br />

optimal focusing position according to the degree-<strong>of</strong>-focus<br />

in<strong>for</strong>mation estimated in the analysis module. There are five<br />

different focusing techniques, such as manual focusing<br />

(MF), infrared aut<strong>of</strong>ocusing (IRAF), through-the-lens<br />

aut<strong>of</strong>ocusing (TTLAF), semi-digital aut<strong>of</strong>ocusing (SDAF),<br />

and fully digital aut<strong>of</strong>ocusing (FDAF). 4–7 Table I briefly<br />

summarizes and compares those techniques.<br />

The FDAF systems usually involve restoration and fusion<br />

methods in the control module which operates using<br />

prior in<strong>for</strong>mation like point spread function (PSF), gradients,<br />

multiple source inputs, etc. to obtain the details about<br />

out-<strong>of</strong>-focus blur in images. Image fusion-based aut<strong>of</strong>ocus-<br />

Figure 2. Representation <strong>of</strong> the schematic <strong>of</strong> the a wave front coding system, b proposed FA system.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 369


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Table I. Comparison <strong>of</strong> conventional AF systems with the proposed system.<br />

Aut<strong>of</strong>ocusing<br />

technique Analysis module Control module<br />

Focusing<br />

accuracy<br />

Hardware<br />

specifications<br />

Manual Human decision Manual Subject to human operation Low shutter speed f/2-f/8.0<br />

IRAF Calculating the time <strong>of</strong> IR travel Moving the focusing lens High High shutter speed f/3.5-f/5.6<br />

TTLAF Minimizing a phase difference Moving the focusing lens Very high under good conditions High shutter speed f/2-f/11<br />

SDAF Calculating high frequency <strong>of</strong> image Moving the focusing lens Acceptable Medium shutter speed f/2-f/28<br />

FDAF Estimating PSF, blur models Restoration filters and fusion methods Acceptable NIL<br />

Proposed method Color channel shifting Multiple filter aperture FA Acceptable 30 to 1 / 4,000 sec. f/5.6-f/22<br />

ing methods have focused on operation <strong>of</strong> multiple source<br />

images using wavelet or discrete cosine trans<strong>for</strong>mations<br />

(DCT) 8,9 with a priori obtained camera PSF. Other methods<br />

use pyramid-based representation to decompose the source<br />

images into different spatial scales and orientations. 10,11<br />

Similar results, although with more artifacts and less visual<br />

stability, can be achieved by using a set <strong>of</strong> basis functions. 12<br />

Another technique similar to pyramid representation approach<br />

has been based on wavelet trans<strong>for</strong>m to decompose<br />

the image into various subbands. 13,14 The output is generated<br />

by selecting one <strong>of</strong> the decomposed subbands such that<br />

the selected subband has maximum energy. Restorationbased<br />

techniques have been carried out to overcome the out<strong>of</strong>-focus<br />

problem. However, restoration <strong>of</strong> images with different<br />

depth <strong>of</strong> fields tend to cause reblurring and ringing<br />

artifacts in the region with low depth <strong>of</strong> field or in-focus<br />

regions. 15,16 Even with equal depth <strong>of</strong> field the nature <strong>of</strong><br />

restoration poses a serious limitation to the visual quality <strong>of</strong><br />

the restored images. Another drawback is the slow convergence<br />

process <strong>of</strong> the iterative framework.<br />

The main contribution <strong>of</strong> the proposed method is listed<br />

below:<br />

(a) Multiple apertures and corresponding sensors can<br />

enhance depth in<strong>for</strong>mation.<br />

(b) Focusing process is inherently designed in accordance<br />

with color in<strong>for</strong>mation.<br />

(c) Neither image restoration nor blur identification is<br />

necessary.<br />

Figure 3. General single aperture model.<br />

(d) Set <strong>of</strong> images with multiple apertures and focus<br />

settings can be generated using a single image with<br />

channel shifting,<br />

(e) Fusion algorithm involves separate feature-based<br />

fusion and color blending consistency to preserve<br />

the channel dependencies.<br />

(f) Proposed algorithm does not need trans<strong>for</strong>mation<br />

or convolution operations.<br />

Recently, images obtained at different shutter speeds<br />

were combined into an image in which full dynamic range is<br />

preserved. 17 The proposed approach extends and generalizes<br />

the standard fusion approach to color images. The proposed<br />

approach does not need multiple source images captured at<br />

different aperture settings. Instead we derive different source<br />

images from a single out-<strong>of</strong>-focus image to obtain various<br />

positions <strong>of</strong> focal points. 18–20 For each focal point three color<br />

channels are aligned and the corresponding images are used<br />

<strong>for</strong> fusion.<br />

MULTIPLE FILTER-APERTURE (FA) MODEL<br />

An aperture <strong>of</strong> a lens can adjust the amount <strong>of</strong> incoming<br />

light accepted through the lens. It can also control the focal<br />

length, camera-to-object distance, and depth <strong>of</strong> field. Generally,<br />

the center <strong>of</strong> an aperture is aligned on the optical axis<br />

<strong>of</strong> the lens. Any controlled aperture accepts light from various<br />

angles depending on the object position. Correspondingly,<br />

the convergence pattern on the imaging plane <strong>for</strong>ms<br />

either a point or a circular region as shown in Figure 3. For<br />

370 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 4. Aperture shifted from the center.<br />

Figure 5. Multiple aperture set-up <strong>for</strong> red and blue channel filters.<br />

objects placed at near-, mid-, and far-focal distance the image<br />

convergence takes place either in front, on, or behind the<br />

CCD/CMOS sensor. However, this image convergence as<br />

well as the blur in<strong>for</strong>mation can only be represented in a<br />

bi-axial plane as shown in Fig. 3.<br />

An interesting alternative <strong>for</strong> tri-axial representation <strong>of</strong><br />

the image and out-<strong>of</strong>-focus blur was found to be achieved<br />

using non-centric aperture as shown in Figure 4. For a noncentric<br />

aperture located either on the upper or lower part <strong>of</strong><br />

the optical axis, the convergence pattern was found to be<br />

split between these axes. The split difference between the<br />

patterns will give another dimension to the conventional biaxial<br />

plane making it a tri-axial representation. For the objects<br />

at the same positions (near, in, far focal distances), the<br />

convergence pattern <strong>of</strong> the channel aperture <strong>for</strong>m an overlapping<br />

convergence on the CCD/CMOS sensor. For instance,<br />

the near focal distance object converges on the upper<br />

part <strong>of</strong> the optical axis where, at the same position, the far<br />

focal distance object converges on the lower part. If these<br />

overlapping channels are exactly aligned, then we will have a<br />

focused pattern in the image.<br />

An extension <strong>of</strong> the above approach will be to use a lens<br />

with two apertures on either side <strong>of</strong> the optical axis. An<br />

interesting phenomenon that can be observed is that, <strong>for</strong> the<br />

near and far focused objects, the convergence pattern lies on<br />

opposite sides <strong>for</strong> each aperture in reverse order <strong>for</strong> each<br />

channel. For example, the red aperture can have nearfocused<br />

convergence on the top and far-focused convergence<br />

on the bottom whereas the blue aperture has far-focused<br />

convergence on the top and near-focused convergence at the<br />

bottom, as shown in Figure 5. This phenomenon is called<br />

the filter-aperture (FA) extraction. The out-<strong>of</strong>-focus blur is<br />

now distributed among the color channels <strong>for</strong>ming the<br />

image.<br />

Now we extend the above multiple aperture convergence<br />

to a typical RGB image scenario. To obtain an RGB<br />

image using the multiple aperture configurations we need to<br />

obtain R, G, and B channel convergence patterns separately.<br />

This can be done using three apertures in a Bayer pattern<br />

where the images are individually obtained on the sensor <strong>for</strong><br />

the three apertures and later combined to <strong>for</strong>m the RGB<br />

image. Evidently, multiple apertures provide additional<br />

depth in<strong>for</strong>mation <strong>of</strong> objects at different distances. Since any<br />

color image is composed <strong>of</strong> three channels, we have used<br />

three apertures and, correspondingly, three filters (see<br />

Figure 6).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 371


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 6. Multiple FA model showing the convergence pattern <strong>for</strong> the R, G, and B color channels.<br />

The main advantage <strong>of</strong> the FA model is that it can provide<br />

an alternative method <strong>for</strong> blur estimation in aut<strong>of</strong>ocusing<br />

applications. Images acquired by using a normal lens<br />

have uni<strong>for</strong>m or spatially variant blur confined on all channels.<br />

However, in the proposed algorithm, by using three<br />

filtered sensors the aut<strong>of</strong>ocusing problem turns into the<br />

alignment <strong>of</strong> R, G, and B channels with various depths <strong>of</strong><br />

field. The out-<strong>of</strong>-focusing phenomenon with single and<br />

multiple aperture lenses are compared in Figure 7. As shown<br />

in Fig. 7(b) the out <strong>of</strong> focus blur is modeled as a misalignment<br />

<strong>of</strong> three color channels <strong>of</strong> R, G, and B.<br />

DIGITAL AUTOFOCUSING ALGORITHM<br />

The proposed algorithm uses the image obtained from the<br />

multiple FA configurations <strong>for</strong> the aut<strong>of</strong>ocusing application.<br />

The proposed aut<strong>of</strong>ocusing algorithm consists <strong>of</strong> the following<br />

procedures to obtain a well-restored image: (i) salient<br />

feature computation, (ii) color channel shifting and alignment<br />

<strong>for</strong> selected pixels, and (iii) s<strong>of</strong>t decision fusion and<br />

blending.<br />

Salient Focus Measure<br />

The feature saliency computation process contains a family<br />

<strong>of</strong> functions that estimate saliency in<strong>for</strong>mation. In practice,<br />

these functions can operate on individual pixels or on a local<br />

region <strong>of</strong> pixels. When combining images having different<br />

focus measures, <strong>for</strong> instance, a desirable saliency measure<br />

would provide a quantitative measure that increases when<br />

features are better focused. Various saliency measures, including<br />

variance and gradients, have been employed and<br />

Figure 7. Comparison <strong>of</strong> out-<strong>of</strong>-focus blurs <strong>for</strong> a single aperture model and the proposed multiple aperture<br />

models: a and b out-<strong>of</strong>-focus image captured using ordinary camera and proposed FA system under same<br />

focal settings, c restored result using the regularized restoration method, d restored result using the proposed<br />

channel shifting and fusion algorithm.<br />

372 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

validated <strong>for</strong> related applications. The saliency function only<br />

selects the frequencies in the focused image that will be attenuated<br />

due to defocusing. One way to detect a high frequency<br />

component is to apply the following absolute<br />

Laplacian operator as<br />

2 L k =<br />

2 2 L k<br />

+ 2 2 L k<br />

.<br />

x y<br />

The second derivatives in the x and y directions <strong>of</strong>ten have<br />

opposite signs and tend to cancel each other. In the case <strong>of</strong><br />

textured images, this phenomenon may frequently occur and<br />

the Laplacian behaves in an unstable manner. However, this<br />

problem can be overcome by using the absolute Laplacian as<br />

in Eq. (1). In order to accommodate <strong>for</strong> possible variations<br />

in the size <strong>of</strong> texture elements, we compute the partial derivative<br />

using a variable spacing between pixels <strong>for</strong> computing<br />

the derivatives. Hence a discrete approximation to the<br />

modified Laplacian, ML k i,j, <strong>for</strong> pixel intensity, Ii,j, is<br />

given as<br />

ML k i,j = 2Ii,j − Ii −1,j − Ii +1,j<br />

1<br />

+ 2Ii,j − Ii,j −1 − Ii,j +1. 2<br />

Finally, the focus measure at a point i,j is computed as the<br />

sum <strong>of</strong> modified Laplacian values, in a small window around<br />

i,j, that are greater than a prespecified threshold value,<br />

i+N<br />

fi,j = <br />

j+N<br />

<br />

p=i−N q=j−N<br />

ML k p,q <strong>for</strong> ML k p,q T 1 .<br />

The heuristically determined threshold value T 1 in the range<br />

40–60 provides acceptable results in most cases. The parameter<br />

N represents the window size <strong>for</strong> computing the focus<br />

measure. In contrast to region-based aut<strong>of</strong>ocusing methods,<br />

we typically use a smaller window <strong>of</strong> size, e.g., N=1. Equation<br />

(3) can be referred to as sum modified Laplacian (SML)<br />

which is used as an intermediate image estimate <strong>for</strong> determining<br />

focus in<strong>for</strong>mation.<br />

3<br />

Figure 8. Schematic <strong>of</strong> channel alignment procedure <strong>for</strong> R, G, and B<br />

channels.<br />

Color Channel Shift and Alignment<br />

For shifting and aligning color channels we need to find the<br />

optimal pixel-<strong>of</strong>-interest at different positions in the image<br />

according to their focal measures. These pixels-<strong>of</strong>-interest<br />

can be referred to as a focal point pixels. The term “focal<br />

point pixel” refers to a pixel-<strong>of</strong>-interest around which channel<br />

shifting and alignment is carried out. For a given image,<br />

the SML measure can be used to determine the focal point<br />

region whose focal measure is significantly lower than other<br />

regions <strong>of</strong> the image. Then <strong>for</strong> a given region, we select the<br />

focal point pixel either from the center <strong>of</strong> region or the pixel<br />

with the lowest focus measure. Similar operations can be<br />

per<strong>for</strong>med <strong>for</strong> different selected focal point regions in different<br />

neighborhood. Hence<strong>for</strong>th, <strong>for</strong> a corresponding focal<br />

point pixel, we per<strong>for</strong>m channel alignment and remove the<br />

out-<strong>of</strong>-focus blur in that given neighborhood (see Figure 8).<br />

For a given particular image captured by using FA configuration,<br />

the out-<strong>of</strong>-focus blur was just confined to channels<br />

on either side <strong>of</strong> the green channel as shown in Figure 9.<br />

As can be seen from the figure, the green channel suffers<br />

minimal blur distortion as the sensor was placed at the center<br />

whereas the red and the blue channels have maximal blur<br />

distortion. The proposed aut<strong>of</strong>ocusing technique uses the<br />

green channel as the reference and aligns the red and the<br />

blue channels to the green channel <strong>for</strong> any particular location,<br />

such as<br />

I RGB = S r,c I R + I B + I G ,<br />

where S r,c represents the shift operator and the shift vector<br />

r,c represents the amount <strong>of</strong> shift in row and column directions<br />

<strong>for</strong> the respective red and blue channels with respect<br />

to the reference focal point on the green channel. If the shift<br />

vectors are not identical, we can generalize the above equation<br />

as<br />

I RGB = SI R r 1 ,c 1 + I B r 2 ,c 2 + I G .<br />

The shift vectors on the same sensor filter are linearly dependent.<br />

For a particular reference channel it is possible to<br />

estimate the exact number <strong>of</strong> shift vectors using the sensor<br />

filter configurations. For example, in our experiments the<br />

green channel has been used as reference, hence the red and<br />

blue pixels are misaligned by a pattern corresponding to the<br />

sensor filter as shown in Figure 10.<br />

S<strong>of</strong>t Decision Fusion and Blending<br />

In order to merge images with multiple focal point planes,<br />

image fusion is required on multiple channel images. Un<strong>for</strong>tunately,<br />

when the channel-shifted images are directly fused,<br />

misalignment or misregistration is unavoidable. The pixels<br />

<strong>of</strong> different channel aligned images, when fused together,<br />

may sometimes tend to overlap or get missed because <strong>of</strong> the<br />

channel shifting. This problem can be overcome by applying<br />

an inverse shift operation to the images with respect to a<br />

reference image. The reference image has to be chosen from<br />

one <strong>of</strong> the several channel shifted images extracted using<br />

channel shifting and alignment. In the proposed approach<br />

we choose the reference image as the one that will have a<br />

focal point located approximately in the center <strong>of</strong> the image,<br />

−1<br />

I k = S r,cIr I k ,I r ,<br />

where I r represents the reference image <strong>for</strong> registration and<br />

S −1 represents the inverse shift operation. After selection <strong>of</strong><br />

4<br />

5<br />

6<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 373


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 10. Row and column shift vectors <strong>for</strong> color channel shifting and<br />

alignment with reference G channel.<br />

Figure 9. Multiple FA model image: a R, b G, and c B channels,<br />

respectively.<br />

the focal point using the SML operator and channel alignment<br />

by channel shifting, the appropriate regions need to be<br />

selected from the given image. Given the location <strong>of</strong> the<br />

pixel-<strong>of</strong>-interest <strong>for</strong> channel alignment, we simply select an<br />

approximate region area that is defined on its neighborhood.<br />

But <strong>for</strong> a more efficient fusion process we could isolate the<br />

region around that pixel using neighborhood connectivity.<br />

For a given pixel-<strong>of</strong>-interest and the eight-neighborhood<br />

connectivity, we can extract the region more accurately <strong>for</strong><br />

the purpose <strong>of</strong> image fusion as<br />

Figure 11. Experimental set up: a digital camera with the FA model;<br />

b interior configuration <strong>of</strong> the FA; and c the R, G, and B sensor filters.<br />

374 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Table II. A sample set <strong>of</strong> shift vectors estimated <strong>for</strong> different locations.<br />

Table III. Hardware configuration <strong>for</strong> the multiple FA system.<br />

Region Red channel r ,c Blue channel r ,c<br />

Hardware title<br />

Specifications<br />

R 1 upper left 10,4 8,5<br />

R 2 upper middle 7,6 8,5<br />

R 3 upper right 10,5 7,6<br />

R 4 center left 9,4 10,5<br />

R 5 center middle 9,3 9,3<br />

R 6 center right 8,6 10,2<br />

R 7 lower left 10,4 9,4<br />

R 8 lower middle 11,6 9,4<br />

R 9 lower right 11,5 8,3<br />

i+N<br />

F k = <br />

j+N<br />

<br />

x=i−N y=j−N<br />

f p xs i , ...,s i+k ,yt i , ...,t j+k ,<br />

where F k represents the region around the pth focal point<br />

pixel, f p and s,t represent the neighborhood connectivity.<br />

Even though the SML operator can provide an accurate<br />

measure, we need to extract the specific region from the<br />

7<br />

Digital camera Nikon D-100<br />

R, G, B filters Green-Kodak-Wratten Filter No. 58<br />

Blue-Kodak-Wratten Filter No. 47<br />

Red-Kodak-Wratten Filter No. 25<br />

Focusing<br />

APO-Symmar-L-150-5.6,11,22<br />

f-5.6, f-11, f-22<br />

Sensor<br />

23.7 15.6 mm RGB CCD; 6.31 million total<br />

pixels<br />

Lens mounting<br />

Schneider Apo-Tele-Xenar<br />

Relative aperture focal length −5.6/ 250<br />

Shutter speed<br />

30 to 1 / 4,000 sec. and bulb<br />

Color mode<br />

Triple mode <strong>for</strong> R, G, and B channels<br />

image <strong>for</strong> fusion. One <strong>of</strong> disadvantages <strong>of</strong> the FA model is<br />

that, <strong>for</strong> the channel-aligned images with closely located focal<br />

points, the SML operator does not always per<strong>for</strong>m well.<br />

Hence, we used a color-based region segmentation algorithm<br />

<strong>for</strong> extracting selective regions from the channel-<br />

Figure 12. Experimental results: a the source image; b the focal point location <strong>for</strong> channel shifting and<br />

alignment; c the image after channel shifting to new focal point location; and d the final image fused from<br />

a and c.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 375


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 13. Experimental results: a the source image; b objects closer to the left side <strong>of</strong> the image are<br />

focused and object to the right side are out-<strong>of</strong>-focus; c similar set up with the focus on the right side; and d<br />

the final image fused from b and c.<br />

aligned images if the SML results are not good enough.<br />

When fusing color images, features such as edge and textures<br />

should be preserved, and also the color blending consistency<br />

should be maintained. The fusion process is per<strong>for</strong>med on<br />

each level <strong>of</strong> the channel-aligned images in conjunction with<br />

SML to generate the composite image C. The reconstruction<br />

process integrates in<strong>for</strong>mation from different levels as<br />

and<br />

I ck = F k · I ak + 1−F k I bk<br />

−1<br />

I a = S r,cIr I a ,I r ,<br />

8<br />

9<br />

where I ck represents the reconstructed image from two input<br />

images I ak and I bk . The variable k represents the regions<br />

extracted based on their respective focal measure. The inverse<br />

shifting operation is described in Eq. (9) where I r represents<br />

the reference image and r,cI r the corresponding<br />

shift vectors with respect to I r . A typical problem <strong>of</strong> image<br />

fusion is the appearance <strong>of</strong> unnatural borders between the<br />

different decisions regions due to overlapping blur at focus<br />

boundaries. To combat this, s<strong>of</strong>t decision blending can be<br />

employed using smoothing or low pass filtering <strong>of</strong> the saliency<br />

parameter F k . In this paper Gaussian smoothing has<br />

been used to obtain the desired effect <strong>of</strong> blending. This creates<br />

weighted decision regions where a linear combination <strong>of</strong><br />

pixels in the two images A and B are used to generate corresponding<br />

pixels in the fused image C. We then have<br />

376 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 14. Experimental results: a the source image; b–c channel shifted images <strong>for</strong> focal points at near<br />

and center positions; and d the final image obtained by fusing a–c.<br />

I ck = F˜k · I ak + 1−F˜kI bk ,<br />

10<br />

where F˜k is now a smoothed version <strong>of</strong> its <strong>for</strong>mer self. At<br />

times there can be missing pixels in the fused image which<br />

are not selected using the SML. The number <strong>of</strong> missing<br />

pixels varies from image to image, but is always confined to<br />

a very small portion <strong>of</strong> the entire image. The missing pixels<br />

have to be replaced with pixels from any one <strong>of</strong> the available<br />

channel aligned images. One simple way to find an appropriate<br />

replacement is to get the location <strong>of</strong> the missing pixel<br />

in the image and then match it with the image whose focal<br />

point pixel is nearest to the respective missing pixel:<br />

Ix,y = min disIx,y,f p x,y.<br />

11<br />

EXPERIMENTAL RESULTS<br />

Dataset Simulation and Experiments<br />

In order to demonstrate the per<strong>for</strong>mance <strong>of</strong> the proposed<br />

algorithm, we used test images captured using the proposed<br />

multiple FA model with multiple out-<strong>of</strong>-focus objects in the<br />

background. The experimental setup is shown in Figure 11<br />

which represents the camera used <strong>for</strong> the experiments along<br />

with the multiple FA configurations <strong>of</strong> the camera and the<br />

sensor filter. The hardware specifications used <strong>for</strong> the system<br />

are listed in Table III. Experiments were per<strong>for</strong>med on an<br />

RGB image <strong>of</strong> size 640480. Here, each image contains<br />

multiple objects at different distances from the camera.<br />

Figure 12(a) represents a test image with low depth-<strong>of</strong>-field,<br />

where focus is on the objects close to the camera lens. The<br />

channels aligned <strong>for</strong> the focal point are shown in Fig. 12(b).<br />

The image after channel shifting is shown in Fig. 12(c). The<br />

blue object in the back <strong>of</strong> the astronaut was out <strong>of</strong> focus in<br />

Fig. 12(a), which is now in-focus in Fig. 12(c), whereas the<br />

other regions <strong>of</strong> the image tend to get defocused. The fused<br />

image <strong>of</strong> Figs. 12(a) and 12(c) is shown in Fig. 12(d). Similar<br />

results with multiple objects are shown in Figures 13 and<br />

14. The selected focal point <strong>for</strong> the channel alignment<br />

and shifting are represented in Figures 15(f) and 15(g).<br />

Table IV. Image quality comparisons <strong>for</strong> the various aut<strong>of</strong>ocusing methods.<br />

Aut<strong>of</strong>ocusing<br />

method<br />

Prior<br />

in<strong>for</strong>mation<br />

Mode<br />

Input<br />

frames Operation RMSE PSNR<br />

Wiener filter PSF Gray 1 Pixel based 12.35 23.36<br />

Iterative filter NIL Gray 1 Pixel based 8.56 26.32<br />

Constrained least<br />

Edge Gray 1 Pixel based 9.56 25.10<br />

square filter<br />

Pyramid fusion NIL Gray, Color At least 2 Window based 5.68 28.42<br />

and Pixel based<br />

Wavelet fusion NIL Gray, Color At least 2 Window based 5.02 29.95<br />

and Pixel based<br />

Proposed NIL Color 1 Window based<br />

and Pixel based<br />

8.06 26.41<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 377


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Figure 15. Experimental results: a the source image; b–d channel shifted images <strong>for</strong> focal points in right,<br />

center, and left positions; e–g the SML results <strong>for</strong> selected regions; and h the final image obtained fusing<br />

b–d.<br />

Figures 15(b)–15(d) illustrate the results <strong>of</strong> SML operator<br />

<strong>for</strong> a selected region. These figures represent images which<br />

have different out-<strong>of</strong>-focus regions obtained from single<br />

source images using channel shifting. Figure 15(e) represents<br />

the fused image from Figs. 15(b)–15(d). The resulting fused<br />

image contains in-focus regions from respective images. The<br />

above set <strong>of</strong> results illustrates the feasibility <strong>of</strong> the proposed<br />

fusion based algorithm <strong>for</strong> aut<strong>of</strong>ocusing.<br />

378 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Maik et al.: Color shift model-based segmentation and fusion <strong>for</strong> digital aut<strong>of</strong>ocusing<br />

Per<strong>for</strong>mance Comparison<br />

For measuring the per<strong>for</strong>mance <strong>of</strong> the multiple FA configurations,<br />

various test images were captured using the proposed<br />

system as well as the ordinary Nikon D-100 camera.<br />

The test images were then processed <strong>for</strong> out-<strong>of</strong>-focus removal<br />

using the proposed channel shifting and fusion algorithm<br />

and the ordinary camera images were restored using<br />

some state <strong>of</strong> the art restoration methods, including Wiener<br />

filter, regularized iterative restoration, constrained least<br />

squares filter, as well as some existing fusion-based methods<br />

including pyramid decomposition and wavelet methods. The<br />

per<strong>for</strong>mance metric in the <strong>for</strong>m <strong>of</strong> PSNR and RMSE were<br />

obtained <strong>for</strong> the test images using the above algorithms as<br />

given in Table IV. As can be seen in the table, the images<br />

captured using the multiple FA configurations tend to have<br />

some degradation when compared to conventional camera<br />

images when there is no out-<strong>of</strong>-focus blurs. But with the<br />

out-<strong>of</strong>-focus blur the image quality <strong>of</strong> the conventional camera<br />

images tend to drastically reduce due to processing by<br />

restoration and is more or less comparable with the restored<br />

images using the color channel shifting and fusion. However,<br />

the fusion methods tend to give slightly higher image quality,<br />

but they require multiple source input images <strong>for</strong> achieving<br />

good per<strong>for</strong>mance, whereas the proposed method can<br />

achieve it with just a single source input image making our<br />

method more suitable and efficient <strong>for</strong> increasing potential<br />

applications.<br />

For aligning the blue channel with the green channel<br />

the pixels have to be shifted in an upward direction and<br />

towards the left or diagonally to the left and vice versa <strong>for</strong><br />

the red channel. In our experiments we tried precomputing<br />

the shift vectors at nine different locations on a test image<br />

manually using the above convention. We found that the<br />

shift vectors differ slightly <strong>for</strong> different regions on the image,<br />

as shown in Table II. These shift vectors were then used<br />

accordingly <strong>for</strong> various test images based on the location <strong>of</strong><br />

the focal point pixel in one <strong>of</strong> the nine regions. The corresponding<br />

shift vectors were then used to align the channels.<br />

CONCLUSIONS<br />

In this paper, we proposed an aut<strong>of</strong>ocusing algorithm which<br />

restores an out-<strong>of</strong>-focus image with multiple, differently out<strong>of</strong>-focus<br />

objects. A novel FA configuration is proposed <strong>for</strong><br />

modeling out-<strong>of</strong>-focus blur in images. The proposed algorithm<br />

starts with a single input image and multiple source<br />

images with different apertures are generated using channel<br />

shifting. The fusion is carried out <strong>for</strong> segmented regions<br />

from each source image using the SML operator. The s<strong>of</strong>t<br />

decision fusion algorithm overcomes undesired artifacts in<br />

the region <strong>of</strong> merging in the fused images. Experimental<br />

results show that the proposed algorithm works well <strong>for</strong> the<br />

images with multiple out-<strong>of</strong>-focus objects.<br />

ACKNOWLEDGMENTS<br />

This research was supported by Seoul Future Contents Convergence<br />

(SFCC) Cluster established by Seoul R&BD Program<br />

and by the Korea <strong>Science</strong> and Engineering Foundation<br />

(KOSEF) through the National Research Laboratory Program<br />

funded by the Ministry <strong>of</strong> <strong>Science</strong> and Technology<br />

(M103 0000 0311-06J0000-31110).<br />

REFERENCES<br />

1 Y. Ishihara and K. Tanigaki, “A high photosensitivity IL-CCD image<br />

sensor with monolithic resin lens array”, in Proc. IEEE Integrated<br />

Electronic and Digitial Microscopy (IEEE Press, Piscataway, NJ, 1983) pp.<br />

497–500.<br />

2 J. Tanida, R. Shogenji, Y. Kitumara, K. Yamada, M. Miyamoto, and S.<br />

Miyatake, “Color image with an integrated compound imaging system”,<br />

Opt. Express 18, 2109–2117 (2003).<br />

3 E. R. Dowski and G. E. Johnson, “Wavefront coding system: amodern<br />

method <strong>of</strong> achieving high per<strong>for</strong>mance and/or low cost imaging<br />

systems”, Proc. SPIE 3779, 137–145 (1999).<br />

4 S. Kim and J. K. Paik, “Out-<strong>of</strong>-focus blur estimation and restoration <strong>for</strong><br />

digital aut<strong>of</strong>ocusing system”, Electron. Lett. 34, 1217–1219 (1998).<br />

5 J. Shin, V. Maik, J. Lee, and J. Paik, “Multi-object digital aut<strong>of</strong>ocusing<br />

using image fusion”, Lect. Notes Comput. Sci. 3708, 806–813 (2005).<br />

6 V. Maik, J. Shin, and J. Paik, “Regularized image restoration by means <strong>of</strong><br />

fusion <strong>for</strong> digital aut<strong>of</strong>ocusing”, Lecture Notes in Artificial Intelligence<br />

3802, 929–934 (2005).<br />

7 G. Ligthart and F. Groen, “A comparison <strong>of</strong> different aut<strong>of</strong>ocus<br />

algorithms”, IEEE Int. Conf. Pattern Recognition (IEEE Press,<br />

Piscasataway, NJ, 1992) pp. 597–600.<br />

8 M. Subbarao, T. C. Wei, and G. Surya, “Focused image recovery from<br />

two defocused images recorded with different camera settings”, IEEE<br />

Trans. Image Process. 4, 1613–1628 (1995).<br />

9 M. Matsuyama, Y. Tanji, and M. Tanka, “Enhancing the ability <strong>of</strong><br />

NAS-RIF algorithm <strong>for</strong> blind image deconvolution”, Proc. IEEE Int.<br />

Conf. Circuits and Systems, vol. 4 (IEEE Press, Piscataway, NJ, 2000) pp.<br />

553–556.<br />

10 A. Tekalp and H. Kaufman, “On statistical identification <strong>of</strong> a class <strong>of</strong><br />

linear space-invariant image blurs using non minimum-phase ARMA<br />

models”, IEEE Trans. Acoust., Speech, Signal Process. ASSP36,<br />

1360–1363 (1988).<br />

11 K. Kodama, H. Mo, and A. Kubota, “All-in-focus image generation by<br />

merging multiple differently focused images in three-dimensional<br />

frequency domain”, Lect. Notes Comput. Sci. 3768, 303–314 (2005).<br />

12 K. Aizawa, K. Kodama, and A. Kubota, “Producing object-based special<br />

effects by fusing multiple differently focused images”, IEEE Trans.<br />

Circuits Syst. Video Technol. 10, 323–330 (2000).<br />

13 L. Bogoni and M. Hansen, “Pattern selective color image fusion”, Int. J.<br />

Pattern Recognit. Artif. Intell. 34, 1515–1526 (2001).<br />

14 S. Li, J. T. Kwok, and Y. Wang, “Combination <strong>of</strong> images with diverse<br />

focuses using the spatial frequency”, Int. J. Inf. Fusion 2, 169–176<br />

(2001).<br />

15 Z. Zhang and R. S. Blum, “A categorization <strong>of</strong> multiscaledecomposition-based<br />

image fusion schemes with a per<strong>for</strong>mance study<br />

<strong>for</strong> a digital camera application”, Proc. IEEE 87, 1315–1326 (1999).<br />

16 V. Maik, J. Shin, J. Lee, and J. Paik, “Pattern selective image fusion <strong>for</strong><br />

multi-focus image reconstruction”, Lect. Notes Comput. Sci. 3691,<br />

677–684 (2005).<br />

17 A. Kubota, K. Kodama, and K. Aizawa, “Registration and blur estimation<br />

method <strong>for</strong> multiple differently focused images”, Proc. IEEE Int. Conf.<br />

Image Processing, vol. 2 (IEEE Press, Piscataway, NJ, 1999) pp. 515–519.<br />

18 S. K. Lee, S. H. Lee, and J. S. Choi, “Depth measurement using frequency<br />

analysis with an active projection”, Proc. IEEE Conf. Image Processing,<br />

vol. 3 (IEEE Press, Piscataway, NJ, 1999) 906–909.<br />

19 G. Piella, “A general framework <strong>for</strong> multi resolution image fusion: From<br />

pixels to regions”, J. Inf. Fusion 4, 259–280 (2003).<br />

20 T. Adelson and J. Y. Wang, “Single lens stereo with a plenoptic camera”,<br />

IEEE Trans. Pattern Anal. Mach. Intell. 2, 99–106 (1992).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 379


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 380–385, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

Error Spreading Control in Image Steganographic<br />

Embedding Schemes Using Unequal Error Protection<br />

Ching-Nung Yang, Guo-Jau Chen and Tse-Shih Chen<br />

CSIE Dept., National Dong Hwa University, #1, Da Husueh Rd., Sec. 2, Hualien, Taiwan<br />

E-mail: cnyang@mail.ndhu.edu.tw<br />

Rastislav Lukac<br />

The Edward S. Rogers Sr. Department <strong>of</strong> ECE, University <strong>of</strong> Toronto, 10 King’s College Road, Toronto,<br />

Ontario, M5S 3G4 Canada<br />

Abstract. A steganographic scheme proposed by van Dijk and<br />

Willems can alter a relatively small amount <strong>of</strong> bits to hide the secret<br />

compared to other schemes while reducing the distortion and improving<br />

the resistance against the steganalysis. However, one bit<br />

error in the embedding scheme by van Dijk and Willems may result<br />

in multibit error when extracting the hidden data. This problem is<br />

called as error spreading. It is observed that only some single-bit<br />

errors suffer from error spreading. In this paper, we propose a new<br />

steganographic solution which takes advantage <strong>of</strong> unequal error<br />

protection codes and allows <strong>for</strong> the different protection <strong>of</strong> different<br />

secret bits. Thus the proposed solution can effectively protect bits<br />

which could suffer from error spreading. In addition, it saves parity<br />

bits, thus greatly reducing the amount <strong>of</strong> bit alterations compared to<br />

the relevant previous schemes. Experimentation using various test<br />

images indicates that the proposed solution achieves the trade<strong>of</strong>f<br />

between the per<strong>for</strong>mance and the protection <strong>of</strong> the embedded<br />

secret. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4380<br />

INTRODUCTION<br />

Steganography is a method <strong>of</strong> hiding or embedding the secret<br />

message into a cover media to ensure that an unintended<br />

party will not be aware <strong>of</strong> the existence <strong>of</strong> the embedded<br />

secret. Popular steganographic techniques <strong>for</strong> visual<br />

data protection embed the secret message such as a binary<br />

image by manipulating the least significant bit (LSB) plane<br />

<strong>of</strong> a cover image, thus producing the so-called stegoimage.<br />

In the simplest <strong>for</strong>m, the data embedding can be realized by<br />

c 0 s with c 0 denoting the least significant bit <strong>of</strong> a pixel from<br />

the cover image and s denoting the secret bit.<br />

A more efficient steganographic method was proposed<br />

by van Dijk and Willems. 1 Their coded LSB method attempts<br />

to reduce the distortion when the noise (e.g., due to faulty<br />

communication channels) or active attacks by the third party<br />

(intentional modifications <strong>of</strong> some insignificant bits in a<br />

cover image to prevent the extraction <strong>of</strong> hidden secret by the<br />

authorized user) are introduced into the stegoimage. However,<br />

in some situations, one error bit produced during the<br />

stegoimage transmission phase or by active attacks <strong>of</strong>ten results<br />

in two or more errors when decoding (extracting) the<br />

Received Oct. 28, 2006; accepted <strong>for</strong> publication Mar. 22, 2007.<br />

1062-3701/2007/514/380/6/$20.00.<br />

embedded message. This phenomenon, known as an error<br />

spreading problem, affects the clearness <strong>of</strong> the extracted secret<br />

image. A straight<strong>for</strong>ward application <strong>of</strong> error correction,<br />

i.e., using stegoencoding to hide the secret first and then<br />

adding the parity to provide the error correcting (EC) capability,<br />

will inevitably increase the required amount <strong>of</strong> bit<br />

alterations. Thus, the risk <strong>of</strong> being detected will increase.<br />

Zhang and Wang 2 proposed a new stegoencoding approach<br />

combining the coded LSB and EC capability simultaneously<br />

to address the error spreading problem. Their solution has<br />

the same error correcting capability <strong>for</strong> all protected bits, but<br />

according to our observation the error spreading does not<br />

affect all spatial locations <strong>of</strong> the secret image.<br />

In this paper, we propose a more reasonable solution,<br />

unequal error protection (UEP) codes, to obtain the different<br />

protection ability <strong>for</strong> nonaffected and affected bits and to<br />

save parity bits. It will be shown that the proposed solution<br />

outper<strong>for</strong>ms the previous relevant solutions in terms <strong>of</strong> the<br />

trade<strong>of</strong>f between the per<strong>for</strong>mance and the protection <strong>of</strong> the<br />

embedded secret.<br />

The rest <strong>of</strong> this paper is organized as follows. In the<br />

Coded LSB Scheme Section, the coded LSB scheme is described.<br />

In the EC Codes Based Error Correction: Zhang-<br />

Wang Scheme Section, Zhang-Wang stegoencoding based on<br />

EC codes is presented to show a solution <strong>for</strong> the error<br />

spreading problem. Our scheme based on UEP codes is proposed<br />

in the UEP Codes Based Error Correction the Proposed<br />

Scheme Section. Motivation and design characteristics<br />

are discussed in detail. In the Comparison and Experimental<br />

Results Section, the proposed method is tested using a variety<br />

<strong>of</strong> test images. The effect <strong>of</strong> UEP codes-based data embedding<br />

is evaluated and compared with the previous approaches.<br />

Finally, conclusions are drawn in the Conclusions<br />

Section.<br />

CODED LSB SCHEME<br />

In the plain LSB embedding scheme, the secret bits are hidden<br />

by simply replacing LSBs <strong>of</strong> the cover pixels. Due to the<br />

noiselike appearance <strong>of</strong> the LSB plane <strong>of</strong> natural images,<br />

embedding n bits implies, in average, the alteration <strong>of</strong> n/2<br />

original LSBs. To reduce the number <strong>of</strong> altered LSBs and<br />

380


Yang et al.: Error spreading control in image steganographic embedding schemes using unequal error protection<br />

Table I. Eight cosets <strong>for</strong> the coded LSB scheme with l=3, n=7 using 7,4 Hamming code with Gx=x 3 +x 2 +1.<br />

L 0<br />

L 1<br />

L 2<br />

L 3<br />

L 4<br />

L 5<br />

L 6<br />

L 7<br />

0000000, 0001101, 0011010, 0010111, 0110100, 0111001, 0101110, 0100011,<br />

1101000, 1100101, 1110010, 1111111, 1011100, 1010001, 1000110, 1001011<br />

0000001, 0001100, 0011011, 0010110, 0110101, 0111000, 0101111, 0100010,<br />

1101001, 1100100, 1110011, 1111110, 1011101, 1010000, 1000111, 1001010<br />

0000010, 0001111, 0011000, 0010101, 0110110, 0111011, 0101100, 0100001,<br />

1101010, 1100111, 1110000, 1111101, 1011110, 1010011, 1000100, 1001001<br />

0000011, 0001110, 0011001, 0010100, 0110111, 0111010, 0101101, 0100000,<br />

1101011, 1100110, 1110001, 1111100, 1011111, 1010010, 1000101, 1001000<br />

0000100, 0001001, 0011110, 0010011, 0110000, 0111101, 0101010, 0100111,<br />

1101100, 1100001, 1110110, 1111011, 1011000, 1010101, 1000010, 1001111<br />

0000101, 0001000, 0011111, 0010010, 0110001, 0111100, 0101011, 0100110,<br />

1101101, 1100000, 1110111, 1111010, 1011001, 1010100, 1000011, 1001110<br />

0000110, 0001011, 0011100, 0010001, 0110010, 0111111, 0101000, 0100101,<br />

1101110, 1100011, 1110100, 1111001, 1011010, 1010111, 1000000, 1001101<br />

0000111, 0001010, 0011101, 0010000, 0110011, 0111110, 0101001, 0100100,<br />

1101111, 1100010, 1110101, 1111000, 1011011, 1010110, 1000001, 1001100<br />

preserve the original features such as edges and fine details<br />

<strong>of</strong> the cover image, the coded LSB scheme <strong>of</strong> Ref. 1 divides<br />

the secret image into chips <strong>of</strong> l bits. Each l-bit chip is then<br />

embedded into LSBs <strong>of</strong> n pixels using n,k cyclic codes<br />

where n is the in<strong>for</strong>mation code length and l=n−k.<br />

k<br />

Let G 1 x= i=0 g 1,i x i l<br />

and G 2 x= i=0 g 2,i x i be two binary<br />

polynomials with degree k and l, respectively, so that<br />

G 1 x·G 2 x=x n +1. Using n,k cyclic codes with the generating<br />

function Gx=G 2 x, it is possible to construct 2 l<br />

code sets which consist <strong>of</strong> unique 2 k codewords. Thus, each<br />

code set can be used to describe one l-bit secret chip by<br />

choosing the nearest codeword to represent the embedded<br />

secret, as depicted in algorithm 1.<br />

Algorithm 1<br />

Inputs: secret message <strong>of</strong> l bits s l−1 s l−2 ...s 0 , and cover image<br />

O with n LSBs c n−1 c n−2 ...c 0 .<br />

Output: stegoimage O with n LSBs c n−1 c n−2 ...,c 0 .<br />

Step 1: Choose one cyclic n,k code with the generating<br />

function Gx=g l x l + ¯ +g 1 x+g 0 and then select any<br />

k-tuples as the input to construct a code set (coset) <strong>of</strong> 2 k<br />

codewords. Choose one n-tuple codeword that does not<br />

appear in this coset, and then add an unused n-tuple to<br />

all the codewords in the coset to construct another coset.<br />

Step 2: Repeat step 1, until all 2 n codewords are used. The<br />

process generates 2 l different cosets L 0 ,L 1 ,...,L 2 l −1 that<br />

include 2 k codewordsineachcoset.<br />

Step 3: Encrypt the secret bits s l−1 ,s l−2 ,...,s 0 by choosing<br />

the coset L i with i= l−1 i=0 s i 2 i . Then, find the codeword<br />

c n−1 c n−2 ...c 0 in the coset L i such that the Hamming distance<br />

between c n−1 c n−2 ...c 0 and c n−1 c n−2 ...c 0 is minimum.<br />

Step 4: Deliver n LSBs c n−1 c n−2 ...c 0 to the corresponding<br />

pixels in the embedded stegoimage O.<br />

As shown in Ref. 2, the efficiency <strong>of</strong> the steganographic<br />

schemes can be demonstrated using the so-called embedding<br />

rate (ER) which is defined as the number <strong>of</strong> embedded bits<br />

per pixel, i.e., ER=l/n. Another suitable criterion is the socalled<br />

embedding efficiency (EE) which is calculated as the<br />

number <strong>of</strong> altered bits per pixel, i.e., EE=l/l alt where l alt<br />

denotes the average LSB alteration when l secret bits are<br />

embedded into n LSBs. The value l alt can be calculated from<br />

all codewords in the cosets by a computer program. The ER<br />

parameter is suitable <strong>for</strong> discussing the embedded capacity<br />

whereas the EE parameter is used to evaluate the distortion<br />

in the cover image. It is obvious that <strong>for</strong> the plain LSB embedding<br />

scheme ER=l/n=n/n=100% and EE=1/l alt<br />

=n/n/2=200%. For the comparison purposes, the values<br />

corresponding to algorithm 1 (coded LSB scheme with<br />

l=3 and n=7) are provided below.<br />

Let us assume algorithm 1 with n=7 and l=3, i.e., the<br />

objective is to embed three secret bits into seven LSBs. The<br />

above setting implies that k=4, resulting in x 7 +1=x 4 +x 3<br />

+x 2 +1x 3 +x 2 +1 and Gx=x 3 +x 2 +1. After the first two<br />

steps in algorithm 1, we construct eight cosets with sixteen<br />

codewords in each coset (see Table I). Suppose that 101<br />

denotes the secret and 0001110 denotes the original set <strong>of</strong><br />

LSBs. We use the coset L 5 to find the codeword 1001110 that<br />

has the minimum Hamming distance equal to one from<br />

0001110 while altering only one LSB to embed three secret<br />

bits. The embedding rate and embedding efficiency are<br />

ER=3/7=42.9% and EE=3/0.875=343%, respectively.<br />

Note that l alt =0.875 <strong>for</strong> Table I. It is easy to see that the plain<br />

LSB embedding scheme ER=100% ,EE=200% has larger<br />

embedded capacity than the coded LSB scheme. On the<br />

other hand, the coded LSB scheme modifies fewer bits, thus<br />

reducing the distortion in the cover image.<br />

However, the coded LSB scheme suffers from the error<br />

spreading problem. The occurrence <strong>of</strong> one error bit in the<br />

encoded LSBs may cause more than one bit error during the<br />

secret message extraction process. Considering the above example,<br />

we can use 0000000 to carry the secret 000. Suppose<br />

that there is one error bit, <strong>for</strong> example, 0010000. Then, according<br />

to Table I, the extracted secret is 111, i.e., there are<br />

three error bits. However, employing the error pattern<br />

0000001 results in the extracted secret 001 which corresponds<br />

to only one error bit (no error spreading). This sug-<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 381


Yang et al.: Error spreading control in image steganographic embedding schemes using unequal error protection<br />

Table II. Eight cosets <strong>for</strong> the EC-based coded LSB scheme Zhang-Wang Scheme with l=3,N E =11 and 1-error correcting capability using 7,4<br />

Hamming code and 11,7 shorten Hamming code.<br />

L 0<br />

L 1<br />

L 2<br />

L 3<br />

L 4<br />

L 5<br />

L 6<br />

L 7<br />

00000000000, 10010001101, 10110011010, 00100010111, 11110110100, 01100111001<br />

01000101110, 11010100011, 01111101000, 11101100101, 11001110010, 01011111111<br />

10001011100, 00011010001, 00111000110, 10101001011<br />

11100000001, 01110001100, 01010011011, 11000010110, 00010110101, 10000111000<br />

10100101111, 00110100010, 10011101001, 00001100100, 00101110011, 10111111110<br />

01101011101, 11111010000, 11011000111, 01001001010<br />

01010000010, 11000001111, 11100011000, 01110010101, 10100110110, 00110111011<br />

00010101100, 10000100001, 00101101010, 10111100111, 10011110000, 00001111101<br />

11011011110, 01001010011, 01101000100, 11111001001<br />

10110000011, 00100001110, 00000011001, 10010010100, 01000110111, 11010111010<br />

11110101101, 01100100000, 11001101011, 01011100110, 01111110001, 11101111100<br />

00111011111, 10101010010, 10001000101, 00011001000<br />

10100000100, 00110001001, 00010011110, 10000010011, 01010110000, 11000111101<br />

11100101010, 01110100111, 11011101100, 01001100001, 01101110110, 11111111011<br />

00101011000, 10111010101, 10011000010, 00001001111<br />

01000000101, 11010001000, 11110011111, 01100010010, 10110110001, 00100111100<br />

00000101011, 10010100110, 00111101101, 10101100000, 10001110111, 00011111010<br />

11001011001, 01011010100, 01111000011, 11101001110<br />

11110000110, 01100001011, 01000011100, 11010010001, 00000110010, 10010111111<br />

10110101000, 00100100101, 10001101110, 00011100011, 00111110100, 10101111001<br />

01111011010, 11101010111, 11001000000, 01011001101<br />

00010000111, 10000001010, 10100011101, 00110010000, 11100110011, 01110111110<br />

01010101001, 11000100100, 01101101111, 11111100010, 11011110101, 01001111000<br />

10011011011, 00001010110, 00101000001, 10111001100<br />

gests that if the error falls in the right three positions, i.e.,<br />

0000100, 0000010, and 0000001, then there is still only one<br />

error in the extracted secret and the damage is not expanded.<br />

However, <strong>for</strong> the other four error patterns 1000000,<br />

0100000, 0010000, and 0001000, the decoded secret is 110,<br />

011, 111, and 101, respectively, indicating that the extracted<br />

secret suffers from more than one error bit. Since such errors<br />

affect the quality <strong>of</strong> the extracted secret message, the scheme<br />

should be improved by using the error correcting mechanism,<br />

such as one described below based on EC codes.<br />

EC CODES BASED ERROR CORRECTION:<br />

ZHANG-WANG SCHEME<br />

Following the previous approach, a coded LSB scheme with<br />

ER=l/n is constructed using the generating function<br />

Gx=G 2 x. Then, 2 k n-tuple vectors in each coset are encoded<br />

into an N E ,n cyclic EC code with N E denoting the<br />

code length <strong>of</strong> EC codes. Although in this improved coded<br />

LSB scheme the embedding rate is reduced to ER=l/N E , the<br />

scheme now has the error correcting capability <strong>of</strong> N E ,n<br />

cyclic EC codes.<br />

Let us consider the previous example with one-error<br />

correcting capability and parameters n=7, l=3 and N E =11.<br />

A 7,4 Hamming code is used to embed three secret bits<br />

and a 11,7 shorten Hamming code is used to achieve one<br />

error correcting capability. For example, embedding the secret<br />

000 into the 7-tuple 0001101 first and then appending<br />

the parity 1001 <strong>for</strong>ms the codeword 10010001101 which is<br />

listed as the second codeword in the coset L 0 (see Table II<br />

listing all eight generated cosets). If one error occurs in the<br />

sixth position, resulting in the codeword 10010101101, then<br />

the minimum Hamming distance is associated with the<br />

codeword 10010001101 from L 0 and the extracted secret is<br />

000. Because <strong>of</strong> the error correcting capability <strong>of</strong> the 11,7<br />

shorten Hamming code, the error is always corrected no<br />

matter where the error bit occurs, thus overcoming the error<br />

spreading problem. On the other hand, the approach is less<br />

efficient than the conventional method, as it reduces the<br />

embedding rate from 42.9% to 3/11=27.3% and also decreases<br />

the embedding efficiency from 343% to<br />

3/2.625=114% (note that l alt =2.625 <strong>for</strong> Table II). There<strong>for</strong>e,<br />

the different correction mechanism is needed. Since<br />

only the error in the first k bits <strong>of</strong> an n-tuple will produce<br />

additional errors in the secret extraction phase, it should be<br />

sufficient to ensure the validity <strong>of</strong> the first k bits instead <strong>of</strong><br />

all n bits.<br />

UEP CODES BASED ERROR CORRECTION: THE<br />

PROPOSED SCHEME<br />

UEP codes, a category <strong>of</strong> EC codes, allow different protection<br />

<strong>for</strong> different bit locations. 3,4 In practice, some in<strong>for</strong>mation<br />

bits are protected against a greater number <strong>of</strong> errors<br />

than other, less significant, in<strong>for</strong>mation bits. Basically, a UEP<br />

code can be denoted as n,k,d 1 ,d 2 ,...,d k . By employing<br />

UEP codes to protect the message, the occurrence <strong>of</strong> no<br />

more than ⌊d i −1/2⌋ errors in the transmitted codeword<br />

does not affect the correctness <strong>of</strong> the ith bit in the decoded<br />

message.<br />

It was noted that the first k bits <strong>of</strong> vectors need an<br />

enhanced protection to prevent the error spreading. There<strong>for</strong>e,<br />

we propose to apply UEP codes to assure the correctness<br />

<strong>of</strong> these k bits and reduce the number <strong>of</strong> redundant<br />

parity bits. The main difference between the UEP-based<br />

scheme and the EC-based scheme relates to the use <strong>of</strong><br />

382 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Yang et al.: Error spreading control in image steganographic embedding schemes using unequal error protection<br />

Table III. Eight cosets <strong>for</strong> the UEP-based coded LSB scheme the proposed scheme with l=3, N U =11 using 10,7,3333222 UEP code.<br />

L 0<br />

L 1<br />

L 2<br />

L 3<br />

L 4<br />

L 5<br />

L 6<br />

L 7<br />

0000000000*11000011011000011010010001011110001101000100111001<br />

000010111011001000110001101000110110010110011100100101111111<br />

1001011100010101000100010001101101001011<br />

0010000001*11100011001010011011011001011010101101010110111000<br />

001010111111101000100011101001111110010010111100110111111110<br />

1011011101011101000000110001111111001010<br />

010000001010000011111100011000000001010111001101100000111011<br />

010010110010001000010101101010100110011111011100000001111101<br />

1101011110000101001101010001001001001001<br />

011000001110100011101110011001001001010011101101110010111010<br />

011010110110101000000111101011101110011011111100010011111100<br />

1111011111001101001001110001011011001000<br />

1000000100*01000010010000011110110001001100001100001100111101<br />

100010101001001001111001101100010110000100011101101101111011<br />

0001011000110101010110010000100101001111<br />

101000010101100010000010011111111001001000101100011110111100<br />

101010101101101001101011101101011110000000111101111111111010<br />

0011011001111101010010110000110111001110<br />

110000011000000010110100011100100001000101001100101000111111<br />

110010100000001001011101101110000110001101011101001001111001<br />

0101011010100101011111010000000001001101<br />

111000011100100010100110011101101001000001101100111010111110<br />

111010100100101001001111101111001110001001111101011011111000<br />

0111011011101101011011110000010011001100<br />

N U ,n,d 1 ,d 2 ,...,d k UEP codes with N U denoting the<br />

code length <strong>of</strong> UEP codes instead <strong>of</strong> N E ,n EC codes. Since<br />

the value <strong>of</strong> N U is smaller than N E , the proposed UEP-based<br />

coded LSB scheme will have the higher embedding rate<br />

while still providing the same protection <strong>of</strong> the secret to the<br />

error spreading.<br />

As be<strong>for</strong>e, let us consider the scenario with one error<br />

correcting capability and parameters n=7, l=3, and<br />

N U =10. Suppose that the 10,7,3333222 UEP code is<br />

used to ensure the protection against errors. The corresponding<br />

eight cosets are listed in Table III. Assuming that<br />

the embedded secret and the encoded result are, <strong>for</strong> instance,<br />

000 and 0000000000, respectively, one error can occur in the<br />

following cases:<br />

• The presence <strong>of</strong> the error in the 7th bit (from right)<br />

implies the codeword 0001000000. As shown in Table<br />

III, there is only one codeword 0000000000 in the coset<br />

L 0 with the unit Hamming distance to 0001000000. In<br />

this case, the recovered secret is 000, i.e., no processing<br />

error.<br />

• If the error affects the third bit (from right) beyond the<br />

error correcting capability <strong>of</strong> the considered UEP code,<br />

then 0000000000 in the coset L 0 and 1000000100 in the<br />

coset L 4 are the two codewords with the unit Hamming<br />

distance to the codeword 0000000100 under consideration.<br />

The decoding process can result in the extracted<br />

secret 000 or 100, respectively. Thus, even in the latter<br />

situation (i.e., 100), there is still only one error, suggesting<br />

no error spreading.<br />

• Finally, the alteration <strong>of</strong> the 8th bit (from right) due to<br />

the error implies 0010000000 which will be decoded as<br />

0000000000 in the coset L 0 or 0010000001 in the coset<br />

L 1 . This suggests that the secret can be extracted as 000<br />

or 001, respectively. Similar to the previous case, even<br />

when 001 is used as the extracted secret, the proposed<br />

method still overcomes the error spreading problem.<br />

It is evident that when no more than one error falls in<br />

the first four bits <strong>of</strong> the original 7-bit vector, the use <strong>of</strong> UEP<br />

codes ensures that the first four bits will be correctly decoded<br />

and the error will be corrected, as ⌊3−1/2⌋=1, i.e.,<br />

the Hamming distance between the first four bits <strong>of</strong> two<br />

codewords is three providing one error correcting capability<br />

<strong>for</strong> the first four bits. However, no error spreading is also<br />

observed in the situations when one error occurs in other<br />

places. This is due to the fact that a single error in other<br />

places will result, in the worst case, in a single error in the<br />

decoded secret bits. The achieved embedding rate and embedding<br />

efficiency are ER=l/N U =30% and EE=l/l alt<br />

=133%. Note that l alt =2.25 <strong>for</strong> Table III.<br />

By employing the familiar representation used in UEP<br />

codes, the 11,7 shorten Hamming code can be represented<br />

as 11,7,3333333. There<strong>for</strong>e, if we protect the first four<br />

bits only, then we can save one redundant checking bit by<br />

using the 11,7,3333222 UEP code. Note that the errorcorrecting<br />

capability <strong>of</strong> N U ,n,d 1 ,d 2 ,...,d k UEP codes is<br />

not better compared to N E ,n,d EC codes. However, UEP<br />

codes have better embedding rate and embedding efficiency<br />

and also overcome the error spreading problem.<br />

COMPARISON AND EXPERIMENTAL RESULTS<br />

Different analytical tools, such as the sample pair analysis 5<br />

and image quality metrics, 6 are used to analyze the<br />

steganographic solutions. To resist the various attacks on the<br />

stegoimage while still providing the required per<strong>for</strong>mance,<br />

an ideal steganographic scheme should be constructed by<br />

considering the relation between the cover image and the<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 383


Yang et al.: Error spreading control in image steganographic embedding schemes using unequal error protection<br />

Table IV. Comparison <strong>of</strong> the conventional, EC-based, and UEP-based coded LSB schemes <strong>for</strong> n=2 to 12.<br />

Conventional scheme Zhang-Wang EC-based scheme Proposed UEP-based scheme<br />

n,k,l ER=l/n N E ,n,d ER=l/N E N U ,n,d 1 ,d 2 ,...,d k ER=l/N U<br />

2,1,1 50.0% 5,2,3 20.0% 4,2,32 25.0%<br />

4,1,3 75.0% 7,4,3 42.9% 6,4,3222 50.0%<br />

5,4,1 20.0% 9,5,4 11.1% 8,5,33332 12.5%<br />

6,4,2 33.3% 10,6,4 20.0% 9,6,333322 22.2%<br />

7,4,3 42.9% 11,7,4 27.3% 10,7,3333222 30.0%<br />

8,4,4 50% 12,8,4 33.3% 11,8,33332222 36.4%<br />

9,4,5 55.6% 13,9,4 38.5% 12,9,333322222 41.7%<br />

10,4,6 60% 14,10,4 42.9% 13,10,3333222222 46.2%<br />

11,4,7 63.6% 15,11,4 46.7% 14,11,33332222222 50.0%<br />

12,4,8 66.7% 17,12,5 47.1% 15,12,333322222222 53.3%<br />

stegoimage (or the relations between the stegopixel and the<br />

original pixel) and simultaneously the scheme should allow<br />

to achieve the high embedding rates. There<strong>for</strong>e, the schemes<br />

under consideration are evaluated here in terms <strong>of</strong> the embedding<br />

rate, embedding efficiency, and the peak-signal-tonoise<br />

(PSNR) ratio calculated using the original cover image<br />

and its stegoversion.<br />

Table IV shows the embedding rates and codes used in<br />

the conventional scheme, EC-based scheme and UEP-based<br />

scheme <strong>for</strong> n=2 to 12. As it can be seen from the listed<br />

results, all UEP-based schemes have the shorter code length<br />

than the EC-based schemes. Both these schemes have the<br />

ability to correct the first k bits when no larger than one<br />

error occurs, and avoid the error spreading problem. Table V<br />

shows the detail comparison <strong>for</strong> these three schemes with<br />

n,k,l=7,4,3. Code-based schemes address the error<br />

spreading problem at the cost <strong>of</strong> their smaller ER and EE.<br />

The UEP-based scheme, with ER=30.0% and EE=133%,<br />

needs 3334 pixels in a cover image to embed 1000<br />

=333430% secret bits while altering 750 LSBs<br />

=33342.25/12 within these embedded pixels; and, as<br />

can be seen, it outper<strong>for</strong>ms the EC-based scheme.<br />

In order to compare the distortion caused by the<br />

schemes under consideration, the well-known 259259 test<br />

gray-scale images “Baboon”, “Barb”, “Boat”, “Elaine”,<br />

“Mena”, and “Peppers” have been used as the cover images.<br />

The secret NDHU (National Dong Hwa University) “logo”<br />

and “text” gray-scale images to be embedded are shown in<br />

Figure 1. Secret images with size 5959 left, 4747 middle, and<br />

4949 pixels right.<br />

Table V. ER and EE <strong>for</strong> the conventional, EC-based, and UEP-based coded LSB schemes<br />

with n,k,l=7,4,3.<br />

Coded LSB scheme ER EE Embedding <strong>of</strong> 1000 bits<br />

Number <strong>of</strong> pixels needed Altered LSBs<br />

Conventional 42.9% 342% 2334 292<br />

EC-based 27.3% 114% 3667 877<br />

UEP-based 30.0% 133% 3334 750<br />

Table VI. PSNRdB between the cover image and its stegoimage <strong>for</strong> the conventional,<br />

EC-based, and UEP-based schemes.<br />

Coded LSB scheme Conventional EC-based UEP-based<br />

Cover image<br />

Baboon 57.173 54.671 54.691<br />

Barb 57.148 54.687 54.675<br />

Boat 57.181 54.696 54.677<br />

Elaine 57.143 54.675 54.681<br />

Mena 57.151 54.683 54.710<br />

Peppers 57.179 54.693 54.671<br />

Figure 1. Note that due to the different embedding rates <strong>for</strong><br />

different schemes, the secret images <strong>of</strong> 5959, 4747, and<br />

4949 pixels <strong>for</strong> the conventional, EC-based, and UEPbased<br />

schemes, respectively, have been used to ensure fair<br />

comparisons. The achieved PSNR values are listed in Table<br />

VI. The results indicate that the considered schemes produce<br />

high-quality stegoimages and the highest PSNR was<br />

achieved by the conventional scheme due to the higher EE<br />

(343% versus 114% in the EC-based scheme and 133% in<br />

the UEP-based scheme). This suggests that adding the error<br />

correcting capability does not distort the stegoimage seriously<br />

and that employing the error correcting codes (EC or<br />

UEP) in the coded LSB embedding scheme constitutes a<br />

reasonable and practical solution to overcome the error<br />

spreading problem.<br />

384 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Yang et al.: Error spreading control in image steganographic embedding schemes using unequal error protection<br />

Figure 2. Recovered secret images with BER=2%, 4%, and 8% <strong>for</strong> errors<br />

placed in random positions.<br />

To further study the error spreading problem, the two<br />

types <strong>of</strong> error patterns, namely random errors and worst<br />

errors have been added into the LSBs <strong>of</strong> stegoimages. The<br />

first type (random errors) means that the errors are randomly<br />

distributed in n-bit vector. However, the second type<br />

(worst errors) means that the errors occur in the worst positions<br />

(the first k bits <strong>of</strong> the original n-bit vector) where will<br />

cause error spreading. Figures 2 and 3 show the corresponding<br />

results obtained by extracting the embedded secret images<br />

from the noise corrupted stegoimages. Visual inspection<br />

<strong>of</strong> the results reveals that in a noisy environment the<br />

schemes based on UEP and EC codes have comparable per<strong>for</strong>mance<br />

and clearly outper<strong>for</strong>m the conventional coded<br />

LSB scheme. Moreover, since the proposed UEP-based<br />

scheme has higher ER than the EC-based scheme, it can be<br />

concluded that our solution provides a trade<strong>of</strong>f between the<br />

data embedding per<strong>for</strong>mance and the protection <strong>of</strong> the embedded<br />

secret.<br />

CONCLUSIONS<br />

A refined steganographic solution was introduced. Using<br />

UEP codes, we overcame the error spreading problem in the<br />

coded LSB steganographic scheme originally proposed by<br />

van Dijk and Willems. Our solution has the same correction<br />

effect as the Zhang-Wang EC-based scheme while allowing<br />

<strong>for</strong> lower embedding rates. This suggests that the solution<br />

Figure 3. Recovered secret images with BER=2%, 4%, and 8% <strong>for</strong> errors<br />

placed in the worst positions.<br />

proposed in this paper embeds the same secret message with<br />

the higher efficiency and produces less distortion in the generated<br />

stegoimage. The proposed solution is suitable <strong>for</strong> applications,<br />

such as transmission <strong>of</strong> the private digital materials<br />

(e.g., documents or signature images) through public<br />

and wireless networks, where data hiding and protection<br />

against communication errors are required or recommended.<br />

ACKNOWLEDGMENT<br />

This work was supported in part by TWISC@NCKU, National<br />

<strong>Science</strong> Council under the Grants NSC 94-3114-P-<br />

006-001-Y.<br />

REFERENCES<br />

1 M. van Dijk and F. Willems, “Embedding in<strong>for</strong>mation in grayscale<br />

images”, Proc. 22nd Symp. In<strong>for</strong>m. Theory in the Benelux (Elsevier,<br />

Netherlands, 2001), pp. 147–154.<br />

2 X. Zhang and S. Wang, “Stego-encoding with error correction<br />

capability”, IEICE Trans. Fundamentals E88-A, 3663–3667 (2005).<br />

3 W. J. van Gils, “Two topics on linear unequal error protection codes:<br />

bounds on their length and cyclic code classes”, IEEE Trans. Inf. Theory<br />

29, 866–876 (1983).<br />

4 M. C. Lin, C. C. Lin, and S. Lin, “Computer search <strong>for</strong> binary cyclic UEP<br />

codes <strong>of</strong> odd length up to 65”, IEEE Trans. Inf. Theory 36, 924–935<br />

(1990).<br />

5 S. Dumitrescu, X. Wu, and Z. Wang, “Detection <strong>of</strong> LSB steganography<br />

via sample pair analysis”, IEEE Trans. Signal Process. 51, 1995–2007<br />

(2003).<br />

6 I. Avcibas, N. Memon, and B. Sankur, “Steganalysis using image quality<br />

metrics”, IEEE Trans. Image Process. 12, 221–229 (2003).<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 385


<strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology® 51(4): 386–390, 2007.<br />

© <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 2007<br />

In Situ X-ray Investigation <strong>of</strong> the Formation <strong>of</strong> Metallic<br />

Silver Phases During the Thermal Decomposition <strong>of</strong> Silver<br />

Behenate and Thermal Development <strong>of</strong><br />

Photothermographic Films<br />

B. B. Bokhonov, M. R. Sharafutdinov and B. P. Tolochko<br />

Institute <strong>of</strong> Solid State Chemistry and Mechanochemistry, Russian Academy <strong>of</strong> <strong>Science</strong>, Kutateladze 18,<br />

Novosibirsk 630128, Russia<br />

E-mail: bokhonov@solid.nsk.su<br />

L. P. Burleva and D. R. Whitcomb <br />

Health Group, Eastman Kodak Company, Oakdale, Minnesota 55128<br />

Abstract. Metallic silver <strong>for</strong>mation, resulting from the thermal decomposition<br />

<strong>of</strong> silver behenate, AgBe, and from the thermally induced<br />

reduction <strong>of</strong> AgBe incorporated into a photothermographic<br />

imaging construction, has been compared by in situ x-ray investigation.<br />

In the case <strong>of</strong> the thermal decomposition <strong>of</strong> individual AgBe<br />

crystals, the main factor that determines the growth <strong>of</strong> the silver<br />

particles is the change in the AgBe crystal structure, leading to the<br />

<strong>for</strong>mation <strong>of</strong> intermediate mesomorphic phases that still retain characteristic<br />

layer structure. By contrast, development <strong>of</strong> AgBecontaining<br />

photothermographic films generates silver particles by<br />

the reduction <strong>of</strong> intermediate silver complexes, which are in a liquid<br />

state during the development process. The silver nanoparticles resulting<br />

from these processes exhibit different sizes and morphologies<br />

that are important <strong>for</strong> optimizing the optical properties <strong>of</strong><br />

photothermographic films. © 2007 <strong>Society</strong> <strong>for</strong> <strong>Imaging</strong> <strong>Science</strong> and<br />

Technology.<br />

DOI: 10.2352/J.<strong>Imaging</strong>Sci.Technol.200751:4386<br />

INTRODUCTION<br />

Silver behenate, AgO 2 C 22 H 43 2 , is one <strong>of</strong> the fundamental<br />

components <strong>of</strong> photothermographic materials because it<br />

provides the silver ions <strong>for</strong> reduction in the thermal development<br />

process that leads to the <strong>for</strong>mation <strong>of</strong> a visible<br />

image. 1–4 In the literature, there are a large number <strong>of</strong> reports<br />

devoted to the investigation <strong>of</strong> the phase changes <strong>of</strong><br />

long, saturated-chain silver carboxylates, including silver<br />

behenate, in thermal systems as well as the effect <strong>of</strong> individual<br />

components added to “dry silver” photothermographic<br />

<strong>for</strong>mulations. 5–11 The x-ray investigation <strong>of</strong><br />

silver carboxylates with carbon atoms from 2 to 22 12 showed<br />

that all <strong>of</strong> these crystal structures fall into the triclinic class<br />

and contain two molecules in the unit cell. Among the<br />

dominant characteristics <strong>of</strong> the silver carboxylate crystal<br />

structures, which are defined by their significant anisotropic<br />

physical and chemical properties, 1,13 is the presence <strong>of</strong> a lay-<br />

<br />

IS&T Member<br />

Received Jan. 17, 2007; accepted <strong>for</strong> publication Mar. 22, 2007.<br />

1062-3701/2007/514/386/5/$20.00.<br />

ered structure in which a double layer <strong>of</strong> silver ions separates<br />

a double layer <strong>of</strong> long methylene chains. For example, the<br />

solid-state crystal structure <strong>of</strong> silver stearate (AgSt,<br />

AgO 2 C 22 H 43 2 ) shows that the molecules are actually<br />

dimers connected together <strong>for</strong>ming a polymer. 3<br />

Thermally induced phase changes in the silver carboxylate<br />

crystals have been investigated by various analytical<br />

methods, such as NMR, IR, conductivity, DSC, and XRD.<br />

The temperatures <strong>of</strong> the multiple-phase transitions <strong>for</strong> silver<br />

carboxylates having various chain lengths have been<br />

characterized. 5,14–17 Upon transition from the crystalline<br />

state to the isotropic liquid, the silver carboxylates undergo<br />

up to six to seven phase changes <strong>of</strong> the following sequence:<br />

crystal state→curd→super curd SUC→sub-waxy SW<br />

→waxy W→super waxy SUW→sub-neat SN→neat<br />

N→isotropic liquid. 5,18 It may be relevant that the phase<br />

changes in the silver carboxylate from the crystalline state<br />

into the super curd (SUC) or sub-waxy (SW) phase occur in<br />

the 120–125°C range, the temperature at which the thermal<br />

development in photothermography is normally carried out.<br />

X-ray diffraction, calorimetric, and IR methods were<br />

used in the investigation <strong>of</strong> the structural changes in the<br />

silver stearate crystal lattice. 5 It was shown that, upon heating,<br />

the silver stearate structure proceeded through a series<br />

<strong>of</strong> mesomorphic states. That is, the first phase transition<br />

occurred at 122°C, which was associated with a packing<br />

disorder <strong>of</strong> the aliphatic chains, manifested by a significant<br />

decrease in the separation between silver ion layers. It was<br />

proposed that increasing the temperature above 130°C leads<br />

to further disorder and the breakup <strong>of</strong> the silver ion layers,<br />

and it is responsible <strong>for</strong> the onset <strong>of</strong> the thermal decomposition<br />

reaction <strong>of</strong> the silver stearate, resulting in the <strong>for</strong>mation<br />

<strong>of</strong> metallic silver and paraffin byproducts.<br />

The structure trans<strong>for</strong>mation <strong>of</strong> polycrystalline silver<br />

behenate was also studied by x-ray diffraction during in situ<br />

heating. 15 In contrast to the results reported in Ref. 5, these<br />

authors 15 observed an increase in the interlayer spacing dur-<br />

386


Bokhonov et al.: In situ x-ray investigation <strong>of</strong> the <strong>for</strong>mation <strong>of</strong> metallic silver phases during the thermal decomposition <strong>of</strong> silver behenate...<br />

ing the heating <strong>of</strong> silver behenate crystals. The authors also<br />

indicated that heating silver behenate over 120°C irreversibly<br />

trans<strong>for</strong>ms it from a crystalline to an amorphous state.<br />

Further, at 138–142°C, the first phase changes are observed,<br />

established by the appearance <strong>of</strong> diffraction peaks at the<br />

smaller 2 Bragg angles, which correspond to an increase in<br />

interlayer distance in the silver behenate structure. In agreement<br />

with these results, upon heating above 145°C, the silver<br />

behenate crystals trans<strong>for</strong>m into a liquid-crystalline state<br />

and generate metallic silver phases at 180°C. The authors <strong>of</strong><br />

this report consider that the initial stage <strong>of</strong> heating is the<br />

disordering <strong>of</strong> the silver behenate aliphatic chains. However,<br />

despite the agreement in the explanation <strong>of</strong> the structure<br />

trans<strong>for</strong>mations occurring in the silver behenate 15 and silver<br />

stearate, 5 there is a significant difference in the explanation<br />

<strong>of</strong> the subsequent structure changes in the phase trans<strong>for</strong>mations<br />

<strong>of</strong> these silver carboxylates. While heating silver<br />

stearate decreases its interlayer spacing, 5 heating silver<br />

behenate crystals initially proceeds through an amorphous<br />

phase followed by an increased distance between the layers. 15<br />

Such a contradictory heating behavior between the silver<br />

stearate and behenate seems to be quite surprising because<br />

<strong>of</strong> the close similarities between the silver stearate and<br />

behenate structures (C 18 and C 22 chain length, respectively)<br />

and their phase-trans<strong>for</strong>mation temperatures. In addition,<br />

we have recently reported the thermal decomposition <strong>of</strong> silver<br />

myristate, AgMy, AgO 2 C 14 H 27 2 under conditions<br />

similar to the AgSt and noted similar behavior. 5 Considering<br />

the importance <strong>of</strong> AgBe as a material <strong>for</strong> photothermographic<br />

imaging products and the contradiction between<br />

the trend observed <strong>for</strong> thermal decomposition <strong>of</strong><br />

AgMy and AgSt (in solids and in photothermographic films)<br />

relative to the interlayer spacing differences reported <strong>for</strong> the<br />

AgBe, we have continued the systematic investigation <strong>of</strong> the<br />

effect <strong>of</strong> increasing the chain length on the thermal properties<br />

<strong>of</strong> the AgBe component in this series. Once the reasons<br />

<strong>for</strong> the <strong>for</strong>mation <strong>of</strong> the solid products from these chemical<br />

reactions are better understood, novel routes to achieve control<br />

<strong>of</strong> these processes should be possible, and<br />

photothermographic properties can be further improved. In<br />

this work, we show the results <strong>of</strong> our in situ x-ray diffraction<br />

investigation related to the <strong>for</strong>mation <strong>of</strong> metallic silver from<br />

the thermal decomposition <strong>of</strong> pure silver behenate, as well as<br />

from the thermal development <strong>of</strong> photothermographic materials<br />

based on silver behenate.<br />

EXPERIMENTAL<br />

The synthesis <strong>of</strong> silver behenate was carried out by the exchange<br />

reaction between sodium behenate and silver nitrate,<br />

as typically practiced. 5 Photothermographic films were prepared<br />

from pure AgBe (not a mixture <strong>of</strong> chain lengths as is<br />

common in photothermography) and pre<strong>for</strong>med AgBr,<br />

along with the normal additional imaging components, as<br />

described elsewhere. 19<br />

X-ray experiments were carried out at the time-resolved<br />

diffractometry station—channel 5b <strong>of</strong> VEPP-3, BINP<br />

=1.506 Å. Transmission mode was used <strong>for</strong> small-angle<br />

Figure 1. Phthalazine and 4-methylphthalic acid.<br />

scattering (SAXS). X-ray patterns were obtained on the onecoordinate<br />

detector OD-3 with a 0.01° angular resolution<br />

and a 30 s recording time per frame. Samples were heated at<br />

1°C/min in a special tube furnace, and sample temperatures<br />

were controlled by a thermocouple.<br />

RESULTS AND DISCUSSION<br />

Despite the fact that the thermal decomposition <strong>of</strong> silver<br />

carboxylates and the development <strong>of</strong> photothermographic<br />

films both produce solid products <strong>of</strong> metallic silver, the<br />

chemical trans<strong>for</strong>mations occurring within these processes<br />

are completely different. If the thermal decomposition <strong>of</strong><br />

silver carboxylate proceeds according to the following<br />

scheme: 7<br />

AgO 2 C n H 2n−1 2 → 2Ag + 2CO 2 +C 2n−1 H 22n−1<br />

with the <strong>for</strong>mation <strong>of</strong> metallic silver and paraffin, then the<br />

development stages <strong>of</strong> photothermographic films are more<br />

complicated. Thermally induced reduction <strong>of</strong> the silver ions<br />

during film development can be summarized in a more simplified<br />

<strong>for</strong>m as follows:<br />

AgO 2 C n H 2n−1 2 → silver ion intermediates<br />

→ 2Ag + 2HO 2 C n H 2n−1 .<br />

This reduction reaction is the result <strong>of</strong> preliminary exposure<br />

<strong>of</strong> the silver halide in the film, which <strong>for</strong>ms active latent<br />

image centers that catalyze the thermal development step at<br />

110–130°C. The silver ion intermediates are reduced at the<br />

latent image centers, resulting in crystalline silver particles,<br />

which comprise the visible image. It is generally accepted<br />

that the silver ion source <strong>for</strong> silver particle <strong>for</strong>mation at the<br />

latent image center is the silver carboxylate, 1–4 which is not<br />

light sensitive in the visible region <strong>of</strong> the spectrum, and that<br />

the transport <strong>of</strong> silver ions from the silver carboxylate to the<br />

latent image center is from the thermally initiated <strong>for</strong>mation<br />

<strong>of</strong> the various silver complexes obtained with the components<br />

added into the <strong>for</strong>mulation (developers, toners, and<br />

antifoggants). 1–6<br />

A recent investigation into the composition <strong>of</strong> phase<br />

changes occurring during the development process raised<br />

doubts about the source <strong>of</strong> the silver ions <strong>for</strong>ming the visible<br />

image coming only from the silver carboxylate phase. 20 Contrary<br />

to all other literature, 1–4 the reduction <strong>of</strong> a model system<br />

composition (AgBe/AgBr with phthalazine (PHZ,<br />

Figure 1) 4-methyl-phthalic acid (4MPA, Fig. 1), and developer)<br />

resulted in a significant decrease 45% in the x-ray<br />

peak intensity <strong>for</strong> the AgBr phase. This change was proposed<br />

1<br />

2<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 387


Bokhonov et al.: In situ x-ray investigation <strong>of</strong> the <strong>for</strong>mation <strong>of</strong> metallic silver phases during the thermal decomposition <strong>of</strong> silver behenate...<br />

to be related to the contribution <strong>of</strong> silver ions from the AgBr<br />

in the <strong>for</strong>mation <strong>of</strong> the metallic silver particles <strong>of</strong> the image.<br />

As discussed below, in the full photothermographic imaging<br />

<strong>for</strong>mulation, we see no change in the AgBr signal.<br />

The in situ investigation <strong>of</strong> the structural and phase<br />

changes in the thermal decomposition <strong>of</strong> silver behenate and<br />

the development <strong>of</strong> the photothermographic films prepared<br />

from it showed that the processes accompanying the thermal<br />

<strong>for</strong>mation <strong>of</strong> the silver particles are significantly<br />

different.<br />

In Situ Investigation <strong>of</strong> the Phase Formed in the Process<br />

<strong>of</strong> Thermal Decomposition <strong>of</strong> AgBe<br />

The change in the diffraction characteristics <strong>of</strong> AgBe in the<br />

small angle region 2=0.4–10° during in situ heating<br />

from 20–220°C is shown in Figure 2(a).<br />

Increasing the temperature through this range is accompanied<br />

by a change in the AgBe x-ray diffraction pattern that<br />

is due to phase trans<strong>for</strong>mations occurring within the heated<br />

powder. As the temperature is increased, the reflections <strong>of</strong><br />

the high temperature phases shift to the high diffraction<br />

angle regions, which confirm the decrease in the interlayer<br />

spacing in the silver carboxylate structure. Analysis <strong>of</strong> the<br />

diffraction data <strong>of</strong> these phases in this temperature interval<br />

allows us to separate out at least six phases that have different<br />

structural characteristics, Fig. 2(b).<br />

It should be noted that the similar structural changes in<br />

AgBe and AgMy, with the <strong>for</strong>mation <strong>of</strong> intermediate phases,<br />

were established previously. 5,6,21 According to the authors, 5<br />

the first three phases correspond to the transitions within<br />

the AgBe crystalline state, but then at 155°C, silver behenate<br />

trans<strong>for</strong>ms into a liquid-crystalline material.<br />

The x-ray diffraction patterns <strong>of</strong> the intermediate<br />

phases <strong>for</strong>med during heating, Fig. 2(b), include at least two<br />

series <strong>of</strong> layer reflections, which are evidence <strong>for</strong> the <strong>for</strong>mation<br />

<strong>of</strong> a two-dimensional structure. Increasing the temperature<br />

above 230°C leads to the disappearance <strong>of</strong> the diffraction<br />

image <strong>of</strong> a layered structure. At the same time in the<br />

small 2 angle region, an increase in SAXS intensity is observed,<br />

the maximum <strong>of</strong> which is at 1.17°, Fig. 2(c).<br />

Subsequent heating to 250°C corresponds to a change<br />

in the shape <strong>of</strong> the small angle scattering peak, which appears<br />

as a decrease in the intensity <strong>of</strong> the SAXS maximum.<br />

Simultaneously, peaks appear in the small angle scattering<br />

angles at 20.8°.<br />

The in situ x-ray diffraction <strong>of</strong> the thermal decomposition<br />

<strong>of</strong> AgBe in the wide-angle region is (WAXS,<br />

2=25–55°). Figure 3 showed that heating the powder to<br />

140°C does not appear to significantly change the diffraction<br />

pattern. Upon heating to higher than 145°C, the<br />

crystal-phase reflections <strong>of</strong> AgBe disappear, and beginning at<br />

230°C broad reflections are observed because <strong>of</strong> the (111)<br />

and (200) planes <strong>of</strong> metallic silver, Fig. 3, the intensity <strong>of</strong><br />

which increases as the temperature is increased. Given that<br />

the first phase transition in silver behenate occurs at 128°C,<br />

the presence in the x-ray diffraction <strong>of</strong> the crystalline phase<br />

Figure 2. a Change in the x-ray diffraction pattern <strong>of</strong> silver behenate<br />

during in situ heating. b X-ray diffraction pattern <strong>for</strong> the initial 20°C<br />

and intermediate phases <strong>for</strong>med in the heating <strong>of</strong> AgBe. c SAXS <strong>of</strong><br />

AgBe.<br />

<strong>of</strong> AgBe is evidence that the first phase transition occurs<br />

from one crystalline state to another.<br />

388 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


Bokhonov et al.: In situ x-ray investigation <strong>of</strong> the <strong>for</strong>mation <strong>of</strong> metallic silver phases during the thermal decomposition <strong>of</strong> silver behenate...<br />

Figure 3. WAXS <strong>of</strong> AgBe during in situ thermal decomposition.<br />

Figure 5. Change in the x-ray diffraction pattern <strong>of</strong> photothermographic<br />

films during thermal development.<br />

Figure 4. Change in the x-ray diffraction pattern <strong>of</strong> AgBe during the<br />

development <strong>of</strong> photothermographic films: Initial decrease in the AgBe<br />

layer peak intensities with a simultaneous increase in the signal intensity <strong>of</strong><br />

the small-angle scattering peaks.<br />

In situ X-ray Diffraction Investigation <strong>of</strong> Phase<br />

Formation During Development <strong>of</strong> Photothermographic<br />

Films Prepared with AgBe<br />

The in situ x-ray investigation <strong>of</strong> the <strong>for</strong>mation <strong>of</strong> phases<br />

during the development <strong>of</strong> photothermographic films prepared<br />

with AgBe showed that the <strong>for</strong>mation <strong>of</strong> the silver<br />

phases occurs at temperatures significantly lower than the<br />

temperature <strong>of</strong> the first phase trans<strong>for</strong>mation 126°C.<br />

There are no shifts in the AgBe reflections observed during<br />

heating the photothermographic films from 20–80°C. After<br />

80°C, however, the intensity <strong>of</strong> the x-ray diffraction peaks<br />

related to the layered structure <strong>of</strong> AgBe (001) decrease,<br />

which corresponds to the simultaneous increase in the intensity<br />

in the small-angle scattering region <strong>of</strong> SAXS<br />

2=0.4–1.2°, Figure 4.<br />

The in situ x-ray diffraction pattern over 24–54° 2<br />

showed that the reflections that were due to the metallic<br />

silver appear as low as 80°C, Figure 5. Upon increasing the<br />

temperature (or the development time), the peak intensity <strong>of</strong><br />

the silver reflections increases.<br />

It is important to note that the in situ investigation <strong>of</strong><br />

the thermal development <strong>of</strong> films did not reveal any kind <strong>of</strong><br />

additional reflections from intermediate solid phases. More<br />

significantly, the reflection intensity <strong>for</strong> silver bromide (200)<br />

up to and after processing remained completely unchanged,<br />

Fig. 5.<br />

Comparing the half-widths <strong>of</strong> the silver reflections<br />

(111) and (200) recorded during the decomposition <strong>of</strong><br />

AgBe, Fig. 3, and the developed photothermographic films,<br />

Fig. 5, provide clear evidence that the size <strong>of</strong> the silver crystals<br />

in the developed films are significantly larger than that<br />

<strong>for</strong>med in the process <strong>of</strong> the thermal decomposition <strong>of</strong> pure<br />

AgBe, similar to that observed in AgMy. 6<br />

All <strong>of</strong> these results on the thermal decomposition <strong>of</strong><br />

AgBe and thermal development <strong>of</strong> photothermographic<br />

films can be summarized as follows.<br />

The <strong>for</strong>mation <strong>of</strong> metallic silver phases from the thermal<br />

decomposition <strong>of</strong> pure AgBe occurs through the <strong>for</strong>mation<br />

<strong>of</strong> a series <strong>of</strong> intermediate mesomorphic phases. The<br />

<strong>for</strong>mation <strong>of</strong> silver particles, established by the appearance in<br />

the x-ray diffraction pattern <strong>of</strong> in situ heated signal in the<br />

small-angle scattering, proceeds after the destruction <strong>of</strong> the<br />

AgBe layer structure.<br />

The development <strong>of</strong> the photothermographic films<br />

shows the <strong>for</strong>mation <strong>of</strong> silver phases at 80°C that correspond<br />

to the decreasing intensity <strong>of</strong> the silver behenate layer<br />

structure reflections with a simultaneous increase in the intensity<br />

<strong>of</strong> the SAXS. In particular, it must be emphasized<br />

that the <strong>for</strong>mation <strong>of</strong> the silver phase does not proceed<br />

through any change in the intensity <strong>of</strong> the silver halide<br />

peaks, and it is clear that the silver particles <strong>for</strong>m from the<br />

reduction <strong>of</strong> silver ions originating only from the AgBe crystals.<br />

J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007 389


Bokhonov et al.: In situ x-ray investigation <strong>of</strong> the <strong>for</strong>mation <strong>of</strong> metallic silver phases during the thermal decomposition <strong>of</strong> silver behenate...<br />

Overall, all <strong>of</strong> these results are in good agreement with<br />

previous reports. 22,23 That is, the initial stages <strong>of</strong> thermal<br />

decomposition <strong>of</strong> individual silver carboxylate <strong>for</strong>m nanosized<br />

2–5 nm particles <strong>of</strong> silver, which subsequently agglomerate<br />

up to 10–15 nm, crystallizing on the lateral<br />

planes <strong>of</strong> the silver carboxylate crystals. In our opinion, this<br />

stage <strong>of</strong> silver particle growth is the cause <strong>of</strong> the curve shape<br />

changes in the small-angle scattering in which a decrease in<br />

intensity and a shift to the small-angle region <strong>of</strong> the SAXS<br />

maxima was observed. Finally, it should be noted that the<br />

difference between the thermal behavior <strong>of</strong> pure AgBe described<br />

in here and Ref. 16 could be attributed to the differences<br />

in the preparation procedures. The same effect may<br />

influence the x-ray results during the study <strong>of</strong> the role <strong>of</strong><br />

AgBr in the photothermographic process. 15<br />

CONCLUSIONS<br />

The differences in the diffraction data during the development<br />

<strong>of</strong> photographic films and the thermal decomposition<br />

<strong>of</strong> pure AgBe are related to the differences in the chemical<br />

trans<strong>for</strong>mations in these processes: in contrast to the thermal<br />

decomposition <strong>of</strong> pure AgBe, development <strong>of</strong> the<br />

photothermographic films generates silver particles by the<br />

reduction <strong>of</strong> intermediate silver complexes, which are in the<br />

liquid state (not observable by x-ray diffraction). In the case<br />

<strong>of</strong> the thermal decomposition <strong>of</strong> individual AgBe crystals,<br />

the main factor that determines the growth <strong>of</strong> the silver<br />

particles is the change in the structure, leading to the <strong>for</strong>mation<br />

<strong>of</strong> intermediate mesomorphic phases, which still retain<br />

the characteristic layer structure.<br />

ACKNOWLEDGMENT<br />

The authors gratefully thank T. Blanton (Eastman Kodak<br />

Company) <strong>for</strong> helpful discussions.<br />

REFERENCES<br />

1 P. J. Cowdery-Corvan and D. R. Whitcomb, in Handbook <strong>of</strong> <strong>Imaging</strong><br />

<strong>Material</strong>s, edited by A. S. Diamond and D. S. Weiss (Marcel Dekker,<br />

New York, 2002), p. 473.<br />

2 D. H. Klosterboer, in Neblette’s Eighth Edition: <strong>Imaging</strong> Processes and<br />

<strong>Material</strong>s, edited by J. M. Sturge, V. Walworth, and A. Shepp (Van<br />

Nostrand-Reinhold, New York, 1989), Chap. 9, p. 279.<br />

3 V. M. Andreev, E. P. Fokin, Yu. I. Mikhailov, and V. V. Boldyrev, Zhur.<br />

Nauch. Priklad. Fotogr. Kinemat. 24, 311 (1979).<br />

4 T. Maekawa, M. Yoshikane, H. Fujimura, and I. Toya, J. <strong>Imaging</strong> Sci.<br />

Technol. 45, 365 (2001).<br />

5 B. B. Bokhonov, M. R. Sharafutdinov, B. P. Tolochko, L. P. Burleva, and<br />

D. R. Whitcomb, J. <strong>Imaging</strong> Sci. Technol. 49, 389 (2005).<br />

6 B. B. Bokhonov, L. P. Burleva, A. A. Sidelnikov, M. R. Sharafutdinov, B.<br />

P. Tolochko, and D. R. Whitcomb, J. <strong>Imaging</strong> Sci. Technol. 47, 89 (2003).<br />

7 V. M. Andreev, L. P. Burleva, and V. V. Boldyrev, J. Sib. Branch Acad. Sci.<br />

USSR 5(5), 3 (1984).<br />

8 D. R. Whitcomb and R. D. Rogers, J. <strong>Imaging</strong> Sci. Technol. 43, 517–520<br />

(1999).<br />

9 D. R. Whitcomb and M. Rajeswaran, J. <strong>Imaging</strong> Sci. Technol. 47, 107<br />

(2003).<br />

10 D. R. Whitcomb and R. D. Rogers, Inorg. Chim. Acta 256, 263 (1997).<br />

11 P. Z. Velinzon, S. I. Gaft, O. A. Karekina, N. K. Ryasinskaya, and I. G.<br />

Chezlov, Zhur. Nauch. i Priklad. Fotogr. 48(3), 35–45 (2003).<br />

12 V. Vand, A. Aitken, and R. K. Campbell, Acta Crystallogr. 2, 398–403<br />

(1949).<br />

13 A. E. Gvozdev, Ukr. Fiz. Zh. (Russ. Ed.) 24, 1856 (1979).<br />

14 M. Ikeda, Photograph. Sci. Eng. 24(6), 277 (1980).<br />

15 I. Geuens, I. Vanwelkenhuysen, and R. Gijbels, Proc. 2000 International<br />

Symposium on Silver Halide <strong>Imaging</strong> (IS&T, Springfield, VA, 2000) pp.<br />

203–233.<br />

16 K. Binnemans, R. V. Deun, B. Thijs, I. Vanwelkenhuysen, and I. Geuens,<br />

Chem. Mater. 16, 2021 (2004).<br />

17 X. Liu, S. Liu, J. Zhang, and W. Cao, Thermochim. Acta 440, 1 (2006).<br />

18 V. M. Andreev, L. P. Burleva, B. B. Bokhonov, and Y. I. Mikhailov, Izv.<br />

Sib. Otd. AN SSSR, Ser. Khim. Nauk. 2(4), 58 (1983).<br />

19 C. Zou, J. B. Philip, S. M. Shor, M. C. Skinner, and C. P. Zhou, US<br />

Patent No. 5,434,043 (1995).<br />

20 H. Strijckers, J. <strong>Imaging</strong> Sci. Technol. 47, 100 (2003).<br />

21 T. N. Blanton, S. Zdzieszynski, M. Nikolas, and S. Misture, Powder Diffr.<br />

48, 27 (2005).<br />

22 B. B. Bokhonov, L. P. Burleva, and D. R. Whitcomb, J. <strong>Imaging</strong> Sci.<br />

Technol. 43, 505 (1999).<br />

23 B. B. Bokhonov, L. P. Burleva, D. R. Whitcomb, and M. R. V. Sahyun,<br />

Microsc. Res. Tech. 42, 152 (1998).<br />

390 J. <strong>Imaging</strong> Sci. Technol. 514/Jul.-Aug. 2007


IS&T Corporate Members<br />

IS&T Corporate Members provide significant financial support, thereby assisting the <strong>Society</strong> in achieving its goals <strong>of</strong><br />

disseminating in<strong>for</strong>mation and providing pr<strong>of</strong>essional services to imaging scientists and engineers. In turn, the <strong>Society</strong><br />

provides a number <strong>of</strong> material benefits to its Corporate Members. For complete in<strong>for</strong>mation on the Corporate<br />

Membership program, contact IS&T at info@imaging.org.<br />

Sustaining Corporate Members<br />

Adobe Systems Inc.<br />

345 Park Avenue<br />

San Jose, CA 95110-2704<br />

Canon USA Inc.<br />

One Canon Plaza, Lake Success<br />

New York, NY 11042-1198<br />

Eastman Kodak Company<br />

343 State Street<br />

Rochester, NY 14650<br />

Hewlett-Packard Company<br />

1501 Page Mill Road<br />

Palo Alto, CA 94304<br />

Lexmark International, Inc.<br />

740 New Circle Road NW<br />

Lexington, KY 40511<br />

Xerox Corporation<br />

Wilson Center <strong>for</strong> Research and<br />

Technology<br />

800 Phillips Road<br />

Webster, NY 14580<br />

Supporting Corporate Members<br />

Fuji Photo Film Company, Ltd.<br />

210 Nakanuma, Minami-ashigara<br />

Kanagawa 250-0193 Japan<br />

Konica Minolta Holdings Inc.<br />

No. 1 Sakura-machi<br />

Hino-shi, Tokyo 191-8511 Japan<br />

TREK, Inc./TREK Japan KK<br />

11601 Maple Ridge Road<br />

Medina, NY 14103-0728<br />

Pitney Bowes<br />

35 Waterview Drive<br />

Shelton, CT 06484<br />

Donor Corporate Members<br />

ABBY USA S<strong>of</strong>tware House, Inc.<br />

47221 Fremont Blvd.<br />

Fremont, CA 94538<br />

Axis Communications AB<br />

Embdalavägen 14<br />

SE-223 69 Lund, Sweden<br />

Ball Packaging Europe GmbH<br />

Technical Center Bonn<br />

Friedrich-Woehler-Strasse 51<br />

D-53117 Bonn, Germany<br />

Cheran Digital <strong>Imaging</strong> & Consulting, Inc.<br />

798 Burnt Gin Road<br />

Gaffney, SC 29340<br />

Clariant Produkte GmbH<br />

Division Pigments & Additives<br />

65926 Frankfurt am Main Germany<br />

Felix Schoeller Jr. GmbH & Co. KG<br />

Postfach 3667<br />

D-49026 Osnabruck, Germany<br />

Ferrania SpA<br />

Viale Martiri Della Liberta’ 57<br />

Ferrania (Savona) I-17014<br />

GretagMacbeth<br />

Logo GmbH & Co. KG<br />

Westfälischer H<strong>of</strong> Garbrock 4<br />

48565 Steinfurt, Germany<br />

GR8moments Limited<br />

Units 15-16 Town Yard Industrial Estate<br />

Station Street<br />

Leek, Staf<strong>for</strong>dshire<br />

England, ST13 8BF<br />

Hallmark Cards, Inc.<br />

Chemistry R & D<br />

2501 McGee, #359<br />

Kansas City, MO 64141-6580<br />

ILFORD <strong>Imaging</strong> Switzerland GmbH<br />

Route de l’Ancienne Papeterie 1<br />

CH-1723 Marly, Switzerland<br />

MediaTek Inc.<br />

No. 1 Dusing Rd., 1<br />

Hsinchu 300 R.O.C, Taiwan<br />

Pantone, Inc.<br />

590 Commerce Blvd.<br />

Carlstadt, NJ 07072-3098<br />

Quality Engineering Associates (QEA), Inc.<br />

99 South Bed<strong>for</strong>d Street, #4<br />

Burlington, MA 01803<br />

The Ricoh Company, Ltd.<br />

16-1 Shinei-cho, Tsuzuki-ku<br />

Yokohama 224-0035 Japan<br />

Sharp Corporation<br />

492 Minosho-cho, Yamatokoriyama<br />

Nara 639-1186 Japan<br />

Sony Corporation/<br />

Sony Research Center<br />

6-7-35 Kita-shinagawa<br />

Shinagawa, Tokyo 141 Japan<br />

*as 7/1/07


<strong>Journal</strong> <strong>of</strong> the <strong>Imaging</strong> <strong>Society</strong> <strong>of</strong> Japan VOL.46 NO.3<br />

2007<br />

CONTENTS<br />

Original Papers<br />

Analysis <strong>of</strong> the Magnetic Force Acting on the Toner in the Black Image Area and White Image Area in the<br />

Magnetic Printer (2)N. KOKAJI ...1722<br />

A Model <strong>for</strong> Electrostatic Discharge in the Toner Layer during the Transfer Process<br />

M. MAEDA, K. NISHIWAKI, K. MAEKAWA and M. TAKEUCHI ...1788<br />

<strong>Imaging</strong> Today<br />

The Law <strong>of</strong> Environmental Standard, Safety Standard and Energy Saving<br />

Introduction H. YAMAZAKI, T. TAKEUCHI, K. NAGATO, K. MARUYAMA and K. SUZUKI ...18414<br />

The Notification Systems <strong>of</strong> New Chemical Substances in the World T. YAMAMOTO ...18515<br />

The Environmental Regulations in Japan H. SATO ...19222<br />

The Trend <strong>of</strong> European Product Environmental Legislation and Eco-labels<br />

R. IWANAGA, K. FUJISAWA and A. MATSUMOTO ...19929<br />

Practical Side <strong>of</strong> the Environmentally Conscious Technology <strong>for</strong> the Product<br />

T. BISAIJI, K. YASUDA, T. ARAI, K. SUZUKI, K. AKATANI and M. HASEGAWA ...20737<br />

Lectures in <strong>Science</strong><br />

Introduction <strong>of</strong> Optics (I)<br />

The Behavior <strong>of</strong> a Beam <strong>of</strong> Light upon Reflection or Refraction at a Plane Surface <br />

H. MUROTANI ....21646<br />

Meeting Reports 22353<br />

Announcements 22454<br />

Guide <strong>for</strong> Authors 22959<br />

Contents <strong>of</strong> J. Photographic <strong>Society</strong> <strong>of</strong> Japan23060<br />

Contents <strong>of</strong> J. Printing <strong>Science</strong> and Technology <strong>of</strong> Japan23161<br />

Contents <strong>of</strong> J. Inst. Image Electronics Engineers <strong>of</strong> Japan 23262<br />

Contents <strong>of</strong> <strong>Journal</strong> <strong>of</strong> <strong>Imaging</strong> <strong>Science</strong> and Technology 23363<br />

Essays on <strong>Imaging</strong><br />

The <strong>Imaging</strong> <strong>Society</strong> <strong>of</strong> Japan<br />

c/o Tokyo Polytechnic University, 2-9-5, Honcho, Nakano-ku, Tokyo, 1648768 Japan<br />

Phone :033373-9576 Fax :033372-4414 E-mail :


Advanced Measurement Systems <strong>for</strong> All R&D and<br />

Quality Control Needs in Electrophotography,<br />

Inkjet and Other Printing Technologies<br />

PDT ® -2000 series<br />

Electrophotographic characterization,<br />

uni<strong>for</strong>mity mapping, and defect<br />

detection <strong>for</strong> large and small <strong>for</strong>mat<br />

OPC drums<br />

PDT ® -1000L<br />

PDT ® -1000<br />

ECT-100 TM<br />

OPC drum coating thickness<br />

gauge<br />

Electrophotographic<br />

Component Testing<br />

MFA-2000 TM<br />

Magnetic field distribution<br />

analysis in mag roller magnets<br />

DRA-2000 TM<br />

Semi-insulating components testing<br />

including charge rollers, mag rollers,<br />

transfer rollers, transfer belts, and<br />

print media<br />

TFS-1000 TM<br />

Toner fusing latitude testing<br />

Objective Print Quality<br />

Analysis <strong>for</strong> All Digital<br />

Printing Technologies<br />

IAS ® -1000<br />

Fully-automated high volume print<br />

quality testing<br />

Scanner-based high speed print<br />

quality analysis<br />

Scanner IAS ®<br />

Personal IAS ®<br />

Handheld series <strong>for</strong> print quality,<br />

distinctness <strong>of</strong> image (DOI), and<br />

color measurements. Truly portable;<br />

no PC connection required<br />

PocketSpec TM<br />

DIAS TM<br />

Quality Engineering Associates, Inc.<br />

99 South Bed<strong>for</strong>d Street #4, Burlington, MA 01803 USA<br />

Tel: +1 (781) 221-0080 • Fax: +1 (781) 221-7107 • info@qea.com • www.qea.com


imaging.org<br />

your source <strong>for</strong> imaging technology conferences<br />

UPCOMING CONFERENCES<br />

Join us in Anchorage, Alaska!<br />

September 16-21, 2007<br />

NIP23 Sessions<br />

NIP23<br />

23rd International Conference on<br />

Digital Printing Technologies<br />

www.imaging.org/conferences/nip23<br />

• Advanced <strong>Material</strong>s and Nanoparticles in <strong>Imaging</strong><br />

• Color <strong>Science</strong> and Image Processing<br />

• Digital Art<br />

• Electronic Paper and Paper-like Displays<br />

• Environmental Issues<br />

• Fusing Curing and Drying<br />

• Image Permanence<br />

• Ink Jet Printing <strong>Material</strong>s<br />

• Ink Jet Printing Processes<br />

• Media <strong>for</strong> Digital Printing and Displays<br />

Come and learn about...<br />

• Photo-electronic <strong>Material</strong>s and Devices<br />

• Print and Image Quality<br />

• Printing Systems Engineering and Optimization<br />

• Production Digital Printing<br />

• Security and Forensic Printing<br />

• Textile & Industrial Printing<br />

• Thermal Printing<br />

• Toner Based Printing <strong>Material</strong>s<br />

• Toner Based Printing Processes<br />

DF 2007 Sessions<br />

• Industrial and Commercial Applications<br />

• <strong>Material</strong>s and Substrates<br />

• New and Novel Direct Write Methods<br />

• Printed Architectural Components<br />

• Printed Electronics and Devices<br />

• Printing <strong>of</strong> Biomaterials<br />

• Plus: Joint Session Intellectual Property Panel on “Future<br />

and Limitations <strong>of</strong> Ink Jet and Electrophotography”<br />

Digital<br />

Fabrication 2007<br />

www.imaging.org/conferences/df2007<br />

For more in<strong>for</strong>mation visit the conference website or visit: www.imaging.org/conferences; or contact us at info@imaging.org

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!