Advanced Color Image Processing and Analysis
Advanced Color Image Processing and Analysis
Christine Fernandez-Maloigne
Editor
123
Editor
Christine Fernandez-Maloigne
Xlim-SIC Laboratory
University of Poitiers
11 Bd Marie et Pierre Curie
Futuroscope
France
v
vi Preface
Each color carries its own signature, its own vibration. . . its own universal
language built over millennia! The Egyptians of Antiquity gave to the principal
colors a symbolic value system resulting from the perception they had of natural
phenomena in correlation with these colors: the yellow of the sun, the green of
the vegetation, the black of the fertile ground, the blue of the sky, and the red of
the desert. For religious paintings, the priests generally authorized only a limited
number of colors: white, black, the three basic colors (red, yellow and blue), or their
combinations (green, brown, pink and gray). Ever since, the language of color has
made its way through time, and today therapeutic techniques use colors to convey
this universal language to the unconscious, to open doors to facilitate the cure.
In the scientific world, although the fundamental laws of physics were discovered
in the 1930s, colorimetrics had to await the rise of data processing to be able to use
the many matrix algebra applications that it implies.
In the numerical world, color is of vital importance, as it is necessary to code and
to model, while respecting the basic phenomena of the perception of its appearance,
as we recall in Chaps. 1 and 2. Then color is measured numerically (Chap. 3),
moves from one peripheral to another (Chap. 4), is handled (Chaps. 5–7), to
extract automatically discriminating information from the images and the videos
(Chaps. 8–11) to allow an automatic analysis. It is also necessary to specifically
protect this information, as we show in Chap. 12, to evaluate its quality, with
the metrics and standardized protocols described in Chap. 13. It is with the two
applications in which color is central, the field of art and the field of medicine, that
we conclude this work (Chaps. 14 and 15), which has brought together authors from
all the continents.
Whether looked at as a symbol of joy or of sorrow, single or combined, color is
indeed a symbol of union! Thanks to it, I met many impassioned researchers from
around the world who became my friends, who are like the members of a big family,
rich in colors of skin, hair, eyes, landscapes, and emotions. Each chapter of this will
deliver to you a part of the enigma of digital color imaging and, within filigree, the
stories of all these rainbow meetings. Good reading!
Contents
1 Fundamentals of Color.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 1
M. James Shyu and Jussi Parkkinen
2 CIECAM02 and Its Recent Developments .. . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 19
Ming Ronnier Luo and Changjun Li
3 Colour Difference Evaluation . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 59
Manuel Melgosa, Alain Trémeau, and Guihua Cui
4 Cross-Media Color Reproduction and Display Characterization .. . . . 81
Jean-Baptiste Thomas, Jon Y. Hardeberg, and Alain Trémeau
5 Dihedral Color Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 119
Reiner Lenz, Vasileios Zografos, and Martin Solli
6 Color Representation and Processes with Clifford Algebra . . . . . . . . . . . 147
Philippe Carré and Michel Berthier
7 Image Super-Resolution, a State-of-the-Art Review
and Evaluation .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 181
Aldo Maalouf and Mohamed-Chaker Larabi
8 Color Image Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 219
Mihai Ivanovici, Noël Richard, and Dietrich Paulus
9 Parametric Stochastic Modeling for Color Image
Segmentation and Texture Characterization . . . . . . . .. . . . . . . . . . . . . . . . . . . . 279
Imtnan-Ul-Haque Qazi, Olivier Alata, and Zoltan Kato
10 Color Invariants for Object Recognition .. . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 327
Damien Muselet and Brian Funt
11 Motion Estimation in Colour Image Sequences . . . . .. . . . . . . . . . . . . . . . . . . . 377
Jenny Benois-Pineau, Brian C. Lovell, and Robert J. Andrews
vii
viii Contents
Index . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 499
Chapter 1
Fundamentals of Color
The ability of human beings to perceive color is fantastic. Not only does it make
it possible for us to see the world in a more vibrant way, but it also creates the
wonder that we can express our emotions by using various colors. In Fig. 1.1, the
colors on the wooden window are painted with the meaning of bringing prosperity.
In a way, we see the wonderful world through the colors as a window. There are
endless ways to use, to interpret, and even to process color with the versatility that
is in the nature of color. However, to better handle the vocabulary of color, we need
to understand its attributes first. How to process as well as analyze color images
for specific purposes under various conditions is another important subject which
further extends the wonder of color.
In the communication between humans, color is a fundamental property of
objects. We learn different colors in our early childhood and this seems to be obvious
for us. However, when we start to analyze color more accurately and, for example,
want to measure color accurately, it is not so obvious anymore. For accurate color
measurement, understanding, and management, we need to answer the question:
What is color?
In color vocabulary, black and white are the first words to be used as color names
[2]. After them when the language develops, come red and yellow. The vocabulary
is naturally related to the understanding of nature. Therefore in ancient times, the
color names were related to the four basic elements of the world, water, air, fire,
and earth [9]. In ancient times, the color theory was developed by philosophers like
Plato and Aristotle. For the later development of color theory, it is notable that white
was seen as a basic color. Also the color mixtures were taken into theories, but each
basic color was considered to be a single and separate entity [14].
Also from the point of view of the revolution of color theory by Newton [20], it is
interesting to note that Aristotle had a seven basic color scale, where colors crimson,
4 M.J. Shyu and J. Parkkinen
a
1
0.9
0.8
B
0.7
G
0.6
R
0.5
Y
0.4
M
0.3
C
0.2
0.1
0
380 430 480 530 580 630 680 730
Fig. 1.2 (a) A set of color spectra (x-axis: wavelength from 380 to 730 nm, y-axis: reflectance
factor) and (b) the corresponding colors
violet, leek-green, deep blue, and gray or yellow formed the color scale from black
to white [9]. Aristotle also explains the color sensation so, that color sets the air in
movement and that movement extends from object to the eye [24].
From these theories, one can see that already in ancient times, there exists the
idea of some colors to be mixtures of primary colors and seven primary colors.
Also, it is easy to understand the upcoming problems of Newton’s description of
colors, when the view was that each primary color is a single entity and the color
sensation was seen as a kind of mechanical contact between light and the eye. The
ancient way of thinking was strong until the Seventeenth century.
In the middle of the Seventeenth century, the collected information was enough
to break the theory of ancient Greek about light and color. There were a number
of experiments by prism and color in the early Seventeenth century. The credit
for the discovery of the nature of light as a spectrum of wavelengths is given to
Isaac Newton [20]. The idea that colors are formed as a combination of different
component rays, which are immaterial by nature, was revolutionary at Newton’s
time. It broke the strong influence of ancient Greek thinking. This revolutionary
idea was not easily accepted. A notable person was Johann Wolfgang von Goethe,
who was still in the Nineteenth century opposing Newton’s theory strongly [10].
Newton also presented colors in a color circle. In his idea, there were seven
basic colors: violet, indigo, blue, green, yellow, orange, and red [14]. In the spectral
approach to color as shown in Fig. 1.2, the wavelength scale is linear and continuing
1 Fundamentals of Color 5
both ends, UV from short wavelengths and IR from long wavelengths. From the first
look, the circle form is not natural for this physical signal. However, when the hu-
man perception of the different wavebands is considered, the circle form seems to be
a good way to represent colors. The element, which connects the both ends of visible
spectrum into a circle, is purple, which includes both red and violet part of spectrum.
The first to present colors in a circle form was the Finnish mathematician and
astronomer Sigfrid Forsius in 1611 [14]. There are two different circle representa-
tion and both are based on idea to move from black to white through different color
steps. Since Forsius and Newton, there are a number of presentations of colors on a
circle. The circular form is used for the small number of basic colors. For continuous
color tones, three-dimensional color coordinate systems form other shapes like cone
(HSV) and cube (RGB).
Next important phase in the development of color science was the Nineteenth
century. At that time, theories of human color vision were developed. In 1801,
the English physicist and physician Thomas Young restated an earlier hypothesis
by the English glassmaker George Palmer from the year 1777 [14]. According to
these ideas, there are three different types of color-sensitive cells in the human
retina. In their model, these cells are sensitive to red, green, and violet and to
other colors which are mixtures of these principal pure colors. German physicist
Hermann Helmholz studied this model further. He also provided the first estimates
of spectral sensitivity curves for the retinal cells. This is known as the Young—
Helmholz theory of color vision.
In the mid-Nineteenth century the Young—Helmholz theory was not fully
accepted and in the mid-1870s German physician and physiologist Karl Hering
presented his theory of human color vision [14]. His theory was based on four fun-
damental colors: red, yellow, green, and blue. This idea is the basis for the opponent
color theory, where red—green and blue—yellow form opponent color pairs.
Both theories, the Young—Helmholz theory of color vision and the Hering
opponent color theory, seemed to give a valid explanation to many observations
about the human color vision. However, they were different even in the number
of principal or fundamental colors. German physiologist Johannes von Kries
proposed a solution to this confusion. He explained that the Young—Helmholz
theory explained color vision on retinal color-sensitive cells level, and the Hering’s
opponent color theory was explaining color processes later in visual pathway [14].
This description was not accepted for some years, but currently it is seen as the basic
view about the human color vision.
These ideas were bases for human color vision models, for the trichromatic color
theories, and the standards of representing colors on a three-dimensional space.
However, basing color representation and management on trichromatic theory of
human color vision is very restrictive in many ways. Standard three-dimensional
color coordinates are useful in many practical settings, where color is managed
for humans to look at, especially under fixed illumination. However, there are also
several drawbacks in the color representation based on human color vision.
Current level of measurement accuracy has led to a situation where in
the equations for calculating color coordinates or color differences have become
6 M.J. Shyu and J. Parkkinen
complicated. There are number of parameters, many without explaining the theory,
but fitting the measurements to correspond the model. Furthermore, there are
a number of issues, which cannot be managed by trichromatic color models.
These include, e.g., fluorescence, metamerism, animal color vision, and transfer of
accurate color information. To overcome these drawbacks, spectral color science
has increased interest and is used more and more in color science.
As mentioned above, the basis of color is light, a physical signal of electromag-
netic radiation. This radiation is detected by some detection system. If the system is
human vision, then we consider traditional color. If we do not restrict the detection
system, we consider the physical signal, color spectrum. Kuehni separates these
approaches into color and spectral color.
where l(λ ) is the spectrum of illumination, r(λ ) is the reflectance spectrum of the
object, si (λ ) is the sensitivity of the ith detector, and n is the number of detectors.
If the detector system has only one detector, it sees only intensity differences
and not colors. Or we can also say that the detector sees only intensities of one
color, i.e., color corresponding to the sensitivity s(λ ). For color sensation, at least
two detectors w ith different wavelength sensitivities are needed (n ≥ 2). The ratio
of these different detector responses gives the color information. In the human eye,
there are three types of wavelength-sensitive cone-cells (n = 3). These cells collect
the color information from the incoming signal and human visual system converts
it into color we see.
When we consider the color of an object, an essential part of color detection is
the illumination. Since the color signal is originally light reflected (or radiated, or
transmitted) from an object, the color of the illumination also affects to the detected
object’s color, term l(λ )r(λ ) in (1.1). A schematic drawing of detection of object
color is shown in Fig. 1.3.
1 Fundamentals of Color 7
Here we have the approach that the color information is carried by the electro-
magnetic signal coming from the object and reaching the detector system. For this
approach, we can set to the color signal certain assumptions.
Reflectance spectrum r(λ ) (or color spectrum l(λ )r(λ )) can be represented as a
function r: Λ → R, which satisfies
The proposition can be set due to the physical properties of the electromagnetic
radiation. It means that reflectance (radiance or transmittance) spectra and color
spectra can be thought as members of the square integrable function space,
L2 . Since in practice the spectrum is formed as discrete measurements of the
continuous signal, the spectra are represented as vectors in the space Rn . If spectra
are represented in a low-dimensional space, they lose information, which causes
problems like metamerism.
Using the vector space approach to the color, there are some questions to consider
related to the color representation:
– What are the methods to manage color accurately?
– What is the actual dimensionality of color information?
– How to select the dimensions to represent color properly?
In the case of standard color coordinates, the dimensionality has been selected to be
three. This is based on the models of the human color vision. Models are based on
the assumption that there are three types of color sensitivity functions in the human
retina.
In the spectral approach, originally the color signal is treated by using linear
models [17, 18, 22, 23, 25]. The most popularly used and the standard method is the
principal component analysis (PCA).
In this view, colors are represented as inner products between color spectrum and
basis spectra of defined coordinate system. This approach unifies the ground of the
different methods of the color representation and analysis. The basis spectra can be
defined, e.g., by human response curves of three colors or by the interesting colors
using some learning algorithm, depending on the needs and applications.
8 M.J. Shyu and J. Parkkinen
In the case of human eye n = 3 in (1.1) and si (λ )’s are marked as x̄(λ ), ȳ(λ ), z̄(λ )
and called color matching functions [27]. This leads to the tristimulus values X, Y,
and Z
X=k l(λ ) r(λ )x̄(λ ) d λ
Y=k l(λ ) r(λ )ȳ(λ ) d λ (1.3)
Z=k l(λ ) r(λ )z̄(λ ) d λ
k = 100/ l(λ )ȳ(λ ) d λ
Moreover, three elements are involved for a human to perceive color on an object:
light source, object, and observer. The physical property of the light source and
the surface property of the object can be easily measured in their spectral power
distribution with optical instruments. However, the observer’s sensation of color
cannot be measured directly by instruments since there is no place to gather a direct
reading of perception. Equation (1.3) represents an implicit way to describe the
human color perception in a numerical way which makes it possible to bring the
human color perception into a quantitative form and to further compute or process it.
This implicit model to describe human color perception can be observed by the
color-matching phenomena of two physically (spectrally) different objects which
appear as the same color to the human eye, in the following equations:
1 Fundamentals of Color 9
2.5
x2
y2
2 z2
x10
1.5 y10
z10
0.5
0
380 420 460 500 540 580 620 660 700 740 780
Fig. 1.4 Color matching functions for CIE standard observer in 2 and 10◦ -degree viewing angles
l(λ ) r1 (λ ) x̄(λ ) d λ = l(λ ) r2 (λ ) x̄(λ ) d λ
l(λ ) r1 (λ ) ȳ(λ ) d λ = l(λ ) r2 (λ ) ȳ(λ ) d λ (1.4)
l(λ ) r1 (λ ) z̄(λ ) d λ = l(λ ) r2 (λ ) z̄(λ ) d λ
Due to the integral operation in the equations, there can be two sets of different
spectral reflectance of two objects that cause the equality to happen, i.e., make
them appear as the same color. Furthermore, with the known (measurable) physical
stimuli in the equations, if the unknown color-matching functions (x̄(λ ), ȳ(λ ), z̄(λ ))
can be derived for the human visual system, it is possible to predict whether two
objects of different spectral power distribution would appear as equal under this
human visual color-matching model.
It was the Commission International de l’Eclairage (CIE) that in 1924 took the
initiative to set up a Colorimetry Study Committee to coordinate the derivation of the
color-matching functions [6]. Based on experimental color-mixture data and not on
any particular theory of the color vision process, a set of color-matching functions
for use in technical Colorimetry was first presented to the Colorimetry Committee
at the 1931 CIE sessions [6]. This “1931 Standard Observer” as it was then called
was based on observations made with colorimeters using field sizes subtending
2 degrees. In 1964, the CIE took a further step to standardizing a second set of
color-matching functions as the “1964 Standard Observer” which used field sizes
subtending 10 degrees. With these two sets of color-matching functions, shown in
Fig. 1.4, it is possible to compute human color perception and subsequently open
up promising research in the world of color science based on the model of human
vision.
10 M.J. Shyu and J. Parkkinen
1.5 Metamerism
Fig. 1.5 Example of metamerism: two different reflectance curves from a metameric pair that
could appear as the same color under specific illumination
1 Fundamentals of Color 11
Another aspect related to the color appearance under two different conditions
is the color constancy. It is a phenomenon, where the observer considers the object
color the same under different illuminations [3]. It means that the color is understood
to be the same although the color signal reaching the eye is different under different
illuminations. The color constancy can be seen as related to a color-naming problem
[8]. The color constancy is considered in the Retinex theory, which is the basis, e.g.,
for the illumination change normalized color image analysis method [15].
In the color constancy, the background and the context, where the object is seen,
are important for constant color appearance. If we look at a red paper under white
illumination on a black background, it looks the same as that of a white paper
under red illumination on a black background [8]. This implicit model to describe
human color perception can be observed by the color-matching phenomena of two
physically different objects that appear to be the same color to human eyes.
The measurement of color can be done in various ways. In printing and publishing,
the reflection densitometer has been used historically in prepress and pressroom
operations for color quality control. ISO standard 5/3 for Density Measurement—
Spectral Conditions defines a set of weightings indicating the standard spectral
response for Status A, Status M, and Status T filters [1]. Reflectance density (DR ) is
calculated from spectral reflectance according to the following equation:
where
r(λ ) is the reflectance value at wavelength λ of the object measured
Π(λ ) is the spectral product at wavelength λ for the appropriate density response
It is well known that densitometers can be used to evaluate print characteristics
such as consistency of color from sheet to sheet, color uniformity across the sheet,
and color matching of the proof. According to (1.5), one can find that for two prints
of the same ink, if the reflectance values r(λ ) are the same, it is certain that the
density measures will be the same, i.e., the color of the prints will appear to be the
same. However, it is also known that for two inks whose narrow-band density values
have been measured as identical could appear as different colors to the human eye
if their spectral characteristics are different in the insensitive dead zone of the filter
[5]. It must be pointed out that due to the spectral product at each wavelength, prints
even with the same density values but not with the same ink have not necessarily
the same spectral reflectance values, i.e., they can appear as different colors to the
human eye. Since the spectral product in densitometry is not directly related to
12 M.J. Shyu and J. Parkkinen
Fig. 1.6 (a) Gray patches with the same color setting appear as the same color. (b) The same
patches in the center appear as different levels of gray due to the “simultaneous contrast effect”
where the background influence makes the central color patches appear different
human visual response, the density measure can only guarantee the equality of the
physical property of the same material, not the perceptual attribute of the color that
appears.
There are similarities and differences between Densitometry and Colorimetry.
Both involve integration with certain spectral weightings, but only the spectral
weighting in the color-matching functions in Colorimetry is directly linked to the
responsivity of human color vision. The measurement of color in the colorimetric
way defined in (1.3) is therefore precisely related to the perceptual attribute of
human color vision.
On the other hand, the resulting values of Colorimetry are more into the percep-
tual measurements of human color response. By definition in (1.4), if the spectral
reflectance of r1 (λ ) and r2 (λ ) are exactly the same, this “spectral matching” method
can of course create the sensation of two objects of the same color. However, it is
not necessary to constrain the reflectance of the two objects to be exactly the same,
as long as the integration results are the same, the sensation of color equality would
occur, which is referred as “colorimetric matching.” These two types of matching
are based on the same physical properties of the light source and the same adaptation
status of the visual system, which is usually referred as “fundamental Colorimetry”
(or simple CIE XYZ tristimulus system).
Advanced Colorimetry usually refers to the color processing that goes beyond
the matching between simple solid color patches or pixels, where spatial influence,
various light sources, different luminance levels, different visual adaptation, and
various appearance phenomena are involved in a cross media environment. These
are the areas on which active research into Color Imaging focuses and the topics
covered in the subsequent chapters. One example is shown in Fig. 1.6a where all
the gray patches are painted with the same R, G, and B numbers and appear
as the same color in such circumstances. However, the same gray patches with
different background color patches now appear as different levels of gray as shown
in Fig. 1.6b. This so-called “simultaneous contrast effect” gives a good example of
how “advanced Colorimetry” has to deal with subjects beyond the matching among
simple color patches where spatial influence and background factors, etc. are taken
into consideration.
1 Fundamentals of Color 13
L∗ = 116 (Y/Yn)1/3 − 16
a∗ = 500 (X/Xn )1/3 − (Y/Yn )1/3 (1.7)
b∗ = 200 (Y/Yn )1/3 − (Z/Zn)1/3
where Y/Yn , X/Xn , and Z/Zn > 0.008856; more details in CIE 15.24
X, Y, and Z are tristimulus values of the object measured
Xn , Yn , and Zn are the tristimulus values of a reference white object
Fig. 1.7 Gray scales in physical and perceptual linear space: (a) a gray scale with a linear
increment of the reflectance factor (0.05) and (b) a gray scale with a visually linear increment
of the L* (Lightness) value in the CIELAB coordinate
14 M.J. Shyu and J. Parkkinen
At the end of this chapter, we have a short philosophical discussion about color. In
general texts and discussions the term “color” is not used rigorously. It can mean the
human sensation, it can mean the color signal reflected from an object, and it can
be a property of an object itself. In the traditional color approach, it is connected to
the model of human color vision. Yet the same vocabulary is used when considering
the animal vision, although the animal (color) vision systems may vary very much
from that of human. In order to analyze and manage the color, we need to define
1 Fundamentals of Color 15
Fig. 1.8 Colorful banners are used in Japanese traditional buildings (Photographed by M. James
Shyu in Kyoto, Japan)
color well. In this chapter, and in the book, the spectral approach is described in
addition to the traditional color representation. In the spectral approach, color means
the color signal originated from the object and reaching the color detection system.
Both approaches are used in this book depending on the topic of the chapter.
In the traditional color science, black and white, and gray levels in between, are
called achromatic light. This means that they differ from each other only by radiant
intensity, or luminous intensity in photometrical terms. Other light is chromatic.
Hence, one may say that black, white, and gray are not colors. This is a meaningful
description only, if we have a fixed detection system, which is well defined. In the
traditional color approach, the human color vision is considered to be based on a
fixed detection system. Studies about human cone sensitivities and cone distribution
show that this is not the case [12].
In the spectral approach, the “achromaticity” of black, white, and gray levels
is not so obvious. In the spectral sense, the ultimate white is the equal energy
white, for which the spectrum intensity is a constant maximum value over the whole
wavelength range. When we start to decrease the intensity for some wavelength, the
spectrum changes and at a certain point the spectrum represents a color in traditional
means. If we consider white not to be a color, we have to define “epsilon” for each
wavelength by which the change from the white spectrum makes it a color. Also,
white spectrum can be seen as the limit of sequence of color spectra. This means
that in the traditional color approach, the limit of sequence of colors is not color.
Blackness, whiteness, and grayness are also dependent on detection system.
Detected signal looks white, when all the wavelength-sensitive sensors give the
16 M.J. Shyu and J. Parkkinen
a
1
EE white
Limited white
0
380 430 480 530 580 630 680 730 780
b
1
sensor A1
sensor A2
sensor A3
0
380 430 480 530 580 630 680 730 780
c
1
sensor B1
sensor B2
sensor B3
0
380 430 480 530 580 630 680 730 780
Fig. 1.9 White is a relative attribute. (a) Two spectra, equal energy white (blue line) and a
spectrum which look white for sensors A, but colored to sensors B (red line). (b) Sensors A,
sensitivity functions have the same shape as the “limited white” (red line) spectrum on (a).
(c) Sensors B, sensitivity functions does not match with spectrum “limited white” (red line) in (a)
maximum response. In Fig. 1.9a there are two color signals, which both “looks
white” for the theoretical color detection system given in Fig. 1.9b. But if we change
the detector system to one shown in Fig. 1.9c, the other white is a colored signal,
since not all the sensors have maximum input.
1 Fundamentals of Color 17
With this small discussion, we want to show that in the color science there is a
need and development into direction of generalized color. In this approach, color is
not restricted to the human visual system, but its basis is in a measurable and well-
defined color signal. Signal, which originates from the object, reaches the color
detection system, and carries the full color information of the object. The traditional
color approach is shown to be powerful tool to manage color for human vision.
The well defined models are useful tools also in the future, but the main restriction,
uncertainty in understanding of detection system, needs much research also in the
future.
References
21. George Palmer, Theory of Colors and Vision”, from Selected papers on Colorimetry—
Fundamentals. In D.L. MacAdam (ed.), SPIE Milestone Series Volume MS77, pp. 5–8, SPIE
Optical Engineering Press, 1993 (Originally printed by Leacroft 1777 reprinted from Sources
of Color Science, pp. 40–47, MIT Press, 1970)
22. Parkkinen JPS, Jaaskelainen T, Oja E (1985) Pattern recognition approach to color measure-
ment and discrimination Acta Polytechnica Scandinavica: Appl Phys 1(149):171–174
23. Parkkinen JPS, Hallikainen J, Jaaskelainen T (1989) Characteristic spectra of Munsell colors.
J Opt Soc Am A 6:318–322
24. Wade NJ (1999) A natural history of vision MIT Press, Cambridge, MA, USA 2nd printing
25. Wandell B (1985) The synthesis and analysis of color images, NASA Technical Memorandum
86844. Ames Research Center, California, USA, pp 1–34
26. William David Wright, The CIE Contribution to Colour Technology, 1931 to 1987, pp. 2–5, in
Inter-Society Color Council News, Number 368, July/August, 1997
27. Wyszecki G, Stiles W (1982) Color science: concepts and methods, quantitative data and
formulae 2nd edition Wiley, New York
28. Zollinger H (1999) Color: a multidisciplinary approach. Wiley, Weinheim
Chapter 2
CIECAM02 and Its Recent Developments
The reflection is for the colors what the echo is for the sounds
Joseph Joubert
Abstract The development of colorimetry can be divided into three stages: colour
specification, colour difference evaluation and colour appearance modelling. Stage 1
considers the communication of colour information by numbers. The second stage
is colour difference evaluation. While the CIE system has been successfully applied
for over 80 years, it can only be used under quite limited viewing conditions,
e.g., daylight illuminant, high luminance level, and some standardised view-
ing/illuminating geometries. However, with recent demands on crossmedia colour
reproduction, e.g., to match the appearance of a colour or an image on a display
to that on hard copy paper, conventional colorimetry is becoming insufficient. It
requires a colour appearance model capable of predicting colour appearance across
a wide range of viewing conditions so that colour appearance modelling becomes
the third stage of colorimetry. Some call this as advanced colorimetry. This chapter
will focused on the recent developments based on CIECAM02.
2.1 Introduction
The development of colorimetry [1] can be divided into three stages: colour
specification, colour difference evaluation and colour appearance modelling. Stage 1
considers the communication of colour information by numbers. The Commission
Internationale de l’Eclairage (CIE) recommended a colour specification system in
1931 and later, it was further extended in 1964 [2]. The major components include
standard colorimetric observers, or colour matching functions, standard illuminants
and standard viewing and illuminating geometry. The typical colorimetric measures
are the tristimulus value (X,Y, Z), chromaticity coordinates (x, y), dominant wave-
length, and excitation purity.
The second stage is colour difference evaluation. After the recommendation
of the CIE specification system in 1931, it was quickly realised that the colour
space based on chromaticity coordinates was far from a uniform space, i.e., two
pairs of stimuli having similar perceived colour difference would show large
difference of the two distances from the chromaticity diagram. Hence, various
uniform colour spaces and colour difference formulae were developed. In 1976,
the CIE recommended CIELAB and CIELUV colour spaces [2] for presenting
colour relationships and calculating colour differences, More recently, the CIE
recommended the CIEDE2000 [3] for evaluating colour differences.
While the CIE system has been successfully applied for over 80 years, it can only
be used under quite limited viewing conditions, for example, daylight illuminant,
high luminance level, and some standardised viewing/illuminating geometries.
However, with recent demands on cross-media colour reproduction, for example,
to match the appearance of a colour or an image on a display to that on hard
copy paper, conventional colorimetry is becoming insufficient. It requires a colour
appearance model capable of predicting colour appearance across a wide range of
viewing conditions so that colour appearance modelling becomes the third stage of
colorimetry. Some call this as advanced colorimetry.
A great deal of research has been carried out to understand colour appearance
phenomena and to model colour appearance. In 1997, the CIE recommended a
colour appearance model designated CIECAM97s [4,5], in which the “s” represents
a simple version and the “97” means the model was considered as an interim model
with the expectation that it would be revised as more data and better theoretical un-
derstanding became available. Since then, the model has been extensively evaluated
by not only academic researchers but also industrial engineers in the imaging and
graphic arts industries. Some shortcomings were identified and the original model
was revised. In 2002, a new model: CIECAM02 [6, 7] was recommended, which is
simpler and has a better accuracy than CIECAM97s.
The authors previously wrote an article to describe the developments of
CIECAM97s and CIECAM02 [8]. The present article will be more focused on
the recent developments based on CIECAM02. There are six sections in this
chapter. Section 2.2 defines the viewing conditions and colour appearance terms
used in CIECAM02. Section 2.3 introduces some important colour appearance data
2 CIECAM02 and Its Recent Developments 21
sets which were used for deriving CIECAM02. In Sect. 2.4, a brief introduction
of different chromatic adaptation transforms (CAT) leading to the CAT02 [8],
embedded in CIECAM02, will be given. Section 2.5 gives various visual phenomena
predicted by CIECAM02. Section 2.6 summarises some recent developments of the
CIECAM02. For example, the new uniform colour spaces based on CIECAM02
by Luo et al. (CAM02-UCS, CAM02-SCD and CAM02-LCD) [9] will be covered.
Xiao et al. [10–12] extended CIECAM02 to predict the change in size of viewing
field on colour appearance, known as size effect. Fu et al. [13] has extended the
CIECAM02 for predicting colour appearances of unrelated colours presented in
mesopic region. Finally, efforts were paid to modify the CIECAM02 in connection
with international color consortium (ICC) profile connection space for the colour
management [14]. In the final section, the authors point out a concept of the
universal model based on CIECAM02.
Proximal field
Background
Surround
The aim of the colour appearance model is to predict the colour appearance under
different viewing conditions. Various components in a viewing field have an impact
on the colour appearance of a stimulus. Hence, the accurate definition of each
component of the viewing field is important. Figures 2.2–2.4 are three configurations
considered in this chapter: colour patches for related colours, images for related
colours, and patches for unrelated colours. The components in each configuration
will be described below. Note that in the real world, objects are normally viewed
in a complex context of many stimuli; they are known as “related” colours. An
“unrelated colour” is perceived by itself, and is isolated, either completely or
partially, from any other colours. Typical examples of unrelated colours are signal
lights, traffic lights, and street lights, viewed in a dark night.
[Link] Stimulus
In Figs. 2.2 and 2.4 configurations, the stimulus is a colour element for which a
measure of colour appearance is required. Typically, the stimulus is taken to be
a uniform patch of about 2◦ angular subtense. A stimulus is first defined by the
tristimulus values (X,Y, Z) measured by a tele-spectroradiometer (TSR) and then
normalised against those of reference white so that Y is the percentage reflection
factor.
In Fig. 2.3 configuration, the stimulus becomes an image. The pixel of each
image is defined by device independent coordinates such as CIE XYZ or CIELAB
values.
2 CIECAM02 and Its Recent Developments 23
stimulus
Dark surround
In Fig. 2.2 configuration, the reference white is used for scaling lightness (see later)
of the test stimulus. It is assigned to have a lightness of 100. It is measured by a
TSR again to define the tristimulus values of the light source (XW ,YW , ZW ) in cd/m2
unit. The parameter of LW (equal to YW ) in the model defines the luminance of the
24 M.R. Luo and C. Li
light source. When viewing unrelated colours, there is no such element. For viewing
images, the reference white will be the white border (about 10 mm) surrounding the
image.
The reference white in this context can be considered as the “adopted white”
i.e., the measurement of “a stimulus that an observer who is adapted to the viewing
environment would judge to be perfectly achromatic and to have a reflectance factor
of unity (i.e., have absolute colorimetric coordinates that an observer would consider
to be the perfect white diffuser)” [ISO 12231] For viewing an image, there could
be some bright areas such as a light source or specularly reflecting white objects,
possibly illuminated by different sources. In the latter case, the “adapted white” (the
actual stimulus which an observer adapted to the scene judges to be equivalent to a
perfect white diffuser) may be different from the adopted white measured as above.
[Link] Background
[Link] Surround
A surround is a field outside the background in Fig. 2.2 configuration, and outside
the white border (reference white) in Fig. 2.3. Surround includes the entire room or
the environment. Figure 2.4 configuration has a surround in complete darkness.
Surround is not measured directly, rather the surround ratio is determined and
used to assign a surround. The surround ratio, SR , can be computed:
where LSW is the luminance of the surround white and LDW is the luminance of the
device white. LSW is a measurement of a reference white in the surround field while
LDW is a measurement of the device white point for a given device, paper or peak
white. If SR is 0, then a dark surround is appropriate. If SR is less than 0.2, then a dim
surround should be used while an SR of greater than or equal to 0.2 corresponds to
an average surround. Different surround “average,” “dim,” “dark” leads to different
2 CIECAM02 and Its Recent Developments 25
parameters (F: incomplete adaptation factor; Nc: chromatic induction factor and c:
impact of surround) used in CIECAM02. Table 2.1 define SR values in some typical
examples in real applications.
For Fig. 2.2 configuration, adapting field is the total environment of the colour
element considered, including the proximal field, the background and the surround,
and extending to the limit of vision in all directions. For Fig. 2.3 image configura-
tion, it can be approximated the same as background, i.e., approximate an L∗ of 50.
The luminance of adapting field is expressed as LA , which can be approximated
by LW × Yb /100, or by Lb .
Another parameter is also very important concerning the range of illumination from
the source. It is well known that rods and cones in our eyes are not uniformly
distributed on the retina. Inside the foveola (the central 1◦ field of the eye), there
are only cones; outside, there are both cones and rods; in the area beyond about
40◦ from the visual axis, there are nearly all rods and very few cones. The rods
provide monochromatic vision under low luminance levels; this scotopic vision is
in operation when only rods are active, and this occurs when the luminance level is
less than about 0.1 cd/m2. Between this level and about 10 cd/m2 , vision involves a
mixture of rod and cone activities, which is referred to as mesopic vision. It requires
luminance of about 10 cd/m2 for photopic vision in which only cones are active.
26 M.R. Luo and C. Li
Colour appearance models based on colour vision theories have been developed to
fit various experimental data sets, which were carefully generated to study particular
colour appearance phenomena. Over the years, a number of experimental data sets
were accumulated to test and develop various colour appearance models. Data sets
investigated by CIE TC 1-52 CAT include: Mori et al. [16] from the Color Science
Association of Japan, McCann et al. [17] and Breneman [18] using a haploscopic
matching technique; Helson et al. [19], Lam and Rigg [20] and Braun and Fairchild
[21] using the memory matching technique; and Luo et al. [22, 23] and Kuo and
Luo [24] using the magnitude estimation method. These data sets, however, do
not include visual saturation correlates. Hence, Juan and Luo [25, 26] investigated
a data set of saturation correlates using the magnitude estimation method. The
data accumulated played an important role in the evaluation of the performance
of different colour appearance models and the development of the CIECAM97s and
CIECAM02.
1 In
this chapter we will use for simplified terms “D65” and “A” instead of the complete official
CIE terms: “CIE standard illuminant D65” and “CIE standard illuminant A”.
2 CIECAM02 and Its Recent Developments 29
a different illuminant (e.g., A). The following is divided into two parts: light and
chromatic adaptation, and the historical developments of Bradford transform [20],
CMCCAT2000 [27] and CAT02.
Adaptation can be divided into two: light and chromatic. The former is the
adaptation due to the change of light levels. It can be further divided into two: light
adaptation and dark adaptation. Light adaptation is the decrease in visual sensitivity
upon an increase in the overall level of illumination. An example occurs when
entering a bright room from a dark cinema. Dark adaptation is opposite to light
adaptation and occurs, for example, when entering a dark cinema from a well-lit
room.
normal. This is caused by the fact that most of coloured objects in the real world
are more or less colour constant (they do not change their colour appearance under
different illuminants). The most obvious example is white paper always appears
white regardless of which illuminant it is viewed under. The second stage is called
the “adaptive shift” and it is caused by physiological changes and by a cognitive
mechanism, which is based upon an observer’s knowledge of the colours in the
scene content in the viewing field. Judd [28] stated that “the processes by means of
which an observer adapts to the illuminant or discounts most of the effect of non-
daylight illumination are complicated; they are known to be partly retinal and partly
cortical”.
The von Kries coefficient law is the oldest and widely used to quantify chromatic
adaptation. In 1902, von Kries [29] assumed that, although the responses of the three
cone types (RGB)2 are affected differently by chromatic adaptation, the spectral
sensitivities of each of the three cone mechanisms remain unchanged. Hence,
chromatic adaptation can be considered as a reduction of sensitivity by a constant
factor for each of the three cone mechanisms. The magnitude of each factor depends
upon the colour of the stimulus to which the observer is adapted. The relationship,
given in (2.2), is known as the von Kries coefficient law.
Rc = α · R,
Gc = β · G,
Bc = γ · B, (2.2)
where Rc , Gc , Bc and R, G, B are the cone responses of the same observer, but
viewed under test and reference illuminants, respectively. α , β and γ are the von
Kries coefficients corresponding to the reduction in sensitivity of the three cone
mechanisms due to chromatic adaptation. These can be calculated using (2.3).
Rwr Gwr Bwr
α= ; β= ; γ= , (2.3)
Rw Gw Bw
where
R Rc G Gc B Bc
= , = , = , (2.4)
Rw Rwr Gw Gwr Bw Bwr
2 Inthis chapter the RGB symbols will be used for the cone fundamentals, in other chapters the
reader will find the LMS symbols. The use of RGB here should not be confused with the RGB
primaries used in visual colour matching.
2 CIECAM02 and Its Recent Developments 31
Here Rwr , Gwr , Bwr , and Rw , Gw , Bw are the cone responses under the reference and
test illuminants, respectively. Over the years, various CATs have been developed but
most are based on the von Kries coefficient law.
In 1985, Lam and Rigg accumulated a set of corresponding colour pairs. They used
58 wool samples that had been assessed twice by a panel of five observers under D65
and A illuminants. The memory-matching technique was used to establish pairs of
corresponding colours. In their experiment, a subgroup of colours was first arranged
in terms of chroma and hue, and each was then described using Munsell H V/C
coordinates. The data in H V/C terms, were then adjusted and converted to CIE
1931 XYZ values under illuminant C. Subsequently, the data under illuminant C
were transformed to those under illuminant D65 using the von Kries transform.
They used this set of data to derive a chromatic transform known as BFD transform
now. The BFD transform can be formulated as the following:
Step 1:
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
R X 0.8951 0.2664 0.1614
1
⎝ G ⎠ = MBFD ⎝ Y ⎠ with MBFD = ⎝ −0.7502 1.7135 0.0367 ⎠ .
Y
B Z 0.0389 −0.0685 1.0296
Step 2:
⎛ ⎞ ⎛ ⎞⎛ ⎞
Rc Rwr /Rw R
⎝ Gc ⎠ = ⎝ Gwr /Gw ⎠⎝ G ⎠ with
p
Bc Bwr /Bw sign(B)|B| p
Step 3:
⎛ ⎞ ⎛ ⎞
Xc Y Rc
⎝ Yc ⎠ = M −1 ⎝ Y Gc ⎠ .
BFD
Zc Y Bc
Note that the BFD transform is a nonlinear transform. The exponent p
in step 2 for calculating the blue corresponding spectral response can be
considered as a modification of the von Kries type of transform. The BFD
transform performs much better than the von Kries transform. In 1997, Luo
32 M.R. Luo and C. Li
and Hunt [30] in 1997 modified the step 2 in the above BFD transform by
introducing an adaptation factor D. The new step 2 becomes,
Step 2’
⎛ ⎞ ⎛ ⎞
Rc [D(Rwr /Rw ) + 1 − D]R
⎝ Gc ⎠ = ⎝ [D(Gwr /Gw ) + 1 − D]G ⎠,
Bc [D(Bwr /Bwp ) + 1 − D]sign(B)|B| p
where
1/4
D = F − F/[1 + 2LA + L2A /300].
The transform consisting of Step 1, Step 2’ and Step 3 was then recommended by
the colour measurement committee (CMC) of the society of dyers and colourists
(SDC) and, hence, was named as the CMCCAT97. This transform is included
in the CIECAM97s for describing colour appearance under different viewing
conditions. The BFD transform was originally derived by fitting only one data set,
Lam and Rigg. Although it gave a reasonably good fit to many other data sets,
it predicted badly the McCann data set. In addition, the BFD and CMCCAT97
include an exponent p for calculating the blue corresponding spectral response. This
causes uncertainty in reversibility and complexity in the reverse mode. Li et al.
[31] addressed this problem and provided a solution by including an iterative
approximation using the Newton method. However, this is unsatisfactory in imaging
applications where the calculations need to be repeated for each pixel. Li et al.
[27] gave a linearisation version by optimising the transform to fit all the available
data sets, rather than just the Lam and Rigg set. The new transform, named
CMCCAT2000, is given below.
[Link] Cmccat2000
Step 1:
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
R X 0.7982 0.3389 −0.1371
⎝ G ⎠ = M00 ⎝ Y ⎠ with M00 = ⎝ −0.5918 1.5512 0.0406 ⎠ .
B Z 0.0008 0.0239 0.9753
Step 2:
⎛ ⎞ ⎛ ⎞
Rc [D(Yw /Ywr )(Rwr /Rw ) + 1 − D]R
⎝ Gc ⎠ = ⎝ [D(Yw /Ywr )(Gwr /Gw ) + 1 − D]G ⎠
Bc [D(Yw /Ywr )(Bwr /Bw ) + 1 − D]B
with
D = F{0.08 log10 [0.5(LA1 + LA2 )] + 0.76 − 0.45(LA1 − LA2 )/(LA1 + LA2 )}.
2 CIECAM02 and Its Recent Developments 33
Step 3:
⎛ ⎞ ⎛ ⎞
Xc Rc
⎝ Yc ⎠ = M −1 ⎝ Gc ⎠ .
00
Zc Bc
The CMCCAT2000 not only overcomes all the problems with respect to reversibility
discussed above, but also gives a more accurate prediction than other transforms of
almost all the available data sets.
During and after the development of the CMCCAT2000, scientists decided to
drop the McCann et al. data set because the experiment was carried out under a
very chromatic adapting illuminant. Its viewing condition is much different from all
the other corresponding data sets. Hence, it would be better to optimising the linear
chromatic adaptation transform via fitting all the corresponding data sets without
the McCann et al. data set. The new matrix obtained by the authors, now named the
CAT02 matrix, is given by
⎛ ⎞
0.7328 0.4296 −0.1624
M02 = ⎝ −0.7036 1.6975 0.0061 ⎠ ,
0.0030 0.0136 0.9834
which was first included in the appendix of our paper [32] in 2002. At the same
time, Nathan Moroney (Chair of CIETC8-01 at that time) proposed a new formula
for D function:
1 −LA −42
D = F 1− e 92 . (2.5)
3.6
The CMCCAT2000 with the new matrix and D formula given by (2.5) becomes the
CAT02.
At a later stage, CIE TC 8-01 Colour Appearance Modelling for Colour
Management Systems had to choose a linear chromatic transform for CIECAM02.
Multiple candidates such as CMCCAT2000 [27], the sharp chromatic transform
[33] developed by Finlayson et al., and CAT02 [6–8] were proposed for use as
a von Kries type transform. All had similar levels of performance with respect
to the accuracy of predicting various combinations of previously derived sets of
corresponding colours. In addition to the sharpening of the spectral sensitivity
functions, considerations used to select the CIE transform included the degree
of backward compatibility with CIECAM97s and error propagation properties by
combining the forward and inverse linear CAT, and the data sets which were
used during the optimisation process. Finally, CAT02 was selected because it is
compatible with CMCCAT97 and was optimised using all available data sets except
the McCann et al. set, which includes a very chromatic adapting illuminant.
Figure 2.6 illustrates 52 pairs of corresponding colours predicted by CIECAM02
(or its chromatic adaptation transform, CAT02) from illuminant A (open circles
of vectors) to SE (open ends of vectors) plotted in the CIE u v chromaticity
diagram for the 2◦ observer. The open circle colours have a value of L∗ equal
34 M.R. Luo and C. Li
Fig. 2.6 The corresponding colours predicted by the CIECAM02 from illuminant A (open circles
of vectors) to illuminant SE (open ends of vectors) plotted in CIE u v chromaticity diagram for the
CIE 1931 standard colorimetric observer. The plus (+) and the dot (•) represent illuminants A and
SE , respectively
Hunt [34] studied the effect of light and dark adaptation on colour perception
and collected data for corresponding colours via a visual colorimeter using the
haploscopic matching technique, in which each eye was adapted to different viewing
conditions and matches were made between stimuli presented in each eye.
2 CIECAM02 and Its Recent Developments 35
The results revealed a visual phenomena known as Hunt effect [34]. It refers to
the fact that the colourfulness of a colour stimulus increases due to the increase
of luminance. This effect highlights the importance of considering the absolute
luminance level in colour appearance models, which is not considered in traditional
colorimetry.
Stevens and Stevens [35] asked observers to make magnitude estimations of the
brightness of stimuli across various adaptation conditions. The results showed
that the perceived brightness contrast increased with an increase in the adapting
luminance level according to a power relationship.
Bartleson and Breneman [36] found that the perceived contrast in colourfulness
and brightness increased with increasing illuminance level from dark surround, dim
surround to average surround. This is an important colour appearance phenomenon
to be modelled, especially for the imaging and graphic arts industries where, on
many occasions, it is required to reproduce images on different media under quite
distinct viewing conditions.
The lightness contrast effect [37] reflects that the perceived lightness increases when
colours are viewed against a darker background and vice versa. It is a type of
simultaneous contrast effect considering the change of colour appearance due to
different coloured backgrounds. This effect has been widely studied and it is well
known that a change in the background colour has a large impact on the perception
of lightness and hue. There is some effect on colourfulness, but this is much smaller
than the effect on lightness and hue [37].
When a grey scale is illuminated by a light source, the lighter neutral stimuli
will exhibit a certain amount of the hue of the light source and the darker
stimuli will show its complementary hue, which is known as the Helson–Judd
effect [39]. Thus for tungsten light, which is much yellower than daylight, the
lighter stimuli will appear yellowish, and the darker stimuli bluish. This effect is
not modelled by CIECAM02.
Recently, several extensions to the CIECAM02 have been made, which have
widened the applications of the CIECAM02. In this section, the extensions for
predicting colour discrimination data sets, size effects and unrelated colour appear-
ance in the mesopic region. Besides, recent developments from CIETC8-11 will be
reported as well.
average 2.5 ΔEab ∗ units, are a combined data set used to develop the CIE 2000 colour
where ΔJ , ΔaM and ΔbM are the differences of J , aM and bM between the “standard”
and “sample” in a pair. Here, KL is a lightness parameter and is given in Table 2.2.
Three colour spaces named CAM02-LCD, CAM02-SCD and CAM02-UCS were
developed for large, small and combined large and small differences, respectively.
The corresponding parameters in (2.6) and (2.7) are listed in Table 2.2.
The three new CIECAM02 based colour spaces, together with the other spaces
and formulae were also tested by Luo et al. [9]. The results confirmed that CAM02-
SCD and CAM02-LCD performed the best for small and large colour difference data
sets. When selecting one UCS to evaluate colour differences across a wide range,
CAM02-UCS performed the second best across all data sets. The authors have been
recommending using CAM02-UCS for all applications.
Figure 2.7 shows the relationship between CIECAM02 J and CAM02-UCS J’ and
Fig. 2.8 shows the relationship between CIECAM02 M and CAM02-UCS M’. It can
be seen that CIECAM02 J is less than CAM02-UCS J’ except at the two ends, while
CIECAM02 M is greater than CAM02-UCS M’ except when M = 0. Thus in order
to have a more uniform space, CIECAM02 J should be increased and CIECAM02
M should be decreased.
The experimental colour discrimination ellipses used in the previous studies
[44, 45] were also used for comparing different colour spaces. Figures 2.9 and
2.10 show the ellipses plotted in CIELAB and CAM02-UCS spaces, respectively.
The size of the ellipse was adjusted by a single factor in each space to ease
visual comparison. For perfect agreement between the experimental results and a
uniform colour space, all ellipses should be constant radius circles. Overall, it can
be seen that the ellipses in CIELAB (Fig. 2.9) are smaller in the neutral region
and gradually increase in size as chroma increases. In addition, the ellipses are
38 M.R. Luo and C. Li
orientated approximately towards the origin except for those in the blue region in
CIELAB space. All ellipses in CAM02-UCS (Fig. 2.10) are approximately equal-
sized circles. In other words, the newly developed CAM02-UCS is much more
uniform than CIELAB.
2 CIECAM02 and Its Recent Developments 39
Fig. 2.11 The flow chart of size effect correction model based on CIECAM02
Finally in Step 4, the colour appearance attributes J , C and H for the target
stimulus size θ are predicted using the formulae:
C = KC × C, (2.9)
H = H. (2.10)
The earlier experimental results [10] were used to derive the above model.
Figure 2.12 shows the corrected attributes J of 25◦ , 35◦ and 45◦ , respectively,
plotted against J at 2◦ viewing field. The thick solid line is the corrected J when
viewing field is 25◦ ; the dotted line corresponds to the J with viewing angle being
35◦ . The dashed line is the J with viewing angle of 45◦ . The thin solid line is the
45◦ line where J = J . The trend is quite clear as shown in Fig. 2.12, i.e., an increase
of lightness for a larger viewing field. For example, when J = 60 with a size of 2◦ , J
values are 62.9, 65.7 and 68.5 for sizes of 25◦ , 35◦ and 45◦ , respectively. However,
when J = 10 with a size of 2◦ , J s become 16.6, 22.9 and 29.2 for 25◦ , 35◦ and 45◦ ,
respectively. This implies that the large effect is mainly occurred for the dark colour
region.
Figure 2.13 shows the corrected attributes C of 25◦ , 35◦ and 45◦, respectively
plotted against C at 2◦ viewing field. Vertical axis is the size effect corrected C .
The thick solid line is the corrected C when viewing angle is 25◦ ; the dotted line
corresponds to the C with viewing angle being 35◦ . The dashed line is the C with
viewing angle of 45◦ . The thin solid line is the 45◦ line where C = C . Again, a clear
trend in Fig. 2.13 is shown that an increase of chroma for a larger viewing field. For
example, when C is 60 with a size of 2◦ , C values are 68.4, 73.2 and 78.0 for sizes
42 M.R. Luo and C. Li
of 25◦ , 35◦ and 45◦ , respectively. However, when C is 10 with a size of 2◦ , C s
become 11.4, 12.2 and 13.0 for 25◦ , 35◦ and 45◦, respectively. This implies that the
large effect in mainly occurrs in the high chroma region.
its own. Brightness and colourfulness were found to decrease with decreases of
both luminance level and stimulus size. The results were used to further extend
CIECAM02 for predicting unrelated colours under both photopic and mesopic
conditions. The model includes parameters to reflect the effects of luminance level
and stimulus size. The model is described below:
Inputs:
Measure or calculate the luminance L and chromaticity x,y of the test colour stimu-
lus corresponding to CIE colour-matching functions (2◦ or 10◦ ). The parameters are
the same as CIECAM02 except that the test illuminant is equal energy illuminant
(SE , i.e., XW = YW = ZW = 100), and LA = 1/5 of the adapting luminance, and the
surround parameters are set as those under the dark viewing condition. As reported
by Fu et al. [13], when there is no reference illuminant to compare with (such as
assessing unrelated colours), SE illuminant can be used by assuming no adaptation
takes place for unrelated viewing condition.
Step 1: Using the CIECAM02 (Steps 0–8, Step 10, ignore the calculation of Q and
s) to predict the (cone) achromatic signal A, colourfulness (M) and hue
(HC ).
Step 2: Modify the achromatic signal A since there is a contribution from rod
response using the formula:
Here, kA depends on luminance level and viewing angle size of the colour
stimulus.
Step 3: Modify the colourfulness M predicted from CIECAM02 using the follow-
ing formula:
Mnew = kM M.
Here, kM depends on luminance level and viewing angle size of the colour
stimulus.
Step 4: Predict the new brightness using the formula:
0.01 to 1000 cd/m2 , and LA was set at one fifth of these values. The ratio Yb /Yw set
at 0.2. Figure 2.15 shows the brightness and colourfulness changes, for the same red
colour, predicted by the new model for different stimulus sizes ranging from 0.2◦
to 40◦ . The luminance level (L) was set at 0.1 cd/m2 . It can be seen that brightness
and colourfulness increase when luminance increases up to around 100 cd/m2 , and
they also increase when stimulus size increases. These trends reflect the phenomena
found in Fu et al.’s study, i.e. when luminance level increases, colours become
brighter and more colourful, and larger colours appear brighter and more colourful
than smaller sized colours; however, below a luminance of 0.1 cd/m2 and above a
luminance of 60 cd/m2 , and below a stimulus size of 0.5◦ and above a stimulus size
of 100 , these results are extrapolations, and must be treated with caution.
2 CIECAM02 and Its Recent Developments 45
J = 100(A/Aw)cz
gives a problem for some colours. In fact Li and Luo [47] have shown that Aw > 0,
but for some colours, the achromatic signal
A = 2Ra + Ga + (1/20)Ba − 0.305 Nbb
can be negative; thus, the ratio in the bracket for the J function is negative which
gives problem when computing J. At the beginning, it has been suggested that the
source of the problem is the CAT02 transform which, for certain colours, predicts
negative tristimulus values. Several approaches have been made on modifying the
CAT02 matrix. Brill and Süsstrunk [48–50] found that the red and green CAT02
primaries lie outside the HPE triangle and called this as the “Yellow-Blue” problem.
They suggested that the last row of the CAT02 matrix can be changed to 0, 0, 1. The
changed matrix is denoted by MBS . It has been found that for certain colours, using
46 M.R. Luo and C. Li
matrix MBS works well, but using matrix M02 does not. However, this repair seems
to correct neither the prediction of negative tristimulus values for the CAT02 nor the
failure of CIECAM02.
Another suggestion is equivalent to set Ra ≥ 0.1, i.e., if Ra < 0.1, then set Ra =
0.1, if Ra ≥ 0.1, then Ra does not change. Similar considerations are applied to
Ga and Ba . Thus, under this modification, the achromatic signal A is non-negative.
However, this change causes new problem with the inverse model.
Li et al. [51] gave a mathematical approach for obtaining CAT02 matrix.
The approach has two constraints. The first one is to ensure the CAT02 predict
corresponding colours with non-negative tristimulus values under all the illuminants
considered for all colours located on or inside the CIE chromaticity locus. The
second one is to fit all the corresponding colour data sets. This approach indeed
ensures the CAT02 with the new matrix predicts corresponding colours with non-
negative tristimulus values which is important in many applications. However, this
approach does not solve the mathematical failure problem for the CIECAM02.
Recently, Li et al. [14] proposed a mathematical approach for ensuring the
achromatic signal A being non-negative, at the same time the CIECAM02 should
fit all the colour appearance data sets. Finally the problem is formulated as a
constrained non-linear optimisation problem. By solving the optimization problem,
a new CAT02 matrix was derived. With this new matrix, it was found that the
mathematical failure problem of the CIECAM02 is overcome for all the illuminants
considered. Besides, they also found that if the CAT02 with the HPE matrix, the
mathematical failure problem is also overcome for any illuminant. More important,
the HPE matrix makes the CIECAM02 simpler. All the new matrices are under the
evaluation of the CIE TC8-11.
The ICC has developed and refined a comprehensive and rigorous system for colour
management [52]. In an ICC colour management work flow, an input colour is
mapped from a device colour space into a colorimetric description for specific
viewing conditions (called the profile connection space—PCS). The PCS is selected
as either CIE XYZ or Lab space under illuminant D50 and the 2◦ observer.
Generally speaking, the input and output devices have different gamuts and, hence,
a gamut mapping is involved. Gamut mapping in XYZ space can cause problems
because of the perceptual non-uniformity of that colour space. Lab space is not
a good space for gamut mapping since lines of constant hue are not generally
straight lines, especially in the blue region [53]. CIECAM02 has been shown to
have a superior perceptual uniformity as well as better hue constancy [40]. Thus,
the CIECAM02 space has been selected as the gamut mapping space.
However, the ICC PCS can contain non-physical colours, which cause problems
when transforming to CIECAM02 space, for example, in the Lightness function J
defined above and the calculation of the parameter defined by
2 CIECAM02 and Its Recent Developments 47
Kuo et al. [55] found that the sum of the first row of the HPE matrix (eq. (12)) is
different from unity, which causes a non-zero value of a and b when transforming the
test light source to the reference (equal-energy) light source under full adaptation.
Hence, a slight change to the matrix should be made. For example, the top right
element −0.07868 could be changed to −0.07869. In fact, Kuo et al. [55] suggested
changing each element in the first row slightly.
2.7 Conclusion
This chapter describes the CIECAM02 in great details. Furthermore, more recent
works have been introduced to extend its functions. Efforts were made to reduce the
problems such as mathematical failure for the computation of the lightness attribute.
48 M.R. Luo and C. Li
Note if D is greater than one or less than zero, set it to one or zero,
respectively.
Yw Yw Yw
DR = D · + 1 − D, DG = D · + 1 − D, DB = D · + 1 − D,
Rw Gw Bw
FL = 0.2 k4 · (5LA ) + 0.1(1 − k4)2 · (5LA )1/3 ,
2 CIECAM02 and Its Recent Developments 49
where k = 1
5·LA +1 .
0.2
Yb √ 1
n= , z = 1.48 + n, Nbb = 0.725 · , Ncb = Nbb ,
Yw n
⎛ ⎞ ⎛ ⎞ ⎛ ⎞ ⎛ ⎞
Rwc DR · R w Rw Rwc
⎝ Gwc ⎠ = ⎝ DG · Gw ⎠ , ⎝ Gw ⎠ = MHPE · M −1 · ⎝ Gwc ⎠ ,
CAT02
Bwc DB · B w Bw Bwc
⎛ ⎞
0.7328 0.4296 −0.1624
MCAT02 = ⎝ −0.7036 1.6975 0.0061 ⎠ ,
0.0030 0.0136 0.9834
⎛ ⎞
0.38971 0.68898 − 0.07868
MHPE = ⎝ −0.22981 1.18340 0.04641 ⎠ ,
0.00000 0.00000 1.00000
⎛
0.42 ⎞
FL ·Rw
⎜ 100 ⎟
Raw = 400 · ⎝ 0.42 ⎠ + 0.1,
FL ·Rw
100 + 27.13
50 M.R. Luo and C. Li
⎛ ⎞
FL ·Gw 0.42
⎜ 100 ⎟
Gaw = 400 · ⎝ ⎠ + 0.1,
FL ·Gw 0.42
100 + 27.13
⎛ ⎞
FL ·Bw 0.42
⎜ 100 ⎟
Baw = 400 · ⎝ ⎠ + 0.1,
FL ·Bw 0.42
100 + 27.13
Baw
Aw = 2 · Raw + Gaw + − 0.305 · Nbb .
20
Note that all parameters computed in this step are needed for the following
calculations. However, they depend only on surround and viewing condi-
tions; hence, when processing pixels of image, they are computed once for
all. The following computing steps are sample dependent.
Step 1: Calculate (sharpened) cone responses (transfer colour-matching functions
to sharper sensors)
⎛ ⎞ ⎛ ⎞
R X
⎝ G ⎠ = MCAT02 · ⎝ Y ⎠ ,
B Z
If R is negative, then
2 CIECAM02 and Its Recent Developments 51
Table 2.4 Unique hue data Red Yellow Green Blue Red
for calculation of hue
quadrature i 1 2 3 4 5
hi 20.14 90.00 164.25 237.53 380.14
ei 0.8 0.7 1.0 1.2 0.8
Hi 0.0 100.0 200.0 300.0 400.0
⎛ 0.42 ⎞
−FL ·R
⎜ 100 ⎟
Ra = −400 · ⎝ 0.42 ⎠ + 0.1
−FL ·R
100 + 27.13
12 · Ga Ba
a = Ra − + ,
11 11
(R + Ga − 2 · Ba)
b= a ,
9
b
h = tan−1
a
Step 10: Calculate the correlates of chroma (C), colourfulness (M) and
saturation (s)
50000 1/2
· Nc · Ncb · et · a2 + b2
t= 13
,
Ra + Ga + 21
20 · Ba
J 0.5
C = t 0.9 · · (1.64 − 0.29n)0.73 ,
100
M = C · FL0.25 ,
0.5
M
s = 100 · .
Q
Input: J or Q; C, M or s; H or h
Output: X,Y, Z ( under test illuminant Xw ,Yw , Zw )
Illuminants, viewing surrounds and background parameters are the same as
those given in the forward mode. See notes at the end of this Appendix calculat-
ing/defining the luminance of the adapting field and surround conditions.
Step 0: Calculate viewing parameters
Compute all FL , n, z, Nbb = Nbc , Rw , Gw , Bw , D, DR , DG , DB , Rwc , Gwc , Bwc ,
Rw , Gw , Bw Raw , Gaw , Baw and Aw using the same formulae as in Step 0 of
the Forward model. They are needed in the following steps. Note that all
data computed in this step can be used for all samples (e.g., all pixels for an
image) under the viewing conditions. Hence, they are computed once for
all. The following computing steps are sample dependent.
Step 1: Obtain J, C and h from H, Q, M, s
The entering data can be in different combination of perceived correlates,
i.e., J or Q; C, M, or s; and H or h. Hence, the followings are needed to
convert the others to J, C, and h.
Step 1–1: Compute J from Q (if start from Q)
2
c·Q
J = 6.25 · .
(Aw + 4) · FL0.25
2 CIECAM02 and Its Recent Developments 53
M
C= (if start from M)
FL0.25
4 J 0.5
Q= · · (Aw +4.0) · FL0.25
c 100
s 2
Q
and C = 100 · ( F 0.25 ) (if start from s)
L
Step 1–3: Calculate h from H (if start from H)
The correlate of hue (h) can be computed by using data in Table 2.4 in
the Forward mode.
Choose a proper i (i = 1,2,3 or 4) so that Hi ≤ H < Hi+1 .
p1
p4 = ,
sin(h)
460
p2 · (2 + p3) · 1403
b= 220 cos(h) 27 ,
p4 + (2 + p3) · 1403 · sin(h) − 1403 + p3 · 6300
1403
cos(h)
a = b· .
sin(h)
Step 8: Calculate X, Y and Z (for the coefficients of the inverse matrix, see the note
at the end of the Appendix)
⎛ ⎞ ⎛ ⎞
X R
⎝ Y ⎠ = M −1 · ⎝ G ⎠ .
CAT02
Z B
Notes to Appendix
1. It is recommended to use the matrix coefficients given below for the inverse
−1 −1
matrix MCAT02 and MHPE :
⎛ ⎞
1.096124 −0.278869 0.182745
−1
MCAT02 = ⎝ 0.454369 0.473533 0.072098 ⎠ ,
−0.009628 −0.005698 1.015326
⎛ ⎞
1.910197 −1.112124 0.201908
M −1 = ⎝ 0.370950 0.629054 −0.000008 ⎠
HPE
0.000000 0.000000 1.000000
2. For implementing the CIECAM02, the testing data and the corresponding results
from the forward and reverse modes can be found from reference 7.
3. The LA is computed using (2.11)
EW Yb LW ·Yb
LA = · = , (2.11)
π YW YW
where Ew = π ·Lw is the illuminance of reference white in lux unit; Lw the luminance
of reference white in cd/m2 unit, Yb the luminance factor of the background and Yw
the luminance factor of the reference white.
References
1. Luo MR (1999) Colour science: past, present and future. In: MacDonald LW and Luo MR
(Eds) Colour imaging: vision and technology. Wiley, New York, 384–404
2. CIE Technical Report (2004) Colorimetry, 3rd ed. Publication 15:2004, CIE Central Bureau,
Vienna.
56 M.R. Luo and C. Li
3. Luo MR, Cui GH, Rigg B (2001) The development of the CIE 2000 colour difference formula.
Color Res Appl 26:340-350.
4. Luo MR, Hunt RWG (1998) The structure of the CIE 1997 colour appearance model
(CIECAM97s). Color Res Appl 23:138–146
5. CIE (1998) The CIE 1997 interim colour appearance model (simple version), CIECAM97s.
CIE Publication 131, CIE Central Bureau, Vienna, Austria.
6. Moroney N, Fairchild MD, Hunt RWG, Li C, Luo MR, Newman T (2002) The CIECAM02
color appearance model, Proceedings of the 10th color imaging conference, IS&T and SID,
Scottsdale, Arizona, 23–27
7. CIE (2004) A colour appearance model for colour management systems: CIECAM02, CIE
Publication 159 CIE Central Bureau, Vienna, Austria
8. Luo MR and Li CJ (2007) CIE colour appearance models and associated colour spaces, Chapter
11 of the book: colorimetry-understanding the CIE System. In: Schanda J (ed) Wiley, New York
9. Luo MR, Cui GH, Li CJ and Rigg B (2006) Uniform colour spaces based on CIECAM02
colour appearance model. Color Res Appl 31:320–330
10. Xiao K, Luo MR, Li C, Hong G (2010) Colour appearance prediction for room colours, Color
Res Appl 35:284–293
11. Xiao K, Luo MR, Li CJ, Cui G, Park D (2011) Investigation of colour size effect for colour
appearance assessment, Color Res Appl 36:201–209
12. Xiao K, Luo MR, Li CJ (2012) Color size effect modelling, Color Res Appl 37:4–12
13. Fu CY, Li CJ, Luo MR, Hunt RWG, Pointer MR (2007) Quantifying colour appearance for
unrelated colour under photopic and mesopic vision, Proceedings of the 15th color imaging
conference, IS&T and SID, Albuquerque, New Mexico, 319–324
14. Li CJ, Chorro-Calderon E, Luo MR, Pointer MR (2009) Recent progress with extensions to
CIECAM02, Proceedings of the 17th color imaging conference, IS&T and SID, Albuquerque,
New Mexico 69–74
15. CIE Publ. 17.4:1987, International lighting vocabulary, the 4th edition
16. Mori L, Sobagaki H, Komatsubara H, Ikeda K (1991) Field trials on CIE chromatic adaptation
formula. Proceedings of the CIE 22nd session, 55–58
17. McCann JJ, McKee SP, Taylor TH (1976) Quantitative studies in Retinex theory: a comparison
between theoretical predictions and observer responses to the ‘color mondrian ’ experiments.
Vision Res 16:445–458
18. Breneman EJ (1987) Corresponding chromaticities for different states of adaptation to complex
visual fields. J Opt Soc Am A 4:1115–1129
19. Helson H, Judd DB, Warren MH (1952) Object-color changes from daylight to incandescent
filament illumination. Illum Eng 47:221–233
20. Lam KM (1985) Metamerism and colour constancy. Ph.D. thesis, University of Bradford, UK
21. Braun KM, Fairchild MD (1996) Psychophysical generation of matching images for cross-
media colour reproduction. Proceedings of 4th color imaging conference, IS&T, Springfield,
Va., 214–220
22. Luo MR, Clarke AA, Rhodes PA, Schappo A, Scrivener SAR, Tait C (1991) Quantifying colour
appearance. Part I. LUTCHI colour appearance data. Color Res Appl 16:166–180
23. Luo MR, Gao XW, Rhodes PA, Xin HJ, Clarke AA, Scrivener SAR (1993) Quantifying colour
appearance, Part IV: transmissive media. Color Res Appl 18:191–209
24. Kuo WG, Luo MR, Bez HE (1995) Various chromatic adaptation transforms tested using new
colour appearance data in textiles. Color Res Appl 20:313–327
25. Juan LY, Luo MR (2000) New magnitude estimation data for evaluating colour appearance
models. Colour and Visual Scales 2000, NPL, 3-5 April, UK
26. Juan LY, Luo MR (2002) Magnitude estimation for scaling saturation. Proceedings of 9th
session of the association internationale de la couleur (AIC Color 2001), Rochester, USA,
(June 2001), Proceedings of SPIE 4421, 575–578
27. Li CJ, Luo MR, Rigg B, Hunt RWG (2002) CMC 2000 chromatic adaptation transform:
CMCCAT2000. Color Res Appl 27:49–58
2 CIECAM02 and Its Recent Developments 57
28. Judd DB (1940), Hue, saturation, and lightness of surface colors with chromatic illumination.
J Opt Soc Am 30:2–32
29. Kries V (1902), Chromatic adaptation, Festschrift der Albrecht-Ludwig-Universitat (Fribourg),
[Translation: MacAdam DL, Sources of Color Science, MIT Press, Cambridge, Mass. (1970)]
30. Luo MR, Hunt RWG (1998) A chromatic adaptation transform and a colour inconstancy index.
Color Res Appl 23:154–158
31. Li CJ, Luo MR, Hunt RWG (2000) A revision of the CIECAM97s Model. Color Res Appl
25:260–266
32. Hunt RWG, Li CJ, Juan LY, Luo MR (2002), Further improvements to CIECAM97s. Color
Res Appl 27:164–170
33. Finlayson GD, Süsstrunk S (2000) Performance of a chromatic adaptation transform based on
spectral sharpening. Proceedings of IS&T/SID 8th color imaging conference, 49–55
34. Hunt RWG (1952) Light and dark adaptation and perception of color. J Opt Soc Am
42:190–199
35. Stevens JC, Stevens SS (1963) Brightness functions: effects of adaptation. J. Opt Soc Am
53:375–385
36. Bartleson CJ, Breneman EJ (1967) Brightness perception in complex fields. J. Opt Soc Am
57:953–957
37. Luo MR, Gao XW, Sciviner SAR (1995) Quantifying colour appearance, Part V, Simultaneous
contrast. Color Res Appl 20:18–28
38. Wyszecki G, Stiles WS (1982) Color Science: concepts and methods, Quantitative data and
formulae. Wiley, New York
39. Helson H (1938) Fundamental problems in color vision. I. The principle governing changes in
hue, saturation, and lightness of non-selective samples in chromatic illumination. J Exp Psych
23:439–477
40. CIE Publ. 152:2003, Moroney N, Han Z (2003) Field trials of the CIECAM02 colour
appearance, Proceedings of the 25th session of the CIE, San Diego D8-2–D8-5.
41. Tastl I, Bhachech M, Moroney N, Holm J (2005) ICC colour management and CIECAM02,
Proceedings of the 13th of CIC, p 318
42. Gury R, Shaw M (2005) Dealing with imaginary color encodings in CIECAM02 in an ICC
workflow. Proceedings of the 13th of CIC, pp 217–223
43. Li CJ, Luo MR, Cui GH (2003) Colour-difference evaluation using colour appearance models.
The 11th Color Imaging Conference, IS&T and SID, Scottsdale, Arizona, November, 127–131
44. Luo MR, Rigg B (1986) Chromaticity–discrimination ellipses for surface colours. Color Res
Appl 11:25–42
45. Berns RS, Alman DH, Reniff L, Snyder GD, Balonon-Rosen MR (1991) Visual determi-
nation of suprathreshold color-difference tolerances using probit analysis. Color Res Appl
16:297–316
46. Hunt RWG (1952) Measuring colour, 3rd edition, Fountain Press, Kingston-upon-Thames,
1998
47. Li CJ, Chorro-Calderon E, Luo MR, Pointer MR (2009) Recent progress with extensión to
CIECAM02, Seventeenth Colour Imaging Conference, Final Program and Proceedings, 69–74
48. Brill MH (2006) Irregularity in CIECAM02 and its avoidance. Color Res Appl 31(2):142–145
49. Brill MH, Susstrunk S (2008) Repairing gamut problems in CIECAM02: a progress report.
Color Res Appl 33(5):424–426
50. Süsstrunk S, Brill M (2006) The nesting instinct: repairing non nested gamuts in CIECAM02.
14th SID/IS&T color imaging conference
51. Li CJ, Perales E, Luo MR, Martı́nez-Verdú F, A Mathematical approach for predicting non-
negative tristimulus values using the CAT02 chromatic adaptation transform, Color Res Appl
(in press)
52. ISO 15076-1 (2005) Image technology, colour management-Architecture, profile format and
data structure-Part I: based on ICC.1:2004-10, [Link]
53. Moroney N (2003) A hypothesis regarding the poor blue constancy of CIELAB. Color Res
Appl 28(5):371–378
58 M.R. Luo and C. Li
54. W. Gill GW (2008) A solution to CIECAM02 numerical and range issues, Proceedings of the
16th of color imaging conference, IS&T and SID, Portland, Oregan 322–327
55. Kuo CH, Zeise E, Lai D (2006) Robust CIECAM02 implementation and numerical experiment
within an ICC workflow. Proceedings of the 14th of CIC, pp 215–219
56. Hunt RWG, Li CJ, Luo MR (2002) Dynamic cone response functions for modes of colour
appearance. Color Res Appl 28:82–88
57. Paula J, Alessi P (2008) Private communication pursuit of scales corresponding to equal
perceptual brightness, personal correspondence
Chapter 3
Colour Difference Evaluation
M. Melgosa ()
Departamento de Optica, Facultad de Ciencias, Universidad de Granada, Spain
e-mail: mmelgosa@[Link]
A. Trémeau
Laboratory Hubert Curien, UMR CNRS 5516, Jean Monnet University, Saint-Etienne, France
e-mail: [Link]@[Link]
G. Cui
VeriVide Limited, Leicester, LE19 4SG, United Kingdom
e-mail: [Link]@[Link]
3.1 Introduction
From two homogeneous colour stimuli, we can ask ourselves what is the magnitude
of the perceived colour difference between them. Of course, this question may
also be asked in the case of more complex stimuli like two colour images.
In fact, to achieve a consistent answer to the previous question, we must first
specify the experimental observation conditions: for example, size of the stimuli,
background behind them, illuminance level, etc. It is well known that experimental
illuminating and viewing conditions (the so-called “parametric effects”) play an
important role on the magnitude of perceived colour differences, as reported by
the International Commission on Illumination (CIE) [1]. Specifically, to avoid the
spread of experimental results under many different observation conditions, in 1995
the CIE proposed [2] to analyze just 17 “colour centers” well distributed in colour
space (Table 3.1), under a given set of visual conditions similar to those usually
found in industrial practice, which are designated as “reference conditions,” and are
as follows:
Illumination: D65 source
Illuminance: 1000 lx
Observer: Normal colour vision
Background field: Uniform, neutral grey with L∗ = 50
While ΔV is the result of a subjective measurement like the average of the visual
assessments performed by a panel of observers, using a specific method and working
under fixed observation conditions, ΔE is an objective measurement which can be
currently performed using colorimetric instrumentation. Obviously, the main goal is
to achieve a ΔE analogous to ΔV for any colour pair in colour space and under any
visual set of observational conditions. In this way, complex tasks like visual pass/fail
decisions in a production chain could be done in a completely automatic way
62 M. Melgosa et al.
Fig. 3.1 A yellow colour pair of textile samples, together with a grey scale for visual assessment
of the colour difference in such a pair. A colour mask may be employed to choose a colour pair
in the grey scale, or to have a test pair with the same size than those in the grey scale. Photo from
Dr. Michal Vik, Technical University of Liberec, Czech Republic
Fig. 3.2 Visual versus instrumental color-difference evaluation: example of quality control using
a colorimeter. Photo from “Precise Color Communication”, Konica-Minolta Sensing, Inc., 1998
(Fig. 3.2). However, it must be recognized that this is a very ambitious goal, because
in fact it is intended to predict the final answer of our visual system, currently
unknown in many aspects. Anyway, important advances have been produced in
colour-difference measurement, as will be described in the next section.
3 Colour Difference Evaluation 63
∗
with visual judgments. While the CIELAB colour-difference formula ΔEab had the
advantage that it was very similar to the Adams–Nickerson (ANLAB40) formula, al-
ready adopted by several national industrial groups, the CIELUV colour-difference
∗
formula ΔEuv had the advantage of a linear chromaticity diagram, particularly
useful in lighting applications. It can be said that the CIELAB colour-difference
formula was soon accepted by industry: while in 1977 more than 20 different
colour-difference formulas were employed in the USA industry, 92% of these
industries had adopted CIELAB in 1992 [23]. Because there are no fixed scale
factors between the results provided by two different colour-difference formulas,
the uniformity of practice (standardization) achieved by CIELAB was an important
achievement for industrial practice. It should be said that a colour difference
between 0.4 and 0.7 CIELAB units is approximately a just noticeable or threshold
difference, although even lower values of colour differences are sometimes managed
by specific industries. Colour differences between contiguous samples in colour
atlases (e.g., Munsell Book of Color) are usually greater than 5.0 CIELAB units,
being designated as large colour differences.
After the proposal of CIELAB, many CIELAB-based colour-difference for-
mulas were proposed with considerable satisfactory results [24]. Among these
CIELAB-based formulas, it is worth mentioning the CMC [25] and BFD [26]
colour-difference formulas. The CMC formula was recommended by the Colour
Measurement Committee of the Society of Dyers and Colourists (UK), and inte-
grated into some ISO standards. CIELAB lightness, chroma, and hue differences
are properly weighted in the CMC formula, which also includes parametric factors
dependent on visual conditions (e.g., the CMC lightness differences have half value
for textile samples). In 1995 the CIE proposed the CIE94 colour-difference formula
[27], which may be considered a simplified version of CMC. CIE94 was based
on most robust trends in three reliable experimental datasets, proposing simple
corrections to CIELAB (linear weighting functions of the average chroma for the
CIELAB chroma and hue differences), as well as parametric factors equal to 1.0
under the so-called “reference conditions” (see Introduction). It can be said that
CIE94 adopted a versatile but too conservative approach adopting only the most
well-known CIELAB corrections, like the old chroma-difference correction already
suggested by McDonald for the ANLAB formula in 1974 [28].
In 2001 the CIE recommended its last colour-difference formula, CIEDE2000
[29]. From a combined dataset of reliable experimental data containing 3,657 colour
pairs from four different laboratories, the CIEDE2000 formula was developed [30].
The CIEDE2000 formula has the same final structure as the BFD [26] formula.
Five corrections to CIELAB were included in CIEDE2000: A weighting function
for lightness accounting for the “crispening effect” produced by an achromatic
background with lightness L∗ = 50; a weighting function for chroma identical to
the one adopted by the previous CIE94 formula; a weighting function for hue which
is dependent on both hue and chroma; a correction of the a∗ coordinate for neutral
colours; and a rotation term which takes account of the experimental chroma and
hue interaction in the blue region. The most important correction to CIELAB in
CIEDE2000 is the chroma correction [31]. CIEDE2000 also includes parametric
3 Colour Difference Evaluation 65
First, for each one of the two colour samples, designated as “b” (“batch”) and “s”
(“standard”), a localized modification of the CIELAB coordinate a∗ is made:
L = L∗ (3.3)
∗
a = (1 + G) a (3.4)
b = b∗ (3.5)
⎛ ⎞
∗7
Cab
G = 0.5 ⎝1 − ⎠ (3.6)
∗7 + 257
Cab
where the upper bar means arithmetical mean of standard and batch. Transformed
a , b are used in calculations of transformed chroma and hue angle, in the usual
way [21]:
!
C = a 2 + b 2 (3.7)
h = arctan ba (3.8)
SC = 1 + 0.045C (3.14)
66 M. Melgosa et al.
SH = 1 + 0.015 C T (3.15)
T = 1 − 0.17 cos h − 30◦ + 0.24 cos 2h
+0.32 cos 3h + 6◦ − 0.20 cos 4h − 63◦ (3.16)
120
b*
90
60
30
a*
0
−30
−60
−60 −30 0 30 60 90 120
Fig. 3.3 Experimental colour discrimination ellipses in CIELAB a∗ b∗ for the BFD and
RIT-DuPont datasets (red), compared with predictions made by the CIEDE2000 colour-difference
formula (black) [30]
where the symbol “Δ” indicates differences between batch and standard samples in
the colour pair. For each one of the two samples (and also the reference white), the
next equations based on CIELAB L∗ , a∗ , b∗ coordinates are applied:
68 M. Melgosa et al.
where the new e and f coordinates are the result of a rotation and re-scaling of the
CIELAB a∗ b∗ coordinates, and L99d is not too different to CIELAB lightness L∗ .
!
G= e2 + f 2 (3.25)
C99d = 22.5 ln (1 + 0.06G) (3.26)
∗ . Finally:
where this new chroma C99d is a compression of CIELAB chroma Cab
In 2006, on the basis of the CIECAM02 colour appearance model [14], three new
Euclidean colour-difference formulas were proposed [37] for small (CAM02-SCD),
large (CAM02-LCD), and all colour differences (CAM02-UCS). In these CAM02
formulas a non-linear transformation to CIECAM02 lightness J, and a logarithmic
compression to the CIECAM02 colourfulness M were applied. The corresponding
equations are as follows:
$ 2
ΔECAM02 = Δ J KL + (Δa )2 + (Δb )2 (3.30)
(1 + 100 c1) J
J = (3.31)
1 + c1 J
M = (1/c2 ) ln (1 + c2M) (3.32)
a = M cos (h) (3.33)
b = M sin (h) (3.34)
where J, M, and h are the CIECAM02 lightness, colourfulness, and hue angle values,
respectively. In addition, the ΔJ , Δa , and Δb are the J , a , and b differences
between the standard and batch in a colour pair. Finally, the parameter KL has
values 0.77, 1.24, and 1.00 for the CAM02-LCD, CAM02-SCD, and CAM02-UCS
formulas, respectively, while c1 = 0.007 for all these formulas, and c2 has values
0.0053, 0.0363, and 0.0228, for the CAM02-LCD, CAM02-SCD, and CAM02-UCS
formulas, respectively [37]. The results achieved by these CAM02 formulas are very
encouraging: embedded uniform colour space in the CIECAM02 colour appearance
3 Colour Difference Evaluation 69
model can be useful to make successful predictions of colour differences; that is,
colour difference may be a specific aspect of colour appearance. Berns and Xue
have also proposed colour-difference formulas based on the CIECAM02 colour
appearance model [38].
OSA-UCS is a noteworthy empirical colour system for large colour differences
developed in 1974 by the Optical Society of America’s committee on Uniform Color
Scales [39]. In this system, the straight lines radiating from any colour sample are
geodesic lines with uniform colour scales. Thus, OSA-UCS was adopted to develop
a CIE94-type colour-difference formula, valid under D65 illuminant and CIE
1964 colorimetric observer [40]. This formula was latter refined with chroma and
lightness compressions, achieving an Euclidean colour-difference formula based
also on the OSA-UCS space [41]. The equations conducting to this Euclidean
formula, noted as ΔEE , are as follows:
ΔEE = (ΔLE )2 + (ΔGE )2 + (ΔJE )2 (3.35)
LE = b1L ln 1 + baLL (10 LOSA ) with aL = 2.890, bL = 0.015 (3.36)
with J and G coordinates defined, for the D65 illuminant, from the transformations:
J 2 (0.5735 LOSA + 7.0892) 0
=
G 0 −2 (0.7640 LOSA + 9.2521)
70 M. Melgosa et al.
⎛ ⎞
ln A/B
0.1792 0.9837 ⎜ 0.9366 ⎟
⎝ ⎠ (3.44)
0.9482 −0.3175 ln B/C
0.9807
⎛ ⎞ ⎡ ⎤⎛ ⎞
A 0.6597 0.4492 −0.1089 X10
⎝ B ⎠ = ⎣ −0.3053 1.2126 0.0927 ⎦ ⎝ Y10 ⎠ (3.45)
C −0.0374 0.4795 0.5579 Z10
where C indicates the arithmetical average of the chroma of the two samples in the
colour pair.
Finally, Berns’ models adopted the Euclidean distance as the measure of colour
differences, and all the previous parameters were optimized [13] to achieve a
minimum deviation between visual and computed colour differences for the RIT-
DuPont dataset [6]:
ΔE E = (ΔLE )2 + (ΔaE )2 + (ΔbE )2 (3.53)
Recently, Shen and Berns [43] have developed an Euclidean colour space IPT-
EUC, claiming it as a potential candidate for a unique colour model for both
describing colour and measuring colour differences.
Euclidean colour spaces can be also developed by either analytical or compu-
tational methods to map the non-linear, non-uniform colour spaces to linear and
uniform colour spaces based on the clues provided by different colour difference
formulas optimized for reliable experimental datasets [44, 45].
PF/3 is a combined index which was proposed by Guan and Luo [7] from
previous metrics suggested by Luo and Rigg [26], which in turn employed the γ and
CV metrics proposed by Alder et al. [46] and the VAB metric proposed by Shultze
[47]. The corresponding defining equations are as follows:
' (2
1 N
ΔEi ΔEi
log10 (γ ) = ∑ log10
N i=1
− log10
ΔVi ΔVi
(3.54)
N
2 ∑ ΔEi /ΔVi
1 N
(ΔEi − FΔVi ) i=1
VAB = ∑
N i=1 ΔEi FΔVi
with F =
N
(3.55)
∑ ΔVi /ΔEi
i=1
N
∑ ΔEi ΔVi
1 N
(ΔEi − f ΔVi ) 2
∑
i=1
CV = 100 2 with f = N
(3.56)
N i=1 ΔEi ∑ ΔVi2
i=1
100 (γ − 1) + VAB + CV
PF/3 = 100
(3.57)
3
where N indicates the number of colour pairs (with visual and computed differences
ΔVi and ΔEi , respectively), F and f are factors adjusting the ΔEi and ΔVi values
to the same scale, and the upper bar in a variable indicates the arithmetical mean.
For perfect agreement between ΔEi and ΔVi , CV and VAB should equal zero and γ
should equal one, in such a way that PF/3 should equal zero. A higher PF/3 value
indicates worse agreement. Guan and Luo [7] state that PF/3 gives roughly the
typical error in the predictions of ΔVi as a percentage: for example, a 30% error in
all pairs corresponds approximately to γ of 1.3, VAB of 0.3, and CV of 30 leading to
PF/3 of 30.
The decimal logarithm of γ is the standard deviation of the log10 (ΔEi /ΔVi ). This
metric was adopted because ΔEi values should be directly proportional to ΔVi , and
the ratio ΔEi /ΔVi should be constant. The standard deviation of the values of this
ratio could be used as a measure of agreement, but this would give rise to anomalies
which can be avoided by considering the logarithms of the ΔEi /ΔVi values [25].
Natural logarithms have sometimes been employed to define the γ index, but the
standard version of PF/3 uses decimal logarithms. The VAB and CV values express
the mean square root of the ΔEi values with respect to the ΔVi values (scaled by the
F or f coefficients), normalized to appropriate quantities. Therefore, the VAB and
CV indices could be interpreted as two coefficients of variations, and the F and f
factors as slopes of the plot of ΔEi against ΔVi (although they are not exactly the
slope of the linear-regression fit). In earlier papers [25, 26], the product-moment
correlation coefficient r was also employed as another useful measure to test the
relationship between ΔEi and ΔVi . However, the r coefficient was not included in
3 Colour Difference Evaluation 73
final PF/3 definition because it was found to be quite inconsistent with the other
three indices for different experimental and theoretical datasets [7,16,47]. The main
reason to propose the PF/3 index was that sometimes different measures led to
different conclusions; for example, one formula performed the best according to
CV while using VAB other different formula provided the most accurate prediction.
Thus, it was considered useful to avoid making a decision as to which of the metrics
was the best, and provide a single value to evaluate the strength of the relationship
between ΔEi and ΔVi [16]. Anyway, although PF/3 was widely employed in recent
colour-difference literature, other indices have been also employed in this field.
For example, the “wrong decision” percentage [48] is employed in acceptability
experiments, the coefficient of variation of tolerances was employed by Alman
et al. [49] and the linear correlation coefficients also continue being used by some
researchers [50, 51].
Any flaw in γ , CV, or VAB is immediately transferred to PF/3, which is by
definition an eclectic index. In addition, PF/3 cannot be used to indicate the
significance of the difference between two colour-difference formulas with respect
to a given set of visual data, because the statistical distribution followed by PF/3 is
unknown. This last point is an important shortcoming for the PF/3 index, because
the key question is not just to know that a colour-difference formula has a lower
PF/3 than other for a given set of reliable visual data, but to know whether these
two colour-difference formulas are or not statistically significant different for these
visual data. From the scientific point of view, it is not reasonable to propose a new
colour-difference formula if it is not significantly better than previous formulas, for
different reliable visual datasets. In addition, industry is reluctant to change colour-
difference formulas they are familiar to, in such a way that these changes must be
based on the achievement of statistically significant improvements.
In a recent paper [52] the STRESS index has been suggested as a good alternative
to PF/3 for colour-difference evaluation. STRESS comes from multidimensional
scaling [53], and is defined as follows:
) *
2 1/2
∑ (ΔEi − F1ΔVi ) ∑ ΔEi2
STRESS = 100 with F1 = (3.58)
∑ F1 ΔVi
2 2
∑ ΔEi ΔVi
OSA-GP
CAM02-UCS
CAM02-SCD
DIN99d
CIEDE2000
CIE94
CMC
CIELAB
0 10 20 30 40
STRESS
Fig. 3.4 Computed STRESS values using different colour-difference formulas for the combined
dataset employed at CIEDE2000 development [30]
have been suggested [56] concluding that, for the experimental dataset employed at
CIEDE2000 development [29, 30], only a few colour pairs with very small colour
differences have a low degree of consistency.
Most complex images are not made up of large uniform fields. Therefore,
discrimination and appearance of fine patterned colour images differ from similar
measurements made using large homogeneous fields [57]. Direct application
of previously mentioned colour-difference formulas to predict complex image
difference (e.g., using a simple pixel by pixel comparison method) does not give
satisfactory results. Colour discrimination and appearance is a function of spatial
pattern. In general, as the spatial frequency of the target goes up (finer variations
in space), colour differences become harder to see, especially differences along the
blue-yellow direction. So, if we want to apply a colour-difference formula to colour
images, the patterns of the image have to be taken into account.
Different spatial colour-difference metrics have been suggested, the most famous
one being the one proposed by Zhang and Wandell [58] in 1996, known as S-
CIELAB. S-CIELAB is a “perceptual colour fidelity” metric. It measures how
accurate the reproduction of a colour image is to the original when viewed by a
human observer. S-CIELAB is a spatial extension of CIELAB, where two input
images are processed like made in the human visual system, before conventional
CIELAB colour differences are applied pixel by pixel. Specifically, the steps
followed by S-CIELAB are as follows: (1) each pixel (X,Y, Z) in the input images
is translated to an opponent colour space, consisting of one luminance and two
chrominance components; (2) each one of these three components is passed through
a spatial filter that is selected according to the spatial sensitivity of the human visual
system to this component, taken into account visual conditions; (3) the filtered
images are transformed back into the CIE X,Y, Z format; (4) finally, the colour
differences can be computed using the conventional CIELAB colour-difference
formula, and the average of these colour differences for all pixels could then be
used to represent the difference between two complex images. In fact, this idea can
be applied using at the end any colour-difference formula; for example, in 2003
Johnson and Fairchild [59] applied the S-CIELAB framework replacing CIELAB
by the CIEDE2000 colour-difference formula. Recently, Johnson et al. [60] have
also pointed out that, for image difference calculations, the ideal opponent colour
space would be both linear and orthogonal, such that the linear filtering is correct
and any spatial processing on one channel does not affect the others, proposing a
new opponent colour space and corresponding spatial filters specifically designed
for image colour-difference calculations.
The evaluation of colour differences in complex images requires the corre-
sponding images be carefully selected, as suggested by standardization organisms,
avoiding potential bias from some kind of images [61]. Experimental methods
76 M. Melgosa et al.
employed to compare image quality must also be carefully considered [62]. While
some results indicate a clear advantage of S-CIELAB with respect to CIELAB
analyzing colour differences in complex images [63], other results [64] suggest no
clear improvements using spatial colour-difference models, and results dependent
on image content. Recent CIE Publication 199–2011 [65] provides useful informa-
tion related to methods for evaluating colour differences in images.
Colour differences have been an active field of research since the 1950s trying
to respond to industrial requirements in important topics like colour control,
colour reproduction, etc. CIE-proposed colour-difference formulas have played an
important positive role in the communication between buyers and sellers, as well as
among different industries. The CIE recommendations of CIE94 and CIEDE2000
colour-difference formulas in 1995 and 2001, respectively, are eloquent examples
of significant work and advances in this scientific area. Currently, research on
colour differences continues, in particular, within some CIE Technical Committees
in Divisions 1 and 8, as shown by the following examples: CIE TC1–55 (chairman:
M. Melgosa) is working on the potential proposal of a uniform colour space
for industrial colour-difference evaluation; CIE TC1–57 (chairman: A. Robertson)
“Standards in colorimetry” has proposed the CIEDE2000 colour-difference formula
as a CIE standard; CIE TC1–63 (chairman: K. Richter) has studied the range of
validity of the CIEDE2000 colour-difference formula, concluding with the proposal
of the new CIE TC1–81 (chairman: K. Richter) to analyze the performance of
colour-difference formulas for very small colour differences (visual thresholds); CIE
TC8–02 (chairman: M.R. Luo) studied colour differences in complex images [65].
Another important aspect in colour-difference research is the need of new
reliable experimental datasets which can be used to develop better colour-difference
formulas. New careful determinations of visual colour differences under well-
defined visual conditions, together with their corresponding uncertainties, are highly
desirable [66]. At the same time it is also very convenient to avoid an indiscriminate
use of new colour-difference formulas, which should affect negatively industrial
colour communication. New colour-difference formulas are only interesting if they
can prove a statistically significant improvement with respect to previous ones, for
several reliable experimental datasets.
There is an increasing activity aimed at incorporating colour-appearance models
into practical colour-difference specification. For example, a colour appearance
model could incorporate the effects of the background and luminance level on
colour-difference perception, in such a way that the associated colour-difference
formula could be applied to a wide set of visual conditions, in place of just a
given set of “reference conditions”. A colour appearance model would also make
it possible to directly compare colour differences measured for different viewing
conditions or different observers. Colour appearance models would also make it
3 Colour Difference Evaluation 77
References
1. CIE Publication 101 (1993) Parametric effects in colour-difference evaluation. CIE Central
Bureau, Vienna
2. Witt K (1995) CIE guidelines for coordinated future work on industrial colour-difference
evaluation. Color Res Appl 20:399–403
3. Kuehni RG (2009) Variability in estimation of suprathreshold small color differences. Color
Res Appl 34:367–374
4. MacAdam DL (1942) Visual sensitivities to color differences in daylight. J Opt Soc Am
32:247–274
5. Shen S, Berns RS (2011) Color-difference formula performance for several datasets of small
color differences based on visual uncertainty. Color Res Appl 36:15–26
6. Berns RS, Alman DH, Reniff L, Snyder GD, Balonon-Rosen MR (1991) Visual determi-
nation of suprathreshold color-difference tolerances using probit analysis. Color Res Appl
16:297–316
7. Guan S, Luo MR (1999) Investigation of parametric effects using small colour-differences.
Color Res Appl 24:331–343
8. ISO 105-A02:1993 Tests for Colour Fastness-Part A02: Gray Scale for Assessing Change
in Colour, International Organization for Standardization Geneva, Switzerland. [Link]
[Link]
9. AATCC Committee RA36, AATCC Evaluation Procedure 1 (2007) Gray scale for color
change. AATCC, NC, Research Triangle Park. [Link]
10. Fastness Tests Co-ordinating Committee (F.T.C.C.) Publication XI (1953) The development of
the geometric grey scales for fastness assessment. J Soc Dyers Colour 69:404–409
11. Cárdenas LM, Shamey R, Hinks D (2009) Development of a novel linear gray scale for visual
assessment of small color differences. AATCC Review 9:42–47
12. Montag ED, Wilber DC (2003) A comparison of color stimuli and gray-scale methods of color
difference scaling. Color Res Appl 28:36–44
13. Berns RS (2008) Generalized industrial color-difference based on multi-stage color vision and
line-element integration. Óptica Pur Appl 41:301–311
78 M. Melgosa et al.
14. CIE Publication 159:2004 (2004) A colour appearance model for colour management systems:
CIECAM02. CIE Central Bureau, Vienna
15. Fairchild MD (2005) Colour Appearance Models, 2nd edn. Wiley, New York
16. Luo MR (2002) Development of colour-difference formulae. Rev Prog Color 32:28–39
17. McDonald R (1982) A review of the relationship between visual and instrumental assessment
of colour difference, part 1. J Oil Colour Chem Assoc 65:43–53
18. McDonald R (1982) A review of the relationship between visual and instrumental assessment
of colour difference, part 2. J Oil Colour Chem Assoc 65:93–106
19. Wit K (2007) CIE color difference metrics. In: Schanda J (ed) Chapter 4 in Colorimetry-
Understanding the CIE System, Wiley, New York
20. CIE Publication 13.3 (1995) Method of measuring and specifying colour rendering properties
of light sources. CIE Central Bureau, Vienna
21. CIE 15:2004 (2004) Colorimetry, 3rd edn. CIE Central Bureau, Vienna
22. Robertson AR (1990) Historical development of CIE recommended color difference equations.
Color Res Appl 15:167–170
23. Kuehni RG (1990) Industrial color-difference: progress and problems. Color Res
Appl 15:261–265
24. Melgosa M (2000) Testing CIELAB-based color-difference formulas. Color Res
Appl 25:49–55
25. Clarke FJJ, McDonald R, Rigg B (1984) Modification to the JPC79 colour-difference formula.
J Soc Dyers Colour 100:128–132
26. Luo MR, Rigg B (1987) BFD(l:c) colour-difference formula. Part 1 – Development of the
formula. J Soc Dyers Colour 103:86–94
27. CIE Publication 116 (1995) Industrial colour-difference evaluation. CIE Central Bureau,
Vienna
28. McDonald R (1974) The effect of non-uniformity in the ANLAB color space on the
interpretation of visual colour differences. J Soc Dyers Colour 90:189–198
29. CIE Publication 142 (2001) Improvement to industrial colour-difference evaluation. CIE
Central Bureau, Vienna
30. Luo MR, Cui G, Rigg B (2001) The development of the CIE 2000 colour-difference formula:
CIEDE2000. Color Res Appl 26:340–350
31. Melgosa M, Huertas R, Berns RS (2004) Relative significance of the terms in the CIEDE2000
and CIE94 color-difference formulas. J Opt Soc Am A 21:2269–2275
32. Sharma G, Wu W, Dalal EN (2005) The CIEDE2000 color-difference formula: implementation
notes, supplementary test data, and mathematical observations. Color Res Appl 30:21–30
33. Kuehni RG (2002) CIEDE2000, milestone or final answer? Color Res Appl 27:126–128
34. 6176 DIN (2000) Farbmetrische Bestimmung von Farbabständen bei Köroerfarben nach der
DIN-99-Formel. DIN Deutsche Institut für Normung e.V, Berlin
35. Cui G, Luo MR, Rigg B, Roesler G, Witt K (2002) Uniform colour spaces based on the DIN99
colour-difference formula. Color Res Appl 27:282–290
36. Kuehni RG (1999) Towards an improved uniform color space. Color Res Appl 24:253–265
37. Luo MR, Cui G, Li C (2006) Uniform colour spaces based on CIECAM02 colour appearance
model. Color Res Appl 31:320–330
38. Xue Y (2008) Uniform color spaces based on CIECAM02 and IPT color difference equations.
MD Thesis, Rochester Institute of Technology, Rochester, NY
39. MacAdam DL (1974) Uniform color scales. J Opt Soc Am 64:1691–1702
40. Huertas R, Melgosa M, Oleari C (2006) Performance of a color-difference formula based on
OSA-UCS space using small-medium color differences. J Opt Soc Am A 23:2077–2084
41. Oleari C, Melgosa M, Huertas R (2009) Euclidean color-difference formula for small-medium
color differences in log-compressed OSA-UCS space. J Opt Soc Am A 26:121–134
42. Ebner F, Fairchild MD (1998) Development and testing of a color space (IPT) with improved
hue uniformity. In: Proceedings of 6th Color Imaging Conference, 8–13, IS&T, Scottsdale, AZ
43. Shen S (2008) Color difference formula and uniform color space modeling and evaluation. MD
Thesis, Rochester Institute of Technology, Rochester, NY
3 Colour Difference Evaluation 79
44. Thomsen K (2000) A Euclidean color space in high agreement with the CIE94 color difference
formula. Color Res Appl 25:64–65
45. Urban P, Rosen MR, Berns RS, Schleicher D (2007) Embedding non-euclidean color
spaces into Euclidean color spaces with minimal isometric disagreement. J Opt Soc Am A
24:1516–1528
46. Alder C, Chaing KP, Chong TF, Coates E, Khalili AA, Rigg B (1982) Uniform chromaticity
scales – New experimental data. J Soc Dyers Colour 98:14–20
47. Schultze W The usefulness of colour-difference formulae for fixing colour tolerances. In:
Proceedings of AIC/Holland (Soesterberg 1972) 254–265
48. McLaren K (1970) Colour passing—Visual or instrumental? J Soc Dyers Colour 86:389–392
49. Alman DH, Berns RS, Snyder GD, Larsen WA (1989) Performance testing of color-difference
metrics using a color tolerance dataset. Color Res Appl 14:139–151
50. Gibert JM, Dagà JM, Gilabert EJ, Valldeperas J and the Colorimetry Group (2005) Evaluation
of colour difference formulae. Color Technol 121:147–152
51. Attridge GG, Pointer MR (2000) Some aspects of the visual scaling of large colour differences-
II. Color Res. Appl 25:116–122
52. Garcı́a PA, Huertas R, Melgosa M, Cui G (2007) Measurement of the relationship between
perceived and computed color differences. J Opt Soc Am A 24:1823–1829
53. Coxon APM (1982) The user’s guide to multidimensional scaling. London, Heinemann
54. Melgosa M, Garcı́a PA, Gómez-Robledo L, Shamey R, Hinks D, Cui G, Luo MR (2011) Notes
on the application of the standardized residual sum of squares index for the assessment of intra-
and inter-observer variability in color-difference experiments. J. Opt. Soc. Am. A. 28:949–953
55. Melgosa M, Huertas R, Berns RS (2008) Performance of recent advanced color-difference
formulas using the standardized residual sum of squares index. J. Opt. Soc. Am. A 25:1828–
1834
56. Morillas S, Gómez-Robledo L, Huertas R, Melgosa M (2009) Fuzzy analysis for detection of
inconsistent data in experimental datasets employed at the development of the CIEDE2000
colour difference formula. J Mod Optic 56:1447–1456
57. Wandell BA (1996) Photoreceptor sensitivity changes explain color appearance shifts induced
by large uniform background in dichoptic matching. Vis Res 35:239–254
58. Zhang XM, Wandell BA (1996) A spatial extension to CIELAB for digital color image
reproduction. Proc Soc Information Display 27:731–734
59. Johnson GM, Fairchild MD (2003) A top down description of S-CIELAB and CIEDE2000.
Color Res Appl 28:425–435
60. Johnson GM, Song X, Montag E, Fairchild MD (2010) Derivation of a color space for image
color difference measurements. Color Res Appl 35:387–400
61. International Standardization Organization (ISO) Graphic technology—Prepress digital data
exchange. Part 1, ISO 12640–1 (1997), Part 2, ISO 12640–2 (2004), Part 3 ISO 12640–3 (2007)
62. International Standardization Organization (ISO) (2005) Photography—Psychophysical exper-
imental method to estimate image quality. Parts 1, 2 and 3, ISO 20462
63. Aldaba MA, Linhares JM, Pinto PD, Nascimento SM, Amano K, Foster DH (2006) Visual
sensitivity to color errors in images of natural scenes. Vis Neurosci 23:555–559
64. Lee DG (2008) A colour-difference model for complex images on displays. Ph.D. Thesis,
University of Leeds, UK
65. CIE Publication 199:2011 (2011) Methods for evaluating colour differences in images. CIE
Central Bureau, Vienna
66. Melgosa M (2007) Request for existing experimental datasets on color differences. Color Res
Appl 32:159
67. Huang Z, Xu H, Luo MR, Cui G, Feng H (2010) Assessing total differences for effective
samples having variations in color, coarseness, and glint. Chinese Optics Letters 8:717–720
68. Dekker N, Kirchner EJJ, Supèr R, van den Kieboom GJ, Gottenbos R (2011) Total appearance
differences for metallic and pearlescent materials: Contributions from color and texture. Color
Res Appl 36:4–14
Chapter 4
Cross-Media Color Reproduction and Display
Characterization
The purest and most thoughtful minds are those which love
color the most
John Ruskin
4.1 Introduction
Digital images today are captured and reproduced using a plethora of different
imaging technologies (e.g., digital still cameras based on CMOS or CCD sensors,
Plasma or Liquid Crystal Displays, inkjet, or laser printers). Even within the
same type of imaging technology, there are many parameters which influence the
processes, resulting in a large variation in the color behavior of these devices.
It is therefore a challenge to achieve color consistency throughout an image
reproduction workflow, even more so since such image reproduction workflows tend
to be highly distributed and generally uncontrolled. This challenge is relevant for a
wide range of users, from amateurs of photography to professionals of the printing
industry. And as we try to advocate in this chapter, it is also highly relevant to
researchers within the field of image processing and analysis.
In the next section we introduce the field of cross-media color reproduction,
including a brief description of current standards for color management, the
concept of colorimetric characterization of imaging devices, and color gamut
mapping. Then, in Sect. 4.3 we focus on state of the art and recent research in
the characterization of displays. In Sect. 4.4, we consider methods for inverting
display characterization models; this is an essential step in cross-media color
reproduction, before discussing quality factors, based on colorimetric indicators,
briefly in Sect. 4.5. Finally, in Sect. 12.5 we draw some conclusions and outline
some directions for further research.
When using computers and digital media technology to acquire, store, process,
and reproduce images of colored objects or scenes, a digital color space is used,
typically RGB, describing each color as a combination of variable amounts of the
primaries red, green, and blue. Since most imaging devices speak RGB one may
think that there is no problem with this. However, every individual device has its
own definition of RGB, i.e., for instance for output devices such as displays, for
the same input RGB values, different devices will produce significantly different
colors. It usually suffices to enter the TV section of an home electronics store to be
reminded of this fact.
So, therefore, the RGB color space is usually not standardized, and every individ-
ual imaging device has its own definition of it, i.e., its very own relationship between
the displayed or acquired real-world color and the corresponding RGB digital color
space. Achieving color consistency throughout a complex and distributed color
4 Cross-Media Color Reproduction and Display Characterization 83
reproduction workflow with several input and output devices is therefore a serious
challenge; achieving such consistency defines the research field of cross-media color
reproduction.
The main problem is thus to determine the relationships between the different
devices’s color languages, analogously to color dictionaries. As we will see in the
next sections, a standard framework has been defined (color management system),
in which dictionaries (profiles) are defined for all devices; between their native color
language and a common, device-independent language. Defining these dictionaries
by characterizing the device’s behavior is described in Sect. 4.2.2, while Sect. 4.2.3
addresses the problem of when a device simply does not have rich enough
vocabulary to reproduce the colors of a certain image.
However, there is still a long way to go, when it comes to software development
(integration of CMS in operating systems, user-friendliness, simplicity, etc.), re-
search in cross-media color reproduction (better color consistency, gamut mapping,
color appearance models, etc.), and standardization. Color management is a very
active area of research and development, though limited by our knowledge on
the human perception process. Thus in the next sections, we will briefly review
different approaches to the colorimetric characterization of image acquisition and
reproduction devices.
Successful cross-media color reproduction needs the calibration and the character-
ization of each color device. It further needs a color conversion algorithm, which
permits to convert color values from one device to another.
In the literature, the distinction between calibration and characterization can
vary substantially, but the main idea usually remains the same. For instance, some
authors will consider a tone response curve establishment as a part of the calibration,
others as a part of the characterization. These difference does not mean too much in
practice and is just a matter of terminology. Let us consider the following definition:
The calibration process put a device in a fixed state, which will not change with
time. For a color device, it consists in setting up the device. Settings can be position,
brightness, contrast, and sometimes primaries and gamma, etc.
The characterization process can be defined as understanding and modeling the
relationship between the input and the output, in order to control a device for a given
calibration set-up. For a digital color device, this means either to understand the
relationship between a digital value input and a produced color for an output color
device (printer, display) or, in the case of an input color device (camera, scanner), to
understand the relationship between the acquired color and the digital output value.
Usually, a characterization model is mostly static, and is relying on the capability of
the device to remain in a fixed state, thus on the calibration step.
As stated above, the characterization of a color device is a modeling step, which
permits to relate the digital value that characterizes the device and the actual color
defined in a standard color space, such as CIEXYZ. There are different approaches
to modeling a device.
One can consider a physical approach, which will aim to determine a set of
physical parameters of a device, and uses these in a physical model based on
the technology definition. Such an approach has been extensively used for CRT
displays, and also it is quite common for cameras. In this case, the resulting
accuracy will be constrained by how well the device fits the model hypothesis and
how accurate the related measurements were taken. Commonly a physical device
model consists in a two steps process. First, a linearization of the intensity response
curves of the individual channels, i.e., the relation between the digital value and
4 Cross-Media Color Reproduction and Display Characterization 85
the corresponding intensity of light. The second step is typically colorimetric linear
transform (i.e., a 3x3 matrix multiplication). The characteristics of the colorimetric
transform is based on the chromaticity of the device primaries.
Another approach consists in fitting a data set with any numerical model. In this
case, the accuracy will depend on the number of data, on their distribution and on
the interpolation method used. Typically a numerical model would require more
measurement, but would make no assumption on the device behavior. We can note
that the success of such a model will depend also on the capacity of the model to fit
with the technology anyway.
For a numerical method, depending on the interpolation method used, one have
to provide different sets of measures in order to optimize the model determination.
This implies to first define which color space is used to make all the measures.
The CIEXYZ color space seems at first to be the best choice considering that some
numerical method would use its vectorial space properties successfully, particularly
additivity, in opposition with CIELAB. An advantage is that it is absolute and can
be used as an intermediary color space to a uniform color space, CIELAB, which is
recommended by the CIE for measuring the color difference when we will evaluate
the model accuracy (the Δ E in CIELAB color space). However, since we define the
error of the model, and often the cost function of the optimization process as an
Euclidean distance in CIELAB, this color space can be a better choice.
These sets of measures can be provided using a specific (optimal) color chart, or
a first approach can be to use a generic color chart, which allows to define a first
model of characterization.
However, it has been shown that it is of major importance to have a good
distribution of the data everywhere in the gamut of the device and more particularly
on the faces and the edges of the gamut, which is roughly fitting with the edges and
faces of the RGB-associated cube. These faces and edges define the color gamut
of the color device. The problem with acquisition device such as cameras is that
the lighting conditions are changing, and it is difficult to have a dedicated data set
of patches to measure for every possible condition. Thus, optimized color charts
have been designed, for which the spectral characteristics of the color patches are
designed carefully.
Another possibility is that, based on a first rough or draft model, one can provide
an optimal data set to measure, which takes into account the nonlinearity of the
input device. There are several methods to minimize errors due to the nonlinear
response of devices. By increasing the number of patches, we can tighten the mesh’s
sampling. This method can be used to reach a lower error. Unfortunately, it might
not improve much the maximum error. To reduce it, one can decide to over-sample
some particular area of the color space. The maximum error is on the boundaries
of the gamut, since there are fewer points to interpolate, and in the low luminosity
areas, as our eyes can easily see small color differences in dark colors. Finally, one
can solve this nonlinearity problem by using a nonlinear data set distribution, which
provides a quite regular sampling in the CIELAB color space.
86 J.-B. Thomas et al.
where ρ is the actual digital value output, ν is the nonlinearized value, L(λ ), R(λ ),
S(λ ) are the spectral power distribution of the illuminant, the spectral reflectance of
the object, and the spectral sensitivity of the sensor, including a color filter.
The input device calibration includes the setup of the time exposure, the
illumination (for a scanner), the contrast setup, the color filters, etc.
In the case of input devices, let us call the forward transform the transform which
relates the acquired color with the digital value, e.g., conversion from CIEXYZ to
RGB. Meanwhile the inverse transform will estimate the acquired color given digital
value caught by the device, e.g., converts from RGB to CIEXYZ.
The input device characterization can be done using a physical modeling or a
combination of numerical methods. In the case of a physical modeling, the tone
response curves will have to be retrieved; the spectral transmission of the color
filters may have to be retrieved too, in order to determine their chromaticities, thus
establishing the linear transform between intensity linearized values and the digital
values. This last part requires usually a lot of measurements, and may require the use
of a monochromator or an equivalent expensive tool. In order to reduce this set of
measurements, one needs to make some assumptions and to set some constraints to
solve the related inverse problem. Such constraints can be the modality of the spec-
tral response of a sensor or that the sensor response curve can be fitted with just a
few of the first Fourier coefficients, see, e.g., [10,36,76]. Such models would mostly
use the CIEXYZ color space or another space which has the additivity property.
Johnson [51] gives good advice for achieving a reliable color transformation for
both scanners and digital cameras. In his paper, one can find diverse characterization
procedures, based on the camera colorimetric evaluation using a set of test images.
The best is to find a linear relationship to map the output values to the input
target (each color patch). The characterization matrix, once more, provides the
transformation applied to the color in the image. In many cases, the regression
analysis shows that the first order linear relationship is not satisfactory and a
higher order relationship or even nonlinear processing is required (log data, gamma
correction, or S-shape, e.g.). Lastly, if a matrix cannot provide the transformation,
then a look-up table (LUT) will be used. Unfortunately, the forward transform can
be complicated and quite often produces artifacts [51]. Possible solutions to the
problems of linear transformations encountered by Johnson are least-squares fitting,
4 Cross-Media Color Reproduction and Display Characterization 87
nonlinear transformations, or look-up tables with interpolation. In the last case, any
scanned pixel can be converted into tristimulus values via the look-up table(s) and
interpolation is used for intermediate points which do not fall in the table itself. This
method is convenient for applying a color transformation when a first order solution
is not relevant. It can have a very high accuracy level if the colors are properly
selected.
The colorimetric characterization of a digital camera was analyzed by [45].
An investigation was done to determine the influence of the polynomial used
for interpolation and the possible correlation between the RGB channels. The
channel independence allows us to separate the contribution of spectral radiance
from the three channels. Hong et al. [45] also checked the precision of the model
with respect to the training samples’ data size provided and the importance of
the color precision being either 8 or 12 bits. According to the authors, there are
two categories of color characterization methods: either spectral sensitivity based
(linking the spectral sensitivity to the CIE color-matching functions) or color target
based (linking color patches to the CIE color-matching functions). These two
solutions lead to the same results, but the methods and devices used are different.
Spectral sensitivity analysis requires special equipment like a radiance meter and a
monochromator; while a spectrophotometer is the only device needed for the color
target-based solution. Typical methods like 3D look-up tables with interpolation
and extrapolation, least square polynomials modeling and neural networks can be
used for the transformation between RGB and CIEXYZ values, but in this article,
polynomial regression is used. As for each experiment only one parameter (like
polynomial order, number of quantization levels, or size of the training sample)
changes, the Δ Eab∗ difference is directly linked to the parameter.
Articles published on this topic are rare, but characterization of other input
devices with a digital output operates the same way. Noriega et al. [67] and
[37] further propose different transformation techniques. These articles discuss the
colorimetric characterization of a scanner and a negative film. In the first article
[67], the authors decided to use least squares fitting, LUTs and distance-weighted
interpolation. The originality comes from the use of the Mahalanobis distance used
to perform the interpolation. The second article [37] deals with the negative film
characterization. Distance-weighted interpolation, Gaussian interpolation neural
networks, and nonlinear models have been compared using Principal Component
Analysis. In these respective studies, the models were trained with the Mahalanobis
distance (still using the color difference as a cost function) and neural networks.
An output device in this context is any device that will reproduce a color, such as
printers, projection systems, or monitors. In this case, the input to the device is a
digital value, and we will call the forward transform the transform that predicts the
color displayed for a given input, e.g., RGB to CIEXYZ. The inverse or backward
transform will then define which digital value we have to input to the device to
reproduce a wanted color, e.g., CIEXYZ to RGB.
88 J.-B. Thomas et al.
The characterization approach for output devices and media is similar to that of
input devices. One has to determine a model based on more or less knowledge of
the physical behavior of the device, and more or less measurement of color patches
and mathematical approximation/interpolation. Since displays are covered in depth
in Sect. 4.3, we will here briefly discuss printer characterization.
We can distinguish between two kinds of printer characterization models, the
computational and the physical ones. Typically, for a 4-colorant CMYK printer
the computational approach consists in building a grid in four dimensions, a
multidimensional look-up table (mLUT). The estimation of the resulting color
for a given colorant combination will then be calculated by multidimensional
interpolation in the mLUT. An important design trade-off for such modeling is
between the size of the mLUT and the accuracy of the interpolation.
The physical models attempt to imitate the physics involved in the printing
device. Here also these models can be classified into two subtypes with regard to
the assumptions they make and their complexity [90]: regression-based and first-
principal models. Regression-based models are rather simple and works with a few
parameters to predict a printer output while first-principal model will closely imitate
the physics of the printing process by taking into account multiple light interactions
between the paper and the ink layers, for instance. Regression-based models are
commonly used to model the behavior of digital printing devices.
During the last century, printing technology has evolved and the printer models
as well. Starting from a single-colorant printing device the Murray-Davies model
predicts the output spectral reflectances of a single-colorant coverage value knowing
the spectral reflectance of the paper and maximum colorant coverage value. This
model was extended to color by [64]. The prediction of a colorant combination is
the summation of all the colorants involved in the printing process weighted by
their coverage on the paper. All the colorants are referring to all the primaries (cyan,
magenta, and yellow in case of a CMY printer) plus all the combination between
them plus the paper, these colors are called the Neugebauer primaries (NP). Later
the interaction of light penetrating and scattering into the paper was added to these
models by [95], as form of an exponent known as the n factor. For more information
about printer characterization, refer, e.g., to [38].
A color gamut is the set of all colors that can be produced by a given device or
that are present in a given image. Although these sets are in principle discrete,
gamuts are most often represented as volumes or blobs in a 3D color space using a
gamut boundary descriptor [7]. When images are to be reproduced between different
devices, the problem of gamut mismatch has to be addressed. This is usually referred
to as color gamut mapping. There is a vast amount of literature about the gamut-
mapping problem, see, for instance, a recent book by [63].
4 Cross-Media Color Reproduction and Display Characterization 89
To keep the image appearance, some constraints are usually considered while
doing a gamut mapping:
• Preserve the gray axis of the image and aim for maximum luminance contrast.
• Reduce the number of out-of-gamut colors.
• Minimize hue shifts.
• Increase the saturation.
CIELAB is one of the most often used color spaces for gamut mapping, but there
are deficiencies in the uniformity of hue angles in the blue region. To prevent this
shift, one can use Hung and Berns’ data to correct the CIELAB color space [21].
To map a larger source gamut into a smaller destination gamut of a device with
a reduced lightness dynamic range, often a linear lightness remapping process is
applied. It suffers from a global reduction in the perceived lightness contrast and an
increase in the average lightness of the remapped image. It is of utmost importance
to preserve the lightness contrast. An adaptive lightness rescaling process has been
developed by [22]. The lightness contrast of the original scene is increased before
the dynamic range compression is applied to fit the input lightness range into the
destination gamut. This process is known as a sigmoidal mapping function, the
shape of this function aids in the dynamic range mapping process by increasing
the image contrast and by reducing the low-end textural defects of hard clipping.
We can categorize different types of pointwise gamut-mapping technics (See
Fig. 4.1); gamut clipping only changes the colors outside the reproduction gamut
while gamut compression changes all colors from the original gamut. The knee
function rescaling preserves the chromatic signal through the central portion of the
gamut, while compressing the chromatic signal near the edges of the gamut. The
sigmoid-like chroma mapping function has three linear segments; the first segment
preserves the contrast and colorimetry, the second segment is a mid-chroma boost
(increasing chroma), and the last segment compresses the out-of-gamut chroma
values into the destination gamut.
Spatial gamut mapping has become an active field of research in the recent years
[35, 56]. In contrast to the conventional color gamut-mapping algorithms, where
the mapping can be performed once and for all and stored as a look-up table, e.g.,
in an ICC profile, the spatial algorithms are image dependent by nature. Thus,
the algorithms have to be applied for every single image to be reproduced, and
make direct use of the gamut boundary descriptors many times during the mapping
process.
Quality assessment is also required for the evaluation of gamut-mapping
algorithms, and extensive work has been carried out on subjective assessment
[32]. This evaluation is long, tiresome, and even expensive. Therefore, objective
assessment methods are preferable. Existing work on this involves image quality
metrics, e.g., by [17, 44]. However, these objective methods can still not replace
subjective assessment, but can be used as a supplement to provide a more thorough
evaluation.
Recently, [4] presented a novel, computationally efficient, iterative, spatial
gamut-mapping algorithm. The proposed algorithm offers a compromise between
90 J.-B. Thomas et al.
the colorimetrically optimal gamut clipping and the most successful spatial meth-
ods. This is achieved by the iterative nature of the method. At iteration level
zero, the result is identical to gamut clipping. The more we iterate, the more we
approach an optimal, spatial, gamut-mapping result. Optimal is defined as a gamut-
mapping algorithm that preserves the hue of the image colors as well as the spatial
ratios at all scales. The results show that as few as five iterations are sufficient
to produce an output that is as good or better than that achieved in previous,
computationally more expensive, methods. Unfortunately, the method also shares
some of the minor disadvantages of other spatial gamut-mapping algorithms: halos
and desaturation of flat regions for particularly difficult images. There is therefore
much work left to be done in this direction, and one promising idea is to incorporate
knowledge of the strength of the edges.
Fig. 4.2 3D look-up table for a characterization process from RGB to CIELAB
appeared. Many methods have been borrowed from printers or camera though, but
the way to reproduce colors and the assumptions one can do are different when
talking about displays, so the results or the explanation of why a model is good or
not are slightly different. We propose to discuss the state of the art and the major
trends about display colorimetric characterization in this section.
Many color characterization methods or models exist; we can classify them in three
groups. In a first one, we find the models, which tend to model physically the color
response of the device. They are often based on the assumption of independence
between channels and of chromaticity constancy of primaries. Then, a combination
of the primary tristimulus at the full intensity weighted by the luminance response
of the display relatively to a digital input can be used to perform the colorimetric
transform. The second group can be called numerical models. They are based on
a training data set, which permits optimization of the parameters of a polynomial
function to establish the transform. The last category consists of 3D LUT-based
models. Some other methods can be considered as hybrid. They can be based
on a data set and assume some physical properties of the display, such as in the
work of [16].
The models in the 3D LUT group are based on the measurement of a defined
number of color patches, i.e., we know the transformation between the input values
92 J.-B. Thomas et al.
(i.e., RGB input values to a display device) and output values (i.e., CIEXYZ or
CIELAB values) measured on the screen by a colorimeter or spectrometer in a
small number of color space locations (see Fig. 4.2). Then this transformation
is generalized to the whole space by interpolation. Studies assess that these
methods can achieve accurate results [11, 80], depending on the combination of
the interpolation method used [2,5,18,53,66], the number of patches measured, and
on their distribution [80] (note that some of the interpolation methods cited above
cannot be used with a non-regular distribution). However, to be precise enough, a lot
of measurements are typically required, i.e., a 10× 10 × 10 grid of patches measured
in [11]. Note that such a model is technology independent since no assumptions are
made about the device but that the display will always have the same response at the
measurement location. Such a model needs high storage capacity and computational
power to handle the 3D data. The computational power is usually not a problem
since Graphic Processor Units can perform this kind of task easily today [26]. The
high number of measurements needed is a greater challenge.
The numerical models suppose that the transform can be approximated by a set
of equations, usually an n-order polynomial function. The parameters are retrieved
using an n-order polynomial regression process based on measurements. The
number of parameters required involves a significant number of measurements,
depending on the order of the polynomial function. The advantage of these models is
that they take into account channel interdependence by applying cross components
factors in the establishment of the function [54,55,83]. More recently, an alternative
method has been proposed by [89] who removed the three-channel crosstalk from
the model, considering that the inter-channel dependence is only due to two-channel
crosstalk, thus reducing the required number of measurements. They obtained
results as accurate as when considering the three-channel crosstalk.
Radial basis function (RBF) permits to use a sum of low-order polynomials
instead of one high-order polynomial and has been used successfully in different
works [26, 27, 79, 80]. Mostly polyharmonic splines are used, which include thin
plate splines (TPS) that [75] used for printers too. TPS are a subset of polyharmonic
splines (bi-harmonic splines). Sharma and Shaw [75] recalled the mathematical
framework and presented some applications and results for printer characterization.
They showed that using TPS, they achieved a better result than in using local
polynomial regression. They showed that by using a smoothing factor, error in
measurement impact can be avoided at the expense of the computational cost that
optimize this parameter, similar results were observed by [26]. However, [75] did
study neither data distribution influence (but they stated that the data distribution
can improve the accuracy in their conclusion) nor the use of other kernels for
interpolation. This aspect has been studied by [26], in which main improvements
were in the optimization of the selection of the data used to build the model in an
iterative way.
4 Cross-Media Color Reproduction and Display Characterization 93
Physical models are historically widely used for displays, since the CRT technology
follows well the assumptions cited above [13, 19, 29]. Such a model typically
first aims to linearize the intensity response of the device. This can be done by
establishing a model that assumes the response curve to follow a mathematical
function, such as a gamma law for CRT [13, 14, 28, 74], or an S-shaped curve
for LCD [58, 59, 94]. Another way to linearize the intensity response curve is to
generalize measurements by interpolation along the luminance for each primary
[68]. The measurement of the luminance can be done using a photometer. Some
approaches propose as well a visual response curve estimation, where the 50%
luminance point for each channel is determined by the user to estimate the gamma
value [28]. This method can be generalized to the retrieval of more luminance levels
in using half-toned patches [62, 65]. Recently a method to retrieve the response
curve of a projection device using an uncalibrated camera has been proposed by [8]
and extended by [62]. Note that it has been assumed that the normalized response
curve is equivalent for all the channels, and that only the gray level response curve
can be retrieved. In the case of a doubt about this assumption, it is useful to retrieve
the three response curves independently. Since visual luminance matching for the
blue channel is a harder task, it is of use to perform an intensity matching for the
red and green channel, and a chromaticity matching or gray balancing for the blue
one [57]. This method should not be used with projectors though, since they show a
large chromaticity shift with the variation of input for the pure primaries.
A model has been defined by [91, 92] for DLP projectors using a white segment
in the color wheel. In their model, the characteristics of the luminance of the white
channel is retrieved with regard to additive property of the display, given the four-
tuplet (R, G, B,W ) from an input (dr , dg , db ).
The second step of these models is commonly the use of a 3 × 3 matrix
containing primary tristimulus values at full intensity to build the colorimetric
transform from luminance to an additive independent color space. The primaries
can be estimated by measurement of the device channels at full intensity, using
a colorimeter or a spectroradiometer, assuming their chromaticity constancy. In
practice this assumption does not hold perfectly, and the model accuracy suffers
from that. The major part of the non-constancy of primaries can be corrected
by applying a black offset correction [50]. Some authors tried to minimize the
chromaticity non-constancy in finding the best chromaticity values of primaries
(optimizing the components of the 3 × 3 matrix) [30]. Depending on the accuracy
required, it is also possible to use generic primaries such as sRGB for some
applications [8], or data supplied by the manufacturer [28]. However, the use of
a simple 3 × 3 matrix for the colorimetric transform leads to inaccuracy due to
the lack of channel independence and of chromaticity constancy of primaries. An
alternative approach has been derived in the masking model and modified masking
model, which takes into account the cross-talk between channels [83]. Furthermore,
the lack of chromaticity constancy can be critical, particularly for LCD technology,
which has been shown to fail this assumption [20, 58]. The piecewise linear model
94 J.-B. Thomas et al.
assuming variation in chromaticity (PLVC) [34] is not subject to this effect, but has
not been widely used since [68] demonstrated that among the models they tested in
their article, the PLVC and the piecewise linear-assuming chromaticity constancy
(PLCC) models were of equivalent accuracy for the CRT monitors they tested. With
the last one requiring less computation, it has been more used than the former one.
These results have been confirmed in studies on CRT technology [68,69], especially
with a flare correction [50, 86]. On DLP technology when there is a flare correction,
results can be equivalent; however, PLVC can give better results on LCDs [86].
Other models exist, such as the two-steps parametric model proposed by [16].
This model assumes separation between chromaticity and intensity, and is shown to
be accurate, with average Δ Eab ∗ ’s around 1 or below for one DLP projector and a
CRT monitor. The luminance curve is retrieved, as for other physical models, but
the colorimetric transform is based on 2D interpolation in the chromaticity plane
based on a set of saturated measured colors.
Physical models are easily invertible, do not require a lot of measurements, require
a little computer memory, and do not require high computing power. So, they
can be used in real time. Moreover, the assumptions of channel independence
and chromaticity constancy are appropriate for the CRT technology. However,
these assumptions (and others such as spatial uniformity, both in luminance and
in chromaticity, view angle independence, etc.) do not fit so well with some of
today’s display technologies. For instance, the colorimetric characteristic of a part
of an image in a Plasma Display is strongly dependent of what is happening in the
surrounding [25] for energy economy reasons. In LC technology, which has become
the leader for displays market, these common assumptions are not valid. Making
such assumptions can reduce drastically the accuracy of the characterization. For
instance, a review of problems faced in LC displays has been done by [94]. Within
projection systems, the large amount of flare induces a critical chromaticity shift of
primaries.
In the same time, the computing power has become less and less a problem. Some
models not used in practice because of their complexity can now be highly beneficial
for display color characterization. This section provides definitions, analysis, and
discussion about display color characterization models. We do not detail hybrid
methods or numerical methods in this section because they show less interest for
modeling purpose, and we do prefer to refer the reader to the papers cited above. 3D
LUT-based method are more considered in the part concerning model’s inversion.
In 1983, [28] wrote what is considered to be the pioneer article in the area of
physical models for display characterization. In this work, the author stated that a
power function can be used, but is not the best to fit with the luminance response
curve of a CRT device. Nevertheless, the well-known “gamma” model that considers
a power function to approximate the luminance response curve of a CRT display is
still currently widely used.
Whichever shape the model takes, the principle remains the same. First, it
estimates the luminance response of the device for each channel, using a set of
functions monotonically increasing such as (4.2). Note that the results of these
functions can also be estimated with any interpolation method, since the problem
of monotonicity that can arise during the inversion process is taken into account.
This step is followed by a colorimetric transform.
We review here two types of models. The models of the first type are based
on functions, the second type is the PLCC model. This model is based on
linear interpolation of the luminance response curve and its accuracy has been
demonstrated by [68] who found it the best among the models they tested (except in
front of the PLVC model for chromatic accuracy).
96 J.-B. Thomas et al.
0
0 50 100 150 200 250
b
100
80
60
40
20
0
0 50 100 150 200 250
c
150
100
50
0
0 50 100 150 200 250
For function-based model, the function used is the power function for CRT
devices, which is still the most used, even if it has been shown that it does not
fit well LC technology [33]. It has been shown that for other technologies, there is
no reason to try to fit the device response with a gamma curve, especially for an
LCD technology that shows an S-shaped response curve in most cases (Fig. 4.3)
and an S-curve model can be defined [58, 59, 94]. However, the gamma function is
still often used, mainly because it is easy to estimate the response curve with a few
number of measurements, or using estimations with a visual matching pattern.
The response in luminance for a set of digital values input to the device can be
expressed as follows:
YR = fr (Dr )
YG = fg (Dg )
YB = fb (Db ), (4.2)
4 Cross-Media Color Reproduction and Display Characterization 97
where fr , fg , and fb are functions that give the YR ,YG , and YB contribution in
luminance of each primary independently for a digital input Dr , Dg , Db . Note that
for CRT devices, after normalization of the luminance and digital value, the function
can be the same for each channel. This assumption is not valid for LCD technology
[73], and is only a rough approximation for DLP-based projection systems, as seen,
for instance, in the work of [72].
For a CRT, for the channel h ∈ {r, g, b}, this function can be expressed as
1 Note that [68] added a term to this equation, which became Log(YH ) = a + bh × Log(dh ) +
ch .(Log(dh ))2 .
98 J.-B. Thomas et al.
dα (α − β )xα +β −1 + α Cxα −1
g(d) = , g (d) = . (4.6)
dβ + C (xβ + C)2
To ensure the monotonicity of the functions for the S-curve models I and II,
some constraints on the parameters have to be applied. We let the reader refer to the
discussion in the original article [59] for that matter.
For the PLCC model, the function f is approximated by a piecewise linear
interpolation between the measurements. The approximation is valid for a large
enough amount of measurements (16 measurements per channel in [68]). This
model is particularly useful when no information is available about the shape of
the display luminance response curve.
A colorimetric transform is then performed from the (YR ,YG ,YB ) “linearized”
luminance to the CIEXYZ color tristimulus.
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
X Xr,max Xg,max Xb,max YR
⎣ Y ⎦ = ⎣ Yr,max Yg,max Yb,max ⎦ × ⎣ YG ⎦ , (4.7)
Z Zr,max Zg,max Zb,max YB
where the matrix components are the tristimulus colorimetric values of each
primary, measured at their maximum intensity.
Using such a matrix for the colorimetric transform supposes perfect additivity
and chromaticity constancy of primaries. These assumptions have been shown to be
acceptable for CRT technology [19, 29].
The channel interdependence observed in CRT technology is mainly due to an
insufficient power supply and an inaccuracy of the electron beams, which meet
inaccurately the phosphors [54]. In LC technology, it comes from the overlapping
of the spectral distribution of primaries (the color filters), and from the interferences
between the capacities of two neighboring subpixels [72, 94]. In DLP-DMD
projection devices, there is still some overlapping between primaries and inaccuracy
at the level of the DMD mirrors.
Considering the assumption of chromaticity constancy, it appears that when there
is a flare [54], either a black offset (internal flare) or an ambient flare (external flare),
added to the signal, the assumption of chromaticity constancy is not valid anymore.
Indeed, the flare is added to the output signal and the lower the luminance level of the
primaries, the more the flare is a significant fraction of the resulting stimulus. This
leads to a hue shift toward the black offset chromaticity. Often the flare has a “gray”
(nearly achromatic) chromaticity; thus, the chromaticities of the primaries shift to a
“gray” chromaticity (Fig. 4.4, left part). Note that the flare “gray” chromaticity does
not necessarily correspond to the achromatic point of the device (Fig. 4.4). In fact,
4 Cross-Media Color Reproduction and Display Characterization 99
Fig. 4.4 Chromaticity tracking of primaries with variation of intensity. The left part of the figure
shows it without black correction. On the right, one can see the result with a black correction
performed. All devices tested in our PLVC model study are shown, a-PLCD1, b-PLCD2, c-PDLP,
d-MCRT, e-MLCD1, f-MLCD2. Figures from [86]
100 J.-B. Thomas et al.
in the tested LCD devices (Fig. 4.4a, b, e, f), we can notice the same effect as in
the work of [61]: the black level chromaticity is bluish because of the poor filtering
power of the blue filter in the low wavelength.
The flare can be taken all at once as the measured light for an input (dr,k , dg,k , db,k )
= (0, 0, 0) to the device. Then it includes ambient and internal flare.
The ambient flare comes from any light source reflecting on the display screen.
If the viewing conditions do not change it remains constant, can be measured and
taken into account, or can be simply removed in setting up a dark environment (note
that for a projection device, there is always an amount of light that lights the room,
coming from the bulb through the ventilation hole).
The internal flare, which is the major part of chromaticity inconstancy at least
in CRT technology [54], is coming from the black level. In CRT technology, it has
been shown that in setting the brightness to a high level, the black level increases to a
non-negligible value [54]. In LC technology, the panel let an amount of light passing
through due to a leakage of the crystal to stop all the light. In DLP technology, an
amount of light can be not absorbed by the “black absorption box,” and is focused
on the screen via the lens.
On Fig. 4.4, one can see the chromaticity shift to the flare chromaticity with the
decreasing of the input level. We have performed these measurements in a dark
room, then the ambient flare is minimized, and only the black level remains. After
black level subtraction, the chromaticity is more constant (Fig. 4.4), and a new
model can be set up in taking that into account [43, 50, 54, 55].
The gamma models reviewed above have been extended in adding an offset term.
Then the GOG can become a gain-offset-gamma-offset (GOGO) model [46,54,55].
The previous equation (4.2) becomes:
where c is a term containing all the different flares in presence. If we consider the
internal offset bh as null, the model becomes gain-gamma-offset (GGO) [46].
A similar approach can be used for the PLCC model. When the black correction
[50] is performed, we name it PLCC* in the following. The colorimetric transform
used then is (4.9) that permits to take the flare into account during the colorimetric
transformation. For the S-curve models, the black offset is taken into account in the
matrix formulation in the original papers.
If we consider that mathematically, the linear transform from the linearized
RGB to CIEXYZ needs to associate the origin of RGB to the origin of CIEXYZ in
order to respect the vectorial space property of additivity and homogeneity. Thus,
the original transform of the origin of RGB to CIEXYZ needs to be translated of
[−Xk −Yk − Zk ]. However, in doing that we modify the physical reality and we need
to translate the result of the transformation of [XkYk Zk ]. We can formulate these
transforms such as in (4.9).
102 J.-B. Thomas et al.
⎡ ⎤
⎡ ⎤ ⎡ ⎤ YR
X Xr,max − Xk Xg,max − Xk Xb,max − Xk Xk ⎢ YG ⎥
⎣ Y ⎦ = ⎣ Yr,max − Yk Yg,max − Yk Yb,max − Yk Yk ⎦ × ⎢ ⎥ . (4.9)
⎣ YB ⎦
Z Zr,max − Zk Zg,max − Zk Zb,max − Zk Zk
1
technologies [86]. This model does not consider the channel interdependence, but
does model the chromaticity shift of the primaries. In this section, we recall the
principles of this model, and some features that characterize it.
Knowing the tristimulus values of X, Y , and Z for each primary as a function of
the digital input, assuming additivity, the resulting color tristimulus values can be
expressed as the sum of tristimulus values for each component (i.e., primary) at the
given input level. Note that in order not to add several times the black level, it is
removed from all measurements used to define the model. Then, it is added to the
result, to return to a correct standard observer color space [50, 69]. The model is
summarized and generalized in (9.3) for N primaries, and illustrated in (4.11) for a
three primaries RGB device, following an equivalent formulation as the one given
by [50].
For an N primary device, we consider the digital input to the ith primary,
di (mi ), with i an integer ∈ [0, N], and mi an integer limited by the resolution
of the device (i.e., mi ∈ [0, 255] for a channel coded on 8 bits). Then, a color
CIEXYZ(. . . , di (mi ), . . .) can be expressed by:
i=N−1
X(. . . , di (mi ), . . .) = ∑ [X(di ( j)) − Xk ] + Xk ,
i=0, j=mi
i=N−1
Y (. . . , di (mi ), . . .) = ∑ [Y (di ( j)) − Yk ] + Yk ,
i=0, j=mi
i=N−1
Z(. . . , di (mi ), . . .) = ∑ [Z(di ( j)) − Zk ] + Zk (4.10)
i=0, j=mi
with Xk ,Yk , Zk the color tristimulus coming out from a (0, . . . , 0) input.
We illustrate this for a three primaries RGB device, with each channel coded on
8 bits. The digital input are dr (i), dg ( j), db (l), with i, j, l integers ∈ [0, 255]. In this
case, a CIEXYZ(dr (i), dg ( j), db (l)) can be expressed by:
tion is less true for CRT technology. Then the accuracy decreases (depending on
the device properties). More precisely, [68, 69] stated that chromaticity error is
lower for the PLVC than for the PLCC in low luminance. This is due to the
setting of primaries colorimetric values at maximum intensity in the PLCC. Both
models show inaccuracy for high luminance colors due to channel interdependence.
Jimenez Del Barco et al. [50] found that for CRT technology, the higher level of
brightness in the settings leads to a non-negligible amount of light for a (0,0,0)
input. This light should not be added three times, and they proposed a correction
for that.2 They found that the PLVC model was more accurate in medium to high
luminance colors. Inaccuracy is more important in low luminance, due to inaccuracy
of measurements, and in high luminance, due to channel dependencies. Thomas
et al. [86] demonstrated that this model is more accurate than usual linear models
(PLCC, GOGO) for LCD technology, since it takes into account the chromaticity
shift of primaries that is a key features for characterizing this type of display. More
results for this model are presented in the next chapter.
2 Equations (4.10) and (4.11) are based on the equation proposed by [50], and take that into account.
4 Cross-Media Color Reproduction and Display Characterization 105
• Empirical methods based on 3-D LUT (look-up table) can be inverted directly
[11], using the same geometrical structure. In order to have a better accuracy,
however, it is common to build another geometrical structure to yield the inverse
model. For instance, it is possible to build a draft model to define a new set of
color patches to be measured [80].
The computational complexity required to invert these models makes them seldom
used in practice, except the full 3-D LUT, whose major drawback is that it requires
a lot of measurements. However, these models do have the possibility to take into
account more precisely the device color-reproduction features, such as interaction
between channels or chromaticity inconstancy of the primaries. Thus, they are often
more accurate than the models of the first category.
Models such as the PLCC, the black-corrected PLCC*, the GOG, or GOGO models
[13, 14, 29, 50, 54, 55, 68] are easily inverted since they are based on linear algebra
and on simple functions. For these models, it is sufficient to invert the matrix of
(4.7). Then we have:
⎡ ⎤ ⎡ ⎤−1 ⎡ ⎤
YR Xr,max Xg,max Xb,max X
⎣ YG ⎦ = ⎣ Yr,max Yg,max Yb,max ⎦ × ⎣ Y ⎦ . (4.12)
YB Zr,max Zg,max Zb,max Z
Once the linearized {YR ,YG ,YB } have been retrieved, the intensity response curve
function is inverted as well to retrieve the {dr , dg , db } digital values. This task is
easy for a gamma-based model or for an interpolation-based one. However, for some
models such as the S-curve I, an optimization process can be required (note that this
response curve can be used to create a 1D LUT).
When the inversion becomes more difficult, it is of use to set an optimization process
using the combination of the forward transform and the color difference (often the
euclidean distance) in a perceptually uniform color space, such as CIELAB, as cost
function. This generally leads to better results than usual linear models, depending
on the forward model, but is computationally expensive, and cannot be implemented
in real time. It is then of use to set a 3-D LUT based on the forward model. Note that
it does not mean that an optimization process is useless, since it can help to design
a good LUT.
106 J.-B. Thomas et al.
Fig. 4.5 The transform between RGB and CIELAB is not linear. Thus while using a linear
interpolation based on data regularly distributed in RGB, the accuracy is not the same everywhere
in the colorspace. This figure shows a plot of regularly distributed data in a linear space (blue dot,
left) and the resulting distribution after a cubic root transform (that mimics CIELAB transform)(red
dots, right)
Such a model is defined by the number and the distribution of the color patches
used in the LUT, and by the interpolation method used to generalize the model
to the entire space. In this subsection, we review some basic tools and methods.
We distinguish works on displays from more general works, which have been
performed in this way either in a general purpose or especially for printers. One of
the major challenges for printers is the problem of measurement, which is really
restrictive, and many works have been carried out in using a 3-D LUT for the
color characterization of these devices. Moreover, since printer devices are highly
nonlinear, their colorimetric models are complex. So it has been customary in the
last decade to use a 3-D complex LUT for the forward model, created by using
an analytical forward model, both to reduce the amount of measurements and to
perform the color space transform in a reasonable time. The first work we know
about creating a LUT based on the forward model is a patent from [81]. In this
work, the LUT is built to replace the analytical model in the forward direction. It
is based on a regular grid designed in the printer CMY color space, and the same
LUT is used in the inverse direction, simply in switching the domain and co-domain.
Note that in displays, the forward model is usually computationally simple and that
we need only to use a 3-D LUT for the inverse model. The uniform mapping of
the CMY space leads to a nonuniform mapping in CIELAB space for the inverse
direction, and it is common now to resample this space to create a new LUT. To do
that, a new grid is usually designed in CIELAB and is inverted after gamut mapping
of the points located outside the gamut of the printer. Several algorithms can be used
to redistribute the data [24, 31, 41] and to fill the grid [9, 77, 88].
4 Cross-Media Color Reproduction and Display Characterization 107
Returning to displays, let us call source space the independent color space
(typically CIELAB or alternatively CIEXYZ), the domain from where we want to
move, and destination space, the RGB color space, the co-domain, where we want
to move to. If we want to build a grid, we then have two classical approaches to
distribute the patches in the source space, using the forward model. One can use
directly a regular distribution in RGB and transform it to CIELAB using the forward
model; this approach is the same as used by [81] for printers, and leads to a non-
uniform mapping of the CIELAB space, which can lead to a lack of homogeneity
of the inverse model depending on the interpolation method used (See Fig. 4.5). An
other approach can be to distribute the patches regularly in CIELAB, following a
given pattern, such as an hexagonal structure [80] or any of the methods used in
printers [24, 31, 41]. Then, an optimization process using the forward model can be
performed for each point to find the corresponding RGB values. The main idea of
the method and the notation used in this document are the following:
• One can define a regular 3-D grid in the destination color space (RGB).
• This grid defines cubic voxels. Each one can be split into five tetrahedra (See
Fig. 4.6).
• This tetrahedral shape is preserved within the transform to the source space
(either CIEXYZ or CIELAB).
• Thus, the model can be generalized to the entire space, using tetrahedral
interpolation [53]. It is considered in this case that the color space has a linear
behavior within the tetrahedron (e.g., the tetrahedron is small enough).
The most used way to define such a grid is to take directly a linear distribution
of points on each digital dr , dg , and db axis as seeds and to fill up the rest of
the destination space. A tetrahedral structure is then built with these points. The
built structure is used to retrieve any RGB value needed to display a specific color
inside the device’s gamut. The more points are used to build the grid, the more
the tetrahedra will be small and the interpolation accurate. Each vertex is defined
by Vi, j,k = (Ri , G j , Bk ), where Ri = di , G j = d j , Bk = dk , and di , d j , dk ∈ [0, 1]
are the possible normalized digital values, for a linear distribution. i ∈ [0, Nr − 1],
j ∈ [0, Ng − 1], and k ∈ [0, Nb − 1] are the indexes (integers) of the seeds of the grid
along each primary, and Nr (resp. Nb , Ng ) is the number of steps along channel R
(resp. G, B).
Once this grid has been built, we define the tetrahedral structure for the
interpolation following [53]. Then, we use the forward model to transform the
structure into CIELAB color space. An inverse model has been built. According to
the nonlinearity of the CIELAB transform, the size of the tetrahedra is not anymore
the same as it was in RGB. In the following section, a modification of this framework
is proposed that makes this grid more homogeneous in the source color space where
we perform the interpolation; this should lead to a better accuracy, following [41].
Let us consider the PLVC model inversion as an example. This model inversion
is not as straightforward as the matrix-based models previously defined. For a three
primaries display, according to [68], it can be performed defining all subspaces
defined by the matrices of each combinations of measured data (note that the
108 J.-B. Thomas et al.
Fig. 4.6 The two ways to split a cubic voxel in 5 tetrahedra. These two methods are combined
alternatively when splitting the cubic grid to guarantee that no coplanar segments are crossing
intercepts have to be subtracted, and once all the contributions are known, they have
to be added). One can perform an optimization process for each color [50], or define
a grid in RGB, such as described above, which will allow us to perform the inversion
using 3D interpolation. Note that Post and Calhoun have proposed to define a full
LUT considering all colors. They said themselves that it is inefficient. Defining a
reduced regular grid in RGB leads to the building of an irregular grid in CIELAB
due to the nonlinear transform. This irregular grid could lead to inaccuracy or a lack
of homogeneity in interpolation, especially if it is linear. Some studies addressed
this problem [84, 85]. They built an optimized LUT, based on a customized RGB
grid.
4 Cross-Media Color Reproduction and Display Characterization 109
4.5.1 Purpose
Fig. 4.7 Evaluation of a forward model scheme. A digital value is sent to the model and to
the display. A value is computed and a value is measured. The difference between these values
represents the error of the model in a perceptually pseudo-uniform color space
We can see a duality between two types of display characterization methods and
goals: the consumer, end-user purpose, which intends only to keep the meaning and
aesthetic unchanged through the color workflow, and the accurate professional one,
which aims to have a very high colorimetric fidelity through the color workflow. We
see also through these examples that the constraints and the needs are not necessarily
going in the opposite direction.
In the next section, we will relate the quality of a model with colorimetric
objective indicators.
4.5.2 Quality
Once a model is set up, there is a need to evaluate its quality to confirm we are within
the accuracy we wanted. In this section, we discuss how to use objective indicators
for assessing quality.
[Link] Evaluation
Table 4.1 This table shows the set of thresholds one can use to assess the
quality of a color characterization model, depending on the purpose
∗
Δ Eab Professional Consumer
Mean Δ Eab∗ Max Δ Eab∗ Mean Δ Eab∗
a good choice, since it will cover the whole device possibility. This can also be a
good choice for the comparison of one method over different devices. However, if
one wants to relate the result to the visual interpretation of the signal throughout
the whole gamut of the device, it might be judicious to select an equiprobably
distributed data set in a perceptual color space. This means that most of the data
will fall into low digital values.
Once we have an estimation of the model failure, we would like to be able to say how
it is good or not for a given purpose. The ideal colorimetric case is to have an error
below the just noticeable difference3(JND). Kang [52] stated on page 167 of his
book that the JND is of 1 Δ Eab∗ unit. Mahy et al. [60] study assessed that the JND is
∗
of 2.3 Δ Eab units. Considering that the CIELAB color space is not perfectly uniform,
it is impossible to give a perfect threshold with an euclidean metric.4 Moreover,
these thresholds have been defined for simultaneous pair comparison of uniform
color patches. This situation almost never fit with a display use, it may then not be
the best choice when comparing color display devices.
∗ thresholds for color imaging devices, many thresholds have
In the case of Δ Eab
been used [1, 42, 70, 78]. Stokes et al. [82] found a perceptibility acceptance for
pictorial images of an average of 2.15 units. Catrysse et al. [23] used a threshold of
3 units. Gibson and Fairchild [39] found acceptable a characterized display that has
a prediction error average of 1.98 and maximum of 5.57, while the non-acceptable
has at the best an average of 3.73 and a maximum of 7.63 using Δ E94 ∗ .
3A JND is the smallest detectable difference between a starting and secondary level of a particular
sensory stimulus, in our case two color samples.
4 The JND while using Δ E ∗ should be closer to one than with other metrics but has still been
00
defined for simultaneous pair comparison of uniform color patches.
112 J.-B. Thomas et al.
quality color reproduction, and a consumer color reproduction, which aims only
at the preservation of the intended meaning, and relate the purpose with objective
indicators.
Considering the professional reproduction, let us consider the following rule of
thumb. If we want to reach a good accuracy, we need to consider two indicators:
the average and the maximum error. Let us consider the average: from 0 to 1 is
considered good, from 1 to 3 acceptable, and over 3 not acceptable. If now we
consider the maximum, from 0 to 3 is good, from 3 to 6 is acceptable, over is not
acceptable. If we compare this scale with the rule of thumb used by [42], it makes
sense since below three it is hardly perceptible, the same if we look at the work
of [1]. If we look at the JND proposed by [52] or [60] it seems to make sense
since in both cases, the good is under the JND. In this case we would prefer results
to be good, and it may be possible to discard a couple model/display if it does
not satisfy this condition. In the case of this professional reproduction, it could be
better to use the maximum error to discard a couple model/display. Considering the
consumer prediction, we propose to consider that from 0 to 3 it is good, from 3 to 6
it is acceptable, and over 6 it is not acceptable. In this case we would rather accept
methods that shows average results up to 6, since it should not spoil the meaning of
the reproduction. This is basically the same as the rule of thumb proposed by [42],
perceptible but acceptable being the basic idea of preserving the intended meaning.
Table 4.2 Qualitative interpretation of different models based on Table 4.1. The efficiency of a
model is dependent on several factors: the purpose, the number of measurements, the nature of the
data to measure, the computational cost, its accuracy, etc. All these parameters depend strongly on
each display
Polyharmonic
Model PLVC Bala PLCC* splines GOGO
Type of 54 (CIEXYZ) 1–3 visual 54 (Y) 216 (CIEXYZ) 3–54 (Y)
measurement measures tasks measures 3 measures measures 3
for 1–3 (CIEXYZ) (CIEXYZ)
pictures
Technology Dependent Dependent Dependent Independent CRT
Purpose Professional or Consumer Professional or Professional Consumer
Consumer Consumer
References
19. Brainard DH (1989) Calibration of a computer-controlled color monitor. Color Res Appl
14:23–34
20. Brainard DH, Pelli DG, Robson T (2002) Display characterization. Wiley, New York
21. Braun G, Ebner F, Fairchild M (1998) Color gamut mapping in a hue linearized cielab
colorspace. In: The proceedings of the IS&T/SID sixth color imaging conference: color
science, systems and applications, Springfield (VA), pp 346–350
22. Braun GJ, Fairchild MD (1999) Image lightness rescaling using sigmoidal contrast enhance-
ment functions. J Electron Imag 8(4):380
23. Catrysse PB, Wandell BA, *a PBC, W BA, B E, Gamal AE (1999) Comparative analysis of
color architectures for image sensors. In: Sampat N, Yeh T (eds) proceedings of SPIE, vol 3650,
pp 26–35
24. Chan JZ, Allebach JP, Bouman CA (1997) Sequential linear interpolation of multidimensional
functions. IEEE Trans Image Process 6:1231–1245
25. Choi SY, Luo MR, Rhodes PA, Heo EG, Choi IS (2007) Colorimetric characterization model
for plasma display panel. J Imag Sci Technol 51(4):337–347
26. Colantoni P, Thomas JB (2009) A color management process for real time color reconstruction
of multispectral images. In: Lecture notes in computer science, 16th Scandinavian conference,
SCIA, vol 5575, pp 128–137
27. Colantoni P, Stauder J, Blonde L (2005) Device and method for characterizing a colour device.
European Patent 05300165.7
28. Cowan WB (1983) An inexpensive scheme for calibration of a colour monitor in terms of CIE
standard coordinates. SIGGRAPH Comput Graph 17(3):315–321
29. Cowan W, Rowell N (1986) On the gun independency and phosphor constancy of color video
monitor. Color Res Appl 11:S34–S38
30. Day EA, Taplin L, Berns RS (2004) Colorimetric characterization of a computer-controlled
liquid crystal display. Color Res Appl 29(5):365–373
31. Dianat S, Mestha L, Mathew A (2006) Dynamic optimization algorithm for generating inverse
printer map with reduced measurements. Proceedings of international conference on acoustics,
speech and signal processing, IEEE 3
32. Dugay F, Farup I, Hardeberg JY (2008) Perceptual evaluation of color gamut mapping
algorithms. Color Res Appl 33(6):470–476
33. Fairchild M, Wyble D (1998) Colorimetric characterization of the Apple Studio display
(flat panel LCD). Munsell color science laboratory technical report
34. Farley WW, Gutmann JC (1980) Digital image processing systems and an approach to the
display of colors of specified chrominance. Technical report HFL-80-2/ONR-80, Virginia
Polytechnic Institute and State University, Blacksburg, VA
35. Farup I, Gatta C, Rizzi A (2007) A multiscale framework for spatial gamut mapping. Image
Process, IEEE Trans 16(10):2423–2435
36. Finlayson G, Hordley S, Hubel P (1998) Recovering device sensitivities with quadratic
programming. In: The proceedings of the IS&T/SID sixth color imaging conference: color
science, systems and applications, Springfield (VA): The society for imaging science and
technology, pp 90–95
37. Gatt A, Morovic J, Noriega L (2003) Colorimetric characterization of negative film for digital
cinema post-production. In: Color imaging conference, IS&T, pp 341–345
38. Gerhardt J (2007) Spectral color reproduction: model based and vector error diffusion
approaches. PhD thesis, Ecole Nationale Superieure des Telecommunications and Gjøvik
University College, URL [Link]
39. Gibson JE, Fairchild MD (2000) Colorimetric characterization of three computer displays
(LCD and CRT). Munsell color science laboratory technical report
40. Green P (ed) (2010) Color management: understanding and using ICC profiles. Wiley,
Chichester
41. Groff RE, Koditschek DE, Khargonekar PP (2000) Piecewise linear homeomorphisms: The
scalar case. In: IJCNN (3), pp 259–264
116 J.-B. Thomas et al.
42. Hardeberg J (1999) Acquisition and reproduction of colour images: colorimetric and multi-
spectral approaches. These de doctorat, Ecole Nationale Superieure des Telecommunications,
ENST, Paris, France
43. Hardeberg JY, Seime L, Skogstad T (2003) Colorimetric characterization of projection displays
using a digital colorimetric camera. In: Projection displays IX,, SPIE proceedings, vol 5002,
pp 51–61
44. Hardeberg J, Bando E, Pedersen M (2008) Evaluating colour image difference metrics for
gamut-mapped images. Coloration Technology 124:243–253
45. Hong G, Luo MR, Rhodes PA (2001) A study of digital camera colorimetric characterization
based on polynomial modeling. Color Res Appl 26(1):76–84
46. IEC:61966–3 (1999) Color measurement and management in multimedia systems and equip-
ment, part 3: equipment using CRT displays. IEC
47. International Color Consortium (2004) Image technology colour management – architecture,
profile format, and data structure. Specification ICC.1.2004–10
48. International Commission on Illumination (1996) The relationship between digital and colori-
metric data for computer-controlled CRT displays. CIE, Publ 122
49. Ishii A (2002) Color space conversion for the laser film recorder using 3-d lut. SMPTE J
16(11):525–532
50. Jimenez Del Barco L, Diaz JA, Jimenez JR, Rubino M (1995) Considerations on the calibration
of color displays assuming constant channel chromaticity. Color Res Appl 20:377–387
51. Johnson T (1996) Methods for characterizing colour scanners and digital cameras. Displays
16(4)
52. Kang HR (ed) (1997) Color technology for electronic imaging devices. SPIE Press, iSBN:
978-0819421081
53. Kasson LM, Nin SI, Plouffe W, Hafner JL (1995) Performing color space conversions with
three-dimensional linear interpolation. J Electron Imag 4(3):226–250
54. Katoh N, Deguchi T, Berns R (2001) An accurate characterization of CRT monitor
(i) verification of past studies and clarifications of gamma. Opt Rev 8(5):305–314
55. Katoh N, Deguchi T, Berns R (2001) An accurate characterization of CRT monitor (II) proposal
for an extension to CIE method and its verification. Opt Rev 8(5):397–408
56. Kimmel R, Shaked D, Elad M, Sobel I (2005) Space-dependent color gamut mapping:
A variational approach. IEEE Trans Image Process 14(6):796–803
57. Klassen R, Bala R, Klassen N (2005) Visually determining gamma for softcopy display.
In: Procedeedings of the thirteen’s color imaging conference, IS&T/SID, pp 234–238
58. Kwak Y, MacDonald L (2000) Characterisation of a desktop LCD projector. Displays
21(5):179–194
59. Kwak Y, Li C, MacDonald L (2003) Controling color of liquid-crystal displays. J Soc Inform
Disp 11(2):341–348
60. Mahy M, Eycken LVV, Oosterlinck A (1994) Evaluation of uniform color spaces developed
after the adoption of cielab and cieluv. Color Res Appl 19(2):105–121
61. Marcu GG, Chen W, Chen K, Graffagnino P, Andrade O (2001) Color characterization issues
for TFT-LCD displays. In: Color imaging: Device-independent color, color hardcopy, and
applications VII, SPIE, SPIE Proceedings, vol 4663, pp 187–198
62. Mikalsen EB, Hardeberg JY, Thomas JB (2008) Verification and extension of a camera-based
end-user calibration method for projection displays. In: CGIV, pp 575–579
63. Morovic J (2008) Color gamut mapping. Wiley, Chichester
64. Neugebauer HEJ (1937) Die theoretischen Grundlagen des Mehrfarbendruckes. Zeitschrift für
wissenschaftliche Photographie, Photophysik und Photochemie 36(4):73–89
65. Neumann A, Artusi A, Zotti G, Neumann L, Purgathofer W (2003) Interactive perception based
model for characterization of display device. In: Color imaging IX: processing, hardcopy, and
applications IX, SPIE Proc., vol 5293, pp 232–241
66. Nielson GM, Hagen H, Müller H (eds) (1997) Scientific visualization, overviews, method-
ologies, and techniques. IEEE Computer Society, see [Link]
Visualization-Overviews-Methodologies-Techniques/dp/0818677775
4 Cross-Media Color Reproduction and Display Characterization 117
94. Yoshida Y, Yamamoto Y (2002) Color calibration of LCDs. In: Tenth color imaging con-
ference, IS&T - The society for imaging science and technology, pp 305–311, Scottsdale,
Arizona, USA
95. Yule JAC, Nielsen WJ (1951) The penetration of light into paper and its effect on halftone
reproductions. In: TAGA Proceedings, vol 3, p 65
Chapter 5
Dihedral Color Filtering
Abstract Linear filter systems are used in low-level image processing to analyze
the visual properties of small image patches. We show first how to use the theory
of group representations to construct filter systems that are both steerable and are
minimum mean squared error solutions. The underlying groups are the dihedral
groups and the permutation groups and the resulting filter systems define a transform
which has many properties in common with the well-known discrete Fourier
transform. We also show that the theory of extreme value distributions provides
a framework to investigate the statistical properties of the vast majority of the
computed filter responses. These distributions are completely characterized by only
three parameters and in applications involving huge numbers of such distributions,
they provide very compact and efficient descriptors of the visual properties of the
images. We compare these descriptors with more conventional methods based on
histograms and show how they can be used for re-ranking (finding typical images in
a class of images) and classification.
5.1 Introduction
Linear filter systems are used in low-level image processing to analyze the visual
properties of small image patches. The patches are analyzed by computing the
similarity between the patch and a number of fixed filter functions. These similarity
values are used as descriptors of the analyzed patch. The most important step in this
approach is the selection of the filter functions. Two popular methods to select them
are the minimum mean squared error (MMSE) criterion and the invariance/steerable
filter approach. The MMSE method, like the jpeg transform coding, selects those
filters that allow a reconstruction of the original patch with a minimal statistical
error. The invariance, or the more general steerable, filter system assumes that
the patches of interest come in different variations and that one fixed selection of
filters can detect both, if a given patch is of interest and if this is the case which
variant of these patches it is. A typical example is an edge detector that is used to
detect if a given patch was an edge and in the case of an edge it gives a possible
estimate of its orientation (see [4, 15] for a comparison of different types of local
descriptors). For the case of digital color images, we will show how the theory of
group representations can be used to construct such steerable filter systems and that,
under fairly general conditions, these filters systems are of the MMSE type.
The general results from representation theory show that the filter functions
implement a transform which has many properties in common with the well-known
discrete Fourier transform. One of these properties that is of interest in practical
applications is the existence of fast transforms. As is the case for the Fourier
transform where the DFT can be computed efficiently by using the FFT, it can be
shown here that the basic color filter operations can be optimized by computing
intermediate results. Apart of the speedup achievable by reducing the number of
necessary arithmetic operations, we will also see that the bulk of the computations
are simple additions and subtractions which make these filters suitable for hardware
implementations and applications where a huge number of images have to be
processed. A typical example of such a task is image retrieval from huge image
databases. Such databases can contain millions or billions of images that have to
be indexed and often it is also necessary to retrieve images from such a database at
very high speed. We will therefore illustrate some properties of these filter systems
by investigating properties of image databases harvested from websites.
In our illustrations, we will show that the theory of extreme value distributions
provides a framework to investigate the statistical properties of the vast majority
of the computed filter responses. These distributions are completely characterized
by only three parameters and in applications involving huge numbers of such
distributions, they provide very compact and efficient descriptors of the visual
properties of the images. We will also compare these descriptors with more
conventional methods based on histograms and show how they can be used for re-
ranking (finding typical images in a class of images) and classification.
5 Dihedral Color Filtering 121
In the following, we will only consider digital color images defined on a square grid.
This is the most important case in practice, but similar results can be derived for
images on hexagonal grids. We also assume that the pixel values are RGB vectors.
Our first goal is to identify natural transformations that modify spatial configurations
of RGB vectors.
We recall that a group is a set of elements with a combination rule that maps a
pair of elements to another element in the set. The combination rule is associative,
every element has an inverse and there is a neutral element. More information about
group theory can be found in every algebra textbook and especially in [3]. In the
following, we will only deal with dihedral groups. Such a group is defined as the
set of all transformations that map a regular polygon into itself. In the following,
we will use the dihedral group D4 of the symmetry transformations of the square to
describe the transformation rules of the grid on which the images are defined. We
will also use the group D3 formed by the symmetry transformations of a triangle
to describe the modifications of RGB vectors. A description of the usage of the
dihedral groups to model the geometric properties of the sensor grid can be found
in [8, 9]. This was then extended to the investigation of color images in [10, 11].
Here we will not describe the application of the same ideas in the study of RGB
histograms which can be found in [13].
It can be shown that the dihedral group Dn of the n-sided regular polygon consists
of 2n elements. The symmetry group D4 of the square grid has eight elements: the
four rotations ρk with rotation angles k = kπ /2, k = 0, . . . , 3 and the reflection σ
on one of the diagonals combined with one of the four rotations. The elements are
thus given by ρ0 , . . . , ρ3 , ρ0 σ , . . . , ρ3 σ . This is a general property of the dihedral
groups where we find for the hexagonal grid the corresponding transformation
group D6 consisting of twelve elements, six rotations and six rotations combined
with a reflection. For the RGB vectors, we consider the R, G, and B channel as
represented by corner points of an equilateral triangle. The symmetry group of the
RGB space is thus the group D3 . It has six elements ρ0 , ρ1 , ρ2 , ρ0 σ , ρ1 σ , ρ2 σ where
the ρk now represent rotations with rotation angle k · 120◦ and σ is the reflection on
one fixed symmetry axis of the triangle. Algebraically, this group is identical to the
permutation group S(3) of three elements.
We introduce the following notation to describe the general situation: For a
group G with elements g and a set Z with elements z we say that G operates on Z
if there is a mapping (G, Z) → Z such that (g, z)
→ gz and - - (g2 g1 )z = g2 (g1 z); we
also say that G is a transformation group. Furthermore, -Z- denotes the number of
- -
elements in the set and -G- is the number of the group elements. In the context of
the dihedral groups the set Z consists -of the - n corner points of the regular polygon,
the transformation group is Dn and -Dn - = 2n. In the following, we need more
general point configurations than the corner points of the polygon and therefore
we introduce X as a set of points in the 2D-plane and we will use x for its elements.
Such a set X is the collection of points on which the filter functions are defined. As a
122 R. Lenz et al.
The mapping p
→ p(g,h) is a linear mapping of pattern space and we can therefore
describe it by a matrix T(g, h) such that p(g,h) = T(g, h)p. We calculate:
and we see that the matrices satisfy T(g1 g2 , h1 h2 ) = T(g1 , h1 )T(g2 , h2 ). This shows
that this rule defines a (matrix) representation of the group which is a mapping from
the group into a space of matrices such that the group operation maps to matrix
multiplication.
Matrices describe linear mappings between vector spaces in a given coordinate
system. Changing the basis in the vector space gives a new description of the
same linear transformations by different matrices. Changing the basis in the pattern
space P using a matrix B will replace the representation matrices T(g, h) by
the matrices BT(g, h)B−1 . It is therefore natural to construct matrices B that
simplify the matrices BT(g, h)B−1 for all group elements (g, h) simultaneously. The
following theorem from the representation theory of finite groups collects the basic
results that give a complete overview over the relevant properties of these reduced
matrices (see [3, 19, 20] for details):
Theorem 1. – We can always find a matrix B such that all BT(g, h)B−1 are block-
diagonal with blocks Tm of minimum size.
(4) (3)
– These smallest blocks are of the form T(g, h) = Ti (g) ⊗ T j (h) where ⊗
denotes the Kronecker product of matrices.
(4) (3)
– The dimensions of Ti (g) and T j (h) are one or two.
(4) (3)
– Both, the Ti and the T j , are representations and from their transformation
properties follows that it is sufficient to know them for one rotation and the
reflection: T(.) (ρ k σ l ) = (T(.) (ρ ))k (T(.) (σ ))l ).
For the group D3 operating on RGB vectors, it is easy to see that the trans-
formation (R, G, B)
→ R + G + B is a projection on a one-dimensional subspace
invariant under all elements in D3 . The first block of the matrices is therefore
one-dimensional and we have T(3) (h) = 1 for all h ∈ D3 . This defines the trivial
representation of the group and the one-dimensional subspace of the RGB space
defined by this transformation property is the space of all gray value vectors. The
other block is two-dimensional and given by the orthogonal complement to this
one-dimensional invariant subspace. This two-dimensional complement defines the
space of complementary colors. For the group D4 , a complete list of its smallest
representations can be found in Table 5.1.
Given an arbitrary set X closed under D4 , the tools from the representation
theory of finite groups provide algorithms to construct the matrix B such that the
transformed matrices BT(g, h)B−1 are block-diagonal with minimum-sized blocks.
For details, we refer again to the literature [3, 9, 19, 20].
124 R. Lenz et al.
Next, we use the tensor representation construction and find for the structure of the
full 48-dimensional space the decomposition
5.3 Illustration
We will now illustrate some properties of these filter systems with the help of a
single image. We use an image of size 192 × 128 and filters of size 4 × 4. The
properties of the filter results are of course depending on the relation between
the resolution of the original image and the filter size where images with a high
resolution will on average contain more homogeneous regions. We selected the
combination of a small image size and small filter size since we will later use this
combination in an application where we will investigate the statistical properties of
databases consisting of very many thumbnails harvested by an image search engine
from the internet.
In Fig. 5.2, we see the original image with the two parrots in the upper left corner
and the result of the 48 different filters. The first 16 filter results are computed from
the intensity channel and the remaining from the 32 color-opponent combinations
corresponding to the splitting of the original 48-dimensional vector space described
in Table 5.2. The images show the magnitude of the filter responses. In the case
of the intensity-based filters, this means that a pattern and its inverted copy will
produce the same filter result in this figure. The colormap used in Fig. 5.2 and the
following figures is shown below the filter results. From the figure, we can see that
the highest filter responses are obtained by the filters that are related to the spatial
averaging, i.e., those belonging to vector spaces Vti and Vti . We see also that the
intensity filters have in general higher responses than their corresponding color-
opponent filters. From the construction (see (5.2)), we saw that these 48 filters
come in 24 packages of length one, two, and four where all filters in a package have
the same transformation properties and the norm of the filter vectors is invariant
under spatial and color transformations. The norm of these 24 filter packages is
shown in Fig. 5.3. Again, we see the highest response for the three spatial averaging
filters (besides the original) and the three spatial averaging filters combined with
the color-opponent colors (last two in the third row and first in the fourth row).
In both, Figs. 5.2 and 5.3 the filter results are scaled such that the highest value for
all filter responses is one. This makes it possible to see the relative importance of
the different filters and filter packages. In Fig. 5.4, we show the magnitude of the
126 R. Lenz et al.
Fig. 5.4 Selected line and edge filters. Dark corresponds to large magnitude and light corresponds
to low magnitude
filter response vectors for four typical filter packages related to the vector spaces of
types Vai ,V2i ,Vac , and V2c . Visually, they correspond to line and edge filters in the
intensity and the color-opponent images.
We call such a vector a feature vector and the corresponding vector space the feature
space.
128 R. Lenz et al.
We now divide the pattern space P into the smallest subspaces under the
transformation group introduced above. It is then easy to see that the filter systems
defined as the projection operators onto these invariant subspaces have two prop-
erties that are of importance in applications: invariance and steerability. Consider
a filter system F that is a projection on such an invariant subspace. From the
construction, we know that a pattern p in this subspace can be described by a coor-
dinate vector F(p) and since the subspace is invariant we find for the transformed
pattern p(g,h) a transformation matrix T(g, h) such that F(p(g,h) ) = T(g, h)F(p).
From the general theory, it can also be shown that we can always choose 3 the 3
coordinate system such that the matrices T(g, h) are orthonormal. The norm 3F(p)3
of the feature vector is thus invariant under all group transformations. Due to the
symmetry of the scalar product we can also apply the transformations of the group to
the filter functions and thus modify their behavior. The feature vectors obtained from
the modified filter systems are also related via the matrix multiplication with T(g, h)
and we see that we can generate all possible modifications of these feature vectors
from any given instance of it. A formal description is the following: a filter system
is steerable if it satisfies the following condition:
1 2 1 2 1 2
4 h) F, p for an L × L matrix T(g,
F, p(g,h) = F, T(g, h)p = T(g, 4 h).
If we collect all filter coefficients in the matrix F, then a steerable filter system
4 h)F.
satisfies the matrix equations FT(g, h) = T(g,
For a fixed pattern p0 , we can generate its orbit (D4 ⊗ D3 )p0 in pattern space
4 h)F(p0 ) define
and if F is a steerable filter system then the feature vectors T(g,
an orbit1 in feature
2 space. Steerable filters have the advantage that the actual filter
vector F, p0 is computed once all the transformed versions can be computed with
4 h)F.
the closed form expression T(g,
In summary, we constructed filter systems F with the following properties:
computed from the intensity channel, whereas the remaining (17–48) are related
to the two-dimensional complementary color distributions. The first block (1–3 in
intensity, 17–22 in complementary color) is computed by spatial averaging, the
second (4–6 and 23–28) to line-like structures, the third (7–10, 29–44) to edges,
and the last (11–12, 45–48) is related to the one-dimensional representations p and
m (see (5.1)) from the inner orbit. The left diagram shows the values computed from
the [Link] images, the right diagram is based on the DPChallenge images. We
see that the non-negativity of the first filter results leads to significant correlations
with the other filter results. This leads to the structures in the first column and the
last rows in the figures in the left column. The structure in the remaining part of the
matrices can be clearly seen in the two images in the right column.
In Fig. 5.7, the structure of the full second-order moment matrices of the filtered
results is shown. On the left side of the figure the full matrices are shown; in the
right column, the rows and columns related to the averanging filters are removed to
enhance the visibility of the structure in the remaining part. In the upper row, the
results for the [Link] and in the lower row the DPChallenge database were used.
We may further explore the statistical properties of the filters with the help of the
following simple model: consider a black-box unit U with input X the pixel values
from a finite-sized window in a digital image (a similar analogy can be applied to
the receptive fields of a biological vision system). The purpose of this black box is to
5 Dihedral Color Filtering 131
Fig. 5.6 Log Diagonal of the second-order moment matrices after filtering (a) [Link]
(b) DPChallenge
measure the amount of some non-negative quantity X(t) that changes over 5
time. We
write this as u(t) = U(X(t)). We also define an accumulator s(n) = 0n u(t)dt that
accumulates the measured output from the unit until it reaches a certain threshold
s(n) = Max(n) (X) or a certain period of time, above which the accumulator is reset
to zero and the process is restarted.
If we consider u(t), s(n) as stochastic processes and select a finite number N
of random samples u1 , . . . , uN , then their joint distribution J(u1 , . . . , uN ) and the
distribution Y (sN ) of sN , depend on the original distribution F(XN ). At this point,
we may pose two questions:
1. When N → ∞ is there a limiting form of Y (s) → Φ (s)?
2. If there exists such a limit distribution, what are the properties of the black-box
unit U and of J(u1 , . . . , uN ) that determines the form of Φ (s)?
132 R. Lenz et al.
Fig. 5.7 Second-order moment matrices (a) Full Matrix [Link] (b) No averaging filters
[Link] (c) Full Matrix DPChallenge (d) No averaging filters DPChallenge
In [1], the authors have demonstrated that under certain conditions on Y (s) the
possible limiting forms of Φ (s) are given by the three distribution families:
μ −s
Φ (s) = exp − exp , ∀s Gumbel,
σ
) *
s−μ k
Φ (s) = 1 − exp − , s > μ Weibull,
σ
) *
s − μ −k
Φ (s) = exp − , s > μ Fréchet, (5.3)
σ
where μ , σ , k are the location, scale, and shape parameters of the distributions,
respectively. The particular choice between the three families in (5.3) is determined
by the tail behavior of F(X) at large X. In this case, we use as units U the black
box that computes the absolute value of the filter result vectors from the irreducible
5 Dihedral Color Filtering 133
Fig. 5.8 Image type and model distribution in EVT parameter space
representations of the dihedral groups. The filter vectors not associated with the
trivial representation are of the form s = ∑(xi − x j ) where xi , x j are pixel values.
We can therefore expect that these filter values are usually very small and that high
values will appear very seldom. In addition, these sums are calculated over a small,
finite neighborhood, and for this reason, the random variables are highly correlated.
In short, the output for each filter has a form similar to the sums described in [1],
and so it is possible to use the EVT to model their distribution.
We may now analyze which types of images are assigned to each submodel
in (5.3). For economy of space, we only illustrate a single filter (an intensity
edge filter) on the dataset described in Sect. 5.7.1, but the results generalize to all
filters and different datasets. We omit the μ parameter since it usually exhibits
very little variation and the most important behavior is observed in the other two
parameters. First of all, if we look at Fig. 5.8 we see a correlated dispersion in the
two axes, with the Fréchet images spanning only a very small region of the space
at low σ , k, and well separated from 2-parameter and 3-parameter Weibull. Also
notice how the Fréchet set typically includes images with near-uniform colored
regions with smooth transitions between them, or alternatively very coarse-textured,
homogeneous regions with sharp boundaries. High frequency textures seem to be
relatively absent from the Fréchet, and on average the image intensities seem to be
lower in the Fréchet set than in the Weibulls.
On the other hand, the 2-parameter and 3-parameter Weibull clusters are
intermixed, with the 2-parameter mostly restricted to the lower portion of the space.
For smaller σ , k values, the 2-parameter Weibull images exhibit coarser textures,
with the latter becoming more fine-grained as σ , k increase in tandem. Also, there
134 R. Lenz et al.
Fig. 5.9 A comparison between the extrema and other regions of a filtered image (a) Original
image (b) Edge filter result (c) Tails (maxima) (d) Mode (e) Median (f) Synthesis
From the construction of the filters follows that they can be implemented as a
combination of three basic transforms: one operating on the RGB vectors, one
for the four-point, and one for the eight-point orbit. These filters are linear and
they are therefore completely characterized by three matrices of sizes 3 × 3, 4 ×
4, and 8 × 8. The rows of these matrices define projection operators onto the
spaces Vxy introduced above and they can be computed using algorithms from the
representation theory of finite groups.
We illustrate the basic idea with the help of the transformation matrix related
to the RGB transformation. We already noted that the sum R+G+B is invariant
under permutations of the RGB channels. It follows that the vector 1 1 1
defines a projection onto this invariant subspace. We also know that the orthogonal
complement to this one-dimensional subspace defines a two-dimensional
subspace
that cannot be reduced further. Any two vectors orthogonal to 1 1 1 can therefore
be used to fill the remaining two rows of the RGB transformation matrix. Among the
possible choices we mention here two: the Fourier transform and an integer-valued
transform. The Fourier transform is a natural choice considering the interpretation of
the RGB vector as three points on a triangle. In this case, the remaining two matrix
rows are given by cos(2kπ /3) and sin(2kπ /3). Since the filters are applied to a large
number of vectors, it is important to find implementations
that
are computationally
efficient. One solution is given by the two vectors 1 −1 0 and 1 1 −2 . The
resulting transformation matrix has the advantage that the filters can be implemented
using addition and subtraction only. We can furthermore reduce the number of
operations by computing intermediate sums. One solution is to compute RG =
R+G first and combine that afterward to obtain RG+B and RG-2B. The complete
transformation can therefore be computed with the help of five operations instead
of the six operations required by a direct implementation of the matrix–vector
operation.
For the four- and the eight-point orbit transforms, we can use the general tools of
representation theory to find the integer-valued transform matrices:
⎛ ⎞
1 1 1 1 1 1 1 1
⎜ 1 1 −1 −1 1 1 −1 −1 ⎟
⎛ ⎞ ⎜ ⎟
⎜ 1 −1 1 −1 1 −1 1 −1 ⎟
1 1 1 1 ⎜ ⎟
⎜ −1 1 −1 1 ⎟ ⎜ ⎟
⎜ ⎟ ⎜ −1 1 1 −1 −1 1 1 −1 ⎟
⎝ −1 −1 1 1 ⎠ and ⎜ ⎟. (5.4)
⎜ 0 −1 −1 0 0 1 1 0 ⎟
⎜ ⎟
1 −1 −1 1 ⎜ 1 0 0 −1 −1 0 0 1 ⎟
⎜ ⎟
⎝ −1 0 0 −1 1 0 0 1 ⎠
0 1 −1 0 0 −1 1 0
Also here we see that we can use intermediate sums to reduce the number of
operations necessary. This is an example of a general method to construct fast-group
theoretical transforms similar to the FFT implementation of the Fourier transform.
More information can be found in [18].
136 R. Lenz et al.
under the group D4 are related to scaling. In the simplest case of the three averaging
intensity filters, one gets as filter results the vector (F1 , F2 , F3 ) representing the
average intensity over the three different rings in the 4 × 4 patch. Assuming that
the scaling is such that one can, on average, interchange all three orbits then one
can treat the three filter results F1 , F2 , F3 as function values defined on the corners
of a triangle. In that case, one can use the same strategy as for the RGB components
and apply the D3 -based filters to the vectors (F1 , F2 , F3 ). The first filter will then
compute the average value over three different scales. Its visual effect is a blurring.
The second filter computes the difference between the intensities in the inner four-
pixel patch and the intensities on the eight points on the next orbit. The visual
property is a center-surround filter. Finally, the third filter computes the difference
between the two inner and the outer orbit. Also, here we can convert the vectors with
the last two filter results to polar coordinates to obtain a magnitude “blob-detector”
and an angular phase-like result. We use the same color coding as for the edge-
filter illustration in Fig. 5.10 and show the result of these three scale-based filters
in Fig. 5.11. We don’t describe this construction in detail but we only show its
effect on the values of the diagonal elements in the second-order moment matrix. In
Fig. 5.12, we show the logarithms of the absolute values of the diagonal elements
in the second-order moment matrix computed from the filtered patches as before
(marked by crosses) and the result of the scaling operation (given by the circles).
For the first three filters, we see that the value of the first component increased
138 R. Lenz et al.
significantly while the values of the other two decreased correspondingly. This is a
typical effect that can also be observed for the other filter packages. We conclude
the descriptions of these generalizations by remarking that this is an illustration
showing that one can use the representation theory of the group D4 ⊗ D3 ⊗ D3 to
incorporate scaling properties into the framework.
Among possible applications that can be based on the presented filter systems we
will here illustrate the usefulness in an image classification experiment, where we
try to separate classes of images downloaded from the Internet. A popular approach
for large-scale image classification is to combine global or local image histograms
with a supervised learning algorithm. Here we derive a 16 bins histogram for each
filter package, resulting in a 16 × 24 representation of each image. The learning
algorithm is the Support Vector Machine implementation SVMlight described in [6].
For simplicity and reproducibility reasons, all experiments are carried out with
default settings.
Two-class classification results are illustrated for the keyword pairs garden–
beach and andy warhol–claude monet. We believe these pairs to be representa-
tive examples of various tasks that can be encountered in image classification.
5 Dihedral Color Filtering 139
Table 5.3 Two-class classification accuracy for various filter packages, and the overall descriptor
Filter package
Keyword pair 1:3 4:6 7:10 13:15 16:18 19:22 ALL ALL+EVT
Garden–beach 0.78 0.77 0.79 0.69 0.68 0.67 0.80 0.79
Andy warhol–claude monet 0.66 0.81 0.81 0.80 0.82 0.83 0.89 0.84
The Picsearch1 image search service is queried with each keyword, and 500
thumbnail images (maximum size 128 pixels) are saved from each search result.
Based on recorded user statistics, we only save the most popular images in each
category, which we assume will increase the relevance of each class. The popularity
estimate is based on the ratio between how many times an image has been clicked
and viewed in the public search interface. For each classification task, we create
a training set containing every second image from both keywords in the pair, and
remaining images are used for evaluating the classifier.
Classification results are given in Table 5.3. The overall result for the entire
descriptor (the entire 16 × 24 representation) is shown in the second to last column,
and classification results in earlier columns are based on selected filter packages.
The last column summarizes the classification results obtained from the EVT-
parameter descriptions of the distributions. The classification accuracy is given by
the proportion of correctly labeled images. A value of 0.75 means that 75% of the
images were labeled with a correct label. We conclude that the best classification
result is obtained when the entire descriptor is used. But as the table indicates, the
importance of different filter packages varies with the image categories in use. We
see, for instance, that the color content of an image (captured in filter packages
13–24) is more important for the andy warhol–claude monet classification result,
than for garden–beach.
We illustrate the classification result by plotting subsets of classified images. The
result based on the entire descriptor, for each keyword pair respectively, can be seen
in Fig. 5.13a, b. Each sub-figure shows the 10+10 images than obtained the most
positive and most negative score from the Support Vector Machine. Similar plots
for selected filter packages are shown in Figs. 5.14–5.19.
In closing, we briefly illustrate the practical application of the Extreme Value
theory models in the above classification examples and as an alternative represen-
tation to histograms. The input data vector to the SVM in this case contains the
three parameters: location, scale, shape estimated by fitting the EVT models to each
1 [Link]
140 R. Lenz et al.
Fig. 5.13 Classification examples based on the entire descriptor (filter results 1–24) (a) beach
(top) vs garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
Fig. 5.14 Classification examples based on filter package: 1–3 (intensity mean) (a) beach (top) vs
garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
of the 24 filters packages. Compared with the histogram from before, we are now
only using a 3 × 24-dimensional vector for the full filter descriptor, as opposed to a
16 × 24-dimensional vector. This leads to a much reduced data representation, faster
training and classification steps, and no need to optimally set the number of bins.
First, in Fig. 5.20 we show the comparative results for a single filter classification
on the andy warhol–claude monet set. We can see that the EVT, even with its lower
dimensionality, is equally or sometimes even more accurate than the histogram
representations. In terms of absolute accuracy numbers, the EVT scores for the full-
filter descriptor are shown in the last column of Table 5.3. As it is obvious, these
scores are very close to the histogram-based results.
5 Dihedral Color Filtering 141
Fig. 5.15 Classification examples based on filter package: 4–6 (intensity lines) (a) beach (top) vs
garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
Fig. 5.16 Classification examples based on filter package: 7–10 (intensity edges) (a) beach (top)
vs garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
Fig. 5.17 Classification examples based on filter package: 13–15 (color mean) (a) beach (top) vs
garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
Fig. 5.18 Classification examples based on filter package: 16–18 (color lines) (a) beach (top) vs
garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
5.8 Summary
We started from the obvious observations that the pixels of digital images are located
on grids and that, on average, the three color channels are interchangeable. These
two properties motivated the application of tools from the representation theory of
5 Dihedral Color Filtering 143
Fig. 5.19 Classification examples based on filter package: 19–22 (color edges) (a) beach (top) vs
garden (bottom) (b) andy warhol (top) vs claude monet (bottom)
Fig. 5.20 Two-class classification accuracy comparison between the EVT and the histogram
representations for the andy warhol–claude monet set. We only use a single filter paclage at a
time
finite groups and we showed that in this framework, we can explain how steerable
filter systems and MMSE-based transform-coding methods are all linked to those
group theoretical symmetry properties. Apart from these theoretical properties,
the representation theory also provides algorithms that can be used to construct the
filter coefficients automatically and it also shows how to create fast filter systems
using the same principles as the FFT-implementations of the DFT. We also sketched
briefly how the group structure can be used to define natural bins for the histogram
descriptors of orientation parameters. A generalization that includes simple scaling
properties was also sketched.
144 R. Lenz et al.
Fig. 5.21 Retrieval results from the four classes using EVT (a) beach (top) vs garden (bottom)
(b) andy warhol (top) vs claude monet (bottom)
References
1. Bertin E, Clusel M (2006) Generalised extreme value statistics and sum of correlated variables.
J Phys A: Math Gen 39(24):7607–7619
2. Datta R, Li J, Wang JZ (2008) Algorithmic inferencing of aesthetics and emotion in natural
images: An exposition. In: 2008 IEEE International Conference on Image Processing, ICIP
2008, pp 105–108, San Diego, CA
3. Fässler A, Stiefel EL (1992) Group theoretical methods and their applications. Birkhäuser,
Boston
4. Freeman WT, Adelson EH (1991) The design and use of steerable filters. IEEE Trans Pattern
Anal Mach Intell 13(9):891–906
5. Hubel DH (1988) Eye, brain, and vision. Scientific American Library, New York
6. Joachims T (1999) Making large-scale support vector machine learning practical. MIT Press,
Cambridge, pp 169–184
7. Lenz R (1990) Group theoretical methods in image processing. Lecture notes in computer
science (Vol. 413). Springer, Heidelberg
8. Lenz R (1993) Using representations of the dihedral groups in the design of early vision filters.
In: Proceedings of international conference on acoustics, speech, and signal processing, pp
V165–V168. IEEE
9. Lenz R (1995) Investigation of receptive fields using representations of dihedral groups. J Vis
Comm Image Represent 6(3):209–227
10. Lenz R, Bui TH, Takase K (2005) Fast low-level filter systems for multispectral color images.
In: Nieves JL, Hernandez-Andres J (eds) Proceedings of 10th congress of the international
colour association, vol 1, pp 535–538. International color association
11. Lenz R, Bui TH, Takase K (2005) A group theoretical toolbox for color image operators. In:
Proceedings of ICIP 05, pp III–557–III–560. IEEE, September 2005
12. Lenz R, Carmona PL (2009) Octahedral transforms for 3-d image processing. IEEE Trans
Image Process 18(12):2618–2628
13. Lenz R, Carmona PL (2010) Hierarchical s(3)-coding of rgb histograms. In: Ranchordas A
et al (eds) Selected papers from VISAPP 2009, vol 68 of Communications in computer and
information science. Springer, Berlin, pp 188–200
14. Lowe DG (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis
60(2):91–110
15. Mikolajczyk K, Schmid C (2005) A performance evaluation of local descriptors. IEEE Trans
Pattern Anal Mach Intell 27(10):1615–1630
16. Oliva A, Torralba A (2006) Building the Gist of a Scene: The Role of Global Image Features
in Recognition. Progress in Brain Research 155:23-36
17. Olshausen BA, Field DJ (1996) Emergence of simple-cell receptive field properties by learning
a sparse code for natural images. Nature 381(6583):607–609
18. Rockmore D (2004) Recent progress and applications in group FFTs. In: Byrnes J, Ostheimer
G (eds) Computational noncommutative algebra and applications. Kluwer, Dordrecht
19. Serre J-P (1977) Linear representations of finite groups. Springer, New York
20. Terras A (1999) Fourier analysis on finite groups and applications. Cambridge University
Press, Cambridge
21. Thorpe S, Fize D, Marlot C (1996) Speed of processing in the human visual system. Nature
381(6582):520–522
22. Yosida K (1980) Functional analysis. Springer, Berlin
Chapter 6
Color Representation and Processes
with Clifford Algebra
P. Carré ()
Laboratory XLIM-SIC, UMR CNRS 7252, University of Poitiers, France
e-mail: [Link]@[Link]
M. Berthier
Laboratory MIA (Mathématiques, Images et Applications), University of La Rochelle, France
e-mail: [Link]@[Link]
6.1 Introduction
In this first section, we start with a brief description of basic concepts of Quaternion
and Clifford algebra.
q = S(q) + V(q)
where r(x, y), g(x, y), and b(x, y) are the red, green, and blue components of the
image.
From a colour described in RGB colour space with a quaternion vector q ∈ P,
HSV colour space coordinates can be found as well with operations on quaternions.
We consider that Value is the norm of the colour’s orthogonal projection vector
(q.μgrey )μgrey on the grey axis μgrey (this axis can be defined such that μgrey = i+√j+k
3
.
Saturation and Hue are represented on the orthogonal plan to the grey axis which
crosses (q.μgrey )μgrey . The Saturation is the distance between the colour vector q
and the grey axis μgrey , and Hue is the angle between the colour vector q and a
colour vector ν taken anywhere on the plan orthogonal to μgrey and which sets the
reference zero Hue angle. This reference Hue value is often taken to represent the
red colour vector, so we decided arbitrarily to associate the red colour vector or any
other one to the ν vector and gave it a zero Hue value (Fig. 6.1). Hue is the angle
between this reference colour vector and the colour vector q.
If q is a colour vector, then Value V , Saturation S, and Hue H can be given with
the grey-axis μgrey ∈ S∩P and the reference colour vector ν ∈ S∩P with elementary
quaternionic operations as below
⎧ μν qν μ |
⎪
⎪ H=tan−1 |q−
⎪
⎪ |q−ν qν |
⎨
S=| 12 (q + μ qμ )| . (6.2)
⎪
⎪
⎪
⎪
⎩ V =| 12 (q − μ qμ )|
After a first view of several possible manipulations on colour image encoded with
quaternions, we focus on applications which can be done with. We propose to study
low level operations on colour images like filtering operations for instance.
6 Color Representation and Processes with Clifford Algebra 151
[Link] Definition
The Clifford algebra framework allows to encode geometric transformations via al-
gebraic formulas. Let us fix an orthonormal basis (e1 , e2 , e3 ) of the vector space R3 .
We embed this space in a larger 8-dimensional vector space, denoted R3,0 , with
basis given by a unit 1, the three vectors e1 , e2 , e3 and the formal products e1 e2 ,
e2 e3 , e1 e3 , and e1 e2 e3 . The key point is that elements of R3,0 can be multiplied: the
product of ei and e j is, for example, ei e j . The rules of multiplication are given by:
e2i = 1, ei e j = −e j ei .
uv = u · v + u ∧ v, (6.5)
where · denotes the scalar product and u ∧ v is the bivector generated by u and v.
Since the ei ’s are orthogonal, then
ei e j = ei ∧ e j .
The linear combinations of the elements ei ∧ e j are called grade-2 entities termed
bivectors. They encode pieces of two-dimensional vector subspaces of R3 with
a magnitude and an orientation. In these algebras, multivectors which are the
extension of vectors to higher dimensions, are the basic elements. One example
is that we often represent vectors as one-dimensional directed quantities (also
represented by arrows), they are thus represented in geometric algebras by 1-vectors.
As their dimension is one, they are said 1-graded. By extension, in geometric
6 Color Representation and Processes with Clifford Algebra 153
algebra, there are grade-2 entities termed bivectors which are plane segments
endowed with orientation. In general, a k-dimensional oriented entity is known as a
k-vector. For an overview on geometric algebras see [4–7] for instance. In geometric
algebra, oriented subspaces are basic elements, as vectors in a m-dimensional linear
vector space V m . These oriented subspaces are called blades, and the term k-blade
is used to describe a k-dimensional homogeneous subspace. A multivector is then a
linear combination of blades.
Any multivector M1 ∈ Rn,0 is so described by the following equation:
n
M= ∑ Mk (6.6)
k=0
with Mk the k-vector part of any multivector M i.e., the grade k operator.
The Geometric product is an associative law and distributive over the addition of
multivectors. In general, the result of the geometric product is a multivector. As we
have said, if used on 1-vectors a and b, this is the sum of the inner and outer product:
ab = a.b + a ∧ b. Note also that the geometric product is not commutative.
This product is used to construct k-dimensional subspace elements from inde-
pendent combinations of blades. For example, multiplying with this product the two
independent 1-vectors e1 and e2 gets the bi-vector e12 . And if you multiply again
this bivector by the third 1-vector in V 3 , e3 , you get the trivector e123 . The basis of
R3,0 algebra is so given by (e0 , e1 , e2 , e3 , e23 , e31 , e12 , e123 ). Here e0 stands for the
element of grade 0, it is so the scalar part.
• External or Wedge product: The wedge product is denoted by ∧.
Wedge product can then be described using geometric product as follow with
A a s-graded multivector and B a r-graded multivector, both in Rn,0 :
• Inner Product: This product also called interior product denoted by ., is used
to give the notion of orthogonality between two multivectors [4]. Let A be a
a-vector and B be a b-vector, then A.B is the subspace of B, with a (b − a)
dimension, orthogonal to subspace A. If b < a then A.B = 0 and A is orthogonal
to B. For 1-vectors, the inner product equals the scalar product used in linear
algebra V m .
Ar .Bs = Ar Bs |r−s| . (6.8)
• Scalar Product: This product denoted by ∗, is used to define distances and
modulus.
A ∗ B ≡ AB0 .
1 Inthe following, little letters will be used to represent 1-vectors whereas bolded capital letters
will stand for any multivectors.
154 P. Carré and M. Berthier
We will now introduce how geometric algebras can be associated with colour image
processing, but first of all, we give a survey on what have been done already linking
image processing and geometric algebra.
We propose to use the geometric transformations to express hue, saturation and
value colour information from any colour pixel m incoded as a 1-vector of R3,0 in
RGB colour space. This resolution is performed using only algebraic expressions in
R3,0 and is a generalization of what was already done in the quaternionic formalism.
A colour image is seen as a function f from R2 with values in the vector part of the
Clifford algebra R3,0 :
Let ϑ be the 1-vector carrying the grey level axis, r carries the pur red vector and
m represents any colour vector.
• Value is the modulus of the projection of m with respect to ϑ , it is then
expressed by:
V = |(m.ϑ )ϑ −1 |. (6.10)
• Saturation is the distance from the vector m to the grey level axis ϑ , it is then the
modulus of the rejection of m with respect to ϑ :
S = |(m ∧ ϑ )ϑ −1 |. (6.11)
• To reach the hue, we need to define a colour which hue is zero, let ϑ2 be the
1-vector that represents H = 0. A general agreement is to say that pur red has
a null hue, ϑ2 is then r’s rejection with respect to ϑ . Therefore, H is the angle
between ϑ2 and mϑ⊥ = (m ∧ ϑ )ϑ −1 which is the m’s rejection with respect to ϑ .
The hue can then be given by:
−1 m⊥
H = cos ·ϑ .
|m⊥ |
6 Color Representation and Processes with Clifford Algebra 155
Fig. 6.4 Hue’s modification: (a) original image (b) modified image
We thus have formulated the hue, saturation and value of a RGB colour vector
using only algebraic expressions. From these concepts, we can define colour
transform by using only algebraic expressions.
Performing the translation operator along the grey axis ϑ with coefficient α ∈
R on every pixel of an image will result on an alteration of the general value or
brightness of the original image. The result of such an alteration of the brightness
seems more contrasted and warmer than the original image.
∗
m = m + αϑ = I ϑ + Seϑ T ϑ2 + αϑ → I = I + α .
Here the rotation operator is applied on each pixels f [m, n] of the image around
the grey axis ϑ , this is shown in Fig. 6.4. The result is an alteration of the hue of the
original image.
θ θ θ θ θ θ
m = e−ϑ 2 meϑ 2 = Ie−ϑ 2 ϑ eϑ 2 + Se−ϑ 2 eϑ T ϑ2 eϑ 2
m = I ϑ + Seϑ (T+θ ) ϑ2 −→ T = T + θ .
Figure 6.4 shows this kind of hue modification. The original image (Fig. 6.4a)
has been modified by the rotation around the greyscale axis with an angle of π /3.
So the red roof is now green, the green threshold is now blue, etc. Note that one can
imagine to choose any other colour vector for the rotation axis than the grey one to
perform an other operation than hue alteration.
The translation operator, associated with the weight β ∈ R on its saturation axis
f⊥ [x, y], is applied to perform an alteration of an image’s saturation. The saturation
axis f⊥ [x, y] is the rejection of f [x, y] with respect to ϑ .
∗ ∗
m = I ϑ + Seϑ T ϑ2 = m + β eϑ T ϑ2
∗
m = I ϑ + (S + β )eϑ T ϑ2 −→ S = S + β .
156 P. Carré and M. Berthier
Fig. 6.5 Saturation’s modification: (a) original image (b) modified image
Figure 6.5 illustrates this saturation’s alteration where the original image
(Fig. 6.4a) has been altered as described in the precedent equation to get the result
(Fig. 6.4b). The operation is equivalent to:
• Get pale or washed colours in opposition to the original image when the
saturation level is lowered as in the Fig. 6.5.
• Whereas when the saturation level is uppered, colours seem to be more vivid than
in the original image.
Note that all of these colour transformations can be done because colours are
encoded on the vector part of a R3,0 multivector. In fact, reflections, translations,
rotations and rejections are defined only for simple multivectors that is to say
information included in such multivectors is described on one and only one grade.
In this section, we introduced geometric algebras and studied several of their
properties which will help us through the next section where we will describe how
to use them to process digital images. We also showed that embedding colours into
R3,0 algebra allows to perform image alteration using algebraic formalisation only.
From these concepts, we now propose to study colour edge detection with Clifford
algebra.
The formalisation of colour information into R3,0 allows to define more complex
colour processings. Here, we describe how to use the geometric concepts studied
before to define algebraically spatial approaches such as colour edge detection.
As we have seen, Sangwine [8] defined a method to detect colour edges by
using colour pixels incoded in pur quaternions. The idea is to define a convolution
operation for quaternions and apply specific filters to perform the colour edge
detection. This method can be written with the geometric algebra formalism where
colours are represented by 1-vectors of R3,0 :
and where h1 and h2 are a couple of filters which are used to perform a reflection
of every colour√with respect to the 1-vector associated with the greyscale axis ϑ =
(e1 + e2 + e3)/ 3 defined as:
⎡ ⎤ ⎡ ⎤
1 1 1 1 1 1
1⎣ 1⎣
h1 = 0 0 0⎦ and h2 = 0 0 0 ⎦. (6.13)
6 6 −1 −1 −1
ϑ ϑϑ ϑ ϑ ϑ
Figure 6.6 illustrates the saturation filtering scheme. We observe the detection
of all chromatic edges. But, the main drawback of this method is still that it is
based on a saturation measurement only. In fact, when edges contain achromaticity
information only, this approach is not able to detect them properly.
The geometric product allow us to fulfill the drawback of the previous method
by describing geometrically every colour pixel f [x, y] with respect to the greyscale
axis. This description is given by this geometric product f [x, y]ϑ where f [x, y]ϑ is
broken into two terms for every pixels:
Fig. 6.6 Saturation gradient approach: (a) original image, (b) saturation gradient (c) edges
selected by maxima extraction
In order to compute the geometric product and the Sangwine filtering, we use
this couple of filters
⎛ ⎞ ⎛ ⎞
1 1 1 1 1 1
v=⎝ 0 1 0 ⎠ and u=⎝ 0 ϑ 0 ⎠.
ϑϑ ϑ ϑ ϑ ϑ −1
−1 −1
g[m, n] = {[ϑ f [m+1, n+1]ϑ −1 + f [m+1, n−1]]+[ϑ f [m, n + 1]ϑ −1 + f [m, n − 1]]
+[ϑ f [m − 1, n + 1]ϑ −1 + f [m − 1, n − 1]]} + f [m, n]ϑ . (6.16)
Fig. 6.7 (a) Original image with chromatic and achromatic information; (b) bivector part | f [x, y] ∧
ϑ |; (c) Prewitt filtering applied on the scalar part and combined with the achromatic mask; (d) final
result
with
⎡ ⎤
−1 −1 −1
G1 = ⎣ 0 0 0 ⎦
1 1 1
and G2,3,4 rotated version of G1 .
Then, to store just achromatic information, we use the mask defined with the
modulus of the bivectial part. The result gives the gradient of pixels which do not
contain chromaticity information (Fig. 6.7c). The last step is to combine this value
gradient to the saturation one defined before. We can use different techniques to
merge the two gradients as the maximum operator to preserve only the maximum
value between those two gradients (Fig. 6.7d).
160 P. Carré and M. Berthier
Fig. 6.8 Gradient examples on colour images: (a) (b) (c) original images; (d) (e) (f) final gradient
Figure 6.8 shows results on classical digital colour processing images. One can
note that the computer graphics image (Fig. 6.8a) points up that achromatic regions
are here well detected (Fig. 6.8d). The following house image (Fig. 6.8b) also
includes achromatic areas such as the gutters and the windows frame which appear
in the calculated gradient (Fig. 6.8e).
We now describe the use of these new concepts for the definition of a colour
Fourier transform.
The Fourier transform is well known to be an efficient tool for analyzing signals
and especially grey level images. When dealing with nD images, as colour images,
it is not so clear how to define a Fourier transform which is more than n Fourier
transforms computed marginally. The first attempt to define such a transform is due
to S. Sangwine and T. Ells [10] who proposed to encode the colour space RGB by
the space of imaginary quaternions H0 . For a function f from R2 to H0 representing
a colour image, the Fourier Transform is given by
Fμ f (U) = f (X) exp(− μ X,U)dX, (6.18)
R2
6 Color Representation and Processes with Clifford Algebra 161
To our knowledge, the only generalizations of the usual Fourier transform using
quaternions and concerning image processing are those proposed by Sangwine et al.
and by Bülow. The first one is clearly motivated by colour analysis and the second
one aims at detecting two-dimensional symmetries in grey-level images.
Several constructions have been proposed in the context of Clifford algebras.
In [13], a definition is given using the algebras R2,0 and R3,0 in order to introduce
the concept of 2D analytic signal. A definition appears also in [14] which is
mainly applied to analyse frequencies of vector fields. With the same Fourier kernel
Mawardi and Hitzer in [15] establish in [15] an uncertainty principle for multivector
functions. The reader may find in [16] a construction using the Dirac operator
and applications to Gabor filters. Let us also mention, from a different viewpoint,
reference [17] where generalized Fourier descriptors are defined by considering the
action of the motion group of R2 that is the semidirect product of the groups R2 and
SO(2).
The final part of this section describes the generalization of the Fourier Transform
by using Bivectors of the Clifford Algebra R3,0 . We start by illustrating how
Quaternion are used to define colour Fourier Transform.
162 P. Carré and M. Berthier
Note that i and j can be replaced by arbitrary pur imaginary quaternions. The choice
of this formula is justified by the following equality:
Fij f (U) = Fcc f (U) − iFsc f (U) − jFcs f (U) + kFss f (U),
where
Fcc f (U) = f (X) cos(2π u1 x1 ) cos(2π u2x2 )dX
R2
and similar expressions involving sinus and cosinus for Fsc f , Fcs f and Fss f .
We refer the reader to [18] for details and applications to analytic signals.
In order to understand what the Fourier coefficients stand for, we studied the digital
caracterization of the Discrete Quaternionic Fourier Transform (DQFT). The colour
6 Color Representation and Processes with Clifford Algebra 163
Fourier spectrum presented some symmetries due to zero scalar spatial part of any
colour image exactly as it was well known that the spectrum of a real signal by a
complex Fourier transform (CFT) had hermitian properties of symmetry. Even if
the spatial information of a colour image is using pure quaternions only, applying
a DQFT on an image results in full quaternions (i.e., with scalar part non zero).
We wanted to find, after Inverse DQFT, a space where scalar part is zero in order
to avoid any loss of information as the spatial colour image is coded on a pure
quaternion matrix which real part automatically set to zero.
Let
F[o, p] = Fr [o, p] + Fi[o, p]i + Fj [o, p] j + Fk [o, p]k (6.21)
1
∑ ∑ e2μπ ( M + N ) F[o, p]
om pn
f [m, n] = √ (6.22)
MN o p
We can see that the real part must be odd and all the imaginary parts must be
even. This is a direct extension of the antihermitian property of the complex Fourier
transform of imaginary signal.
When studying the complex spectrum domain, several notions are helpful such
as the modulus and the angle.
A Fourier coefficient, F[o, p] = q0 + iq1 + jq2 + kq3 can be written as:
F[o, p] = q = |q|eνϕ
164 P. Carré and M. Berthier
with |q| the QFT modulus, ϕ ∈ R the QFT phase and ν ∈ H0 ∩ H1 the QFT axis.
Figure 6.9 illustrates this polar representation of the Fourier coefficient for the image
Lena. We can see that the Fourier coefficient has a modulus similar to that Greyscale
image. It is more difficult to give interpretation of the information contained in the
angle or the axis.
In order to try to give an interpretation of the information contained in the
quaternionic spectrum of colour images, we can study spatial atoms associated with
a pulse (Dirac) in the frequency domain.
Initialization could be done in two different ways:
• F[o, p] = Kr .δo0 ,p0 [o, p] − Kr δ−o0 ,−p0 [o, p]. Initialization done on the real part
of the spectrum, leading to odd oscillations on the spatial domain linked to the
direction μ parameter of the Fourier transform. Complex colours are obtained
in the RGB colour space after modifying this μ parameter and normalising it
because it always needs to stay a pure unit quaternion.
Fr [o0 , p0 ] = Kr and Fr [−o0, −p0 ] = −Kr is associated with:
o m p n
0 0
f [m, n] = 2 μ (Kr ) sin 2π + . (6.26)
M N
Initializing a pair of constants on the real component leads to a spatial oscillation
following the same imaginary component(s) as those included in the direction μ
(Fig. 6.10b).
• F[o, p] = e.(Ke .δo0 ,p0 [o, p] + Ke δ−o0 ,−p0 [o, p]) with e = i, j or k. Initialization
done on the imaginary part of the spectrum, leading to even oscillations on the
spatial domain independently from the μ parameter of the Fourier transform.
Complex colours in the RGB colour space are reached by initialization on several
imaginary components weighted as in the additive colour synthesis theory.
With e = i, j, k, Fe [o0 , p0 ] = Fe [−o0 , −p0 ] = Ke is associated with:
o m p n
0 0
f [m, n] = e 2(Ke ) cos 2π + . (6.27)
M N
Initializing a pair of constants on any imaginary component with any direction μ
leads to a spatial oscillation on the same component (Fig. 6.10a).
6 Color Representation and Processes with Clifford Algebra 165
The coordinates (o0 , p0 ) and (−o0 , −p0 ) of the two initialization points in the
Fourier domain affect the orientation and the frequency of the oscillations in the
spatial domain as it does so with greyscale image in complex Fourier domain.
Orientation of the oscillations can be changed as shown in Fig. 6.10c.
Below we outline the different ways of defining a Clifford Fourier Transform.
f = f 0 + f 1 e1 + f 2 e2 + f 3 e3 +
+ f23 i3 e1 + f31 i3 e2 + f12 i3 e3 + f123 i3
166 P. Carré and M. Berthier
For the special case of a function f from R2 to C2 = R0,2 ⊗ C, the kernel can be
made explicit and
1
F± f (U) = exp(±U ∧ X) f (X)dX. (6.32)
2π R2
Let us remark that exp(±U ∧ X) is the exponential of a bivector, i.e., a spinor. This
construction allows to introduce two-dimensional Clifford Gabor filters (see [16] for
details).
As the reader may notice, there are many ways to generalize the usual definition
of the complex Fourier transform. In all the situations mentioned above the
multiplication is non commutative and as a consequence the position of the kernel
in the integral is arbitrary. We may in fact distinguish two kinds of approaches: the
first ones deal with so called bivectors (see below) and the second ones involve the
pseudoscalar e1 e2 e3 of the Clifford algebra R3,0 . The rest of this paper focus on
the first approaches. The purpose of the last part of this chapter is to propose a well
founded mathematical definition that explains why it is necessary to introduce those
bivectors and their role in the definition. Before going into details, we recall some
mathematical notions.
We start by some considerations about the theory of abstract Fourier transform and
then introduce basic notions on Clifford algebras and spinor groups. The main result
of this section is the description of the Spin(3) and Spin(4) characters. From the
mathematical viewpoint, defining a Fourier Transform requires to deal with group
actions. For example, in the classical one-dimensional formula
+∞
F f (u) = f (x) exp(−iux)dx (6.33)
−∞
6 Color Representation and Processes with Clifford Algebra 167
the involved group is the additive group (R, +). This is closely related to the
well-known Shift Theorem
where fα (x) denotes the function x
−→ f (x + α ), which reflects the fact that a
translation of a vector α produces a multiplication by exp(iuα ). The correspondance
α
−→ exp(iuα ) is a so-called character of the additive group (R, +).
More precisely, a character of an abelian group G is a map ϕ : G −→ S1 that
preserves the composition laws of both groups. Here S1 is the multiplicative group
of unit complex numbers. It is a special case, for abelian groups, of the notion
of irreducible unitary representations, [19]. The abstract definition of a Fourier
Transform for an (abelian) additive group G and a function f from G to C is given by
F f (ϕ ) = f (x)ϕ (−x)dν (x), (6.35)
G
parametrized by U = (u1 , . . . , un ). They form the group (Rn , +). Applying the above
formula to this situation leads to the usual Fourier Transform
F f (U) = f (X) exp(−iU, X)dX. (6.36)
Rn
It is classical, see [19], that considering the group of rotations SO(2, R), resp. the
group Zn , and the corresponding characters yields to the Fourier series theory, resp.
the discrete Fourier Transform.
One of the ingredients of the construction of the Colour Fourier transform is the
notion of Spin characters which extends the notion of characters to maps from R2
to spinor groups representing rotations.
[Link] Rotation
In the same way that characters of the Fourier transform for grey-level images are
maps form R2 to the rotation group S1 of the complex plane C, we want to define
characters for colour images as maps from R2 to the rotation group acting on the
space of colours, chosen in the sequel to be RGB. The Clifford algebra framework
is particulary well adapted to treat this problem since it allows to encode geometric
transformations via algebraic formulas.
Rotations of R3 correspond to specific elements of R3,0 , namely those given by
τ ⊥v := τ −1 vτ . (6.37)
and the unit 1 (the vectors e1 , e2 , e3 and e4 are elements of an orthonormal basis
of R4 ). As before, the multiplication rules are given by e2i = 1 and ei e j = −e j ei .
The corresponding spinor group Spin(4) is the cross product of two copies of
Spin(3) and acts as rotations on vectors of R4 by formula (6.37).
One fundamental remark is that every spinor τ of Spin(3), resp. Spin(4), can be
written as the exponential of a bivector of R3,0 , resp. R4,0 , i.e.,
1 i
τ=∑ B (6.38)
i≥0 i!
for some bivector B. This means precisely that the Lie exponential map is onto (see
[20] for a general theorem on compact connected Lie groups). As an example, the
spinor
(1 + n2n1 ) n2 ∧ n1
τ=! = exp (θ /2)
2(1 + n1 · n2 ) |n2 ∧ n1 |
is the rotation of R3 that sends by formula (6.37) the unit vector n1 to the unit vector
n2 leaving the plane (n1 , n2 ) globally invariant. In the above expression, θ is the
angle between n1 and n2 and |n2 ∧ n1 | is the magnitude of the bivector n2 ∧ n1 .
The aim here is to compute the group morphisms (i.e., maps preserving the
composition laws) from the additive group R2 to the spinor group of the involved
Clifford algebra. We don’t detail the proofs since they need specific tools on Lie
algebras (see [21] for explanations). In the sequel, we denote S23,0 , resp. S24,0 , the set
of unit bivectors of the algebra R3,0 , resp. R4,0 . Let us first treat the case of Spin(3)
characters.
6 Color Representation and Processes with Clifford Algebra 169
Theorem 1 (Spin (3) Characters). The group morphisms of the additive group R2
to Spin(3) are given by the maps that send (x1 , x2 ) to:
1
exp (x1 u1 + x2 u2 )B , (6.39)
2
The reader may find in [21] the complete description of the rotations in the space
R4 . The classification is given as follows.
• Simple rotations are exponential of simple bivectors that is exterior products of
two vectors. These rotations turn only one plane.
170 P. Carré and M. Berthier
Before examining Clifford Colour Fourier, it is useful to rewrite the usual definition
of the complex Fourier transform in the language of Clifford algebras.
with values in the group of unit complex numbers which is in fact the group Spin(2)
of the Clifford algebra R2,0 . Considering the complex valued function f = f1 + i f2
as a map in the vector part of this algebra, i.e.,
We give first a general definition for a function f from R2 with values in the vector
part of the Clifford algebra R4,0 :
f : (x1 , x2 ) −→ f1 (x1 , x2 )e1 + f2 (x1 , x2 )e2 + f3 (x1 , x2 )e3 + f4 (x1 , x2 )e4 . (6.45)
It is defined on R4 × S24,0 .
Let us give an example. The vector space H of quaternions can be identified with
the vector space R4 under the correspondance: e1 ↔ i, e2 ↔ j, e3 ↔ k, and e4 ↔ 1.
It then can be shown that
1
Di j = − (e1 + e2 )(e3 − e4 ).
4
For most of the applications to colour image processing that will be investigated
below, it is sufficient to consider a transform that can be applied to functions with
values in the vector part of the algebra R3,0 . Such a function is given by
It is defined on R2 .
As an example, let us mention that (under the above identification of H with R4 )
Fμ f (u1 , u2 ) = CF Dμ f (u1 , u2 ),
D μ = ( μ1 e1 + μ2 e2 + μ3 e3 ) ∧ e4
f = f D + f ⊥D , (6.49)
where fD , resp. f⊥D , is the parallel part, resp. the orthogonal part, of f with respect
to the bivector D. Simple computations show that
CF f (u1 , u2 , u3 , u4 , D)
= fD (x1 , x2 ) exp [−(x1 (u1 + u3 ) + x2(u2 + u4))D] dx1 dx2
R2
+ f⊥D (x1 , x2 ) exp [−(x1 (u1 + u3 ) + x2(u2 + u4 ))I4 D] dx1 dx2 . (6.50)
R2
or equivalently
f (x1 , x2 ) = v1 [( f (x1 , x2 ) · v1 ) + ( f (x1 , x2 ) · (v1 D))D]
+ v3 [( f (x1 , x2 ) · v3 ) + ( f (x1 , x2 ) · (v3 I4 D))I4 D]
= v1 [α (x1 , x2 ) + β (x1, x2 )D] + v3 [γ (x1 , x2 ) + δ (x1 , x2 )I4 D] (6.53)
Since D2 = (I4 D)2 = −1, the terms in the brackets can be identified with complex
numbers α (x1 , x2 ) + iβ (x1 , x2 ) and γ (x1 , x2 ) + iδ (x1 , x2 ) on which a usual complex
FFT can be applied. Let us denote α ; (u1 , u2 ) + iβ;(u1 , u2 ) and γ;(u1 , u2 ) + iδ;(u1 , u2 )
the results. The Clifford Fourier transform of f in the direction D is given by
CF D f (u1 , u2 ) = v1 α; (u1 , u2 ) + β;(u1 , u2 )D + v3 γ;(u1 , u2 ) + δ;(u1 , u2 )I4 D .
(6.54)
For the applications treated below, it will be clear how to choose the unit vectors
v1 and v3 .
The Clifford Fourier transform defined by equation (6.46) is left invertible. Its
inverse is given by
CF −1 g(x1 , x2 )
= g(u1 , u2 , u3 , u4 , D)⊥ϕ(u1 ,u2 ,u3 ,u4 ,D) (x1 , x2 )du1 du2 du3du4 dν (D), (6.55)
R4 ×S24,0
174 P. Carré and M. Berthier
where ν is a measure on the set S24,0 . The inversion formula for the Clifford Fourier
transform (6.48) (colour image definition) is much more simpler.
Proposition 2 (Inverse Clifford Fourier Transform for Colour Images). The
Clifford Fourier transform defined by equation (6.48) is invertible. Its inverse is
given by
CF −1
D g(x1 , x2 ) = g(u1 , u2 )⊥ϕ(u1 ,u2 ,0,0,D) (x1 , x2 )du1 du2 . (6.56)
R2
Remark the analogy (change of signs in the spin characters) with the usual
inversion formula.
Since this chapter is mainly devoted to colour image processing and for sake of
simplicity, we describe now properties concerning the only transformation (6.48).
Shift Theorem
where f is as in (6.47).
Proposition 3 (Shift Theorem for Colour Images). The Clifford Fourier Trans-
form of the function f(α1 ,α2 ) in the direction D is given by
It is well known that a function f defined on R2 is real if and only if the usual
Fourier coefficients satisfy
where F is the usual Fourier transform and the overline denotes the complex
conjugacy. This property, called hermitian symmetry is important when dealing with
frequencies filtering. Note that the precedent equation implies that
With the quaternionic Fourier transform, we noted that the colour Fourier
coefficients satisfied an anti-hermitian symmetry.
The next proposition generalizes this hermitian property to the Clifford Fourier
transform for color images.
Proposition 4 (Generalized Hermitian Symmetry for Colour Images). Let f be
given as in (6.47), then the e4 term in
CF D f (u1 , u2 )⊥ϕ(u1 ,u2 ,0,0,D) (x1 , x2 ) + CF D f (−u1 , −u2 )⊥ϕ(−u1 ,−u2 ,0,0,D) (x1 , x2 )
(6.61)
Energy Conservation
The following statement is an analog of the usual Parceval equality satisfied by the
usual Fourier transform.
Proposition 5 (Clifford Parceval Equality). Let f be given as in (6.47), then
(CF D f (u1 , u2 ))2 du1 du2 = ( f (x1 , x2 ))2 dx1 dx2 (6.62)
R2 R2
whenever one term is defined (and thus both terms are defined).
Let us recall that for a vector u of the algebra R4,0 , u2 = Q(u) where Q is the
Euclidean quadratic form on R4 .
Colour Bivector
Hue Bivector
In the middle image, the green and blue high frequencies are removed while
the red ones are preserved (I4 D = e2 ∧ e3 ). The low-pass filter removes all high
frequencies of the right image excepted those of the red petals.
6.4 Conclusion
Hypercomplex or quaternions numbers have been used recently for both greyscale
and colour image processing. Geometric algebra allows to handle geometric entities
such as scalars, vectors, or bivectors independently. These entities are handled with
the help algebraic expressions such as products (inner, outer, geometric, . . . ) for
instance and rules over these products allow to affect or modify entities.
This chapter presents how quaternion and geometric algebra is used as a new
formalism to perform colour image processing.
The first section reminds us how to use quaternions to process colour infor-
mation, and how the three components of a colour pixel split to the vectorial
part of R3,0 multivector. This condition is required to apply and define geometric
operations algebraically on colour vectors such as translations and rotations for
instance. After that, we illustrate that the R3,0 algebra is convenient to analyse and/or
alter geometrically colour in images with operations tools defined algebraically. For
that, we gave examples with alteration of the global hue, saturation, or value of
colour images. After this description of some basic issues of colour manipulations,
we shows different existent filtering approaches for using quaternions with colour
images, and we proposed to generalize approaches already defined with quaternions
and enhanced them with this new formalism. Illustrations proved it gave more
accurate colour edge detection.
The second section introduces the discrete quaternionic Fourier transform pro-
posed by Sangwine and by Bülow, and the conditions on the quaternionic spectrum
to enable manipulations into this frequency domain without loosing information
when going back to the spatial domain. This parts gives some interpretation of
the quaternionic Fourier space. We conclude on a geometric approach using group
6 Color Representation and Processes with Clifford Algebra 179
actions for the Clifford colour fourier transform. The idea is to generalize the
usual definition based on the characters of abelian groups by considering group
morphisms from R2 to spinor groups Spin(3) and Spin(4). The transformation is
parameterized by a bivector and a quadratic form, the choice of which is related to
the application to be treated.
References