Basic Engineering Data Collection and Analysis
Basic Engineering Data Collection and Analysis
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Basic Engineering
Data Collection
and Analysis
Stephen B. Vardeman
Iowa State University
J. Marcus Jobe
Miami University
Ames, Iowa
Basic Engineering Data Collection and © 2001 Stephen B. Vardeman and J. Marcus Jobe
Analysis
Basic Engineering Data Collection and Analysis is available under a Creative
Stephen B. Vardeman and J. Marcus Jobe Commons Attribution NonCommercial ShareAlike 4.0 International license.
Sponsoring Editor: Carolyn Crockett
You may share and adapt the material, so long as you provide appropriate credit
Marketing: Chris Kelly
to the original authors, do not use the material for commercial purposes, and
Editorial Assistant: Ann Day any adaptations or remixes of the material which you create are distributed
Production Editor: Janet Hill under the same license as the original.
Production Service: Martha Emry
Permissions: Sue Ewing
Cover Design/Illustration: Denise Davidson Originally published by Brooks/Cole Cengage Learning in 2001.
Interior Design: John Edeen Published online by Iowa State University Digital Press in 2023.
Interior Illustration: Bob Cordes
Print Buyer: Kristina Waller
Typesetting: Eigentype Compositors Library of Congress Catalog Number: 00-040358
ISBN-13: 978-0-534-36957-6 (print)
ISBN-10: 0-534-36957-X (print)
ISBN: 978-1-958291-03-0 (PDF)
https://doi.org/10.31274/isudp.2023.127
Iowa State University is located on the ancestral lands and territory of the Baxoje
(bah-kho-dzhe), or Ioway Nation. The United States obtained the land from the
Meskwaki and Sauk nations in the Treaty of 1842. We wish to recognize our
obligations to this land and to the people who took care of it, as well as to the
17,000 Native people who live in Iowa today.
Soli Deo Gloria
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Preface
v
vi Preface
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Pedagogical Features
Pedagogical and practical features include:
■ Precise exposition
■ A logical two-color layout, with examples delineated by a color rule
■ Margin notes naming formulas and calling attention to some main issues of
discussion
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
The Exercises
There are far more exercises in this text than could ever be assigned over several
semesters of teaching from this book. Exercises involving direct application of
section material appear at the end of each section, and answers for most of them
appear at the end of the book. These give the reader immediate reinforcement that
the mechanics and main points of the exposition have been mastered. The rich sets of
Chapter Exercises provide more. Beyond additional practice with the computations
of the chapter, they add significant insight into how engineering statistics is done
and into the engineering implications of the chapter material. These often probe
what kinds of analyses might elucidate the main features of a scenario and facilitate
substantive engineering progress, and ponder what else might be needed. In most
cases, these exercises were written after we had analyzed the data and seriously
considered what they show in the engineering context. These come from a variety
of engineering disciplines, and we expect that instructors will find them to be not
only useful for class assignments but also for lecture examples to many different
engineering audiences.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Ancillaries
Several types of ancillary material are available to support this text.
■ The CD packaged with the book provides PowerPointTM visuals and audio
presenting solutions for selected Section Exercises.
■ For instructors only, a complete solutions manual is available through the
local sales representative.
■ The publisher also maintains a web site supporting instruction using Basic
Engineering Data Collection and Analysis at www.brookscole.com.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Acknowledgments
There are many who deserve thanks for their kind help with this project. People at
Duxbury Thomson Learning have been great. We especially thank Carolyn Crockett
for her encouragement and vision in putting this project together. Janet Hill has
been an excellent Production Editor. We appreciate the help of Seema Atwal with
the book’s ancillaries, and are truly pleased with the design work overseen by
Vernon Boes.
Acknowledgments ix
First class help has also come from outside of Duxbury Thomson Learning.
Martha Emry of Martha Emry Production Services has simply been dynamite to
work with. She is thorough, knowledgeable, possessed of excellent judgment and
unbelievably patient. Thanks Martha! And although he didn’t work directly on this
project, we gratefully acknowledge the meticulous work of Chuck Lerch, who wrote
the solutions manual and provided the answer section for Statistics for Engineering
Problem Solving. We have borrowed liberally from his essentially flawless efforts
for answers and solutions carried over to this project. We are also grateful to Jimmy
Wright and Victor Chan for their careful work as error checkers. We thank Tom
Andrika for his important contributions to the development of the PowerPoint/audio
CD supplement. We thank Tiffany Lynn Hagemeyer for her help in preparing the
MINITAB, JMP, and Excel data files for download. Andrew Vardeman developed
the web site, providing JMP, MINITAB, and Excel help for the text, and we aprreci-
ate his contributions to this effort. John Ramberg, University of Arizona; V. A.
Samaranayake, University of Missouri at Rolla; Paul Joyce, University of Idaho;
James W. Hardin, Texas A & M; and Jagdish K. Patel, University of Missouri at
Rolla provided helpful reviews of this book at various stages of completion, and we
thank them.
It is our hope that this book proves to be genuinely useful to both engineering
students and working engineers, and one that instructors find easy to build their
courses around. We’ll be glad to receive comments and suggestions at our e-mail
addresses.
Steve Vardeman J. Marcus Jobe
[email protected] [email protected]
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Contents
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1 Introduction 1
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
2 Data Collection 26
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
B Tables 785
Introduction
T his chapter lays a foundation for all that follows: It contains a road map for the
study of engineering statistics. The subject is defined, its importance is described,
some basic terminology is introduced, and the important issue of measurement is
discussed. Finally, the role of mathematical models in achieving the objectives of
engineering statistics is investigated.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1
2 Chapter 1 Introduction
Table 1.1
Thrust Face Runouts (.0001 in.)
Gears laid
0 10 20 30 40
Runout (.0001 in.)
Gears hung
0 10 20 30 40
Runout (.0001 in.)
Example 1 But how “precise” is this figure? Runout values are variable. So is there any
(continued ) assurance that the difference seen in the present means would reappear in further
testing? Or is it possibly explainable as simply “stray background noise”? Lay-
ing gears is more expensive than hanging them. Can one know whether the extra
expense is justified?
Drawing These questions point to the need for methods of formal statistical inference
inferences from data and translation of those inferences into practical conclusions. Meth-
from data ods presented in this text can, for example, be used to support the following
statements about hanging and laying gears:
1. One can be roughly 90% sure that the difference in long-run mean runouts
produced under conditions like those of the engineer’s study is in the range
3.2 to 7.4
2. One can be roughly 95% sure that 95% of runouts for gears laid under
conditions like those of the engineer’s study would fall in the range
3.0 to 22.2
3. One can be roughly 95% sure that 95% of runouts for gears hung under
conditions like those of the engineer’s study would fall in the range
.8 to 35.0
These are formal quantifications of what was learned from the study of laid
and hung gears. To derive practical benefit from statements like these, the process
engineer had to combine them with other information, such as the consequences
of a given amount of runout and the costs for hanging and laying gears, and had to
apply sound engineering judgment. Ultimately, the runout improvement was great
enough to justify some extra expense, and the laying method was implemented.
The example shows how the elements of statistics were helpful in solving an
engineer’s problem. Throughout this text, the intention is to emphasize that the
topics discussed are not ends in themselves, but rather tools that engineers can use
to help them do their jobs effectively.
Section 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Explain why engineering practice is an inherently 3. Describe the difference between descriptive and
statistical enterprise. (formal) inferential statistics.
2. Explain why the concept of variability has a central
place in the subject of engineering statistics.
1.2 Basic Terminology 5
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Definition 3 An experimental study (or, more simply, an experiment) is one in which the
investigator’s role is active. Process variables are manipulated, and the study
environment is regulated.
Most real statistical studies have both observational and experimental features,
and these two definitions should be thought of as representing idealized opposite
ends of a continuum. On this continuum, the experimental end usually provides
the most efficient and reliable ways to collect engineering data. It is typically
much quicker to manipulate process variables and watch how a system responds
to the changes than to passively observe, hoping to notice something interesting or
revealing.
Inferring In addition, it is far easier and safer to infer causality from an experiment than
causality from an observational study. Real systems are complex. One may observe several
instances of good process performance and note that they were all surrounded by
circumstances X without being safe in assuming that circumstances X cause good
process performance. There may be important variables in the background that are
changing and are the true reason for instances of favorable system behavior. These
so-called lurking variables may govern both process performance and circum-
stances X. Or it may simply be that many variables change haphazardly without
appreciable impact on the system and that by chance, during a limited period of
observation, some of these happen to produce X at the same time that good perfor-
mance occurs. In either case, an engineer’s efforts to create X as a means of making
things work well will be wasted effort.
6 Chapter 1 Introduction
Most engineering studies tend to be of the second type, although some important
engineering applications do involve enumerative work. One such example is the
1.2 Basic Terminology 7
reliability testing of critical components—e.g., for use in a space shuttle. The interest
is in the components actually in hand and how well they can be expected to perform
rather than on any broader problem like “the behavior of all components of this
type.” Acceptance sampling (where incoming lots are checked before taking formal
receipt) is another important kind of enumerative study. But as indicated, most
engineering studies are analytical in nature.
Example 2 The students working on the pelletizing machine were not interested in any partic-
(continued ) ular batch of pellets, but rather in the question of how to make the machine work
effectively. They hoped (or tacitly assumed) that what they learned about making
fuel pellets would remain valid at later times, at least under shop conditions like
those they were facing. Their experimental study was analytical in nature.
Particularly when discussing enumerative studies, the next two definitions are
helpful.
Definition 6 A population is the entire group of objects about which one wishes to gather
information in a statistical study.
Definition 7 A sample is the group of objects on which one actually gathers data. In the
case of an enumerative investigation, the sample is a subset of the population
(and can in some cases include the entire population).
Figure 1.2 shows the relationship between a population and a sample. If a crate of
100 machine parts is delivered to a loading dock and 5 are examined in order to
verify the acceptability of the lot, the 100 parts constitute the population of interest,
and the 5 parts make up a (single) sample of size 5 from the population. (Notice the
word usage here: There is one sample, not five samples.)
Sample
Population
There are several ways in which the meanings of the words population and
sample are often extended. For one, it is common to use them to refer to not only
objects under study but also data values associated with those objects. For example,
if one thinks of Rockwell hardness values associated with 100 crated machine parts,
the 100 hardness values might be called a population (of numbers). Five hardness
values corresponding to the parts examined in acceptance sampling could be termed
a sample from that population.
Example 2 Cyr, Ellson, and Rickard identified eight different sets of experimental conditions
(continued ) under which to run the pelletizing machine. Several production runs of fuel pellets
were made under each set of conditions, and each of these produced its own
percentage of conforming pellets. These eight sets of percentages can be referred
to as eight different samples (of numbers).
Definition 8 Qualitative or categorical data are the values of basically nonnumerical char-
acteristics associated with items in a sample. There can be an order associated
with qualitative data, but aggregation and counting are required to produce
any meaningful numerical values from such data.
Consider again 5 machine parts constituting a sample from 100 crated parts. If each
part can be classified into one of the (ordered) categories (1) conforming, (2) rework,
and (3) scrap, and one knows the classifications of the 5 parts, one has 5 qualitative
data points. If one aggregates across the 5 and finds 3 conforming, 1 reworkable, and
1 scrap, then numerical summaries have been derived from the original categorical
data by counting.
In contrast to categorical data are numerical data.
1.2 Basic Terminology 9
Returning to the crated machine parts, Rockwell hardness values for 5 selected
parts would constitute a set of quantitative measurement data. Counts of visible
blemishes on a machined surface for each of the 5 selected parts would make up a
set of quantitative count data.
It is sometimes convenient to act as if infinitely precise measurement were
possible. From that perspective, measured variables are continuous in the sense
that their sets of possible values are whole (continuous) intervals of numbers. For
example, a convenient idealization might be that the Rockwell hardness of a ma-
chine part can lie anywhere in the interval (0, ∞). But of course this is only an
idealization. All real measurements are to the nearest unit (whatever that unit may
be). This is becoming especially obvious as measurement instruments are increas-
ingly equipped with digital displays. So in reality, when looked at under a strong
enough magnifying glass, all numerical data (both measured and count alike) are
discrete in the sense that they have isolated possible values rather than a continuum
of available outcomes. Although (0, ∞) may be mathematically convenient and
completely adequate for practical purposes, the real set of possible values for the
measured Rockwell hardness of a machine part may be more like {.1, .2, .3, . . .}
than like (0, ∞).
Well-known conventional wisdom is that measurement data are preferable to
categorical and count data. Statistical methods for measurements are simpler and
more informative than methods for qualitative data and counts. Further, there is
typically far more to be learned from appropriate measurements than from qualitative
data taken on the same physical objects. However, this must sometimes be balanced
against the fact that measurement can be more time-consuming (and thus expensive)
than the gathering of qualitative data.
Example 3 Information on 200 pellets was collected. The students could have simply
(continued ) observed and recorded whether or not a given pellet had mass within the specifi-
cations, thereby producing qualitative data. Instead, they took the time necessary
to actually measure pellet mass to the nearest .1 gram—thereby collecting mea-
surement data. A graphical summary of their findings is shown in Figure 1.3.
20
Frequency
10
3 4 5 6 7 8
Mass (g)
Notice that one can recover from the measurements the conformity/noncon-
formity information—about 28.5% (57 out of 200) of the pellets had masses that
did not meet specifications. But there is much more in Figure 1.3 besides this.
The shape of the display can give insights into how the machine is operating and
the likely consequences of simple modifications to the pelletizing process. For
example, note the truncated or chopped-off appearance of the figure. Masses do
not trail off on the high side as they do on the low side. The students reasoned that
this feature of their data had its origin in the fact that after powder is dispensed
into a die, it passes under a paddle that wipes off excess material before a cylinder
compresses the powder in the die. The amount initially dispensed to a given die
may have a fairly symmetric mound-shaped distribution, but the paddle probably
introduces the truncated feature of the display.
Also, from the numerical data displayed in Figure 1.3, one can find a per-
centage of pellet masses in any interval of interest, not just the interval [6.2, 7.0].
And by mentally sliding the figure to the right, it is even possible to project the
likely effects of increasing die size by various amounts.
Definition 10 Univariate data arise when only a single characteristic of each sampled item
is observed.
Definition 11 Multivariate data arise when observations are made on more than one
characteristic of each sampled item. A special case of this involves two
characteristics—bivariate data.
Definition 12 When multivariate data consist of several determinations of basically the same
characteristic (e.g., made with different instruments or at different times),
the data are called repeated measures data. In the special case of bivariate
responses, the term paired data is used.
Definition 13 A (complete) factorial study is one in which several process variables (and
settings of each) are identified as being of interest, and data are collected under
each possible combination of settings of the process variables. The process
variables are usually called factors, and the settings of each variable that are
studied are termed levels of the factor.
For example, suppose there are four factors of interest—call them A, B, C, and D for
convenience. If A has 3 levels, B has 2, C has 2, and D has 4, a study that includes
samples collected under each of the 3 × 2 × 2 × 4 = 48 different possible sets of
conditions would be called a 3 × 2 × 2 × 4 factorial study.
Combining these then produced eight sets of conditions under which data were
collected (see Table 1.2).
Table 1.2
Combinations in a 23 Factorial Study
When many factors and/or levels are involved, the number of samples in a
full factorial study quickly reaches an impractical size. Engineers often find that
1.2 Basic Terminology 13
they want to collect data for only some of the combinations that would make up a
complete factorial study.
Definition 14 A fractional factorial study is one in which data are collected for only some
of the combinations that would make up a complete factorial study.
One cannot hope to learn as much about how a response is related to a given set
of factors from a fractional factorial study as from the corresponding full factorial
study. Some information must be lost when only part of all possible sets of conditions
are studied. However, some fractional factorial studies will be potentially more
informative than others. If only a fixed number of samples can be taken, which
samples to take is an issue that needs careful consideration. Sections 8.3 and 8.4
discuss fractional factorials in detail, including how to choose good ones, taking
into account what part of the potential information from a full factorial study they
can provide.
Example 2 The experiment actually carried out on the pelletizing process was, as indicated
(continued ) in Table 1.2, a full factorial study. Table 1.3 lists four experimental combinations,
forming a well-chosen half of the eight possible combinations. (These are the
combinations numbered 2, 3, 5, and 8 in Table 1.2.)
Table 1.3
Half of the 23 Factorial
Section 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Describe a situation in your field where an observa- 3. What kind of information can be derived from
tional study might be used to answer a question of a single sample of n bivariate data points (x, y)
real importance. Describe another situation where that can’t be derived from two separate sam-
an experiment might be used. ples of, respectively, n data points x and n data
2. Describe two different contexts in your field where, points y?
respectively, qualitative and quantitative data might 4. Describe a situation in your field where paired data
arise. might arise.
14 Chapter 1 Introduction
5. Consider a study of making paper airplanes, where a full factorial and then a fractional factorial data
two different Designs (say, delta versus t wing), two structure that might arise from such a study.
different Papers (say, construction versus typing), 6. Explain why it is safer to infer causality from an
and two different Loading Conditions (with a paper experiment than from an observational study.
clip versus without a paper clip) are of interest in
terms of their effects on flight distance. Describe
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
60°
40°
Angle
past vertical 20° Metal part
horizontal position, released, and allowed to swing through a test part firmly
fixed in a vertical position at the bottom of its arc of motion. The number of
degrees past vertical that the arm traversed after impact with the part provided an
effective measure of brittleness.
follows. The book is to be opened to a page somewhere near the beginning and one
somewhere near the end. The stack between the two pages is to be grasped firmly
between the thumb and index finger and stack thickness read to the nearest .1 mm
using an ordinary ruler. Dividing the stack thickness by the number of sheets in the
stack and recording the result to the nearest .0001 mm will then produce a thickness
measurement.
Figure 1.5 shows a graph of these data and clearly reveals that even repeated
measurements by one person on one book will vary and also that the patterns of
variation for two different individuals can be quite different. (Wendel’s values
are both smaller and more consistent than Gulliver’s.)
Wendel
Gulliver
Example 7 Ignoring the possibility that some property of Gulliver’s book was responsible for
(continued ) his values showing more spread than those of Wendel, it appears that Wendel’s
measuring technique was more precise than Gulliver’s.
The precision of both students’ measurements could probably have been
improved by giving each a binder clip and a micrometer. The binder clip would
provide a relatively constant pressure on the stacks of pages being measured,
thereby eliminating the subjectivity and variation involved in grasping the stack
firmly between thumb and index finger. For obtaining stack thickness, a microm-
eter is clearly a more precise instrument than a ruler.
Maintaining the U.S. reference sets for physical measurement is the business of
the National Institute of Standards and Technology. It is important business. Poorly
calibrated measuring devices may be sufficient for local purposes of comparing
local conditions. But to establish the values of quantities in any absolute sense, or
to expect local values to have meaning at other places and other times, it is essential
to calibrate measurement systems against a constant standard. A millimeter must be
the same today in Iowa as it was last week in Alaska.
The possibility of bias or inaccuracy in measuring systems has at least two im-
portant implications for planning statistical engineering studies. First, the fact that
18 Chapter 1 Introduction
Accuracy and measurement systems can lose accuracy over time demands that their performance
statistical be monitored over time and that they be recalibrated as needed. The well-known
studies phenomenon of instrument drift can ruin an otherwise flawless statistical study.
Second, whenever possible, a single system should be used to do all measuring. If
several measurement devices or technicians are used, it is hard to know whether the
differences observed originate with the variables under study or from differences in
devices or technician biases. If the use of several measurement systems is unavoid-
able, they must be calibrated against a standard (or at least against each other). The
following example illustrates the role that human differences can play.
Section 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Why might it be argued that in terms of producing Explain which of the three aspects of measure-
useful measurements, one must deal first with the ment quality—validity, precision, and accuracy—
issue of validity, then the issue of precision, and this averaging of many measurements can be ex-
only then the issue of accuracy? pected to improve and which it cannot.
2. Often, in order to evaluate a physical quantity 3. Explain the importance of the stability of the mea-
(for example, the mean yield of a batch chemi- surement system to the real-world success of a sta-
cal process run according to some standard plant tistical engineering study.
operating procedures), a large number of measure-
ments of the quantity are made and then averaged.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Mathematical models are themselves not reality, but they can be extremely effective
descriptions of reality. This effectiveness hinges on two somewhat opposing prop-
erties of a mathematical model: (1) its degree of simplicity and (2) its predictive
20 Chapter 1 Introduction
ability. The most powerful mathematical models are those that simultaneously are
simple and generate good predictions. A model’s simplicity allows one to maneuver
within its framework, deriving mathematical consequences of basic assumptions that
translate into predictions of process behavior. When these are empirically correct,
one has an effective engineering tool.
The elementary “laws” of mechanics are an outstanding example of effective
mathematical modeling. For example, the simple mathematical statement that the
acceleration due to gravity is constant,
a=g
yields, after one easy mathematical maneuver (an integration), the prediction that
beginning with 0 velocity, after a time t in free fall an object will have velocity
v = gt
And a second integration gives the prediction that beginning with 0 velocity, a time t
in free fall produces displacement
1 2
d= gt
2
The beauty of this is that for most practical purposes, these easy predictions are quite
adequate. They agree well with what is observed empirically and can be counted
on as an engineer designs, builds, operates, and/or improves physical processes or
products.
Mathematics But then, how does the notion of mathematical modeling interact with the
and statistics subject of engineering statistics? There are several ways. For one, data collection
and analysis are essential in fitting or estimating parameters of mathematical
models. To understand this point, consider again the example of a body in free fall.
If one postulates that the acceleration due to gravity is constant, there remains the
question of what numerical value that constant should have. The parameter g must
be evaluated before the model can be used for practical purposes. One does this by
gathering data and using them to estimate the parameter.
A standard first college physics lab has traditionally been to empirically evalu-
ate g. The method often used is to release a steel bob down a vertical wire running
through a hole in its center and allowing 60-cycle current to arc from the bob through
a paper tape to another vertical wire, burning the tape slightly with every arc. A
schematic diagram of the apparatus used is shown in Figure 1.7. The vertical posi-
1
tions of the burn marks are bob positions at intervals of 60 of a second. Table 1.4
gives measurements of such positions. (We are grateful to Dr. Frank Peterson of
the ISU Physics and Astronomy Department for supplying the tape.) Plotting the
bob positions in the table at equally spaced intervals produces the approximately
quadratic plot shown in Figure 1.8. Picking a parabola to fit the plotted points in-
volves identifying an appropriate value for g. A method of curve fitting (discussed
in Chapter 4) called least squares produces a value for g of 9.79m/sec2 , not far from
the commonly quoted value of 9.8m/sec2 .
1.4 Mathematical Models, Reality, and Data Analysis 21
Paper tape
Arc
Sliding
metal
bob
Bare
Bare
wire
wire
AC Generator
Table 1.4
Measured Displacements of a Bob in Free Fall
1 .8 13 223.8
2 4.8 14 260.0
3 10.8 15 299.2
4 20.1 16 340.5
5 31.9 17 385.0
6 45.9 18 432.2
7 63.3 19 481.8
8 83.1 20 534.2
9 105.8 21 589.8
10 131.3 22 647.7
11 159.5 23 708.8
12 190.5
Notice that (at least before Newton) the data in Table 1.4 might also have been
used in another way. The parabolic shape of the plot in Figure 1.8 could have
suggested the form of an appropriate model for the motion of a body in free fall.
That is, a careful observer viewing the plot of position versus time should conclude
that there is an approximately quadratic relationship between position and time (and
22 Chapter 1 Introduction
700
600
500
Displacement (mm)
400
300
200
100
1
Time ( 60 second)
from that proceed via two differentiations to the conclusion that the acceleration
due to gravity is roughly constant). This text is full of examples of how helpful it
can be to use data both to identify potential forms for empirical models and to then
estimate parameters of such models (preparing them for use in prediction).
This discussion has concentrated on the fact that statistics provides raw material
for developing realistic mathematical models of real systems. But there is another
important way in which statistics and mathematics interact. The mathematical theory
of probability provides a framework for quantifying the uncertainty associated with
inferences drawn from data.
If, for example, five students arrive at the five different laboratory values of g,
questions naturally arise as to how to use them to state both a best value for g
and some measure of precision for the value. The theory of probability provides
guidance in addressing these issues. Material in Chapter 6 shows that probability
Chapter 1 Exercises 23
considerations support using the class average of 9.796 to estimate g and attaching
to it a precision on the order of plus or minus .02m/sec2 .
We do not assume that the reader has studied the mathematics of probability,
so this text will supply a minimal introduction to the subject. But do not lose sight
of the fact that probability is not statistics—nor vice versa. Rather, probability is a
branch of mathematics and a useful subject in its own right. It is met in a statistics
course as a tool because the variation that one sees in real data is closely related
conceptually to the notion of chance modeled by the theory of probability.
Section 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Explain in your own words the importance of mathematical models to engineering practice.
Chapter 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Calibration of measurement equipment is most 5. Describe a situation from your field where a full
clearly associated with which of the following factorial study might be conducted (name at least
concepts: validity, precision, or accuracy? Explain. three factors, and the levels of each, that would
2. If factor A has levels 1, 2, and 3, factor B has appear in the study).
levels 1 and 2, and factor C has levels 1 and 2, list 6. Example 7 concerns the measurement of the thick-
the combinations of A, B, and C that make up a ness of book paper. Variation in measurements is
full factorial arrangement. a fact of life. To observe this reality firsthand,
3. Explain how paired data might arise in a heat measure the thickness of the paper used in this
treating study aimed at determining the best way book ten times. Use the method described imme-
to heat treat parts made from a certain alloy. diately before Example 7. For each determination,
record the measured stack thickness, the number
4. Losen, Cahoy, and Lewis purchased eight spanner
of sheets, and the quotient to four decimal places.
bushings of a particular type from a local machine
If you are using this book in a formal course,
shop and measured a number of characteristics of
be prepared to hand in your results and compare
these bushings, including their outside diameters.
them with the values obtained by others in your
Each of the eight outside diameters was measured
class.
once by two student technicians, with the follow-
ing results. (The units are inches.) Considering 7. Exercise 6 illustrates the reality of variation in
both students’ measurements, what type of data physical measurement. Another exercise that is
are given here? Explain. similar in spirit, but leads to qualitative data, in-
volves the spinning of U.S. pennies. Spin a penny
on a hard surface 20 different times; for each trial,
Bushing 1 2 3 4
record whether the penny comes to rest with heads
Student A .3690 .3690 .3690 .3700
or tails showing. Did all the trials have the same
Student B .3690 .3695 .3695 .3695 outcome? Is the pattern you observed the one you
Bushing 5 6 7 8 expected to see? If not, do you have any possible
Student A .3695 .3700 .3695 .3690 explanations?
Student B .3695 .3700 .3700 .3690
24 Chapter 1 Introduction
8. Consider a situation like that of Example 1 (in- variables include such things as the hardnesses,
volving the heat treating of gears). Suppose that diameters and surface roughnesses of the pistons
the original gears can be purchased from a variety and the hardnesses, and inside diameters and sur-
of vendors, they can be made out of a variety of face roughnesses of the bores into which the pis-
materials, they can be heated according to a va- tons fit. Describe, in general terms, an observa-
riety of regimens (involving different times and tional study to try to determine how to improve
temperatures), they can be cooled in a number of life. Then describe an experimental study and say
different ways, and the furnace atmosphere can why it might be preferable.
be adjusted to a variety of different conditions. A 12. In the context of Exercise 9, it might make sense
number of features of the final gears are of interest, to average the strengths you record. Would you
including their flatness, their concentricity, their expect such an average to be more or less precise
hardness (both before and after heat treating), and than a single measurement as an estimate of the
their surface finish. average strength of this kind of dowel? Explain.
(a) What kind of data arise if, for a single set Argue that such averages can be no more (or less)
of conditions, the Rockwell hardness of sev- accurate than the individual measurements that
eral gears is measured both before and after make them up.
heat treating? (Use the terminology of Sec-
13. A toy catapult launches golf balls. There are a
tion 1.2.) In the same context, suppose that
number of things that can be altered on the con-
engineering specifications on flatness require
figuration of the catapult: The length of the arm
that measured flatness not exceed .40 mm.
can be changed, the angle the arm makes when it
If flatness is measured for several gears and
hits the stop can be changed, the pull-back angle
each gear is simply marked Acceptable or Not
can be changed, the weight of the ball launched
Acceptable, what kind of data are generated?
can be changed, and the place the rubber cord
(b) Describe a three-factor full factorial study that
(used to snap the arm forward) is attached to the
might be carried out in this situation. Name
arm can be changed. An experiment is to be done
the factors that will be used and describe the
to determine how these factors affect the distance
levels of each. Write out a list of all the differ-
a ball is launched.
ent combinations of levels of the factors that
(a) Describe one three-factor full factorial study
will be studied.
that might be carried out. Make out a data
9. Suppose that you wish to determine “the” axial collection form that could be used. For each
strength of a type of wooden dowel. Why might it launch, specify the level to be used of each of
be a good idea to test several such dowels in order the three factors and leave a blank for record-
to arrive at a value for this “physical constant”? ing the observed value of the response vari-
10. Give an example of a 2 × 3 full factorial data able. (Suppose two launches will be made for
structure that might arise in a student study of the each setup.)
breaking strengths of wooden dowels. (Name the (b) If each of the five factors mentioned above is
two factors involved, their levels, and write out all included in a full factorial experiment, a min-
six different combinations.) Then make up a data imum of how many different combinations of
collection form for the study. Plan to record both levels of the five factors will be required? If
the breaking strength and whether the break was there is time to make only 16 launches with
clean or splintered for each dowel, supposing that the device during the available lab period, but
three dowels of each type are to be tested. you want to vary all five factors, what kind of
11. You are a mechanical engineer charged with im- a data collection plan must you use?
proving the life-length characteristics of a hydro-
static transmission. You suspect that important
Chapter 1 Exercises 25
14. As a variation on Exercise 6, you could try using by applying the method in Exercise 6 ten times
only pages in the first four chapters of the book. to stacks of pages from only the first four chap-
If there were to be a noticeable change in the ul- ters. Is there a noticeable difference in precision
timate precision of thickness measurement, what of measurement from what is obtained using the
kind of a change would you expect? Try this out whole book?
2
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Data Collection
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
2.1.1 Measurement
Good measurement is indispensable in any statistical engineering study. An engi-
neer planning a study ought to ensure that data on relevant variables will be col-
lected by well-trained people using measurement equipment of known and adequate
quality.
When choosing variables to observe in a statistical study, the concepts of mea-
surement validity and precision, discussed in Section 1.3, must be remembered. One
practical point in this regard concerns how directly a measure represents a system
property. When a direct measure exists, it is preferable to an indirect measure,
because it will usually give much better precision.
26
2.1 General Principles in the Collection of Engineering Data 27
2.1.2 Sampling
Once it is established how measurement/observation will proceed, the engineer can
consider how much to do, who is to do it, where and under what conditions it is
to be done, etc. Sections 2.2, 2.3, and 2.4 consider the question of choosing what
observations to make, first in enumerative and then in experimental studies. But first,
a few general comments about the issues of “How much?”, “Who?”, and “Where?”.
How much The most common question engineers ask about data collection is “How many
data? observations do I need?” Unfortunately, the proper answer to the question is typically
“it depends.” As you proceed through this book, you should begin to develop some
intuition and some rough guides for choosing sample sizes. For the time being, we
point out that the only factor on which the answer to the sample size question really
depends is the variation in response that one expects (coming both from unit-to-unit
variation and from measurement variation).
This makes sense. If objects to be observed were all alike and perfect measure-
ment were possible, then a single observation would suffice for any purpose. But if
there is increase either in the measurement noise or in the variation in the system
or population under study, the sample size necessary to get a clear picture of reality
becomes larger.
However, one feature of the matter of sample size sometimes catches people a bit
off guard—the fact that in enumerative studies (provided the population size is large),
sample size requirements do not depend on the population size. That is, sample size
requirements are not relative to population size, but, rather, are absolute. If a sample
size of 5 is adequate to characterize compressive strengths of a lot of 1,000 red
clay bricks, then a sample of size 5 would be adequate to characterize compressive
strengths for a lot of 100,000 bricks with similar brick-to-brick variability.
Who should The “Who?” question of data collection cannot be effectively answered without
collect data? reference to human nature and behavior. This is true even in a time when automatic
data collection devices are proliferating. Humans will continue to supervise these
and process the information they generate. Those who collect engineering data must
not only be well trained; they must also be convinced that the data they collect will
be used and in a way that is in their best interests. Good data must be seen as a help
in doing a good job, benefiting an organization, and remaining employed, rather
than as pointless or even threatening. If those charged with collecting or releasing
data believe that the data will be used against them, it is unrealistic to expect them
to produce useful data.
2.1 General Principles in the Collection of Engineering Data 29
Even where those who will gather data are convinced of its importance and are
eager to cooperate, care must be exercised. Personal biases (whether conscious or
subconscious) must not be allowed to enter the data collection process. Sometimes
in a statistical study, hoped-for or predicted best conditions are deliberately or
unwittingly given preference over others. If this is a concern, measurements can be
made blind (i.e., without personnel knowing what set of conditions led to an item
being measured). Other techniques for ensuring fair play, having less to do with
human behavior, will be discussed in the next two sections.
Where should The “Where?” question of engineering data collection can be answered in
data be general terms: “As close as possible in time and space to the phenomenon being
collected? studied.” The importance of this principle is most obvious in the routine monitoring
of complex manufacturing processes. The performance of one operation in such a
process is most effectively monitored at the operation rather than at some later point.
If items being produced turn out to be unsatisfactory at the end of the line, it is rarely
easy to backtrack and locate the operation responsible. Even if that is accomplished,
unnecessary waste has occurred during the time lag between the onset of operation
malfunction and its later discovery.
Example 3 wafers could be identified and eliminated, thus saving the considerable extra
(continued ) expense of further processing. What’s more, the need for adjustments to the
process was signaled in a timely manner.
2.1.3 Recording
The object of engineering data collection is to get data used. How they are recorded
has a major impact on whether this objective is met. A good data recording format
can make the difference between success and failure.
Table 2.1
Mass and Bottom Piece Widths of PVC Bottles
Sample Item Mass (g) Width (mm) Sample Item Mass (g) Width (mm)
First 3 samples
Last 3 samples
31
30
29
Width (mm)
28
27
26
25
24
is easy to see how the two variables are related. If, as in the figure, the recording
symbol is varied over time, it is also easy to track changes in the characteristics
over time. In the present case, width seems to be inversely related to mass, which
appears to be decreasing over time.
Date
Time
Raw Material Lot
Operation
Operator
Machine
Range
Variables Control Chart
1
Zero Equals
2
Measurements
4
Specifications
5
Units
Gage
Sum
Mean
Production Lot(s)
Part and Drawing
Period Covered
Dimension
and values of the experimental variables, and it is also wise to keep track of other
variables that might later prove to be of interest. In the context of routine process
monitoring, data records will be useful in discovering differences in raw material
lots, machines, operators, etc., only if information on these is recorded along with
the responses of interest. Figure 2.2 shows a form commonly used for the routine
collection of measurements for process monitoring. Notice how thoroughly the user
is invited to document the data collection.
Section 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Consider the context of a study on making paper ditions (with a paper clip versus without a paper
airplanes where two different Designs (say delta clip) are of interest with regard to their impact on
versus t wing), two different Papers (say construc- flight distance. Give an operational definition of
tion versus typing), and two different Loading Con- flight distance that you might use in such a study.
2.2 Sampling in Enumerative Studies 33
2. Explain how training operators in the proper use our product—we take 5% samples of every outgo-
of measurement equipment might affect both the ing order, regardless of order size!”?
repeatability and the reproducibility of measure- 4. State briefly why it is critical to make careful oper-
ments made by an organization. ational definitions for response variables in statis-
3. What would be your response to another engi- tical engineering studies.
neer’s comment, “We have great information on
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
properties in any useful way. There is no good way to take information from samples
drawn via these methods and make reliable statements of likely margins of error. The
method introduced next avoids the deficiencies of systematic and judgment-based
sampling.
Methods for actually carrying out the selection of a simple random sample
Mechanical methods include mechanical methods and methods using “random digits.” Mechanical
and simple random methods rely for their effectiveness on symmetry and/or thorough mixing in a
sampling physical randomizing device. So to speak, the slips of paper in the hat need to be of
the same size and well scrambled before sample selection begins.
The first Vietnam-era U.S. draft lottery was a famous case in which adequate
care was not taken to ensure appropriate operation of a mechanical randomizing
device. Birthdays were supposed to be assigned priority numbers 1 through 366 in a
“random” way. However, it was clear after the fact that balls representing birth dates
were placed into a bin by months, and the bin was poorly mixed. When the balls
were drawn out, birth dates near the end of the year received a disproportionately
large share of the low draft numbers. In the present terminology, the first five dates
out of the bin should not have been thought of as a simple random sample of size 5.
Those who operate games of chance more routinely make it their business to know
(via the collection of appropriate data) that their mechanical devices are operating
in a more random manner.
2.2 Sampling in Enumerative Studies 35
1. each digit 0 through 9 has the same chance of appearing at any particular
location in the table one wants to consider, and
2. knowledge of which digit will occur at a given location provides no help in
predicting which one will appear at another.
Table 2.2
Random Digits
12159 66144 05091 13446 45653 13684 66024 91410 51351 22772
30156 90519 95785 47544 66735 35754 11088 67310 19720 08379
59069 01722 53338 41942 65118 71236 01932 70343 25812 62275
54107 58081 82470 59407 13475 95872 16268 78436 39251 64247
99681 81295 06315 28212 45029 57701 96327 85436 33614 29070
36 Chapter 2 Data Collection
12159 66144 05091 13446 45653 13684 66024 91410 51351 22772
30156 90519 95785 47544 66735 35754 11088 67310 19720 08379
Data Display
C2
56 74 43 61 80 22 30 67 35 7
10 69 19 49 8 45 3 37 21 17
2 12 9 14 72
2.2 Sampling in Enumerative Studies 37
Section 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. For the sake of exercise, treat the runout values for 2. Repeat Exercise 1 using statistical or spreadsheet
38 laid gears (given in Table 1.1) as a population software to do the random sampling.
of interest, and using the random digit table (Ta- 3. Explain briefly why in an enumerative study, a sim-
ble B.1), select a simple random sample of 5 of ple random sample is or is not guaranteed to be
these runouts. Repeat this selection process a total representative of the population from which it is
of four different times. (Begin the selection of the drawn.
first sample at the upper left of the table and pro-
ceed left to right and top to bottom.) Are the four
samples identical? Are they each what you would
call “representative” of the population?
38 Chapter 2 Data Collection
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. taxonomy of variables,
2. handling extraneous variables,
3. comparative study,
4. replication, and
5. allocation of resources.
Then Section 2.4 discusses a few generic experimental frameworks for planning a
specific experiment.
Physical process
Response
variable
Managed variables
Concomitant
variables
Some of the variables that are neither primary responses nor managed in an experi-
ment will nevertheless be observed.
Figure 2.4 is an attempt to picture Definitions 2 through 4. In it, the physical process
somehow produces values of a response. “Knobs” on the process represent managed
variables. Concomitant variables are floating about as part of the experimental
environment without being its main focus.
Example 7 The choice Dimond and Dix made to control Drying Time and the Pressure
(continued ) provided a uniform environment for comparing the nine wood/glue combinations.
But strictly speaking, they learned only about joint behavior under their particular
experimental Time and Pressure conditions.
To make projections for other conditions, they had to rely on their expe-
rience and knowledge of material science to decide how far the patterns they
observed were likely to extend. For example, it may have been reasonable to
expect what they observed to also hold up for any drying time at least as long
as the experimental one, because of expert knowledge that the experimental time
was sufficient for the joints to fully set. But such extrapolation is based on other
than statistical grounds.
Example 7 Consider embellishing a bit on the gluing study of Dimond and Dix. Imagine
(continued ) that the students were uneasy about two issues, the first being the possibility that
surface roughness differences in the pieces to be glued might mask the wood/glue
combination differences of interest. Suppose also that because of constraints on
schedules, the strength testing was going to have to be done in two different
sessions a day apart. Measuring techniques or variables like ambient humidity
might vary somewhat between such periods. How might such potential problems
have been handled?
Blocking is one way. If the specimens of each wood type were separated into
relatively rough and relatively smooth groups, the factor Roughness could have
then served as an experimental factor. Each of the glues could have been used the
same number of times to join both rough and smooth specimens of each species.
This would set up comparison of wood/glue combinations separately for rough
and for smooth surfaces.
In a similar way, half the testing for each wood/glue/roughness combination
might have been done in each testing session. Then, any consistent differences
between sessions could be identified and prevented from clouding the comparison
of levels of the primary experimental variables. Thus, Testing Period could have
also served as a blocking variable in the study.
Experimenters usually hope that by careful planning they can account for the
most important extraneous variables via control and blocking. But not all extraneous
variables can be supervised. There are an essentially infinite number, most of which
Randomization cannot even be named. And there is a way to take out insurance against the possibility
and extraneous that major extraneous variables get overlooked and then produce effects that are
variables mistaken for those of the primary experimental variables.
Example 7 Dimond and Dix took the notion of randomization to heart in their gluing study
(continued ) and, so to speak, randomized everything in sight. In the tension strength testing for
a given type of wood, they glued .500 × .500 × 300 blocks to a .7500 × 3.500 × 31.500
board of the same wood type, as illustrated in Figure 2.5.
Each glue was used for three joints on each type of wood. In order to deal
with any unpredicted differences in material properties (e.g., over the extent of
the board) or unforeseen differences in loading by the steel strap used to provide
pressure on the joints, etc., the students randomized the order in which glue was
applied and the blocks placed along the base board. In addition, when it came
time to do the strength testing, that was carried out in a randomly determined
order.
2.3 Principles for Effective Experimentation 43
Metal strap
Wood block
Block position
Wood board
Example 8 In the gear loading study, hanging was the standard method in use at the time
(continued ) of the study. From its records, the company could probably have located some
values for thrust face runout to use as a baseline for evaluating the laying method.
But the choice to run a comparative study, including both laid and hung gears,
put the engineer on firm ground for drawing conclusions about the new method.
A second In a potentially confusing use of language, the word control is sometimes used
usage of to mean the practice of including a standard or no-change sample in an experiment
“control” for comparison purposes. (Notice that this is not the usage in Definition 3.) When
a control group is included in a medical study to verify the effectiveness of a new
drug, that group is either a standard-treatment or no-treatment group, included to
provide a solid basis of comparison for the new treatment.
2.3.4 Replication
In much of what has been said so far, it has been implicit that having more than one
observation for a given setting of experimental variables is a good idea.
Example 10 “like” the two used in the study, they needed to make and test several prototypes
(continued ) for each design.
ISU Professor Emeritus L. Wolins calls the problem of identifying what con-
stitutes replication in an experiment the unit of analysis problem. There must be
replication of the basic experimental unit or object. The agriculturalist who, in order
to study pig blood chemistry, takes hundreds of measurements per hour on one pig,
has a (highly multivariate) sample of size 1. The pig is the unit of analysis.
Without proper replication, one can only hope to be lucky. If experimental error
is small, then accepting conclusions suggested by samples of size 1 will lead to
correct conclusions. But the problem is that without replication, one usually has
little idea of the size of that experimental error.
Section 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Consider again the paper airplane study from Ex- 5. Continuing the paper airplane scenario of Exercise
ercise 1 of Section 2.1. Describe some variables 1 of Section 2.1, discuss the pros and cons of Tom
that you would want to control in such a study. and Juanita flying each of their own eight planes
What are the response and experimental variables twice, as opposed to making and flying two planes
that would be appropriate in this context? Name a of each of the eight types, one time each.
potential concomitant variable here. 6. Random number tables are sometimes used in the
2. In general terms, what is the trade-off that must planning of both enumerative and analytical/ex-
be weighed in deciding whether or not to control a perimental studies. What are the two different ter-
variable in a statistical engineering study? minologies employed in these different contexts,
3. In the paper airplane scenario of Exercise 1 of Sec- and what are the different purposes behind the use
tion 2.1, if (because of schedule limitations, for of the tables?
example) two different team members will make 7. What is blocking supposed to accomplish in an
the flight distance measurements, discuss how the engineering experiment?
notion of blocking might be used. 8. What are some purposes of replication in a statisti-
4. Again using the paper airplane scenario of Exer- cal engineering study?
cise 1 of Section 2.1, suppose that two students are 9. Comment briefly on the notion that in order for
each going to make and fly one airplane of each a statistical engineering study to be statistically
of the 23 = 8 possible types once. Employ the no- proper, one should know before beginning data col-
tion of randomization and Table B.1 and develop lection exactly how an entire experimental budget
schedules for Tom and Juanita to use in their flight is to be spent. (Is this, in fact, a correct idea?)
testing. Explain how the table was used.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Notice that this definition says nothing about how the combinations of settings
of experimental variables included in the study are structured. In fact, they may
be essentially unstructured or produce data with any of the structures discussed
in Section 1.2. That is, there are completely randomized one-factor, factorial, and
fractional factorial experiments. The essential point in Definition 8 is that all else is
randomized except what is restricted by choice of which combinations of levels of
experimental variables are to be used in the study.
Although it doesn’t really fit every situation (or perhaps even most) in which
Paraphrase of the term complete randomization is appropriate, language like the following is
the definition commonly used to capture the intent of Definition 8. “Experimental units (objects)
of complete are allocated at random to the treatment combinations (settings of experimental
randomization variables). Experimental runs are made in a randomly determined order. And any
post-facto measuring of experimental outcomes is also carried out in a random
order.”
100%
HF
75%
HF
50%
HF
Randomization is a good idea. Its virtues have been discussed at some length.
So it would be wise to point out that using it can sometimes lead to practically
unworkable experimental plans. Dogmatic insistence on complete randomization
can in some cases be quite foolish and unrealistic. Changing experimental variables
according to a completely randomly determined schedule can sometimes be exceed-
ingly inconvenient (and therefore expensive). If the inconvenience is great and the
fear of being misled by the effects of extraneous variables is relatively small, then
backing off from complete to partial randomization may be the only reasonable
course of action. But when choosing not to randomize, the implications of that
choice must be carefully considered.
Example 11 From the discussion of replication in the previous section and present con-
(continued ) siderations of complete randomization, it would seem that the purest method of
conducting the study would be to make a new dilution of HF for each of the rods
as its turn comes for testing. But this would be time-consuming and might require
more acid than was available.
If the investigator had three containers to use for baths but limited acid, an
alternative possibility would be to prepare three different dilutions, one 100%,
one 75%, and one 50% dilution. A given dilution could then be used in testing
all rods assigned to that concentration. Notice that this alternative allows for a
randomized order of testing, but it introduces some question as to whether there
is “true” replication.
Taking the resource restriction idea one step further, notice that even if an
investigator could afford only enough acid for making one bath, there is a way
of proceeding. One could do all 100% concentration testing, then dilute the
acid and do all 75% testing, then dilute the acid again and do all 50% testing.
The resource restriction would not only affect the “purity” of replication but also
prevent complete randomization of the experimental order. Thus, for example, any
unintended effects of increased contamination of the acid (as more and more tests
were made using it) would show up in the experimental data as indistinguishable
from effects of differences in acid concentration.
To choose intelligently between complete randomization (with “true” repli-
cation) and the two plans just discussed, the real severity of resource limitations
would have to be weighed against the likelihood that extraneous factors would
jeopardize the usefulness of experimental results.
blocking variables also have one of these standard structures. The essential points
of Definition 9 are the completeness of each block (in the sense that it contains
each setting of the primary variables) and the randomization within each block. The
following two examples illustrate that depending upon the specifics of a scenario,
Definition 9 can describe a variety of experimental plans.
Example 12 As actually run, Gronberg’s golf ball flight study amounted to a randomized
(continued ) complete block experiment. This is because he hit and recorded flight distances
for all 30 balls on six different evenings (over a six-week period). Note that
this allowed him to have (six different) homogeneous conditions under which to
compare the flight distances of balls having 80, 90, and 100 compression. (The
blocks account for possible changes over time in his physical condition and skill
level as well as varied environmental conditions.)
Notice the structure of the data set that resulted from the study. The settings of
the single primary experimental variable Compression combined with the levels
of the single blocking factor Day to produce a 3 × 6 factorial structure for 18
samples of size 10, as pictured in Figure 2.7.
100
Compression
90
Compression
80
Compression
Operator 1
Block 1 1 Run 1 Run 1 Run 1 Run
Day 1
Operator 2
Block 2 1 Run 1 Run 1 Run 1 Run
Day 1
Operator 2
Block 4 1 Run 1 Run 1 Run 1 Run
Day 2
Table 2.4
Half of a 23 Factorial Run Once in Each of Four Blocks
Example 13 In Section 1.2, the pelletizing machine study examined all eight possible com-
(continued ) binations of Volume, Flow, and Mixture. These are listed in Table 2.5. Imagine
that only half of these eight combinations can be run on a given day, and there
is some fear that daily environmental conditions might strongly affect process
performance. How might one proceed?
There are then two blocks (days), each of which will accommodate four
runs. Some possibilities for assigning runs to blocks would clearly be poor. For
example, running combinations 1 through 4 on the first day and 5 through 8 on
54 Chapter 2 Data Collection
the second would make it impossible to distinguish the effects of Mixture from
any important environmental effects.
What turns out to be a far better possibility is to run, say, the four combinations
listed in Table 2.3 (combinations 2, 3, 5, and 8) on one day and the others on
the next. This is illustrated in Table 2.6. In a well-defined sense (explained in
Chapter 8), this choice of an incomplete block plan minimizes the unavoidable
clouding of inferences caused by the fact all eight combinations of levels of
Volume, Flow, and Mixture cannot be run on a single day.
Table 2.6
A 23 Factorial Run in Two Incomplete Blocks
Table 2.7
A Once-Replicated 23 Factorial Run in Four Incomplete
Blocks
There may be some reader uneasiness and frustration with the “rabbit out of a
hat” nature of the examples of incomplete block experiments, since there has been
no discussion of how to go about making up a good incomplete block plan. Both
the choosing of an incomplete block plan and corresponding techniques of data
analysis are advanced topics that will not be developed until Chapter 8. The purpose
here is to simply introduce the possibility of incomplete blocks as a useful option in
experimental planning.
56 Chapter 2 Data Collection
Section 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. What standard name might be applied to the ex- How does the method you used here differ from
perimental plan you developed for Exercise 4 of what you did in part (a)?
Section 2.3? 3. Once more referring to the paper airplane scenario
2. Consider an experimental situation where the three of Exercise 1 of Section 2.1, suppose that only the
factors A, B, and C each have two levels, and it factors Design and Paper are of interest (all planes
is desirable to make three experimental runs for will be made without paper clips) but that Tom and
each of the possible combinations of levels of the Juanita can make and test only two planes apiece.
factors. Devise an incomplete block plan for this study that
(a) Select a completely random order of experi- gives each student experience with both designs
mentation. Carefully describe how you use Ta- and both papers. (Which two planes will each make
ble B.1 or statistical software to do this. Make and test?)
an ordered list of combinations of levels of the 4. Again in the paper airplane scenario of Exercise 1
three factors, prescribing which combination of Section 2.1, suppose that Tom and Juanita each
should be run first, second, etc. have time to make and test only four airplanes
(b) Suppose that because of physical constraints, apiece, but that in toto they still wish to test all eight
only eight runs can be made on a given day. possible types of planes. Develop a sensible plan
Carefully discuss how the concept of blocking for doing this. (Which planes should each person
could be used in this situation when planning test?) You will probably want to be careful to make
which experimental runs to make on each of sure that each person tests two delta wing planes,
three consecutive days. What possible purpose two construction paper planes, and two paper clip
would blocking serve? planes. Why is this? Can you arrange your plan so
(c) Use Table B.1 or statistical software to ran- that each person tests each Design/Paper combina-
domize the order of experimentation within tion, each Design/Loading combination, and each
the blocks you described in part (b). (Make Paper/Loading combination once?
a list of what combinations of levels of the fac- 5. What standard name might be applied to the plan
tors are to be run on each day, in what order.) you developed in Exercise 4?
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
PROBLEM DEFINITION
STUDY DEFINITION
PHYSICAL PREPARATION
These 12 points are listed in a reasonably rational order, but planning any
real study may involve departures from the listed order as well as a fair amount
of iterating among the steps before they are all accomplished. The need for other
steps (like finding funds to pay for a proposed study) will also be apparent in some
contexts. Nevertheless, steps 1 through 12 form a framework for getting started.
1000
500
has studied and asked questions in order to gain expert knowledge about a system
is he or she then in a position to decide intelligently what is not known about the
system—and thus what data will be of help.
It is often helpful at step 2 to make flowcharts describing an ideal process and/or
the process as it is currently operating. (Sometimes the comparison of the two is
enough in itself to show an engineer how a process should be modified.) During the
construction of such a chart, data needs and variables of potential interest can be
identified in an organized manner.
Work
order
Typesetting Yes
Typesetting
needed?
No
Yes Makeready
Makeready
needed?
No
Photo lab
Masking
Plating
Printing
Cutting Yes
Cutting
needed?
No
Yes Folding
Folding
needed?
No
Ship
Example 15 what might go wrong in the printing process and at what points what data could
(continued ) be gathered in order to monitor and improve process performance.
Step 3 After determining the general arena and physical context of a statistical engi-
neering study, it is necessary to agree on a statement of purpose and scope for the
study. An engineering project team assigned to work on a wave soldering process
for printed circuit boards must understand the steps in that process and then begin to
define what part(s) of the process will be included in the study and what the goal(s)
of the study will be. Will flux formulation and application, the actual soldering,
subsequent cleaning and inspection, and touch-up all be studied? Or will only some
part of this list be investigated? Is system throughput the primary concern, or is it
instead some aspect of quality or cost? The sharper a statement of purpose and scope
can be made at this point, the easier subsequent planning steps will be.
Tooling Material
Figure 2.11 Cause and effect diagram for a molding process. From the Third Symposium
on Taguchi Methods.
c Copyright, American Supplier Institute, Dearborn, Michigan
(U.S.A.). Reproduced by permission under License No. 930403.
62 Chapter 2 Data Collection
Example 16 wheel quality. Without some kind of organization, it would be all but impossible
(continued ) to develop anything like a complete list of important factors in a complex situation
like this.
Step 6 Armed with (1) a list of variables that might influence the response(s) of interest
and some guesses at their relative importance, (2) a solid understanding of the issues
raised in Section 2.3, and (3) knowledge of resource and physical constraints and
time-frame requirements, one can begin to make decisions about which (if any)
variables are to be managed. Experiments have some real advantages over purely
observational studies (see Section 1.2). Those must be weighed against possible extra
costs and difficulties associated with managing both variables that are of interest
and those that are not. The hope is to choose a physically and financially workable
set of managed variables in such a way that the aggregate effects of variables not of
interest and not managed are not so large as to mask the effects of those variables
that are of interest.
Step 7 Choosing experimental levels and then combinations for managed variables
is part of the task of deciding on a detailed data collection protocol. Levels of
controlled and block variables should usually be chosen to be representative of
the values that will be met in routine system operation. For example, suppose the
amount of contamination in a transmission’s hydraulic fluid is thought to affect
time to failure when the transmission is subjected to stress testing, where Operating
Speed and Pressure are the primary experimental variables. It only makes sense to
see that the contamination level(s) during testing are representative of the level(s)
that will be typical when the transmission is used in the field.
With regard to primary experimental variables, one should also choose typical
levels—with a couple of provisos. Sometimes the goal in an engineering experiment
is to compare an innovative, nonstandard way of doing things to current practice.
In such cases, it is not good enough simply to look at system behavior with typical
settings for primary experimental variables. Also, where primary experimental vari-
ables are believed to have relatively small effects on a response, it may be necessary
to choose ranges for the primary variables that are wider than normal, to see clearly
how they act on the response.
Other physical realities and constraints on data collection may also make it
appropriate to use atypical values of managed variables and subsequently extrapolate
experimental results to “standard” circumstances. For example, it is costly enough to
run studies on pilot plants using small quantities of chemical reagents and miniature
equipment but much cheaper than experimentation on a full-scale facility. Another
kind of engineering study in which levels of primary experimental variables are
purposely chosen outside normal ranges is the accelerated life test. Such studies
are done to predict the life-length properties of products that in normal usage would
far outlast any study of feasible length. All that can then be done is to turn up
the stress on sample units beyond normal levels, observe performance, and try to
extrapolate back to a prediction for behavior under normal usage. (For example, if
sensitive electronic equipment performs well under abnormally high temperature
2.5 Preparing to Collect Engineering Data 63
and humidity, this could well be expected to imply long useful life under normal
temperature and humidity conditions.)
After the experimental levels of individual manipulated variables are chosen,
they must be combined to form the experimental patterns (combinations) of man-
aged variables. The range of choices is wide: factorial structures, fractional factorial
structures, other standard structures, and patterns tailor-made for a particular prob-
lem. (Tailor-made plans will, for example, be needed in situations where particular
combinations of factor levels prescribed by standard structures are a priori clearly
unsafe or destructive of company property.)
But developing a detailed data collection protocol requires more than even
choices of experimental combinations. Experimental order must be decided. Explicit
instructions for actually carrying out the testing must be agreed upon and written
down in such a way that someone who was not involved in study planning can carry
out the data collection. A timetable for initial data collection must be developed.
In all of this, it must be remembered that several iterations of data collection and
analysis (all within given budget constraints) may be required in order to find a
solution to the original engineering problem.
Section 5 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Either take an engineering system and response tomer Satisfaction and make a cause-and-effect di-
variable that you are familiar with from your field agram showing a variety of variables that may po-
or consider, for example, the United Airlines pas- tentially affect the response. How might such a
senger flight system and the response variable Cus- diagram be practically useful?
Chapter 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Use Table B.1 and choose a simple random sam- carefully how you do this. If you use the table,
ple of n = 8 out of N = 491 widgets. Describe begin in the upper left corner.
carefully how you label the widgets. Begin in the (b) What are some purposes of the randomization
upper left corner of the table. Then use spread- used in part (a)?
sheet or statistical software to redo the selection. 4. A sanitary engineer wishes to compare two meth-
2. Consider a potential student project concerning ods for determining chlorine content of Cl2 -
the making of popcorn. Possible factors affecting demand-free water. To do this, eight quite dif-
the outcome of popcorn making include at least ferent water samples are split in half, and one
the following: Brand of corn, Temperature of corn determination is made using the MSI method and
at beginning of cooking, Popping Method (e.g., another using the SIB method. Explain why it
frying versus hot air popping), Type of Oil used could be said that the principle of blocking was
(if frying), Amount of Oil used (if frying), Batch used in the engineer’s study. Also argue that the
Size, initial Moisture Content of corn, and Person resulting data set could be described as consisting
doing the evaluation of a single batch. Using these of paired measurement data.
factors and/or any others that you can think of, an- 5. A research group is testing three different meth-
swer the following questions about such a project: ods of electroplating widgets (say, methods A, B,
(a) What is a possible response variable in a pop- and C). On a particular day, 18 widgets are avail-
corn project? able for testing. The effectiveness of electroplat-
(b) Pick two possible experimental factors in this ing may be strongly affected by the surface texture
context and describe a 2 × 2 factorial data of the widgets. The engineer running the exper-
structure in those variables that might arise in iment is able to divide the 18 available widgets
such a study. into three groups of 6 on the basis of surface tex-
(c) Describe how the concept of randomization ture. (Assume that widgets 1–6 are rough, widgets
might be employed. 7–12 are normal, and widgets 13–18 are smooth.)
(d) Describe how the concept of blocking might (a) Use Table B.1 or statistical software in an
be employed. appropriate way and assign each of the treat-
3. An experiment is to be performed to compare the ments to 6 widgets. Carefully explain exactly
effects of two different methods for loading gears how you do the assignment of levels of treat-
in a carburizing furnace on the amount of distor- ments A, B, and C to the widgets.
tion produced in a heat treating process. Thrust (b) If equipment limitations are such that only
face runout will be measured for gears laid and one widget can be electroplated at once, but
for gears hung while treating. it is possible to complete the plating of all 18
(a) 20 gears are to be used in the study. Randomly widgets on a single day, in exactly what order
divide the gears into a group (of 10) to be laid would you have the widgets plated? Explain
and a group (of 10) to be hung, using either where you got this order.
Table B.1 or statistical software. Describe (c) If, in contrast to the situation in part (b), it is
Chapter 2 Exercises 65
possible to plate only 9 widgets in a single observed values of the two responses could
day, make up an appropriate plan for plating be entered for each experimental run.
9 on each of two consecutive days. (b) Suppose that it is feasible to make the runs
(d) If measurements of plating effectiveness are listed in your answer to part (a) in a com-
made on each of the 18 widgets, what kind of pletely randomized order. Use a mechanical
data structure will result from the scenario in method (like slips of paper in a hat) to arrive at
part (b)? From the scenario in part (c)? a random order of experimentation for your
6. A company wishes to increase the light intensity study. Carefully describe the physical steps
of its photoflash cartridge. Two wall thicknesses you follow in developing this order for data
1 00 00 collection.
( 16 and 18 ) and two ignition point placements are
under study. Two batches of the basic formulation 9. Use Table B.1 and
used in the cartridge are to be made up, each (a) Select a simple random sample of 7 widgets
batch large enough to make 12 cartridges. Discuss from a production run of 619 widgets (begin
how you would recommend running this initial at the upper left corner of the table and move
phase of experimentation if all cartridges can be left to right, top to bottom). Tell how you la-
made and tested in a short time period by a single beled the widgets and name which ones make
technician. Be explicit about any randomization up your sample.
and/or blocking you would employ. Say exactly (b) Beginning in the table where you left off in
what kinds of cartridges you would make and test, (a), select a second simple random sample of
in what order. Describe the structure of the data 7 widgets. Is this sample the same as the first?
that would result from your study. Is there any overlap at all?
7. Use Table B.1 or statistical software and 10. Redo Exercise 9 using spreadsheet or statistical
(a) Select a simple random sample of 5 widgets software.
from a production run of 354 such widgets. 11. Consider a study comparing the lifetimes (mea-
(If you use the table, begin at the upper left sured in terms of numbers of holes drilled before
corner and move left to right, top to bottom.) failure) of two different brands of 8-mm drills in
(b) Select a random order of experimentation for drilling 1045 steel. Suppose that steel bars from
a context where an experimental factor A has three different heats (batches) of steel are avail-
two levels; a second factor, B, has three lev- able for use in the study, and it is possible that the
els; and two experimental runs are going to different heats have differing physical properties.
be made for each of the 2 × 3 = 6 different The lifetimes of a total of 15 drills of each brand
possible combinations of levels of the factors. will be measured, and each of the bars available
Carefully describe how you do this. is large enough to accommodate as much drilling
8. Return to the situation of Exercise 8 of the Chap- as will be done in the entire study.
ter 1 Exercises. (a) Describe how the concept of control could be
(a) Name factors and levels that might be used in used to deal with the possibility that different
a three-factor, full factorial study in this situ- heats might have different physical properties
ation. Also name two response variables for (such as hardnesses).
the study. Suppose that in accord with good (b) Name one advantage and one drawback to
engineering data collection practice, you wish controlling the heat.
to include some replication in the study. Make (c) Describe how one might use the concept of
up a data collection sheet, listing all the com- blocking to deal with the possibility that dif-
binations of levels of the factors to be studied, ferent heats might have different physical
and include blanks where the corresponding properties.
3
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Elementary
Descriptive Statistics
Engineering data are always variable. Given precise enough measurement, even
supposedly constant process conditions produce differing responses. Therefore, it is
not individual data values that demand an engineer’s attention as much as the pattern
or distribution of those responses. The task of summarizing data is to describe their
important distributional characteristics. This chapter discusses simple methods that
are helpful in this task.
The chapter begins with some elementary graphical and tabular methods of
data summarization. The notion of quantiles of a distribution is then introduced and
used to make other useful graphical displays. Next, standard numerical summary
measures of location and spread for quantitative data are discussed. Finally comes a
brief look at some elementary methods for summarizing qualitative and count data.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
66
3.1 Elementary Graphical and Tabular Treatment of Quantitative Data 67
Gears laid
0 10 20 30 40
Runout (.0001 in.)
Gears hung
0 10 20 30 40
Table 3.1
Bullet Penetration Depths (mm)
20 30 40 50 60 70
Penetration (mm)
20 30 40 50 60 70
Penetration (mm)
The dot diagrams show the penetrations of the 200 grain bullets to be both
larger and more consistent than those of the 230 grain bullets. (The students
had predicted larger penetrations for the lighter bullets on the basis of greater
muzzle velocity and smaller surface area on which friction can act. The different
consistencies of penetration were neither expected nor explained.)
Dot diagrams give the general feel of a data set but do not always allow the
recovery of exactly the values used to make them. A stem-and-leaf plot carries
much the same visual information as a dot diagram while preserving the original
values exactly. A stem-and-leaf plot is made by using the last few digits of each data
point to indicate where it falls.
0 5 8 9 9 9 9
1 0 0 1 1 1 1 1 1 1 2 2 2 2 3 3 3 3 4 4 4 5 5 5 5 6 7 7 8 9
2 7
3
0
0 5 8 9 9 9 9
1 0 0 1 1 1 1 1 1 1 2 2 2 2 3 3 3 3 4 4 4
1 5 5 5 6 7 7 8 9
2
2 7
3
3
Example 1 Figure 3.3 gives two possible stem-and-leaf plots for the thrust face runouts
(continued ) of laid gears. In both, the first digit of each observation is represented by
the number to the left of the vertical line or “stem” of the diagram. The
numbers to the right of the vertical line make up the “leaves” and give the
second digits of the observed runouts. The second display shows somewhat
more detail than the first by providing “0–4” and “5–9” leaf positions for each
possible leading digit, instead of only a single “0–9” leaf for each leading
digit.
Example 2 Figure 3.4 gives two possible stem-and-leaf plots for the penetrations of 200 grain
(continued ) bullets in Table 3.1. On these, it was convenient to use two digits to the left of
the decimal point to make the stem and the two following the decimal point to
create the leaves. The first display was made by recording the leaf values directly
from the table (from left to right and top to bottom). The second display is a
better one, obtained by ordering the values that make up each leaf. Notice that
both plots give essentially the same visual impression as the second dot diagram
in Figure 3.2.
When comparing two data sets, a useful way to use the stem-and-leaf idea is to
make two plots back-to-back.
70 Chapter 3 Elementary Descriptive Statistics
Example 1 Figure 3.5 gives back-to-back stem-and-leaf plots for the data of Table 1.1 (pg. 3).
(continued ) It shows clearly the differences in location and spread of the two data sets.
Example 1 Table 3.2 gives one possible frequency table for the laid gear runouts. The relative
(continued ) frequency values are obtained by dividing the entries in the frequency column
Table 3.2
Frequency Table for Laid Gear Thrust Face Runouts
Cumulative
Runout Relative Relative
(.0001 in.) Tally Frequency Frequency Frequency
by 38, the number of data points. The entries in the cumulative relative frequency
column are the ratios of the totals in a given class and all preceding classes to the
total number of data points. (Except for round-off, this is the sum of the relative
frequencies on the same row and above a given cumulative relative frequency.)
The tally column gives the same kind of information about distributional shape
that is provided by a dot diagram or a stem-and-leaf plot.
Choosing intervals The choice of intervals to use in making a frequency table is a matter of
for a frequency judgment. Two people will not necessarily choose the same set of intervals. However,
table there are a number of simple points to keep in mind when choosing them. First, in
order to avoid visual distortion when using the tally column of the table to gain an
impression of distributional shape, intervals of equal length should be employed.
Also, for aesthetic reasons, round numbers are preferable as interval endpoints. Since
there is usually aggregation (and therefore some loss of information) involved in the
reduction of raw data to tallies, the larger the number of intervals used, the more
detailed the information portrayed by the table. On the other hand, if a frequency
table is to have value as a summarization of data, it can’t be cluttered with too many
intervals.
After making a frequency table, it is common to use the organization provided
by the table to create a histogram. A (frequency or relative frequency) histogram is
a kind of bar chart used to portray the shape of a distribution of data points.
Example 2 Table 3.3 is a frequency table for the 200 grain bullet penetration depths, and
(continued ) Figure 3.6 is a translation of that table into the form of a histogram.
Table 3.3
Frequency Table for 200 Grain Penetration Depths
Cumulative
Penetration Relative Relative
Depth (mm) Tally Frequency Frequency Frequency
Example 2
(continued )
6
Frequency
4
60 70
Penetration depth (mm)
The vertical scale in Figure 3.6 is a frequency scale, and the histogram is a frequency
histogram. By changing to relative frequency on the vertical scale, one can produce
a relative frequency histogram. In making Figure 3.6, care was taken to
Following these guidelines results in a display in which equal enclosed areas cor-
respond to equal numbers of data points. Further, data point positioning is clearly
indicated by bar positioning on the horizontal axis. If these guidelines are not fol-
lowed, the resulting bar chart will in one way or another fail to faithfully represent
its data set.
Figure 3.7 shows terminology for common distributional shapes encountered
when making and using dot diagrams, stem-and-leaf plots, and histograms.
The graphical and tabular devices discussed to this point are deceptively simple
methods. When routinely and intelligently used, they are powerful engineering
tools. The information on location, spread, and shape that is portrayed so clearly on
a histogram can give strong hints as to the functioning of the physical process that
is generating the data. It can also help suggest physical mechanisms at work in the
Examples of process.
engineering For example, if data on the diameters of machined metal cylinders purchased
interpretations of from a vendor produce a histogram that is decidedly bimodal (or multimodal,
distribution shape having several clear humps), this suggests that the machining of the parts was done
3.1 Elementary Graphical and Tabular Treatment of Quantitative Data 73
on more than one machine, or by more than one operator, or at more than one
time. The practical consequence of such multichannel machining is a distribution
of diameters that has more variation than is typical of a production run of cylinders
from a single machine, operator, and setup. As another possibility, if the histogram
is truncated, this might suggest that the lot of cylinders has been 100% inspected
and sorted, removing all cylinders with excessive diameters. Or, upon marking
engineering specifications (requirements) for cylinder diameter on the histogram,
one may get a picture like that in Figure 3.8. It then becomes obvious that the lathe
turning the cylinders needs adjustment in order to increase the typical diameter.
But it also becomes clear that the basic process variation is so large that this
adjustment will fail to bring essentially all diameters into specifications. Armed
with this realization and a knowledge of the economic consequences of parts failing
to meet specifications, an engineer can intelligently weigh alternative courses of
action: sorting of all incoming parts, demanding that the vendor use more precise
equipment, seeking a new vendor, etc.
Investigating the shape of a data set is useful not only because it can lend insight
into physical mechanisms but also because shape can be important when determining
the appropriateness of methods of formal statistical inference like those discussed
later in this book. A methodology appropriate for one distributional shape may not
be appropriate for another.
Lower Upper
specification specification
Cylinder diameter
Table 3.4
Torques Required to Loosen Two Bolts on Face Plates (ft lb)
1 16 16 18 15 14
2 15 16 19 17 17
3 15 17 20 14 16
4 15 16 21 17 18
5 20 20 22 19 16
6 19 16 23 19 18
7 19 20 24 19 20
8 17 19 25 15 15
9 15 15 26 12 15
10 11 15 27 18 20
11 17 19 28 13 18
12 18 17 29 14 18
13 18 14 30 18 18
14 15 15 31 18 14
15 18 17 32 15 13
16 15 17 33 16 17
17 18 20 34 16 16
3.1 Elementary Graphical and Tabular Treatment of Quantitative Data 75
20 2 2
2
10 15 20
Bolt 3 torque (ft lb)
otherwise, unwanted differential forces might act on the face plate. It is also quite
reasonable that bolt 3 and bolt 4 torques be related, since the bolts were tightened
by different heads of a single pneumatic wrench operating off a single source of
compressed air. It stands to reason that variations in air pressure might affect the
tightening of the bolts at the two positions similarly, producing the big-together,
small-together pattern seen in Figure 3.9.
The previous example illustrates the point that relationships seen on scatterplots
suggest a common physical cause for the behavior of variables and can help reveal
that cause.
In the most common version of the scatterplot, the variable on the horizontal
axis is a time variable. A scatterplot in which univariate data are plotted against time
order of observation is called a run chart or trend chart. Making run charts is one
of the most helpful statistical habits an engineer can develop. Seeing patterns on a
run chart leads to thinking about what process variables were changing in concert
with the pattern. This can help develop a keener understanding of how process
behavior is affected by those variables that change over time.
Diameter Diameter
Joint (inches above nominal) Joint (inches above nominal)
1 −.005 16 .015
2 .000 17 .000
3 −.010 18 .000
4 −.030 19 −.015
5 −.010 20 −.015
6 −.025 21 −.005
7 −.030 22 −.015
8 −.035 23 −.015
9 −.025 24 −.010
10 −.025 25 −.015
11 −.025 26 −.035
12 −.035 27 −.025
13 −.040 28 −.020
14 −.035 29 −.025
15 −.035 30 −.015
+.020 +.020
Diameter
Diameter
.000 .000
–.020 –.020
–.040 –.040
5 10 15 20 25 30
Time of manufacture
Figure 3.10 Dot diagram and run chart of consecutive outer diameters
lathe. Figure 3.10 gives both a dot diagram and a run chart for the data in the
table. In keeping with standard practice, consecutive points on the run chart have
been connected with line segments.
Here the dot diagram is not particularly suggestive of the physical mecha-
nisms that generated the data. But the time information added in the run chart
is revealing. Moving along in time, the outer diameters tend to get smaller until
3.2 Quantiles and Related Graphical Tools 77
part 16, where there is a large jump, followed again by a pattern of diameter gen-
erally decreasing in time. In fact, upon checking production records, Williams
and Markowski found that the lathe had been turned off and allowed to cool down
between parts 15 and 16. The pattern seen on the run chart is likely related to the
behavior of the lathe’s hydraulics. When cold, the hydraulics probably don’t do
as good a job pushing the cutting tool into the part being turned as when they are
warm. Hence, the turned parts become smaller as the lathe warms up. In order
to get parts closer to nominal, the aimed-for diameter might be adjusted up by
about .020 in. and parts run only after warming up the lathe.
Section 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. The following are percent yields from 40 runs of (b) A trick often employed in the analysis of paired
a chemical process, taken from J. S. Hunter’s arti- data such as these is to reduce the pairs to dif-
cle “The Technology of Quality” (RCA Engineer, ferences by subtracting the values of one of the
May/June 1985): variables from the other. Compute differences
(top bolt–bottom bolt) here. Then make and
65.6, 65.6, 66.2, 66.8, 67.2, 67.5, 67.8, 67.8, 68.0, interpret a dot diagram for these values.
68.0, 68.2, 68.3, 68.3, 68.4, 68.9, 69.0, 69.1, 69.2,
69.3, 69.5, 69.5, 69.5, 69.8, 69.9, 70.0, 70.2, 70.4, Piece Top Bolt Bottom Bolt
70.6, 70.6, 70.7, 70.8, 70.9, 71.3, 71.7, 72.0, 72.6,
72.7, 72.8, 73.5, 74.2 1 110 125
2 115 115
Make a dot diagram, a stem-and-leaf plot, a fre- 3 105 125
quency table, and a histogram of these data. 4 115 115
2. Make back-to-back stem-and-leaf plots for the two 5 115 120
samples in Table 3.1. 6 120 120
3. Osborne, Bishop, and Klein collected manufactur- 7 110 115
ing data on the torques required to loosen bolts 8 125 125
holding an assembly on a piece of heavy machin- 9 105 110
ery. The accompanying table shows part of their 10 130 110
data concerning two particular bolts. The torques 11 95 120
recorded (in ft lb) were taken from 15 different 12 110 115
pieces of equipment as they were assembled. 13 110 120
(a) Make a scatterplot of these paired data. Are 14 95 115
there any obvious patterns in the plot? 15 105 105
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
the test had worse scores, and roughly 20% had better scores. This concept is also
useful in the description of engineering data. However, because it is often more
convenient to work in terms of fractions between 0 and 1 rather than in percentages
between 0 and 100, slightly different terminology will be used here: “Quantiles,”
rather than percentiles, will be discussed. After the quantiles of a data set are carefully
defined, they are used to create a number of useful tools of descriptive statistics:
quantile plots, boxplots, Q-Q plots, and normal plots (a type of theoretical Q-Q
plot).
Definition 1 For a data set consisting of n values that when ordered are x1 ≤ x2 ≤ · · · ≤ xn ,
1. if p = i−.5
n
for a positive integer i ≤ n, the p quantile of the data
set is
i − .5
Q( p) = Q = xi
n
i−.5
(The ith smallest data point will be called the n
quantile.)
2. for any number p between .5n and n−.5n
that is not of the form i−.5
n
for
an integer i, the p quantile of the data set will be obtained by linear
interpolation between the two values of Q( i−.5 n
) with corresponding
i−.5
n
that bracket p.
Table 3.6
Ten Paper Towel Breaking
Strengths
Table 3.7
Quantiles of the Paper Towel Breaking Strength
Distribution
i −.5 i −.5
i 10
ith Smallest Data Point, xi = Q 10
Example 5 the smallest 3 data points and half of the fourth smallest are counted as lying to
(continued ) the left of the desired number, and the largest 6 data points and half of the seventh
largest are counted as lying to the right. Thus, the fourth smallest data point must
be the .35 quantile, as is shown in Table 3.7.
To illustrate convention (2) of Definition 1, consider finding the .5 and .93
.5−.45
quantiles of the strength distribution. Since .5 is .55−.45 = .5 of the way from .45
to .55, linear interpolation gives
Definition 3 Q(.25) and Q(.75) are called the first (or lower) quartile and third (or
upper) quartile of a distribution, respectively.
Example 5 Referring again to Table 3.7 and the value of Q(.5) previously computed, for the
(continued ) breaking strength distribution
Definition 4 A quantile plot is a plot of Q( p) versus p. For an ordered data set of size
n containing values x1 ≤ x2 ≤ · · · ≤ xn , such a display is made by first plot-
ting the points ( i−.5
n
, xi ) and then connecting consecutive plotted points with
straight-line segments.
It is because convention (2) in Definition 1 calls for linear interpolation that straight-
line segments enter the picture in making a quantile plot.
3.2 Quantiles and Related Graphical Tools 81
Q( p)
10,000
9,000
8,000
7,000
.1 .2 .3 .4 .5 .6 .7 .8 .9 p
A quantile plot allows the user to do some informal visual smoothing of the plot to
compensate for any jaggedness. (The tacit assumption is that the underlying data-
generating mechanism would itself produce smoother and smoother quantile plots
for larger and larger samples.)
3.2.2 Boxplots
Familiarity with the quantile idea is the principal prerequisite for making boxplots,
an alternative to dot diagrams or histograms. The boxplot carries somewhat less
information, but it has the advantage that many can be placed side-by-side on a
single page for comparison purposes.
There are several common conventions for making boxplots. The one that will
be used here is illustrated in generic fashion in Figure 3.12. A box is made to extend
from the first to the third quartiles and is divided by a line at the median. Then the
interquartile range
Interquartile
range
IQR= Q(.75) − Q(.25)
82 Chapter 3 Elementary Descriptive Statistics
is calculated and the smallest data point within 1.5IQR of Q(.25) and the largest
data point within 1.5IQR of Q(.75) are determined. Lines called whiskers are made
to extend out from the box to these values. Typically, most data points will be within
the interval [Q(.25) − 1.5IQR, Q(.75) + 1.5IQR]. Any that are not then get plotted
individually and are thereby identified as outlying or unusual.
Example 5 Consider making a boxplot for the paper towel breaking strength data. To begin,
(continued )
Q(.25) = 8,572 g
Q(.5) = 9,088 g
Q(.75) = 9,614 g
So
and
1.5IQR = 1,563 g
Then
and
Since all the data points lie in the range 7,009 g to 11,177 g, the boxplot is as
shown in Figure 3.13.
9,088
7,583 10,688
8,572 9,614
A boxplot shows distributional location through the placement of the box and
whiskers along a number line. It shows distributional spread through the extent of
the box and the whiskers, with the box enclosing the middle 50% of the distribution.
Some elements of distributional shape are indicated by the symmetry (or lack
thereof) of the box and of the whiskers. And a gap between the end of a whisker
and a separately plotted point serves as a reminder that no data values fall in that
interval.
Two or more boxplots drawn to the same scale and side by side provide an
effective way of comparing samples.
So
Example 6 Similar calculations for the 200 grain bullet penetration depths yield
(continued )
Q(.25) = 60.25 mm
Q(.5) = 62.80 mm
Q(.75) = 64.35 mm
Q(.75) + 1.5IQR = 70.50 mm
Q(.25) − 1.5IQR = 54.10 mm
Table 3.8
Quantiles of the Bullet Penetration Depth Distributions
Figure 3.14 then shows boxplots placed side by side on the same scale. The
plots show the larger and more consistent penetration depths of the 200 grain
bullets. They also show the existence of one particularly extreme data point in
the 200 grain data set. Further, the relative lengths of the whiskers hint at some
skewness (recall the terminology introduced with Figure 3.7) in the data. And
all of this is done in a way that is quite uncluttered and compact. Many more of
3.2 Quantiles and Related Graphical Tools 85
70
200 Grain
50
230 Grain
bullets
40
30
these boxes could be added to Figure 3.14 (to compare other bullet types) without
visual overload.
Then, recognizing ordered data values as quantiles and letting Q 1 and Q 2 stand for
the quantile functions of the two respective data sets, it is clear from display (3.1)
that
Q 2 ( p) = 2Q 1 ( p) + 1 (3.2)
Table 3.9
Two Small Artificial Data Sets
Q2( p)
15
14
13
12
Data set 1 11
10
3 4 5 6 7 8
9
8
Data set 2
7 2
7 9 11 13 15 17 3 4 5 6 7 Q1( p)
Figure 3.15 Dot diagrams for two Figure 3.16 Q- Q plot for the data
small data sets of Table 3.9
That is, the two data sets have quantile functions that are linearly related. Looking
at either display (3.1) or (3.2), it is obvious that a plot of the points
i − .5 i − .5
Q1 , Q2
5 5
Definition 5 A Q-Q plot for two data sets with respective quantile functions Q 1 and Q 2 is
a plot of ordered pairs (Q 1 ( p), Q 2 ( p)) for appropriate values of p. When two
data sets of size n are involved, the values of p used to make the plot will be
i−.5
n
for i = 1, 2, . . . , n. When two data sets of unequal sizes are involved, the
values of p used to make the plot will be i−.5 n
for i = 1, 2, . . . , n, where n is
the size of the smaller set.
To make a Q-Q plot for two data sets of the same size,
Steps in making 1. order each from the smallest observation to the largest,
a Q-Q plot
2. pair off corresponding values in the two data sets, and
3. plot ordered pairs, with the horizontal coordinates coming from the first data
set and the vertical ones from the second.
When data sets of unequal size are involved, the ordered values from the smaller
data set must be paired with quantiles of the larger data set obtained by interpolation.
3.2 Quantiles and Related Graphical Tools 87
A Q-Q plot that is reasonably linear indicates the two distributions involved have
similar shapes. When there are significant departures from linearity, the character
of those departures reveals the ways in which the shapes differ.
Example 6 Returning again to the bullet penetration depths, Table 3.8 (page 84) gives the
(continued ) raw material for making a Q-Q plot. The depths on each row of that table need
only be paired and plotted in order to make the plot given in Figure 3.17.
The scatterplot in Figure 3.17 is not terribly linear when looked at as a whole.
However, the points corresponding to the 2nd through 13th smallest values in
each data set do look fairly linear, indicating that (except for the extreme lower
ends) the lower ends of the two distributions have similar shapes.
The horizontal jog the plot takes between the 13th and 14th plotted points
indicates that the gap between 43.85 mm and 47.30 mm (for the 230 grain data)
is out of proportion to the gap between 63.55 and 63.80 mm (for the 200 grain
data). This hints that there was some kind of basic physical difference in the
mechanisms that produced the smaller and larger 230 grain penetration depths.
Once this kind of indication is discovered, it is a task for ballistics experts or
materials people to explain the phenomenon.
Because of the marked departure from linearity produced by the 1st plotted
point (27.75, 58.00), there is also a drastic difference in the shapes of the extreme
lower ends of the two distributions. In order to move that point back on line with
the rest of the plotted points, it would need to be moved to the right or down
(i.e., increase the smallest 230 grain observation or decrease the smallest 200
grain observation). That is, relative to the 200 grain distribution, the 230 grain
distribution is long-tailed to the low side. (Or to put it differently, relative to
the 230 grain distribution, the 200 grain distribution is short-tailed to the low
side.) Note that the difference in shapes was already evident in the boxplot in
Figure 3.14. Again, it would remain for a specialist to explain this difference in
distributional shapes.
200 Grain pentration (mm)
70
60
50
20 30 40 50 60
230 Grain penetration (mm)
The Q-Q plotting idea is useful when applied to two data sets, and it is easiest to
explain the notion in such an “empirical versus empirical” context. But its greatest
usefulness is really when it is applied to one quantile function that represents a data
set and a second that represents a theoretical distribution.
Definition 6 A theoretical Q-Q plot or probability plot for a data set of size n and a
theoretical distribution, with respective quantile functions Q 1 and Q 2 , is a plot
of ordered pairs (Q 1 ( p), Q 2 ( p)) for appropriate values of p. In this text, the
values of p of the form i−.5n
for i = 1, 2, . . . , n will be used.
Recognizing Q 1 ( i−.5
n
) as the ith smallest data point, one sees that a theoretical
Q-Q plot is a plot of points with horizontal plotting positions equal to observed data
and vertical plotting positions equal to quantiles of the theoretical distribution. That
is, with ordered data x 1 ≤ x2 ≤ · · · ≤ xn , the points
Ordered pairs i − .5
making a xi , Q 2
n
probability plot
are plotted. Such a plot allows one to ask, “Does the data set have a shape similar to
the theoretical distribution?”
Normal The most famous version of the theoretical Q-Q plot occurs when quantiles for
plotting the standard normal or Gaussian distribution are employed. This is the familiar
bell-shaped distribution. Table 3.10 gives some quantiles of this distribution. In
order to find Q( p) for p equal to one of the values .01, .02, . . . , .98, .99, locate the
entry in the row labeled by the first digit after the decimal place and in the column
labeled by the second digit after the decimal place. (For example, Q(.37) = −.33.)
A simple numerical approximation to the values given in Table 3.10 adequate for
most plotting purposes is
The origin of Table 3.10 is not obvious at this point. It will be explained in
Section 5.2, but for the time being consider the following crude argument to the
effect that the quantiles in the table correspond to a bell-shaped distribution. Imagine
that each entry in Table 3.10 corresponds to a data point in a set of size n = 99. A
possible frequency table for those 99 data points is given as Table 3.11. The tally
column in Table 3.11 shows clearly the bell shape.
The standard normal quantiles can be used to make a theoretical Q-Q plot as
a way of assessing how bell-shaped a data set looks. The resulting plot is called a
normal (probability) plot.
3.2 Quantiles and Related Graphical Tools 89
Table 3.10
Standard Normal Quantiles
.00 .01 .02 .03 .04 .05 .06 .07 .08 .09
.0 −2.33 −2.05 −1.88 −1.75 −1.65 −1.55 −1.48 −1.41 −1.34
.1 −1.28 −1.23 −1.18 −1.13 −1.08 −1.04 −.99 −.95 −.92 −.88
.2 −.84 −.81 −.77 −.74 −.71 −.67 −.64 −.61 −.58 −.55
.3 −.52 −.50 −.47 −.44 −.41 −.39 −.36 −.33 −.31 −.28
.4 −.25 −.23 −.20 −.18 −.15 −.13 −.10 −.08 −.05 −.03
.5 0.00 .03 .05 .08 .10 .13 .15 .18 .20 .23
.6 .25 .28 .31 .33 .36 .39 .41 .44 .47 .50
.7 .52 .55 .58 .61 .64 .67 .71 .74 .77 .81
.8 .84 .88 .92 .95 .99 1.04 1.08 1.13 1.18 1.23
.9 1.28 1.34 1.41 1.48 1.55 1.65 1.75 1.88 2.05 2.33
Table 3.11
A Frequency Table for the Standard Normal Quantiles
−2.80 to −2.30 1
−2.29 to −1.79 2
−1.78 to −1.28 7
−1.27 to −.77 12
−.76 to −.26 17
−.25 to .25 21
.26 to .76 17
.77 to 1.27 12
1.28 to 1.78 7
1.79 to 2.29 2
2.30 to 2.80 1
Example 5 Consider again the paper towel strength testing scenario and now the issue of
(continued ) how bell-shaped the data set in Table 3.6 (page 79) is. Table 3.12 was made using
Tables 3.7 (page 79) and 3.10; it gives the information needed to produce the
theoretical Q-Q plot in Figure 3.18.
Considering the small size of the data set involved, the plot in Figure 3.18
is fairly linear, and so the data set is reasonably bell-shaped. As a practical
consequence of this judgment, it is then possible to use the normal probability
models discussed in Section 5.2 to describe breaking strength. These could be
employed to make breaking strength predictions, and methods of formal statistical
inference based on them could be used in the analysis of breaking strength data.
90 Chapter 3 Elementary Descriptive Statistics
2.0
Standard normal quantile
1.0
–1.0
Special graph paper, called normal probability paper (or just probability
paper), is available as an alternative way of making normal plots. Instead of plotting
points on regular graph paper using vertical plotting positions taken from Table 3.10,
points are plotted on probability paper using vertical plotting positions of the form
i−.5
n
. Figure 3.19 is a normal plot of the breaking strength data from Example 5 made
on probability paper. Observe that this is virtually identical to the plot in Figure 3.18.
Normal plots are not the only kind of theoretical Q-Q plots useful to engineers.
Many other types of theoretical distributions are of engineering importance, and
each can be used to make theoretical Q-Q plots. This point is discussed in more
3.2 Quantiles and Related Graphical Tools 91
99.99
0.01
0.2 0.1 0.05
99.8 99.9
0.5
99
1
98
2
95
5
10
90
20
80
70
30
40
60
50
50
60
40
30
70
80
20
90
10
95
5
98
2
99
1
0.5
99.9 99.8
0.01
Figure 3.19 Normal plot for the paper towel strengths (made on probability paper,
used with permission of the Keuffel and Esser Company)
detail in Section 5.3, but the introduction of theoretical Q-Q plotting here makes it
possible to emphasize the relationship between probability plotting and (empirical)
Q-Q plotting.
92 Chapter 3 Elementary Descriptive Statistics
Section 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. The following are data (from Introduction to Con- (a) Make quantile plots for these two samples.
temporary Statistical Methods by L. H. Koopmans) Find the medians, the quartiles, and the .37
on the impact strength of sheets of insulating ma- quantiles for the two data sets.
terial cut in two different ways. (The values are in (b) Draw (to scale) carefully labeled side-by-side
ft lb.) boxplots for comparing the two cutting meth-
ods. Discuss what these show about the two
Lengthwise Cuts Crosswise Cuts methods.
(c) Make and discuss the appearance of a Q-Q plot
1.15 .89 for comparing the shapes of these two data sets.
.84 .69 2. Make a Q-Q plot for the two small samples in
.88 .46 Table 3.13 in Section 3.3.
.91 .85 3. Make and interpret a normal plot for the yield data
.86 .73 of Exercise 1 of Section 3.1.
.88 .67
4. Explain the usefulness of theoretical Q-Q plotting.
.92 .78
.87 .77
.93 .80
.95 .79
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
The word average, as used in colloquial speech, has several potential technical
meanings. One is the median, Q(.5), which was introduced in the last section. The
median divides a data set in half. Roughly half of the area enclosed by the bars of a
well-made histogram will lie to either side of the median. As a measure of center,
it is completely insensitive to the effects of a few extreme or outlying observations.
For example, the small set of data
2, 3, 6, 9, 10
has median 6, and this remains true even if the value 10 is replaced by 10,000,000
and/or the value 2 is replaced by −200,000.
The previous section used the median as a center value in the making of boxplots.
But the median is not the technical meaning most often attached to the notion of
average in statistical analyses. Instead, it is more common to employ the (arithmetic)
mean.
1X
n
x̄ = x
n i=1 i
The mean is sometimes called the first moment or center of mass of a distribution,
drawing on an analogy to mechanics. Think of placing a unit mass along the number
line at the location of each value in a data set—the balance point of the mass
distribution is at x̄.
and
1
I x̄ = (.37 + .52 + .65 + .92 + 2.89 + 3.62) = 1.495% waste
6
94 Chapter 3 Elementary Descriptive Statistics
Supplier 1 Supplier 2
.37, .52, .65, .89, .99, 1.45, 1.47,
.92, 2.89, 3.62 1.58, 2.27, 2.63, 6.54
and
1
x̄ = (.89 + .99 + 1.45 + 1.47 + 1.58 + 2.27 + 2.63 + 6.54)
8
I = 2.228% waste
Figure 3.20 shows dot diagrams with the medians and means marked. Notice
that a comparison of either medians or means for the two suppliers shows the
supplier 2 waste to be larger than the supplier 1 waste. But there is a substan-
tial difference between the median and mean values for a given supplier. In
both cases, the mean is quite a bit larger than the corresponding median. This
reflects the right-skewed nature of both data sets. In both cases, the center of
mass of the distribution is pulled strongly to the right by a few extremely large
values.
Supplier 1
Q(.5) = .785
0 1 2 3 4 5 6
x = 1.495
Waste (percent)
Supplier 2
Q(.5) = 1.525
0 1 2 3 4 5 6
x = 2.228
Waste (percent)
Example 7 shows clearly that, in contrast to the median, the mean is a mea-
sure of center that can be strongly affected by a few extreme data values. People
sometimes say that because of this, one or the other of the two measures is “better.”
Such statements lack sense. Neither is better; they are simply measures with dif-
ferent properties. And the difference is one that intelligent consumers of statistical
information do well to keep in mind. The “average” income of employees at a com-
pany paying nine workers each $10,000/year and a president $110,000/year can be
described as $10,000/year or $20,000/year, depending upon whether the median or
mean is being used.
R = xn − x1
Notice the word usage here. The word range could be used as a verb to say, “The
data range from 3 to 21.” But to use the word as a noun, one says, “The range is
(21 − 3) = 18.” Since the range depends only on the values of the smallest and
largest points in a data set, it is necessarily highly sensitive to extreme (or outlying)
values. Because it is easily calculated, it has enjoyed long-standing popularity in
industrial settings, particularly as a tool in statistical quality control.
However, most methods of formal statistical inference are based on another mea-
sure of distributional spread. A notion of “mean squared deviation” or “root mean
squared deviation” is employed to produce measures that are called the variance
and the standard deviation, respectively.
96 Chapter 3 Elementary Descriptive Statistics
1 X
n
s2 = (x − x̄)2
n − 1 i=1 i
Example 7 The spreads in the two sets of percentage wastes recorded in Table 3.13 can be
(continued ) expressed in any of the preceding terms. For the supplier 1 data,
Q(.25) = .52
Q(.75) = 2.89
and so
Also,
Further,
1
s2 = ((.37 − 1.495)2 + (.52 − 1.495)2 + (.65 − 1.495)2 + (.92 − 1.495)2
6−1
+ (2.89 − 1.495)2 + (3.62 − 1.495)2 )
= 1.945(% waste)2
so that
√
I s= 1.945 = 1.394% waste
3.3 Standard Numerical Summary Measures 97
and
Further,
1
s2 = ((.89 − 2.228)2 + (.99 − 2.228)2 + (1.45 − 2.228)2 + (1.47 − 2.228)2
8−1
+ (1.58 − 2.228)2 + (2.27 − 2.228)2 + (2.63 − 2.228)2 + (6.54 − 2.228)2 )
= 3.383(% waste)2
so
I s = 1.839% waste
Supplier 2 has the smaller IQR but the larger R and s. This is consistent with
Figure 3.20. The central portion of the supplier 2 distribution is tightly packed.
But the single extreme data point makes the overall variability larger for the
second supplier than for the first.
Proposition 1 For any data set and any number k larger than 1, a fraction of at least 1 − (1/k 2 )
(Chebyschev’s Theorem ) of the data are within ks of x̄.
98 Chapter 3 Elementary Descriptive Statistics
This little theorem says, for example, that at least 34 of a data set is within 2 standard
deviations of its mean. And at least 89 of a data set is within 3 standard deviations of
its mean. So the theorem promises that if a data set has a small standard deviation,
it will be tightly packed about its mean.
Example 7 Returning to the waste data, consider illustrating the meaning of Chebyschev’s
(continued ) theorem with the supplier 1 values. For example, taking k = 2, at least 34 =
1 − ( 12 )2 of the 6 data points (i.e., at least 4.5 of them) must be within 2 standard
deviations of x̄. In fact
and
so simple counting shows that all (a fraction of 1.0) of the data are between these
two values.
Definition 10 Numerical summarizations of sample data are called (sample) statistics. Nu-
merical summarizations of population and theoretical distributions are called
(population or model) parameters. Typically, Roman letters are used as sym-
bols for statistics, and Greek letters are used to stand for parameters.
1 X
N
Population µ= x (3.4)
mean N i=1 i
Comparing this expression to the one in Definition 7, not only is a different symbol
used for the mean but also N is used in place of n. It is standard to denote a
population size as N and a sample size as n. Chapter 5 gives a definition for the
3.3 Standard Numerical Summary Measures 99
mean of a theoretical distribution. But it is worth saying now that the symbol µ will
be used in that context as well as in the context of equation (3.4).
As another example of the usage suggested by Definition 10, consider the vari-
ance and standard deviation. Definition 9 refers specifically to the sample variance
and standard deviation. If a data set represents an entire population, then it is com-
mon to use the lowercase Greek sigma squared (σ 2 ) to stand for the population
variance and to define
1 X
Population N
variance σ2 = (x − µ)2 (3.5)
N i=1 i
The nonnegative square root of σ 2 is then called the population standard devia-
tion, σ . (The division in equation (3.5) is by N , and not the N − 1 that might be
expected on the basis of Definition 9. There are reasons for this change, but they are
not accessible at this point.) Chapter 5 defines a variance and standard deviation for
theoretical distributions, and the symbols σ 2 and σ will be used there as well as in
the context of equation (3.5).
On one point, this text will deviate from the Roman/Greek symbolism conven-
tion laid out in Definition 10: the notation for quantiles. Q( p) will stand for the pth
quantile of a distribution, whether it is from a sample, a population, or a theoretical
model.
Table 3.14
Means and Ranges for a Critical Dimension on Samples of n = 5 Parts
Example 8
(continued )
x
3515
3510
3505
3500
5 10 15 20 25
Sample number
R
5
0
5 10 15 20 25
Sample number
shift were not really systematically any different from the others. Instead, the
person making the measurements for samples 9 through 15 used the gauge in a
fundamentally different way than other employees. The pattern in the x̄ values
was caused by this change in measurement technique.
Terminology and Patterns revealed in the plotting of sample statistics against time ought to alert
causes for patterns an engineer to look for a physical cause and (typically) a cure. Systematic vari-
on plots against ations or cycles in a plot of means can often be related to process variables that
Time come and go on a more or less regular basis. Examples include seasonal or daily
variables like ambient temperature or those caused by rotation of gauges or fixtures.
Instability or variation in excess of that related to basic equipment precision can
sometimes be traced to mixed lots of raw material or overadjustment of equipment
by operators. Changes in level of a process mean can originate in the introduction
of new machinery, raw materials, or employee training and (for example) tool wear.
Mixtures of several patterns of variation on a single plot of some summary statistic
against time can sometimes (as in Example 8) be traced to changes in measurement
calibration. They are also sometimes produced by consistent differences in machines
or streams of raw material.
Plots against Plots of summary statistics against time are not the only useful ones. Plots
process variables against process variables can also be quite informative.
Table 3.15
Mean Joint Strengths for Nine Wood/Glue Combinations
x̄
Wood Glue Mean Joint Shear Strength (lb)
pine white 131.7
pine carpenter’s 192.7
pine cascamite 201.3
fir white 92.0
fir carpenter’s 146.3
fir cascamite 156.7
oak white 257.7
oak carpenter’s 234.3
oak cascamite 177.7
102 Chapter 3 Elementary Descriptive Statistics
Example 9
(continued )
250 Oak
200
Strength (lb)
Pine
150
Fir
100
From the plot, it is obvious that the gluing properties of pine and fir are
quite similar, with pine joints averaging around 40–45 lb stronger. For these
two soft woods, cascamite appears slightly better than carpenter’s glue, both of
which make much better joints than white glue. The gluing properties of oak
(a hardwood) are quite different from those of pine and fir. In fact, the glues
perform in exactly the opposite ordering for the strength of oak joints. All of this
is displayed quite clearly by the simple plot in Figure 3.22.
The two previous examples have illustrated the usefulness of plotting sample
statistics against time and against levels of an experimental variable. Other possi-
bilities in specific engineering situations can potentially help the working engineer
understand and manipulate the systems on which he or she works.
figures on the printout do not match exactly those found earlier. MINITAB simply
uses slightly different conventions for those quantities than the ones introduced in
Section 3.2.
High-quality statistical packages like MINITAB (and JMP, SAS, SPSS, SYS-
TAT, SPLUS, etc.) are widely available. One of them should be on the electronic
desktop of every working engineer. Unfortunately, this is not always the case, and
engineers often assume that standard spreadsheet software (perhaps augmented with
third party plug-ins) provides a workable substitute. Often this is true, but sometimes
it is not.
The primary potential problem with using a spreadsheet as a substitute for sta-
tistical software concerns numerical accuracy. Spreadsheets can and do on occasion
return catastrophically wrong values for even simple statistics. Established vendors
of statistical software have many years of experience dealing with subtle numerical
issues that arise in the computation of even simple summaries of even small data
sets. Most vendors of spreadsheet software seem unaware of or indifferent to these
matters. For example, consider the very small data set
0, 1, 2
The sample variance of these data is easily seen to be 1.0, and essentially any
statistical package or spreadsheet will reliably return this value. However, suppose
100,000,000 is added to each of these n = 3 values, producing the data set
The actual sample variance is unchanged, and high-quality statistical software will
reliably return the value 1.0. However, as of late 1999, the current version of the
leading spreadsheet program returned the value 0 for this second sample variance.
This is a badly wrong answer to an apparently very simple problem.
So at least until vendors of spreadsheet software choose to integrate an es-
tablished statistical package into their products, we advise extreme caution in the
use of spreadsheets to do statistical computations. A good source of up-to-date
information on this issue is the AP Statistics electronic bulletin board found at
http://forum.swarthmore.edu/epigone/apstat-l.
104 Chapter 3 Elementary Descriptive Statistics
Section 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Calculate and compare the means, medians, ranges, Exercise 1 of Section 3.2 and thereby check your
interquartile ranges, and standard deviations of the answers to Exercise 1 here.
two data sets introduced in Exercise 1 of Section 4. Add 1.3 to each of the lengthwise cut impact
3.2. Discuss the interpretation of these values in the strengths referred to in Exercise 1 and then re-
context of comparing the two cutting methods. compute the values of the mean, median, range,
2. Are the numerical values you produced in Exercise interquartile range, and standard deviation. How
1 above most naturally thought of as statistics or as do these compare with the values obtained earlier?
parameters? Explain. Repeat this exercise after multiplying each length-
3. Use a statistical package to compute basic sum- wise cut impact strength by 2 (instead of adding
mary statistics for the two data sets introduced in 1.3).
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Sample fraction
The number of items in the sample with the characteristic
of items with a p̂ = (3.6)
characteristic n
will be used. A given sample can produce many such values of “ p hat” if either a
single characteristic has many possible categories or many different characteristics
are being monitored simultaneously.
3.4 Descriptive Statistics for Qualitative and Count Data 105
Table 3.16 gives counts of sampled connectors falling into the first four
categories (the four defect categories) over the 30-day period. Then, using the
fact that 30 × 100 = 3,000 connectors were inspected over this period,
p̂ A = 3/3000 = .0010
p̂ B = 0/3000 = .0000
p̂ C = 11/3000 = .0037
p̂ D = 1/3000 = .0003
Table 3.16
Counts of Connectors Classified into Four Defect
Categories
Table 3.17
Counts and Fractions of Tools with Various
Problems
Sample mean
The total number of occurrences
occurences per û = (3.7)
unit or item The total number of inspection units or sampled items
3.4 Descriptive Statistics for Qualitative and Count Data 107
is used. û is really closer in meaning to x̄ than to p̂, even though it can turn out to be
a number between 0 and 1 and is sometimes expressed as a percentage and called a
rate.
Although the counts totaled in the numerator of expression (3.7) must all be
integers, the values totaled to create the denominator need not be. For instance,
suppose vinyl floor tiles are being inspected for serious blemishes. If on one occasion
inspection of 1 box yields a total of 2 blemishes, on another occasion .5 box yields
0 blemishes, and on still another occasion 2.5 boxes yield a total of 1 blemish, then
2+0+1
û = = .75 blemishes/box
1 + .5 + 2.5
Example 10 It was possible for a single cable connector to have more than one defect of a
(continued ) given severity and, in fact, defects of different severities. For example, Delva,
Lynch, and Stephany’s records indicate that in the 3,000 connectors inspected,
1 connector had exactly 2 moderately serious defects (along with a single very
serious defect), 11 connectors had exactly 1 moderately serious defect (and no
others), and 2,988 had no moderately serious defects. So the observed rate of
moderately serious defects could be reported as
2 + 11
û = = .0043 moderately serious defects/connector
1 + 11 + 2988
This is an occurrence rate for moderately serious defects( û), but not a fraction
of connectors having moderately serious defects ( p̂).
The difference between the statistics p̂ and û may seem trivial. But it is a point
that constantly causes students confusion. Methods of formal statistical inference
based on p̂ are not the same as those based on û. The distinction between the two
kinds of rates must be kept in mind if those methods are to be applied appropriately.
To carry this warning a step further, note that not every quantity called a
percentage is even of the form p̂ or û. In a laboratory analysis, a specimen may be
declared to be “30% carbon.” The 30% cannot be thought of as having the form of p̂
in equation (3.6) or û in equation (3.7). It is really a single continuous measurement,
not a summary statistic. Statistical methods for p̂ or û have nothing to say about
such rates.
3.4.2 Bar Charts and Plots for Qualitative and Count Data
Often, a study will produce several values of p̂ or û that need to be compared. Bar
charts and simple bivariate plots can be a great aid in summarizing these results.
108 Chapter 3 Elementary Descriptive Statistics
Example 10 Figure 3.23 is a bar chart of the fractions of connectors in the categories A through
(continued ) D. It shows clearly that most connectors with defects fall into category C, having
moderately serious defects but no serious or very serious defects. This bar chart
is a presentation of the behavior of a single categorical variable.
Fraction of connectors
.004
.003
.002
.001
0
A B C D
Connector category
Example 11 Figure 3.24 is a bar chart of the information on tool problems in Table 3.17. It
(continued ) shows leaks to be the most frequently occurring problems on this production run.
.08
.07
.06
Problem rate
.05
.04
.03
.02
.01
0
Type 1 leak
Type 2 leak
Type 3 leak
Missing part 1
Missing part 2
Missing part 3
Bad part 4
Bad part 5
Bad part 6
Wrong part 7
Wrong part 8
Figures 3.23 and 3.24 are both bar charts, but they differ considerably. The
first concerns the behavior of a single (ordered) categorical variable—namely, Con-
nector Class. The second concerns the behavior of 11 different present–not present
categorical variables, like Type 1 Leak, Missing Part 3, etc. There may be some
significance to the shape of Figure 3.23, since categories A through D are arranged
in decreasing order of defect severity, and this order was used in the making of
the figure. But the shape of Figure 3.24 is essentially arbitrary, since the particular
ordering of the tool problem categories used to make the figure is arbitrary. Other
equally sensible orderings would give quite different shapes.
The device of segmenting bars on a bar chart and letting the segments stand
for different categories of a single qualitative variable can be helpful, particularly
where several different samples are to be compared.
Table 3.18
Percents Scrap and Rework in a Turning Operation
Job Number Percent Scrap Percent Rework Job Number Percent Scrap Percent Rework
1 2 25 10 3 18
2 3 11 11 0 3
3 0 5 12 1 5
4 0 0 13 0 0
5 0 20 14 0 0
6 2 23 15 0 3
7 0 6 16 0 2
8 0 5 17 0 2
9 2 8 18 1 5
110 Chapter 3 Elementary Descriptive Statistics
Example 12 Scrap
(continued ) Rework
25
Percent of production
20
15
10
1 5 10 15
Job number
Table 3.19
Defects Per Truck on 26 Production Days
2.5
Defects / Truck
2.0
1.5
1.0
.5
12/2
12/6
12/9
12/13
12/16
12/20
12/23
12/24
12/26
12/27
12/30
1/3
1/6
1/10
Date
.4
20% Reground powder
.2
Small Large
Shot size
Section 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. From your field, give an example of a variable that represented in Table 1.1, what would have been the
is a rate (a) of the form p̂, (b) of the form û, and sample fractions nonconforming p̂? Give a practi-
(c) of neither form. cal reason why having the values in Table 1.1 might
2. Because gauging is easier, it is sometimes tempting be preferable to knowing only the corresponding p̂
to collect qualitative data related to measurements values.
rather than the measurements themselves. For ex- 3. Consider the measurement of the percentage cop-
ample, in the context of Example 1 in Chapter 1, if per in brass specimens. The resulting data will be a
gears with runouts exceeding 15 were considered kind of rate data. Are the rates that will be obtained
to be nonconforming, it would be possible to derive of the type p̂, of the type û, or of neither type?
fractions nonconforming, p̂, from simple “go–no Explain.
go” checking of gears. For the two sets of gears
Chapter 3 Exercises 113
Chapter 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. The accompanying values are gains measured on which later can clog the filters in extruders when
120 amplifiers designed to produce a 10 dB gain. the recycled material is used. The following are the
These data were originally from the Quality Im- amounts (in ppm by weight of aluminum) found
provement Tools workbook set (published by the in bihourly samples of PET recovered at the plant
Juran Institute). They were then used as an exam- over roughly a two-day period.
ple in the article “The Tools of Quality” (Quality
Progress, September 1990). 291, 222, 125, 79, 145, 119, 244, 118, 182, 63,
30, 140, 101, 102, 87, 183, 60, 191, 119, 511,
8.1, 10.4, 8.8, 9.7, 7.8, 9.9, 11.7, 8.0, 9.3, 9.0, 8.2, 120, 172, 70, 30, 90, 115
8.9, 10.1, 9.4, 9.2, 7.9, 9.5, 10.9, 7.8, 8.3, 9.1, 8.4,
9.6, 11.1, 7.9, 8.5, 8.7, 7.8, 10.5, 8.5, 11.5, 8.0, 7.9, (Apparently, the data are recorded in the order in
8.3, 8.7, 10.0, 9.4, 9.0, 9.2, 10.7, 9.3, 9.7, 8.7, 8.2, which they were collected, reading left to right, top
8.9, 8.6, 9.5, 9.4, 8.8, 8.3, 8.4, 9.1, 10.1, 7.8, 8.1, to bottom.)
8.8, 8.0, 9.2, 8.4, 7.8, 7.9, 8.5, 9.2, 8.7, 10.2, 7.9, (a) Make a run chart for these data. Are there any
9.8, 8.3, 9.0, 9.6, 9.9, 10.6, 8.6, 9.4, 8.8, 8.2, 10.5, obvious time trends? What practical engineer-
9.7, 9.1, 8.0, 8.7, 9.8, 8.5, 8.9, 9.1, 8.4, 8.1, 9.5, ing reason is there for looking for such trends?
8.7, 9.3, 8.1, 10.1, 9.6, 8.3, 8.0, 9.8, 9.0, 8.9, 8.1, (b) Ignoring the time order information, make a
9.7, 8.5, 8.2, 9.0, 10.2, 9.5, 8.3, 8.9, 9.1, 10.3, 8.4, stem-and-leaf diagram. Use the hundreds digit
8.6, 9.2, 8.5, 9.6, 9.0, 10.7, 8.6, 10.0, 8.8, 8.6 to make the stem and the other two digits (sep-
arated by commas to indicate the different data
(a) Make a stem-and-leaf plot and a boxplot for points) to make the leaves. After making an
these data. How would you describe the shape initial stem-and-leaf diagram by recording the
of this data set? Does the shape of your stem- data in the (time) order given above, make a
and-leaf plot (or a corresponding histogram) second one in which the values have been or-
give you any clue how a high fraction within dered.
specifications was achieved? (c) How would you describe the shape of the stem-
(b) Make a normal plot for these data and interpret and-leaf diagram? Is the data set bell-shaped?
its shape. (Standard normal quantiles for p = (d) Find the median and the first and third quartiles
.0042 and p = .9958 are approximately −2.64 for the aluminum contents and then find the .58
and 2.64, respectively.) quantile of the data set.
(c) Although the nominal gain for these amplifiers (e) Make a boxplot.
was to be 10 dB, the design allowed gains from (f) Make a normal plot, using regular graph paper.
7.75 dB to 12.2 dB to be considered acceptable. List the coordinates of the 26 plotted points.
About what fraction, p, of such amplifiers do Interpret the shape of the plot.
you expect to meet these engineering specifi- (g) Try transforming the data by taking natural log-
cations? arithms and again assess the shape. Is the trans-
2. The article “The Lognormal Distribution for Mod- formed data set more bell-shaped than the raw
eling Quality Data When the Mean is Near Zero” data set?
by S. Albin (Journal of Quality Technology, April (h) Find the sample mean, the sample range, and
1990) described the operation of a Rutgers Uni- the sample standard deviation for both the
versity plastics recycling pilot plant. The most im- original data and the log-transformed values
portant material reclaimed from beverage bottles from (g). Is the mean of the transformed val-
is PET plastic. A serious impurity is aluminum, ues equal to the natural logarithm of the mean
of the original data?
114 Chapter 3 Elementary Descriptive Statistics
3. The accompanying data are three hypothetical sam- 4. Gaul, Phan, and Shimonek measured the resis-
ples of size 10 that are supposed to represent mea- tances of 15 resistors of 2 × 5 = 10 different types.
sured manganese contents in specimens of 1045 Two different wattage ratings were involved, and
steel (the units are points, or .01%). Suppose that five different nominal resistances were used. All
these measurements were made on standard speci- measurements were reported to three significant
mens having “true” manganese contents of 80, us- digits. Their data follow.
ing three different analytical methods. (Thirty dif- (a) Make back-to-back stem-and-leaf plots for
ferent specimens were involved.) comparing the 14 watt and 12 watt resistance
distributions for each nominal resistance. In a
Method 1 few sentences, summarize what these show.
(b) Make pairs of boxplots for comparing the 14
87, 74, 78, 81, 78, watt and 12 watt resistance distributions for each
77, 84, 80, 85, 78 nominal resistance.
(c) Make normal plots for the 12 watt nominal 20
ohm and nominal 200 ohm resistors. Interpret
Method 2
these in a sentence or two. From the appear-
86, 85, 82, 87, 85, ance of the second plot, does it seem that if
84, 84, 82, 82, 85 the nominal 200 ohm resistances were treated
as if they had a bell-shaped distribution, the
tendency would be to overestimate or to un-
Method 3 derestimate the fraction of resistances near the
nominal value?
84, 83, 78, 79, 85,
82, 82, 81, 82, 79
1
4
Watt Resistors
(a) Make (on the same coordinate system) side- 20 ohm 75 ohm 100 ohm 150 ohm 200 ohm
by-side boxplots that you can use to compare
the three analytical methods. 19.2 72.9 97.4 148 198
(b) Discuss the apparent effectiveness of the three 19.2 72.4 95.8 148 196
methods in terms of the appearance of your di- 19.3 72.0 97.7 148 199
agram from (a) and in terms of the concepts 19.3 72.5 94.1 148 196
of accuracy and precision discussed in Sec- 19.1 72.7 95.1 148 196
tion 1.3. 19.0 72.3 95.4 147 195
(c) An alternative method of comparing two such
19.6 72.9 94.9 148 193
analytical methods is to use both methods of
19.2 73.2 98.5 148 196
analysis once on each of (say) 10 different
19.3 71.8 94.8 148 196
specimens (10 specimens and 20 measure-
ments). In the terminology of Section 1.2, what 19.4 73.4 94.6 147 199
kind of data would be generated by such a 19.4 70.9 98.3 147 194
plan? If one simply wishes to compare the 19.3 72.3 96.0 149 195
average measurements produced by two ana- 19.5 72.5 97.3 148 196
lytical methods, which data collection plan (20 19.2 72.1 96.0 148 195
specimens and 20 measurements, or 10 spec- 19.1 72.6 94.8 148 199
imens and 20 measurements) seems to you
most likely to provide the better comparison?
Explain.
Chapter 3 Exercises 115
1 5-Gram Weighings
2
Watt Resistors
20 ohm 75 ohm 100 ohm 150 ohm 200 ohm Scale 1 Scale 2 Scale 3
20.1 73.9 97.2 152 207 Student 1 5.03, 5.02 5.07, 5.09 4.98, 4.98
19.7 74.2 97.9 151 205 Student 2 5.03, 5.01 5.02, 5.07 4.99, 4.98
20.2 74.6 96.8 155 214 Student 3 5.06, 5.00 5.10, 5.08 4.98, 4.98
24.4 72.1 99.2 146 195
20.2 73.8 98.5 148 202 20-Gram Weighings
20.1 74.8 95.5 154 211
Scale 1 Scale 2 Scale 3
20.0 75.0 97.2 149 197
20.4 68.6 98.7 150 197 Student 1 20.04, 20.06 20.04, 20.04 19.94, 19.93
20.3 74.0 96.6 153 199 Student 2 20.02, 19.99 20.03, 19.93 19.95, 19.95
20.6 71.7 102 149 196 Student 3 20.03, 20.02 20.06, 20.03 19.91, 19.96
19.9 76.5 103 150 207
100-Gram Weighings
19.7 76.2 102 149 210
20.8 72.8 102 145 192 Scale 1 Scale 2 Scale 3
20.4 73.2 100 147 201
20.5 76.7 100 149 257 Student 1 100.06, 100.35 100.25, 100.08 99.87, 99.88
Student 2 100.05, 100.01 100.10, 100.02 99.87, 99.88
(d) Compute the sample means and sample stan- Student 3 100.00, 100.00 100.01, 100.02 99.88, 99.88
dard deviations for all 10 samples. Do these
values agree with your qualitative statements 6. The accompanying values are the lifetimes (in num-
made in answer to part (a)? bers of 24 mm deep holes drilled in 1045 steel
(e) Make a plot of the 10 sample means computed before tool failure) for n = 12 D952-II (8 mm)
in part (d), similar to the plot in Figure 3.22. drills. These were read from a graph in “Computer-
Comment on the appearance of this plot. assisted Prediction of Drill-failure Using In-process
5. Blomquist, Kennedy, and Reiter studied the prop- Measurements of Thrust Force” by A. Thangaraj
erties of three scales by each weighing a standard and P. K. Wright (Journal of Engineering for In-
5 g weight, 20 g weight, and 100 g weight twice dustry, May 1988).
on each scale. Their data are presented in the ac-
companying table. Using whatever graphical and 47, 145, 172, 86, 122, 110, 172, 52, 194, 116,
numerical data summary methods you find helpful, 149, 48
make sense of these data. Write a several-page dis- Write a short report to your engineering manager
cussion of your findings. You will probably want summarizing what these data indicate about the
to consider both accuracy and precision and (to the lifetimes of drills of this type in this kind of appli-
extent possible) make comparisons between scales cation. Use whatever graphical and numerical data
and between students. Part of your discussion might summary tools make clear the main features of the
deal with the concepts of repeatability and repro- data set.
ducibility introduced in Section 2.1. Are the pic-
7. Losen, Cahoy, and Lewis purchased eight spanner
tures you get of the scale and student performances
bushings of a particular type from a local machine
consistent across the different weights?
shop and measured a number of characteristics of
these bushings, including their outside diameters.
Each of the eight outside diameters was measured
116 Chapter 3 Elementary Descriptive Statistics
once by each of two student technicians, with the (a) Find the .84 quantile of the Compound 1 failure
following results (the units are inches): times.
(b) Give the coordinates of the two lower-left
Bushing 1 2 3 4 points that would appear on a normal plot of
Student A .3690 .3690 .3690 .3700 the Compound 1 data.
(c) Make back-to-back stem-and-leaf plots for
Student B .3690 .3695 .3695 .3695
comparing the life length properties of bear-
Bushing 5 6 7 8 ings made from Compounds 1 and 2.
Student A .3695 .3700 .3695 .3690 (d) Make (to scale) side-by-side boxplots for com-
Student B .3695 .3700 .3700 .3690 paring the life lengths for the two compounds.
Mark numbers on the plots indicating the loca-
A common device when dealing with paired data tions of their main features.
like these is to analyze the differences. Subtracting (e) Compute the sample means and standard devi-
B measurements from A measurements gives the ations of the two sets of lifetimes.
following eight values: (f) Describe what your answers to parts (c), (d),
and (e) above indicate about the life lengths of
.0000, −.0005, −.0005, .0005, .0000, .0000, these turbine bearings.
−.0005, .0000
9. Heyde, Kuebrick, and Swanson measured the
(a) Find the first and third quartiles for these dif- heights of 405 steel punches purchased by a com-
ferences, and their median. pany from a single supplier. The stamping machine
(b) Find the sample mean and standard deviation in which these are used is designed to use .500 in.
for the differences. punches. Frequencies of the measurements they
(c) Your mean in part (b) should be negative. Inter- obtained are shown in the accompanying table.
pret this in terms of the original measurement
problem. Punch Height Punch Height
(d) Suppose you want to make a normal plot of the (.001 in.) Frequency (.001 in.) Frequency
differences on regular graph paper. Give the co-
ordinates of the lower-left point on such a plot. 482 1 496 7
8. The accompanying data are the times to failure (in 483 0 497 13
millions of cycles) of high-speed turbine engine 484 1 498 24
bearings made out of two different compounds. 485 1 499 56
These were taken from “Analysis of Single Classi- 486 0 500 82
fication Experiments Based on Censored Samples 487 1 501 97
from the Two-parameter Weibull Distribution” by 488 0 502 64
J. I. McCool (The Journal of Statistical Planning 489 1 503 43
and Inference, 1979). 490 0 504 3
491 2 505 1
Compound 1 492 0 506 0
3.03, 5.53, 5.60, 9.30, 9.92, 493 0 507 0
12.51, 12.95, 15.21, 16.04, 16.84 494 0 508 0
495 6 509 2
Compound 2
(a) Summarize these data, using appropriate purity. Describe the shape of the purity distri-
graphical and numerical tools. How would bution.
you describe the shape of the distribution of (c) The author of the article found it useful to
punch heights? The specifications for punch reexpress the purities by subtracting 99.30
heights were in fact .500 in. to .505 in. Does (remember that the preceding values are in
this fact give you any insight as to the ori- units of .01% above 99.00%) and then tak-
gin of the distributional shape observed in ing natural logarithms. Do this with the raw
the data? Does it appear that the supplier has data and make a second stem-and-leaf dia-
equipment capable of meeting the engineer- gram and a second histogram to portray the
ing specifications on punch height? shape of the transformed data. Do these fig-
(b) In the manufacturing application of these ures look more bell-shaped than the ones you
punches, several had to be placed side-by-side made in part (b)?
on a drum to cut the same piece of material. In (d) Make a normal plot for the transformed values
this context, why is having small variability from part (c). What does it indicate about the
in punch height perhaps even more important shape of the distribution of the transformed
than having the correct mean punch height? values? (Standard normal quantiles for p =
10. The article “Watch Out for Nonnormal Distri- .005 and p = .995 are approximately −2.58
butions” by D. C. Jacobs (Chemical Engineer- and 2.58, respectively.)
ing Progress, November 1990) contains 100 mea- 11. The following are some data taken from the article
sured daily purities of oxygen delivered by a sin- “Confidence Limits for Weibull Regression with
gle supplier. These are as follows, listed in the time Censored Data” by J. I. McCool (IEEE Transac-
order of their collection (read left to right, top to tions on Reliability, 1980). They are the ordered
bottom). The values given are in hundredths of failure times (the time units are not given in the
a percent purity above 99.00% (so 63 stands for paper) for hardened steel specimens subjected to
99.63%). rolling contact fatigue tests at four different values
of contact stress.
63, 61, 67, 58, 55, 50, 55, 56, 52, 64, 73, 57, 63,
81, 64, 54, 57, 59, 60, 68, 58, 57, 67, 56, 66, 60,
.87 × 106 .99 × 106 1.09 × 106 1.18 × 106
49, 79, 60, 62, 60, 49, 62, 56, 69, 75, 52, 56, 61,
58, 66, 67, 56, 55, 66, 55, 69, 60, 69, 70, 65, 56, psi psi psi psi
73, 65, 68, 59, 62, 58, 62, 66, 57, 60, 66, 54, 64, 1.67 .80 .012 .073
62, 64, 64, 50, 50, 72, 85, 68, 58, 68, 80, 60, 60, 2.20 1.00 .18 .098
53, 49, 55, 80, 64, 59, 53, 73, 55, 54, 60, 60, 58, 2.51 1.37 .20 .117
50, 53, 48, 78, 72, 51, 60, 49, 67
3.00 2.25 .24 .135
You will probably want to use a statistical analysis 3.90 2.95 .26 .175
package to help you do the following: 4.70 3.70 .32 .262
(a) Make a run chart for these data. Are there any 7.53 6.07 .32 .270
obvious time trends? What would be the prac- 14.7 6.65 .42 .350
tical engineering usefulness of early detection 27.8 7.05 .44 .386
of any such time trend? 37.4 7.37 .88 .456
(b) Now ignore the time order of data collection
and represent these data with a stem-and-leaf (a) Make side-by-side boxplots for these data.
plot and a histogram. (Use .02% class widths Does it look as if the different stress levels
in making your histogram.) Mark on these the produce life distributions of roughly the same
supplier’s lower specification limit of 99.50% shape? (Engineering experience suggests that
118 Chapter 3 Elementary Descriptive Statistics
different stress levels often change the scale by R. Rossi (Solid State Technology, 1984). (The
but not the basic shape of life distributions.) units were not given in the article.)
(b) Make Q-Q plots for comparing all six dif-
ferent possible pairs of distributional shapes. 5.55, 5.52, 5.45, 5.53, 5.37, 5.22, 5.62, 5.69,
Summarize in a few sentences what these in- 5.60, 5.58, 5.51, 5.53
dicate about the shapes of the failure time
distributions under the different stress levels. (a) Make a dot diagram and a boxplot for these
12. Riddle, Peterson, and Harper studied the perfor- data and compute the statistics x̄ and s.
mance of a rapid-cut industrial shear in a continu- (b) Make a normal plot for these data. How bell-
ous cut mode. They cut nominally 2-in. and 1-in. shaped does this data set look? If you were to
strips of 14 gauge and 16 gauge steel sheet metal say that the shape departs from a perfect bell
and measured the actual widths of the strips pro- shape, in what specific way does it? (Refer to
duced by the shear. Their data follow, in units of characteristics of the normal plot to support
10−3 in. above nominal. your answer.)
14. The article “Thermal Endurance of Polyester
Material Thickness Enameled Wires Using Twisted Wire Specimens”
by H. Goldenberg (IEEE Transactions on Electri-
14 Gauge 16 Gauge cal Insulation, 1965) contains some data on the
lifetimes (in weeks) of wire specimens tested for
2, 1, 1, 1, −2, −6, −1, −2,
thermal endurance according to AIEE Standard
1 in. 0, 0, −2, −1, −2, −1, 57. Several different laboratories were used to
−10, −5, 1 −1, −1, −5 make the tests, and the results from two of the
Machine Setting laboratories, using a test temperature of 200◦ C,
10, 10, 8, 8, −4, −3, −4, −2, follow:
2 in. 8, 8, 7, −3, −3, −3,
7, 9, 11 −3, −4, −4 Laboratory 1 Laboratory 2
(a) Compute sample means and standard devia- 14, 16, 17, 18, 20, 27, 28, 29, 29, 29,
tions for the four samples. Plot the means in 22, 23, 25, 27, 28 30, 31, 31, 33, 34
a manner similar to the plot in Figure 3.22.
Make a separate plot of this kind for the stan- Consider first only the Laboratory 1 data.
dard deviations. (a) Find the median and the first and third quar-
(b) Write a short report to an engineering man- tiles for the lifetimes and then find the .64
ager to summarize what these data and your quantile of the data set.
summary statistics and plots show about the (b) Make and interpret a normal plot for these
performance of the industrial shear. How do data. Would you describe this distribution as
you recommend that the shear be set up in bell-shaped? If not, in what way(s) does it
the future in order to get strips cut from these depart from being bell-shaped? Give the co-
materials with widths as close as possible to ordinates of the 10 points you plot on regular
specified dimensions? graph paper.
(c) Find the sample mean, the sample range, and
13. The accompanying data are some measured resis-
the sample standard deviation for these data.
tivity values from in situ doped polysilicon spec-
Now consider comparing the work of the two dif-
imens taken from the article “LPCVD Process
ferent laboratories (i.e., consider both data sets).
Equipment Evaluation Using Statistical Methods”
Chapter 3 Exercises 119
(d) Make back-to-back stem-and-leaf plots for 16. The accompanying values are representative of
these two data sets (use two leaves for obser- data summarized in a histogram appearing in
vations 10–19, two for observations 20–29, the article “Influence of Final Recrystallization
etc.) Heat Treatment on Zircaloy-4 Strip Corrosion”
(e) Make side-by-side boxplots for these two data by Foster, Dougherty, Burke, Bates, and Worces-
sets. (Draw these on the same scale.) ter (Journal of Nuclear Materials, 1990). Given
(f) Based on your work in parts (d) and (e), which are n = 20 particle diameters observed in a bright-
of the two labs would you say produced the field TEM micrograph of a Zircaloy-4 specimen.
more precise results? The units are 10−2 µm.
(g) Is it possible to tell from your plots in (d)
and (e) which lab produced the more accurate 1.73, 2.47, 2.83, 3.20, 3.20, 3.57, 3.93, 4.30,
results? Why or why not? 4.67, 5.03, 5.03, 5.40, 5.77, 6.13, 6.50, 7.23,
15. Agusalim, Ferry, and Hollowaty made some mea- 7.60, 8.33, 9.43, 11.27
surements on the thickness of wallboard during
its manufacture. The accompanying table shows (a) Compute the mean and standard deviation of
thicknesses (in inches) of 12 different 4 ft × 8 ft these particle diameters.
boards (at a single location on the boards) both (b) Make both a dot diagram and a boxplot for
before and after drying in a kiln. (These boards these data. Sketch the dot diagram on a ruled
were nominally .500 in. thick.) scale and make the boxplot below it.
(c) Based on your work in (b), how would you
describe the shape of this data set?
Board 1 2 3 4 5 6 (d) Make a normal plot of these data. In what
Before Drying .514 .505 .500 .490 .503 .500 specific way does the distribution depart from
After Drying .510 .502 .493 .486 .497 .494 being bell-shaped?
(e) It is sometimes useful to find a scale of mea-
Board 7 8 9 10 11 12
surement on which a data set is reasonably
Before Drying .510 .508 .500 .511 .505 .501
bell-shaped. To that end, take the natural loga-
After Drying .502 .505 .488 .486 .491 .498
rithms of the raw particle diameters. Normal-
plot the log diameters. Does this plot appear
(a) Make a scatterplot of these data. Does there
to be more linear than your plot in (d)?
appear to be a strong relationship between
after-drying thickness and before-drying 17. The data in the accompanying tables are measure-
thickness? How might such a relationship ments of the latent heat of fusion of ice taken from
be of practical engineering importance in the Experimental Statistics (NBS Handbook 91) by
manufacture of wallboard? M. G. Natrella. The measurements were made (on
(b) Calculate the 12 before minus after differ- specimens cooled to −.072◦ C) using two differ-
ences in thickness. Find the sample mean and ent methods. The first was an electrical method,
sample standard deviation of these values. and the second was a method of mixtures. The
How might the mean value be used in running units are calories per gram of mass.
the sheetrock manufacturing process? (Based (a) Make side-by-side boxplots for comparing
on the mean value, what is an ideal before- the two measurement methods. Does there
drying thickness for the boards?) If some- appear to be any important difference in the
how all variability in before-drying thickness precision of the two methods? Is it fair to
could be eliminated, would substantial after- say that at least one of the methods must be
drying variability in thickness remain? Ex- somewhat inaccurate? Explain.
plain in terms of your calculations.
120 Chapter 3 Elementary Descriptive Statistics
(b) The dial bore data might well be termed (c) Find the sample mean, the sample range, and
“paired” data. A common method of anal- the sample standard deviation for the Laser
ysis for such data is to take differences and data.
study those. Compute the ten “notch minus Now consider comparing the two different drilling
non-notch” differences for the dial bore val- methods.
ues. Make a dot diagram for these and then (d) Make back-to-back stem-and-leaf plots for
a boxplot. What physical interpretation does the two data sets.
a nonzero mean for such differences have? (e) Make side-by-side boxplots for the two data
What physical interpretation does a large vari- sets. (Draw these on the same scale.)
ability in these differences have? (f) Based on your work in parts (d) and (e), which
(c) Make a scatterplot of the air spindler notch of the two processes would you say produced
measurements versus the dial bore notch mea- the most consistent results? Which process
surements. Does it appear that the air spindler produced an “average” angle closest to the
and dial bore measurements are strongly re- nominal angle (45◦ )?
lated? As it turns out, each metal part actually had two
(d) How would you suggest trying to determine holes drilled in it and their angles measured. Be-
which of the two gauges is most precise? low are the measured angles of the second hole
20. Duren, Leng and Patterson studied the drilling of drilled in each of the parts made using the Laser
holes in a miniature metal part using two different process. (The data are listed in the same part order
physical processes (laser drilling and electrical as earlier.)
discharge machining). Blueprint specifications on
these holes called for them to be drilled at an angle Laser (Hole B)
of 45◦ to the top surface of the part in question.
43.1, 44.3, 44.5, 46.3, 43.9, 41.9,
The realized angles measured on 13 parts drilled
using each process (26 parts in all) are 43.4, 49.0, 43.5, 47.2, 44.8 ,44.0, 43.9
conversion in the case of the Rockwell readings) (g) Is it possible to tell from your plot (e) which
were machine produced the most accurate results?
Why or why not?
Dial Rockwell 22. Ritchey, Bazan, and Buhman did an experiment
to compare flight times of several designs of pa-
536.6, 539.2, 524.4,
per helicopters, dropping them from the first to
536.6, 526.8, 531.6,
ground floors of the ISU Design Center. The flight
540.5, 534.0, 526.8,
times that they reported for two different designs
531.6 were (the units are seconds)
501.2, 522.0, 531.6, 2.47, 2.45, 2.43, 2.67, 2.69, 3.42, 3.50, 3.29, 3.51, 3.53,
522.0, 519.4, 523.2, 2.48, 2.44, 2.71, 2.84, 2.84 2.67, 2.69, 3.47, 3.40, 2.87
522.0, 514.2, 506.4,
518.1 (a) Find the median and the first and third quar-
tiles for the Design 1 data. Then find the .62
quantile of the Design 1 data set.
Brinell (b) Make and interpret a normal plot for the De-
sign 1 data. Would you describe this distri-
542.6, 526.0, 520.5, bution as bell-shaped? If not, in what way(s)
514.0, 546.6, 512.6, does it depart from being bell-shaped?
516.0, 580.4, 600.0, (c) Find the sample mean, the sample range, and
601.0 the sample standard deviation for the Design 1
data. Show some work.
Consider first only the Dial Rockwell data. Now consider comparing the two different de-
(a) Find the median and the first and third quar- signs.
tiles for the hardness measurements. Then (d) Make back-to-back stem-and-leaf plots for
find the .27 quantile of the data set. the two data sets.
(b) Make and interpret a normal plot for these (e) Make side-by-side boxplots for the two data
data. Would you describe this distribution as sets. (Draw these on the same scale.)
bell-shaped? If not, in what way(s) does it (f) Based on your work in parts (d) and (e), which
depart from being bell-shaped? of the two designs would you say produced
(c) Find the sample mean, the sample range, and the most consistent results? Which design
the sample standard deviation for these data. produced the longest flight times?
Now consider comparing the readings from the (g) It is not really clear from the students’ report
different testers (i.e., consider all three data sets.) whether the data came from the dropping of
(d) Make back-to-back stem-and-leaf plots for one helicopter of each design ten times, or
the two Rockwell data sets. (Use two “leaves” from the dropping of ten helicopters of each
for observations 500–509, two for the obser- design once. Briefly discuss which of these
vations 510–519, etc.) possibilities is preferable if the object of the
(e) Make side-by-side boxplots for all three data study was to identify a superior design. (If
sets. (Draw these on the same scale.) necessary, review Section 2.3.4.)
(f) Based on your work in part (e), which of the
three machines would you say produced the
most precise results?
4
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Describing
Relationships
Between Variables
The methods of Chapter 3 are really quite simple. They require little in the way of
calculations and are most obviously relevant to the analysis of a single engineering
variable. This chapter provides methods that address the more complicated prob-
lem of describing relationships between variables and are computationally more
demanding.
The chapter begins with least squares fitting of a line to bivariate quantitative
data and the assessment of the goodness of that fit. Then the line-fitting ideas are
generalized to the fitting of curves to bivariate data and surfaces to multivariate
quantitative data. The next topic is the summarization of data from full factorial
studies in terms of so-called factorial effects. Next, the notion of data transforma-
tions is discussed. Finally, the chapter closes with a short transitional section that
argues that further progress in statistics requires some familiarity with the subject
of probability.
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
y ≈ β 0 + β1 x (4.1)
123
124 Chapter 4 Describing Relationships Between Variables
Table 4.1
Pressing Pressures and Resultant
Specimen Densities
x, y,
Pressure (psi) Density (g/cc)
2,000 2.486
2,000 2.479
2,000 2.472
4,000 2.558
4,000 2.570
4,000 2.580
6,000 2.646
6,000 2.657
6,000 2.653
8,000 2.724
8,000 2.774
8,000 2.808
10,000 2.861
10,000 2.879
10,000 2.858
4.1 Fitting a Line by Least Squares 125
2.900
2.800
2.600
2.500
It is very easy to imagine sketching a straight line through the plotted points in
Figure 4.1. Such a line could then be used to summarize how density depends upon
pressing pressure. The principle of least squares provides a method of choosing a
“best” line to describe the data.
Definition 1 To apply the principle of least squares in the fitting of an equation for y to
an n-point data set, values of the equation parameters are chosen to minimize
X
n
2
yi − ŷ i (4.2)
i=1
In the context of fitting a line to (x, y) data, the prescription offered by Def-
inition 1 amounts to choosing a slope and intercept so as to minimize the sum of
squared vertical distances from (x, y) data points to the line in question. This notion
is shown in generic fashion in Figure 4.2 for a fictitious five-point data set. (It is the
squares of the five indicated differences that must be added and minimized.)
Looking at the form of display (4.1), for the fitting of a line,
ŷ = β0 + β1 x
126 Chapter 4 Describing Relationships Between Variables
y
A possible
fitted line
y4 – y4
y 3 – y3 is positive
is positive
y 5 – y5
y1 – y1 is negative
is positive
y 2 – y2
is negative
X
n
2
S(β0 , β1 ) = yi − (β0 + β1 xi ) (4.3)
i=1
!
X
n X
n
nβ0 + x i β1 = yi (4.4)
i=1 i=1
and
! !
X
n X
n X
n
x i β0 + xi2 β1 = xi yi (4.5)
i=1 i=1 i=1
For reasons that are not obvious, equations (4.4) and (4.5) are sometimes called
the normal (as in perpendicular) equations for fitting a line. They are two linear
equations in two unknowns and can be fairly easily solved for β0 and β1 (provided
4.1 Fitting a Line by Least Squares 127
there are at least two different xi ’s in the data set). Simultaneous solution of equations
(4.4) and (4.5) produces values of β1 and β0 given by
Slope of the P
xi − x̄ yi − ȳ
least squares b1 = P 2 (4.6)
line, b1 xi − x̄
and
Intercept of
the least b0 = ȳ − b1 x̄ (4.7)
squares line, b0
Notice the notational convention here. The particular numerical slope and intercept
minimizing S(β0 , β1 ) are denoted (not as β’s but) as b1 and b0 .
In display (4.6), somewhat standard practice has been followed (and the sum-
mation notation abused) by not indicating the variable or range of summation (i,
from 1 to n).
Example 1 It is possible to verify that the data in Table 4.1 yield the following summary
(continued ) statistics:
X
xi = 2,000 + 2,000 + · · · + 10,000 = 90,000,
90,000
so x̄ = = 6,000
15
X 2
xi − x̄ = (2,000 − 6,000)2 + (2,000 − 6,000)2 + · · · +
Example 1 Then the least squares slope and intercept, b1 and b0 , are given via equations
(continued ) (4.6) and (4.7) as
5,840
I b1 = = .0000486 (g/cc)/psi
120,000,000
and
I ŷ = 2.375 + .0000487x
Interpretation of sketched on a scatterplot of the (x, y) points from Table 4.1. Note that the slope on
the slope of the this plot, b1 ≈ .0000487 (g/cc)/psi, has physical meaning as the (approximate)
least squares increase in y (density) that accompanies a unit (1 psi) increase in x (pressure).
line The intercept on the plot, b0 = 2.375 g/cc, positions the line vertically and is the
value at which the line cuts the y axis. But it should probably not be interpreted
as the density that would accompany a pressing pressure of x = 0 psi. The point
is that the reasonably linear-looking relation that the students found for pressures
between 2,000 psi and 10,000 psi could well break down at larger or smaller
Extrapolation pressures. Thinking of b0 as a 0 pressure density amounts to an extrapolation
outside the range of data used to fit the equation, something that ought always to
be approached with extreme caution.
2.900
2.800
Density (g/cc)
2.700
2.600
Least squares line
y = 2. 375 + .0000487x
2.500
1
ȳ = (2.558 + 2.570 + 2.580) = 2.5693 g/cc
3
and so to use this as a representative value. But assuming that y is indeed
approximately linearly related to x, the fitted value
might be even better for representing average density for 4,000 psi pressure.
Looking then at the situation for x = 5,000 psi, there are no data with this
x value. The only thing one can do to represent density at that pressure is to ask
Interpolation whether interpolation is sensible from a physical viewpoint. If so, the fitted value
Definition 2 The sample (linear) correlation between x and y in a sample of n data pairs
(xi , yi ) is
P
xi − x̄ yi − ȳ
r=q 2 P 2 (4.8)
P
xi − x̄ · yi − ȳ
Interpreting the The sample correlation always lies in the interval from −1 to 1. Further, it is −1
sample correlation or 1 only when all (x, y) data points fall on a single straight line. Comparison of
130 Chapter 4 Describing Relationships Between Variables
P 2 P 2 1/2
formulas (4.6) and (4.8) shows that r = b1 xi − x̄ / yi − ȳ so that
b1 and r have the same sign. So a sample correlation of −1 means that y decreases
linearly in increasing x, while a sample correlation of +1 means that y increases
linearly in increasing x.
Real data sets do not often exhibit perfect (+1 or −1) correlation. Instead r is
typically between −1 and 1. But drawing on the facts about how it behaves, people
take r as a measure of the strength of an apparent linear relationship: r near +1
or −1 is interpreted as indicating a relatively strong linear relationship; r near 0
is taken as indicating a lack of linear relationship. The sign of r is thought of as
indicating whether y tends to increase or decrease with increased x.
Example 1 For the pressure/density data, the summary statistics in the example following
(continued ) display (4.7) (page 127) produces
5,840
I r=p = .9911
(120,000,000)(.289366)
This value of r is near +1 and indicates clearly the strong positive linear rela-
tionship evident in Figures 4.1 and 4.3.
Definition 3 The coefficient of determination for an equation fitted to an n-point data set
via least squares and producing fitted y values ŷ 1 , ŷ 2 , . . . , ŷ n is
P 2 P 2
yi − ȳ − yi − ŷ i
R =
2
P 2 (4.9)
yi − ȳ
Interpretation R 2 may be interpreted as the fraction of the raw variation in y accounted for
of R2 using Pthe fitted equation.
P That is, providedP the fitted equation includes a constant
term, (yi − ȳ)2 ≥ (yi − ŷ i )2 . Further, (yi − ȳ)2 is a measure of raw variabil-
P
ity in y, while (yi − ŷ i )2 is a measure of variation in y remaining after fitting the
P P
equation. So the nonnegative difference (yi − ȳ)2 − (yi − ŷ i )2 is a measure of
the variability in y accounted for in the equation-fitting process. R 2 then expresses
this difference as a fraction (of the total raw variation).
4.1 Fitting a Line by Least Squares 131
Example 1 Using the fitted line, one can find ŷ values for all n = 15 data points in the original
(continued ) data set. These are given in Table 4.2.
Table 4.2
Fitted Density Values
.289366 − .005153
I R2 = = .9822
.289366
and the fitted line accounts for over 98% of the raw variability in density, reducing
the “unexplained” variation from .289366 to .005153.
R2 as a squared The coefficient of determination has a second useful interpretation. For equa-
correlation tions that are linear in the parameters (which are the only ones considered in this
text), R 2 turns out to be a squared correlation. It is the squared correlation between
the observed values yi and the fitted values ŷ i . (Since in the present situation of
fitting a line, the ŷ i values are perfectly correlated with the xi values, R 2 also turns
out to be the squared correlation between the yi and xi values.)
Example 1 Since ŷ is perfectly correlated with x, this is also the correlation between ŷ and y.
(continued ) But notice as well that
r 2 = (.9911)2 = .9822 = R 2
ei = yi − ŷ i
If a fitted equation is telling the whole story contained in a data set, then its
residuals ought to be patternless. So when they’re plotted against time order of
observation, values of experimental variables, fitted values, or any other sensible
quantities, the plots should look randomly scattered. When they don’t, the patterns
can themselves suggest what has gone unaccounted for in the fitting and/or how the
data summary might be improved.
Table 4.3
Additive Concentrations and Compressive Strengths for Fly Ash Cylinders
0 1221 3 1609
0 1207 3 1627
0 1187 3 1642
1 1555 4 1451
1 1562 4 1472
1 1575 4 1465
2 1827 5 1321
2 1839 5 1289
2 1802 5 1292
Using formulas (4.6) and (4.7), it is possible to show that the least squares
line through the (x, y) data in Table 4.3 is
Table 4.4
Residuals from a Straight-Line Fit to the Fly Ash Data
x y ŷ e = y − ŷ x y ŷ e = y − ŷ
Example 2
(continued ) 300
200
100
Residual, ei
–100
–200
–300
0 1 2 3 4 5
Percent ammonium phosphate, xi
the fitting of a line to Roth’s data. Figure 4.5 is a simple scatterplot of Roth’s
data (which in practice should be made before fitting any curve to such data).
It is obvious from the scatterplot that the relationship between the amount of
ammonium phosphate and compressive strength is decidedly nonlinear. In fact,
a quadratic function would come much closer to fitting the data in Table 4.3.
1800
1700
Compressive strength (psi)
1600
1500
Least squares line
1400
1300
1200
0 1 2 3 4 5
Percent ammonium phosphate
Order of yi 1 2 Technician
observation, i
Interpreting Figure 4.6 shows several patterns that can occur in plots of residuals against
patterns on various variables. Plot 1 of Figure 4.6 shows a trend on a plot of residuals versus
residual plots time order of observation. The pattern suggests that some variable changing in time
is acting on y and has not been accounted for in fitting ŷ values. For example,
instrument drift (where an instrument reads higher late in a study than it did early
on) could produce a pattern like that in Plot 1. Plot 2 shows a fan-shaped pattern on
a plot of residuals versus fitted values. Such a pattern indicates that large responses
are fitted (and quite possibly produced and/or measured) less consistently than small
responses. Plot 3 shows residuals corresponding to observations made by Technician
1 that are on the whole smaller than those made by Technician 2. The suggestion is
that Technician 1’s work is more precise than that of Technician 2.
Normal-plotting Another useful way of plotting residuals is to normal-plot them. The idea is that
residuals the normal distribution shape is typical of random variation and that normal-plotting
of residuals is a way to investigate whether such a distributional shape applies to
what is left in the data after fitting an equation or model.
Example 1 Table 4.5 gives residuals for the fitting of a line to the pressure/density data. The
(continued ) residuals ei were treated as a sample of 15 numbers and normal-plotted (using
the methods of Section 3.2) to produce Figure 4.7.
The central portion of the plot in Figure 4.7 is fairly linear, indicating a gen-
erally bell-shaped distribution of residuals. But the plotted point corresponding to
the largest residual, and probably the one corresponding to the smallest residual,
fail to conform to the linear pattern established by the others. Those residuals
seem big in absolute value compared to the others.
From Table 4.5 and the scatterplot in Figure 4.3, one sees that these large
residuals both arise from the 8,000 psi condition. And the spread for the three
densities at that pressure value does indeed look considerably larger than those at
the other pressure values. The normal plot suggests that the pattern of variation
at 8,000 psi is genuinely different from those at other pressures. It may be that
a different physical compaction mechanism was acting at 8,000 psi than at the
other pressures. But it is more likely that there was a problem with laboratory
technique, or recording, or the test equipment when the 8,000 psi tests were made.
136 Chapter 4 Describing Relationships Between Variables
Example 1 In any case, the normal plot of residuals helps draw attention to an idiosyncrasy
(continued ) in the data of Table 4.1 that merits further investigation, and perhaps some further
data collection.
Table 4.5
Residuals from the Linear Fit to the Pressure/Density
Data
x, Pressure y, Density ŷ e = y − ŷ
2,000 2.486 2.4723 .0137
2,000 2.479 2.4723 .0067
2,000 2.472 2.4723 −.0003
4,000 2.558 2.5697 −.0117
4,000 2.570 2.5697 .0003
4,000 2.580 2.5697 .0103
6,000 2.646 2.6670 −.0210
6,000 2.657 2.6670 −.0100
6,000 2.653 2.6670 −.0140
8,000 2.724 2.7643 −.0403
8,000 2.774 2.7643 .0097
8,000 2.808 2.7643 .0437
10,000 2.861 2.8617 −.0007
10,000 2.879 2.8617 .0173
10,000 2.858 2.8617 −.0037
2.0
Standard normal quantile
1.0
–1.0
–2.0
–.04 –.02 0 .02 .04
Residual quantile
80
A 30-year-old
75
6' 8" student
Height (in.)
70 2 3 2
3
65
60
20 22 24 26 28 30
Age (years)
data alone would have identified the 30-year-old student in Figure 4.8 as unusual.
That would have raised the possibility of that data point strongly influencing both r
and any curve that might be fitted via least squares.
4.1.5 Computing
The examples in this section have no doubt left the impression that computations
were done “by hand.” In practice, such computations are almost always done with
a statistical analysis package. The fitting of a line by least squares is done using a
regression program. Such programs usually also compute R 2 and have an option
that allows the computing and plotting of residuals.
It is not the purpose of this text to teach or recommend the use of any particular
statistical package, but annotated printouts will occasionally be included to show
how MINITAB formats its output. Printout 1 is such a printout for an analysis of
the pressure/density data in Table 4.1, paralleling the discussion in this section.
(MINITAB’s regression routine is found under its “Stat/Regression/Regression”
menu.) MINITAB gives its user much more in the way of analysis for least squares
curve fitting than has been discussed to this point, so your understanding of Printout 1
will be incomplete. But it should be possible to locate values of the major summary
statistics discussed here. The printout shown doesn’t include plots, but it’s worth
noting that the program has options for saving fitted values and residuals for later
plotting.
Analysis of Variance
Source DF SS MS F P
Regression 1 0.28421 0.28421 717.06 0.000
Residual Error 13 0.00515 0.00040
Total 14 0.28937
At the end of Section 3.3 we warned that using spreadsheet software in place of
high-quality statistical software can, without warning, produce spectacularly wrong
answers. The example provided at the end of Section 3.3 concerns a badly wrong
sample variance of only three numbers. It is important to note that the potential
for numerical inaccuracy shown in that example carries over to the rest of the
statistical methods discussed in this book, including those of the present section.
For example, consider the n = 6 hypothetical (x, y) pairs listed in Table 4.6. For
fitting a line to these data via least squares, MINITAB correctly produces R 2 = .997.
But as recently as late 1999, the current version of the leading spreadsheet program
returned the ridiculously wrong value, R 2 = −.81648. (This data set comes from a
posting by Mark Eakin on the “edstat” electronic bulletin board that can be found
at http://jse.stat.ncsu.edu/archives/.)
Table 4.6
6 Hypothetical Data Pairs
x y x y
Section 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. The following is a small set of artificial data. Show (c) Obtain the sample correlation between y and
the hand calculations necessary to do the indicated ŷ for these data and compare it to your answer
tasks. to part (b).
(d) Use the formula in Definition 3 and compute
x 1 2 3 4 5 R 2 for these data. Compare it to the square of
y 8 8 6 6 4 your answers to parts (b) and (c).
(e) Find the five residuals from your fit in part (a).
(a) Obtain the least squares line through these data. How are they portrayed geometrically on the
Make a scatterplot of the data and sketch this scatterplot for (a)?
line on that scatterplot. 2. Use a computer package and redo the computations
(b) Obtain the sample correlation between x and y and plotting required in Exercise 1. Annotate your
for these data. output, indicating where on the printout you can
140 Chapter 4 Describing Relationships Between Variables
find the equation of the least squares line, the value (e) Based on your analysis of these data, what
of r , the value of R 2 , and the residuals. average molecular weight would you predict
3. The article “Polyglycol Modified Poly (Ethylene for an additional reaction run at 188◦ C? At
Ether Carbonate) Polyols by Molecular Weight Ad- 200◦ C? Why would or wouldn’t you be willing
vancement” by R. Harris (Journal of Applied Poly- to make a similar prediction of average molec-
mer Science, 1990) contains some data on the effect ular weight if the reaction is run at 70◦ C?
of reaction temperature on the molecular weight of 4. Upon changing measurement scales, nonlinear re-
resulting poly polyols. The data for eight experi- lationships between two variables can sometimes
mental runs at temperatures 165◦ C and above are be made linear. The article “The Effect of Experi-
as follows: mental Error on the Determination of the Optimum
Metal-Cutting Conditions” by Ermer and Wu (The
Pot Temperature, x (◦ C) Average Molecular Weight, y Journal of Engineering for Industry, 1967) con-
tains a data set gathered in a study of tool life in
165 808 a turning operation. The data here are part of that
176 940 data set.
188 1183
205 1545 Cutting Speed, x (sfpm) Tool Life, y (min)
220 2012
800 1.00, 0.90, 0.74, 0.66
235 2362
700 1.00, 1.20, 1.50, 1.60
250 2742
600 2.35, 2.65, 3.00, 3.60
260 2935
500 6.40, 7.80, 9.80, 16.50
Use a statistical package to help you complete the 400 21.50, 24.50, 26.00, 33.00
following (both the plotting and computations):
(a) What fraction of the observed raw variation in (a) Plot y versus x and calculate R 2 for fitting a
y is accounted for by a linear equation in x? linear function of x to y. Does the relationship
(b) Fit a linear relationship y ≈ β0 + β1 x to these y ≈ β0 + β1 x look like a reasonable explana-
data via least squares. About what change in tion of tool life in terms of cutting speed?
average molecular weight seems to accompany (b) Take natural logs of both x and y and repeat
a 1◦ C increase in pot temperature (at least over part (a) with these log cutting speeds and log
the experimental range of temperatures)? tool lives.
(c) Compute and plot residuals from the linear re- (c) Using the logged variables as in (b), fit a lin-
lationship fit in (b). Discuss what they suggest ear relationship between the two variables us-
about the appropriateness of that fitted equa- ing least squares. Based on this fitted equation,
tion. (Plot residuals versus x, residuals versus what tool life would you predict for a cutting
ŷ, and make a normal plot of them.) speed of 550? What approximate relationship
(d) These data came from an experiment where the between x and y is implied by a linear approx-
investigator managed the value of x. There is imate relationship between ln(x) and ln(y)?
a fairly glaring weakness in the experimenter’s (Give an equation for this relationship.) By the
data collection efforts. What is it? way, Taylor’s equation for tool life is yx α = C.
4.2 Fitting Curves and Surfaces by Least Squares 141
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
y ≈ β 0 + β1 x (4.11)
y ≈ β 0 + β1 x + β 2 x 2 + · · · + βk x k (4.12)
The least squares fitting of equation (4.12) to a set of n pairs (xi , yi ) is conceptually
only slightly more difficult than the task of fitting equation (4.11). The function of
k + 1 variables
X
n X
n
2
S(β0 , β1 , β2 , . . . , βk ) = (yi − ŷ i ) = 2
yi − (β0 + β1 xi + β2 xi2 + · · · + βk xik )
i=1 i=1
y ≈ β 0 + β1 x + β2 x 2 (4.13)
to the data of Table 4.3. Printout 2 shows the MINITAB run. (After entering x
and y values from Table 4.3 into two columns of the worksheet, an additional
column was created by squaring the x values.)
Analysis of Variance
Source DF SS MS F P
Regression 2 658230 329115 48.78 0.000
Residual Error 15 101206 6747
Total 17 759437
Source DF Seq SS
x 1 21
x**2 1 658209
Figure 4.9 shows the fitted curve sketched on a scatterplot of the (x, y) data.
Although the quadratic curve is not an altogether satisfactory summary of Roth’s
data, it does a much better job of following the trend of the data than the line
sketched in Figure 4.5.
4.2 Fitting Curves and Surfaces by Least Squares 143
1800
1700
1500
1400
1300
1200
0 1 2 3 4 5
Percent ammonium phosphate
The previous section showed that when fitting a line to (x, y) data, it is helpful
to quantify the goodness of that fit using R 2 . The coefficient of determination can
also be used when fitting a polynomial of form (4.12). Recall once more from
Definition 3 that
P P
(yi − ȳ)2 − (yi − ŷ i )2
Coefficient of R =
2
P (4.14)
determination (yi − ȳ)2
is the fraction of the raw variability in y accounted for by the fitted equation.
Calculation by hand from formula (4.14) is possible, but of course the easiest way
to obtain R 2 is to use a computer package.
Example 3 Consulting Printout 2, it can be seen that the equation ŷ = 1242.9 + 382.7x −
(continued ) 76.7x 2 produces R 2 = .867. So 86.7% of the raw variability in compressive
strength is accounted for using the fitted quadratic. The sample√ correlation be-
tween the observed strengths yi and fitted strengths ŷ i is + .867 = .93.
Comparing what has been done in the present section to what was done in
Section 4.1, it is interesting that for the fitting of a line to the fly ash data, R 2
obtained there was only .000 (to three decimal places). The present quadratic is
a remarkable improvement over a linear equation for summarizing these data.
A natural question to raise is “What about a cubic version of equation (4.12)?”
Printout 3 shows some results of a MINITAB run made to investigate this possi-
bility, and Figure 4.10 shows a scatterplot of the data and a plot of the fitted cubic
144 Chapter 4 Describing Relationships Between Variables
Example 3 equation. (x values were squared and cubed to provide x, x 2 , and x 3 for each y
(continued ) value to use in the fitting.)
Regression Analysis
Source DF SS MS F P
Regression 3 723197 241066 93.13 0.000
Residual Error 14 36240 2589
Total 17 759437
1800
1700
Compressive strength (psi)
1400
1300
1200
0 1 2 3 4 5
Percent ammonium phosphate
Figure 4.10 Scatterplot and fitted cubic for the fly ash
data
4.2 Fitting Curves and Surfaces by Least Squares 145
R 2 for the cubic equation is .952, somewhat larger than for the quadratic.
But it is fairly clear from Figure 4.10 that even a cubic polynomial is not totally
satisfactory as a summary of these data. In particular, both the fitted quadratic in
Figure 4.9 and the fitted cubic in Figure 4.10 fail to fit the data adequately near
an ammonium phosphate level of 2%. Unfortunately, this is where compressive
strength is greatest—precisely the area of greatest practical interest.
The example illustrates that R 2 is not the only consideration when it comes to
judging the appropriateness of a fitted polynomial. The examination of plots is also
important. Not only scatterplots of y versus x with superimposed fitted curves but
plots of residuals can be helpful. This can be illustrated on a data set where y is
expected to be nearly perfectly quadratic in x.
gt 2
displacement =
2
one expects
2
1 1
y ≈ g t0 + (x − 1)
2 60
g x 2 1 x g 1 2
= + g t0 − + t − (4.15)
2 60 60 60 2 0 60
g 2 g 1 g 1 2
= x + t0 − x+ t0 −
7200 60 60 2 60
ŷ = b0 + b1 x + b2 x 2 (4.16)
7200b2 . This is in fact how the value 9.79 m/sec2 , quoted in Section 1.4, was
obtained.
A multiple linear regression program fits equation (4.16) to the bob drop data
giving
(from which g ≈ 9790 mm/sec2 ) with R 2 that is 1.0 to 6 decimal places. Residuals
for this fit can be calculated using Definition 4 and are also given in Table 4.7.
Figure 4.11 is a normal plot of the residuals. It is reasonably linear and thus not
remarkable (except for some small suggestion that the largest residual or two may
not be as extreme as might be expected, a circumstance that suggests no obvious
physical explanation).
4.2 Fitting Curves and Surfaces by Least Squares 147
2.0
–1.0
–2.0
–.3 –.2 –.1 0 .1 .2 .3
Residual quantile
.30
Residual
–.30
5 10 15 20 25
Point number, x
Example 4 pattern suggested by Figure 4.12 reappear consistently, it would indicate that
(continued ) something in the mechanism generating the 60 cycle current may cause cycles
1
to be alternately slightly shorter then slightly longer than 60 sec. The practical
implication of this would be that if a better determination of g were desired, the
regularity of the AC current waveform is one matter to be addressed.
What if a Examples 3 and 4 (respectively) illustrate only partial success and then great
polynomial success in describing an (x, y) data set by means of a polynomial equation. Situations
doesn’t fit like Example 3 obviously do sometimes occur, and it is reasonable to wonder what
(x, y) data? to do when they happen. There are two simple things to keep in mind.
For one, although a polynomial may be unsatisfactory as a global description
of a relationship between x and y, it may be quite adequate locally—i.e., for
a relatively restricted range of x values. For example, in the fly ash study, the
quadratic representation of compressive strength as a function of percent ammonium
phosphate is not appropriate over the range 0 to 5%. But having identified the region
around 2% as being of practical interest, it would make good sense to conduct a
follow-up study concentrating on (say) 1.5 to 2.5% ammonium phosphate. It is quite
possible that a quadratic fit only to data with 1.5 ≤ x ≤ 2.5 would be both adequate
and helpful as a summarization of the follow-up data.
The second observation is that the terms x, x 2 , x 3 , . . . , x k in equation (4.12) can
be replaced by any (known) functions of x and what we have said here will remain
essentially unchanged. The normal equations will still be k + 1 linear equations
in β0 , β1 , . . . , βk , and a multiple linear regression program will still produce least
squares values b0 , b1 , . . . , bk . This can be quite useful when there are theoretical
reasons to expect a particular (nonlinear but) simple functional relationship between
x and y. For example, Taylor’s equation for tool life is of the form
y ≈ αx β
for y tool life (e.g., in minutes) and x the cutting speed used (e.g., in sfpm). Taking
logarithms,
This is an equation for ln(y) that is linear in the parameters ln(α) and β involving
the variable ln(x). So, presented with a set of (x, y) data, empirical values for α and
β could be determined by
y ≈ β 0 + β1 x 1 + β2 x 2 + · · · + βk x k (4.17)
to the data using the least squares principle. This is pictured for a k = 2 case in
Figure 4.13, where six (x 1 , x2 , y) data points are pictured in three dimensions, along
with a possible fitted surface of the form (4.17). To fit a surface defined by equation
(4.17) to a set of n data points (x1i , x2i , . . . , xki , yi ) via least squares, the function
of k + 1 variables
X
n X
n
2
S(β0 , β1 , β2 , . . . , βk ) = (yi − ŷ i )2 = yi − (β0 + β1 x1i + · · · + βk xki )
i=1 i=1
x2
x1
ŷ = b0 + b1 x1 + b2 x2 + b3 x3
Table 4.8
Brownlee’s Stack Loss Data
i, x2i , x3i ,
Observation x1i , Cooling Water Acid yi ,
Number Air Flow Inlet Temperature Concentration Stack Loss
1 80 27 88 37
2 62 22 87 18
3 62 23 87 18
4 62 24 93 19
5 62 24 93 20
6 58 23 87 15
7 58 18 80 14
8 58 18 89 14
9 58 17 88 13
10 58 18 82 11
11 58 19 93 12
12 50 18 89 8
13 50 18 86 7
14 50 19 72 8
15 50 19 79 8
16 50 20 80 9
17 56 20 82 15
4.2 Fitting Curves and Surfaces by Least Squares 151
Interpreting with R 2 = .975. The coefficients in this equation can be thought of as rates of
fitted coefficients change of stack loss with respect to the individual variables x 1 , x2 , and x3 , holding
from a multiple the others fixed. For example, b1 = .80 can be interpreted as the increase in stack
regression loss y that accompanies a one-unit increase in air flow x 1 if inlet temperature x2
and acid concentration x3 are held fixed. The signs on the coefficients indicate
whether y tends to increase or decrease with increases in the corresponding x. For
example, the fact that b1 is positive indicates that the higher the rate at which the
plant is run, the larger y tends to be (i.e., the less efficiently the plant operates).
The large value of R 2 is a preliminary indicator that the equation (4.18) is an
effective summarization of the data.
Source DF SS MS F P
Regression 3 795.83 265.28 169.04 0.000
Residual Error 13 20.40 1.57
Total 16 816.24
Source DF Seq SS
air 1 775.48
water 1 18.49
acid 1 1.86
Unusual Observations
Obs air stack Fit StDev Fit Residual St Resid
10 58.0 11.000 13.506 0.552 -2.506 -2.23R
Example 5 In the context of the nitrogen plant, it is sensible to ask whether all three variables,
(continued ) x1 , x2 , and x3 , are required to adequately account for the observed variation in
y. For example, the behavior of stack loss might be adequately explained using
only one or two of the three x variables. There would be several consequences
of practical engineering importance if this were so. For one, in such a case, a
simple or parsimonious version of equation (4.17) could be used in describing
the oxidation process. And if a variable is not needed to predict y, then it is
possible that the expense of measuring it might be saved. Or, if a variable doesn’t
seem to have much impact on y (because it doesn’t seem to be essential to include
it when writing an equation for y), it may be possible to choose its level on purely
economic grounds, without fear of degrading process performance.
As a means of investigating whether indeed some subset of x 1 , x2 , and x3
is adequate to explain stack loss behavior, R 2 values for equations based on all
possible subsets of x1 , x2 , and x3 were obtained and placed in Table 4.9. This
shows, for example, that 95% of the raw variability in y can be accounted for
using a linear equation in only the air flow variable x 1 . Use of both x1 and the
water temperature variable x 2 can account for 97.3% of the raw variability in
stack loss. Inclusion of x 3 , the acid concentration variable, in an equation already
involving x 1 and x2 , increases R 2 only from .973 to .975.
If identifying a simple equation for stack loss that seems to fit the data well
is the goal, the message in Table 4.9 would seem to be “Consider an x 1 term first,
and then possibly an x2 term.” On the basis of R 2 , including an x3 term in an
equation for y seems unnecessary. And in retrospect, this is entirely consistent
with the character of the fitted equation (4.18): x 3 varies from 72 to 93 in the
original data set, and this means that ŷ changes only a total amount
ŷ = b0 + b1 x1 + b2 x2 + b3 x3
these can and should go through thorough residual analyses before they are
adopted as data summaries. As an example, consider a fitted equation involving
4.2 Fitting Curves and Surfaces by Least Squares 153
Table 4.9
R2 ’s for Equations Predicting Stack Loss
Equation Fit R2
y ≈ β0 + β1 x1 .950
y ≈ β0 + β2 x2 .695
y ≈ β0 + β3 x3 .165
y ≈ β0 + β1 x1 + β2 x2 .973
y ≈ β0 + β1 x1 + β3 x3 .952
y ≈ β0 + β2 x2 + β3 x3 .706
y ≈ β0 + β1 x1 + β2 x2 + β3 x3 .975
x1 and x2 . A multiple linear regression program can be used to produce the fitted
equation
Dropping variables (Notice that b0 , b1 , and b2 in equation (4.19) differ somewhat from the corre-
from a fitted sponding values in equation (4.18). That is, equation (4.19) was not obtained
equation typically from equation (4.18) by simply dropping the last term in the equation. In general,
changes coefficients the values of the coefficients b will change depending on which x variables are
and are not included in the fitting.)
Residuals for equation (4.19) can be computed and plotted in any number
of potentially useful ways. Figure 4.14 shows a normal plot of the residuals and
three other plots of the residuals against, respectively, x 1 , x2 , and ŷ. There are
no really strong messages carried by the plots in Figure 4.14 except that the
data set contains one unusually large x 1 value and one unusually large ŷ (which
corresponds to the large x1 ). But there is enough of a curvilinear “up-then-down-
then-back-up-again” pattern in the plot of residuals against x 1 to suggest the
possibility of adding an x12 term to the fitted equation (4.19).
You might want to verify that fitting the equation
y ≈ β0 + β1 x1 + β2 x2 + β3 x12
with corresponding R 2 = .980 and residuals that show even less of a pattern than
those for the fitted equation (4.19). In particular, the hint of curvature on the plot
of residuals versus x1 for equation (4.19) is not present in the corresponding plot
for equation (4.20). Interestingly, looking back over this example, one sees that
fitted equation (4.20) has a better R 2 value than even fitted equation (4.18), in
154 Chapter 4 Describing Relationships Between Variables
Standard normal quantile
2.0 2.0
1.0 1.0
2
Residual
2
0 0
–1.0 –1.0
–2.0 –2.0
2.0 2.0
1.0 1.0
2 2
Residual
Residual
2 2
0 0
–1.0 –1.0
–2.0 –2.0
20 25 30 10 20 30
Inlet temperature, x2 Fitted Stack Loss, y
Figure 4.14 Plots of residuals from a two-variable equation fit to the stack loss data
( ŷ = −42.00 − .78x1 + .57x2 )
Example 5 spite of the fact that equation (4.18) involves the process variable x 3 and equation
(continued ) (4.20) does not.
Equation (4.20) is somewhat more complicated than equation (4.19). But
because it still really only involves two different input x’s and also eliminates the
slight pattern seen on the plot of residuals for equation (4.19) versus x 1 , it seems
an attractive choice for summarizing the stack loss data. A two-dimensional rep-
resentation of the fitted surface defined by equation (4.20) is given in Figure 4.15.
The slight curvature on the plotted curves is a result of the x 12 term appearing in
equation (4.20). Since most of the data have x 1 from 50 to 62 and x2 from 17 to
24, the curves carry the message that over these ranges, changes in x 1 seem to
produce larger changes in stack loss than do changes in x 2 . This conclusion is
consistent with the discussion centered around Table 4.9.
4.2 Fitting Curves and Surfaces by Least Squares 155
35
30
20 x2 = 28 x2 = 20
15 x2 = 24
10
50 55 60 65 70 75
Air flow, x1
Common residual The plots of residuals used in Example 5 are typical. They are
plots in multiple
regression
1. normal plots of residuals,
All of these can be used to help assess the appropriateness of surfaces fit to multivari-
ate data, and they all have the potential to tell an engineer something not previously
discovered about a set of data and the process that generated them.
Earlier in this section, there was a discussion of the fact that an “x term” in
the equations fitted via least squares can be a known function (e.g., a logarithm)
of a basic process variable. In fact, it is frequently helpful to allow an “x term” in
equation (4.17) (page 149) to be a known function of several basic process variables.
The next example illustrates this point.
156 Chapter 4 Describing Relationships Between Variables
x1 = canard placement in inches above the plane defined by the main wing
x2 = tail placement in inches above the plane defined by the main wing
(The front-to-rear positions of the three surfaces were constant throughout the
study.)
A straightforward least squares fitting of the equation
y ≈ β 0 + β1 x 1 + β2 x 2
to these data produces R 2 of only .394. Even the addition of squared terms in
both x1 and x2 , i.e., the fitting of
y ≈ β0 + β1 x1 + β2 x2 + β3 x12 + β4 x22
Table 4.10
Lift/Drag Ratios for 9 Canard/Tail Position Combinations
x1 , x2 , y,
Canard Position Tail Position Lift/Drag Ratio
−1.2 −1.2 .858
−1.2 0.0 3.156
−1.2 1.2 3.644
0.0 −1.2 4.281
0.0 0.0 3.481
0.0 1.2 3.918
1.2 −1.2 4.136
1.2 0.0 3.364
1.2 1.2 4.018
4.2 Fitting Curves and Surfaces by Least Squares 157
Regression Analysis
Analysis of Variance
Source DF SS MS F P
Regression 3 5.4771 1.8257 2.97 0.136
Residual Error 5 3.0724 0.6145
Total 8 8.5495
(After reading x1 , x2 , and y values from Table 4.10 into columns of MINITAB’s
worksheet, x1 x2 products were created and y fitted to the three predictor variables
x1 , x2 , and x1 x2 in order to create this printout.)
Figure 4.16 shows the nature of the fitted surface (4.21). Raising the canard
(increasing x1 ) has noticeably different predicted impacts on y, depending on the
value of x2 (the tail position). (It appears that the canard and tail should not be
lined up—i.e., x1 should not be near x2 . For large predicted response, one wants
small x1 for large x2 and large x1 for small x2 .) It is the cross-product term x1 x2
in relationship (4.21) that allows the response curves to have different characters
for different x2 values. Without it, the slices of the fitted (x 1 , x2 , ŷ) surface would
be parallel for various x 2 , much like the situation in Figure 4.15.
x2 = 1.2
Fitted lift / Drag ratio, y
4.0
3.0
x2 = 0
x2 = –1.2
2.0
1.0
–1.2 0 1.2
Canard position, x1
Example 6 Although the main new point of this example has by now been made, it
(continued ) probably should be mentioned that equation (4.21) is not the last word for fitting
the data of Table 4.10. Figure 4.17 gives a plot of the residuals for relationship
(4.21) versus canard position x1 , and it shows a strong curvilinear pattern. In fact,
the fitted equation
1.0
Residual
–1.0
–1.2 0 1.2
Canard position, x1
x2
Dots show (x1, x2) locations
of fictitious data points
20
The region
15 × with 1 ≤ x1 ≤ 5
and 10 ≤ x2 ≤ 20
10
(3,15) is unlike the
5 (x1, x2) pairs for the data
1 2 3 4 5 x1
that when several different x variables are involved, it is difficult to tell whether a
particular (x1 , x2 , . . . , xk ) vector is a large extrapolation. About all one can do is
check to see that it comes close to matching some single data point in the set on
each coordinate x1 , x2 , . . . , xk . It is not sufficient that there be some point with x 1
value near the one of interest, another point with x 2 value near the one of interest,
etc. For example, having data with 1≤ x 1 ≤ 5 and 10≤ x2 ≤ 20 doesn’t mean that
the (x1 , x2 ) pair (3, 15) is necessarily like any of the pairs in the data set. This fact
is illustrated in Figure 4.18 for a fictitious set of (x 1 , x2 ) values.
The influence Another potential pitfall is that the fitting of curves and surfaces via least squares
of outlying can be strongly affected by a few outlying or extreme data points. One can try to
data vectors identify such points by examining plots and comparing fits made with and without
the suspicious point(s).
Example 5 Figure 4.14 earlier called attention to the fact that the nitrogen plant data set
(continued ) contains one point with an extreme x 1 value. Figure 4.19 is a scatterplot of
(x1 , x2 ) pairs for the data in Table 4.8 (page 150). It shows that by most qualitative
standards, observation 1 in Table 4.8 is unusual or outlying.
If the fitting of equation (4.20) is redone using only the last 16 data points in
Table 4.8, the equation
and R 2 = .942 are obtained. Using equation (4.23) as a description of stack loss
and limiting attention to x 1 in the range 50 to 62 could be considered. But it
is possible to verify that though some of the coefficients (the b’s) in equations
(4.20) and (4.23) differ substantially, the two equations produce comparable ŷ
values for the 16 data points with x1 between 50 and 62. In fact, the largest
difference in fitted values is about .4. So, since point 1 in Table 4.8 doesn’t
160 Chapter 4 Describing Relationships Between Variables
Example 5
Water temperature, x2
(continued ) 25
2
20
2
2 3
15
50 55 60 65 70 75 80
Air flow, x1
Figure 4.19 Plot of (x1 , x2 ) pairs for the stack loss data
radically change predictions made using the fitted equation, it makes sense to
leave it in consideration, adopt equation (4.20), and use it to describe stack loss
for (x1 , x2 ) pairs interior to the pattern of scatter in Figure 4.19.
Replication and A third warning has to do with the notion of replication (first discussed in
surface fitting Section 2.3). It is the fact that the fly ash data of Example 3 has several y’s for
each x that makes it so clear that even the quadratic and cubic curves sketched
in Figures 4.9 and 4.10 are inadequate descriptions of the relationship between
phosphate and strength. The fitted curves pass clearly outside the range of what look
like believable values of y for some values of x. Without such replication, what is
permissible variation about a fitted curve or surface can’t be known with confidence.
For example, the structure of the lift/drag data set in Example 6 is weak from this
viewpoint. There is no replication represented in Table 4.10, so an external value for
typical experimental precision would be needed in order to identify a fitted value as
obviously incompatible with an observed one.
The nitrogen plant data set of Example 5 was presumably derived from a
primarily observational study, where no conscious attempt was made to replicate
(x1 , x2 , x3 ) settings. However, points number 4 and 5 in Table 4.8 (page 150) do
represent the replication of a single (x1 , x2 , x3 ) combination and show a difference
in observed stack loss of 1. And this makes the residuals for equation (4.20) (which
range from −2.0 to 2.3) seem at least not obviously out of line.
Section 9.2 discusses more formal and precise ways of using data from studies
with some replication to judge whether or not a fitted curve or surface misses some
observed y’s too badly. For now, simply note that among replication’s many virtues
is the fact that it allows more reliable judgments about the appropriateness of a fitted
equation than are otherwise possible.
The possibility The fourth caution is that the notion of equation simplicity ( parsimony) is
of overfitting important for reasons in addition to simplicity of interpretation and reduced expense
involved in using the equation. It is also important from the point of view of typically
giving smooth interpolation and not overfitting a data set. As a hypothetical example,
4.2 Fitting Curves and Surfaces by Least Squares 161
consider the artificial, generally linear (x, y) data plotted in Figure 4.20. It would be
possible to run a (wiggly) k = 10 version of the polynomial (4.12) through each of
these points. But in most physical problems, such a curve would do a much worse
job of predicting y at values of x not represented by a data point than would a simple
fitted line. A tenth-order polynomial would overfit the data in hand.
Empirical models As a final point in this section, consider how the methods discussed here fit
and engineering into the broad picture of using models for attacking engineering problems. It must
be said that physical theories of physics, chemistry, materials, etc. rarely produce
equations of the forms (4.12) or (4.17). Sometimes pertinent equations from those
theories can be rewritten in such forms, as was possible with Taylor’s equation for
tool life earlier in this section. But the majority of engineering applications of the
methods in this section are to the large number of problems where no commonly
known and simple physical theory is available, and a simple empirical description
of the situation would be helpful. In such cases, the tool of least squares fitting of
curves and surfaces can function as a kind of “mathematical French curve,” allowing
an engineer to develop approximate empirical descriptions of how a response y is
related to system inputs x1 , x2 , . . . , xk .
Section 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Return to Exercise 3 of Section 4.1. Fit a quadratic 2. Here are some data taken from the article “Chemi-
relationship y ≈ β0 + β1 x + β2 x 2 to the data via thermomechanical Pulp from Mixed High Den-
least squares. By appropriately plotting residuals sity Hardwoods” by Miller, Shankar, and Peterson
and examining R 2 values, determine the advis- (Tappi Journal, 1988). Given are the percent NaOH
ability of using a quadratic rather than a linear used as a pretreatment chemical, x 1 , the pretreat-
equation to describe the relationship between x ment time in minutes, x 2 , and the resulting value
and y. If a quadratic fitted equation is used, how of a specific surface area variable, y (with units of
does the predicted mean molecular weight at 200◦ C cm3 /g), for nine batches of pulp produced from a
compare to that obtained in part (e) of the earlier mixture of hardwoods at a treatment temperature
exercise? of 75◦ C in mechanical pulping.
162 Chapter 4 Describing Relationships Between Variables
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
equation can adequately describe the data and have intuitively appealing and under-
standable interpretations. The use of simple plots and residuals will be discussed,
as tools helpful in assessing whether such a simple structure holds.
The discussion begins with the 2-factor case, then considers three (or, by anal-
ogy, more) factors. Finally, the special case where each factor has only two levels is
discussed.
The ȳ i. and ȳ . j are row and column averages when one thinks of the ȳ i j laid out in
a two-dimensional format, as shown in Figure 4.21.
Example 7 Joint Strengths for Three Different Joint Types in Three Different Woods
Kotlers, MacFarland, and Tomlinson studied the tensile strength of three differ-
ent types of joints made on three different types of wood. Butt, lap, and beveled
joints were made in nominal 100 × 400 × 1200 pine, oak, and walnut specimens
using a resin glue. The original intention was to test two specimens of each Joint
Type/Wood Type combination. But one operator error and one specimen failure
not related to its joint removed two of the original data points from consideration
and gave the data in Table 4.11. These data have complete 3 × 3 factorial struc-
164 Chapter 4 Describing Relationships Between Variables
Factor B
Level 1 Level 2 Level J
Factor A
Table 4.12
Sample Means for Nine Wood/Joint Combinations
Wood
ture. Collecting y’s for the nine different combinations into separate samples and
calculating means, the ȳ i j ’s are as presented in tabular form in Table 4.12 and
Interaction plotted in Figure 4.22. This figure is a so-called interaction plot of these means.
Plot The qualitative messages given by the plot are as follows:
1. Joint types ordered by strength are “beveled is stronger than lap, which
in turn is stronger than butt.”
2400
2200
2000
Mean stress at failure, y (psi)
1800
1600
Beveled
Lap
1400
1200 Butt
1000
800
600
Pine Oak Walnut
Wood
The row and column average means ( ȳi· ’s and ȳ· j ’s, respectively) might be
taken as measures of average response behavior at different levels of the factors in
question. If so, it then makes sense to use the differences between these and the
grand average mean ȳ.. as measures of the effects of those levels on mean response.
This leads to Definition 5.
Definition 5 In a two-way complete factorial study with factors A and B, the fitted main
effect of factor A at its ith level is
ai = ȳ i. − ȳ ..
b j = ȳ . j − ȳ ..
Example 7 Simple arithmetic and the ȳ’s in Table 4.12 yield the fitted main effects for the
(continued ) joint strength study of Kotlers, MacFarland, and Tomlinson. First for factor A
(the Joint Type),
These fitted main effects quantify the first two qualitative messages carried by
the data and listed as (1) and (2) before Definition 5. For example,
a2 > a3 > a1
says that beveled joints are strongest and butt joints the weakest. Further, the fact
that the ai ’s and b j ’s are of roughly the same order of magnitude says that the
Joint Type and Wood Type factors are of comparable importance in determining
tensile strength.
A difference between fitted main effects for a factor amounts to a difference be-
tween corresponding row or column averages and quantifies how different response
behavior is for those two levels.
which indicates that pine joint average strength is about 467 psi less than oak
joint average strength.
168 Chapter 4 Describing Relationships Between Variables
In some two-factor factorial studies, the fitted main effects as defined in Defini-
tion 5 pretty much summarize the story told by the means ȳ i j , in the sense that
Display (4.24) implies, for example, that the pattern of mean responses for level 1
of factor A is the same as for level 2 of A. That is, changing levels of factor B (from
say j to j 0 ) produces the same change in mean response for level 2 as for level 1
(namely, b j 0 − b j ). In fact, if relation (4.24) holds, there are parallel traces on an
interaction plot of means.
Example 7 To illustrate the meaning of expression (4.24), the fitted effects for the Joint
(continued ) Type/Wood Type data have been used to calculate 3 × 3 = 9 values of ȳ .. +
ai + b j corresponding to the nine experimental combinations. These are given in
Table 4.13.
For comparison purposes, the ȳ i j from Table 4.12 and the ȳ .. + ai + b j from
Table 4.13 are plotted on the same sets of axes in Figure 4.23. Notice the parallel
traces for the ȳ .. + ai + b j values for the three different joint types. The traces for
yij
2400
Beveled
y.. + ai + bj
2200
2000
Stress at failure (psi)
1800
1600 Lap
1400
Butt
1200
1000
800
600
Pine Oak Walnut
Wood
Table 4.13
Values of ȳ.. + ai + bj for the Joint Strength Study
Wood
the ȳ i j values for the three different joint types are not parallel (particularly when
walnut is considered), so there are apparently substantial differences between the
ȳ i j ’s and the ȳ .. + ai + b j ’s.
When relationship (4.24) fails to hold, the patterns in mean response across
levels of one factor depend on the levels of the second factor. In such cases, the
differences between the combination means ȳ i j and the values ȳ .. + ai + b j can
serve as useful measures of lack of parallelism on the plots of means, and this leads
to another definition.
Definition 6 In a two-way complete factorial study with factors A and B, the fitted inter-
action of factor A at its ith level and factor B at its jth level is
abi j = ȳ i j − ( ȳ .. + ai + b j )
Interpretation of The fitted interactions in some sense measure how much pattern the combination
interactions in a means ȳ i j carry that is not explainable in terms of the factors A and B acting
two-way separately. Clearly, when relationship (4.24) holds, the fitted interactions abi j are all
factorial study small (nearly 0), and system behavior can be thought of as depending separately on
level of A and level of B. In such cases, an important practical consequence is that it
is possible to develop recommendations for levels of the two factors independently
of each other. For example, one need not recommend one level of A if B is at its
level 1 and another if B is at its level 2.
Consider a study of the effects of factors Tool Type and Turning Speed on the
metal removal rate for a lathe. If the fitted interactions are small, turning speed
recommendations that remain valid for all tool types can be made. However, if
the fitted interactions are important, turning speed recommendations might vary
according to tool type.
170 Chapter 4 Describing Relationships Between Variables
Example 7 Again using the Joint Type/Wood Type data, consider calculating the fitted in-
(continued ) teractions. The raw material for these calculations already exists in Tables 4.12
and 4.13. Simply taking differences between entries in these tables cell-by-cell
yields the fitted interactions given in Table 4.14.
It is interesting to compare these fitted interactions to themselves and to
the fitted main effects. The largest (in absolute value) fitted interaction (ab23 )
corresponds to beveled walnut joints. This is consistent with one visual message
in Figures 4.22 and 4.23: This Joint Type/Wood Type combination is in some
sense most responsible for destroying any nearly parallel structure that might
otherwise appear. The fact that (on the whole) the abi j ’s are not as large as the
ai ’s or b j ’s is consistent with a second visual message in Figures 4.22 and 4.23:
The lack of parallelism, while important, is not as important as differences in
Joint Types or Wood Types.
Table 4.14
Fitted Interactions for the Joint Strength Study
Wood
Example 7 has proceeded “by hand.” But using a statistical package can make
the calculations painless. For example, Printout 6 illustrates that most of the results
of Example 7 are readily available in MINITAB’s “General Linear Model” routine
(found under the “Stat/ANOVA/General Linear Model” menu). Comparing this
printout to the example does bring up one point regarding the fitted effects defined
in Definitions 5 and 6. Note that the printout provides values of only two (of three)
Joint main effects, two (of three) Wood main effects, and four (of nine) Joint × Wood
Fitted effects interactions. These are all that are needed, since it is a consequence of Definition 5
sum to zero that fitted main effects for a given factor must total to 0, and it is a consequence of
Definition 6 that fitted interactions must sum to zero across any row or down any
column of the two-way table of factor combinations. The fitted effects not provided
by the printout are easily deduced from the ones that are given.
That is, ȳ.. , the fitted main effects, and the fitted interactions provide a decomposition
or breakdown of the combination sample means into interpretable pieces. These
pieces correspond to an overall effect, the effects of factors acting separately, and
the effects of factors acting jointly.
Taking a hint from the equation fitting done in the previous two sections, it
makes sense to think of (4.25) as a fitted version of an approximate relationship,
y ≈ µ + αi + β j + αβi j (4.26)
y ≈ µ + αi + β j
and there are other simplified versions of equation (4.26) that also have appealing
interpretations. For example, the simplified version of equation (4.26),
y ≈ µ + αi
Residuals e = y − ŷ
(and should look like noise if the simplified equation is an adequate description of
the data set). Further, the fraction of raw variation in y accounted for in the fitting
process is (as always)
P P
Coefficient of (y − ȳ)2 − (y − ŷ)2
determination R2 = P (4.27)
(y − ȳ)2
where the sums are over all observed y’s. (Summation notation is being abused even
further than usual, by not even subscripting the y’s and ŷ’s.)
Table 4.15
Golf Ball Flight Distances for Four Compression/Evening Combinations
Evening (B)
1 2
180 192 196 180
193 190 192 195
80 197 182 191 197
189 192 194 192
187 179 186 193
Compression (A)
180 175 190 185
185 190 195 167
100 167 185 180 180
162 180 170 180
170 185 180 165
174 Chapter 4 Describing Relationships Between Variables
Example 8 These data have complete two-way factorial structure. The factor Evening is
(continued ) not really of primary interest. Rather, it is a blocking factor, its levels creating
homogeneous environments in which to compare 80 and 100 compression flight
distances. Figure 4.24 is a graphic using boxplots to represent the four samples
and emphasizing the factorial structure.
Calculating sample means corresponding to the four cells in Table 4.15 and
then finding fitted effects is straightforward. Table 4.16 displays cell, row, column,
and grand average means. And based on those values,
80 Compression
Flight distance (yd)
190
100 Compression
180
170
1 2
Evening
Table 4.16
Cell, Row, Column, and Grand Average Means for the Golf Ball Flight Data
Evening (B)
1 2
80 ȳ 11 = 188.1 ȳ 12 = 191.6 189.85
Compression (A)
100 ȳ 21 = 177.9 ȳ 22 = 179.2 178.55
183.00 185.40 184.20
4.3 Fitted Effects for Factorial Data 175
180
100 Compression
1 2
Evening
The fitted effects indicate that most of the differences in the cell means in Ta-
ble 4.16 are understandable in terms of differences between 80 and 100 compres-
sion balls. The effect of differences between evenings appears to be on the order
of one-fourth the size of the effect of differences between ball compressions.
Further, the pattern of flight distances across the two compressions changed rela-
tively little from evening to evening. These facts are portrayed graphically in the
interaction plot of Figure 4.25.
The story told by the fitted effects in this example probably agrees with most
readers’ intuition. There is little reason a priori to expect the relative behaviors of
80 and 100 compression flight distances to change much from evening to evening.
But there is slightly more reason to expect the distances to be longer overall on
some nights than on others.
It is worth investigating whether the data in Table 4.15 allow the simplest
To do so, fitted responses are first calculated corresponding to the three different
possible corresponding relationships
y ≈ µ + αi (4.28)
y ≈ µ + αi + β j (4.29)
y ≈ µ + αi + β j + αβi j (4.30)
176 Chapter 4 Describing Relationships Between Variables
These are generated using the fitted effects. They are collected in Table 4.17
(not surprisingly, the first and third sets of fitted responses are, respectively, row
average and cell means).
Residuals e = y − ŷ for fitting the three equations (4.28), (4.29), and (4.30)
are obtained by subtracting the appropriate entries in, respectively, the third,
fourth, or fifth column of Table 4.17 from each of the data values listed in
Table 4.15. For example, 40 residuals for the fitting of the “A main effects only”
equation (4.28) would be obtained by subtracting 189.85 from every entry in the
upper left cell of Table 4.15, subtracting 178.55 from every entry in the lower
left cell, 189.85 from every entry in the upper right cell, and 178.55 from every
entry in the lower right cell.
Figure 4.26 provides normal plots of the residuals from the fitting of the three
equations (4.28), (4.29), and (4.30). None of the normal plots is especially linear,
but at the same time, none of them is grossly nonlinear either. In particular, the
first two, corresponding to simplified versions of relationship 4.26, are not signif-
icantly worse than the last one, which corresponds to the use of all fitted effects
(both main effects and interactions). From the limited viewpoint of producing
residuals with an approximately bell-shaped distribution, the fitting of any of the
three equations (4.28), (4.29), and (4.30) would appear approximately equally
effective.
The calculation of R 2 values for equations (4.28), (4.29), and (4.30) proceeds
as follows. First, since the grand average of all 40 flight distances is ȳ = 184.2
yards (which in this case also turns out to be ȳ .. ) ,
X
(y − ȳ)2 = (180 − 184.2)2 + · · · + (179 − 184.2)2
0 0 0
Figure 4.26 Normal plots of residuals from three different equations fitted to the golf data
values for the three equations are obtained as the sums of the squared residuals.
For example, using Tables 4.15 and 4.17, for equation (4.29),
X
(y − ŷ)2 = (180 − 188.65)2 + · · · + (179 − 188.65)2
Finally, equation (4.27) is used. Table 4.18 gives the three values of R 2 .
The story told by the R 2 values is consistent with everything else that’s been
said in this example. None of the values is terribly big, which is consistent with
the large within-sample variation in flight distances evident in Figure 4.24. But
Table 4.18
R2 Values for Fitting Equations
(4.28), (4.29), and (4.30) to
Gronberg’s Data
Equation R2
y ≈ µ + αi .366
y ≈ µ + αi + β j .382
y ≈ µ + αi + β j + αβi j .386
178 Chapter 4 Describing Relationships Between Variables
Example 8 considering A (Compression) main effects does account for some of the observed
(continued ) variation in flight distance, and the addition of B (Evening) main effects adds
slightly to the variation accounted for. Introducing interactions into consideration
adds little additional accounting power.
Notation for sample ȳ i jk = the sample mean response when factor A is at level i,
means and their factor B is at level j, and factor C is at level k
averages (for three-way 1 X
factorial data) ȳ ... = ȳ
IJK i, j,k i jk
1 X
ȳ i j. = ȳ
K k i jk
Factor A level
K
l
ve
le
C
2
or
ct
1
Fa
2
1
1 2 J
Factor B level
Table 4.19
Levels of Three Process Variables in a 23 Study of Material Strength
Table 4.20
Sample Mean Strengths for 23 Treatment Combinations
ȳ i j k ,
i, j, k, Sample Mean
Factor A Level Factor B Level Factor C Level Strength (psi)
1 1 1 1520
2 1 1 2450
1 2 1 2340
2 2 1 2900
1 1 2 1670
2 1 2 2540
1 2 2 2230
2 2 2 3230
4.3 Fitted Effects for Factorial Data 181
Factor A level
y112 = 1670 y122 = 2230
le 2
l
ve
C
1 y111 = 1520 y121 = 2340
or
1
ct
Fa
1 2
Factor B level
For example,
1
ȳ 1.. = (1520 + 2340 + 1670 + 2230) = 1940 psi
2·2
is the average mean on the bottom face, while
1
ȳ 11. = (1520 + 1670) = 1595 psi
2
is the average mean on the lower left edge. For future reference, all of the average
sample means are collected here:
Definition 7 In a three-way complete factorial study with factors A, B, and C, the fitted
main effect of factor A at its ith level is
ai = ȳ i.. − ȳ ...
b j = ȳ . j. − ȳ ...
ck = ȳ ..k − ȳ ...
Definition 8 In a three-way complete factorial study with factors A, B, and C, the fitted
2-factor interaction of factor A at its ith level and factor B at its jth level is
abi j = ȳ i j. − ( ȳ ... + ai + b j )
the fitted 2-factor interaction of factor A at its ith level and factor C at its
kth level is
and the fitted 2-factor interaction of factor B at its jth level and factor C at
its kth level is
bc jk = ȳ . jk − ( ȳ ... + b j + ck )
4.3 Fitted Effects for Factorial Data 183
Interpreting two- These fitted 2-factor interactions can be thought of in two equivalent ways:
way interactions
1. as what one gets as fitted interactions upon averaging across all levels of
in a three-way study
the factor that is not under consideration to obtain a single two-way table of
(average) means and then calculating as per Definition 6 (page 169);
2. as what one gets as averages, across all levels of the factor not under consid-
eration, of the fitted two-factor interactions calculated as per Definition 6,
one level of the excluded factor at a time.
Example 9 To illustrate the meaning of Definitions 7 and 8, return to the composite material
(continued ) strength study. For example, the fitted A main effects are
The entire set of fitted effects for the means of Table 4.20 is as follows.
Remember equation (4.25) (page 171). It says that in 2-factor studies, the fitted
grand mean, main effects, and two-factor interactions completely describe a factorial
set of sample means. Such is not the case in three-factor studies. Instead, a new pos-
Interpretation of sibility arises: 3-factor interaction. Roughly speaking, the fitted three-factor interac-
three-way interactions tions in a 3-factor study measure how much pattern the combination means carry that
is not explainable in terms of the factors A, B, and C acting separately and in pairs.
Definition 9 In a three-way complete factorial study with factors A, B, and C, the fitted
3-factor interaction of A at its ith level, B at its jth level, and C at its kth
level is
Example 9 To illustrate the meaning of Definition 9, consider again the composite ma-
(continued ) terial study. Using the previously calculated fitted main effects and 2-factor
interactions,
Similar calculations can be made to verify that the entire set of 3-factor interac-
tions for the means of Table 4.20 is as follows:
Main effects and 2-factor interactions are more easily interpreted than 3-factor
interactions. One insight into their meaning was given immediately before Defi-
A second nition 9. Another is the following. If at the different levels of (say) factor C, the
interpretation fitted AB interactions are calculated and the fitted AB interactions (the pattern of
of three-way parallelism or nonparallelism) are essentially the same on all levels of C, then the
interactions 3-factor interactions are small (near 0). Otherwise, large 3-factor interactions allow
the pattern of AB interaction to change, from one level of C to another.
When beginning the analysis of three-way factorial data, one hopes to discover
a simplified version of equation (4.32) that is both interpretable and an adequate
description of the data. (Indeed, if it is not possible to do so, little is gained by
using the factorial breakdown rather than simply treating the data in question as IJK
unstructured samples.)
As was the case earlier with two-way factorial data, the process of fitting a
simplified version of display (4.32) via least squares is, in general, unfortunately
somewhat complicated. But when all sample sizes are equal (i.e., the data are
4.3 Fitted Effects for Factorial Data 185
Example 9 Looking over the magnitudes of the fitted effects for Kinzer’s composite material
(continued ) strength study, the A and B main effects clearly dwarf the others, suggesting the
possibility that the relationship
y ≈ µ + αi + β j (4.33)
Fitted y values, y
2465 3095
2 2465 3095
Factor A
1625 2255
2
C
or
ct
Fa
1 1625 2255
1
1 Factor B 2
Example 9 All eight fitted values corresponding to equation (4.33) are shown geometrically
(continued ) in Figure 4.29. The fitted values given in the figure might be combined with
product requirements and cost information to allow a process engineer to make
sound decisions about autoclave temperature, autoclave time, and time span.
2 26.0 24.0
Plane design
22.5 21.6
rw L 2
t t
ei igh
gh
1 15.8 18.4
H 1
Pa y
v
pe
ea
1 2
Large Small
Plane size
Table 4.21
Shorthand Names for the 23 Factorial Treatment
Combinations
their high levels. For example, if level 2 of each of factors A, B, and C is designated
the high level, shorthand names for the 23 = 8 different ABC combinations are as
given in Table 4.21. Using these names, for example, ȳ a can stand for a sample mean
where factor A is at its high (or second) level and all other factors are at their low
(or first) levels.
Special relationship A second convenience special to two-level factorial data structures is the fact
between 2p effects that all effects of a given type have the same absolute value. This has already been
of a given type illustrated in Example 9. For example, looking back, for the data of Table 4.20,
This is always the case for fitted effects in 2 p factorials. In fact, if two fitted
effects of the same type are such that an even number of 1 → 2 or 2 → 1 subscript
changes are required to get the second from the first, the fitted effects are equal
(e.g., bc22 = bc11 ). If an odd number are required, then the second fitted effect is
−1 times the first (e.g., bc12 = −bc22 ). This fact is so useful because one needs only
to do the arithmetic necessary to find one fitted effect of each type and then choose
appropriate signs to get all others of that type.
A statistician named Frank Yates is credited with discovering an efficient,
mechanical way of generating one fitted effect of each type for a 2 p study. His
method is easy to implement “by hand” and produces fitted effects with all “2”
subscripts (i.e., corresponding to the “all factors at their high level” combination).
The Yates algorithm The Yates algorithm consists of the following steps.
for computing fitted
2p factorial effects Step 1 Write down the 2 p sample means in a column in what is called Yates
standard order. Standard order is easily remembered by beginning
4.3 Fitted Effects for Factorial Data 189
The last column (made via step 3) gives fitted effects (all factors at level 2), again
in standard order.
Example 9 Table 4.22 shows the use of the Yates algorithm to calculate fitted effects for the
(continued ) 23 composite material study. The entries in the final column of this table are, of
course, exactly as listed earlier, and the rest of the fitted effects are easily obtained
via appropriate sign changes. This final column is an extremely concise summary
of the fitted effects, which quickly reveals which types of fitted effects are larger
than others.
Table 4.22
The Yates Algorithm Applied to the Means of Table 4.20
The Yates algorithm is useful beyond finding fitted effects. For balanced data
sets, it is also possible to modify it slightly to find fitted responses, ŷ, correspond-
ing to a simplified version of a relation like display (4.32). First, the desired (all
factors at their high level) fitted effects (using 0’s for those types not considered)
The reverse Yates are written down in reverse standard order. Then, by applying p cycles of the
algorithm and easy Yates additions and subtractions, the fitted values, ŷ, are obtained, listed in re-
computation of fitted verse standard order. (Note that no final division is required in this reverse Yates
responses algorithm.)
190 Chapter 4 Describing Relationships Between Variables
Example 9 Consider fitting the relationship (4.33) to the balanced data set that led to the
(continued ) means of Table 4.20 via the reverse Yates algorithm. Table 4.23 gives the details.
The fitted values in the final column are exactly as shown earlier in Figure 4.29.
Table 4.23
The Reverse Yates Algorithm Applied to Fitting the "A and B
Main Effects Only" Equation (4.33) to the Data of Table 4.20
The restriction to two-level factors that makes these notational and computa-
tional devices possible is not as specialized as it may at first seem. When an engineer
wishes to study the effects of a large number of factors, even 2 p will be a large num-
The importance ber of conditions to investigate. If more than two levels of factors are considered,
of two-level the sheer size of a complete factorial study quickly becomes unmanageable. Rec-
factorials ognizing this, two-level studies are often used for screening to identify a few (from
many) process variables for subsequent study at more levels on the basis of their
large perceived effects in the screening study. So this 2 p material is in fact quite
important to the practice of engineering statistics.
Section 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Since the data of Exercise 2 of Section 4.2 have Plot these versus level of the NaOH variable,
complete factorial structure, it is possible (at least connecting fitted values having the same level
temporarily) to ignore the fact that the two experi- of the Time variable with line segments, as in
mental factors are basically quantitative and make Figure 4.23. Discuss how this plot compares
a factorial analysis of the data. to the two plots of fitted y versus x 1 made in
(a) Compute all fitted factorial main effects and in- Exercise 2 of Section 4.2.
teractions for the data of Exercise 2 of Section (c) Use the fitted values computed in (b) and find
4.2. Interpret the relative sizes of these fitted ef- a value of R 2 appropriate to the “main effects
fects, using a interaction plot like Figure 4.22 only” representation of y. How does it com-
to facilitate your discussion. pare to the R 2 values from multiple regres-
(b) Compute nine fitted responses for the “main ef- sions? Also use the fitted values to compute
fects only” explanation of y, y ≈ µ + αi + β j .
4.4 Transformations and Choice of Measurement Scale 191
residuals for this “main effects only” represen- (b) The students actually had some physical the-
tation. Plot these (versus level of NaOH, level ory suggesting that the log of the drain time
of Time, and ŷ, and in normal plot form). What might be a more convenient response variable
do they indicate about the present “no interac- than the raw time. Take the logs of the y’s and
tion” explanation of specific area? recompute the factorial effects. Does an inter-
2. Bachman, Herzberg, and Rich conducted a 23 fac- pretation of this system in terms of only main
torial study of fluid flow through thin tubes. They effects seem more plausible on the log scale
measured the time required for the liquid level in than on the original scale?
a fluid holding tank to drop from 4 in. to 2 in. for (c) Considering the logged drain times as the re-
two drain tube diameters and two fluid types. Two sponses, find fitted values and residuals for a
different technicians did the measuring. Their data “Diameter and Fluid main effects only” expla-
are as follows: nation of these data. Compute R 2 appropriate
to such a view and compare it to R 2 that re-
Diameter sults from using all factorial effects to describe
Technician (in.) Fluid Time (sec)
log drain time. Make and interpret appropriate
residual plots.
1 .188 water 21.12, 21.11, 20.80 (d) Based on the analysis from (c), what change in
2 .188 water 21.82, 21.87, 21.78 log drain time seems to accompany a change
1 .314 water 6.06, 6.04, 5.92 from .188 in. diameter to .314 in. diameter?
2 .314 water 6.09, 5.91, 6.01 What does this translate to in terms of raw drain
1 .188 ethylene glycol 51.25, 46.03, 46.09 time? Physical theory suggests that raw time is
2 .188 ethylene glycol 45.61, 47.00, 50.71
inversely proportional to the fourth power of
drain tube radius. Does your answer here seem
1 .314 ethylene glycol 7.85, 7.91, 7.97
compatible with that theory? Why or why not?
2 .314 ethylene glycol 7.73, 8.01, 8.32
3. When analyzing a full factorial data set where the
(a) Compute (using the Yates algorithm or other- factors involved are quantitative, either the surface-
wise) the values of all the fitted main effects, fitting technology of Section 4.2 or the factorial
two-way interactions, and three-way interac- analysis material of Section 4.3 can be applied.
tions for these data. Do any simple interpreta- What practical engineering advantage does the first
tions of these suggest themselves? offer over the second in such cases?
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
0 4 3
0 6 5 5 5 6 6 8 9 8 8
1 4 0 3 0
1 7 5 9 5 6 9 5
2 0
2 9 5 7 9
3 2
3 6
0 0
–1.0 –1.0
–2.0 –2.0
Figure 4.32 Normal plots for discovery times and log discovery times
Power
g(y) = (y − γ )α (4.34)
transformations
Where several samples (and corresponding ȳ and s values) are involved, an empirical
way of investigating whether (1) or (2) above might be useful is to plot ln(s) versus
ln( ȳ) and see if there is approximate linearity. If so, a slope of roughly 1 makes (1)
appropriate, while a slope of δ 6= 1 signals what version of (2) might be helpful.
In addition to this empirical way of identifying a potentially variance-stabilizing
transformation, theoretical considerations can sometimes provide guidance. Stan-
dard theoretical distributions (like those introduced in Chapter 5) have their own
relationships between their (theoretical) means and variances, which can help pick
out an appropriate version of (1) or (2) above.
A power law β β β
y ≈ αx1 1 x2 2 · · · xk k (4.35)
ȳ .... = 6.1550
a2 = .4563 b2 = 1.6488 c2 = 3.2163 d2 = 1.1425
ab22 = .0750 ac22 = .2975 ad22 = .4213
bc22 = .7525 bd22 = .2213 cd22 = .7987
abc222 = .0838 abd222 = .2950 acd222 = .3775 bcd222 = .0900
abcd2222 = .2688
y ≈ µ + β j + γk + δl (4.37)
and
y ≈ µ + β j + γk + δl + βγ jk + γ δkl (4.38)
Table 4.24
Daniel’s 24 Drill Advance Rate Data
Combination y Combination y
Example 12 are suggested. (The five largest fitted effects are, in order of decreasing magnitude,
(continued ) the main effects of C, B, and D, and then the two-factor interactions of C with D
and B with C.) Fitting equation (4.37) to the balanced data of Table 4.24 produces
R 2 = .875, and fitting relationship (4.38) produces R 2 = .948. But upon closer
examination, neither fitted equation turns out to be a very good description of
these data.
Figure 4.33 shows a normal plot and a plot against ŷ for residuals from
a fitted version of equation (4.37). It shows that the fitted version of equation
(4.37) produces several disturbingly large residuals and fitted values that are
systematically too small for responses that are small and large, but too large for
moderate responses. Such a curved plot of residuals versus ŷ in general suggests
that a nonlinear transformation of y may potentially be effective.
The reader is invited to verify that residual plots for equation (4.38) look even
worse than those in Figure 4.33. In particular, it is the bigger responses that are
2.0
Standard normal quantile
1.0
–1.0
–2.0
–1.0 0.0 1.0 2.0 3.0 4.0
Residual quantile
4.0
3.0
Residual
2.0
1.0
0
–1.0
–2.0
2.0 4.0 6.0 8.0 10.0 12.0
Fitted response, y
2.0
–1.0
–2.0
–.2 –.1 0 .1 .2
Residual quantile
.2
.1
Residual
–.1
–.2
1.0 2.0
Fitted response, ln( y)
y 0 .... = 1.5977
a2 = .0650 b2 = .2900 c2 = .5772 d2 = .1633
ab22 = −.0172 ac22 = .0052 ad22 = .0334
bc22 = −.0251 bd22 = −.0075 cd22 = .0491
abc222 = .0052 abd222 = .0261 acd222 = .0266 bcd222 = −.0173
abcd2222 = .0193
198 Chapter 4 Describing Relationships Between Variables
Example 12 For the logged drill advance rates, the simple relationship
(continued )
ln(y) ≈ µ + β j + γk + δl (4.39)
y2 ≈ β0 + β1 x1 + β2 x2 + β3 x12 + β4 x22 + β5 x1 x2
Table 4.25
Yields and Filtration Times in a 32 Factorial Chemical
Process Study
x1 , x2 , y1 , y2 ,
Condensation Amount Yield Filtration
Temperature (◦ C) of B (cc) (g) Time (sec)
90 24.4 21.1 150
90 29.3 23.7 10
90 34.2 20.7 8
100 24.4 21.1 35
100 29.3 24.1 8
100 34.2 22.2 7
110 24.4 18.4 18
110 29.3 23.4 8
110 34.2 21.9 10
(4.40) rates as a statistical engineering success story. But there is the embarrass-
ing fact that upon substituting x 1 = 103.2 and x2 = 30.88 into equation (4.40),
one gets ŷ 2 = −11 sec, hardly a possible filtration time.
Looking again at the data, it is not hard to see what has gone wrong. The
largest response is more than 20 times the smallest. So in order to come close to
fitting both the extremely large and more moderate responses, the fitted quadratic
surface needs to be very steep—so steep that it is forced to dip below the (x1 , x2 )-
plane and produce negative ŷ 2 values before it can “get turned around” and start
to climb again as it moves away from the point of minimum ŷ 2 toward larger x1
and x2 .
One cure for the problem of negative predicted filtration times is to use ln(y2 )
as a response variable. Values of ln(y2 ) are given in Table 4.26 to illustrate the
moderating effect the logarithm has on the factor of 20 disparity between the
largest and smallest filtration times.
Fitting the approximate quadratic relationship
y2 , ln(y2 ),
Filtration Time (sec) Log Filtration Time (ln(sec))
150 5.0106
10 2.3026
8 2.0794
35 3.5553
8 2.0794
7 1.9459
18 2.8904
8 2.0794
10 2.3026
The taking of logs in this example had two beneficial effects. The first was to
cut the ratio of largest response to smallest down to about 2.5 (from over 20), al-
lowing a good fit (as measured by R 2 ) for a fitted quadratic in two variables, x 1 and
x2 . The second was to ensure that minimum predicted filtration time was positive.
Of course, other transformations besides the logarithmic one are also useful in
describing the structure of multifactor data sets. Sometimes they are applied to the
responses and sometimes to other system variables. As an example of a situation
where a power transformation like that specified by equation (4.34) is useful in
understanding the structure of a sample of bivariate data, consider the following.
Table 4.27
Average Grain Diameters and Yield Strengths for Copper Deposits
350
300
250
200
.2 .4 .6 .8
Average grain diameter ( µm)
350
300
250
200
1 1.2 1.4 1.6 1.8 2.0 2.2
Reciprocal square root grain diameter
Section 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. What are benefits that can sometimes be derived will there be important interactions? (In order to
from transforming data before applying standard make this concrete, you may if you wish consider
statistical techniques? the relationship y ≈ kx12 x2−3 . Plot, for at least two
2. Suppose that a response variable, y, obeys an ap- different values of x 2 , y as a function of x1 . Then
proximate power law in at least two quantitative plot, for at least two different values of x 2 , ln(y) as
variables (say, x1 and x2 ). Will there be important a function of x1 . What do these plots show in the
interactions? If the log of y is analyzed instead, way of parallelism?)
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Section 5 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Read again Section 1.4 and the present one. Then Give an example of a deterministic model that is
describe in your own words the difference between useful in your field.
deterministic and stochastic/probabilistic models.
Chapter 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Nicholson and Bartle studied the effect of the wa- (b) Compute the sample correlation between x and
ter/cement ratio on 14-day compressive strength y by hand. Interpret this value.
for Portland cement concrete. The water/cement (c) What fraction of the raw variability in y is
ratios (by volume) and compressive strengths of accounted for in the fitting of a line to the data?
nine concrete specimens are given next. (d) Compute the residuals from your fitted line and
make a normal plot of them. Interpret this plot.
Water/Cement 14-Day Compressive (e) What compressive strength would you predict,
Ratio, x Strength, y (psi) based on your calculations from (a), for speci-
mens made using a .48 water/cement ratio?
.45 2954, 2913, 2923 (f) Use a statistical package to find the least
.50 2743, 2779, 2739 squares line, the sample correlation, R 2 , and
.55 2652, 2607, 2583 the residuals for this data set.
2. Griffith and Tesdall studied the elapsed time in 14
(a) Fit a line to the data here via least squares, mile runs of a Camaro Z-28 fitted with different
showing the hand calculations.
204 Chapter 4 Describing Relationships Between Variables
sizes of carburetor jetting. Their data from six runs (a) What type of data structure did the researchers
of the car follow: employ? (Use the terminology of Section 1.2.)
What was an obvious weakness in their data
Jetting Size, x Elapsed Time, y (sec) collection plan?
(b) Use a regression program to fit the following
66 14.90 equations to these data:
68 14.67
70 14.50 y ≈ β 0 + β1 x 1 + β2 x 2
72 14.53
y ≈ β0 + β1 x1 + β2 ln(x2 )
74 14.79
76 15.02 y ≈ β0 + β1 x1 + β2 ln(x2 ) + β3 x1 ln(x2 )
(a) What is an obvious weakness in the students’ What are the R 2 values for the three differ-
data collection plan? ent fitted equations? Compare the three fitted
(b) Fit both a line and a quadratic equation (y ≈ equations in terms of complexity and apparent
β0 + β1 x + β2 x 2 ) to these data via least ability to predict y.
squares. Plot both of these equations on a scat- (c) Compute the residuals for the third fitted equa-
terplot of the data. tion in (b). Plot them against x1 , x2 , and ŷ.
(c) What fractions of the raw variation in elapsed Also normal-plot them. Do any of these plots
time are accounted for by the two different suggest that the third fitted equation is inade-
fitted equations? quate as summary of these data? What, if any,
(d) Use your fitted quadratic equation to predict an possible improvement over the third equation
optimal jetting size (allowing fractional sizes). is suggested by these plots?
3. The following are some data taken from “Kinet- (d) As a means of understanding the nature of the
ics of Grain Growth in Powder-formed IN-792: A third fitted equation in (b), make a scatterplot
Nickel-Base Super-alloy” by Huda and Ralph (Ma- of y vs. x2 using a logarithmic scale for x2 . On
terials Characterization, September 1990). Three this plot, plot three lines representing ŷ as a
different Temperatures, x1 (◦ K), and three different function of x2 for the three different values of
Times, x2 (min), were used in the heat treating of x1 . Qualitatively, how would a similar plot for
specimens of a material, and the response the second equation differ from this one?
(e) Using the third equation in (b), what mean
y = mean grain diameter (µm) grain diameter would you predict for x 1 =
1500 and x2 = 500?
was measured. (f) It is possible to ignore the fact that the Tem-
perature and Time factors are quantitative and
Temperature, x1 Time, x2 Grain Size, y make a factorial analysis of these data. Do so.
Begin by making an interaction plot similar
1443 20 5 to Figure 4.22 for these data. Based on that
1443 120 6 plot, discuss the apparent relative sizes of the
1443 1320 9 Time and Temperature main effects and the
1493 20 14 Time × Temperature interactions. Then com-
1493 120 17 pute the fitted factorial effects (the fitted main
1493 1320 25
effects and interactions).
1543 20 29 4. The article “Cyanoacetamide Accelerators for the
1543 120 38 Epoxide/Isocyanate Reaction” by Eldin and Ren-
1543 1320 60
ner (Journal of Applied Polymer Science, 1990)
Chapter 4 Exercises 205
reports the results of a 23 factorial experiment. Us- values of ŷ and compute R 2 for the “A main ef-
ing cyanoacetamides as catalysts for an epoxy/iso- fects only” description of impact strength. (The
cyanate reaction, various mechanical properties of formula in Definition 3 works in this context
a resulting polymer were studied. One of these was as well as in regression.)
(d) Now recognize that the experimental factors
y = impact strength (kJ/mm2 ) here are quantitative, so methods of curve and
surface fitting may be applicable. Fit the equa-
The three experimental factors employed and their tion y ≈ β0 + β1 (epoxy/isocyanate ratio) to
corresponding experimental levels were as follows: the data. What eight values of ŷ and value
of R 2 accompany this fit?
Factor A Initial Epoxy/Isocyanate Ratio
0.4 (−) vs. 1.2 (+) 5. Timp and M-Sidek studied the strength of mechan-
ical pencil lead. They taped pieces of lead to a desk,
Factor B Flexibilizer Concentration with various lengths protruding over the edge of the
10 mol % (−) vs. 40 mol % (+) desk. After fitting a small piece of tape on the free
Factor C Accelerator Concentration end of a lead piece to act as a stop, they loaded it
1/240 mol % (−) vs. 1/30 mol% (+) with paper clips until failure. In one part of their
(The flexibilizer and accelerator concentrations are study, they tested leads of two different Diame-
relative to the amount of epoxy present initially.) ters, used two different Lengths protruding over
The impact strength data obtained (one observation the edge of the desk, and tested two different lead
per combination of levels of the three factors) were Hardnesses. That is, they ran a 23 factorial study.
as follows: Their factors and levels were as follows:
Factor A Diameter .3 mm (−) vs. .7 mm (+)
Combination y Combination y
Factor B Length Protruding 3 cm (−) vs.
(1) 6.7 c 6.3 4.5 cm (+)
a 11.9 ac 15.1 Factor C Hardness B (−) vs. 2H (+)
b 8.5 bc 6.7 and m = 2 trials were made at each of the 23 = 8
ab 16.5 abc 16.4 different sets of conditions. The data the students
obtained are given here.
(a) What is an obvious weakness in the researchers’
data collection plan?
Combination Number of Clips
(b) Use the Yates algorithm and compute fitted fac-
torial effects corresponding to the “all high” (1) 13, 13
treatment combination (i.e., compute ȳ ... , a2 , a 74, 76
b2 , etc.). Interpret these in the context of the b 9, 10
original study. (Describe in words which fac-
ab 43, 42
tors and/or combinations of factors appear to
c 16, 15
have the largest effect(s) on impact strength
ac 89, 88
and interpret the sign or signs.)
(c) Suppose only factor A is judged to be of im- bc 10, 12
portance in determining impact strength. What abc 54, 55
predicted/fitted impact strengths correspond to
this judgment? (Find ŷ values using the reverse (a) It appears that analysis of these data in terms
Yates algorithm or otherwise.) Use these eight of the natural logarithms of the numbers of
206 Chapter 4 Describing Relationships Between Variables
clips first causing failure is more straightfor- (a) Compute the fitted 23 factorial effects (main
ward than the analysis of the raw numbers of effects, 2-factor interactions and 3-factor inter-
clips. So take natural logs and compute the fit- actions) corresponding to the following set of
ted 23 factorial effects. Interpret these. In par- conditions: 60 mesh, 500 cc, vibrated cylinder.
ticular, what (in quantitative terms) does the (b) If your arithmetic for part (a) is correct, you
size of the fitted A main effect say about lead should have found that the largest of the fitted
strength? Does lead hardness appear to play effects (in absolute value) are (respectively)
a dominant role in determining this kind of the C main effect, the A main effect, and then
breaking strength? the AC 2-factor interaction. (The next largest
(b) Suppose only the main effects of Diameter are fitted effect is only about half of the smallest
judged to be of importance in determining lead of these, the AC interaction.) Now, suppose
strength. Find a predicted log breaking strength you judge these three fitted effects to summa-
for .7 mm, 2H lead when the length protruding rize the main features of the data set. Interpret
is 4.5 cm. Use this to predict the number of this data summary (A and C main effects and
clips required to break such a piece of lead. AC interactions) in the context of this 3-factor
(c) What, if any, engineering reasons do you have study.
for expecting the analysis of breaking strength (c) Using your fitted effects from (a) and the data
to be more straightforward on the log scale than summary from (b) (A and C main effects and
on the original scale? AC interactions), what fitted response would
6. Ceramic engineering researchers Leigh and Taylor, you have for these conditions: 60 mesh, 500
in their paper “Computer Generated Experimen- cc, vibrated cylinder?
tal Designs” (Ceramic Bulletin, 1990), studied the (d) Using your fitted effects from (a), what average
packing properties of crushed T-61 tabular alumina change in density would you say accompanies
powder. The densities of batches of the material the vibration of the graduated cylinder before
were measured under a total of eight different sets density determination?
of conditions having a 23 factorial structure. The 7. The article “An Analysis of Transformations” by
following factors and levels were employed in the Box and Cox (Journal of the Royal Statistical So-
study: ciety, Series B, 1964) contains a classical unrepli-
Factor A Mesh Size of Powder Particles cated 33 factorial data set originally taken from an
6 mesh (−) vs. 60 mesh (+) unpublished technical report of Barella and Sust.
These researchers studied the behavior of worsted
Factor B Volume of Graduated Cylinder yarns under repeated loading. The response vari-
100 cc (−) vs. 500 cc (+) able was
Factor C Vibration of Cylinder
no (−) vs. yes (+) y = the numbers of cycles till failure
The mean densities (in g/cc) obtained in m = 5
for specimens tested with various values of
determinations for each set of conditions were as
follows:
x 1 = length (mm)
ȳ (1) = 2.348 ȳ a = 2.080
x2 = amplitude of the loading cycle (mm)
ȳ b = 2.298 ȳ ab = 1.980
ȳ c = 2.354 ȳ ac = 2.314 x3 = load (g)
ȳ bc = 2.404 ȳ abc = 2.374
Chapter 4 Exercises 207
The researchers’ data are given in the accompany- to the data. What fraction of the observed vari-
ing table. ability in y = ln(y) does this equation account
for? What change in y 0 seems to accompany a
x1 x2 x3 y x1 x2 x3 y unit (a 1 ln(g)) increase in x30 ?
(c) To carry the analysis one step further, note that
250 8 40 674 300 9 50 438 your fitted coefficients for x10 and x20 are nearly
250 8 45 370 300 10 40 442 the negatives of each other. That suggests that
250 8 50 292 300 10 45 332 y 0 depends only on the difference between x10
250 9 40 338 300 10 50 220 and x20 . To see how this works, fit the equation
250 9 45 266 350 8 40 3,636
250 9 50 210 350 8 45 3,184 y 0 ≈ β0 + β1 (x10 − x20 ) + β2 x30
250 10 40 170 350 8 50 2,000
to the data. Compute and plot residuals from
250 10 45 118 350 9 40 1,568
this relationship (still on the log scale). How
250 10 50 90 350 9 45 1,070
does this relationship appear to do as a data
300 8 40 1,414 350 9 50 566
summary? What power law for y (on the orig-
300 8 45 1,198 350 10 40 1,140 inal scale) in terms of x1 , x2 , and x3 (on their
300 8 50 634 350 10 45 884 original scales) is implied by this last fitted
300 9 40 1,022 350 10 50 360 equation? How does this equation compare to
300 9 45 620 the one from (a) in terms of parsimony?
(d) Use your equation from (c) to predict the life
(a) To find an equation to represent these data, of an additional specimen of length 300 mm, at
you might first try to fit multivariable polyno- an amplitude of 9 mm, under a load of 45 g. Do
mials. Use a regression program and fit a full the same for an additional specimen of length
quadratic equation to these data. That is, fit 325 mm, at an amplitude of 9.5 mm, under
a load of 47 g. Why would or wouldn’t you
y ≈ β0 + β1 x1 + β2 x2 + β3 x3 + β4 x12 + β5 x22 be willing to make a similar projection for an
additional specimen of length 375 mm, at an
+β6 x32 + β7 x1 x2 + β8 x1 x3 + β9 x2 x3
amplitude of 10.5 mm, under a load of 51 g?
to the data. What fraction of the observed vari- 8. Bauer, Dirks, Palkovic, and Wittmer fired tennis
ation in y does it account for? In terms of par- balls out of a “Polish cannon” inclined at an angle
simony (or providing a simple data summary), of 45◦, using three different Propellants and two
how does this quadratic equation do as a data different Charge Sizes of propellant. They observed
summary? the distances traveled in the air by the tennis balls.
(b) Notice the huge range of values of the response Their data are given in the accompanying table.
variable. In cases like this, where the response (Five trials were made for each Propellant/Charge
varies over an order of magnitude, taking log- Size combination and the values given are in feet.)
arithms of the response often helps produce a
simple fitted equation. Here, take (natural) log-
arithms of all of x 1 , x2 , x3 , and y, producing
(say) x10 , x20 , x30 , and y 0 , and fit the equation
set from the article follows. (The data in Section 13. K. Casali conducted a gas mileage study on his
4.1 are the x2 = .01725 data only.) well-used four-year-old economy car. He drove
a 107-mile course a total of eight different times
Cutting Speed, Feed, (in comparable weather conditions) at four differ-
ent speeds, using two different types of gasoline,
x1 (sfpm) x2 (ipr) Tool Life, y (min)
and ended up with an unreplicated 4 × 2 factorial
800 .01725 1.00, 0.90, 0.74, 0.66 study. His data are given in the table below.
700 .01725 1.00, 1.20, 1.50, 1.60
700 .01570 1.75, 1.85, 2.00, 2.20 Speed Gasoline Gallons Mileage
600 .02200 1.20, 1.50, 1.60, 1.60 Test (mph) Octane Used (mpg)
600 .01725 2.35, 2.65, 3.00, 3.60
1 65 87 3.2 33.4
500 .01725 6.40, 7.80, 9.80, 16.50
2 60 87 3.1 34.5
500 .01570 8.80, 11.00, 11.75, 19.00
3 70 87 3.4 31.5
450 .02200 4.00, 4.70, 5.30, 6.00
4 55 87 3.0 35.7
400 .01725 21.50, 24.50, 26.00, 33.00
5 65 90 3.2 33.4
(a) Taylor’s expanded tool life equation is 6 55 90 2.9 36.9
α α
yx1 1 x2 2 = C. This relationship suggests that 7 70 90 3.3 32.4
ln(y) may well be approximately linear in 8 60 90 3.0 35.7
both ln(x 1 ) and ln(x2 ). Use a multiple linear
regression program to fit the relationship (a) Make a plot of the mileages that is useful for
judging the size of Speed × Octane interac-
ln(y) ≈ β0 + β1 ln(x1 ) + β2 ln(x2 ) tions. Does it look as if the interactions are
large in comparison to the main effects?
to these data. What fraction of the raw vari- (b) Compute the fitted main effects and interac-
ability in ln(y) is accounted for in the fitting tions for the mileages, using the formulas of
process? What estimates of the parameters α1 , Section 4.3. Make a plot like Figure 4.23
α2 , and C follow from your fitted equation? for comparing the observed mileages to fit-
(b) Compute and plot residuals (continuing to ted mileages computed supposing that there
work on log scales) for the equation you fit are no Speed × Octane interactions.
in part (a). Make at least plots of residuals (c) Now fit the equation
versus fitted ln(y) and both ln(x1 ) and ln(x2 ),
and make a normal plot of these residuals. Mileage ≈ β0 + β1 (Speed) + β2 (Octane)
Do these plots reveal any particular problems
with the fitted equation? to the data and plot lines representing the pre-
(c) Use your fitted equation to predict first a log dicted mileages versus Speed for both the 87
tool life and then a tool life, if in this machin- octane and the 90 octane gasolines on the
ing application a cutting speed of 550 and a same set of axes.
feed of .01650 is used. (d) Now fit the equation Mileage ≈ β0 + β1
(d) Plot the ordered pairs appearing in the data (Speed) separately, first to the 87 octane data
set in the (x1 , x2 )-plane. Outline a region in and then to the 90 octane data. Plot the two
the plane where you would feel reasonably different lines on the same set of axes.
safe using the equation you fit in part (a) to (e) Discuss the different appearances of the plots
predict tool life. you made in parts (a) through (d) of this exer-
cise in terms of how well they fit the original
210 Chapter 4 Describing Relationships Between Variables
data and the different natures of the assump- α2 , and C follow from your fitted equation?
tions involved in producing them. Using your estimates of α1 , α2 , and C, plot on
(f) What was the fundamental weakness in the same set of (x1 , y) axes the functional re-
Casali’s data collection scheme? A weakness lationships between x 1 and y implied by your
of secondary importance has to do with the fitted equation for x2 equal to 3,000, 6,000,
fact that tests 1–4 were made ten days ear- and then 10,000 psi, respectively.
lier than tests 5–8. Why is this a potential (b) Compute and plot residuals (continuing to
problem? work on log scales) for the equation you fit
14. The article “Accelerated Testing of Solid Film in part (a). Make at least plots of residuals
Lubricants” by Hopkins and Lavik (Lubrication versus fitted ln(y) and both ln(x1 ) and ln(x2 ),
Engineering, 1972) contains a nice example of and make a normal plot of these residuals.
the engineering use of multiple regression. In the Do these plots reveal any particular problems
study, m = 3 sets of journal bearing tests were with the fitted equation?
made on a Mil-L-8937 type film at each combi- (c) Use your fitted equation to predict first a log
nation of three different Loads and three different wear life and then a wear life, if in this appli-
Speeds. The wear lives of journal bearings, y, cation a speed of 20 rpm and a load of 10,000
in hours, are given next for the tests run by the psi are used.
authors. (d) (Accelerated life testing) As a means of
trying to make intelligent data-based predic-
Speed, Load, tions of wear life at low stress levels (and
x1 (rpm) x2 (psi) Wear Life, y (hs)
correspondingly large lifetimes that would be
impractical to observe directly), you might
20 3,000 300.2, 310.8, 333.0 (fully recognizing the inherent dangers of the
20 6,000 99.6, 136.2, 142.4 practice) try to extrapolate using the fitted
20 10,000 20.2, 28.2, 102.7 equation. Use your fitted equation to predict
60 3,000 67.3, 77.9, 93.9 first a log wear life and then a wear life if
60 6,000 43.0, 44.5, 65.9 a speed of 15 rpm and load of 1,500 psi are
used in this application.
60 10,000 10.7, 34.1, 39.1
100 3,000 26.5, 22.3, 34.8 15. The article “Statistical Methods for Controlling
100 6,000 32.8, 25.6, 32.7 the Brown Oxide Process in Multilayer Board
100 10,000 2.3, 4.4, 5.8
Processing” by S. Imadi (Plating and Surface
Finishing, 1988) discusses an experiment con-
(a) The authors expected to be able to describe ducted to help a circuit board manufacturer mea-
wear life as roughly following the relationship sure the concentration of important components
yx1 x2 = C, but they did not find this relation- in a chemical bath. Various combinations of lev-
ship to be a completely satisfactory model. So els of
instead, they tried using the more general rela-
α α
tionship yx 1 1 x2 2 = C. Use a multiple linear x1 = % by volume of component A (a proprietary
regression program to fit the relationship formulation, the major component of which
is sodium chlorite)
ln(y) ≈ β0 + β1 ln(x1 ) + β2 ln(x2 ) and
to these data. What fraction of the raw vari- x2 = % by volume of component B (a proprietary
ability in ln(y) is accounted for in the fitting formulation, the major component of which
process? What estimates of the parameters α1 , is sodium hydroxide)
Chapter 4 Exercises 211
were set in the chemical bath, and the variables regression program. Is this equation the same
one you found in part (b)?
y1 = ml of 1N H2 SO4 used in the first phase (d) If you were to compare the equations for x2
of a titration derived in (b) and (c) in terms of the sum
of squared differences between the predicted
and
and observed values of x2 , which is guaran-
y2 = ml of 1N H2 SO4 used in the second phase teed to be the winner? Why?
of a titration 16. The article “Nonbloated Burned Clay Aggregate
Concrete” by Martin, Ledbetter, Ahmad, and Brit-
were measured. Part of the original data col- ton (Journal of Materials, 1972) contains data
lected (corresponding to bath conditions free of on both composition and resulting physical prop-
Na2 CO3 ) follow: erty test results for a number of different batches
of concrete made using burned clay aggregates.
x1 x2 y1 y2 The accompanying data are compressive strength
measurements, y (made according to ASTM C 39
15 25 3.3 .4
and recorded in psi), and splitting tensile strength
20 25 3.4 .4 measurements, x (made according to ASTM C
20 30 4.1 .4 496 and recorded in psi), for ten of the batches
25 30 4.3 .3 used in the study.
25 35 5.0 .5
30 35 5.0 .3 Batch 1 2 3 4 5
30 40 5.7 .5 y 1420 1950 2230 3070 3060
35 40 5.8 .4 x 207 233 254 328 325
batch of concrete of this type if you were to polymer concentration, x 2 , on percent recoveries
measure a splitting tensile strength of 245 psi? of pyrite, y1 , and kaolin, y2 , from a step of an ore
(g) Compute the residuals from your fitted line. refining process. (High pyrite recovery and low
Plot the residuals against x and against ŷ. kaolin recovery rates were desirable.) Data from
Then make a normal plot of the residuals. one set of n = 9 experimental runs are given here.
What do these plots indicate about the linear-
ity of the relationship between splitting ten- x1 (rpm) x2 (ppm) y1 (%) y2 (%)
sile strength and compressive strength?
(h) Use a statistical package to find the least 1350 80 77 67
squares line, the sample correlation, R 2 , and 950 80 83 54
the residuals for these data. 600 80 91 70
(i) Fit the quadratic relationship y ≈ β0 + β1 x + 1350 100 80 52
β2 x 2 to the data, using a statistical package. 950 100 87 57
Sketch this fitted parabola on your scatterplot 600 100 87 66
from part (a). Does this fitted quadratic ap- 1350 120 67 54
pear to be an important improvement over the 950 120 80 52
line you fit in (c) in terms of describing the
600 120 81 44
relationship of y to x?
(j) How do the R 2 values from parts (h) and (i) (a) What type of data structure did the researcher
compare? Does the increase in R 2 in part (i) use? (Use the terminology of Section 1.2.)
speak strongly for the use of the quadratic (as What was an obvious weakness in his data
opposed to linear) description of the relation- collection plan?
ship of y to x for concretes of this type? (b) Use a regression program to fit the following
(k) If you use the fitted relationship from part equations to these data:
(i) to predict y for x = 245, how does the
prediction compare to your answer for part y1 ≈ β0 + β1 x1
(f)?
(l) What do the fitted relationships from parts y1 ≈ β0 + β2 x2
(c) and (i) give for predicted compressive y1 ≈ β0 + β1 x1 + β2 x2
strengths when x = 400 psi? Do these com-
pare to each other as well as your answers to
What are the R 2 values for the three differ-
parts (f) and (k)? Why would it be unwise to
ent fitted equations? Compare the three fitted
use either of these predictions without further
equations in terms of complexity and appar-
data collection and analysis?
ent ability to predict y1 .
17. In the previous exercise, both x and y were really (c) Compute the residuals for the third fitted
response variables. As such, they were not subject equation in part (b). Plot them against x1 ,
to direct manipulation by the experimenters. That x2 , and ŷ 1 . Also normal-plot them. Do any of
made it difficult to get several (x, y) pairs with these plots suggest that the third fitted equa-
a single x value into the data set. In experimen- tion is inadequate as a summary of these data?
tal situations where an engineer gets to choose (d) As a means of understanding the nature of
values of an experimental variable x, why is it the third fitted equation from part (b), make a
useful/important to get several y observations for scatterplot of y1 vs. x2 . On this plot, plot three
at least some x’s? lines representing ŷ 1 as a function of x2 for
18. Chemical engineering graduate student S. Osoka the three different values of x1 represented in
studied the effects of an agitator speed, x 1 , and a the data set.
Chapter 4 Exercises 213
(e) Using the third equation from part (b), what Factor A Plane Design
pyrite recovery rate would you predict for straight wing (−) vs. t wing (+)
x1 = 1000 rpm and x2 = 110 ppm?
Factor B Nose Weight
(f) Consider also a multivariable quadratic de-
none (−) vs. paper clip (+)
scription of the dependence of y1 on x1 and
x2 . That is, fit the equation Factor C Paper Type
notebook (−) vs. construction (+)
y1 ≈ β0 + β1 x1 + β2 x2 + β3 x12 Factor D Wing Tips
straight (−) vs. bent up (+)
+β4 x22 + β5 x1 x2
The mean flight distances, y (ft), recorded by Fel-
to the data. How does the R 2 value here com- lows for two launches of each plane were as shown
pare with the ones in part (b)? As a means of in the accompanying table.
understanding this fitted equation, plot on a (a) Use the Yates algorithm and compute the fit-
single set of axes the three different quadratic ted factorial effects corresponding to the “all
functions of x 2 obtained by holding x1 at one high” treatment combination.
of the values in the data set. (b) Interpret the results of your calculations from
(g) It is possible to ignore the fact that the speed (a) in the context of the study. (Describe in
and concentration factors are quantitative and words which factors and/or combinations of
to make a factorial analysis of these y1 data. factors appear to have the largest effect(s) on
Do so. Begin by making an interaction plot flight distance. What are the practical impli-
similar to Figure 4.22 for these data. Based cations of these effects?)
on that plot, discuss the apparent relative sizes
of the Speed and Concentration main effects Combination y Combination y
and the Speed × Concentration interactions.
Then compute the fitted factorial effects (the (1) 6.25 d 7.00
fitted main effects and interactions). a 15.50 ad 10.00
(h) If the third equation in part (b) governed y1 , b 7.00 bd 10.00
would it lead to Speed × Concentration inter- ab 16.50 abd 16.00
actions? What about the equation in part (f)? c 4.75 cd 4.50
Explain. ac 5.50 acd 6.00
19. The data given in the previous exercise concern bc 4.50 bcd 4.50
both responses y1 and y2 . The previous analysis abc 6.00 abcd 5.75
dealt with only y1 . Redo all parts of the problem,
replacing the response y1 with y2 throughout. (c) Suppose factors B and D are judged to be
20. K. Fellows conducted a 4-factor experiment, with inert as far as determining flight distance is
the response variable the flight distance of a pa- concerned. (The main effects of B and D and
per airplane when propelled from a launcher fab- all interactions involving them are negligi-
ricated specially for the study. This exercise con- ble.) What fitted/predicted values correspond
cerns part of the data he collected, constituting to this description of flight distance (A and
a complete 24 factorial. The experimental factors C main effects and AC interactions only)?
involved and levels used were as given here. Use these 16 values of ŷ to compute residu-
als, y − ŷ. Plot these against ŷ, levels of A,
levels of B, levels of C, and levels of D. Also
214 Chapter 4 Describing Relationships Between Variables
normal-plot these residuals. Comment on any (b) What is the correlation between x1 and y?
interpretable patterns in your plots. The correlation between x2 and y?
(d) Compute R 2 corresponding to the descrip- (c) Based on (a) and (b), describe how strongly
tion of flight distance used in part (c). (The Thickness and Hardness appear to affect bal-
formula in Definition 3 works in this context listic limit. Review the raw data and specu-
as well as in regression. So does the represen- late as to why the variable with the smaller
tation of R 2 as the squared sample correlation influence on y seems to be of only minor im-
between y and ŷ.) Does it seem that the grand portance in this data set (although logic says
mean, A and C main effects, and AC 2-factor that it must in general have a sizable influence
interactions provide an effective summary of on y).
flight distance? (d) Compute the residuals for the third fitted
21. The data in the accompanying table appear in the equation from (a). Plot them against x 1 , x2 ,
text Quality Control and Industrial Statistics by and ŷ. Also normal-plot them. Do any of
Duncan (and were from a paper of L. E. Simon). these plots suggest that the third fitted equa-
The data were collected in a study of the effec- tion is seriously deficient as a summary of
tiveness of armor plate. Armor-piercing bullets these data?
were fired at an angle of 40◦ against armor plate (e) Plot the (x 1 , x2 ) pairs represented in the data
of thickness x1 (in .001 in.) and Brinell hardness set. Why would it be unwise to use any of the
number x2 , and the resulting so-called ballistic fitted equations to predict y for x 1 = 265 and
limit, y (in ft/sec), was measured. x2 = 440?
22. Basgall, Dahl, and Warren experimented with
x1 x2 y x1 x2 y smooth and treaded bicycle tires of different
widths. Tires were mounted on the same wheel,
253 317 927 253 407 1393 placed on a bicycle wind trainer, and accelerated
258 321 978 252 426 1401 to a velocity of 25 miles per hour. Then pedaling
259 341 1028 246 432 1436 was stopped, and the time required for the wheel
247 350 906 250 469 1327 to stop rolling was recorded. The sample means,
256 352 1159 242 257 950 y, of five trials for each of six different tires were
246 363 1055 243 302 998 as follows:
257 365 1335 239 331 1144
262 375 1392 242 355 1080 Tire Width Tread Time to Stop, y (sec)
255 373 1362 244 385 1276 700/19c smooth 7.30
258 391 1374 234 426 1062 700/25c smooth 8.44
700/32c smooth 9.27
(a) Use a regression program to fit the following 700/19c treaded 6.63
equations to these data:
700/25c treaded 6.87
700/32c treaded 7.07
y ≈ β 0 + β1 x 1
y ≈ β 0 + β2 x 2 (a) Carefully make an interaction plot of times
required to stop, useful for investigating the
y ≈ β 0 + β1 x 1 + β2 x 2
sizes of Width and Tread main effects and
Width × Tread interactions here. Comment
What are the R 2 values for the three differ- briefly on what the plot shows about these
ent fitted equations? Compare the three fitted effects. Be sure to label the plot very clearly.
equations in terms of complexity and appar-
ent ability to predict y.
Chapter 4 Exercises 215
(b) Compute the fitted main effects of Width, 5.00 km/sec detonation velocity, what PETN
the fitted main effects of Tread, and the fit- density would you employ?
ted Width × Tread interactions from the y’s. (g) Compute the residuals from your fitted line.
Discuss how they quantify features that are Plot them against x and against ŷ. Then make
evident in your plot from (a). a normal plot of the residuals. What do these
23. Below are some data read from a graph in the ar- indicate about the linearity of the relationship
ticle “Chemical Explosives” by W. B. Sudweeks between y and x?
that appears as Chapter 30 in Riegel’s Handbook (h) Use a statistical package and compute the
of Industrial Chemistry. The x values are densities least squares line, the sample correlation, R 2 ,
(in g/cc) of pentaerythritol tetranitrate (PETN) and the residuals from the least squares line
samples and the y values are corresponding deto- for these data.
nation velocities (in km/sec). 24. Some data collected in a study intended to reduce
a thread stripping problem in an assembly process
x y x y x y follow. Studs screwed into a metal block were
stripping out of the block when a nut holding
.19 2.65 .50 3.95 .91 5.29 another part on the block was tightened. It was
.20 2.71 .50 3.87 .91 5.11 thought that the depth the stud was screwed into
.24 2.79 .50 3.57 .95 5.33 the block (the thread engagement) might affect
.24 3.19 .55 3.84 .95 5.27 the torque at which the stud stripped out. In the
.25 2.83 .75 4.70 .97 5.30 table below, x is the depth (in 10−3 inches above
.30 3.52 .77 4.19 1.00 5.52 .400) and y is the torque at failure (in lbs/in.).
.30 3.41 .80 4.75 1.00 5.46
.32 3.51 .80 4.38 1.00 5.30 x y x y x y x y
.43 3.38 .85 4.83 1.03 5.59
80 15 40 70 75 70 20 70
.45 3.13 .85 5.32 1.04 5.71
76 15 36 65 25 70 40 65
88 25 30 65 30 60 30 75
(a) Make a scatterplot of these data and comment
on the apparent linearity (or the lack thereof) 35 60 0 45 78 25 74 25
of the relationship between y and x. 75 35 44 50 60 45
(b) Compute the sample correlation between y
and x. Interpret this value. (a) Use a regression program and fit both a linear
(c) Show the “hand” calculations necessary to fit equation and a quadratic equation to these
a line to these data by least squares. Then plot data. Plot them on a scatterplot of the data.
your line on the graph from (a). What are the fractions of raw variability in y
(d) About what increase in detonation velocity accounted for by these two equations?
appears to accompany a unit (1 g/cc) increase (b) Redo part (a) after dropping the x = 0 and
in PETN density? What increase in detona- y = 45 data point from consideration. Do
tion velocity would then accompany a .1 g/cc your conclusions about how best to describe
increase in PETN density? the relationship between x and y change ap-
(e) What fraction of the raw variability in detona- preciably? What does this say about the ex-
tion velocity is “accounted for” by the fitted tent to which a single data point can affect a
line from part (c)? curve-fitting analysis?
(f) Based on your analysis, about what detona- (c) Use your quadratic equation from part (a) and
tion velocity would you predict for a PETN find a thread engagement that provides an op-
density of 0.65 g/cc? If it was your job to timal predicted failure torque. What would
produce a PETN explosive charge with a
216 Chapter 4 Describing Relationships Between Variables
you probably want to do before recommend- (e) About what increase in log grip force appears
ing this depth for use in this assembly pro- to accompany an increase in drag of 10% of
cess? the total possible? This corresponds to what
25. The textbook Introduction to Contemporary Sta- kind of change in raw grip force?
tistical Methods by L. H. Koopmans contains a (f) What fraction of the raw variability in log grip
data set from the testing of automobile tires. A tire force is accounted for in the fitting of a line
under study is mounted on a test trailer and pulled to the data in part (d)?
at a standard velocity. Using a braking mecha- (g) Based on your answer to (d), what log grip
nism, a standard amount of drag (measured in %) force would you predict for a tire of this type
is applied to the tire and the force (in pounds) under these conditions using 40% of the pos-
with which it grips the road is measured. The fol- sible drag? What raw grip force?
lowing data are from tests on 19 different tires (h) Compute the residuals from your fitted line.
of the same design made under the same set of Plot the residuals against x and against ŷ.
road conditions. x = 0% indicates no braking and Then make a normal plot of the residuals.
x = 100% indicates the brake is locked. What do these plots indicate about the linear-
ity of the relationship between drag and log
Drag, x (%) Grip Force, y (lb) grip force?
(i) Use a statistical package to find the least
10 550, 460, 610 squares line, the sample correlation, R 2 , and
20 510, 410, 580 the residuals for these (x, y 0 ) data.
30 470, 360, 480 26. The article “Laboratory Testing of Asphalt Con-
50 390, 310, 400 crete for Porous Pavements” by Woelfl, Wei, Faul-
70 300, 280, 340 stich, and Litwack (Journal of Testing and Evalu-
100 250, 200, 200, 200 ation, 1981) studied the effect of asphalt content
on the permeability of open-graded asphalt con-
(a) Make a scatterplot of these data and comment crete. Four specimens were tested for each of
on “how linear” the relation between y and x six different asphalt contents, with the following
appears to be. results:
In fact, physical theory can be called upon to pre-
dict that instead of being linear, the relationship Asphalt Content, Permeability,
between y and x is of the form y ≈ α exp(βx) x (% by weight) y (in./hr water loss)
for suitable α and β. Note that if natural loga-
rithms are taken of both sides of this expression, 3 1189, 840, 1020, 980
ln(y) ≈ ln(α) + βx. Calling ln(α) by the name 4 1440, 1227, 1022, 1293
β0 and β by the name β1 , one then has a linear 5 1227, 1180, 980, 1210
relationship of the form used in Section 4.1. 6 707, 927, 1067, 822
(b) Make a scatterplot of y 0 = ln(y) versus x. 7 835, 900, 733, 585
Does this plot look more linear than the one 8 395, 270, 310, 208
in (a)?
(c) Compute the sample correlation between y 0 (a) Make a scatterplot of these data and comment
and x “by hand.” Interpret this value. on how linear the relation between y and x
(d) Fit a line to the drags and logged grip forces appears to be. If you focus on asphalt con-
using the least squares principle. Show the tents between, say, 5% and 7%, does linearity
necessary hand calculations. Sketch this line seem to be an adequate description of the re-
on your scatterplot from (b). lationship between y and x?
Chapter 4 Exercises 217
Temporarily restrict your attention to the x = 5, 6, (l) What do the fitted relationships from (c), (i)
and 7 data. and (j) give for predicted permeabilities when
(b) Compute the sample correlation between y x = 2%? Compare these to each other as well
and x “by hand.” Interpret this value. as your answers to (f) and (k). Why would
(c) Fit a line to the asphalt contents and per- it be unwise to use any of these predictions
meabilities using the least squares principle. without further data collection?
Show the necessary hand calculations. Sketch 27. Some data collected by Koh, Morden, and Og-
this fitted line on your scatterplot from (a). bourne in a study of axial breaking strengths (y)
(d) About what increase in permeability appears for wooden dowel rods follow. The students tested
to accompany a 1% (by weight) increase in m = 4 different dowels for each of nine combi-
asphalt content? nations of three different diameters (x1 ) and three
(e) What fraction of the raw variability in perme- different lengths (x2 ).
ability is “accounted for” in the fitting of a
line to the x = 5, 6, and 7 data in part (c)?
x1 (in.) x2 (in.) y (lb)
(f) Based on your answer to (c), what measured
permeability would you predict for a speci- .125 4 51.5, 37.4, 59.3, 58.5
men of this material with an asphalt content .125 8 5.2, 6.4, 9.0, 6.3
of 5.5%? .125 12 2.5, 3.3, 2.6, 1.9
(g) Compute the residuals from your fitted line.
.1875 4 225.3, 233.9, 211.2, 212.8
Plot the residuals against x and against ŷ.
.1875 8 47.0, 79.2, 88.7, 70.2
Then make a normal plot of the residuals.
.1875 12 18.4, 22.4, 18.9, 16.6
What do these plots indicate about the linear-
ity of the relationship between asphalt content .250 4 358.8, 309.6, 343.5, 357.8
and permeability? .250 8 127.1, 158.0, 194.0, 133.0
(h) Use a statistical package and values for x = .250 12 68.9, 40.5, 50.3, 65.6
5, 6, and 7 to find the least squares line, the
sample correlation, R 2 , and the residuals for (a) Make a plot of the 3 × 3 means, ȳ, corre-
these data. sponding to the different combinations of di-
Now consider again the entire data set. ameter and length used in the study, plotting
(i) Fit the quadratic relationship y ≈ β0 + β1 x + ȳ vs. x2 and connecting the three means for
β2 x 2 to the data using a statistical pack- a given diameter with line segments. What
age. Sketch this fitted parabola on your sec- does this plot suggest about how successful
ond scatterplot from part (a). Does this fit- an equation for y that is linear in x2 for each
ted quadratic appear to be an important im- fixed x1 might be in explaining these data?
provement over the line you fit in (c) in terms (b) Replace the strength values with their natural
of describing the relationship over the range logarithms, y 0 = ln(y), and redo the plotting
3≤x ≤8? of part (a). Does this second plot suggest that
(j) Fit the linear relation y ≈ β0 + β1 x to the en- the logarithm of strength might be a linear
tire data set. How do the R 2 values for this fit function of length for fixed diameter?
and the one in (i) compare? Does the larger R 2 (c) Fit the following three equations to the data
in (i) speak strongly for the use of a quadratic via least squares:
(as opposed to a linear) description of the re-
lationship of y to x in this situation? y 0 ≈ β 0 + β1 x 1 ,
(k) If one uses the fitted relationship from (i) to y 0 ≈ β 0 + β2 x 2 ,
predict y for x = 5.5, how does the prediction
compare to your answer for (f)? y 0 ≈ β 0 + β1 x 1 + β2 x 2
218 Chapter 4 Describing Relationships Between Variables
What are the coefficients of determination for analysis. Looking again at your plot from (a),
the three fitted equations? Compare the equa- does it seem that the interactions of Diameter
tions in terms of their complexity and their and Length will be important in describing the
apparent ability to predict y 0 . raw strengths, y? Compute the fitted factorial
(d) Add three lines to your plot from part (b), effects and comment on the relative sizes of
showing predicted log strength (from your the main effects and interactions.
third fitted equation) as a function of x2 for (h) Redo part (g), referring to the graph from part
the three different values of x1 included in (b) and working with the logarithms of dowel
the study. Use your third fitted equation to strength.
predict first a log strength and then a strength 28. The paper “Design of a Metal-Cutting Drilling
for a dowel of diameter .20 in. and length 10 Experiment—A Discrete Two-Variable Problem”
in. Why shouldn’t you be willing to use your by E. Mielnik (Quality Engineering, 1993–1994)
equation to predict the strength of a rod with reports a drilling study run on an aluminum al-
diameter .50 in. and length 24 in.? loy (7075-T6). The thrust (or axial force), y1 , and
(e) Compute and plot residuals for the third equa- torque, y2 , required to rotate drills of various di-
tion you fit in part (c). Make plots of residuals ameters x1 at various feeds (rates of drill penetra-
vs. fitted response and both x 1 and x2 , and tion into the workpiece) x2 , were measured with
normal-plot the residuals. Do these plots sug- the following results:
gest any potential inadequacies of the third
fitted equation? How might these be reme-
Diameter, Feed Rate, Thrust, Torque,
died?
x1 (in.) x2 (in. rev) y1 (lb) y2 (ft-lb)
(f) The students who did this study were strongly
suspicious that the ratio x3 = x12 /x2 is the .250 .006 230 1.0
principal determiner of dowel strength. In .406 .006 375 2.1
fact, it is possible to empirically discover the .406 .013 570 3.8
importance of this quantity as follows. Try
.250 .013 375 2.1
fitting the equation
.225 .009 280 1.0
y 0 ≈ β0 + β1 ln x1 + β2 ln x2 .318 .005 225 1.1
.450 .009 580 3.8
to these data and notice that the fitted coef- .318 .017 565 3.4
ficients of ln x1 and ln x2 are roughly in the .318 .009 400 2.2
ratio of 4 to −2, i.e., 2 to −1. (What does this .318 .009 400 2.1
fitted equation for ln(y) say about y?) Then .318 .009 380 2.1
plot y vs. x 3 and fit the linear equation .318 .009 380 1.9
(a) Use a regression program to fit the following (f) Redo part (e), using y10 as the response vari-
equations to these data: able.
(g) Do your answers to parts (e) and (f) comple-
y10 ≈ β0 + β1 x10 , ment those of part (d)? Explain.
y10 ≈ β0 + β2 x20 , 29. The article “A Simple Method to Study Dispersion
Effects From Non-Necessarily Replicated Data in
y10 ≈ β0 + β1 x10 + β2 x20 Industrial Contexts” by Ferrer and Romero (Qual-
ity Engineering, 1995) describes an unreplicated
What are the R 2 values for the three differ- 24 experiment done to improve the adhesive force
ent fitted equations? Compare the three fitted obtained when gluing on polyurethane sheets as
equations in terms of complexity and appar- the inner lining of some hollow metal parts. The
ent ability to predict y10 . factors studied were the amount of glue used (A),
(b) Compute and plot residuals (continuing to the predrying temperature (B), the tunnel temper-
work on log scales) for the third equation you ature (C), and the pressure applied (D). The exact
fit in part (a). Make plots of residuals vs. fitted levels of the variables employed were not given
y10 and both x10 and x20 , and normal-plot these in the article (presumably for reasons of corporate
residuals. Do these plots reveal any particular security). The response variable was the adhesive
problems with the fitted equation? force, y, in Newtons, and the data reported in the
(c) Use your third equation from (a) to predict article follow:
first a log thrust and then a thrust if a drill of
diameter .360 in. and a feed of .011 in./rev Combination y Combination y
are used. Why would it be unwise to make
a similar prediction for x1 = .450 and x2 = (1) 3.80 d 3.29
.017? (Hint: Make a plot of the (x1 , x2 ) pairs a 4.34 ad 2.82
in the data set and locate this second set of b 3.54 bd 4.59
conditions on that plot.) ab 4.59 abd 4.68
(d) If the third equation fit in part (a) governed y1 , c 3.95 cd 2.73
would it lead to Diameter × Feed interactions
ac 4.83 acd 4.31
for y1 measured on the log scale? To help you
bc 4.86 bcd 5.16
answer this question, plot yb10 vs. x2 (or x20 ) for
abc 5.28 abcd 6.06
each of x 1 = .250, .318, and .406. Does this
equation lead to Diameter × Feed interactions (a) Compute the fitted factorial effects corre-
for raw y1 ? sponding to the “all high” treatment com-
(e) The first four data points listed in the ta- bination.
ble constitute a very small complete factorial (b) Interpret the results of your calculations in
study (an unreplicated 2 × 2 factorial in the the context of the study. Which factors and/or
factors Diameter and Feed). Considering only combinations of factors appear to have the
these data points, do a “factorial” analysis of largest effects on the adhesive force? Suppose
this part of the y1 data. Begin by making an in- that only the A, B, and C main effects and
teraction plot similar to Figure 4.22 for these the B × D interactions were judged to be
data. Based on that plot, discuss the apparent of importance here. Make a corresponding
relative sizes of the Diameter and Feed main statement to your engineering manager about
effects on thrust. Then carry out the arith- how the factors impact adhesive force.
metic necessary to compute the fitted factorial
effects (the main effects and interactions).
220 Chapter 4 Describing Relationships Between Variables
Probability:
The Mathematics
of Randomness
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
221
222 Chapter 5 Probability: The Mathematics of Randomness
factor in statistical engineering studies, and the many small, unnameable causes that
work to produce it are conveniently thought of as chance phenomena. In analytical
contexts, changes in system conditions work to make measured responses vary, and
this is most often attributed to chance.
a random variable.
Following Definition 9 in Chapter 1, a distinction was made between discrete
and continuous data. That terminology carries over to the present context and inspires
two more definitions.
Definition 2 A discrete random variable is one that has isolated or separated possible
values (rather than a continuum of available outcomes).
Random variables that are basically count variables clearly fall under Defi-
nition 2 and are discrete. It could be argued that all measurement variables are
discrete—on the basis that all measurements are “to the nearest unit.” But it is often
mathematically convenient, and adequate for practical purposes, to treat them as
continuous.
A random variable is, to some extent, a priori unpredictable. Therefore, in
describing or modeling it, the important thing is to specify its set of potential values
and the likelihoods associated with those possible values.
Definition 4 To specify a probability distribution for a random variable is to give its set
of possible values and (in one way or another) consistently assign numbers
5.1 (Discrete) Random Variables 223
This text will use the notational convention that a capital P followed by an
expression or phrase enclosed by brackets will be read “the probability” of that
expression. In these terms, a probability function for X is a function f such that
Probability function
for the discrete f (x) = P[X = x]
random variable X
That is, “ f (x) is the probability that (the random variable) X takes the value x.”
Z = the next measured torque for bolt 3 (recorded to the nearest integer)
Example 1 to produce the next bolt 3 torque, then it also makes sense to base a probability
(continued ) function for Z on the relative frequencies in Table 5.1. That is, the probability
distribution specified in Table 5.2 might be used. (In going from the relative
frequencies in Table 5.1 to proposed values for f (z) in Table 5.2, there has been
some slightly arbitrary rounding. This has been done so that probability values
are expressed to two decimal places and now total to exactly 1.00.)
Table 5.1
Relative Frequency Distribution for Measured Bolt 3
Torques
Table 5.2
A Probability Function
for Z
Torque Probability
z f (z)
11 .03
12 .03
13 .03
14 .06
15 .26
16 .09
17 .12
18 .20
19 .15
20 .03
5.1 (Discrete) Random Variables 225
Then the probability function in Table 5.2 is also approximately appropriate for Y .
This point is not so important in this specific example as it is in general: Where
The probability one value is to be selected at random from a population, an appropriate probability
distribution of a distribution is one that is equivalent to the population relative frequency distribution.
single value selected This text will usually express probabilities to two decimal places, as in Table 5.2.
at random from Computations may be carried to several more decimal places, but final probabilities
a population will typically be reported only to two places. This is because numbers expressed to
more than two places tend to look too impressive and be taken too seriously by the
uninitiated. Consider for example the statement “There is a .097328 probability of
booster engine failure” at a certain missile launch. This may represent the results of
some very careful mathematical manipulations and be correct to six decimal places
in the context of the mathematical model used to obtain the value. But it is doubtful
that the model used is a good enough description of physical reality to warrant that
much apparent precision. Two-decimal precision is about what is warranted in most
engineering applications of simple probability.
Properties of a The probability function shown in Table 5.2 has two properties that are necessary
mathematically valid for the mathematical consistency of a discrete probability distribution. The f (z)
probability function values are each in the interval [0, 1] and they total to 1. Negative probabilities or
ones larger than 1 would make no practical sense. A probability of 1 is taken as
indicating certainty of occurrence and a probability of 0 as indicating certainty of
nonoccurrence. Thus, according to the model specified in Table 5.2, since the values
of f (z) sum to 1, the occurrence of one of the values 11, 12, 13, 14, 15, 16, 17, 18,
19, and 20 ft lb is certain.
A probability function f (x) gives probabilities of occurrence for individual val-
ues. Adding the appropriate values gives probabilities associated with the occurrence
of one of a specified type of value for X .
Adding the f (z) entries corresponding to possible values larger than 17 ft lb,
P[Z > 17] = f (18) + f (19) + f (20) = .20 + .15 + .03 = .38
The likelihood of the next torque being more than 17 ft lb is about 38%.
226 Chapter 5 Probability: The Mathematics of Randomness
Example 1 If, for example, specifications for torques were 16 ft lb to 21 ft lb, then the
(continued ) likelihood that the next torque measured will be within specifications is
Suppose further that tool serial numbers begin with some code special to the
tool model and end with consecutively assigned numbers reflecting how many
tools of the particular model have been produced. The symmetry of this situation
suggests that each possible value of W (w = 0, 1, . . . , 9) is equally likely. That
is, a plausible probability function for W is given by the formula
(
.1 for w = 0, 1, 2, . . . , 9
f (w) =
0 otherwise
F(x) = P[X ≤ x]
5.1 (Discrete) Random Variables 227
Cumulative probability X
F(x) = f (z)
function for a discrete
z≤x
variable X
(The sum is over possible values less than or equal to x.) In this discrete case, the
graph of F(x) will be a stair-step graph with jumps located at possible values and
equal in size to the probabilities associated with those possible values.
Example 1 Values of both the probability function and the cumulative probability function
(continued ) for the torque variable Z are given in Table 5.3. Values of F(z) for other z are
also easily obtained. For example,
Table 5.3
Values of the Probability Function and Cumulative
Probability Function for Z
Example 1 F(z)
(continued )
1.0
.5
11 12 13 14 15 16 17 18 19 20 z
X
EX = x f (x) (5.1)
x
.2 .2
.1 .1
11 12 13 14 15 16 17 18 19 20 z 0 1 2 3 4 5 6 7 8 9 w
(Remember the warning in Section 3.3 that µ would stand for both the mean of a
population and the mean of a probability distribution.)
Example 1 Returning to the bolt torque example, the expected (or theoretical mean) value of
(continued ) the next torque is
X
EZ = z f (z)
z
This value is essentially the arithmetic mean of the bolt 3 torques listed in
Table 3.4. (The slight disagreement in the third decimal place arises only because
the relative frequencies in Table 5.1 were rounded slightly to produce Table 5.2.)
This kind of agreement provides motivation for using the symbol µ, first seen in
Section 3.3, as an alternative to EZ.
Example 2 Considering again the serial number example, and the second part of Figure 5.2,
(continued ) if a balance point interpretation of expected value is to hold, EW had better turn
out to be 4.5. And indeed,
It was convenient to measure the spread of a data set (or its relative frequency
distribution) with the variance and standard deviation. It is similarly useful to have
notions of spread for a discrete probability distribution.
Definition 8 The variance of a discrete random variable X (or the variance of its distribu-
tion) is
P P
Var X = (x − EX)2 f (x) = x 2 f (x) − (EX)2 (5.2)
√
The standard deviation of X is Var X√. Often the notation σ 2 is used in
place of Var X, and σ is used in place of Var X .
The variance of a random variable is its expected (or mean) squared distance
from the center of its probability distribution. The use of σ 2 to stand for both the
variance of a population and the variance of a probability distribution is motivated
on the same grounds as the double use of µ.
Example 1 The calculations necessary to produce the bolt torque standard deviation are
(continued ) organized in Table 5.4. So
√ √
I σ = Var Z = 4.6275 = 2.15 ft lb
Except for a small difference due to round-off associated with the creation of
Table 5.2, this standard deviation of the random variable Z is numerically the
same as the population standard deviation associated with the bolt 3 torques in
Table 3.4. (Again, this is consistent with the equivalence between the population
relative frequency distribution and the probability distribution for Z .)
5.1 (Discrete) Random Variables 231
Table 5.4
Calculations for Var Z
Example 2 To illustrate the alternative for calculating a variance given in Definition 8, con-
(continued ) sider finding the variance and standard
P deviation of the serial number variable W .
Table 5.5 shows the calculation of w2 f (w).
Table 5.5 P
Calculations for w 2 f (w)
w f (w) w2 f (w)
0 .1 0.0
1 .1 .1
2 .1 .4
3 .1 .9
4 .1 1.6
5 .1 2.5
6 .1 3.6
7 .1 4.9
8 .1 6.4
9 .1 8.1
28.5
232 Chapter 5 Probability: The Mathematics of Randomness
Example 2 Then
(continued ) X
Var W = w2 f (w) − (EW)2 = 28.5 − (4.5)2 = 8.25
so that
√
I Var W = 2.87
Comparing the two probability histograms in Figure 5.2, notice that the distribu-
tion of W appears to be more spread out than that of Z . Happily, this is reflected
in the fact that
√ √
Var W = 2.87 > 2.15 = Var Z
n!
p x (1 − p)n−x for x = 0, 1, . . . , n
f (x) = x! (n − x)! (5.3)
0 otherwise
Equation (5.3) is completely plausible. In it there is one factor of p for each trial pro-
ducing a go/success outcome and one factor of (1 − p) for each trial producing a no
go/failure outcome. And the n!/x! (n − x)! term is a count of the number of patterns
in which it would be possible to see x go/success outcomes in n trials. The name bi-
nomial distribution derives from the fact that the values f (0), f (1), f (2), . . . , f (n)
are the terms in the expansion of
( p + (1 − p))n
.2 .2 .2
.1 .1 .1
0 1 2 3 4 5 x 0 1 2 3 4 5 x 0 1 2 3 4 5 x
f (x)
n = 10
.3 p = .2
.2
.1
0 1 2 3 4 5 6 7 8 9 10 x
(The trick employed here, to avoid plugging into the binomial probability function
9 times by recognizing that the f (u)’s have to sum up to 1, is a common and
useful one.)
The .62 figure is only as good as the model assumptions that produced it.
If an independent, identical success-failure trials description of shaft production
fails to accurately portray physical reality, the .62 value is fine mathematics
but possibly a poor description of what will actually happen. For instance, say
that due to tool wear it is typical to see 40 shafts in specifications, then 10
reworkable shafts, a tool change, 40 shafts in specifications, and so on. In this
case, the binomial distribution would be a very poor description of U , and the
.62 figure largely irrelevant. (The independence-of-trials assumption would be
inappropriate in this situation.)
The binomial There is one important circumstance where a model of independent, identical
distribution and success-failure trials is not exactly appropriate, but a binomial distribution can still be
simple random adequate for practical purposes—that is, in describing the results of simple random
sampling sampling from a dichotomous population. Suppose a population of size N contains
5.1 (Discrete) Random Variables 235
f (0) = P[V = 0]
= P[first pellet selected is nonconforming and
subsequently the second pellet is also nonconforming]
f (2) = P[V = 2]
= P[first pellet selected is conforming and
subsequently the second pellet selected is conforming]
f (1) = 1 − ( f (0) + f (2))
Then think, “In the long run, the first selection will yield a nonconforming pellet
about 34 out of 100 times. Considering only cases where this occurs, in the long
run the next selection will also yield a nonconforming pellet about 33 out of 99
times.” That is, a sensible evaluation of f (0) is
34 33
f (0) = · = .1133
100 99
236 Chapter 5 Probability: The Mathematics of Randomness
Example 4 Similarly,
(continued )
66 65
f (2) = · = .4333
100 99
and thus
2!
(.34)2 (.66)0 = .1156 ≈ f (0)
0! 2!
2!
(.34)1 (.66)1 = .4488 ≈ f (1)
1! 1!
2!
(.34)0 (.66)2 = .4356 ≈ f (2)
2! 0!
Calculation of the mean and variance for binomial random variables is greatly
simplified by the fact that when the formulas (5.1) and (5.2) are used with the
expression for binomial probabilities in equation (5.3), simple formulas result. For
X a binomial (n, p) random variable,
Mean of the X
n
n!
binomial (n, p) µ = EX = x p x (1 − p)n−x = np (5.4)
distribution x=0
x!(n − x)!
Variance of the X
n
n!
binomial (n, p) σ = Var X =
2
(x − np)2 p x (1 − p)n−x = np(1 − p) (5.5)
distribution x=0
x!(n − x)!
5.1 (Discrete) Random Variables 237
Example 3 Returning to the machining of steel shafts, suppose that a binomial distribution
(continued ) with n = 10 and p = .2 is appropriate as a model for
EU = (10)(.2) = 2 shafts
√ p
Var U = 10(.2)(.8) = 1.26 shafts
(
p(1 − p)x−1 for x = 1, 2, . . .
f (x) = (5.6)
0 otherwise
Formula (5.6) makes good intuitive sense. In order for X to take the value x,
there must be x − 1 consecutive no-go/failure results followed by a go/success. In
formula (5.6), there are x − 1 terms (1 − p) and one term p. Another way to see
that formula (5.6) is plausible is to reason that for X as above and x = 1, 2, . . .
1 − F(x) = 1 − P[X ≤ x]
= P[X > x]
= P[x no-go/failure outcomes in x trials]
That is,
Simple relationship for
the geometric (p)
1 − F(x) = (1 − p)x (5.7)
cumulative probability
function
238 Chapter 5 Probability: The Mathematics of Randomness
f(x)
.5
.4
p = .5
.3
.2
.1
1 2 3 4 5 6 7 8 9 x
f(x)
p = .25
.3
.2
.1
1 2 3 4 5 6 7 8 9 10 11 12 13 x
by using the form of the binomial (x, p) probability function given in equation
(5.3). Then for x = 2, 3, . . . , f (x) = F(x) − F(x − 1) = −(1 − F(x)) + (1 −
F(x − 1)). This, combined with equation (5.7), gives equation (5.6).
The name geometric derives from the fact that the values f (1), f (2), f (3), . . .
are terms in the geometric infinite series for
1
p·
1 − (1 − p)
Suppose that testing begins on a production run in this plant, and let
P[the first or second cell tested has the first short] = P[T = 1 or T = 2]
= f (1) + f (2)
= (.01) + (.01)(1 − .01)
= .02
P[at least 50 cells are tested without finding a short] = P[T > 50]
= (1 − .01)50
= .61
Like the binomial distributions, the geometric distributions have means and
variances that are simple functions of the parameter p. That is, if X is geometric ( p),
Mean of the ∞
X 1
geometric (p) µ = EX = x p(1 − p)x−1 = (5.8)
distribution x=1
p
and
Variance of the ∞
X 2
1 1− p
geometric (p) σ = Var X =
2
x− p(1 − p)x−1 = (5.9)
distribution x=1
p p2
Example 5 Formula (5.8) is an intuitively appealing result. If there is only 1 chance in 100 of
(continued ) encountering a shorted battery at each test, it is sensible to expect to wait through
100 tests on average to encounter the first one.
Definition 11 The Poisson (λ) distribution is a discrete probability distribution with prob-
ability function
−λ x
e λ
f (x) = for x = 0, 1, 2, . . . (5.10)
x!
0 otherwise
for λ > 0.
The form of equation (5.10) may initially seem unappealing. But it is one that
has sensible mathematical origins, is manageable, and has proved itself empirically
useful in many different “rare events” circumstances. One way to arrive at equation
(5.10) is to think of a very large number of independent trials (opportunities for
occurrence), where the probability of success (occurrence) on any one is very small
and the product of the number of trials and the success probability is λ. One is
then led to the binomial (n, λn ) distribution. In fact, for large n, the binomial (n, λn )
probability function approximates the one specified in equation (5.10). So one
might think of the Poisson distribution for counts as arising through a mechanism
that would present many tiny similar opportunities for independent occurrence or
nonoccurrence throughout an interval of time or space.
The Poisson distributions are right-skewed distributions over the values x =
0, 1, 2, . . . , whose probability histograms peak near their respective λ’s. Two dif-
ferent Poisson probability histograms are shown in Figure 5.5. λ is both the mean
5.1 (Discrete) Random Variables 241
f(x)
λ = 1.5
.3
.2
.1
0 1 2 3 4 5 6 7 8 x
f (x)
λ = 3.0
.3
.2
.1
0 1 2 3 4 5 6 7 8 9 10 11 x
and the variance for the Poisson (λ) distribution. That is, if X has the Poisson (λ)
distribution, then
X∞
Mean of the e−λ λx
Poisson (λ) µ = EX = x =λ (5.11)
distribution x=0
x!
and
Variance of the ∞
X e−λ λx
Poisson (λ) Var X = (x − λ)2 =λ (5.12)
distribution x=0
x!
Fact (5.11) is helpful in picking out which Poisson distribution might be useful in
describing a particular “rare events” situation.
is then
−3.87
e (3.87)s
for s = 0, 1, 2, . . .
f (s) = s!
0 otherwise
M = the number of students entering the ISU library between 12:00 and
12:01 next Tuesday
and, for example, the probability that between 10 and 15 students (inclusive)
arrive at the library between 12:00 and 12:01 would be evaluated as
Section 1 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. A discrete random variable X can be described (b) If seven of the ten subjects correctly identify
using the probability function the artificial sweetener, is this outcome strong
evidence of a taste difference? Explain.
x 2 3 4 5 6 3. Suppose that a small population consists of the
N = 6 values 2, 3, 4, 4, 5, and 6.
f (x) .1 .2 .3 .3 .1 (a) Sketch a relative frequency histogram for this
population and compute the population mean,
(a) Make a probability histogram for X. Also plot µ, and standard deviation, σ .
F(x), the cumulative probability function (b) Now let X = the value of a single number se-
for X . lected at random from this population. Sketch
(b) Find the mean and standard deviation of X . a probability histogram for this variable X and
2. In an experiment to evaluate a new artificial sweet- compute EX and Var X .
ener, ten subjects are all asked to taste cola from (c) Now think of drawing a simple random sample
three unmarked glasses, two of which contain reg- of size n = 2 from this small population. Make
ular cola while the third contains cola made with tables giving the probability distributions of the
the new sweetener. The subjects are asked to iden- random variables
tify the glass whose content is different from the
other two. If there is no difference between the X = the sample mean
taste of sugar and the taste of the new sweetener,
the subjects would be just guessing. S 2 = the sample variance
(a) Make a table for a probability function for
(There are 15 different possible unordered sam-
X = the number of subjects correctly ples of 2 out of 6 items. Each of the 15 possible
identifying the artificially samples is equally likely to be chosen and has
sweetened cola its own corresponding x̄ and s 2 .) Use the tables
and make probability histograms for these ran-
under this hypothesis of no difference in taste. dom variables. Compute EX and Var X . How
do these compare to µ and σ 2 ?
244 Chapter 5 Probability: The Mathematics of Randomness
4. Sketch probability histograms for the binomial dis- each histogram, mark the location of the mean
tributions with n = 5 and p = .1, .3, .5, .7, and .9. and indicate the size of the standard deviation.
On each histogram, mark the location of the mean
and indicate the size of the standard deviation. 8. A process for making plate glass produces an av-
erage of four seeds (small bubbles) per 100 square
5. Suppose that an eddy current nondestructive eval- feet. Use Poisson distributions and assess proba-
uation technique for identifying cracks in critical bilities that
metal parts has a probability of around .20 of detect- (a) a particular piece of glass 5 ft × 10 ft will
ing a single crack of length .003 in. in a certain ma- contain more than two seeds.
terial. Suppose further that n = 8 specimens of this (b) a particular piece of glass 5 ft × 5 ft will con-
material, each containing a single crack of length tain no seeds.
.003 in., are inspected using this technique. Let W
be the number of these cracks that are detected. Use 9. Transmission line interruptions in a telecommu-
an appropriate probability model and evaluate the nications network occur at an average rate of one
following: per day.
(a) P[W = 3] (a) Use a Poisson distribution as a model for
(b) P[W ≤ 2]
(c) E W X = the number of interruptions in the next
(d) Var W five-day work week
(e) the standard deviation of W
and assess P[X = 0].
6. In the situation described in Exercise 5, suppose (b) Now consider the random variable
that a series of specimens, each containing a sin-
gle crack of length .003 in., are inspected. Let Y Y = the number of weeks in the next four
be the number of specimens inspected in order to in which there are no interruptions
obtain the first crack detection. Use an appropriate
probability model and evaluate all of the following: What is a reasonable probability model for
(a) P[Y = 5] Y ? Assess P[Y = 2].
(b) P[Y ≤ 4]
(c) EY 10. Distinguish clearly between the subjects of prob-
(d) Var Y ability and statistics. Is one field a subfield of the
(e) the standard deviation of Y other?
7. Sketch probability histograms for the Poisson dis- 11. What is the difference between a relative fre-
tributions with means λ = .5, 1.0, 2.0, and 4.0. On quency distribution and a probability distribution?
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
and such that for all a ≤ b, one is willing to assign P[a ≤ X ≤ b] according
to
Z b
P[a ≤ X ≤ b] = f (x) dx (5.14)
a
f(x)
Shaded area gives
P[2 ≤ X ≤ 6]
0 2 6 10 14 18 22 x
Example 8 The Random Time Until a First Arc in the Bob Drop Experiment
Consider once again the bob drop experiment first described in Section 1.4 and
revisited in Example 4 in Chapter 4. In any use of the apparatus, the bob is almost
certainly not released exactly “in sync” with the 60 cycle current that produces
the arcs and marks on the paper tape. One could think of a random variable
Y = the time elapsed (in seconds) from bob release until the first arc
1
as continuous with set of possible values (0, 60 ).
What is a plausible probability density function for Y ? The symmetry of this
situation suggests that probability density should be constant over the interval
1
(0, 60 ) and 0 outside the interval. That is, for any two values y1 and y2 in
1
(0, 60 ), the probability that Y takes a value within a small interval around y1 of
length dy (i.e., f (y1 ) dy approximately) should be the same as the probability
that Y takes a value within a small interval around y2 of the same length dy (i.e.,
f (y2 ) dy approximately). This forces f (y1 ) = f (y2 ), so there must be a constant
1
probability density on (0, 60 ).
Now if f (y) is to have the form
(
c for 0 < y < 1
f (y) = 60
0 otherwise
f (y)
Total area under the
graph of f (y) must be 1
c
0 1 y
60
(
60 for 0 < y < 1
I f (y) = 60
(5.15)
0 otherwise
One point about continuous probability distributions that may at first seem coun-
terintuitive concerns the probability associated with a continuous random variable
For X a continuous assuming a particular prespecified value (say, a). Just as the mass a continuous mass
random variable, distribution places at a single point is 0, so also is P[X = a] = 0 for a continuous
P[X = a] = 0 random variable X . This follows from equation (5.14), because
Z a
P[a ≤ X ≤ a] = f (x) dx = 0
a
One consequence of this mathematical curiosity is that when working with contin-
uous random variables, you don’t need to worry about whether or not inequality
signs you write are strict inequality signs. That is, if X is continuous,
Cumulative probability Z x
function for a F(x) = P[X ≤ x] = f (t) dt (5.16)
continuous variable −∞
F(x) is obtained from f (x) by integration, and applying the fundamental theorem
of calculus to equation (5.16)
Another relationship d
between F(x) and f(x)
F(x) = f (x) (5.17)
dx
Example 8 The cumulative probability function for Y , the elapsed time from bob release
(continued ) until first arc, is easily obtained from equation (5.15). For y ≤ 0,
Z y Z y
F(y) = P[Y ≤ y] = f (t) dt = 0 dt = 0
−∞ −∞
That is,
0 if y ≤ 0
I F(y) =
60y if 0 < y ≤ 1/60
1 if 1
60
<y
A plot of F(y) is given in Figure 5.8. Comparing Figure 5.8 to Figure 5.7
shows that indeed the graph of F(y) has slope 0 for y < 0 and y > 60 1
and
slope 60 for 0 < y < 60 . That is, f (y) is the derivative of F(y), as promised by
1
equation (5.17).
5.2 Continuous Random Variables 249
F(y)
0 1 y
60
Z ∞
EX = x f (x) dx (5.18)
−∞
Formula (5.18) is perfectly plausible from at least two perspectives. First, the
probability in a small interval around x of length dx is approximatelyP f (x) dx.
So multiplying this by x and summing as in Definition 7, one has x f (x) dx,
and formula (5.18) is exactly the limit of such sums as dx gets small. And second,
in mechanics the center of mass of a continuous mass distribution is of the form
given in equation (5.18) except for division by a total mass, which for a probability
distribution is 1.
250 Chapter 5 Probability: The Mathematics of Randomness
Example 8 Thinking of the probability density in Figure 5.7 as an idealized histogram and
(continued ) thinking of the balance point interpretation of the mean, it is clear that EY had
1
better turn out to be 120 for the elapsed time variable. Happily, equations (5.18)
and (5.15) give
Z ∞ Z 0 Z 1/60 Z ∞
µ = EY = y f (y) dy = y · 0 dy + y · 60 dy + y · 0 dy
−∞ −∞ 0 1/60
1/60
1
= 30y 2 = sec
0 120
Definition 14 The variance of a continuous random variable X (sometimes called the vari-
ance of its probability distribution) is
Z ∞ Z ∞
Var X = (x − EX) f (x) dx
2
= x f (x) dx − (EX)
2 2
(5.19)
−∞ −∞
√
The standard deviation of X is Var X√. Often the notation σ 2 is used in
place of Var X, and σ is used in place of Var X .
Example 8 Return for a final time to the bob drop and the random variable Y . Using formula
(continued ) (5.19) and the form of Y ’s probability density,
Z 0 2 Z 1/60 2
1 1
σ = Var Y =
2
y− · 0 dy + y− · 60 dy
−∞ 120 0 120
3 1/60
1
Z ∞ 2 60 y −
1 120
+ y− · 0 dy =
1/60 120 3 0
1 1 2
=
3 120
1 2 2
f (x) = p e−(x−µ) /2σ for all x (5.20)
2
2πσ
for σ > 0.
It is not necessarily obvious, but formula (5.20) does yield a legitimate proba-
bility density, in that the total area under the curve y = f (x) is 1. Further, it is also
the case that
Z ∞
Normal distribution 1 2 2
mean and variance
EX = xp e−(x−µ) /2σ dx = µ
−∞ 2
2πσ
and
Z ∞
1 2 2
Var X = (x − µ)2 p e−(x−µ) /2σ dx = σ 2
−∞ 2
2πσ
That is, the parameters µ and σ 2 used in Definition 15 are indeed, respectively, the
mean and variance (as defined in Definitions 13 and 14) of the distribution.
Figure 5.9 is a graph of the probability density specified by formula (5.20). The
bell-shaped curve shown there is symmetric about x = µ and has inflection points
at µ − σ and µ + σ . The exact form of formula (5.20) has a number of theoretical
origins. It is also a form that turns out to be empirically useful in a great variety of
applications.
In theory, probabilities for the normal distributions can be found directly by
integration using formula (5.20). Indeed, readers with pocket calculators that are
preprogrammed to do numerical integration may find it instructive to check some
of the calculations in the examples that follow, by straightforward use of formulas
(5.14) and (5.20). But the freshman calculus methods of evaluating integrals via
antidifferentiation will fail when it comes to the normal densities. They do not have
antiderivatives that are expressible in terms of elementary functions. Instead, special
normal probability tables are typically used.
252 Chapter 5 Probability: The Mathematics of Randomness
f(x)
µ – 2σ µ – σ µ µ + σ µ + 2σ x
The use of tables for evaluating normal probabilities depends on the following
relationship. If X is normally distributed with mean µ and variance σ 2 ,
Z b Z (b−µ)/σ
1 2 2 1 2
P[a ≤ X ≤ b] = p e−(x−µ) /2σ dx = √ e−z /2 dz (5.21)
a 2πσ 2 (a−µ)/σ 2π
where the second inequality follows from the change of variable or substitution
z = x−µ
σ
. Equation (5.21) involves an integral of the normal density with µ = 0
and σ = 1. It says that evaluation of all normal probabilities can be reduced to the
evaluation of normal probabilities for that special case.
Definition 16 The normal distribution with µ = 0 and σ = 1 is called the standard normal
distribution.
Relation between The relationship between normal (µ, σ 2 ) and standard normal probabilities
normal (µ, σ 2 ) is illustrated in Figure 5.10. Once one realizes that probabilities for all normal
probabilities and distributions can be had by tabulating probabilities for only the standard normal
standard normal distribution, it is a relatively simple matter to use techniques of numerical integration
probabilities to produce a standard normal table. The one that will be used in this text (other forms
are possible) is given in Table B.3. It is a table of the standard normal cumulative
probability function. That is, for values z located on the table’s margins, the entries
in the table body are
Z z
1 2
8(z) = F(z) = √ e−t /2 dt
−∞ 2π
(8 is routinely used to stand for the standard normal cumulative probability function,
instead of the more generic F.)
5.2 Continuous Random Variables 253
Normal
( µ , σ 2) P[a ≤ X ≤ b]
density
µ – 2σ µ–σ a µ µ+σ b µ + 2σ x
Equal areas!
Standard
normal a –µ b –µ
σ ≤Z≤ σ
density P
–2 –1 0 1 2 z
a –µ b –µ
σ σ
(The tabled value is .9608, but in keeping with the earlier promise to state final
probabilities to only two decimal places, the tabled value was rounded to get .96.)
After two table look-ups and a subtraction,
And a single table look-up and a subtraction yield a right-tail probability like
As the table was used in these examples, probabilities for values z located
on the table’s margins were found in the table’s body. The process can be run in
254 Chapter 5 Probability: The Mathematics of Randomness
–2 –1 0 1 2 –2 –1 0 1 2
–.89 –z z
reverse. Probabilities located in the table’s body can be used to specify values z
on the margins. For example, consider locating a value z such that
The last part of Example 9 amounts to finding the .975 quantile for the standard
normal distribution. In fact, the reader is now in a position to understand the origin
of Table 3.10 (see page 89). The standard normal quantiles there were found by
looking in the body of Table B.3 for the relevant probabilities and then locating
corresponding z’s on the margins.
In mathematical symbols, for 8(z), the standard normal cumulative probability
function, and Q z ( p), the standard normal quantile function,
)
8(Q z ( p)) = p
(5.22)
Q z (8(z)) = z
Relationships (5.22) mean that Q z and 8 are inverse functions. (In fact, the rela-
tionship Q = F −1 is not just a standard normal phenomenon but is true in general
for continuous distributions.)
Relationship (5.21) shows how to use the standard normal cumulative probabil-
ity function to find general normal probabilities. For X normal (µ, σ 2 ) and a value
5.2 Continuous Random Variables 255
x associated with X , one converts to units of standard deviations above the mean
via
with a normal distribution with µ = 137.2 and σ = 1.6. And further suppose the
probability that the next jar filled is below declared weight (i.e., P[W < 135.0])
is of interest. Using formula (5.23), w = 135.0 is converted to units of standard
deviations above µ (converted to a z-value) as
135.0 − 137.2
z= = −1.38
1.6
This model puts the chance of obtaining a below-nominal fill level at about 8%.
As a second example, consider the probability that W is within 1 gram of
nominal (i.e., P[134.0 < W < 136.0]). Using formula (5.23), both w1 = 134.0
and w2 = 136.0 are converted to z-values or units of standard deviations above
the mean as
134.0 − 137.2
z1 = = −2.00
1.6
136.0 − 137.2
z2 = = −.75
1.6
256 Chapter 5 Probability: The Mathematics of Randomness
P[134.0 < W < 136.0] = .20 P[–2.0 < Z < –.75] = .20
So then
The preceding two probabilities and their standard normal counterparts are shown
in Figure 5.12.
The calculations for this example have consisted of starting with all of the
quantities on the right of formula (5.23) and going from the margin of Table B.3
to its body to find probabilities for W . An important variant on this process is to
instead go from the body of the table to its margins to obtain z, and then—given
only two of the three quantities on the right of formula (5.23)—to solve for the
third.
For example, suppose that it is easy to adjust the aim of the filling process
(i.e., the mean µ of W ) and one wants to decrease the probability that the next
jar is below the declared weight of 135.0 to .01 by increasing µ. What is the
minimum µ that will achieve this (assuming that σ remains at 1.6 g)?
Figure 5.13 shows what to do. µ must be chosen in such a way that w =
135.0 becomes the .01 quantile of the normal distribution with mean µ and
standard deviation σ = 1.6. Consulting either Table 3.10 or Table B.3, it is easy
to determine that the .01 quantile of the standard normal distribution is
z = Q z (.01) = −2.33
5.2 Continuous Random Variables 257
135.0 µ w
135.0 − µ
−2.33 =
1.6
i.e.,
I µ = 138.7 g
1
e−x/α for x > 0
f (x) = α (5.24)
0 otherwise
for α > 0.
Figure 5.14 shows plots of f (x) for three different values of α. Expression
(5.24) is extremely convenient, and it is not at all difficult to show that α is both the
mean and the standard deviation of the exponential (α) distribution. That is,
Mean of the
Z ∞
1
exponential (α) µ = EX = x e−x/α dx = α
0 α
distribution
and
Variance of the Z ∞
1
exponential (α) σ 2 = Var X = (x − α)2 e−x/α dx = α 2
distribution 0 α
f (x)
2.0 α = .5
1.5 α = 1.0
1.0 α = 2.0
.5
0
1.0 2.0 3.0 4.0 5.0 x
T = the waiting time (in minutes) until the first student passes through the door
A possible model for T is the exponential distribution with α = .08. Using it, the
probability of waiting more than 10 seconds ( 16 min) for the first arrival is
1 1
P T > =1− F = 1 − 1 − e−1/6(.08) = .12
6 6
f (t)
10
5 P[T > 1
6
] = .12
.1 1 .2 t
6
Geometric and The exponential distribution is the continuous analog of the geometric distribu-
exponential tion in several respects. For one thing, both the geometric probability function and
distributions the exponential probability density decline exponentially in their arguments x. For
another, they both possess a kind of memoryless property. If the first success in a
series of independent identical success-failure trials is known not to have occurred
through trial t0 , then the additional number of trials (beyond t0 ) needed to produce
the first success is a geometric ( p) random variable (as was the total number of
trials required from the beginning). Similarly, if an exponential (α) waiting time is
known not to have been completed by time t0 , then the additional waiting time to
260 Chapter 5 Probability: The Mathematics of Randomness
(
0 if x < 0
F(x) = β
(5.26)
1 − e−(x/α) if x ≥ 0
Weibull (α, β) 0 if x < 0
probability f (x) =
β x β−1 e−(x/α)β
(5.27)
density if x > 0
αβ
Weibull (α, β)
µ = E X = α0 1 + β1 (5.28)
mean
and variance
2
Weibull (α, β)
σ 2 = Var X = α 2 0 1 + β2 − 0 1 + β1 (5.29)
variance
5.2 Continuous Random Variables 261
f (x)
2.0 α = .5
β = .5
α = 1.0
1.0
α = 4.0
0
1.0 2.0 3.0 4.0 5.0 x
f (x)
2.0 α = .5
β =1
α = 1.0
1.0
α = 4.0
0
1.0 2.0 3.0 4.0 5.0 x
f (x)
3.0
β =4
α = .5
2.0
α = 1.0
1.0
α = 4.0
0
1.0 2.0 3.0 4.0 5.0 x
R∞
where 0(x) = 0 t x−1 e−t dt is the gamma function of advanced calculus. (For
integer values n, 0(n) = (n − 1)!.) These formulas for f (x), µ, and σ 2 are not par-
ticularly illuminating. So it is probably most helpful to simply realize that β controls
the shape of the Weibull distribution and that α controls the scale. Figure 5.16 shows
plots of f (x) for several (α, β) pairs.
Note that β = 1 gives the special case of the exponential distributions. For
small β, the distributions are decidedly right-skewed, but for β larger than about
3.6, they actually become left-skewed. Regarding distribution location, the form of
the distribution mean given in equation (5.28) is not terribly revealing. It is perhaps
more helpful that the median for the Weibull (α, β) distribution is
Weibull (α, β)
median Q(.5) = αe−(.3665/β) (5.30)
262 Chapter 5 Probability: The Mathematics of Randomness
So, for example, for large shape parameter β the Weibull median is essentially α.
And formulas (5.28) through (5.30) show that for fixed β the Weibull mean, median,
and standard deviation are all proportional to the scale parameter α.
Under the assumption that S can be modeled using a Weibull distribution with
the suggested characteristics, suppose that P[S ≤ 400] is needed. Using equation
(5.30),
428 = αe−(.3665/8.8)
α = 446
f(s)
Weibull density
β = 8.8, α = 446
Section 2 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. The random number generator supplied on a cal- (b) What adjustment to the grinding process (hold-
culator is not terribly well chosen, in that values ing the process standard deviation constant)
it generates are not adequately described by a dis- would increase the fraction of journal diam-
tribution uniform on the interval (0, 1). Suppose eters that will be in specifications? What ap-
instead that a probability density pears to be the best possible fraction of jour-
( nal diameters inside ± .0005 in. specifications,
k(5 − x) for 0 < x < 1 given the σ = .0004 in. apparent precision of
f (x) = the grinder?
0 otherwise
(c) Suppose consideration was being given to pur-
chasing a more expensive/newer grinder, capa-
is a more appropriate model for X = the next value
ble of holding tighter tolerances on the parts it
produced by this random number generator.
produces. What σ would have to be associated
(a) Find the value of k.
with the new machine in order to guarantee that
(b) Sketch the probability density involved here.
(when perfectly adjusted so that µ = 2.0000)
(c) Evaluate P[.25 < X < .75].
the grinder would produce diameters with at
(d) Compute and graph the cumulative probability
least 95% meeting 2.0000 in. ± .0005 in. spec-
function for X, F(x).
ifications?
(e) Calculate EX and the standard deviation of X .
5. The mileage to first failure for a model of military
2. Suppose that Z is a standard normal random vari-
personnel carrier can be modeled as exponential
able. Evaluate the following probabilities involv-
with mean 1,000 miles.
ing Z :
(a) Evaluate the probability that a vehicle of this
(a) P[Z < −.62] (b) P[Z > 1.06]
type gives less than 500 miles of service be-
(c) P[−.37 < Z < .51] (d) P[|Z | ≤ .47]
fore first failure. Evaluate the probability that
(e) P[|Z | > .93] (f) P[−3.0< Z <3.0]
it gives at least 2,000 miles of service before
Now find numbers # such that the following state-
first failure.
ments involving Z are true:
(b) Find the .05 quantile of the distribution of
(g) P[Z ≤ #] = .90 (h) P[|Z | < #] = .90
mileage to first failure. Then find the .90 quan-
(i) P[|Z | > #] = .03
tile of the distribution.
3. Suppose that X is a normal random variable with
6. Some data analysis shows that lifetimes, x (in 106
mean 43.0 and standard deviation 3.6. Evaluate the
revolutions before failure), of certain ball bearings
following probabilities involving X:
can be modeled as Weibull with β = 2.3 and α =
(a) P[X < 45.2] (b) P[X ≤ 41.7]
80.
(c) P[43.8 < X ≤ 47.0] (d) P[|X − 43.0| ≤ 2.0]
(a) Make a plot of the Weibull density (5.27)
(e) P[|X− 43.0|>1.7]
for this situation. (Plot for x between 0 and
Now find numbers # such that the following state-
200. Standard statistical software packages like
ments involving X are true:
MINITAB will have routines for evaluating this
(f) P[X < #] = .95 (g) P[X ≥ #] = .30
density. In MINITAB look under the “Calc/
(h) P[|X − 43.0| > #] = .05
Probability Distributions/Weibull” menu.)
4. The diameters of bearing journals ground on a (b) What is the median bearing life?
particular grinder can be described as normally dis- (c) Find the .05 and .95 quantiles of bearing life.
tributed with mean 2.0005 in. and standard devia-
tion .0004 in.
(a) If engineering specifications on these diame-
ters are 2.0000 in. ± .0005 in., what fraction
of these journals are in specifications?
264 Chapter 5 Probability: The Mathematics of Randomness
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
This section will further discuss the importance of this method. First, some
additional points about probability plotting are made in the familiar context where
f (x) is the standard normal density (i.e., in the context of normal plotting). Then
the general applicability of the idea is illustrated by using it in assessing the appro-
priateness of exponential and Weibull models. In the course of the discussion, the
importance of probability plotting to process capability studies and life data analysis
will be indicated.
A way of determining whether or not the students’ data support the use of
a normal model for U is to make a normal probability plot. Table 5.6 presents
the data collected by Ash, Davison, and Miyagawa. Table 5.7 shows some of the
calculations used to produce the normal probability plot in Figure 5.18.
Table 5.6
Weights of 100 U.S. Nickels
4.81 1 5.00 12
4.86 1 5.01 10
4.88 1 5.02 7
4.89 1 5.03 7
4.91 2 5.04 5
4.92 2 5.05 4
4.93 3 5.06 4
4.94 2 5.07 3
4.95 6 5.08 2
4.96 4 5.09 3
4.97 5 5.10 2
4.98 4 5.11 1
4.99 7 5.13 1
Table 5.7
Example Calculations for a Normal Plot of
Nickel Weights
i − .5 i − .5
i xi Qz
100 100
Example 13
(continued ) Standard normal quantile
2.0 3
2
4 3
2 5 4
3 75
0.0 9 7
4 73
3 44
32 3
2 2
–2.0
At least up to the resolution provided by the graphics in Figure 5.18, the plot
is pretty linear for weights above, say, 4.90 g. However, there is some indication
that the shape of the lower end of the weight distribution differs from that of a
normal distribution. Real nickels seem to be more likely to be light than a normal
model would predict. Interestingly enough, the four nickels with weights under
4.90 g were all minted in 1970 or before (these data were collected in 1988). This
suggests the possibility that the shape of the lower end of the weight distribution
is related to wear patterns and unusual damage (particularly the extreme lower
tail represented by the single 1964 coin with weight 4.81 g).
But whatever the origin of the shape in Figure 5.18, its message is clear. For
most practical purposes, a normal model for the random variable
will suffice. Bear in mind, though, that such a distribution will tend to slightly
overstate probabilities associated with larger weights and understate probabilities
associated with smaller weights.
Much was made in Section 3.2 of the fact that linearity on a Q-Q plot indicates
equality of distribution shape. But to this point, no use has been made of the fact
that when there is near-linearity on a Q-Q plot, the nature of the linear relationship
gives information regarding the relative location and spread of the two distributions
involved. This can sometimes provide a way to choose sensible parameters of a
theoretical distribution for describing the data set.
For example, a normal probability plot can be used not only to determine whether
some normal distribution might describe a random variable but also to graphically
pick out which one might be used. For a roughly linear normal plot,
5.3 Probability Plotting 267
The line eye-fit to the plot further suggests appropriate values for the mean and
standard deviation: µ ≈ 10.8 and σ ≈ 2.1. (Direct calculation with the data in
Table 5.8 gives a sample mean and standard deviation of, respectively, l̄ ≈ 10.9
and s ≈ 1.9.)
Table 5.8
Measured Thread Lengths for 25 U-Bolts
Thread Length
(.001 in. over Nominal) Tally Frequency
6 1
7 0
8 3
9 0
10 4
11 10
12 0
13 6
14 1
268 Chapter 5 Probability: The Mathematics of Randomness
Example 14
(continued ) 3.0
2.0
Standard normal quantile
1.0
Intercept ≈ 10.8
0
–1.0
Difference ≈ 2.1
–2.0
Line eye-fit to plot
–3.0
6 7 8 9 10 11 12 13 14 15 16
Thread length (.001 in. above nominal)
99 1
(2.3%)
98 2 + 2σ
95 5
90 10
(15.9%)
+σ
80 20
70 30
60 40
50 50 σ
40 60
30 70
20 80
–σ
(15.9%)
10 90
5 95
–2 σ
2 98
(2.3%)
1 99
0.5 99.5
S
T
E 0.2 99.8 –3σ
P
(0.135%)
1 VALUE
2 FREQUENCY
Figure 5.20 Thread length data plotted on a capability analysis form (used with permission of
Reynolds Metals Company)
270 Chapter 5 Probability: The Mathematics of Randomness
0
0 8 8 8 9 9
1 0 0 0 0 0 2 2 2 2 3 3 4 4 4 4 4 4
1 5 6 7 7 7 7 8 8 9 9
2 0 1 2 2 2 2 3 4
2 6 7 8 9 9 9
3 0 2 2 2 4 4
3 6 6 7 7
4 2 3
4 5 6 7 8 8
5
5
6
6
7 0
7
8
8 7
x = − ln(1 − p)
I That is, − ln(1 − p) = Q( p), the p quantile of this distribution. Thus, for data
x1 ≤ x2 ≤ · · · ≤ xn , an exponential probability plot can be made by plotting the
ordered pairs
Points to plot
i − .5
for an exponential xi , − ln 1 − (5.31)
probability plot n
Figure 5.22 is a plot of the points in display (5.31) for the service time data. It
shows remarkable linearity. Except for the fact that the third- and fourth-largest
service times (both 48 seconds) appear to be somewhat smaller than might be
predicted based on the shape of the exponential distribution, the empirical service
time distribution corresponds quite closely to the exponential distribution shape.
5
Exponential quantile
0
0 10 20 30 40 50 60 70 80 90
About 7.5 About 24
Data quantile
Example 15 As was the case in normal-plotting, the character of the linearity in Figure
(continued ) 5.22 also carries some valuable information that can be applied to the modeling
of the random variable T . The positioning of the line sketched onto the plot
indicates the appropriate location of an exponentially shaped distribution for T ,
and the slope of the line indicates the appropriate spread for that distribution.
As introduced in Definition 17, the exponential distributions have positive
density f (x) for positive x. One might term 0 a threshold value for the dis-
tributions defined there. In Figure 5.22 the threshold value (0 = Q(0)) for the
exponential distribution with α = 1 corresponds to a service time of roughly 7.5
seconds. This means that to model a variable related to T with a distribution
exactly of the form given in Definition 17, it is
S = T − 7.5
That is, an exponential model for S ought to have an associated spread that is
16.5 times that of the exponential distribution with α = 1.
So ultimately, the data in Figure 5.21 lead via exponential probability plotting
to the suggestion that
S = T − 7.5
= the excess of the next time required to complete a postage stamp sale
over a threshold value of 7.5 seconds
be described with the density
1 e−(s/16.5) for s > 0
f (s) = 16.5 (5.32)
0 otherwise
.06
Probability density
.04
Density for T
.02
Density for
S = T – 7.5
10 20 30 40 50
Service time (sec)
Points to plot !
for a fixed β i − .5 1/β
xi , − ln 1 − (5.33)
Weibull plot n
P[Y ≤ y] = P[X ≤ e y ]
y /α)β
= 1 − e−(e
Reading α and β If data in hand are consistent with a (0-threshold) Weibull (α, β) model, a reasonably
from a 0-threshold linear plot with
Weibull plot
1. slope β and
2. horizontal axis intercept equal to ln(α)
may be expected.
3
3 9.4
4 5.3
4 9.2, 9.4
5 1.3, 2.0, 3.2, 3.2, 4.9
5 5.5, 7.1, 7.2, 7.5, 9.2
6 1.0, 2.4, 3.8, 4.3
6 7.3, 7.7
stress. They were taken from Statistical Models and Methods for Lifetime Data
by J. F. Lawless.
Consider the Weibull modeling of
Table 5.9 shows some of the calculations needed to use display (5.35) to produce
Figure 5.25. The near-linearity of the plot in Figure 5.25 suggests that a (0-
threshold) Weibull distribution might indeed be used to describe R. A Weibull
shape parameter of roughly
1 − (−4)
I β ≈ slope of the fitted line ≈ ≈ 9.6
4.19 − 3.67
and thus
I α ≈ 59
appears appropriate.
276 Chapter 5 Probability: The Mathematics of Randomness
2.0
Line eye-fit to plot
1.0
About 3.67
0
ln(–ln(1 – p))
About 4.19
–1.0
About 4.08
–2.0
–3.0
–4.0
.99
0
1.
.95
.90
.80
.50
.70
.60
.50
.40
.30
.20
.10
.05
.04
.03
.02
.01
1 2 3 4 5 10 20 30 40 50 100 200 300400500 1,000 10,000
Section 3 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. What is the practical usefulness of the technique of 2. Explain how an approximate mean µ and standard
probability plotting? deviation σ can be read off a plot of standard normal
quantiles versus data quantiles.
278 Chapter 5 Probability: The Mathematics of Randomness
3. Exercise 3 of Section 3.2 refers to the chemical (b) Use the method of display (5.35) and investi-
process yield data of J. S. Hunter given in Exercise gate whether the Weibull distribution might be
1 of Section 3.1. There you were asked to make a used to describe bearing load life. If a Weibull
normal plot of those data. description is sensible, read appropriate param-
(a) If you have not already done so, use a computer eter values from the plot. Then use the form
package to make a version of the normal plot. of the Weibull cumulative probability function
(b) Use your plot to derive an approximate mean given in Section 5.2 to find the .05 quantile of
and a standard deviation for the chemical pro- the bearing load life distribution.
cess yields. 5. The data here are from the article “Fiducial Bounds
4. The article “Statistical Investigation of the Fatigue on Reliability for the Two-Parameter Negative Ex-
Life of Deep Groove Ball Bearings” by J. Leiblein ponential Distribution,” by F. Grubbs (Technomet-
and M. Zelen (Journal of Research of the National rics, 1971). They are the mileages at first failure
Bureau of Standards, 1956) contains the data given for 19 military personnel carriers.
below on the lifetimes of 23 ball bearings. The units
are 106 revolutions before failure. 162, 200, 271, 320, 393, 508, 539, 629,
706, 777, 884, 1008, 1101, 1182, 1462,
17.88, 28.92, 33.00, 41.52, 42.12, 45.60, 1603, 1984, 2355, 2880
48.40, 51.84, 51.96, 54.12, 55.56, 67.80,
68.64, 68.64, 68.88, 84.12, 93.12, 98.64, (a) Make a histogram of these data. How would
105.12, 105.84, 127.92, 128.04, 173.40 you describe its shape?
(b) Plot points (5.31) and make an exponential
(a) Use a normal plot to assess how well a normal probability plot for these data. Does it appear
distribution fits these data. Then determine if that the exponential distribution can be used
bearing load life can be better represented by to model the mileage to failure of this kind of
a normal distribution if life is expressed on the vehicle? In Example 15, a threshold service
log scale. (Take the natural logarithms of these time of 7.5 seconds was suggested by a similar
data and make a normal plot.) What mean and exponential probability plot. Does the present
standard deviation would you use in a normal plot give a strong indication of the need for a
description of log load life? For these parame- threshold mileage larger than 0 if an exponen-
ters, what are the .05 quantiles of ln(life) and tial distribution is to be used here?
of life?
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
Example 17 the data displayed in Table 3.4 (see page 74) and Figure 3.9 suggest, for exam-
(continued ) ple, that a sensible value for P[X = 18 and Y = 18] might be 34 1
, the relative
frequency of this pair in the data set. Similarly, the assignments
2
P[X = 18 and Y = 17] =
34
P[X = 14 and Y = 9] = 0
Table 5.10
f (x, y) for the Bolt Torque Problem
y x 11 12 13 14 15 16 17 18 19 20
20 2/34 2/34 1/34
19 2/34
18 1/34 1/34 1/34 1/34 1/34
17 2/34 1/34 1/34 2/34
16 1/34 2/34 2/34 2/34
15 1/34 1/34 3/34
14 1/34 2/34
13 1/34
Properties of a The probability function given in tabular form in Table 5.10 has two properties
joint probability that are necessary for mathematical consistency. These are that the f (x, y) values
function for X and Y are each in the interval [0, 1] and that they total to 1. By summing up just some
of the f (x, y) values, probabilities associated with X and Y being configured in
patterns of interest are obtained.
Example 17 Consider using the joint distribution given in Table 5.10 to evaluate
(continued )
P[X ≥ Y ] ,
P[|X − Y | ≤ 1] ,
and P[X = 17]
Take first P[X ≥ Y ], the probability that the measured bolt 3 torque is at least
as big as the measured bolt 4 torque. Figure 5.27 indicates with asterisks which
possible combinations of x and y lead to bolt 3 torque at least as large as the
5.4 Joint Distributions and Independence 281
bolt 4 torque. Referring to Table 5.10 and adding up those entries corresponding
to the cells that contain asterisks,
P[X ≥ Y ] = f (15, 13) + f (15, 14) + f (15, 15) + f (16, 16)
+ f (17, 17) + f (18, 14) + f (18, 17) + f (18, 18)
+ f (19, 16) + f (19, 18) + f (20, 20)
1 1 3 2 1 17
= + + + + ··· + =
34 34 34 34 34 34
Similar reasoning allows evaluation of P[|X − Y | ≤ 1]—the probability that
the bolt 3 and 4 torques are within 1 ft lb of each other. Figure 5.28 shows
combinations of x and y with an absolute difference of 0 or 1. Then, adding
probabilities corresponding to these combinations,
P[|X − Y | ≤ 1] = f (15, 14) + f (15, 15) + f (15, 16) + f (16, 16)
+ f (16, 17) + f (17, 17) + f (17, 18) + f (18, 17)
18
+ f (18, 18) + f (19, 18) + f (19, 20) + f (20, 20) =
34
x 11 12 13 14 15 16 17 18 19 20
y
20 *
19 * *
18 * * *
17 * * * *
16 * * * * *
15 * * * * * *
14 * * * * * * *
13 * * * * * * * *
Figure 5.27 Combinations of bolt 3
and bolt 4 torques with x ≥ y
x 11 12 13 14 15 16 17 18 19 20
y
20 * *
19 * * *
18 * * *
17 * * *
16 * * *
15 * * *
14 * * *
13 * * *
Figure 5.28 Combinations of bolt 3
and bolt 4 torques with |x − y| ≤ 1
282 Chapter 5 Probability: The Mathematics of Randomness
Example 17 Finally, P[X = 17], the probability that the measured bolt 3 torque is 17 ft lb,
(continued ) is obtained by adding down the x = 17 column in Table 5.10. That is,
Finding marginal In bivariate problems like the present one, one can add down columns in a two-
probability functions way table giving f (x, y) to get values for the probability function of X , f X (x). And
using a bivariate one can add across rows in the same table to get values for the probability function
joint probability of Y , f Y (y). One can then write these sums in the margins of the two-way table.
function So it should not be surprising that probability distributions for individual random
variables obtained from their joint distribution are called marginal distributions.
A formal statement of this terminology in the case of two discrete variables is
next.
Definition 20 The individual probability functions for discrete random variables X and
Y with joint probability function f (x, y) are called marginal probability
functions. They are obtained by summing f (x, y) values over all possible
values of the other variable. In symbols, the marginal probability function for
X is
X
f X (x) = f (x, y)
y
X
f Y (y) = f (x, y)
x
Example 17 Table 5.11 is a copy of Table 5.10, augmented by the addition of marginal
(continued ) probabilities for X and Y . Separating off the margins from the two-way table
produces tables of marginal probabilities in the familiar format of Section 5.1.
For example, the marginal probability function of Y is given separately in Table
5.12.
5.4 Joint Distributions and Independence 283
Table 5.11
Joint and Marginal Probabilities for X and Y
y x 11 12 13 14 15 16 17 18 19 20 f Y (y)
20 2/34 2/34 1/34 5/34
19 2/34 2/34
18 1/34 1/34 1/34 1/34 1/34 5/34
17 2/34 1/34 1/34 2/34 6/34
16 1/34 2/34 2/34 2/34 7/34
15 1/34 1/34 3/34 5/34
14 1/34 2/34 3/34
13 1/34 1/34
f X (x) 1/34 1/34 1/34 2/34 9/34 3/34 4/34 7/34 5/34 1/34
Table 5.12
Marginal
Probability
Function for Y
y f Y (y)
13 1/34
14 3/34
15 5/34
16 7/34
17 6/34
18 5/34
19 2/34
20 5/34
Distribution 1 Distribution 2
x x
y 1 2 3 y 1 2 3
3 .4 0 0 .4 3 .16 .16 .08 .4
2 0 .4 0 .4 2 .16 .16 .08 .4
1 0 0 .2 .2 1 .08 .08 .04 .2
.4 .4 .2 .4 .4 .2
example, in the bolt (X ) torque situation, a technician who has just loosened bolt
3 and measured the torque as 15 ft lb ought to have expectations for bolt 4 torque
(Y ) somewhat different from those described by the marginal distribution in Table
5.12. After all, returning to the data in Table 3.4 that led to Table 5.10, the relative
frequency distribution of bolt 4 torques for those components with bolt 3 torque
of 15 ft lb is as in Table 5.13. Somehow, knowing that X = 15 ought to make a
probability distribution for Y like the relative frequency distribution in Table 5.13
more relevant than the marginal distribution given in Table 5.12.
Table 5.13
Relative Frequency Distribution for Bolt 4
Torques When Bolt 3 Torque Is 15 ft lb
Definition 21 For discrete random variables X and Y with joint probability function f (x, y),
the conditional probability function of X given Y = y is the function of x
f (x, y)
f X |Y (x | y) = X
f (x, y)
x
5.4 Joint Distributions and Independence 285
f (x, y)
f Y |X (y | x) = X
f (x, y)
y
and
The conditional
f (x, y)
probability function f Y |X (y | x) = (5.37)
for Y given X = x f X (x)
And formulas (5.36) and (5.37) are perfectly sensible. Equation (5.36) says
Finding conditional that starting from f (x, y) given in a two-way table and looking only at the row
distributions from specified by Y = y, the appropriate (conditional) distribution for X is given by
a joint probability the probabilities in that row (the f (x, y) values) divided by their sum ( f Y (y) =
P
function x f (x, y)), so that they are renormalized to total to 1. Similarly, equation (5.37)
says that looking only at the column specified by X = x, the appropriate conditional
distribution for Y is given by the probabilities in that column divided by their sum.
Example 17 To illustrate the use of equations (5.36) and (5.37), consider several of the condi-
(continued ) tional distributions associated with the joint distribution for the bolt 3 and bolt 4
torques, beginning with the conditional distribution for Y given that X = 15.
From equation (5.37),
f (15, y)
f Y |X (y | 15) =
f X (15)
conditional distribution for Y given in Table 5.14. Comparing this to Table 5.13,
indeed formula (5.37) produces a conditional distribution that agrees with
intuition.
286 Chapter 5 Probability: The Mathematics of Randomness
y f Y |X (y | 15)
1 9 1
13 ÷ =
34 34 9
1 9 1
14 ÷ =
34 34 9
3 9 3
15 ÷ =
34 34 9
2 9 2
16 ÷ =
34 34 9
2 9 2
17 ÷ =
34 34 9
f (18, y)
f Y |X (y | 18) =
f X (18)
Consulting Table 5.11 again leads to the conditional distribution for Y given that
X = 18, shown in Table 5.15. Tables 5.14 and 5.15 confirm that the conditional
distributions of Y given X = 15 and given X = 18 are quite different. For exam-
ple, knowing that X = 18 would on the whole make one expect Y to be larger
than when X = 15.
Table 5.15
The Conditional
Probability Function for
Y Given X = 18
y f Y |X (y | 18)
14 2/7
17 2/7
18 1/7
20 2/7
To make sure that the meaning of equation (5.36) is also clear, consider the
conditional distribution of the bolt 3 torque (X ) given that the bolt 4 torque is 20
5.4 Joint Distributions and Independence 287
f (x, 20)
f X |Y (x | 20) =
f Y (20)
Table 5.16
The Conditional Probability
Function for X Given Y = 20
x f X |Y (x | 20)
2 5 2
18 ÷ =
34 34 5
2 5 2
19 ÷ =
34 34 5
1 5 1
20 ÷ =
34 34 5
The bolt torque example has the feature that the conditional distributions for Y
given various possible values for X differ. Further, these are not generally the same
as the marginal distribution for Y . X provides some information about Y , in that
depending upon its value there are differing probability assessments for Y . Contrast
this with the following example.
and
Example 18 Intuition dictates that (in contrast to the situation of X and Y in Example 17) the
(continued ) variables U and V don’t furnish any information about each other. Regardless of
what value U takes, the relative frequency distribution of bolt 4 torques in the hat
is appropriate as the (conditional) probability distribution for V , and vice versa.
That is, not only do U and V share the common marginal distribution given in
Table 5.17 but it is also the case that for all u and v, both
fU |V (u | v) = fU (u) (5.38)
and
f V |U (v | u) = f V (v) (5.39)
Equations (5.38) and (5.39) say that the marginal probabilities in Table 5.17
also serve as conditional probabilities. They also specify how joint probabilities
for U and V must be structured. That is, rewriting the left-hand side of equation
(5.38) using expression (5.36),
f (u, v)
= fU (u)
f V (v)
That is,
(The same logic applied to equation (5.39) also leads to equation (5.40).) Ex-
pression (5.40) says that joint probability values for U and V are obtained by
multiplying corresponding marginal probabilities. Table 5.18 gives the joint prob-
ability function for U and V .
Table 5.17
The Common Marginal
Probability Function for U
and V
u or v fU (u) or f V (v)
13 1/34
14 3/34
15 5/34
16 7/34
17 6/34
18 5/34
19 2/34
20 5/35
5.4 Joint Distributions and Independence 289
Table 5.18
Joint Probabilities for U and V
v u 13 14 15 16 17 18 19 20 f V (v)
5 15 25 35 30 25 10 25
20 5/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
2 6 10 14 12 10 4 10
19 2/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
5 15 25 35 30 25 10 25
18 5/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
6 18 30 42 36 30 12 30
17 6/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
7 21 35 49 42 35 14 35
16 7/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
5 15 25 35 30 25 10 25
15 5/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
3 9 15 21 18 15 6 15
14 3/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
1 3 5 7 6 5 2 5
13 1/34
(34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2 (34)2
Example 18 suggests that the intuitive notion that several random variables are
unrelated might be formalized in terms of all conditional distributions being equal to
their corresponding marginal distributions. Equivalently, it might be phrased in terms
of joint probabilities being the products of corresponding marginal probabilities. The
formal mathematical terminology is that of independence of the random variables.
The definition for the two-variable case is next.
Definition 22 Discrete random variables X and Y are called independent if their joint prob-
ability function f (x, y) is the product of their respective marginal probability
functions. That is, independence means that
If formula (5.41) does not hold, the variables X and Y are called dependent.
(Formula (5.41) does imply that conditional distributions are all equal to their cor-
responding marginals, so that the definition does fit its “unrelatedness” motivation.)
U and V in Example 18 are independent, whereas X and Y in Example 17
are dependent. Further, the two joint distributions depicted in Figure 5.29 give an
example of a highly dependent joint distribution (the first) and one of independence
Independence of (the second) that have the same marginals.
observations in The notion of independence is a fundamental one. When it is sensible to model
statistical studies random variables as independent, great mathematical simplicity results. Where
290 Chapter 5 Probability: The Mathematics of Randomness
engineering data are being collected in an analytical context, and care is taken to
make sure that all obvious physical causes of carryover effects that might influence
successive observations are minimal, an assumption of independence between
observations is often appropriate. And in enumerative contexts, relatively small
(compared to the population size) simple random samples yield observations that
can typically be considered as at least approximately independent.
Example 18 Again consider putting bolt torques on slips of paper in a hat. The method of torque
(continued ) selection described earlier for producing U and V is not simple random sam-
pling. Simple random sampling as defined in Section 2.2 is without-replacement
sampling, not the with-replacement sampling method used to produce U and V .
Indeed, if the first slip is not replaced before the second is selected, the proba-
bilities in Table 5.18 are not appropriate for describing U and V . For example,
if no replacement is done, since only one slip is labeled 13 ft lb, one clearly
wants
1
f (13, 13) =
(34)2
f V |U (13 | 13) = 0
1
f V |U (13 | 13) = f V (13) =
34
99
f V |U (13 | 13) =
3,399
5.4 Joint Distributions and Independence 291
f (u, v)
f V |U (v | u) =
fU (u)
so that
f (u, v) = f V |U (v | u) fU (u)
99 1
f (13, 13) = ·
3,399 34
99 1
≈
3,399 34
and so
1 1
f (13, 13) ≈ ·
34 34
For this hypothetical situation where the population size N = 3,400 is much
larger than the sample size n = 2, independence is a suitable approximate de-
scription of observations obtained using simple random sampling.
Where several variables are both independent and have the same marginal
distributions, some additional jargon is used.
For example, the joint distribution of U and V given in Table 5.18 shows U and V
to be iid random variables.
The standard statistical examples of iid random variables are successive mea-
surements taken from a stable process and the results of random sampling with
When can replacement from a single population. The question of whether an iid model is
observations be appropriate in a statistical application thus depends on whether or not the data-
modeled as iid? generating mechanism being studied can be thought of as conceptually equivalent
to these.
292 Chapter 5 Probability: The Mathematics of Randomness
ZZ
P[(X, Y ) ∈ R] = f (x, y) dx dy (5.42)
R
Imagine that the true value of S will be measured with a (very imprecise) analog
stopwatch, producing the random variable
f(s, r)
r
10
0 s
–10 10 20 30
0
Example 19 mean s and standard deviation .5.) Thus, equation (5.43) specifies a mathemati-
(continued ) cally legitimate joint probability density.
To illustrate the use of a joint probability density in finding probabilities, first
consider evaluating P[R > S]. Figure 5.31 shows the region in the (s, r )-plane
where f (s, r ) > 0 and r > s. It is over this region that one must integrate in
order to evaluate P[R > S]. Then,
ZZ
P[R > S] = f (s, r ) ds dr
R
Z ∞ Z ∞
= f (s, r ) dr ds
0 s
Z ∞ Z ∞
1 −s/16.5 1 2
= e √ e−(r −s) /2(.25) dr ds
0 16.5 s 2π(.25)
Z ∞
1 −s/16.5 1
= e ds
0 16.5 2
1
=
2
(once again using the fact that the integral in braces is a normal (mean s and
standard deviation .5) probability).
As a second example, consider the problem of evaluating P[S > 20]. Figure
5.32 shows the region over which f (s, r ) must be integrated in order to evaluate
P[S > 20]. Then,
ZZ
P[S > 20] = f (s, r ) ds dr
R
Z ∞ Z ∞
= f (s, r ) dr ds
20 −∞
Z ∞ Z ∞
1 −s/16.5 1 2
= e √ e−(r −s) /2(.25) dr ds
20 16.5 −∞ 2π(.25)
Z ∞
1 −s/16.5
= e ds
20 16.5
= e−20/16.5
≈ .30
5.4 Joint Distributions and Independence 295
r
r
20
Region
3 where
10 s > 20
2
0 10 20 s
Figure 5.31 Region where f (s, r) > 0 Figure 5.32 Region where f (s, r) > 0
and r > s and s > 20
The last part of the example essentially illustrates the fact that for X and Y with
joint density f (x, y),
Z x Z ∞
F(x) = P[X ≤ x] = f (t, y) dy dt
−∞ −∞
Definition 25 The individual probability densities for continuous random variables X and
Y with joint probability density f (x, y) are called marginal probability
densities. They are obtained by integrating f (x, y) over all possible values of
the other variable. In symbols, the marginal probability density function for
X is
Z ∞
f X (x) = f (x, y) dy (5.44)
−∞
Z ∞
f Y (y) = f (x, y) dx (5.45)
−∞
296 Chapter 5 Probability: The Mathematics of Randomness
Compare Definitions 20 and 25 (page 282). The same kind of thing is done
for jointly continuous variables to find marginal distributions as for jointly discrete
variables, except that integration is substituted for summation.
Example 19 Starting with the joint density specified by equation (5.43), it is possible to arrive
(continued ) at reasonably explicit expressions for the marginal densities for S and R. First
considering the density of S, Definition 25 declares that for s > 0,
Z ∞
1 −s/16.5 1 2
f S (s) = e √ e−(r −s) /2(.25) dr
−∞ 16.5 2π(.25)
1 −s/16.5
= e
16.5
That is, the form of f (s, r ) was chosen so that (as suggested by Example 15)
S has an exponential distribution with mean α = 16.5.
The determination of f R (r ) is conceptually no different than the determi-
nation of f S (s), but the details are more complicated. Some work (involving
completion of a square in the argument of the exponential function and recogni-
tion of an integral as a normal probability) will show the determined reader that
for any r ,
Z ∞
1 2
f R (r ) = √ e−(s/16.5)−((r −s) /2(.25)) ds
0 16.5 2π(.25)
1 1 1 r
I = 1−8 − 2r exp − (5.46)
16.5 33 2,178 16.5
fR(r)
–5 0 5 10 15 20 25 30 r
The marginal density for R derived from equation (5.43) does not belong to
any standard family of distributions. Indeed, there is generally no guarantee that the
process of finding marginal densities from a joint density will produce expressions
for the densities even as explicit as that in display (5.46).
Definition 26 For continuous random variables X and Y with joint probability density
f (x, y), the conditional probability density function of X given Y = y,
is the function of x
f (x, y)
f X |Y (x | y) = Z ∞
f (x, y) dx
−∞
f (x, y)
f Y |X (y | x) = Z ∞
f (x, y) dy
−∞
and
f(x, y)
y
x
Expressions (5.47) and (5.48) are formally identical to the expressions (5.36) and
Geometry of (5.37) relevant for discrete variables. The geometry indicated by equation (5.47) is
conditional that the shape of f X |Y (x | y) as a function of x is determined by cutting the f (x, y)
densities surface in a graph like that in Figure 5.34 with the Y = y-plane. In Figure 5.34,
the divisor in equation (5.47) is the area of the shaded figure above the (x, y)-plane
below the f (x, y) surface on the Y = y plane. That division serves to produce a
function of x that will integrate to 1. (Of course, there is a corresponding geometric
story told for the conditional distribution of Y given X = x in expression (5.48)).
Example 19 In the service time example, it is fairly easy to recognize the conditional distribu-
(continued ) tion of R given S = s as having a familiar form. For s > 0, applying expression
(5.48),
f (s, r ) 1 −s/16.5
f R|S (r | s) = = f (s, r ) ÷ e
f S (s) 16.5
1
I
2
f R|S (r | s) = √ e−(r −s) /2(.25) (5.49)
2π(.25)
That is, given that S = s, the conditional distribution of R is normal with mean
s and standard deviation .5.
This realization is consistent with the bell-shaped cross sections of f (s, r )
shown in Figure 5.30. The form of f R|S (r | s) given in equation (5.49) says that
the measured excess service time is the true excess service time plus a normally
distributed measurement error that has mean 0 and standard deviation .5.
5.4 Joint Distributions and Independence 299
It is evident from expression (5.49) (or from the way the positions of the bell-
shaped contours on Figure 5.30 vary with s) that the variables S and R ought to be
called dependent. After all, knowing that S = s gives the value of R except for a
normal error of measurement with mean 0 and standard deviation .5. On the other
hand, had it been the case that all conditional distributions of R given S = s were
the same (and equal to the marginal distribution of R), S and R should be called
independent. The notion of unchanging conditional distributions, all equal to their
corresponding marginal, is equivalently and more conveniently expressed in terms
of the joint probability density factoring into a product of marginals. The formal
version of this for two variables is next.
Definition 27 Continuous random variables X and Y are called independent if their joint
probability density function f (x, y) is the product of their respective marginal
probability densities. That is, independence means that
If expression (5.50) does not hold, the variables X and Y are called dependent.
Example 20 Residence Hall Depot Counter Service Times and iid Variables
(Example 15 revisited )
Returning once more to the service time example of Jenkins, Milbrath, and Worth,
consider the next two excess service times encountered,
To the extent that the service process is physically stable (i.e., excess service times
can be thought of in terms of sampling with replacement from a single population),
an iid model seems appropriate for S1 and S2 . Treating excess service times as
300 Chapter 5 Probability: The Mathematics of Randomness
Example 20 marginally exponential with mean α = 16.5 thus leads to the joint density for S1
(continued ) and S2 :
1
e−(s1 +s2 )/16.5 if s1 > 0 and s2 > 0
f (s1 , s2 ) = (16.5)2
0 otherwise
Section 4 Exercises ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
1. Explain in qualitative terms what it means for two What is the mean of this conditional distribu-
random variables X and Y to be independent. What tion?
advantage is there when X and Y can be described 3. A laboratory receives four specimens having iden-
as independent? tical appearances. However, it is possible that (a
2. Quality audit records are kept on numbers of major single unknown) one of the specimens is contam-
and minor failures of circuit packs during burn-in inated with a toxic material. The lab must test the
of large electronic switching devices. They indicate specimens to find the toxic specimen (if in fact one
that for a device of this type, the random variables is contaminated). The testing plan first put forth
by the laboratory staff is to test the specimens one
X = the number of major failures at a time, stopping when (and if) a contaminated
specimen is found.
and
Define two random variables
Y = the number of minor failures
X = the number of contaminated specimens
can be described at least approximately by the ac-
and
companying joint distribution.
Y = the number of specimens tested
/
y x 0 1 2 Let p = P[X = 0] and therefore P[X = 1] =
0 .15 .05 .01 1 − p.
1 .10 .08 .01
(a) Give the conditional distributions of Y given
X = 0 and X = 1 for the staff’s initial test-
2 .10 .14 .02
ing plan. Then use them to determine the joint
3 .10 .08 .03
probability function of X and Y . (Your joint
4 .05 .05 .03 distribution will involve p, and you may sim-
ply fill out tables like the accompanying ones.)
(a) Find the marginal probability functions for
both X and Y — f X (x) and f Y (y).
(b) Are X and Y independent? Explain. y f Y |X (y | 0) y f Y |X (y | 1)
(c) Find the mean and variance of X—EX and
Var X. 1 1
(d) Find the mean and variance of Y —EY and 2 2
Var Y . 3 3
(e) Find the conditional probability function for Y , 4 3
given that X = 0—i.e., that there are no major
circuit pack failures. (That is, find f Y |X (y | 0).)
5.4 Joint Distributions and Independence 301
the cumulative probability function for T is (d) The probability that the system has failed by
time t is
F(t) = 1 − P[X > t and Y > t]
P[X ≤ t and Y ≤ t]
so that your answer to (b) can be used to find
the distribution for T . Use your answer to (b) Find this probability using your answer to
and some differentiation to find the probability part (a).
density for T . What kind of distribution does (e) Now, as before, let T be the time until the
T have? What is its mean? system fails. Use your answer to (d) and some
Suppose now that the system is a parallel system differentiation to find the probability density
(i.e., one that fails only when both subsystems fail). for T . Then calculate the mean of T .
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
U = g(X, Y, . . . , Z ) (5.51)
In some special simple cases, it is possible to figure out exactly what distribution U
inherits from X, Y, . . . , Z .
Table 5.19
Relative Frequency Distribution of Plate
Thicknesses
Table 5.20
Relative Frequency Distribution of Slot
Widths
quency distribution in Table 5.19; a relative frequency distribution for the slot
widths measured on a lot of machined blocks is given in Table 5.20.
If a plate is randomly selected and a block is separately randomly selected,
a natural joint distribution for the random variables
U =Y −X
Notice that taking the extremes represented in Tables 5.19 and 5.20, U is guaran-
teed to be at least .153 − .150 = .003 in. but no more than .156 − .148 = .008 in.
In fact, much more than this can be said. Looking at Table 5.21, one can see that
the diagonals of entries (lower left to upper right) all correspond to the same value
of Y − X. Adding probabilities on those diagonals produces the distribution of
U given in Table 5.22.
304 Chapter 5 Probability: The Mathematics of Randomness
f X (x) .4 .3 .3
Table 5.22
The Probability Function for the
Clearance U = Y − X
u f (u)
.003 .06
.004 .12 = .06 + .06
.005 .26 = .08 + .06 + .12
.006 .26 = .08 + .12 + .06
.007 .22 = .16 + .06
.008 .08
Example 21 involves a very simple discrete joint distribution and a very simple
function g—namely, g(x, y) = y − x. In general, exact complete solution of the
problem of finding the distribution of U = g(X, Y, . . . , Z ) is not practically possi-
ble. Happily, for many engineering applications of probability, approximate and/or
partial solutions suffice to answer the questions of practical interest. The balance
of this section studies methods of producing these approximate and/or partial de-
scriptions of the distribution of U , beginning with a brief look at simulation-based
methods.
The authors further give some uncertainty values associated with each of the terms
appearing on the right side of equation (5.52) for an example set of measured
values of the variables. These are given in Table 5.23.
Table 5.23
Reported Uncertainties in the Measured Inputs
to Collector Efficiency
Example 22 Plugging the measured values from Table 5.23 into formula (5.52) produces
(continued ) a measured efficiency of about .44. But how good is the .44 value? That is, how
do the uncertainties associated with the measured values affect the reliability of
the .44 figure? Should you think of the calculated solar collector efficiency as .44
plus or minus .001, or plus or minus .1, or what?
One way of approaching this is to ask the related question, “What would
be the standard deviation of Efficiency if all of C through To were independent
random variables with means approximately equal to the measured values and
standard deviations related to the uncertainties as, say, half of the uncertainty
values?” (This “two sigma” interpretation of uncertainty appears to be at least
close to the intention in the original article.)
Printout 1 is from a MINITAB session in which 100 normally distributed
realizations of variables C through To were generated (using means equal to
measured values and standard deviations equal to half of the corresponding
uncertainties) and the resulting efficiencies calculated. (The routine under the
“Calc/Random Data/Normal” menu was used to generate the realizations of
C through To . The “Calc/Calculator” menu was used to combine these val-
ues according to equation (5.52). Then routines under the “Stat/Basic Statis-
tics/Describe” and “Graph/Character Graphs/Stem-and-Leaf” menus were used
to produce the summaries of the simulated efficiencies.) The simulation produced
a roughly bell-shaped distribution of calculated efficiencies, possessing a mean
value of approximately .437 and standard deviation of about .009. Evidently,
if one continues with the understanding that uncertainty means something like
“2 standard deviations,” an uncertainty of about .02 is appropriate for the nominal
efficiency figure of .44.
5 41 58899
10 42 22334
24 42 66666777788999
39 43 001112233333444
(21) 43 555556666777889999999
40 44 00000011122333444
5.5 Functions of Several Random Variables 307
23 44 555556667788889
8 45 023344
2 45 7
1 46 0
The beauty of Example 22 is the ease with which a simulation can be employed
to approximate the distribution of U . But the method is so powerful and easy to use
that some cautions need to be given about the application of this whole topic before
going any further.
Practical cautions Be careful not to expect more than is sensible from a derived probability distri-
bution (“exact” or approximate) for
U = g(X, Y, . . . , Z )
The output distribution can be no more realistic than are the assumptions used
to produce it (i.e., the form of the joint distribution and the form of the function
g(x, y, . . . , z)). It is all too common for people to apply the methods of this section
using a g representing some approximate physical law and U some measurable
physical quantity, only to be surprised that the variation in U observed in the real
world is substantially larger than that predicted by methods of this section. The fault
lies not with the methods, but with the naivete of the user. Approximate physical
laws are just that, often involving so-called constants that aren’t constant, using
functional forms that are too simple, and ignoring the influence of variables that
aren’t obvious or easily measured. Further, although independence of X, Y, . . . , Z
is a very convenient mathematical property, its use is not always justified. When
it is inappropriately used as a model assumption, it can produce an inappropriate
distribution for U . For these reasons, think of the methods of this section as useful
but likely to provide only a best-case picture of the variation you should expect
to see.
EU = a0 + a1 EX + a2 EY + · · · + an EZ (5.53)
308 Chapter 5 Probability: The Mathematics of Randomness
and variance
Example 21 Consider again the situation of the clearance involved in placing a steel plate
(continued ) in a machined slot on a steel block. With X , Y , and U being (respectively) the
plate thickness, slot width, and clearance, means and variances for these variables
can be calculated from Tables 5.19, 5.20, and 5.22, respectively. The reader is
encouraged to verify that
Now, since
U = Y − X = (−1)X + 1Y
so that
√
I Var U = .0013 in.
It is worth the effort to verify that the mean and standard deviation of the clearance
produced using Proposition 1 agree with those obtained using the distribution of
U given in Table 5.22 and the formulas for the mean and variance given in Section
5.1. The advantage of using Proposition 1 is that if all that is needed are EU
and Var U , there is no need to go through the intermediate step of deriving the
5.5 Functions of Several Random Variables 309
1 1 1
X= X1 + X2 + · · · + Xn
n n n
The mean of an
1 1 1 1
average of n iid EX = EX1 + EX2 + · · · + E X n = n µ =µ (5.55)
random variables n n n n
and
1 2
2
1 2
The variance of an Var X = n
Var X 1 + n1 Var X 2 + · · · + Var X n
average of n iid n
1 2 2 σ2 (5.56)
random variables =n σ =
n n
Since σ 2 /n is decreasing in n, equations (5.55) and (5.56) give the reassuring picture
of X having a probability distribution centered at the population mean µ, with spread
that decreases as the sample size increases.
Example 23 consider what means and standard deviations are associated with the probability
(continued ) distributions of the sample average, S, of first the next 4 and then the next 100
excess service times.
S1 , S2 , . . . , S100 are, to the extent that the service process is physically stable,
reasonably modeled as independent, identically distributed, exponential random
variables with mean α = 16.5. The exponential distribution with mean α = 16.5
has variance equal to α 2 = (16.5)2 . So, using formulas (5.55) and (5.56), for the
first 4 additional service times,
E S = α = 16.5 sec
s
p α2
Var S = = 8.25 sec
4
E S = α = 16.5 sec
s
p α2
Var S = = 1.65 sec
100
Relationships (5.55) and (5.56), which perfectly describe the random behavior
of X under random sampling with replacement, are also approximate descriptions of
the behavior of X under simple random sampling in enumerative contexts. (Recall
Example 18 and the discussion about the approximate independence of observations
resulting from simple random sampling of large populations.)