Análisis de Componentes Principales y Funciones Empíricas Ortogonales
Análisis de Componentes Principales y Funciones Empíricas Ortogonales
general,forzado por de
las fases el aumento dede
los modos gases de
variabilidad natural co
Análisis de Componentes
siguen exactamente un mismoPrincipales y por lo
patrón siempre),
combinación
Funciones deEmpíricas
los modos descritos acá (más otros).
Ortogonales
n cierta semejanza a ENOS, el cual debe reflejar una parte
que no se captura con el modo principal (otro ‘sabor’ de
teratura). Este modo, por construcción, debe estar en
ue su sentido físico debe interpretarse con cuidado. En
e variabilidad natural como ENOS no son canónicos (no
atrón siempre), por lo que puede decirse que son una
s acá (más otros).
Aplicaciones
Ejemplo de motivación
fi
∑
x′k = ekm um k = 1, . . . , K
m=1
Los vectores dependientes del tiempo um(t) representan las nuevas variables
buscadas.
El coe cientes ekm indica el peso (contribución) del modo m en la variable xk.
fi
Equation
th
11.2 expresses the transformation of a (K × 1) data vector x to a vector u of
. Ifm[E] eigenvector, ineeigenvectors
contains all K ,
mwhich the original data jointly
of [Sx ] (assuming it isexhibit maximum
nonsingular) variability.
as its columns,
Equivalentemente,
vector u will alsolas nuevas variables um1).seEquation
componen de las variables originales…
esulting Puthave dimension
another way, (K the×eigenvectors 11.2 sometimes
define a new iscoordinate
called system
analysis formula forthe x# , data.
expressing that the
In particular,!K data can be analyzed, or summarized in
the !orthogonal matrix [E] whose columns are
T !
s of the principal components. em x = the
um =Reversing xk ! m = 1! "in" "Equation
ekmtransformation ! M# 11.2, the (11.1)
# (Equation 9.49) defines the rigid rotation
x can be reconstructed from the principal components according to
k=1
Notice
De formathatmatricial:
each of the M eigenvectorsx# = [E] contains u ( one elementu = %E& T !
x!
pertaining to each of the K
(11.5)
$K×1% $K×K%$K×1%
variables, xk . Similarly, each realization of the mth principal component in Equation 11.1
ch is is computed
obtained from from a particular
Equation
which set
by of
11.2simultaneous
is the observations
multiplying on theof the byvariables
left K
matrix-notation [E] xk . That
and using
representation theof is,
M each
= K of lin
the
ogonality M principal
Para ser estrictos,
property components
of this cuandois
matrix a sort
hablamos
(Equation of weighted
9.42). deThe average
componentes of
reconstruction the # values. (PCs),
principales
of xx expressed Although nosthe
11ofun the form of Equation 11.1 (i.e., here the matrix [E] is square, wi
k
C centraremos
H A P T E R ! en Principal Component (EOF) Analysis
weights
Equation 11.5(the ekm s) do
is sometimes not sum
called
subconjunto thetode1,Mtheir
synthesis squares
formula.
modos, dobastante
ojaláIf because
the full set of
of the
menor M a=scaling
KK(M PCs <<convention
K).
sed#einm #the = synthesis,
1. (Note that the reconstruction
a fixed scalingis convention
complete and forexact, since )memRm2of=the
the weights 1 (cf.
linear combi-
ation La11.4).
ideaines
nations que
M < estos
IfEquation K PCs pocos
11.1 (usually
allowsM modos describan
corresponding
the maximum al M
tovariance
the máximo
largest eleigenvalues)
constraintsistema original.
defining are
the PCs to be
or
, the Esreconstruction
meaningful.)
decir, If the is data
approximate,
sample consists of n observations (and therefore of n data vectors
x, or n rows in the data matrix !M[X]), there will be n values for each of the principal
!x# ≈ [E]
x ≈ e u u! ( k = 1! " " " ! K! (11.6a) (11.6b)
components, or new variables, $K×1%um .$K×M%
k Each$M×1%
km of these constitutes a single-number index of the
m
m=1
resemblance between the eigenvector em and the corresponding individual data vector x.
but Geometrically,
En improves
PCA, se buscan as the the first eigenvector,
también
number Mmodosof eused
1 , points
linealmente
PCs (or, in the accurately,
directionentre
independientes
more (inasthe r(u i, uj) = 0
sí:theK-dimensional
sum of the
!
space of
corresponding x ) in which the because
eigenvalues, data vectorsof jointly 11.4)
Equation exhibitincreases.
the mostBecausevariability.
[E] Thisonly
has first
> Se requiere
eigenvector is de one
the unaassociated
transformación with lineal
the del espacio
largest eigenvalue, que$diagonalize
. The second la eigenvector
matriz de
M columns, and operates on a truncated PC vector u of dimension1 (M × 1), Equation 11.6
covarianza (o with
correlación) y que asocie máxima varianza a los primerosbe modos.
ise2called
, associated
the truncated thesynthesis
second-largest
formula. eigenvalue
The original $2 , (in
is constrained
the case of to Equation perpendicular
11.5) or
to e1 (Equation
approximated (for 9.48),
Equation but 11.6)
subject to this constraint
uncentered data x canit easily
will align
be in the direction
obtained by adding in
Ver caso 2D…
which
back thethe
vectorx! vectors
of sample exhibit
means; their
thatnext
is, bystrongest
reversingvariations. Subsequent eigenvectors
Equation 9.33.
GF601 - semestre otoño 2022
nonzero eigenvalues, where K is the number of rows and columns in [A]. Each eigenvector
AnMatemáticamente,
eigenvalue !, andlaanmatriz
will be dimensioned %Keigenvector,
× 1&.E If(EOFs)
[A] is e seof a square
singular
obtiene at least
dematrixone of
resolver [A]itsun
are a scalar will
eigenvalues
problema andauto-
de nonzer
be
vector,zero,
valores with
parathe unacorresponding
respectively, satisfying
matriz cuadradaeigenvectors
the equation being arbitrary.
AKxK (operador lineal): Synonymous terminology that
M A T R I X also
is sometimes A L Gused
E B Rfor
A Reigenvalues
EVIEW and eigenvectors includes characteristic values 42
and characteristic vectors, latent values"A#e and latent
= !e$vectors, and proper values and proper(9.46a
vectors.
T
transformation !E"
Because each eigenvector x defines a rigid rotation
is defined to have of unitthelength,
K-dimensional coordinate
the dot product axes of
of any
or Si
equivalently
A no es singular y, porisloThistanto, K “territory”
called
eigenvector anwith
eigenspace.
itself one. If,invertible,
space in covers
addition,existen
thethe same matrixpares [A]deisvalores
as (λi) ythen
the original
symmetric, vectores
coordinate
its
but using
(ei)eigenvectors
propios. themutually
Si are different
además set of axesso
A esorthogonal,
simétrica, defined
losthat by the
vectores solutions
propios seránto Equation
ortogonales:9.46.
The K eigenvalue-eigenvector %"A# − !"I#&e
pairs
!
=
contain 0$the same information as the (9.46b
matrix [
from which they were computed, T and so 1$ cani =bej regarded as a transformation of [A]. Th
where 0 equivalence
is a vector consisting entirely ei eof
j =zeros. For every eigenvalue and eigenvector (9.48) pa
can be expressed, again for 0$[A]i $=symmetric,
j' as the spectral decomposition,
that can Jordan
be found to satisfy Equation 9.46, any scalar multiple of the eigenvector, ce, wi
decomposition,
also satisfy
Orthogonal the vectors
equation of together
unit length withare that
said eigenvalue.
to be orthonormal. Consequently, for definiteness
(This terminology has
is usual
Denothing to require
forma to
más thatthethe
do general:
with eigenvectors
Gaussian, or “normal”
!A" havedistribution.)
= !E"!#"!E" unit T length, The orthonormality property is (9.50
analogous to Equation 8.66, expressing theorthogonality of the sine and cosine functions.
For many statistical applications, eigenvalues $1 0and ··· 0
0 eigenvectors are calculated for real (9.47
"e" = 1'
(not containing complex or imaginary numbers) 2 0 · · · 0matrices,
0 $symmetric
T such as covariance
or correlation matrices. Eigenvalues = and!E" 0 0 $3 of
eigenvectors · · ·such
0 matrices
!E" have
&sign, a number of (9.50
This restriction removes the ambiguity only % up% to% a change in since
%% is that their eigenvaluesif a vector
important and remarkable properties. The %% of %%these
first % properties
satisfies Equation 9.46 then its negative, −e will also. % %
and eigenvectors are real-valued. Also, as 0just0 noted, 0 · · ·the eigenvectors of symmetric
Ifmatrices
[A] is are nonsingular there will be K eigenvalue-eigenvector pairs ! and e wi
$
orthogonal. That is, their dot products with eachK other are zero, so kthat theyk
nonzero eigenvalues,
are mutually where KinisK-dimensional
perpendicular the number ofspace. rows and columns in [A]. Each eigenvecto
will be Oftenso that !#" denotes
dimensioned × a diagonal matrix whoseatnonzero elements are the K eigenvalu
the %K × K& matrix [E] is formed, the K columns of which are the eigenvectorswill b
%K 1&. If [A] is singular least one of its eigenvalues
of [A]. It is illuminating to consider also the equivalent of Equation 9.50 in summati
zero, ewith
k . theis,corresponding eigenvectors being arbitrary. Synonymous terminology th
That
notation,
is sometimes also used for eigenvalues and eigenvectors
GF601 -"semestre otoño 2022 # includes characteristic value
k
sum of these !Ek " matrices, where the weights are the corresponding eigenvalues. Hence
the
La matriz valores propiosofΛacorresponde
spectraldedecomposition matrix is analogous toAthe
a la matriz Fourier decomposition
(covarianza o correlación) of
en a
function
el nuevoor data series
espacio (Equation
de nido 8.62a),
por E. Es decir,with the eigenvalues
contiene playing the pero
la misma información, role of the
vista
Fourier amplitudes
desde otro ángulo. and
Λ esthe !Ek " matrices
diagonal porque corresponding to the cosine functions.
la base E es ortogonal.
Other consequences of the equivalence of the information on the two sides of Equa-
tion 9.50que
Puesto A y Λtocuanti
pertain the eigenvalues. The conjunta
can la varianza first of these is
del sistema, se cumple que:
K
' K
'
tr!A" = ak&k = $k = tr!#"% (9.52)
k=1 k=1
This
Si A relationship
es la matriz deis particularly
covarianza deimportant whenla[A]
K variables, sumais de
a covariance matrix,
los K valores in será
propios which
case
igualitsa diagonal
la suma deelements ak&k aredethe
las varianzas K variances.
todas Equation
las variables 9.52 says the sum of these
originales.
variances is given by the sum of the eigenvalues of the covariance matrix.
EnThe
estesecond consequence
caso, los of Equation
valores propios 9.50 forde
—ordenados themayor
eigenvalues
a menor—is corresponden a
las varianzas de los modos independientes (anti-correlacionados entre sí) y dan
K
(
cuenta de la variabilidad conjunta
det!A" = K variables.
de las $k = det!#"& (9.53)
k=1
→ Esto es precisamente lo que se busca!
which is consistent with the property that at least one of the eigenvalues of a singu-
Los modos quedan entonces de nidos por los vectores propios ei. Estos representan
lar matrix (that has zero determinant) will be zero. A real symmetric matrix with all
una base ortogonal
eigenvalues y serán
nonnegative las EOF
is called buscadas.
positive definite.
gridpoints, then the dimensionality of the data vector k!k x is given by the produc
nto a rectangular array of numbers having n rows, each corresponding to
observation, and with
En la práctica… may
firsteach ofseem
K elements a oflittle
the K columns x are strange at nfirst,
observations
containing all of ittheisfirst
observations conventional
variable, the in second
multivariat
K el
observations
iables. This arrangement the of
n ×the
alsoofconvenient second
from the
K numbers variable, and the
standpoint
in the multivariateofdatalast K elements
arranging of x will be calcu
the covariances obse
a matrix, Lth variable.
the Equation 9.4Since
into the L different
a square arrayvariables
called the generally
samplewill be measured
covariance in un
matrix,
it will almost
always be appropriate to base the PCA of such data on the correlat
T
xThe x x · · · x
1 dimension
1"1 1"2 of [R], 1"K and of the matrix of s1!1eigenvectors
s1!2 s1!3 · · ·[E], s1!'will then be !K
xApplication
2 x2"1 xof2"2PCA to this
T · · · x2"K kind of correlation s2!1 matrixs2!2 s2!3will· · produce
· s2!' principal c
successively
maximizing the joint variance s of the
s L variables
s · · · s in a way that co
T
!X$ = x3 = x3"1 x3"2 · · · x3"K %
#S$ = (9.2)3!1 3!2 3!3
3!' %
correlations
both between and among these %variables
% %at the K %
locations. This
% %
%% %% is sometimes
procedure %
%% %% called combined
% %% %% %%
PCA, or CPCA.
%%
xnT Figure xn"111.5
xn"2 ·illustrates
· · xn"K the structure ofsk!1 k!2 sk!3 · · · matrix
the scorrelation sk!' (left) and
of eigenvectors (right) for PCA of vector field data. The first K rows of [
th th
or observations ofthe That is, the
thecorrelations
form shown in covariance
Equation
between the9.1first
shave
k!' is displayed
been
of the stacked
L in
variablesthe at
k row
these and
locations
' colum
and
d a rectangular array, called The
matrix. a matrix,
sample withcovariance
n rows andmatrix, K columns.
or variance-covariance matrix, is
he first of theresolver
Permite two subscripts of the scalar
el problema elements of a matrix denotes
de auto-valores
to the sample (Pearson) correlation matrix (see Figure 3.25), with the re
}
and para
the second indicates
la matriz de the column
covarianza number
S = n so,
-1 XXT : for example, x 3"2 is the
ervations of the second corresponding
of the K [R variables.
1,1]
elements
[R1,2In of[Rthe
] this book ] two matrices
1,Lmatrices will be being given by Equation F
1/2 e1 e2 e3 e4 eM Var
sk!' /#&s
uare brackets, as a pictorial = E(&s
S Ereminder
k!k Λ'!'that($ the . The
symbolK covariances
within represents sk!k in the diagonal positions b
}
ay. left and lower-right corners of the sample covariance matrix are sim
[R ] [R ] [R ] Se
ix [X] in Equation 9.2 corresponds
variances.
[R] =
2,1 exactly
The 2,2 to a conventional
remaining, data
off-diagonal,
2,L table
[E] = elements are covariances amon Var
splay,Lainsolución
which each entrega
columnuna matriz
pertains one of the Λ
to diagonal variables
KxK considered,
resents one of the n and the values
observations. Its below can
contents andalso to be thevisualized
left of theor diagonal positions duplicat
de valores propios y una matriz EKxK cuyas
and K-dimensional
rically within an abstract to the right. space, with each of the n rows
columnas
point.
K = vectores
contienen
The simplest
2 [Link]
example
The pairo of
losisvariance-covariance
The correspondientes
numbers
EOFs.
a L,1
[R data
]
in
matrix
each of the
[RL,L] matrix
[RL,2] for bivariate
rows locates a
it describes how the observations are dispersed around their (vec
is also known as the dispersio
data, which
point
} Var
plane. The collection of these n points on the plane defines a scatterplot
ata.
K-dimensional space defined by the K variables. The diagonal element
FIGURE 11.5 Illustration GF601of- semestre
the structures of the correlation matrix and of the matrix of
otoño 2022
for the maximum amount of the joint variability of x (and therefore also of x)—are
uniquely defined by the eigenvectors of the covariance matrix of x, [S]. In particular, the
Las thseries de tiempo de las variables nuevas (um) resultarán de la proyección de los
m principal component, um is obtained as the projection of the data vector x! onto the
datos
mth originales X esobre
eigenvector, , la base de nida por E :
m
K
!
um = eTm x! = ekm xk! ! m = 1! " " " ! M# (11.1)
k=1
Notice that each of the M eigenvectors contains one element pertaining to each of the K
Las ejes de la nueva base (em) se orientan de manerath que el 1 er modo (m=1) explique
variables, xk . Similarly, each realization of the m principal component in Equation 11.1
la is computed
máxima from a particular
variabilidad [Link] of the K variables xk . That is, each of
the M principal components is a sort of weighted average of the xk values. Although the
El weights
2do eje (the ekm s) dode
se orienta notmanera
sum to tal
1, their
que squares do because
sea ortogonal of ythe
al 1ro scaling convention
explique también el
#em # =de
máximo 1. varianza
(Note that a fixed(luego
posible scaling
delconvention
1ro). Y así for the weights em of the linear combi-
sucesivamente.
nations in Equation 11.1 allows the maximum variance constraint defining the PCs to be
Demeaningful.)
forma análoga,If the udata
1 essample consists
la nueva serieofde n observations
tiempo con (and mayor therefore of n(componente
varianza data vectors
x, or n rows
principal). u2 esinla the
seriedata
conmatrix [X]),mayor
segunda there varianza
will be ne values for each de
independiente of uthe
1. uprincipal
3 …
components, or new variables, um . Each of these constitutes a single-number index of the
Lasresemblance
varianzasbetween
de cada the modo son elos
eigenvector m and the corresponding individual data vector x.
valores propios encontrados en la matriz
diagonalGeometrically, the first
Λ. Por lo tanto, eigenvector, ae1la, points
Λ corresponde matrizindethe direction (in
covarianza the K-dimensional
de U.
space of x! ) in which the data vectors jointly exhibit the most variability. This first
eigenvector is the one associated with the largest eigenvalue, $1 . The second eigenvector
e2 , associated with the second-largest eigenvalue $2 , is constrained to be perpendicular
to e1 (Equation 9.48), but subject to this constraint
GF601 - semestre otoño 2022
it will align in the direction in
fi
'm 'm
Rm2 = K × 100% = K × 100%& (11.4)
! !
'k sk(k
k=1 k=1
472 Here
CHR
Cuidado 2P T E R ! 11 Principal Component (EOF) Analysis
Acon is el
used in the(no
léxico same
es sense that is familiar from linear regression (see Section 6.2).
estándar)…
The total variation exhibited by the original data is completely represented in (or accounted
for by) the full set of K um ’s, in the sense that the sum of the variances of the centered
dataTABLE
x# (and11.2 A partialalso
therefore guideoftothe
synonymous terminology
uncentered associated
variables x), )kwith
sk(k ,PCA.
is equal to the sum of
variances )emm 'm . ofEigenvector
the Eigenvectors, the principal component variables u.
Principal Principal Component
Equation 11.2 expresses Elements,theetransformation
k!m of a (K ×um1) data vector
Components, x# to
Elements, ui!ma vector u of
PCs. If [E] contains all K eigenvectors of [Sx ] (assuming it is nonsingular) as its columns,
the EOFs
resulting vector u will Loadings
also have dimension Empirical Orthogonal
(K × 1). Equation 11.2 Scores
sometimes is called
Variables
the analysis formula for x# , expressing that the data can be analyzed, or summarized in
Modes
terms of of
theVariation
principal Coefficients
components. Reversing the transformationAmplitudes in Equation 11.2, the
dataPattern
x# canVectors Pattern from
be reconstructed Coefficients
the principal components according Expansion
to Coefficients
Principal Axes Empirical Orthogonal Coefficients
Weights x# = [E] u ( (11.5)
$K×1% $K×K%$K×1%
Principal Vectors
Proper Functions
which is obtained from Equation 11.2 by multiplying on the left by [E] and using the
Principal Directions
orthogonality property of this matrix (Equation 9.42). The reconstruction of x# expressed
by Equation 11.5 is sometimes called the synthesis formula. If the full set of M = K PCs
is used in the synthesis, the reconstruction is otoño
GF601 - semestre 2022 and exact, since ) R 2 = 1 (cf.
complete
fi
Volvamos al caso de 2 variables, particularmente útil pues su transformación lineal se
puede visualizar.
X = [x1, x2]
r = 0.92
S = 185.47 110.84
110.84 77.58
E = 0.85 -0.53
0.53 0.85
Λ = 254.8 0
0 8.29
λ1/(λ1+λ2) = 0.968
(S y Λ en °F2)
En este caso x1, con mayor varianza, tiene mayor incidencia en la orientación de las
nuevas coordenadas de nidas por e1 y e2.
}
xT x [Rx ] · · [R · x ]
[R ] Second
2 [R] = 2"1 2,12"2 2,22"K 2,L
[E] = Variable
T
!X$ = x
3
= x3"1 x3"2 · · · x3"K %
(9.2)
% % % %
%% %% %% %%
xnT
[RL,1] [RL,2]
xn"1 xn"2 · · · xn"K
[RL,L]
} Lth
Variable
X30×100
R100×100
En la práctica, sólo se tienen min (n-1, K) modos efectivos (los restantes son nulos).
general, las fases de los modos de variabilidad natural como ENOS no son canónicos (no
siguen exactamente
Ejemplo: Modos un1,mismo
2 y 3 patrón
de lassiempre),
anomalíaspor mensuales
lo que puedededecirse que son una
TSM (global)
combinación de los modos descritos acá (más otros).
5. Compare (cuantitativamente) la serie de tiempo del modo principal de TSM’ con el indice
típico de ENOS, calculado como la TSM’ media en la región ‘Niño 3.4’. Para apoyar esta
comparación, haga un nuevo mapa de correlación que muestre la influencia de este índice
en la TSM’ de todo el globo. ¿Cómo contrasta el modo principal de TSM’, con el que
obtiene a partir de series normalizadas de TSM’? Haga nuevamente un mapa de
correlación para describir su resultado (procure usar la misma paleta la colores y rango en
todos los gráficos anteriores).
El figura siguiente muestra la correlación entre la TSM’ de todo el globo y el índice Niño 3.4. El
patrón es casi idéntico al del modo principal de TSM’ descrito en el punto anterior. De hecho
ambas series tienen una muy alta correlación (0.92), como también se puede apreciar en el
diagrama dispersión entre las observaciones de ambas series.
Los tres modos principales que se obtienen a partir de los datos de TSM’ normalizadas se
muestran a continuación. Se pueden reconocer los modos obtenidos en primera instancia
(datos sin normalizar), sobre todos las configuraciones ‘ENOS’, que en este caso emergen
como como el 2do y 3er componentes. Cabe destacar la importancia de este modo natural de
variabilidad, que emerge en el PCA aunque
GF601 se anule otoño
- semestre la distribución
2022 de varianza de la TSM con
PSA).
SLP’ mensual 1960-2014 (solo HS)
ta parte le daremos un uso más pragmático al análisis EOF. Queremos estimar la precipit
(P) en un número grande de localidades en Chile
GF601 - semestre para el periodo anterior a 1979. Para
otoño 2022
fi
but somewhat poleward of earlier results (Kidson 1975; near 408S and below normal activity within the circum-
Rogers and van Loon 1982). We can have more con- polar trough near Antarctica. Thus, stronger (weaker)
Introduction fidence in the stability of the present As withresults,
other aspects
as they of are
the atmospheric
westerlies circulation,
near 558S imply more (fewer) cyclones in the
The regular passage of based on 15 years
anticyclones, cyclones,of recent
and weather system behavior can vary considerably
state-of-the-art numerical circumpolar regions from
the time-averaged picture. Marked variations in the lo-
and fewer (more) in middle lati-
ir associated fronts accounts for
analyses. most of the shorter- tudes.
m ‘‘weather’’ variationsThe seensecond
in middle latitudes. cation and strength of storm tracks can The
occurrelation
within between PC 1 and anticyclones (Fig.
pressure EOF (Fig. 6b) shows a center 7d) is less coherent, the main response being a corre-
ather systems develop in preferred regions where at- individual seasons (e.g., Lau 1988). Anomalous local
of action
spheric and physiographic in the
features southeast
combine to pro-Pacific near
weather 608S,can
patterns 1208Woccur that lation systems
when weather with subtropical
depart highs to the north of NZ and in
e conditions favoringistheir
out growth.
of phase Thewith variations
time-aver- in the
from theirsubtropical
usual [Link]
For example, during the
the Indian 1982 The time series for pressure PC 1 is
Ocean.
d behavior of weatheron one side
systems has and extensively onwinter
beenAntarctica whenWhile
the other. very few cyclones crossed
the south- New Zealand
also correlated with cyclone PC 2 (r 5 20.60, Fig. 3b)
east Pacific
cumented. In the Southern center(SH),
Hemisphere is common
cy- (NZ), rainfall was well below normal around the coun-
to all studies, some studies and anticyclone PC 5 (r 5 20.49, Fig. 4e). However,
nes tend to form and(e.g.,
intensify in middle latitudes try. In contrast, a preference for cyclones to track across
Kidson 1988b) show additional wavelike structure
d near the principal upper-tropospheric jet streams, NZ during the winters of 1990–92asproduced the cyclone EOFs are not well defined, these corre-
unusually
for Fthis mode
IG. 7. (a)–(c)involving
Distributionextrema near NZcoefficient
of the correlation and thebetween
Ant- lations
cyclone yield
track littleanomalies
density physicalandinsight beyond
the three leading representing
pressure PCs. (d)–
d migrate eastward and poleward as they mature and wet and stormy conditions (Sinclair 1996b).
arctic peninsula. Eigenvector Previous
ay (Taljaard 1967; Jones and Simmonds 1993; Sin-
3 (Fig. studies
6c) describes
(f)ofAs
storm atrackabut
in (a)–(c) for anticyclones.
general
variability association
have fo- between low (high) pressure anom-
ir 1994, 1995). On thewave trainanticyclones
other hand, extendinghave from cused
the Pacific across
GF601
on variations in-South
semestre
transient aliesotoño
eddy activity 2022
and its
and (anti-) cyclones.
3. SPATIAL STRUCTURE
Ejemplos en la literatura: modos de variabilidad de baja frecuencia en el Hemisferio Sur
Figure 1(a) and (b) shows the spatial patterns of PSA1 and PSA2 modes. They are the second and third
EOF patterns of seasonal mean 500-hPa height anomalies for the SH with all seasons pooled together.
Figure 1. (a) EOF 2 (PSA1) and (b) EOF 3 (PSA2) of the 500-hPa seasonal mean height anomalies. Contour interval five
non-dimensional units. Zero contours are omitted, positive loadings are shaded. (c) Same as (a) but for EOF 3 of the 200-hPa eddy
streamfunction anomalies (psi200) with zonal means removed, and (d) same as (c) but for EOF 4
Atlantic with weak negative correlations located to the south (Figure 3(d)). The pattern (Figure 3(d)) is
GF601 - semestre
quite similar otoño
to that of2022
Figure 2(d) south of about 10°N except that correlations were stronger in the
Ejemplos en la literatura: modos de variabilidad de baja frecuencia en el Hemisferio Sur
ABSTRACT
The characteristics of subseasonal circulation variability over the South Pacific are examined using 10-day
lowpass-filtered 700-hPa geopotential height NCEP–NCAR reanalysis data. The extent to which the variability
in each season is characterized by recurrent geographically fixed circulation regimes and/or oscillatory behavior
is determined. Two methods of analysis (a K-means cluster analysis and a cross-validated Gaussian mixture
model) both indicate three to four geographically fixed circulation regimes in austral fall, winter, and (to some
extent) spring. The spatial regime structures are found to be quite similar in each season; they resemble the so-
called Pacific–South American (PSA) patterns discussed in previous studies and often referred to as PSA 1 and
PSA 2. Oscillatory behavior is investigated using singular spectrum analysis. This identifies a predominantly
stationary wave with a period of about 40 days and a spatial structure similar to PSA 1; it is most pronounced
in winter and spring and exhibits a noticeable eastward drift as it decays. The power spectrum of variability is
otherwise well approximated by a red spectrum, together with enhanced broader-band 15–30-day variability.
The results presented herein indicate that low-frequency variability over the South Pacific is not dominated
by a propagating wave whose quadrature phases are PSA 1 and PSA 2, as hitherto described. Rather, it is found
that the variability is well described by the occurrence of three to four geographically fixed circulation regimes,
with a (near) 40-day oscillation that is predominantly stationary in space. The potential subseasonal predictability
implied by this duality is discussed. Only during austral spring is a strong correlation found between El Niño
and the frequency of occurrence of the circulation regimes.
• Identi camos las EOFs como los vectores propios del sistema S E = E Λ.
• Si las variables originales son valores en un campo 2D (o ND), las EOFs indican la
in uencia (dirección y amplitud) de cada modo en una región (patrones
espaciales).
fi
fl
N pixeles representan un
área pequeña.
[R] =
[R2,1] [R2,2] [R2,L]
[E] = } Second
Variable
FIGURE 11.5 Illustration of the structures of the correlation matrix and of the matrix of eigenvectors
En PCA
for este of
caso,
vectorsi field
las variables son de
data. The basic datadiversa
consist naturaleza (con unidades
of multiple observations of L y magnitudes
variables at each of
locations, soesthe
distintivas),
K dimensions oflaboth
fundamental [R] and [E] are
normalización uso×de
o el!KL alaThe
KL". matriz R.
correlation matrix consists of
!K ×K" submatrices containing the correlations between sets of the L variables jointly at the K locations.
The submatrices located on the diagonal of [R] are the ordinary correlation matrices for each of the
L variables. The off-diagonal submatrices contain
GF601 correlation
- semestre coefficients, but are not symmetrical and
otoño 2022
PC1, - SOI
• Una vez hecho el análisis PC/EOF en una cierta región, se puede ver la in uencia
de un cierto modo en un área mayor o en otra región. Ésto se puede evaluar
calculando las correlaciones entre un cierto modo y la variable x en la región de
interés, o mediante regresión lineal utilizando las PCs como predictores.
• De forma similar, se puede expandir el periodo abarcado por las PCs (o rellenar
algún periodo faltante) utilizando algún predictor de referencia que explique buena
parte de la varianza de los PCs.
Ver ejemplos…
fl
Uso de distintos dominios, interpolaciones y reconstrucciones
ANUARY 2013 GARREAUD ET AL.
EOF1 (U850,region)
R2k (%)
o, el espectro de valores propios no discrimina claramente estos modos, ni los distingue 3de
(Figura 2). Para explicar el grueso de la varianza conjunta de Z300 se deben considerar 88
%). 2
PCA de Z300 en HS
1
0
0 10 20 30 40 50 60 70 80
componentes
Baja
Figura 2: Varianza explicada, y error típico (North et al.) dein
los uencia
primeros 88de Z300
modos sobre Tmax
de variabilidad de Z300.
¿Problemas de datos?
¿In uencia relativa mayor de la TSM en
b. Calcule, ahora, el coeficiente de correlación de Pearson entre las PCs de los 2 modos principales de
la costa?
Z300’ y las anomalías de Tmax (Tmax’). Utilice mapas que indiquen la correlación para cada punto de
grilla (restrinja el análisis a las regiones comprendidas entre 30ºS y 40ºS). Describa brevemente los
resultados en relación a lo obtenido en a.
Figura 3: Coeficiente de correlación entre los modos 1 y 2 de Z300, y la Tmax en distintas regiones de centro y sur
de Chile. Panel de la izquierda muestra la varianza de Tmax explicada por el conjunto de los primeros 88 modos.
fi
Ejemplo: Relación entre la circulación sobre el Pací co y Tmax en Chile
GF601 - Otoño 2018
Figura 4: Patrones espaciales (EOF) del 1er y 2do modo de variabilidad para Z300’ y Tmax (análisis PCA-conjunto).
Este método PCA-conjunto también se podría utilizar para construir un modelo predictivo de Tmax. Es
este caso habría que generar un nuevo grupo de series ‘truncadas” de PCs, proyectando solamente los
datos de Z300 sobre los EOF-conjuntos (considerando los elementos de las EOF correspondientes a
Z300). Las PCs truncadas se utilizarían como predictores en un modelo de regresión de Tmax.
GF601 - semestre otoño 2022
fi
Principal Component (EOF) Analysis
Ejemplo: Reconstrucción de la precip. en Chile
tions
Se exhibited by the
quiere estimar field
P en unor fields grande
número being analyzed, and new
de localidades interpre-
en Chile para el periodo anterior a 1979.
data xguía
nalComo can para
be suggested by the nature
la reconstrucción, se of the linear
usarán combinations
series más largas (1960-2014) de sólo 13
tiveestaciones.
in compressing the data.
nvenient to calculate the PCs as linear combinations EOF − 1 of the anomalies EOF − 2
PC, u1 , is thatconjuntas
• EOF/PCs linear combination
de P of x! having the largest variance. The
al •components
Normalización um !de = 2! 3! " " " , −20
m varianzas are the linear combinations having −20
latitude
• Extension de PCs
−40 −40
les or(proyecciones
PCs—that is,del
thegrupo 1) u of u that will account successively
elements m
amount de PCs of−50x! (and therefore also of x)—are
of the joint variability
• Re-normalización
−50
y theextendidas
eigenvectors of the covariance
−60
matrix of x, [S]. In particular, the
−60
ponent, um is obtained
• Reconstrucción P the projection of the data vector x! onto the
de as
−0.05 0 0.05 −0.05 0 0.05
m, loading loading
scores
0 0
!M
the M! eigenvectors contains one element pertaining to each of the K
xk ≈ ekm um ! k = 1!th" " " !−50K! (11.6b) −50
larly, each m=1
realization of the m principal component
1960 1970 1980 1990 in Equation
2000 2010 11.11960 1970 1980 1990 2000 2010
Se quiere estimar P en un número grande de localidades en Chile para el periodo anterior a 1979.
Como guía para la reconstrucción, se usarán series más largas (1960-2014) de sólo 13
estaciones.
0.8
0.6
r
0.4
0.2
0
1 2 3 4 5 6 7 8 9 10 11 12
componentes
Convención de escala or
M
!
Las EOFs conforman una base ortonormal, es decir: ||e|| = 1 xk! ≈ ekm um ! k=
m=1
Eigenvalue
0.5
2
0.2
0.1
1
0.05
0 0.02
1 2 3 4 5 6 1 2 3 4 5 6
Principal Component Number Principal Component Number
FIGURE 11.8 Graphical displays of eigenvalue spectra; that is, eigenvalue magnitudes as a function
GF601
of the principal component number - semestre
(heavier otoñocircled
lines connecting 2022points), for a K = 6 dimensional
fi
fi
k
λCriterios
2
hereCriterios
is the más
closest eigenvalues
objetivos:
eigenvalue to areλ 2 (approximately)
, and n ∗
is
itude isindependent
applied to each univariate
grid point Gaussian distribu
to account for the
j objetivos: k converging longitudes poleward. The data over the NH
e number of independent observations in the sample,
north√ of 20 ° N are used to compute EOFs. Note that the
so known as the effective sample size, or the number ˆ k − (khere 2
• cálculo del error a partir de pruebas
degrees of freedom (Trenberth, 1984; Thiébaux and
no
examples n!
paramétricas
presented
( " ∼ N!0#
con
have 2(
datos
also been
k "#
sintéticos
used in Hannachi
• Cálculo
o de del error thea partir dere-sampleos
re-muestreos et al. (2006).
wiers, 1984). Forexperimentos
example,
" ! # 95% de
confidence interval (Monte-Carlo)
Figure 1 shows the spectrum of the covariance matrix
λ2k is>given >
H0: Las H0:
by λ2kPCsLas
1 or PCs2
±representan representan ruido
ruido sample along with their standard errors as given by the first
∗ . The effective
n
ze of a time series of length n involves in general equation of (18) with sample size n = 3 × 52 = 156.
• Análysis asintótico The leading two !
eigenvalues " nondegenerate and
seem
e autocorrelation structure of the series. For $ example,
separated from the rest,
2 overall
but 2 the spectrum looks
e sumDistribución
• y
of the>autocorrelation error muestral
function,
North et al.: Error típicok≥1 1 + de
2 λ:δλ ∼ λ (2/n*)
ρ(k), 1.2 ( k ∼ N ( k # ( k '
ovides a measure of the decorrelation time, and an in general smooth, which n makes truncation difficult.
(se asume Normalidad en
timate% of n∗ is given by (Thiébaux and
laZwiers,
dist. de 1984);
x) Figure 2 shows the first two EOFs. These EOFs explain
$n−1 &−1
n* : grados
= n 1 + 2 k=1 (1 −Note de libertad
k/n)ρ(k) however . (obs. thatindependientes)
there is a bias in the sample eigenvalues for finit
Eigenvalue spectrum
Another alternative is Equations
to use Monte 11.15 Carlo simulations,
and 11.16 are30large-sample approximations. In particu
ee for example
n*
Regla Björnsson
= general
n((1 - ρde 2 and +
1 )/(1
North Venegas
ρ 2 ))
et1 al.: 1997).(Wilks)
2 This can
e achieved by forming eigenvalues
surrogate data by will be overestimated
resampling a (will tend to be larger than their po
art ofSithe
el data
error típico
using terparts)
δλ ∼ λ (2/n)
randomisation. and
An0.5 the
es-1smallest
example would eigenvalues
25 will tend to be underestimated, an
n* = n(1select
e to randomly + 2 increase
aΣnsubsample
(1 - k/n)with ρ(k))
and decreasing
apply EOFs, sample size.
comparable o mayor a la diferencia
Eigenvalue (%)
20
en select another subsample etc. This operation, which
n beentre λ
repeated vecinos,
many times,
Using
éstos Equation
representarán
yields various
11.16a to construct a standard Gaussian variate provide
realisations
(Hannachi etfor al.)the distribution of the relative 15 error of the eigenvalue estimate,
themultipletos
eigenelements from which
degenerados de EOFs. one can estimate
e uncertainties. Another example would be to fix a
bsetDos
of variables then scramble them by breaking √
the
10
# $ %
o más modos “confundidos” (dentro ˆ
hronological order then apply EOFs, and so on. Further n!(k − (k " − 0 n (ˆ k − (k
ontedel error)
Carlo serán combinaciones
alternatives exist to assess uncertainty z = on √ 5
lineales = ∼ N!0# 1"'
2( k
2 (k
e spectrum
de los modosof the covariance
reales (de matrix. One could for
la población). 0
xample scramble blocks of the data, for example two- 0 10 20 30 40
three-year blocks of monthly Equation data11.17
keepingimplies thus some that Rank
arts of the autocorrelation structure, (e.g. Peng and Fife
Figureotoño
GF601 - semestre 1. Spectrum,
2022 in percentage, of the covariance matrix of winter
¿Para qué ?
Los V&V propios y descomposición lineal de X entrega información muy útil, pero
introduce limitaciones no necesariamente presentes en la naturaleza (ortogonalidad).
→ Como regla general, hay que tener cuidado al interpretar los modos secundarios.
ER = E(M) T
UR = ERT X
(a)
x1 x2
x2
e1
e1 + +
– ++
EOFs
e2
e2
x1
(b)
x1 x2
REOFs (orthogonal)
~
e2
x2 ~
e1
+ +
+
~
e1
~
e2 –
x1
(c)
~
e2 x2 ~
e1
+ +
+
REOFs (oblicua) ~
e1
~
e2 –
x1
EOFs
R-EOFs
(10 pcs)
PC2, -SOI
(series filtradas)
De forma a similar a PCA, CCA es una metodología que permite caracterizar múltiples
variables a partir de un grupo más pequeño de nuevas variables representativas del
conjunto, en este caso llamadas ‘variables canónicas’.
Aunque con resultados diferentes, esta técnica guarda cierta similitud con el análisis
EOF/PC conjunto.
C
# $ #
Var matrix¼of the J variables in y. ¼
variance-covariance $ The ½matrices !S , and !Syx # contain ð1
#
w S w , v S w , w RC # ½ I # xy
thetienen
Si se covariances between
dos grupos deall x e y,oflas
combinations
variables thevariables of x and theseelements
elementscanónicas of y, and
construyen
T
are related according to !S # = !S las .variables
#
[Rcomo
C ] is unadiagonal
the
A
transformación
CCA
matrix
transforms
lineal
ofxy
pairs
the
of
decanonical
yx
original centered
contenidas en cada uno de estos
correlations,
data vectors x!
and y !
into sets of new
grupos: 2 v and w , defined by the 3 dot products
variables, called canonical variates,r m0 0m & & & 0
C1
6 0# I r C2 ! 0 &&& 0 7
T !6 7 J%(
vm = am x 6
=0 a m"i 0 xi " m = 1" ' '
rC3 & & & 0 7 ' " min$I" (12.2a)
½RC # ¼ 6 i=1 7: ð1
6 .. .. .. . . .. 7
and 4. . . . . 5
0# J 0 0 &&& r CM
wm = bTm y! = bm"j yj! " m = 1" ' ' ' " min$I" J%& (12.2b)
e definition of the canonical vectors is reminiscent
j=1 of PCA, which finds a new orthonormal basi
multivariate data set (the eigenvectors of its covariance matrix), subject to a variance-maximizin
This construction of the canonical variates is similar to that of the principal components
In De
CCA, two new
(Equation
umanera bases are
11.1),
equivalente inadefined
that
los eachbyisthe
EOFs, a
los canonical
linear
coe vectors
combination
cientes am,i
a
(a
y b
m and
sort
m,j
b
ofm . However,
weighted
conforman una these
average)
nueva basis
of v
m
therbase
orthogonal
elements
de
nor
of of
the
dimensión
unit length.
respective
m, y serán
The
data
los
canonical
vectors
vectores x! variates
and y! are
. These
canónicos
the projections
vectors
asociados
ofgrupo
the
ofa weights,
cada amcentered
and
de bm , dat
and y0are called
onto the canonical
the canonical vectors.
vectors One be
and can data- and canonical-vector
expressed in matrix formpair need not
through the have the for
analysis
variables.
same dimension as the other. The vectors x! and am each have I elements, and the vectors
y! and bm each have J elements. The
v number
¼ ½A # of pairs,
T
x 0 M, of canonical variates that can ð1
De forma matricial:
be extracted from the two data sets
ðM x is
1Þ equal
ðM xtoIÞthe
ðI xsmaller
1Þ of the dimensions of x and y;
that is, M = min$I" J %.
The canonical vectors am and bm are the unique choices that result in the canonical
variates having the properties
T
w ¼ ½B# y0 : ð1
Corr!v1 " w1 # ≥ðMCorr!v
x 1Þ 2 "ðMw2x# JÞ≥ ·ðJ· ·x≥1ÞCorr!vM " wM # ≥ 0( (12.3a)
$
k = motoño 2022
rC "- semestre
GF601
fi
Se busca representar ambos grupos mediante componentes principales, es decir un
subconjunto con un mismo número de variables canónicas.
fi
v !Sv " !Svw " !I" !RC "
Var = = # (12.4a)
w nen un!Súnico
Las propiedades anteriores de par de vectores
wv " !Sv "
canónicos.
!RC " !I"
fi
where
standardized
where the!I$e
the
Since !I× ×$II"= diagonal
"diagonal
variables $f would
$ = matrix
be
matrix
1, this (Equation 9.31)
#Dx$$ (Equation
transformation
#D 9.31)
ensures contains
contains
unit the
the standard
standard
variances for deviations
deviations
the canonical
xa∗ = a #D $%
x hold for the canonical vectors bm(12.6)
m m
Y seofof the x variables,
x variables,
thevariates; and a similar
quethat is,and a similar equation equation m would
wouldm
hold for the canonical vectors b and
cumple ∗ m and
the!J!J××JJ""diagonal
the diagonal matrix
matrix #D #Dy$$ containing am = am #D
containing thex $%standard
the standard deviations deviations of
of the
the (12.6)
yy variables.
variables.
where the !Iof×whether
Regardless I " diagonal
a CCA matrix
is y #Dx $ (Equation
computed using 9.31) contains the standard variables,
deviations
T T −1/2 standardized −1/2 or unstandardized
T
Regardless
of !Iof× whether
var$v
x variables, %= a CCA
amsimilar is"acomputed
= em$ !S using standardized eorm unstandardized
=theem estandard
m = vectors variables,
1# deviations
bm (12.28)
thethe mand m equation
where the I " diagonal !S
matrix "would
(Equation !S hold
"!Sxxfor " the canonical and
resulting canonical xx
correlations #Dare x the xx
same. xx9.31) contains
the
of resulting
thethe!J x× J " canonical
diagonal
variables, and correlations
matrix
a similar#D $ are the would
containing
equation same. the holdstandard for the deviations of vectors
the easily.
y variables.
bm The
Correlations between the original and canonical
y variables cancanonical
be calculated and
the Correlations
!Jbecause
Regardless
× J " of !S between
whether
diagonal
−1/2
" matrix the
isCCA
a original
symmetric
#Dis computed
$ and
and canonical
containing the
using the variablesdeviations
eigenvectors
standardized
standard ecan or be mutually
are calculated
unstandardized
of the y easily. The An
orthogonal.
variables,
variables.
correlations between corresponding
xx y original and canonical variables, sometimes called
m
correlations
the obvious
resulting
Regardless
La relación of
entre between
analogous
canonical
whether
las corresponding
a equation
correlations
CCA
variables is canare
computed
canónicas original
be the written
using
y las and
same. for canonical
the variances
standardized
originales variables,
or
suele var$w sometimes
%.
unstandardized
caracterizarse called
variables,
homogeneous correlations, are given by m
homogeneous
the Correlations
resulting correlations,
between
canonical theare givenareand
original
correlations bythecanonical
same. variables can be calculated easily. The
mediante su correlación. T T −1
correlations between corresponding
Correlations between the original corr#v mand % x
original=
T canonical
$ a and
Tm #S canonical
variables
x%x $ #D x −1 can
$ variables,
be sometimes
calculated easily. called
(12.7a)
The
corr#vm!1×I" % x $ = !1×I" am #S!I×I" $#D $ (12.7a)
homogeneous
correlations correlations,
between are given
corresponding !1×I"
by
original and x%x !I×I"
canonical
!1×I" !I×I" !I×I"
x Desviación
variables, standard (diagonal)
sometimes called
Correlaciones
homogeneous
and
homogéneas:
correlations, are given byT T −1
and corr#v m % x $ = a m #S x%x $ #D x $ (12.7a)
!1×I"TT TT !I×I" !I×I"−1
corr#v % x
corr#wmm % yT $ = bmTm #S
$ = a!1×I"
#Sx%x $$#D
#D x
−1
$$−1 & (12.7a)
(12.7b)
y%y y
corr#w!1×I" % y $ = !1×I" bm #S y%y $#D
!I×I" !I×I"$ & (12.7b)
and m
!1×J"
!1×J"
!1×J" !J×J"
!1×J" !J×J" !J×J"
y
!J×J"
and
These equations specify vectors of correlations,
T T between −1 the m th
canonical variable vm
These equations corr#w % y $ = b #S $ #D $ & th (12.7b)
of the Ispecify
originalvectors of!1×J"
mcorrelations, m betweeny%y y the m canonical variable v
and each variables x
corr#wmx%,yand i , and
T betweenT
bm #Sy%ythe
$ =between
!1×J" the canonical
−1
$#Dcanonical variable w m and each
(12.7b)of
m
y$ &
!J×J" !J×J"
and each of the I original variables
the J original variables yk . Similarly, variable w and each of
!1×J"the vectors
i
!1×J" !J×J" of heterogeneous
!J×J"
correlations,
m between
the
theJ
These original
equations
canonical variables
specify
variables yand
k . Similarly,
vectors
the of
other the vectors
correlations,
original of heterogeneous
between
variables are the m th correlations, between
canonical variable vm
Correlaciones heterogéneas:
the
andcanonical
These each of the
equations variables
original
I specify and the other
variables
vectors original
i , and
of xcorrelations, betweenvariables the
between arethe mthvariable
canonical canonical and eachvmof
wmvariable
T T −1
the each
and J original
of thevariables
I originalyvariables corr#vxm,% and
k . Similarly,
ythe
T
$= aTm #Sx%y
vectors
between ofthe$heterogeneous
#Dcanonical
y $−1 correlations,
variable w and (12.8a)
between
each of
corr#vm!1×J" i% y $ = a #S $#D $ m (12.8a)
!1×I"
the Jcanonical
the original variables yand the other
k . Similarly, !1×J"
original
the vectors m variables
!1×I" !I×J"
x%y
!I×J" y
!J×J"
are
of heterogeneous
!J×J"
correlations, between
the
andcanonical variables and the other original T variables
T are−1
and corr#vm % y $ = am #Sx%y $#Dy $ (12.8a)
−1
corr#v
corr#wmm!1×J" %%yxTT$$ = = a!1×I"
T
bmTm #S
#S!I×J"
x%y
y%x
$$#D
#D yx$
!J×J"$ −1
& (12.8a)
(12.8b)
T T −1
corr#w!1×J" % x $ = !1×I"
!1×J" bm #S
!1×J" !I×J"
!J×I" $#D x$ &
!I×I"
!J×J" (12.8b)
and m
!1×J"
y%x
!1×J" !J×I" !I×I"
and The canonical vectors am and bm are T chosen T to maximize −1 correlations between the
The canonical
resulting canonicalvectorsvariates amcorr#w
v and
and
GF601 mm% -x
bw, are
but $= chosenbmotoño
(unlike
semestre
#SPCA)
to $#D
maximize
y%x2022 $ & or correlations
xmay may not be between (12.8b)
the
particularly
−1
and tr!ã ã T T
2 m m " 2 tr!b̃ b̃
m m"
Rm !x" = −1 Rm !y" = (12.12a)%
tr!#Sx&x $"#Y $ = #B̃$ #W$ %
! T T
tr!#Sy&y $" (12.10b)
!J×n" !J×J" !J×n"
11$820
7$466 6$743 correlations; reversing7$740
−1$921 7$923 3$840
−1$372 6$743 −12$95 45$47
!Ã#−1 = " and !B̃#−1 = b̄1 b̄T1 = !
the signs 7$740
on the second canonical
(12.16b) 59$91
vectors 61$36
would put pos
11$820 6$743 weights on the maxima 7$923
and 3$840
negative
$ " !7$740"
weights of 7$923# =
comparable
!
magnitudes
""
Ejemplo Wilks: T 7$923
7$740 61$36 61$36 mini
on
59$91 62$77 the
ibutions made by the canonical variates to the respective b̄1 b̄1 = ! covariance " matrices
!7$740" for = !
7$923# ""
nderlying data depend 7$923 61$36 62$77
ibutions made by the on the outer
canonical products
variates to of
thethe columns
respectiveT
b̃2 b̃2 The
of these matrices
−1$372
covariance
= canonical vectors
! " matrices
!−1$372"
(terms
for 1$882 5$279"
3$840# = to Ithaca tem-
! $
summations of Equations 12.11); that is,
TABLE
nderlying data depend on the outer products of the columns 12.1
T of3$840
−1$372 a
these matrices (termsm (corresponding 5$279
1$882 14$75
5$279
x = [Tx, of Tn] T (Ithaca) peratures) b̃ 2 b̃2bm
and =(corresponding!−1$372" to Canandaigua3$840# =
temperatures) for the $
summations ! Equations
" 12.11); that is, ! " 3$840 5$279 14$75
7$466 partition
55$74of the
88$25covariance matrix in Equation 12.13 with I = J = 2.
T
y ã=1 ã[Tx, Therefore the proportions of the Ithaca temperature variance described
1 =! Tn] " !7$466" 11$820# = ! shown are the "eigenvalues" (12.17a)
T (Canadaigua) Also ' (cf. Example 12.3) and the canonical
11$820
7$466 variates
Therefore 88$25
55$74 139$71
88$25
(Equation
the which 12.12a)
proportions are
of
m
the roots.
Ithaca temperature variance described
ã1 ã1T =! " !7$466" 11$820# =correlations,
! are" their
" square (12.17a)
11$820
−1$921 variates 88$25 139$71
(Equation
3$690 −12$95 12.12a) areb
T
ã2 ã2 = ! " !−1$921" 6$743# =! a " " (12.17b)a2 b2
6$743
−1$921 −12$95
3$690 −12$95
1
45$47 2
1
55$74 + 139$71
T
ã2 ã2 = ! !−1$921" 6$743# = ! (Ithaca) " R 1 &x' =
(Canandaigua) (Ithaca)
(12.17b)
55$74 + 139$71
= 0$798
(Canandaigua)
6$743 " −12$95 45$47 " 2 59$52 185$47
T 7$740" T max 59$91 61$36 .0923 1 &x' =
R.0946 −&1618 = 0$798
−&1952
b̄1 b̄1 = ! !7$740" 7$923# = ! " " (12.17c)
59$52 + 185$47
7$923
7$740 61$36
and=T min59$91 61$3662$77 .0263 .0338 .1022 .1907
b̄1 b̄T1 =! !7$740" 7$923# " (12.17c)
7$923 " and
!61$36 62$77 "
0.938 0.593
T −1$372 ' m 1$882 5$279
b̃2 b̃2 = ! " !−1$372" 3$840# =! √ " $ 3$690 + 45$47
(12.17d)
T 3$840
−1$372 r =5$279 14$75
Cm 1$882m 5$279
' 0.969R 2
&x' = 0.770 = 0$202"
b̃2 b̃2 = !−1$372" 3$840# = $ 2 (12.17d)
3$690+
59$52 +185$47
45$47
3$840 5$279 14$75 2
R2 &x' = = 0$202"
fore the proportions of the Ithaca temperature variance described by its two canonical 59$52 + 185$47
es (Equation
fore 12.12a)ofare
the proportions and thevariance
the Ithaca temperature corresponding
describedvarianceby its two fractions
canonicalfor Canandaigua are
es (Equation 12.12a) are and the corresponding variance fractions for Canandaigua are
55$74 + 139$71 2 59$91 + 62$77
R12 &x' = 55$74 + 139$71 = 0$798 R1 &y' =(12.18a) = 0$880"
59$52 + 185$47 2 59$91 +
61$85 77$5862$77
2
R1 &x' = = 0$798 R 1 &y' =
(12.18a) = 0$880"
59$52 + 185$47 61$85 + 77$58
and
and
3$690 + 45$47 2 1$882 + 14$75
R22 &x' =
3$690 + 45$47 = 0$202" R2 &y' =(12.18b) = 0$120$
59$52 + 185$47 = 0$202"
R22 &x' = 2 1$882
(12.18b)
+ 14$75
61$85 77$58
♦
59$52 + 185$47 R 2 &y' = = 0$120$
61$85 + 77$58
he corresponding variance fractions for Canandaigua are
♦GF601 - semestre
he corresponding variance fractions for Canandaigua are otoño 2022
scores
0
r
correspondientes para reconstruir las PCs del grupo mayor entre 1960 y 2014. Finalmente, u
0.4
las PCs extendidas y EOFs del grupo mayor para reconstruir la precipitación.
0.2
Se utilizaron las correlaciones canónicas para crear modelos simples de la variables canónica
segundo 0 grupo. Estos modelos son simplemente v^k = rk uk. Las series extendidas (1960-2014)
1 2 3 4 5 6 7
son, lógicamente, muy buenas estimaciones de las 8series
9 10 11
originales 12
en el 13
caso de los prim
componentes
modos, y van empeorando en el caso de modos mayores. Como ejemplo, la siguiente figura mu
la serie original (negro) y estimada (extendida, rojo) del modo #5. Como se vio (figura 1), este m
la media. Este sesgo resulta de que la función ‘cannoncorr’ entrega variables canónicas
tiene asociado una correlación canónica de ~ 0.94, por lo que la reconstrucción hacia atrás
Extensión modo 1
das en 0 aunque los PCs utilizadas para calcularlas no lo sean. tiempo es aún bastante confiable.
Modelo modo 5
c. Proyecte las PCs del grupo menor de estaciones sobre los vectores canónicos corr
Primera variable canonica − Modo 1 para obtener variables canónicas extendidas (1960-2014).
Segunda variable canonica − modo 5 Grafique la serie de
scores
0 definición, correlación 1.00 en el periodo común. Sin embargo, se tuvo que corregir un pe
−0.5 −0.5
−1 −1
−1.5 −1.5
−2 −2
−2.5 −2.5
1960 1965 1970 1975 1980 1985 1990 1995 2000 2005 2010 2015 1960 1965 1970 1975 1980 1985 1990 1995 2000 2005 2010 2015
La figura siguiente muestra los mapas de anomalías de P para el periodo 1967-69. Los paneles de la
izquierda
Ejemplo:y centro indican las anomalías
Reconstrucción resultantes
de la precip. de las reconstrucciones con PCA conjunta (Tarea
en Chile
4) y PCA/CCA, respectivamente. La reconstrucción PCA/CCA es, en este caso, más verosímil
comparado con las anomalías obtenidas del grupo de estaciones de referencia (panel derecho).
Se quiere estimar P en un número grande de localidades en Chile para el periodo anterior a
Como se describió en la tarea anterior, la reconstrucción utilizando PCA conjunta muestra una
1979. Como guía para la reconstrucción, se usarán series más largas (1960-2014) de sólo 13
dispersión importante en las estaciones de Chile central que no se aprecia en los datos de referencia.
estaciones.
Como en EOF/PCA, esta técnica se puede utilizar para relacionar dos variables
espacialmente distribuidas o disponibles en una grilla regular.
°
10
Ejemplo Wilks: r = .79
ER ! 12 Canonical Correlation Analysis (CCA)
r = .79
(b)
(a) (b)
C2,2(Z500)
C2,2(SST) 18% C2,2(Z500)
°N
23%
50 23%
87
°
30
–74
I I
NP –95 NP
–95 1 0° GM 60 °N GM 60 °N
40 °
40 ° 20 °
20 °
75
r = .79 75
(b)
C2,2(Z500) –67
–67
23%
FIGURE 12.1 Homogeneous correlation maps for a pair of canonical variables pertainin
age winter sea-surface temperatures (SSTs) in the northern Pacific Ocean, and (b) hemis
FIGURE 12.1 Homogeneous correlation 500 mbmaps for The
heights. a pair of canonical
pattern variablesinpertaining
of SST correlation to (a)
the left-hand aver-
panel (and its negativ
ated with
age winter sea-surface temperatures (SSTs) in the
thePNA patternPacific
northern of 500 mb heightand
Ocean, correlations shown in the
(b) hemispheric right-hand panel.
winter
correlation for this pair of canonical variables is 0.79. From Wallace et al. (1992).
500 mb heights.
–95 The pattern
I
NP of SST correlation in the left-hand panel (and its negative) are associ-
GM 60 °N
ated with the PNA pattern of 500 mb40height
° correlations shown in the right-hand panel. The canonical
20 °
correlation for this75pair of canonical variables is 0.79. From Wallace et al. (1992).
CCA parameters may be unstable (i.e., exhibit large variations from batc
for small samples (e.g., Bretherton et al. 1992; Cherry 1996; Friedrerichs
GF601 - semestre otoño 2022
Ejemplo Wilks (PCA/CCA para pronóstico521estacional; Barnston 1994)
CCA APPLIED TO FIELDS
(a)
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180
30 30 30 15
522
0 30 15
00 30
40 S
0
0
0 0
15
15 40 S CHAPTER ! 12 Canonical Correlation Analysis (CCA)
45 30 15 30 45 45
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180
(b)
–45 –60
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180 –49 –60
0 0 30 15 –45
30 –45 –30
60 N –15 60 N
–15
–15
–15
JJA 0
–30
15 15 15
50 N
0 0
50 N –30
–15 0 15
15 15
40 N 15 40 N
15 15 30 30
30 –30 0 0
30 N 15
30 N
30 15 0 –30 30 30 30
0
15
20 N –15
0
0 30
30
15 30
20 N
30
0 –45 –45
–15 –15 15 0
10 N –75 15 45 30 –1510 N
30 0
–60 –15 0 30 –45
EQ 0 EQ
–60 –15 –45
–30 –150 –15 45 30 0
10 S 0 10 S
0 –45 –15 0 0 45 30
30 0 30
20 S
60 –60 –15 60 20 S 45
45
60 –45 –45 –15
–15 –15
–30 –15 0 45 30 45
30 56
30 S –30 –15 0 30 S
–15 0 0 0 45
0 –15 –15
–15 0 0
30
56
40 S 0 40 S
0
30 15
–15
0 15 15 15
30
15 15 15
JFM
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180
(c)
FIGURE 12.3 Spatial display of the first canonical vector for predicted U.S. JFM surfac
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180
0
0
30 30 30
tures. A portion of the corresponding canonical vector for the predictors is shown in Figure
–15
60 N
–15
–15
15
30
SON
60 N
Barnston (1994).
–30 –45 –15
50 N 50 N
0 15
15 –30 15 15
–45 15 30
40 N 40 N
15 15 15 45
45 –45
30 N
30
0
–45
–15
–30
0
–30
0 0 15
0
15
–45 60 –45
–45
EQ three maps expresses the six elements of a1 in terms of the original 235 spatial
–30 –15
through the corresponding elements of the eigenvector matrix [E] for the predic
10 S 15 –30
–30 0 10 S
0 –30 –30 0 –30 30
–45 15 45
45 –60 –60 –15 20 S
20 S –30 30
45
30 S
60 –45
0
–15
–45 –45
–15 –15
–15
–30
–15
–30
–30 0
–15
–15
–30
15 30
15
45
30
30 S The most prominent feature in Figure 12.2 is the progressive evolution of inc
0 15
negative values in the eastern tropical Pacific, which clearly represents an in
–15 15
40 S 15 15 40 S
0 15 0 30
45 15 30 30 15 30 0 15 15 0 0 15
160 w 140 w 120 w 100 w 80 w 60 w 40 w 20 w 00 20 E 40 E 60 E 80 E 100 E 120 E 140 E 160 E 180
El Niño (warm) event when v1 < 0, and development of a La Niña (cold) ev
FIGURE 12.2 Spatial displays of portions of the first canonical vector for predictor sea-surface v1 > 0, in the spring, summer, and fall before the JFM season to be forecast.
temperatures, in the three seasons preceding the JFM for which U.S. surface temperatures are forecast. Figure 12.3 shows the first canonical predictand vector for the JFM forecast,
GF601 - semestre
The corresponding canonical vector for this predictand is shown in Figure 12.3. From Barnston (1994).
otoño 2022
menos de variabilidad
s siguientes sobresalientes,
(Figura 2). Para explicarcomo los ocurren
el grueso en escalas
de la varianza tiempo mayor
conjunta de Z300([Link]
ENSO o considerar
deben 1 88
ello,Ejemplo:
odos el espectro de
(95.1%). Relación entre
valores propios la circulación
no discrimina (Z300
claramente estos ) y la
modos, Tmax
ni los endeChile central (solo
distingue mes de
tes (Figura 2). Para explicar el grueso de la varianza conjunta de Z300 se deben considerar 88 0
.1%).enero) 0 10 20 30 40 50
componentes
gura 1:espaciales
atrones Patrones espaciales (EOF)
(EOF) del 1er y 2dodel 1er de
modo y 2do modo dedevariabilidad
variabilidad Z300’. de Z300’.
- Otoño 2018
4
R2k (%)
0
0 10 20 30 40 50 60 70 80
componentes
Figura 3: Coeficiente de correlación entre los modos 1 y 2 de Z30
Varianza explicada, y error típico (North et al.) de los primeros 88GF601
modos de- variabilidad
semestredeotoño
Z300. 2022
de Chile. Panel de la izquierda muestra la varianza de Tmax explic
Ejemplo: Relación entre la circulación (Z300) y la Tmax en Chile central (solo mes de
GF601 - Otoño 2018
enero)
Análisis PCA-conjunto:
Patrones espaciales
(EOF) del 1er y 2do modo
de variabilidad para Z300’
y Tmax.
Figura 4: Patrones espaciales (EOF) del 1er y 2do modo de variabilidad para Z300’ y Tmax (análisis PCA-conjunto).
Este método PCA-conjunto también se podría utilizar para construir un modelo predictivo de Tmax. Es
este caso habría que generar un nuevo grupo de series ‘truncadas” de PCs, proyectando solamente los
datos de Z300 sobre los EOF-conjuntos (considerando los elementos de las EOF correspondientes a
Z300). Las PCs truncadas se utilizarían como predictores en un modelo de regresión de Tmax.
Análisis PCA/CCA:
Correlaciones homogéneas
entre las variables canónicas y
variables originales
2. PCA/CCA
correspondientes (Z300 y Tmax
a la izquierda y derecha, a. En esta parte, analizaremos la covarianza de Z300 y Tmax utilizando PCA/CCA. Para ello, calcule las
respectivamente) EOF/PCs de Z300’ (1a) y de Tmax’ en el centro-sur de Chile (30-40ºS), separadamente. Para lo que
sigue,
Figura utilizaremos homogéneas
6: Correlaciones un mismo número
entre lasde PCs encanónicas
variables ambos grupos de variables.
y variables Éstos deben explicar
originales correspondientes (Z300 al
y
GF601 - semestre otoño 2022
menos el 95% de varianza conjunta de cada grupo de variables. ¿Cuantos modos se retienen?
Técnicas de agrupamiento (Clustering)
> Contrasta con el análisis discriminante, en el cual se tiene una conocimiento priori
de los grupos.
Esta métrica medirá la distancia entre elementos de un grupo, la cual debe ser
pequeña respecto de la distancia entre grupos.
fi
fi
En general, se de nirá una matriz de distancia ∆ con elementos di,j = d(xi, xj)
fi
Agrupamiento jerárquico
Método de aglomeración
En cada paso, los dos grupos más cercanos entre sí se fusionan en un nuevo grupo.
linkage
Simple Completo
it begins with n single-member groups, and merges two groups at each step, until all th
dataMétodo
are in de
a single
mínima group n − 1 steps. However, the criterion for choosing which pa
after(Ward)
varianza
of groups to merge at each step is that, among all possible ways of merging two group
theElpair to bedemerged
criterio is chosen
agrupación thatenminimizes
utilizado cada pasothe sum→ofG),squared
(G+1 distances la
busca minimizar between th
points and al
varianza theinterior
centroids
de losofgrupos.
their respective groups, summed over the resulting group
That is, among all possible ways of merging two of G + 1 groups to make G groups, tha
merger
Para is made
cada thatfusión
nueva minimizes
posible entre 2 grupos, se calcula:
G ng G ng K
!! !!!
W= "xi − x̄g "2 = !xi"k − x̄g"k #2 $ (14.7
g=1 i=1 g=1 i=1 k=1
In order to implement Ward’s method to choose the best pair from G + 1 group
to merge, Equation
Luego, entre 14.7 must
las G(G+1)/2 be calculated
posibles fusiones,for all of the
se escoge G!G
la con + 1#/2
menor W. possible pairs o
existing groups. For each trial pair, the centroid, or group mean, for the trial merge
group is que
Notar recomputed using the en
W irá aumentando datacada
for iteración…
both of the previously separate groups, before th
squared distances are calculated. In effect, Ward’s method minimizes the sum, over the K
desde G = N (etapa inicial) => W = O
dimensions of x, of within-groups variances. At the first (n-group) stage this variance
zero, and
hastaat the
G =last
1 ( (1-group)
nal) stage this
=> W variance is tr%Sx &, so that W = n tr%Sx &. For dat
= N Tr(S)
vectors whose elements have incommensurate units, operating on nondimensionalize
values (dividing by standard deviations) will prevent artificial domination of the procedur
by one or a few of the K variables.
Criterios de corte
En caso contrario, la distancia entre los grupos fusionados en cada etapa puede
servir para de nir objetivamente un criterio de corte.
Con grupos bien de nidos (i.e., distancias cortas intra-grupo y grandes entre-grupos),
la distancia aumenta en cada etapa de forma paulatina mientras se aglomeren
elementos de grupos bien de nidos, y “explota” en la etapa en que se mezclan los
grupos.
Este salto no es tan claro con grupos poco de nidos, pero se puede aplicar un criterio
de corte usando de la tasa de aumento en la distancia de fusión.
fi
fi
fi
fi
fi
TABLE 13.1 Average July temperature !! F" and precipitation (inches) for locations in three regions
of the United States. Averages are for the period 1951–1980, from Quayle and Presnell (1991).
Group 1: Southeast U.S. (O) Group 2: Central U.S. (X) Group 3: Northeast U.S. !+"
membership in Group 1 vs. Group 2. This problem might arise if the stations in
Table 13.1 represented the core portions of their respective climatic regions, and on the
basis of these data we wanted GF601
to classify stations
- semestre not 2022
otoño listed in this table as belonging to
Ejemplo con K = 2 (Wilks)
with seven
Vínculo: groups. These seven groupseven
completo are though
indicated in Figure
the added
Vínculo: 14.5b,
points may
simple which
be quite canother
far from be com
poin
pared with the complete-linkage result in Figure 14.4. The clusters denoted G2 and G6 i
(a) (b)
Figure 14.4 occur also in Figure 14.5b. However, one long and thin group has develope
in Figure 14.5b, composed of stations from G3 ! G4 , and G5 . This result illustrates th 1.00
or groups are accumulated that are close to a point at one edge or another of a group
Standardized Precipitation
0.75
even though the added points may be quite far from other points in the same group. ♦ 1
Distance
0.50
(a) (b) 0
0.25
+++ +
×
–1 + + ++
+
1.00 0.00
1 5 10 15 20 25
2 –2 –1
Stage Number Standardized
FIGURE 14.5 Clustering of the data in Table 13.1, using single linka
Standardized Precipitation
function of stage, showing a large jump after 22 stages. (b) The seven c
0.75 illustrating the chaining phenomenon.
1
Distance
0.50
×
0
×
0.25
+++ + ×× ×
×
× ×
–1 + + ++ ×
+
0.00
1 5 10 15 20 25 –2 –1 0 1 2
Stage Number Standardized Temperature
2500 2500
Precipitacion anual (mm)
1500 1500
1000 1000
500 500
0 0
−200 0 200 400 −200 0 200 400
Diferencia pp Julio−Enero (mm) Diferencia pp Julio−Enero (mm)
2500 2500
Precipitacion anual (mm)
1500 1500
1000 1000
500 500
0 0
−200 0 200 400 −200 0 200 400
Diferencia pp Julio−Enero (mm) Diferencia pp Julio−Enero (mm)
En este caso, los elementos de los grupos corresponden a las observaciones (eje
temporal) de una cierta variable, la dimensión del problema (K) a los valores de la
variable en el espacio.
fi
años
G1 G2 G3 G4
SLP’ mensual:
Vínculo completo, correlación
SLP’mensual:
SLP’ mensual Distancia: Correlación Vínculo: Completo
G1correlación
Vínculo completo, G2 G3 G4
G1 G2 G3 G4
inicio convergencia
GF601 - semestre otoño 2022
Algoritmo
(3) cambia xi
de grupo
Utilizando un métrica de distancia euclidiana, los resultados son idénticos ya sea se use un vínculo de
tipo centroide o completo. En ambos casos, uno de los grupos resultante confunde los datos de las
Ejemplo con K = 2 (Régimen de precipitación observado en tres estaciones de Chile)
estaciones #2 y #3. Los otros 2 grupos se construyen con observaciones de la estación #1. Con la
distancia de mahalanobis, se obtienen grupos que tienden a deformarse en la dirección de covarianza
entre ambas variables y el resultado es peor que con distancia euclidiana. Como ejemplo, cabe notar
que en el agrupamiento con vínculo de tipo centroide, el grupo 2 resultante (rojo) comparte c. Calcule ahora 3 grupos usando el método k-means. Compare con el punto anterio
observaciones de las tres estaciones.
0
−200 0 200 400
0
−200 0 200 400
K-means
Diferencia pp Julio−Enero (mm) Diferencia pp Julio−Enero (mm)
Linkage: complete. Distance: euclidean Linkage: complete. Distance: mahalanobis K−means clustering
2500 2500
Precipitacion anual (mm)
1500 1500
1000 1000
2000
500 500
1000
500
0
−300 −200 −100 0 100 200 300 400
Diferencia pp Julio−Enero (mm)
Figura 8: Mapas de los 3 principales modos para el anali. de EOFS de SLP’ en la primera
fila , mismo resultado pero con la matriz de correlacion en la segunda fila
PCA
4
GF601 - semestre otoño 2022
Análisis de agrupamiento de días con precipitación en Santiago en base al contenido
de agua en la atmósfera
Pregunta 6:
dispersion,
origin horizontal
than point B in distances
Figure 9.2 are
when lessmeasured
unusual than according vertical to ones
the relative
Mahalano
n. scatter.
For a Although
fixed
origin than point B in Figure 9.2 when measured according to the Mahalan
point
MahalanobisAfixed
is closer
distance to theDcenter
2
, EquationofD2the distribution
9.7 defines according
an ellipse to
(Distancia de Mahalanobis) For a Mahalanobis distance , Equation 9.7 defines an ellips
uncorrelated, a distance
(a) measure
is more(b)that
statistical distance on unusual
distance, it reflects
the distance
statistical plane, than on point
and thethat Bellipse
plane, in
andthe thatiscontext
also is
ellipse a established
circle
also if by
if s1"1
a circle s= the
1"1 =
s2"2s2"2.p
and so is9.7statistically
Equationfurther
9.7 to from
threeby the origin.
dimensions by addingterm a thirdforterm for x3 , the set ofpoi
po
er can be defined
x2 simply
Equation as
Because
to
the
three
points
dimensions
distance in
D 2 x
Figure
constitute
caso
2 9.2a
an
adding
2D are
ellipsoid
sin
a third
uncorrelated,
that will
correlación be a
x ,
distance
spherical
the set of
3 if all three varian
measure t
distance D2 constitute an
blimp-like ellipsoid that will be spherical if all three varianc
unusualness in the contextif oftwothe variances are
B nearly
data scatter canequal but smaller
be defined than the
simply as third, an
%2 $x2 − x̄2 %2 A blimp-like if two twovariances
variancesare
In general
nearlyequal
are nearly
theAvariables
equal
within
but smaller
and larger than the third.
a2 multivariate
than the third, an
2 vector x will not be unc
data
+ " twoB variances are nearly equal
these correlations (9.7)
and
2
Dmust
larger
$x
= also be1 − x̄ than
1 %
accounted
the
$x 2 − x̄
third. 2
+ for when defining
%
" distances in terms
s2"2 In general the variables within a
or probability density. Figure
θ
multivariate
s1"1 data
9.2b illustratess2"2 vector x will
the situation in two notdimensions
be unco
R A N Dthese
OM V correlations
E C T O R Smust A Nalso
points be
Aaccounted
D MFigure
from TR 9.2aI Chave for rotated
E Sbeen when arounddefining distances
the origin throughinanterms
angle %o
x x1
or probability1 density. Figure
in the two 9.2bbeing
variables illustrates
relatively the situation
strongly positively incorrelated.
two dimensions,Again poin
points from Figurethe (a) origin in a statistical sense, although
9.2a have been rotated around (b)in
theorder to calculate the actual Mahalan
origin through an angle %,
in terms of the variables x1 and x2 it would be necessary to use an equation of
(b) 9.4.4 Mahalanobis Distance, Revisited
in the two variables being relatively
x2
the origin in a statistical sense, Dalthough2
A = a
strongly positively correlated.
&x
1"1 1 in
− order
x̄1 $ 2
+ to
2a calculate
&x
1"2 1 −
x2
x̄1 $&xthe
2 − actual
x̄2
Again point
B $ + a Mahalano
&x
2"2 2 − x̄2 '
$2
Este vector se de ne de manera tal que la distancia entre sus medias respecto de la
matriz de covarianza se maximice.
aT [x(n) − media(Xg)]
D2 = aT [media(X1) − media(X2)]
δ1(n) = aT x(n)
> La proyecciones sobre a son análogas a las nuevas variables en PCA o CCA.
aT x(n) ≥ m
fi
fi
G1: T, P en southeast US
G2: T, P en central US
532 CHAPTER ! 13 Discrimination and Classification
TABLE 13.1 Average July temperature !! F" and precipitation (inches) for locations in three regions
of the United States. Averages are for the period 1951–1980, from Quayle and Presnell (1991).
Group 1: Southeast U.S. (O) Group 2: Central U.S. (X) Group 3: Northeast U.S. !+"
membership in Group 1 vs. Group 2. This problem might arise if the stations in
Table 13.1 represented the core portions
GF601 of theirotoño
- semestre respective
2022 climatic regions, and on the
N1 =10; N2 =9
SEPARATING TWO POPULATIONS
→ Spool = (9 S1 + 8 S2)/17
= 1.76 0.37 δ1
0.37 0.85
7
×
→ a = 0.62 4 ×
2.20 × × × ×
×
× ×
3 ×
76 78 80 82
Temperature, °F
F IGURE 13.1 Illustration of the geometry of linear discriminant analysis applied to the south
(circles)GF601
and central (Xs) U.S.
- semestre data2022
otoño in Table 13.1. The (vector) means of the two groups of d
símbolos ‘+’ y ‘o’, respectivamente. Los colores, por su parte, señalan la clasificac
Otro ejemplo 2D: análisis. En efecto, las proyecciones sobre el vector discriminante (a) de solo dos
grupo de #1 resultaron menores al valor crítico m (ver línea segmentada) y fuero
2 variables en dos erróneamente
estaciones clasificados en el grupo #2.
de observación Analisis Discriminante
2500
2000
1000
a.x = m
500
0
−100 0 100 200 300 400
Diferencia pp Julio−Enero (mm)
De forma similar a lo evaluado con PCA y CCA, usemos un campo cuyos puntos de
grilla de nan las K variables, y el paso temporal las N observaciones.
Tomemos las anomalías anuales de TSM entre 1960 y 2012 en un dominio tropical, y
pre-de namos 2 grupos:
G1: El Niño
G2: No El Niño.
Usemos para ellos los años de El Niño desde 1960: {1963 1965 1969 1972 1977 1982
1987 1991 1992 1993 1994 1997 2002 2004 2006 2009}
Problema: se debe elegir un dominio espacial no muy grande, para poder calcular la
inversa de Spool.
> Como en CCA, este problema se puede abordar usando componentes principales.
Para determinar los vectores discriminantes A = (a1, a2, aj, ...), también se debe tomar
en cuenta una medida de covariabilidad entre grupos SB.
Análogamente a Sg …
Spool−1 SB E = E L
La regla de clasi cación para una observación x(n) queda determinada por su
distancia al grupo g en el base de nida por A:
fi
≠