Part 4: Time Series II
|
|
|
EOF Analysis |
|
Principal Component |
|
Rotated EOF |
|
Complex EOF |
Empirical Orthogonal
Function Analysis
|
|
|
Empirical Orthogonal Function (EOF)
analysis attempts to find a relatively small number of independent variables
(predictors; factors) which convey as much of the original information as
possible without redundancy. |
|
EOF analysis can be used to explore the
structure of the variability within a data set in a objective way, and to
analyze relationships within a set of variables. |
|
EOF analysis is also called principal
component analysis or factor analysis. |
What Does EOF Analysis
do?
|
|
|
In brief, EOF analysis uses a set of
orthogonal functions (EOFs) to represent a time series in the following way: |
|
|
|
|
|
|
|
Z(x,y,t) is the original time series as
a function of time (t) and space (x, y). |
|
EOF(x, y) show the spatial structures (x, y) of the major factors that
can account for the temporal variations of Z. |
|
PC(t) is the principal component that tells you how the amplitude of
each EOF varies with time. |
Slide 4
An Example
Another View of the
Rotation
Rotation of Coordinates
|
|
|
Suppose the Pacific SSTs are described
by values at grid points: x1, x2, x3, ...xN.
We know that the xi’s are probably correlated with each other. |
|
Now, we want to determine a new set of
independent predictors zi to describe the state of Pacific SST,
which are linear combinations of xi: |
|
|
|
|
|
|
|
Mathematically, we are rotating the old
set of variable (x) to a new set of variable (z) using a projection matrix
(e): |
Determine the Projection
Coefficients
|
|
|
The EOF analysis asks that the
projection coefficients are determined in such a way that: |
|
(1) z1 explains the maximum possible amount of the variance
of the x’s; |
|
(2) z2 explains the maximum possible amount of the
remaining variance |
|
of the x’s; |
|
(3) so forth for the remaining z’s. |
|
|
|
With these requirements, it can be
shown mathematically that the projection coefficient functions (eij)
are the eigenvectors of the covariance matrix of x’s. |
|
The fraction of the total variance
explained by a particular eigenvector is equal to the ratio of that
eigenvalue to the sum of all eigenvalues. |
Slide 9
Covariance Matrix
|
|
|
The EOF analysis has to start from
calculating the covariance matrix. |
|
For our case, the state of the Pacific
SST is described by values at model grid points Xi. |
|
Let’s assume the observational network
in the Pacific has 10 grids in latitudinal direction and 20 grids in
longitudinal direction, then there are 10x20=200 grid points to describe the
state of pacific SST. So we have 200 state variables: |
|
Xm(t), m =1, 2, 3, 4, …, 200 |
|
In our case, there are monthly
observations of SSTs over these 200 grid points from 1900 to 1998. So we have
N (12*99=1188) observations at each Xm: |
|
Xmn = Xm(tn),
m=1, 2, 3, 4, …., 200 |
|
n=1, 2, 3, 4, …..,
1188 |
|
|
Covariance Matrix – cont.
|
|
|
The covariance between two state
variables Xi and Xj is: |
|
|
|
|
|
|
|
The covariance matrix is as following: |
Eigenvectors of a
Symmetric Matrix
|
|
|
Any symmetric matrix R can be
decomposed in the following way through a diagonalization, or eigenanalysis: |
|
|
|
|
|
Where E is the matrix with the
eigenvectors ei as its columns, and L is the matrix with the
eigenvalues li, along its diagonal and zeros elsewhere. |
|
The set of eigenvectors, ei,
and associated eigenvalues, li, represent a coordinate
transformation into a coordinate space where the matrix R becomes diagonal. |
|
|
Orthogonal Constrains
|
|
|
There are orthogonal constrains been
build in in the EOF analysis: |
|
The principal components (PCs) are orthogonal in time. |
|
There are no simultaneous temporal correlation between any two
principal components. |
|
(2) The EOFs are orthogonal in space. |
|
There are no spatial correlation between any two EOFs. |
|
The second orthogonal constrain is
removed in the rotated EOF analysis. |
Mathematic Background
|
|
|
I don’t want to go through the
mathematical details of EOF analysis. Only some basic concepts are described
in the following few slids. |
|
Through mathematic derivations, we can
show that the empirical orthogonal functions (EOFs) of a time series Z(x, y,
t) are the eigenvectors of the covarinace matrix of the time series. |
|
The eigenvalues of the covariance
matrix tells you the fraction of variance explained by each individual EOF. |
Some Basic Matrix
Operations
|
|
|
A two-dimensional data matrix X: |
|
|
|
|
|
|
|
The transpose of this matrix is XT: |
|
|
|
|
|
The inner product of these two
matrices: |
Correlation Matrix
|
|
|
Sometime, we use the correlation
matrix, in stead of the covariance matrix, for EOF analysis. |
|
For the same time series, the EOFs
obtained from the covariance matrix will be different from the EOFs obtained
from the correlation matrix. |
|
The decision to choose the covariance
matrix or the correlation matrix depends on how we wish the variance at each
grid points (Xi) are weighted. |
|
In the case of the covariance matrix
formulation, the elements of the state vector with larger variances will be
weighted more heavily. |
|
With the correlation matrix, all
elements receive the same weight and only the structure and not the amplitude
will influence the principal components. |
|
|
Correlation Matrix –
cont.
|
|
|
The correlation matrix should be used
for the following two cases: |
|
The state vector is a combination of
things with different units. |
|
(2) The variance of the state vector
varies from point to point so much that this distorts the patterns in the
data. |
|
|
How to Get Principal
Components?
|
|
|
If we want to get the principal
component, we project a single eigenvector onto the data and get an amplitude
of this eigenvector at each time, eTX: |
|
|
|
|
|
|
|
|
|
For example, the amplitude of EOF-1 at
the first measurement time is calculated as the following: |
Presentations of EOF –
Variance Map
|
|
|
There are several ways to present EOFs.
The simplest way is to plot the values of EOF itself. This presentation can
not tell you how much the real amplitude this EOF represents. |
|
One way to represent EOF’s amplitude is
to take the time series of principal components for an EOF, normalize this
time series to unit variance, and then regress it against the original data
set. |
|
This map has the shape of the EOF, but
the amplitude actually corresponds to
the amplitude in the real data with which this structure is associated. |
|
If we have other variables, we can
regress them all on the PC of one EOF and show the structure of several
variables with the correct amplitude relationship, for example, SST and
surface vector wind fields can both be regressed on PCs of SST. |
Presentations of EOF –
Correlation Map
|
|
|
Another way to present EOF is to
correlate the principal component of an EOF with the original time series at
each data point. |
|
This way, present the EOF structure in
a correlation map. |
|
In this way, the correlation map tells you what are the
co-varying part of the variable (for example, SST) in the spatial domain. |
|
In this presentation, the EOF has no
unit and is non-dimensional. |
Using SVD to Get
EOF&PC
|
|
|
We can use Singular Value Decomposition
(SVD) to get EOFs, eigenvalues, and
PC’s directly from the data matrix, without the need to calculate the
covariance matrix from the data first. |
|
If the data set is relatively small,
this may be easier than computing the covariance matrices and doing the
eigenanalysis of them. |
|
If the sample size is large, it may be
computationally more efficient to use the eigenvalue method. |
What is SVD?
|
|
|
Any m by n matrix A can be factored
into |
|
|
|
|
|
|
|
The columns of U (m by m) are the EOFs |
|
The columns of V (n by n) are the PCs. |
|
The diagonal values of S are the
eigenvalues represent the amplitudes of the EOFs, but not the variance explained by the EOF. |
|
The square of the eigenvalue from the
SVD is equal to the eigenvalue from the eigen analysis of the covariance
matrix. |
An Example – with SVD
method
An Example – With
Eigenanalysis
How Many EOFs Should We
Retain?
|
|
|
There are no definite ways to decide
this. Basically, we look at the eigenvalue spectrum and decide: |
|
The 95% significance errors in the
estimation of the eigenvalues is: |
|
|
|
|
|
If the eigenvalues of adjacent EOF’s are closer together than this
standard error, then it is unlikely that their particular structures are
significant. |
|
(2) Or we can just look at the slope of the eigenvalue spectrum. |
|
We would look for a place in the eigenvalue spectrum where it levels
off so that successive eigenvalues are indistinguishable. We would not
consider any eigenvectors beyond this point as being special. |
|
|
An Example
Rotated EOF
|
|
|
The orthogonal constrain on EOFs
sometime cause the spatial structures of EOFS to have significant amplitudes
all over the spatial domain. |
|
č We can not get localized EOF structures. |
|
č Therefore, we want to relax the spatial orthogonal constrain |
|
on EOFs (but still keep the temporal orthogonal
constrain). |
|
č We apply the Rotated EOF analysis. |
|
|
|
To perform the rotated EOF analysis, we
still have to do the regular EOF first. |
|
č We then only keep a few EOF modes for the rotation. |
|
č We “rotated” these selected few EOFs to form new EOFs (factors). |
|
based on some criteria. |
|
č These criteria determine how “simple” the new factors are. |
Criteria for the Rotation
|
|
|
Basically, the criteria of rotating
EOFs is to measure the “simplicity” of the EOF structure. |
|
Basically, simplicity of structure is
supposed to occur when most of the elements of the eigenvector are either of
order one (absolute value) or zero, but not in between. |
|
There are two popular rotation
criteria: |
|
Quartimax Rotation |
|
Varimax Rotation |
|
|
Quartimax and Varimax
Rotation
|
|
|
Ouartimax Rotation |
|
It seeks to rotate the original EOF matrix into a new EOF matrix for
which the variance of squared elements of the eigenvectors is a maximum. |
|
|
|
|
|
|
|
Varimax Rotation (more popular than the
Quartimax rotation) |
|
It seeks to simplify the individual EOF factors. |
|
|
|
|
|
|
|
The criterion of simplicity of the complete factor matrix is defined
as the maximization of the sum of the simplicities of the individual factors. |
Reference For the
Following Examples
|
|
|
The following few examples are from a
recent paper published on Journal of Climate: |
Example 1:
North Atlantic SST Variability
Example 2: Indian Ocean
SST Variability
Example 3:
SLP Variability
(Arctic Oscillation)
Example
4:
Low-Dimensional Variability
(Variance Based)
Slide 35