Given
sets of variates denoted
, ...,
, a quantity called the Covariance Matrix is defined
by
where
and
are the Means of
and
, respectively.
An individual element
of the Covariance Matrix is called the covariance of the two variates
and
, and provides a measure of how strongly correlated these variables are. In fact, the derived quantity
 |
(4) |
where
,
are the Standard Deviations, is called the Correlation of
and
. Note that if
and
are taken from the same set of
variates (say,
), then
 |
(5) |
giving the usual Variance
. The covariance is also symmetric since
 |
(6) |
For two variables, the covariance is related to the Variance by
 |
(7) |
For two independent variates
and
,
 |
(8) |
so the covariance is zero. However, if the variables are correlated in some way, then their covariance will be
Nonzero. In fact, if
, then
tends to increase as
increases. If
, then
tends to decrease as
increases.
The covariance obeys the identity
By induction, it therefore follows that
See also Correlation (Statistical), Covariance Matrix, Variance
© 1996-9 Eric W. Weisstein
1999-05-25