Covariance
Let
In terms of
Covariance is a real that gives some sense of how much two variables are linearly related to each other, as well as the scale of these variables. Covariance is a function of a probability space and two random variables.
In it's most reduced form, a random variable that is the product of two other random variables is what fully determines covariance, which can be seen as the mean of this variable.
Not just two probability distributions
It's important to note how you can't tell if two random variables are independent just by looking at both of their probability distributions. What is important is if there is some relationship between where
Interpreting the number
High absolute values of the covariance mean that the values change very much and are both far from their respective means at the same time. If the sign of the covariance is positive, then both variables tend to take on high values simultaneously. If the sign is negative, then one variable tends to taken on a relatively high value at the times where the other takes on a relatively low value and visa versa.
Independence and covariance
Two random variables that are independence have zero covariance! :o
This is a “implies” statement, and not a iff statement. (Dependent variables can have zero covariance).
Relation to correlation
The correlation, another measure, normalized the contribution of each variable in order to remove the effect of scale.
Alternative form
Covariance has another form:
Some visual reasoning:
.jpg)
Covariance (and many operations on distributions) can be best understood by rooting one's focus on the distribution of a 1D random variable just before an operation such as taking an expectiation. For covariance, this means focusing on the random variable
There seems to be a typo in the below image.
More than 3 variables
The covariance of a random vector