2.11. Random Vectors
We have looked at multiple random variables before. Let \(X_1,X_2,\ldots, X_n\) be random variables then we defined the joint distribution function \(F_{X_1,X_2,\ldots,X_n}\):
In case all random variables are continuous the joint probability density function is defined as
To calculate the probability for an event such that the values of the random variables are within a subset \(A\subset\setR^n\) we calculate the multivariate integral:
Note that \(dx_1dx_2\cdots dx_n\) is an infinitesimal small hypervolume element. Multiplying the probability density with this volume element results in a probability, adding these probabilities for all volume elements in the set \(A\) results in the desired probability.
Very often there is the need to characterize all random variables with one entity. Think of situations where there are hundreds or even thousands of random variables involved, a situation that often occurs in practice. For this the random vectors are introduced:
A random vector is a vector whose elements are random variables. With this notation the distribution function can be abbreviated as \(F_{\v X}\) and the probability density function as \(f_{\v X}\). Using the vectorial notation the probability for the event \(\v X\in A\) equals:
For discrete random variables the distribution function is defined equivalently but instead of defining a probability density function the probability mass function is defined:
or equivalently in vectorial notation:
The probability \(\v X\in A\) for a discrete random variable is then a sum:
and for a continuous random variable we get an integral:
Note that \(f_{\v X}(\v x)\) is the probability density that multiplied with the infinitesimal volume \(d\v x\) is a probability. Thus we are summing probabilities and in the limit for inifinitesimal small intervals (\(dx\rightarrow0\)) we are integrating a probability density.
It is possible to define random vectors with a mixture of discrete and continuous random variables. In this lecture series we stick with either pure discrete or pure continuous random vectors.
Consider two random vectors \(\v X\) and \(\v Y\). We define these two random vectors to be independent in case each element \(X_i\) is independent of each element \(Y_j\). Note that elements of \(\v X\) might be dependent! In a formula we can express this notion of independence with:
This formula is again a clear demonstration of the power of using vectors (linear algebra) to describe multivariate statistics where we can write: