6 PROBABILITY AND LINEAR SYSTEM THEORY
This shows that P[X ~ a] = 0 for any a> 0 and hence that
P[X >0] =0.
A similar argument shows that P[X < 0] = 0; thus P[X = 0] = 1.
o
A random n-vector X = (X 1, ... , X n)T is a collection of n random
variables
Xl' ... , X .. To examine its probabilistic behaviour it is not
sufficient to know the distribution of each
Xi because this information
does
not specify how the components interact. In general one needs
to know the
joint distribution function F(a1, .•• , an) which specifies
the probabilities
of events via the formula
P[XI <a1,···,Xn<anJ =F(au ... ,an).
The random variables Xl' ... , Xn are independent if
F(a1,···, an) = F l(adF 2(a2)··· Fn(an)
where Fi is the distribution of Xi. This is the only case in
which knowledge
of F 1, ... , F n suffices to determine F. On the other
hand, knowledge of
F always determines the distribution of each Xi
(the so-called marginal distribution) since, for example,
F1(ad=P[Xl <a1,X2 < oo, ... ,Xn< ooJ
= F(a1, 00, ... , 00).
Xl' ... , X n have a joint density function f if
F(a1,···,an )= f~w ... f~oo f(x1,···,xn)dxn···dx1·
If the Xi are independent and Xj has density function fj then
f(x 1'·· ., Xn) = fl (xdf2(X2)··· fn(xn)·
If g: [Rn ~ [R is a continuous function then the expectation Eg(X) can
be defined using Stieltjes integrals in a way
that agrees with the usual
expressIOn
Eg(x) = f~oo··-f~oo g(xl,···,xn)f(xl,···,xn)dxl···dxn
valid when X has joint density f. We give the definition for the
bivariate case
n = 2; for n > 2 it is similar but notationally cumber-