0% found this document useful (0 votes)
124 views

Random Variables (R.V.) : Probability Theory

The document discusses random variables and probability distributions. It defines random variables as functions that map outcomes to real numbers in a way that preserves probabilities. Discrete random variables take on countable values, while continuous random variables have probability density functions. It presents properties and operations of random variables, including independence, expectation/mean, and how to represent joint distributions of two random variables. Examples are provided to illustrate key concepts such as cumulative distribution functions.
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
124 views

Random Variables (R.V.) : Probability Theory

The document discusses random variables and probability distributions. It defines random variables as functions that map outcomes to real numbers in a way that preserves probabilities. Discrete random variables take on countable values, while continuous random variables have probability density functions. It presents properties and operations of random variables, including independence, expectation/mean, and how to represent joint distributions of two random variables. Examples are provided to illustrate key concepts such as cumulative distribution functions.
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

Probability Theory

1
3. Random variables (r.v.)
Let (O, K, P) a probability field (or probability o-field).

Definition
A function X: OR with the property that: xeR, {eeOX(e)<x}e K is called a
random variable on this field.
Notations
- X(e) is the value of r.v. for event e;
- {eeOX(e)<x}={X<x};
Properties
Consider (O, K, P) a probability field (or probability o-field) and X, Y: OR two
random variables. Then
1) {eeOX(e)>Y(e)}e K;
2) {eeOX(e)sY(e)}e K;
3) {eeOX(e)=Y(e)}e K.

Definition
Consider (O, K, P) a probability field (or probability o-field) and X: OR a random
variable. The function F:RR given by:
F(x)=P({eeOX(e)<x})=P(X<x)
is named cumulative distribution function (c.d.f.) of X.
Properties
The cumulative distribution function F:RR, of a random variable X, has the following
properties:
1) 1 ) ( F lim =

x
x
and lim F( ) 0
x
x

= ;
2) F(x)e[0, 1] xeR;
3) If a<b P(asX<b)=F(b)-F(a)
4) F monotonically nondecreasing function: asbF(a) sF(b);
5) F is left continuous at any x
0
eR: ) ( F ) ( F lim
0
0
0
x x
x x
x x
=
<



Probability Theory
2
3.2 Discrete random variables. Operations with discrete random variables.
Consider (O, K, P) a probability field (or probability o-field) and X: OR a random
variable.

Definition
If the set X(O)={X(e)eeO} is at most countable (that is, a finite or countable set),
then X is a discrete random variable.
Remark
If X is a discrete random variable then X(O) is either finite - X(O)={x
1
, x
2
, , x
n
}- or
countable - X(O)={x
1
, x
2
, , x
n
,}.

Definition
The distribution of a discrete random variable is defined by the formula P(X=x
i
)=p
i

x
i
eX(O) and can be represented by the following distribution table:
X:
|
|
.
|

\
|
... ...
... ...
2 1
2 1
n
n
p p p
x x x

(or equivalently, X:
I e
|
|
.
|

\
|
i
i
i
p
x
with IcN).
Properties
Consider the distribution of a discrete random variable X, given by P(X=x
i
)=p
i
,
ieIcN. Then:
1) p
i
>0, ieI.
2)

eI
i
p
i
=1.
Proof
1) p
i
is a probability, so p
i
>0, ieI.
2) O= } ) X( {
I
i
i
x = e e
e

I
A
e i
i
and A
i
A
j
=u i=j, i, jeI. Then
1=P(O)=P(

I
A
e i
i
)=

eI
) P(A
i
i
=

eI
i
p
i
.
Example
The distribution table of the random variable X representing the face obtained when a
die is rolled is X:
|
|
|
.
|

\
|
6
1

6
1

6
1

6
1

6
1

6
1
6 5 4 3 2 1
.

Probability Theory
3
Remark
The c.d.f. of a discrete random variable having the distribution table X:
I e
|
|
.
|

\
|
i
i
i
p
x
with IcN can
be computed by the formula: F(x)=

< e } {

x x j i
i
j
p .
Definition
Let X, Y: OR two random variables defined on the same probability field (O, K, P). We
say that X, Y are independent iff
P((X<x)(Y<y)) =P(X<x) P(Y<y) (or P(X<x, Y<y)= P(X<x) P(Y<y))
Consider (O, K, P) a probability field (or probability o-field) and X, Y: OR two random
variables. The function Z: OR
2
defined by Z(e)=(X(e), Y(e)) eeO is called
two-dimensional (or bivariate) random variable (or random vector) and the random
variables X, Y are called marginal random variables (notation Z=(X, Y)).
If the random variables X and Y are discrete, then Z=(X, Y) is a discrete two-dimensional
random variable and the set Z(O) is at least countable.
The c.d.f. of Z is defined by:
F(x, y)=P(X<x, Y<y)=P({eeOX(e)<x, Y(e)<y}) (x, y) e R
2
.
Remark
Using the c.d.f. of random vector Z=(X, Y) we can write
X, Y independent iff F
Z
(x, y)=F
X
(x) F
Y
(y).

The distribution function of a discrete two-dimensional random variable Z=(X, Y).
Let X:
I e
|
|
.
|

\
|
i
i
i
p
x
, IcN and Y:
J i
j
j
q
y
e
|
|
.
|

\
|
, JcN the marginal random variables of
Z=(X, Y). We denote A
i
={eeOX(e)=x
i
}, B
j
={eeOY(e)=y
j
},
C
ij
=A
i
B
j
={eeOX(e)=x
i
, Y(e)=y
j
}, p
ij
=P(C
ij
) for every ieI, jeJ.
Remark
- The set (C
ij
)
(i,j)eIJ
is a partition of O.
- It can be proved that the probabilities p
ij
have the following properties:
- p
ij
>0, (i, j)eIJ;
-

eJ j
ij
p =p
i
, ieI and

eI i
ij
p =p
j
, jeJ;

e J I j i
ij
p
) , (
=1.

Probability Theory
4
As a consequence, the random variable Z=(X, Y) can be given by the matrix:
Z:
J I ) , (

) , (
e
|
|
.
|

\
|
j i
ij
j i
p
y x
.
In the case of finite marginal random variables (I=n, J=m), the random variable Z=(X, Y)
is suitable to be represented by the following table:
Y
X
y
1
y
2
y
j
y
m
P(X=x
i
)
x
1
x
2
p
11
...
.....
p
1j
p
1m

.

p
1
p
2

x
i


...

p
ij



p
i


x
n
p
n1
........ p
nj
.. p
nm
p
n
P(Y=y
j
) q
1
q
2
........

q
j
q
m



Remark
The discrete two-dimensional random variables X, Y are independent iff
p
ij
=p
i
q
j
, (i, j)eIJ.


Proposition 1 (operations with random variables)
Consider X : OR a random variables defined on the probability field (or probability o
field) (O, K, P). Then:
1) aX, aeR,
2) X,
3) X
k
, keN*,
4)
X
1
if 0eX(O),
are also random variables.



Probability Theory
5
Proof
1) If a=0 then xeR we have
{eeO(aX)(e)<x}=

< e > e e e
> e < e e e
0 a if , }
a
x
) ( X {
0 a if , }
a
x
) ( X {
K
K
.
If a=0 then {eeO(aX)(e)<x}=

< e
> e u
0 x if ,
0 x if ,
K
K
. Therefore, (aX) is a random variable.
2) For xeR we can write
{eeO X(e)<x}=

> e < e < e e


s e u
0 x if , } x ) ( X x {
0 x if ,
K
K
.
So X is a random variable.
3) If k=2p, peN* then, xeR,
{eeOX
k
(e)<x}=

> e < e < e e


s e u
0 x if , } x ) ( X x {
0 x if ,
k k
K
K
.
If k=2p+1, peN* then, xeR, {eeOX
k
(e)<x}={eeOX(e)<
k
x }eK.
So X
k
is a random variable.

4) For xeR, we have
{eeO
) ( X
1
e
<x}=

> e < e < e


< e < < e
= e < e
0 if , } 0 ) ( { )} (
1
{
0 x if , } 0 ) (
1
{
0 x if , } 0 ) ( {
x X X
x
X
x
X
K
K
K
e e e e
e e
e e

Proposition 2 (operations with random variables)
Consider X, Y:OR random variables defined on the probability field (or probability o
field) (O, K, P). Then:
5) X-Y, X+Y, XY,
6)
Y
X
if 0eX(O),
7) max{X,Y}, min{X, Y},
are also random variables.
Probability Theory
6
3.3 Continuous random variables.
Definition
Consider (O, K, P) a probability o field. A random variable X: OR is said to be continuous
iff - f: R[0, ) integrable on R such that:
F(x)=
}

x
dt ) t ( f ,
where F is the cumulative distribution function of X.
In this case f is called probability density function (p.d.f.).
Properties (of the p.d.f.)
Let f: R[0, ) a probability density function of a random variable X. Then
1) f(x)>0, xeR.
2)
}


dx ) x ( f =1.
3) F(x)=f(x) provided that F(x) exists.
4) P(asX<b)=P(a<Xsb)=P(asXsb)=P(a<X<b)=
}
b
a
dx ) x ( f with a<b.
Exercise 1.
Consider a continuous random variable X with the p.d.f. of given by: f(x)=

e
otherwise , 0
b] [a, x if , k
.
Compute k and the c.d.f. F(x).

Exercise 2.
Consider a continuous random variable X with the p.d.f. of given by: f(x)=

e
otherwise , 0
3] [0, x if ,
2
kx
.
Compute k, c.d.f. F and P(1<X<2).

3.4. Numerical values associated to random variables.
Definition (expectation, mean value, expected value)
Consider X: OR a random variable on a probability field (or probability o field).
The expectation of r. v. X, denoted E(X), is defined by:
Probability Theory
7
- E(X)=

eI i
i i
p x if X:
I i
i
i
p
x
e
|
|
.
|

\
|
a discrete random variable;
- E(X)=
}


dx ) x ( xf if X: R e
|
|
.
|

\
|
x ,
) x ( f
x
is a continuous random variable.

Properties (expectation)
Consider X, Y: OR discrete random variables defined on the same probability field.
1) E(a)=a, aeR.
2) E(aX)=aE(X), aeR.
3) E(X+Y)=E(X)+E(Y).
4) E(XY)=E(X)E(Y) if X, Y are independent.
Proof
1) We have a:
|
|
.
|

\
|
1
a
, aeR, so E(a)=a1=a.
2) If X:
I i
i
i
p
x
e
|
|
.
|

\
|
a discrete random variable, then aX:
I i
i
i
p
ax
e
|
|
.
|

\
|
so
E(aX)=

eI i
i i
p ax =a

eI i
i i
p x =aE(X).
If X: R e
|
|
.
|

\
|
x ,
) x ( f
x
is a continuous random variable, then aX: R e
|
|
.
|

\
|
x ,
) x ( f
ax
so
E(aX)=
}


dx ) x ( axf =a
}


dx ) x ( xf =aE(X).
We prove the properties 3 and 4 only for discrete case.
3) Let X, Y: OR discrete random variables, X:
I i
i
i
p
x
e
|
|
.
|

\
|
, Y:
J j
j
j
q
y
e
|
|
.
|

\
|
, I, JcN and
X+Y:
J I ) j , i (
ij
j i
p
y x
e
|
|
.
|

\
| +
. Then we can write
E(X+Y)=

e
+
J I ) j , i (
ij j i
p ) y x ( =

e
+
J I ) j , i (
ij j ij i
p y p x =

e e I i J j
ij i
) p x ( +

e e J j I i
ij j
) p y ( =
=

eI i
i i
p x +

eI j
j j
q y =E(X)+E(Y).
Probability Theory
8
4) Let X, Y: OR independent discrete random variables, X:
I i
i
i
p
x
e
|
|
.
|

\
|
, Y:
J j
j
j
q
y
e
|
|
.
|

\
|
,
I, JcN and XY:
J I ) j , i (
j i
j i
q p
y x
e
|
|
.
|

\
|
. Therefore E(XY)=

e J I ) j , i (
j i j i
q p y x
=

e J I ) j , i (
j j i i
) q y )( p x ( =

eI i
i i
) p x (

eJ j
j j
) q y ( =E(X)E(Y).
Definition
The r
th
(initial) moment of a random variable X is defined by:
M
r
=E(X
r
)=

e
|
|
.
|

\
|
|
|
.
|

\
|
}


e
e
r.v. continuous , x ,
) (
: X if ) (
variable random discrete
p
x
: X if
i
i
R
x f
x
dx x f x
p x
r
I i
I i
i
r
i
, with reN*.
It is obvious that m
1
=E(X).

Another important typical value of a random variable is the variance, which measures the
dispersion of the values of X around its mean.

Definition (variance)
The variance of a random variable X: OR is defined by:
Var(X)=E[(X-E(X))
2
].

Remark
If E(X
2
)< then the variance of r. v. X can be computed as follows:
Var(X)=E(X
2
)-E
2
(X)
because Var(X)=E[(X-E(X))
2
]=E[X
2
-2XE(X)+E
2
(X)]=E(X
2
)-2E(X)E(X)+E(E
2
(X))=
=E(X
2
)-2E
2
(X)+E
2
(X)=E(X
2
)-E
2
(X).

Properties (variance)
Consider X, Y: OR, random variables defined on the same probability (o) field
1) Var(a)=0, aeR ;
2) Var(aX)=a
2
Var(X), aeR.
3) Var(XY)=Var(X)+Var(Y) if X, Y are independent random variables.
Probability Theory
9
Proof
1) Var(a)=E(a
2
)-E
2
(a)=a
2
-a
2
=0.
2) Var(aX)=E(a
2
X
2
)-E
2
(aX)=a
2
E(X
2
)-(aE(X))
2
=a
2
[E(X
2
)-E
2
(X)]=a
2
Var(X).
3) As X, Y are independent r. v. we have E(XY)=E(X)E(Y) (*).
Var(XY)=E[(XY)
2
]-E
2
(XY)= E(X
2
+Y
2
2XY)-(E(X) E(Y))
2
=
=E(X
2
)+E(Y
2
) 2E(XY)-(E
2
(X)+ E
2
(Y) 2E(XY))=
=E(X
2
)-E
2
(X)+E(Y
2
)-E
2
(Y) 2[E(XY)-E(X)E(Y)]= Var(X)+Var(Y) from (*).
Definition
Consider X: OR a random variable. The central moment of r
th
order of X is defined by:

r
=E[(X-E(X))
r
] with reN*.
The square root of Var(X) is called the standard deviation of X and we write: o
X
= ) X ( Var .
Remark
1)
1
=0;
2)
2
=Var(X).

Definition (central moment)
Consider X, Y:OR univariate random variables defined on the probability field (or
probability o - field) (O, K, P) and Z=(X, Y) bivariate (two-dimensional) random variable.
The (r, s) order central moment of Z is defined by:
r,s
=E[(X-E(X))
r
(Y-E(Y))
s
]
If Z=(X,Y) is discrete, X:
I i
i
i
p
x
e
|
|
.
|

\
|
, Y:
J j
j
j
q
y
e
|
|
.
|

\
|
, I, JcN then

r,s
=

e

J I ) j , i (
ij
s
j
r
i
p )) Y ( M y ( )) X ( M x ( where p
ij
=P(X=x
i
, Y=y
j
), (i, j)eIJ.
Particular cases
Let Z=(X, Y) bidimensional random variable then:
1)
2,0
=E[(X-E(X))
2
]=Var(X)
2)
0,2
=E[(Y-E(Y))
2
]=Var(Y)
3)
1,1
=E[(X-E(X))(Y-E(Y))]
not
= cov(X, Y) covariance or random variables X and Y.




Probability Theory
10
Properties (covariance)
1) cov(X, Y)=E(XY)-E(X)E(Y);
2) X, Y independent cov(X, Y)=0;
3) cov(X, Y)=cov(Y, X)
4) cov(X, X)=Var(X).
5) cov(X, a)=0 aeR
6) cov(aX, bY)= ab cov(X, Y)
7) cov(X+X, Y)=cov(X, Y)+cov(X, Y).
Consequence
Var(XY)=Var(X)+Var(Y)2cov(X, Y)

Definition (correlation coefficient)
Consider X, Y:OR random variables defined on the probability field (or probability o field)
(O, K, P). The correlation coefficient of random variables X and Y is defined by
(X, Y)=
Y X
Y) , X cov(
o o
=
) Y ( Var ) X ( Var
) Y ( M ) X ( M ) XY ( M
if o
X
, o
Y
=0.
Properties (correlation coefficient)
1) X, Y independent random variables (X, Y)=0;
2) (X, Y)e[-1, 1], X, Y;
3) If Y=aX+b (X, Y)=1;
4) If o
X
=0 then (X, X)=1 and (X, -X)=-1.

Exercise
Consider the discrete bidimensional random variable Z=(X, Y) given by the table:

Probability Theory
11
a)
Y
X
0 y p
i

x a
12
1
b
1
2
1
c
3
2

q
j

4
3
d

b)
Y
X
0 y p
i

x a
6
1
b
1
12
7
c
3
2

q
j

4
3
d

Find x, y a, b, c, d such as E(X)=
3
1
, E(Y)=
4
1
and for the obtained values, compute X, Y, the
cdf F
X
and F
(X, Y)
(0, 2), XY, cov(X, Y), (X, Y), o
X
, o
Y
.










Probability Theory
12
3.5 Several Classical Distributions

3.5.1 Binomial distribution

We say that a random variable X: OR follows a binomial distribution of parameters n and
p, and we write XeBi(n, p) if P(X=k)=
k
n
C p
k
(1-p)
n-k
, k= n , 0 or X:
|
|
.
|

\
|
k n k k
n
q p C
k
, k= n , 0 ,
q=1-p, pe[0, 1].

Remarks
1) p
k
=
k
n
C p
k
(1-p)
n-k
>0 k= n , 0 .
2)

=

n
0 k
k n k k
n
q p C =(q+p)
n
=1
Expectation: E(X)=np
Variance: Var(X)=npq

3.5.2 Poisson distribution
A random variable X: OR is said to be Poisson distributed with parameter >0
- we write XeP()) iff P(X=k)=e
-
! k
k

, keN , or X:
N e

|
|
|
.
|

\
|

k
k
! k
e
k
.
Remarks
1) p
k
=P(X=k)= e
-
! k
k

>0, keN.
2)

e

e

=
N N k
k
k
k
! k
e p = e
-
e

=1.
Expectation: E(X)=
XeP() E(X)=

>


0 k
k
! k
ke =e
-

>

1 k
k
! k
k = e
-


>

1 k
1 k
)! 1 k (
= e
-

>

0 k
k
! k
= e
-
e

=.
Variance: Var(X)=
Probability Theory
13
E(X
2
)=

>


0 k
k
2
! k
e k = e
-

>

1 k
k
)! 1 k (
k = e
-

>

1 k
1 k
)! 1 k (
k =
= e
-
(

>

1 k
1 k
)! 1 k (
) 1 k ( +

>

1 k
1 k
)! 1 k (
)= e
-
(

>

2 k
1 k
)! 2 k (
+ e

)= e
-
( e

+ e

)=(+1).
Hence Var(X)=E(X
2
)-E
2
(X)= (+1)-
2
=.

3.5.3 Hypergeometric distribution
A random variable X: OR has a hypergeometric distribution of parameters n, a and b (it is
written XeH(n, a, b)) iff P(X=k)=
k n-k
a b
n
a+b
C C
C

, ke
0
0, n where a, b, neN,
0
n min{ , n} a = and
ns a+b (or X:
0
k n-k
a b
n
a+b
k 0, n
k
C C
C
e
| |
|

|
|
\ .
).
Remarks
1) p
k
=P(X=k)=
k n-k
a b
n
a+b
C C
C

>0, ke
0
0, n .
2)
0 0 0
n n n
k n-k
a b k n-k n
k a b a+b
n n n
a+b a+b a+b
k=0 k=0 k=0
C C 1 1
p C C C 1
C C C

= = = =

.

Expectation: E(X)= n
a
a b +

Variance: Var(X)= n
1
a b a b n
a b a b a b
+
+ + +
.

3.5.4 Normal distribution

We say that a random variable X: OR follows a normal distribution of parameters m and o
- we write XeN(m, o) - if its p.d.f. has the form:
f(x)=
2
2
2
) m x (
e
2
1
o

t o
, meR, o>0.
Remarks
1) f(x)>0 xeR.
Probability Theory
14
2)
}


dx ) x ( f =
t o 2
1
}

dx e
2
2
2
) m x (
=
t o 2
1
}



o dt 2 e
2
t
=
t
1
}



dt e
2
t
=2
t
1
}

0
t
dt e
2
=
=2
t
1
2
t
=1 where (x-m)=to 2 .

Expectation: E(X)=m
E(X)=
}


dx ) x ( xf =
t o 2
1
}

dx xe
2
2
2
) m x (
t
2
m x
=
o

=
t o 2
1
}



o + o dt 2 e ) m t 2 (
2
t
=
=
t o 2
1
(o 2 )
2
}



dt te
2
t
+m
t o 2
1
}



dt e
2
t
=m

Variance: Var(X)=o
2

Var(X)=
}


dx ) x ( f ) m x (
2
=
}

t o
dx e
2
1
) m x (
2
2
2
) m x (
2
=
=
t o 2
1
}

dx e ) m x (
2
2
2
) m x (
2
t
2
m x
=
o

=
t o 2
1
}



o o dt 2 e t 2
2
t 2 2
=
t
o
2
}



dt )' e ( t
2
t
= =
t
o
2
(t
2
2
t
e

-
}



dt e
2
t
)=
t
o
2
2
}

0
t
dt e
2
=
t
o
2
2
2
t
=o
2
.
Standard deviation (o
X
=o)
The c.d.f. values of a normal distributed random variable are computed using Laplace
function defined by: | :RR, |(z)=
t 2
1
}

z
0
t
2
1
dt e
2
.
Properties
1) |(0)=0;
2) |(-z)=- |(z) (odd function);
3) ) z ( lim
z
|

=0.5;
4) ) z ( lim
z
|

=-0.5;

Probability Theory
15
Cumulative distribution function (F(z)=
2
1
+|(
o
m z
))
F(z)=
}

o

t o
z
2
) m x (
dx e
2
1 2
2
t
m x
=
o

=
t o 2
1
}
o

o
m z
t
2
1
dt e
2
=
=
t 2
1
}

0
t
2
1
dt e
2
+
t 2
1
}
o

m z
0
t
2
1
dt e
2
u
2
t
=
=
t 2
1
}



0
u
du ) 2 ( e
2
+|(
o
m z
)=
=
t
1
}

0
u
du e
2
+|(
o
m z
)=
2
1
+|(
o
m z
).
Remark
It is easy to prove that F(-z)=1-F(z), zeR.

The most important normal distribution is the standard normal N(0,1), with m=0 and o
2
=1:
- we write XeN(0, 1) - if its p.d.f. has the form: f(x)=
2
2
2
1
x
e

t
, xeR
3.5.5 Gamma distribution

We say that a random variable X: OR follows a gamma distribution of parameters a and b
- we write Xe(a, b) - if its p.d.f. has the form:
f(x)=
1
1
, x>0
( )
0 , x 0
x
a
b
a
x e
a b

, a, b>0.

Remarks
1) f(x)>0 xeR.
2)
}


dx ) x ( f =
a
b ) a (
1
I
}

0
b
x
1 a
dx e x
t
b
x
=
=
a
b ) a (
1
I
}


0
t 1 a
bdt e ) bt ( =
a
b ) a (
1
I
}


0
t 1 a a
dt e t b =
=
a
b ) a (
1
I
b
a
I(a)=1
Expectation: E(X)=ab
Probability Theory
16
E(X)=
}


dx ) x ( xf =
a
b ) a (
1
I
}


0
b
x
a
dx e x
t
b
x
=
=
a
b ) a (
1
I
}

0
t a
bdt e ) bt ( =
) a (
b
I
}

0
t a
dt e t =
=
) a (
b
I
I(a+1)=
) a (
b
I
aI(a)=ab.

Variance: Var(X)=ab
2

E(X
2
)=
a
b ) a (
1
I
}


+
0
b
x
1 a
dx e x
t
b
x
=
=
a
b ) a (
1
I
}

+
0
t 1 a
bdt e ) bt ( =
) a (
b
2
I
}

+
0
t 1 a
dt e t =
=
) a (
b
2
I
I(a+2)=
) a (
b
2
I
a(a+1) I(a)=ab
2
(a+1)
Var(X)= ab
2
(a+1)-a
2
b
2
=ab
2
.

3.5.6 Exponential distribution

We say that a random variable X: OR follows a negative exponential distribution of
parameter >0 - we write XeExp() - if its p.d.f. has the form: f(x)=

s
>

0 x , 0
0 x , e
x
.

Remark
XeExp()XeI(1,

1
) so E(X)=

1
and Var(X)=
2
1

.
3.5.7 Beta distribution
We say that a random variable X: OR follows a Beta distribution of parameters a and b
and we write Xe(a, b), if its pdf is:
f(x)=




, a, b>0.
Show that the expectation and variance of X are E(X)=
b a
a
+
and Var(X)= .
) 1 ( ) (
2
+ + + b a b a
ab




Probability Theory
17

3.5.8 Chi-square distribution

We say that a random variable X: OR follows a chi-square distribution of parameters n
and o - we write Xe_
2
(n, o) iff Xe(
2
n
, 2o
2
) . Then the p.d.f. of X is:
f(x)=
2
1
2 2
2
2
1
, x>0
( )(2 )
2
0 , x 0
x n
n
x e
n

o

I o

, neN*, o>0 and E(X)=no


2
, Var(X)=2no
4
.
3.5.9 Student distribution

A random variable X: OR follows a Student distribution of parameter n - we write
XeS(n) if its p.d.f. has the formula
f(x)=
)
2
n
( n
)
2
1 n
(
I t
+
I
2
1 n
2
)
n
x
1 (
+

+ , xeR, neN\{0,1,2}.
It can be proved that E(X)=0 and Var(X)=
2 n
n

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy