PROPERTIES OF EXPECTATIONS
Let X be a r.v. a,b be constants
Then
(a) E (aX + b ) = a E ( x ) + b
(b) Var (aX + b ) = a 2 Var (X )
If X 1 , X 2 ...... X n are any n rvs,
E(X 1 + X 2 + ....... + X n ) = E(X 1 ) + E(X 2 ) + .... + E(X n )
But if X 1 ,.....X n are n indep rvs then
Var (X 1 + X 2 + ..... + X n ) = Var (X 1 ) + Var (X 2 ) + .... + Var (X n )
In particular if X,Y are independent
Var (X + Y ) = Var(X − Y ) = Var (X ) + Var (Y )
Please note : whether we add X and Y or subtract Y from X, we always must add their
variances.
If X,Y are two rvs, we define their covariance
COV (X, Y ) = E[(X − µ 1 )(Y − µ 2 )]
Where µ 1 = E(X ), µ 2 = E(Y )
Th. If X,Y are indep, E(XY ) = E(X )E(Y ) and COV (X, Y ) = 0
100
Sample Mean
Let X 1 , X 2 .....X n be n indep rvs each having the same mean µ and same variance σ 2 .
We define
X 1 + X 2 + ... + X n
X=
n
X is called the mean of the rvs X 1 .....X n . Please note that X is also a rv.
Theorem
1. ()
E X =µ
2. ()
Var X =
σ2
n
.
Proof
(i) ()
EX =
1
n
[E(X 1 ) + E(X 2 ) + .... + E(X n )]
1 µ + µ + ..... + µ
= =µ
n n times
(2) ()
Var X =
1
n2
[Var (X 1 ) + Var(X 2 ) + .... + Var(X n )]
(as the variables are independent)
1 σ 2 + σ 2 + .. + σ 2 nσ 2 σ 2
= = =
n2 n times n2 n
101
Sample variance
Let X 1 ...X n be n indep rvs each having the same mean µ and same variance σ 2 . Let
X1 + X 2 + X n
X= be their sample mean. We define the sample variance as
n
1
(X )
n 2
S2 = i −X
n −1 i =1
Note S 2 is also a r.v.
( )
E S2 =σ 2
Proof. Read it on page 179.
Simulation
To simulate the values taken by a continuous r.v. X, we have to use the following
theorem.
Theorem
Let X be a continuous r.v. with density f(x) and cumulative distribution function F(x). Let
U = F ( X ) . Then U is a r.v. having uniform distribution on (0,1).
In other words, U is a random number. Thus to simulate the value taken by X, we take a
random no U from the table 7 (Now you must put a decimal point before the no) And
solve for X, the equation
F (X ) = U
102
Example 24
Let X have uniform density on (α , β ) . Simulate the values of X using the 3-digit random
numbers.
937, 133, 753, 503, …..
Solution
Since X has uniform density on (α , β ) its density is
1
α <n<β
f (x ) = β −α
0 elsewhere
Thus the cumulative distribution function is
0 x ≤α
F (x ) = x −α
β −α α <x≤β
1 x>β
X −α
F(X ) = means =
β−α
_ _
∴X = α + (β − α )
Hence if = .937, X = α + (β − α ).937
= .133, X = α + (β − α ).133
etc.
Let x have exponential density (with parameter β )
x
−
f (x ) =
1
β
e β
x>0
0 elsewhere
103
Hence the cumulative distribution function is
0 x≤0
F(x ) = − xp
1− e x>0
Thus solving F(X ) = U, (ie) 1 − e
− βx
= U for X, we get
1
X = β ln
1− U
Since U is a random number implies 1-U is also a random number, we can as well use the
formula
1
X = β ln
U
= −β ln U.
Example 25
X has exponential density with parameter 2. Simulate a few values of X.
Solution
The defining equation for X is
X = −2 ln
Taking 3 digit random numbers form table 7 page 595 row 21 col. 3, we get the random
numbers : 913, 516, 692, 007 etc.
The corresponding X values are :
− 2 ln(.913), − 2 ln(.516), − 2 ln (.692 )..........
104
Example 26
The density of a rv X is given by
f (x ) = x − 1 < x < 1
= 0 elsewhere
Simulate a few values of X.
Solution
First let us find the cumulative distribution function F(x).
Case (i) x ≤ 1 In this case F(x) = 0
Case (ii) − 1 < x ≤ 0.
x x
F (x ) = f (t )dt = t dt
−∞ −1
x
1− x2
= − t dt =
−1
2
Case (iii) 0 < x ≤1
x
In this case F ( x ) = f (t )dt
−∞
−1 0 x
= 0 dt + − t dt + tdt
−∞ −1 0
1 x2 1+ x2
= 0+ + =
2 2 2
105
Case (iv) x>1. In this case F(x) =1
Thus
0 x ≤ −1
F(x ) =
2
1− x
2
−1 < x ≤ 0
1+ x 2
2
0 < x ≤1
1 x >1
To simulate a value for X, we have to solve the equation F(x) = U for X
1
Case (i) 0 ≤U <
2
In this case we use the equation
1− x 2
F(x ) = = U(why ?)
2
∴ X = − 1 − 2 U (why ?)
1
Case (ii) ≤U <1
2
In this case we solve for X, the equation
1+ X 2
F(X ) = =U
2
∴ X = + 2U − 1
Thus the defining conditions are :
1
If 0 ≤ U < , X = − 1 − 2U
2
and
1
If ≤ U < 1, x = + 2U − 1
2
106
Let us consider the 3 digit random numbers on page 594 Row 17 Col. 5
726, 282, 272, 022,…….
1
U = .726 ≥ Thus X = + 2 × .726 − 1 = 0.672
2
1
U = .281 < Thus X = − 21 − 2 × .281 = − 0.662
2
Note : Most of the computers have built in programs which generate random deviates
from important distributions. Especially, we can invoke the random deviates from a
standard normal distribution. You may also want to study how to simulate values from a
standard normal distribution by Box-Muller-Marsaglia method given on page 190 of the
text book.
Example 27
Suppose the no of hours it takes a person to learn how to operate a certain machine is a
random variable having normal distribution with µ = 5.8 and σ = 1.2. Suppose it takes
two person to operate the machine. Simulate the time it takes four pairs of persons to
learn how to operate the machine. That is, for each pair, calculate the maximum of the
two learning times.
Solution
We use Box-Muller-Marsaglia Method to generate pairs of values z1 , z 2 taken by a
standard normal distribution. Then we use the formula
x1 = µ + σz1
x 2 = µ + σz 2
to simulate the time taken by a pair of persons.
(where µ = 5.8, σ = 1.2 )
We start with the random numbers from Table 7
107
Page 593, Row 19, Column 4
729, 016, 672, 823, 375, 556, 424, 854
Note
z 1 = − 2 ln (u 2 ) Cos (2πµ 1 )
z 2 = − 2 ln u 2 Sin (2πu 1 )
The angles are expressed in radians.
U1 U2 Z1 Z2 X1 X2
.729 .016 -0.378 -0.991 5.346 4.611
etc.
Review Exercises
5.108. If the probability density of a r.v. X is given by
f (x ) =
(
k 1− x 2 ) 0 < x <1
0 elsewhere
Find the value of k and the probabilities
(a) P(0.1 < X < 0.2)
(b) P(X > 0.5)
Solution
∞ 1
f ( x )dx = 1 gives w s ( )
k 1 − x 2 dx = 1
−∞ 0
1
or k 1 − =1
3
3
∴k =
2
108
The cumulative distribution function F(x) of X is:
Case (i) x ≤ 0 ∴ F (x ) = 0
x
Case (ii) 0 < x ≤ 1 , F(x ) = ( )
k 1 − t 2 dt
0
3 x3
= x− .
2 3
Case (iii) x > 1. F(x ) = 1
∴P(0.1 < X < 0.2 ) = F(0.2) − F(0.1)
(0.2)2 − (0.2) (0.1) − (0.1)
3 3
3 3
= −
2 3 2 3
P(X < 0.5) = 1 − P(X ≤ 0.5)
(0.5) − (0.5)
3
3
= 1 − F(0.5) = 1 −
2 3
5.113: The burning time X of an experimental rocket is a r.v. having the normal
distribution with µ = 4.76 sec and σ = 0.04 sec . What is the prob that this kind of rocket
will burn
(a) <4.66 Sec
(b) > 4.80 se
(c) anywhere from 4.70 to 4.82 sec?
Solution
X − µ 4.66 − 4.76
(a) P(X < 4.66 ) = P <
σ 0.04
= P(Z < −0.25) = 1 − P(Z < 0.25)
= 1 − F (0.25) = 1 − 0.5987 = 04013
109
X − µ 4.80 − 4.76
(b) P(X > 4.80 ) = P >
σ 0.04
= P(Z > 1) = 1 − F (1) = 1 − 0.8413 = 0.1587
(c) P(4.70 < X < 4.82)
4.70 − 4.76 X − µ 4.82 − 4.76
=P < <
0.04 σ 0.04
= P(− 1.5 < Z < 1.5)
= 2F(1.5) − 1 = 2 × 0.9332 − 1 = 0.8664
5.11 The prob density of the time (in milliseconds) between the emission of beta particles
is a r.v. X having the exponential density
0.25e −0.25 x>0
f (x ) =
0 elsewhere
Find the probability that
(a) The time to observe a particle is more than 200 microseconds (=200x 10-3
milliseconds)
(b) The time to observe a particle is < 10 microseconds
Solution
(a) (
P(> 200 micro sec ) = P X > 200 × 10 −3 milli sec )
∞
= [
0.25e −0.25 x dx = − e − 0.25 x ]
∞
200×10 − 3
−3
200×10
−3
= e −50×10
110
(b) P(X < 10 micro sec onds ) = P X < 10 × 10 −3 ( )
10×10 −3
= [
0.25 e − 0.25 x dx = − e −0.25b ] 10×10 − 3
0
0
−3
= 1 − e − 2.5×10
5.120: If n sales people are employed in a door-to-door selling campaign, the gross sales
volume in thousands of dollars may be regarded as a r.v. having the Gamma distribution
1
with α = 100 n and β = . If the sales costs are $5,000 per salesperson, how many
2
sales persons should be employed to maximize the profit.
Solution
For a Gamma distribution µ = αβ = 50 n . Thus (in thousands of dollars) the “average”
profit when n persons are employed.
= T = 50 n − 5n (5 x 1000 per person is the cost per person)
This is a maximum (using calculus) when n = 25.
5.122: Let the times to breakdown for the processors of a parallel processing machine
have joint density
0.04e −0.2 x −0.2 y x > 0, y > 0
f ( x, y ) =
0 elsewhere
where X is the time for the first processor and Y is the time for the 2nd processor. Find
(a) The marginal distributions and their means
(b) The expected value of the sum of the X and Y.
(c) Verify that the mean of a sum is the sum of the means.
111
Solution
(a) Marginal density of X
∞ ∞
= g (x ) = f (x , y )dy = 0.04e − 0.2 x − 0.2 y dy
y = −∞ y =0
∞
− 0.2 x
= 0 .2 e 0.2e − 0.2 y dy = 0.2e − 0.2 x , x > 0
y=0
(and = 0 if x ≤ 0 )
By symmetry, the marginal distribution of Y is
0.2e −0.2 y y>0
h( y ) =
0 elsewhere
1
Since X (& Y) have exponential distributions (with parameters = 5 ) E(X)
0 .2
= E(Y) = 5.
E since f(x,y) = g (x) h (y), X,Y are independent.
∞ ∞
E(X + Y ) = (x + y ) f (x, y )dydx
−∞ −∞
∞ ∞
= (x + y )(0.04)e −0.2 x −0.2 y dydx
x =0 y = 0
∞ ∞
= x.0.04e − 0.2 x − 0.02 y dydx
x =0 y = 0
112
∞ +∞
+ y × 0.04e − 0.2 x − 0.2 y dydx
x =0 y =0
= 5 + 5 = 10 (verify!)
= E(X ) + E(Y )
5.123: Two random variable are independent and each has binomial distribution with
success prob 0.7 and 2 trials.
(a) Find the joint prob distribution.
(b) Find the prob that the 2nd variable is greater than the first.
Solution
Let X,Y be independent and have Binomial distribution with parameters n = 2, and
p = 0.7 Thus
2
P(X = k ) = (0.7 )k (0.3)2− k k = 0,1,2
k
2
P(Y = r ) = (0.7 )r (0.3)2− r r = 0,1,2
r
∴ P(X = k , Y = r ) = P(X = k )P(Y = r ) as X, Y are independent.
2 2
= (0.7 )k + r (.3)4−(k + r )
k r
0 ≤ k, r ≤ 2
113
(b) P(Y > X )
= P(Y = 2, X = 0 or1) + P(Y = 1, X = 0 )
2 2 2
= (0.7 )2 (.3)0 (0.7 )0 (0.3)2 + (0.7 )1 (0.3)1
2 0 1
2 2
+ (0.7 )1 (0.3)1 (0.7 )0 (0.3)2
1 0
5.124 If X1 has mean – 5, variance 3 while X2 has mean 1 and variance 4, and the two are
independent, find
(a) E(3X 1 + 5X 2 + 2)
(b) Var (3X 1 + 5X 2 + 2)
Ans:
(a) 3 (− 5) + 5(1) + 2 = −8
(b) 9 × 3 + 25 × 4 = 127
114