Topic_10_Point_estmation_of_parameters
Topic_10_Point_estmation_of_parameters
Kayondo c 1
1 Introduction
In the field of statistical inference, the concern is on making decisions/drawing
conclusions about populations. Statistical inference methods use information
from samples of the underlying populations for drawing conclusions. This
chapter begins the study of statistical methods that are employed for inference
and in decision making.
Statistical inference involves two crucial areas and these are parameter
estimation and hypothesis testing. As an example of a parameter estimation
problem, suppose that a structural engineer is analyzing the tensile strength of a
component used in an automobile chassis. Since variability in tensile strength is
naturally present between the individual components because of differences in
raw material batches, manufacturing processes, and measurement procedures
(for example), the engineer is interested in estimating the mean tensile strength
of the components. In practice, the engineer will use sample data to compute
a number that is in some sense a reasonable value (or guess) of the true mean.
This number is called a point estimate. We will see that it is possible to establish
the precision of the estimate.
Suppose that we are interested in obtaining a point estimate of a population
parameter. We know that before the data is collected, the observations are
considered to be random variables, say X1 , X2 , ..., Xn . Therefore, any function
of the observation, or any statistic, is also a random variable. For example, the
sample mean X̄ and the sample variance S2 are statistics and they are also random
variables. Since a statistic is a random variable, it has a probability distribution.
We call the probability distribution of a statistic a sampling distribution.
When handling statistical inference problems, it is convenient to have a
general symbol to represent the parameter of interest. We normally use the Greek
symbol θ (theta) to represent the parameter. The objective of point estimation is
to select a single number, based on sample data, that is the most plausible value
for θ. A numerical value of a sample statistic is used as the point estimate.
Suppose X is a random variable with probability distribution f (x), charac-
terized by the unknown parameter θ. Suppose further that X1 , X2 , ..., Xn is a
random sample of size n from X, then statistic Θ̂ = h(X1 , X2 , ..., Xn ) is called a
point estimator of θ. It should be noted that Θ̂ is a random variable since it is
a function of random variables. From a given sample, Θ̂ takes on a particular
numerical value θ̂, called the point estimate of θ.
Definition: A point estimate of some population parameter θ is a single
numerical value θ̂ of a statistic Θ̂. The statistic Θ̂ is called the point estimator.
For example, suppose that X is a random variable that is normally distributed
with an unknown mean µ. The sample mean is a point estimator of the unknown
population mean µ. That is, µ̂ = X̄. Having obtained the sample, the numerical
value x̄ is the point estimate of µ. Suppose that x1 = 30, x2 = 35, x3 = 34 and
2 Engineering Mathematics IV– EMT 2201
30 + 35 + 34 + 36
x̄ = = 33.75.
4
Similarly if the population variance σ2 is also unknown, a point estimator
for σ2 is the sample variance S2 . The sample variance is given as:
n
1 X
S2 = (xi − x̄)2
n−1
i=1
• For µ1 − µ2 , the estimate is µˆ1 − µˆ2 = x¯1 − x¯2 , the difference between the
sample means of two independent random samples.
• For p1 − p2 , the estimate pˆ1 − pˆ2 , the difference between two sample
proportions computed from two independent random samples.
Remark: There may be several different choices for the point estimator of a
parameter. For example, if we wish to estimate the mean of a population, we
might consider the sample mean, the sample median, or perhaps the average of
the smallest and largest observations in the sample (range) as point estimators.
In order to decide which point estimator of a particular parameter is the best
one to use, we need to examine their statistical properties and develop some
criteria for comparing estimators.
H.W. Kayondo c 3
Solution
It follows that:
n n !
1 X 1 X
E(S ) =
2
E
Xi − nX̄ =
2 2
2 2
E(Xi ) − nE(X̄ )
n−1 n−1
i=1 i=1
It follows that:
σ2 = E(Xi2 ) − µ2
Therefore,
E(Xi2 ) = µ2 + σ2 (1)
We shall see in section 5 for sampling distribution of X̄. For now we can use
the results. For a population represented by X with mean µ and variance σ2 and
a sample X1 , X2 , ..., Xn , then:
σ2
E(X̄) = µ and Var(X̄) =
n
4 Engineering Mathematics IV– EMT 2201
It follows that:
Var(X̄) = E(X̄2 ) − (E(X̄))2
Therefore,
σ2
= E(X̄2 ) − µ2
n
This gives:
σ2
E(X̄2 ) = + µ2 (2)
n
Substituting equations (1) and (2) into the last expression for E(S2 ) gives:
n
σ2
!
1 X 1
E(S ) =
2
(µ + σ ) − n(µ + ) =
2 2 2
(nµ2 + nσ2 − nµ2 − σ2 )
n−1 n n−1
i=1
E(S2 ) = σ2
variance of any observation is Var(Xi ) = σ2 . Since Var(X̄) < Var(Xi ) for sample
sizes n ≥ 2, we can conclude that X̄ is a better estimator of µ than a single
observation Xi .
S
σ̂X̄ = √ .
n
normally distributed (or approximately so) for large n, this is a very useful result.
Even in cases in which the point estimator is not normally distributed, we can
state that so long as the estimator is unbiased, the estimate of the parameter
will deviate from the true value by as much as four standard errors at most 6
percent of the time. Thus a very conservative statement is that the true value of
the parameter differs from the point estimate by at most four standard errors.
41.60, 41.48, 42.34, 41.95, 41.86, 42.18, 41.72, 42.26, 41.81, 42.04
A point estimate of the mean thermal conductivity at 100◦ F and 550 watts is
the sample mean given as:
x̄ = 41.924 Btu/hr − f t −◦ F
The standard error of the sample mean is σX̄ = √σn , and since σ is unknown,
we replace it with the sample standard deviation S = 0.284 to obtain the
estimated standard error of X̄ as:
S 0.284
σ̂X̄ = √ = √ = 0.0898.
n 10
Note: The standard error is about 0.2 percent of the sample mean, implying
that we have obtained a relatively precise point estimate of thermal conductivity.
If we can assume that thermal conductivity is normally distributed, 2 times the
standard error is 2σ̂X̄ = 2(0.0898) = 0.1796, and we are highly confident that
the true mean thermal conductivity is with in the interval 41.924 ± 0.1756, or
between 41.744 and 42.104.
This means that the mean square error of Θ̂ is equal to the variance of
the estimator plus the squared bias. If Θ̂ is an unbiased estimator of θ, the
H.W. Kayondo c 7
mean square error of Θ̂ is equal to the variance of Θ̂. The mean square error
is an important criterion for comparing two estimators. Let Θ̂1 and Θ̂2 be two
estimators of the parameter θ, and let MSE(Θ̂1 ) and MSE(Θ̂2 ) be the mean square
errors of Θ̂1 and Θ̂2 . Then the relative efficiency of Θ̂2 to Θ̂1 is defined as:
MSE(Θ̂1 )
MSE(Θ̂2 )
If this relative efficiency is less than 1, we would conclude that Θ̂1 is more
efficient estimator of θ than Θ̂2 , that is, it has a smaller mean square error.
Remark: Sometimes we find that biased estimators are preferable to unbiased
estimators because they have smaller mean square error. That is, we may be able
to reduce the variance of the estimator considerably by introducing a relatively
small amount of bias. As long as the reduction in variance is greater than the
squared bias, an improved estimator from a mean square error viewpoint will
result. For example, Figure 2 shows the probability distribution of a biased
estimator Θ̂1 that has a smaller variance than the unbiased estimator Θ̂2 . An
estimate based on Θ̂1 would more likely be close to the true value of θ than
would an estimate based on Θ̂2 . An estimator Θ̂ that has a mean square error
that is less than or equal to the mean square error of any other estimator, for all
values of the parameter, is called an optimal estimator of θ. Optimal estimators
rarely exist.
Figure 2: A biased estimator Θ̂1 that has smaller variance than the unbiased
estimator Θ̂2 .
Exercise
1. Suppose we have a random sample of size 2n from a population denoted
by X with E(X) = µ and Var(X) = σ2 . Let
2n n
1 X 1X
X̄1 = Xi and X̄2 = Xi
2n n
i=1 i=1
3. Suppose that Θ̂1 and Θ̂2 are unbiased estimators of the parameter θ. We
know that Var(Θ̂1 ) = 10 and Var(Θ̂2 ) = 4. Which estimator is better and in
what sense is it better?
4. Calculate the relative efficiency of the two estimators given in exercise 2.
5. Calculate the relative efficiency of the two estimators given in exercise 3.
6. Suppose that Θ̂1 and Θ̂2 are estimators of the parameter θ. If we know that
E(Θ̂1 ) = θ, E(Θ̂2 ) = θ/2, Var(Θ̂1 ) = 10 and Var(Θ̂2 ) = 4. Which estimator
is better and in which sense?
7. Suppose that Θ̂1 , Θ̂2 and Θ̂3 are estimators of θ. Suppose we know
that E(Θ̂1 ) = E(Θ̂2 ) = θ, E(Θ̂3 ) , θ, Var(Θ̂1 ) = 12, Var(Θ̂2 ) = 10 and
E(Θ̂3 − θ)2 = 6. Compare these three estimators. Which do you prefer?
Justify your choice.
8. Let three random samples of sizes n1 = 20, n2 = 10 and n3 = 8 be taken
from a population with mean µ and variance σ2 . Let S21 , S22 and S23 be
sample variances. Show that
n
9. (a) Show that n1 (Xi − X̄) is a biased estimator of σ2 .
P
i=1
(b) Find the amount of bias in the estimator.
(c) What happens to the bias as the sample size n increases?
10. Let X1 , X2 , ..., Xn be a random sample of size n from a population with
mean µ and variance σ2 .
(a) Show that X̄2 is a biased estimator for µ2 .
(b) Find the amount of bias in this estimator.
(c) What happens to the bias as the sample size n increases?
11. Let X1 , X2 , ..., Xn be a random sample of size n from a normal distribution
with mean µ and variance σ2 . Let Xmin and Xmax be the smallest and largest
observations in the sample.
H.W. Kayondo c 9
13. X̄1 and S21 are sample mean and sample variance from a population with
mean µ1 and variance σ21 . Similarly X̄2 and S22 are sample mean and
sample variance from a second independent population with mean µ2 and
variance σ22 . The sample sizes are n1 and n2 , respectively.
14. Suppose that both populations in exercise 13 have the same variance
σ21 = σ22 = σ2 . Show that
is an unbiased estimator of σ2 .
X̄2 X̄
r̂ = n
λ̂ = n
(1/n) Xi2 − X̄i2 (1/n) Xi2 − X̄i2
P P
i=1 i=1
For illustration using data, consider the data given in the first example of
8
subsection 3.1, for this data, x̄ = 21.65 and x2i = 6639.40, so the moment
P
i=1
estimates are:
(21.65)2 21.65
r̂ = 2
= 1.3 λ̂ = = 0.0599
(1/8)6639.40 − (21.65) (1/8)6639.40 − (21.65)2
P(X1 = x1 , X2 = x2 , ..., Xn = xn ).
12 Engineering Mathematics IV– EMT 2201
That is, L(θ) is just the probability of obtaining the sample values x1 , x2 , ..., xn .
Therefore, in the discrete case, the maximum likelihood estimator is an estimator
that maximizes the probability of occurrence of the sample values.
Example Let X be a Bernoulli random variable. The probability mass function
is:
p (1 − p)1−x for x = 0, 1
( x
f (x; p) =
0, elsewhere.
where p is the parameter to be estimated. The likelihood function of a
random sample of size n is:
i=1
For illustration purposes, suppose that this estimator was applied to the
following scenario: n items are selected at random from a production line, and
each item is judged as either defective ( in which case we set xi = 1) or non
n
defective ( in which case we set xi = 0). Then xi is the number of defective units
P
i=1
in the sample, and p̂ is the sample proportion defective. The parameter p is the
H.W. Kayondo c 13
i=1
n 1
λ̂ = n
=
P X̄
xi
i=1
n n
1 1 −(1/2σ2 ) (xi −µ)2
Y P
√ e−(xi −µ) /(2σ ) =
2 2
L(µσ ) =
2
e i=1
n
X
ln(L(µ, σ2 )) = −(n/2) ln(2πσ2 ) − (2σ2 )−1 (xi − µ)2
i=1
n
∂ ln(L(µ, σ2 )) 1 X
= 2 (xi − µ) = 0
∂µ σ i=1
n
∂ ln(L(µ, σ2 )) n 1 X
=− 2 + 4 (xi − µ)2 = 0
∂σ2 2σ 2σ i=1
The solutions to the above equations yield the maximum likelihood estima-
tors and are given as:
n
1X
µ̂ = X̄ and σˆ2 = (Xi − X̄)2 .
n
i=1
Once again, the maximum likelihood estimators are equal to the moment
estimators.
H.W. Kayondo c 15
n−1 2
E(σˆ2 ) = σ
n
The bias is:
n−1 2 −σ2
E(σˆ2 ) − σ2 = σ − σ2 =
n n
ˆ
Because the bias is negative, σ tends to underestimate the true variance
2
Exercise
1. Consider the Poisson distribution
x
e−λλ
f (x) =
x!, x = 0, 1, 2, ...
16 Engineering Mathematics IV– EMT 2201
(θ + 1)xθ , for 0 ≤ x ≤ 1
(
f (x) =
0, elsewhere.
1 −x/θ
f (x) =xe , 0 ≤ x ≤ ∞, 0 < θ < ∞.
θ2
Find the maximum likelihood estimator for θ.
7. Consider the probability density function:
4 Sampling distributions
Consider a population with mean µ and standard deviation σ. If you chose
a sample from this population, we need to study the characteristics of its statistics.
Definition 2 A statistic is any value that describes the characteristics of a sample. For
instance, the sample mean, X̄, and the sample variance S2 .
If only two data points and the sample mean are known, the values of the
remaining two data points cannot be uniquely determined.
Example 1 A company manager has a total budget of 150, 000 to be completely allocated
to four different projects. How many degrees of freedom does the manager have?
x1 + x2 + x3 + x4 = 150, 000. A fourth project’s budget can be determined from the total
budget and the individual budgets of the other three. For example, if x1 = 40, 000, x2 =
30, 000, x3 = 50, 000, then x4 = 150, 000 − 40, 000 − 30, 000 − 50, 000 = 30, 000. So
there are (n − 1) = 3 degrees of freedom.
sample will have the same normal distribution as the population being sampled
and the mean of these observations is:
n
1 1X
X̄ = (X1 + X2 + ... + Xn ) = Xi
n n
i=1
σ2
has normal distribution with mean µX̄ = µ and variance σ2X̄ = n. Indeed, it is
σ2
important to prove that the mean of X̄ = µ and and the variance of X̄, σ2X̄ = n.
Thus,
n n n
1X 1X 1X 1
µX̄ = E(X̄) = E[ Xi ] = [E(Xi )] = µ = .nµ = µ
n n n n
i i=1 i=1
and
n n n
1X 1 X 1 X 2 1 σ2
Var(X̄) = σ2X̄ = Var[ Xi ] = 2 Var(Xi ) = 2 σ = 2 .nσ2 =
n n i=1 n i=1 n n
i=1
Observe from the table above that the sample mean varies from sample to
sample. Thus, the sample mean is a random variable, denoted by X̄. The value
of the random variable X̄ for a given sample is denoted by x̄. The sample mean,
like the sample median or sample variance is a statistic. A statistic is a random
variable that depends only on the observed sample but not on the population
of interest. Since a statistic is a random variable, it must have a probability
distribution.
The table below shows the sampling distribution of the sample mean X̄.
2 × 16 + 2.5 × 16 = 32 Note that E(X̄) (the mean of the sample mean) is equal to the
population mean and it is generally true. Thus E(X̄) = µ
5.1 Variance
Var(X̄) = E(X̄2 ) − [E(X̄)]2 . E(X̄2 ) = 0.52 × 16 + 1 × 61 + 1.52 × 26 22 × 16 + 2.52 × 16 = 15.5
6 .
Therefore, Var(X̄) = 16.06 − 9
4 = 5
12 .
It can be shown that Var(X̄) = σn . N−n
2
N−1 where σ is the population variance, N the
2
Exercise
From the previous example, list all samples of size two drawn with replacement
and verify that E(X̄) = µ and Var(X̄) = σn where µ is the population mean and
2
Note
Recall that N−n
N−1 is called the finite population correction which is approximately
equal to 1 if n, the sample size is too small compared to the population size
N. Thus if N n, the variance of the distribution of the sample mean X̄ when
sampling without replacement can be approximated by σn which is the same as
2
than the whole population are referred to as Sampling errors. The standard
error of the sample mean depends on the population variance σ2 and the sample
size. The Sampling errors are smaller if you sample from a population with
small variance and if the sample size is large.
The theorem states that if all possible random samples of size n are drawn
with replacement from a finite population of size N with mean µ and standard
deviation σ, then for n sufficiently large, the sampling distribution of the mean
X̄ will be approximately normally distributed with mean µX̄ = µ and standard
x̄−µ
deviation σX̄ = √σn . Hence z = σ/ √n is a value of a standard normal variable Z.
Thus we can calculate for any population, approximate probability P(a <
x̄ < b) using the normal distribution. The normal approximation above will be
good if n ≥ 30 for any finite population. If n < 30, the approximation is good
only if the population is not too different from the normal population. If the
population is known to be normal, the sampling distribution of X̄ will follow a
normal distribution exactly, no matter how small the size of the samples.
Example 2 A soft drink machine is set so that the amount of drink dispensed is a
random variable with mean 200ml and a standard deviation of 15ml. What is the
probability that the average amount dispensed in a random sample of size 36 is at least
204ml.
Solution
n=36, µ = 200, σ = 15. By the CLT, the distribution of X̄ is approximately normal
with mean µx̄ = 200 and S.D σx̄ = √15 = 2.5.
36
Therefore P(X̄ ≥ 204) = P( X̄−200
2.5 ≥
204−200
2.5 ) = P(Z ≥ 1.6) = 0.0548.
Example 3 Given the population 1, 1, 1, 3, 4, 5, 6, 6, 6, and 7. Find the probability that
a random sample of size 36, selected with replacement, will yield a sample mean greater
than 3.8 but less than 4.5 if the mean is measured to the nearest tenth.
Solution
The probability distribution of our population may be written as:
x 1 3 4 5 6 7
P(X = x) 0.3 0.1 0.1 0.1 0.3 0.1
H.W. Kayondo c 21
xp(x) = 1∗0.3+3∗0.1+4∗0.1+5∗0.1+6∗0.3+7∗0.1 = 4.
P
The mean is given by P
The variance is given by x2 p(x) − ( xp(x))2 = 12 ∗ 0.3 + 32 ∗ 0.1 + 42 ∗ 0.1 +
P
52 ∗ 0.1 + 62 ∗ 0.3 + 72 ∗ 0.1 − 16 = 5. The sampling distribution of X̄ may be
approximated by the normal distribution with mean µX̄ = µ = 4 and variance
σ2X̄ = σn = 36
2
5
. Thus, standard deviation σX̄ = 0.373. Therefore, the probability
that X̄ is greater than 3.8 and less than 4.5 is given by the area shown below
σX̄ = 0.37
3.85 4 4.45
The z values that correspond to x¯1 = 3.85 and x¯2 = 4.45 are z1 = 3.85−4
0.373 = −0.40
and z2 = 0.373 = 1.21. Therefore, P(3.8 < X̄ < 4.5) ≈ P(−0.40 < Z < 1.21 = P(Z <
4.45−4
Example 4 Let the first population consist of 3, 5 and 7, while the second population
comprise of 0, 4 and 8. Derive the sampling distribution for X̄1 − X̄2 .
Solution
The possible samples from the two populations with their means are given in
Table 1. There are 81 possible differences (x̄1 − x̄2 ) and these are shown in Table
2. The frequency distribution of X̄1 − X̄2 is given in Table 3 with corresponding
probability histogram shown in Figure 3.
22 Engineering Mathematics IV– EMT 2201
It can observed from the probability histogram in Figure 3, that the sam-
pling distribution of X̄1 − X̄2 may be approximated by a normal curve. This
approximation improves as n1 and n2 increase.
From Table 3, the sampling distributions for X̄1 and X̄2 were obtained and
are shown in Table 4 and from this table, we obtained the quantities:
µX̄1 = 5, µσX̄1 = 1.3333, µX̄2 = 4, µσX̄2 = 5.3333.
From Table 3, we obtained µX̄1 −X̄2 = 1 and σ2X̄ −X̄ = 6.6667.
1 2
From the computations done, we observe that:
σ21 σ22
µX̄1 −X̄2 = µx̄1 − µx̄2 = µ1 − µ2 and σ2X̄ = σ2x̄1 + σ2x̄2 = n1 + n2
1 −X̄2
Remark
For the sampling distribution of X̄1 − X̄2 , it is enough to stop your computation
on Table 3.
Table 1: Means of Random Samples with Replacement from Two Finite Popula-
tions.
From Population 1 From Population 2
No. Sample x̄1 No. Sample x̄2
1 3, 3 3 1 0, 0 0
2 3, 5 4 2 0, 4 2
3 3, 7 5 3 0, 8 4
4 5, 3 4 4 4, 0 2
5 5, 5 5 5 4, 4 4
6 5, 7 6 6 4, 8 6
7 7, 3 5 7 8, 0 4
8 7, 5 6 8 8, 4 6
9 7, 7 7 9 8, 8 8
σ21 σ22
variance σ2X̄ = n1 + n2 . Hence
1 −X̄2
Solution
Example 6 Phones from manufacturer A have a mean guarantee period of 6.5 years and
standard deviation of 0.9 years, while those of manufacturer B have a mean guarantee
24 Engineering Mathematics IV– EMT 2201
period of 6.0 years and a standard deviation of 0.8 years. What is the probability that a
random sample of 36 phones from manufacturer A will have a mean guarantee period
that is at least one year more than the mean guarantee period of a sample of 49 phones
from manufacturer B?
Solution
From the sampling distribution of X̄1 − X̄2 , we have µX̄1 −X̄2 = µ1 − µ2 = 6.5 − 6.0 =
q 2
σ σ2
q
0.5 and σX̄1 −X̄2 = n11 + n22 = 0.81
36 + 49 = 0.189. Since X̄1 − X̄2 = 1, then
0.64
Thus, P(X̄1 − X̄2 > 1.0) = P(Z > 2.65) = 1 − P(Z < 2.65) = 1 − 0.9960 = 0.0040
Remark
When sampling with replacement, the mean of the sample mean µX̄ = µ and
the variance of the sample mean σ2X̄ = σn . When sampling without replacement
2
H.W. Kayondo c 25
the mean of the sample mean µX̄ = µ and the standard deviation of the sample
q
mean σX̄ = √σn N−n
N−1 , where N is the population size and n is the sample size.
q
We define the factor N−nN−1 as the finite population correction factor.
(ii) For samples of size n < 30, the values of S2 fluctuate considerably from
sample to sample, and as such, by estimating σ2 with S2 , the values of Z are
no longer normal. This short coming is resolved by using a distribution
called the t-distribution.
X̄ − µ
t=
√S
n
The probability that a random sample produces a t value falling between any
two specific values is equal to the area under the curve of the t-distribution
between the two ordinates corresponding to the spesified values.
26 Engineering Mathematics IV– EMT 2201
α α
t1−α = −tα 0
tα
Example 7 For example, the t value with 10 degrees of freedom leaving an area of
0.025 to the right if t0.025 = 2.228. Since the t distribution is symmetric about a mean of
zero, we have t1−α = −tα ; that is, the t value leaving an area of 1 − α to the right and
therefore an area of α to the left is equal to the negative t value that leaves an area of α to
the right tail of the distribution, see figure above. For a t distribution with 10 degrees
of freedom we have t0.975 = −t0.025 = −2.228. This means that the t value of a random
sample of size 11 selected from a normal population, will fall between -2.228 and 2.228,
with probability 0.95.
Example 8 Find k such that P(k < T < −1.761) = 0.045, for a random sample of size
15 selected from a normal distribution.
Solution
Note from the tables that t0.05 = 1.761 when ν = 14. Therefore, −t0.05 = −1.761.
Since k is to the left of −t0.05 = −1.761, let k = −tα . Then, 0.045 = 0.05 − α and
α = 0.005. Therefore, with ν = 14 and from the tables, k = −t−0.005 = −2.977.
Therefore, P(−2.977 < T < −1.761) = 0.045.
Example 9 A manufacturer of batteries claims that they will last on average 500 hours.
To maintain this average, he tests 25 batteries 1
each month. If the computed t value falls
between −t0.05 and t0.05 , he is satisfied with his claim. What conclusion should he draw
from a sample that has a mean x̄ = 518 and a standard deviation s = 40 hours if the
distribution of the battery times is approximately normal.
Solution
From the tables, t0.05 = 1.711 for ν = 24. Therefore, the manufacturer is
satisfied with his claim if a sample of 25 bulbs yields a t value between −1.711
and 1.711. If µ = 500, then
518 − 500
t= √ = 2.25,
40/ 25
a value above 1.711. Now, P(T > 2.25) = 0.02. If µ > 500, the value of t computed
from the sample would be more reasonable. Therefore, the manufacturer is
likely to conclude that his batteries are a better product than he thought.
The probability that a random sample produces a χ2 value greater than some
specified value is equal to the area under the curve to the right of this value.
Thus, χ2α represent the χ2 value above which we find an area α as shown below.
α
-
0 χ2α χ2
Example 10 The value of χ20.05 for ν = 7 is 14.067 which is an area of 0.05 to the right
of the χ2 value. Due to lack of symmetry, values for χ21−α are provided in most cases.
For instance, χ20.95 = 2.167
Exercise
1. Random samples of size 2 are drawn, with replacement from the finite
population consisting of 2, 4 and 6.
(a) Assuming that all possible samples are equally likely to occur, con-
struct the sampling distribution of X̄.
(b) Draw a probability histogram for the sampling distribution of X̄.
σ2
(c) Verify that µX̄ = µ and σ2X̄ = n.
(a) Assuming that all possible samples are equally likely to occur, con-
struct the sampling distribution of X̄.
√ p
(b) Verify that µX̄ = µ and σX̄ = (σ/ n) (N − n)/(N − 1).
4. If the standard error of the mean for the sampling distribution of random
samples of size 64 from a large or infinite population is 1.8, how large
must the size of the sample become if the standard error is reduced to 0.8?
6. If all possible samples of size 25 are drawn from a normal population with
mean equal to 60 and a standard deviation of 5.5, what is the probability
that a sample mean X̄ will fall in the interval from µX̄ − 2.5σX̄ to µX̄ + 1.5σX̄ ?
7. If all possible samples of size 5 are drawn from a population with out
replacement. The mean and standard deviation for the population equal to
20 and 2.5, respectively. Given that N = 50 for the population, what is the
probability that a sample mean X̄ will fall in the interval from µX̄ + 0.5σX̄
to µX̄ + 3.5σX̄ ?
9. The marks of 500 students in EPS examination done last year was approxi-
mately normally distributed with a mean of 72% and a standard deviation
of 8%. If 80 random samples of size 16 are drawn from this population
and the means recorded, determine:
(a) the mean and standard error of the sampling distribution of X̄.
(b) the number of sample means that fall between 68% and 76%.
(c) the probability that the sample means fall below 62%.
x 0 1 2 3 4 5
P(X=x) 0.15 0.1 0.3 0.2 0.1 0.15
11. Let X̄1 represent the mean of a sample of size n1 = 2, selected with
replacement from a finite population −1, 0, 1 and 2. Similarly, let X̄2
represent the mean of a sample of size n2 = 2, selected with replacement
from a finite population −1 and 2.
(a) Assuming that all the 64 possible differences x̄1 − x̄2 are equally likely
to occur, construct the sampling distribution of X̄1 − X̄2 .
(b) Construct a probability histogram for the sampling distribution of
X̄1 − X̄2 .
(c) Verify that µX̄1 −X̄2 = µ1 − µ2 and σ2X̄ = σ21 /n1 + σ22 /n2 .
1 −X̄2
13. A market researcher studying the length of time shoppers spend in two
shopping malls observes a sample of 64 shoppers in each mall. The mean
time spent by shoppers in mall A is 50 minutes, while the mean time for
sample of shoppers in mall B is 46 minutes. What is the probability of
observing mean sample difference (X̄A − X̄B ) this large or larger if there is
no difference in the true mean time spent by shoppers in the two malls
and the standard deviation is 13 minutes for both populations?
14. An EPS test was given to two groups of students, i.e group 1 and group
2. The scores for group 1 were normally distributed with mean and
variance of 60 and 100, respectively. Scores for group 2 were also normally
distributed with a mean of 50 and a variance of 121. A random sample of
10 students is selected from group 1 and an independent random sample of
size 11 is selected from group 2. What is the probability that the difference
between sample means (x̄1 − x̄2 ) is:
15. The mean score for an aptitude test done by students intending to do laws
at a certain university was 560 with a standard deviation of 55. What is
the probability that two groups of students selected at random, consisting
of 40 and 55 students, respectively, will differ in their mean scores by:
19. Given a random sample of size 16 from a normal distribution, find k such
that:
20. A researcher claims that the average age for first year students offering civil
engineering in a certain university is 19.0 years. To verify this claim, he
samples 20 students and computes both the sample mean and the sample
standard deviation and finally computes the t statistic. He concludes
that the claim is true if the computed statistic lies between −t0.05 and t0.05 .
Given that the sample for 20 first year students consists of the following:
18.0, 18.4, 20.0, 19.2, 21.0 18.3, 18.9, 19.4, 18.0, 18.5, 20.2, 19.2, 18.0, 19.0,
21.0, 19.5, 18.0, 20.4, 17.9, 22.0.
What conclusion can this researcher make?
32 Engineering Mathematics IV– EMT 2201
H.W. Kayondo c 33
α
Table A.4 Critical Values of the t-Distribution 0 tα
α
v 0.40 0.30 0.20 0.15 0.10 0.05 0.025
1 0.325 0.727 1.376 1.963 3.078 6.314 12.706
2 0.289 0.617 1.061 1.386 1.886 2.920 4.303
3 0.277 0.584 0.978 1.250 1.638 2.353 3.182
4 0.271 0.569 0.941 1.190 1.533 2.132 2.776
5 0.267 0.559 0.920 1.156 1.476 2.015 2.571
6 0.265 0.553 0.906 1.134 1.440 1.943 2.447
7 0.263 0.549 0.896 1.119 1.415 1.895 2.365
8 0.262 0.546 0.889 1.108 1.397 1.860 2.306
9 0.261 0.543 0.883 1.100 1.383 1.833 2.262
10 0.260 0.542 0.879 1.093 1.372 1.812 2.228
11 0.260 0.540 0.876 1.088 1.363 1.796 2.201
12 0.259 0.539 0.873 1.083 1.356 1.782 2.179
13 0.259 0.538 0.870 1.079 1.350 1.771 2.160
14 0.258 0.537 0.868 1.076 1.345 1.761 2.145
15 0.258 0.536 0.866 1.074 1.341 1.753 2.131
16 0.258 0.535 0.865 1.071 1.337 1.746 2.120
17 0.257 0.534 0.863 1.069 1.333 1.740 2.110
18 0.257 0.534 0.862 1.067 1.330 1.734 2.101
19 0.257 0.533 0.861 1.066 1.328 1.729 2.093
20 0.257 0.533 0.860 1.064 1.325 1.725 2.086
21 0.257 0.532 0.859 1.063 1.323 1.721 2.080
22 0.256 0.532 0.858 1.061 1.321 1.717 2.074
23 0.256 0.532 0.858 1.060 1.319 1.714 2.069
24 0.256 0.531 0.857 1.059 1.318 1.711 2.064
25 0.256 0.531 0.856 1.058 1.316 1.708 2.060
26 0.256 0.531 0.856 1.058 1.315 1.706 2.056
27 0.256 0.531 0.855 1.057 1.314 1.703 2.052
28 0.256 0.530 0.855 1.056 1.313 1.701 2.048
29 0.256 0.530 0.854 1.055 1.311 1.699 2.045
30 0.256 0.530 0.854 1.055 1.310 1.697 2.042
40 0.255 0.529 0.851 1.050 1.303 1.684 2.021
60 0.254 0.527 0.848 1.045 1.296 1.671 2.000
120 0.254 0.526 0.845 1.041 1.289 1.658 1.980
∞ 0.253 0.524 0.842 1.036 1.282 1.645 1.960
34 Engineering Mathematics IV– EMT 2201
χα2
Table A.5 Critical Values of the Chi-Squared Distribution 0
α
v 0.995 0.99 0.98 0.975 0.95 0.90 0.80 0.75 0.70 0.50
1 0.04 393 0.03 157 0.03 628 0.03 982 0.00393 0.0158 0.0642 0.102 0.148 0.455
2 0.0100 0.0201 0.0404 0.0506 0.103 0.211 0.446 0.575 0.713 1.386
3 0.0717 0.115 0.185 0.216 0.352 0.584 1.005 1.213 1.424 2.366
4 0.207 0.297 0.429 0.484 0.711 1.064 1.649 1.923 2.195 3.357
5 0.412 0.554 0.752 0.831 1.145 1.610 2.343 2.675 3.000 4.351
6 0.676 0.872 1.134 1.237 1.635 2.204 3.070 3.455 3.828 5.348
7 0.989 1.239 1.564 1.690 2.167 2.833 3.822 4.255 4.671 6.346
8 1.344 1.647 2.032 2.180 2.733 3.490 4.594 5.071 5.527 7.344
9 1.735 2.088 2.532 2.700 3.325 4.168 5.380 5.899 6.393 8.343
10 2.156 2.558 3.059 3.247 3.940 4.865 6.179 6.737 7.267 9.342
11 2.603 3.053 3.609 3.816 4.575 5.578 6.989 7.584 8.148 10.341
12 3.074 3.571 4.178 4.404 5.226 6.304 7.807 8.438 9.034 11.340
13 3.565 4.107 4.765 5.009 5.892 7.041 8.634 9.299 9.926 12.340
14 4.075 4.660 5.368 5.629 6.571 7.790 9.467 10.165 10.821 13.339
15 4.601 5.229 5.985 6.262 7.261 8.547 10.307 11.037 11.721 14.339
16 5.142 5.812 6.614 6.908 7.962 9.312 11.152 11.912 12.624 15.338
17 5.697 6.408 7.255 7.564 8.672 10.085 12.002 12.792 13.531 16.338
18 6.265 7.015 7.906 8.231 9.390 10.865 12.857 13.675 14.440 17.338
19 6.844 7.633 8.567 8.907 10.117 11.651 13.716 14.562 15.352 18.338
20 7.434 8.260 9.237 9.591 10.851 12.443 14.578 15.452 16.266 19.337
21 8.034 8.897 9.915 10.283 11.591 13.240 15.445 16.344 17.182 20.337
22 8.643 9.542 10.600 10.982 12.338 14.041 16.314 17.240 18.101 21.337
23 9.260 10.196 11.293 11.689 13.091 14.848 17.187 18.137 19.021 22.337
24 9.886 10.856 11.992 12.401 13.848 15.659 18.062 19.037 19.943 23.337
25 10.520 11.524 12.697 13.120 14.611 16.473 18.940 19.939 20.867 24.337
26 11.160 12.198 13.409 13.844 15.379 17.292 19.820 20.843 21.792 25.336
27 11.808 12.878 14.125 14.573 16.151 18.114 20.703 21.749 22.719 26.336
28 12.461 13.565 14.847 15.308 16.928 18.939 21.588 22.657 23.647 27.336
29 13.121 14.256 15.574 16.047 17.708 19.768 22.475 23.567 24.577 28.336
30 13.787 14.953 16.306 16.791 18.493 20.599 23.364 24.478 25.508 29.336
40 20.707 22.164 23.838 24.433 26.509 29.051 32.345 33.66 34.872 39.335
50 27.991 29.707 31.664 32.357 34.764 37.689 41.449 42.942 44.313 49.335
60 35.534 37.485 39.699 40.482 43.188 46.459 50.641 52.294 53.809 59.335
36 Engineering Mathematics IV– EMT 2201