0% found this document useful (0 votes)
15 views12 pages

Estimation of Parameters

The document discusses the concepts of unbiased and consistent estimators in statistics, defining unbiasedness as when the expected value of an estimator equals the parameter it estimates, and consistency as the property that the estimator converges to the parameter as sample size increases. It provides a theorem linking unbiasedness and consistency, along with proofs and examples, including the sample mean as an unbiased and consistent estimator of the population mean and the sample variance as a biased estimator of the population variance. Additionally, it explores specific cases and exercises demonstrating these concepts using statistical distributions.

Uploaded by

boranana777
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
15 views12 pages

Estimation of Parameters

The document discusses the concepts of unbiased and consistent estimators in statistics, defining unbiasedness as when the expected value of an estimator equals the parameter it estimates, and consistency as the property that the estimator converges to the parameter as sample size increases. It provides a theorem linking unbiasedness and consistency, along with proofs and examples, including the sample mean as an unbiased and consistent estimator of the population mean and the sample variance as a biased estimator of the population variance. Additionally, it explores specific cases and exercises demonstrating these concepts using statistical distributions.

Uploaded by

boranana777
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

Estimation of Parameters

Unbiased estimator
Let 𝜃̂ be an estimate for the unknown parameter 𝜃 associated with the distribution of random
variable 𝑋. Then 𝜃̂ is an unbiased estimator for 𝜃 if 𝐸(𝜃̂) = 𝜃 ∀ 𝜃.

Consistent estimate
Let 𝜃̂ be an estimate of the parameter 𝜃. We say that 𝜃̂ is a consistent estimate of 𝜃 if
lim 𝑃{|𝜃̂ − θ| > 𝜖} = 0 ∀ 𝜖 > 0 𝑜𝑟 lim 𝑃{|𝜃̂ − θ| ≤ 𝜖} = 1 .
𝑛→∞ 𝑛→∞

Note: The above definition indicates that as the sample size increases, the estimate becomes
better.

Unbiasedness and consistency of estimate can be found using the following theorem:

Theorem
Let 𝜃̂ be an estimate of the parameter 𝜃 based on a sample size 𝑛. If 𝐸(𝜃̂) = 𝜃 and
lim 𝑉(𝜃̂) = 0 then 𝜃̂ is a consistent estimate of 𝜃.
𝑛→∞
Proof:
We shall prove by using Chebyshev’s inequality.
1 2 1 2
∴ lim 𝑃{|𝜃̂ − θ| > 𝜖} ≤ 2 𝐸(𝜃̂ − θ) = 2 𝐸{(𝜃̂ − E(𝜃̂)) + (E(𝜃̂) − θ)}
𝑛→∞ 𝜖 𝜖
(Add and subtract E(𝜃̂))
1
= {𝐸[𝜃̂ − E(𝜃̂)]2 + 2𝐸{[𝜃̂ − E(𝜃̂) ](E(𝜃̂) − θ)} + E(E(𝜃̂) − θ)2 }
𝜖2
1 2
= 2 {𝑉(𝜃̂) + 2[𝐸(𝜃̂) − E(𝜃̂)(E(𝜃̂) − θ)] + (E(𝜃̂) − θ) }
𝜖
1 2
= 2 {𝑉(𝜃̂) + (E(𝜃̂) − θ) } → 0 𝑎𝑠 𝑛 → ∞ using given condition.
𝜖
∴ 𝜃̂ is a consistent estimate of 𝜃.
Exercise
1. Show that sample mean is an unbiased and consistent estimate of population mean.
Solution:
Let 𝑋1 , 𝑋2 , … 𝑋𝑛 be samples taken from the distribution of 𝑋 having mean 𝜇
∑𝑛
𝑖=1 𝑋𝑖
∴ 𝑋̅ = is the sample mean
𝑛

To prove: 𝐸(𝑋̅) = 𝜇 and 𝑉(𝑋̅) = 0


𝑛 𝑛
1 1 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) 𝑛𝜇
̅
𝐸(𝑋) = 𝐸 ( ∑ 𝑋𝑖 ) = ∑ 𝐸(𝑋𝑖 ) = = =𝜇
𝑛 𝑛 𝑛 𝑛
𝑖=1 𝑖=1
That is, the sample mean is an unbiased estimate of the population mean
𝑛 𝑛
∑𝑛𝑖=1 𝑋𝑖 1 1 𝑛𝑉(𝑋) 𝜎 2
̅
𝑉(𝑋) = 𝑉 ( ) = 2 𝑉 (∑ 𝑋𝑖 ) = 2 ∑ 𝑉(𝑋𝑖 ) = =
𝑛 𝑛 𝑛 𝑛2 𝑛
𝑖=1 𝑖=1
2
𝜎
∴ lim 𝑉(𝑋̅) = lim
=0
𝑛→∞ 𝑛→∞ 𝑛
That is, the sample mean is a consistent estimate of the population mean.

2. Show that sample variance 𝑆 2 is not an unbiased estimate of population variance.


Solution:
Let 𝜎 2 be the variance of the distribution of 𝑋, i.e., 𝜎 2 is the population variance
Let 𝑆 2 be the sample variance
To prove: 𝐸(𝑆 2 ) ≠ 𝜎 2
∑𝑛 ̅ 2
𝑖=1(𝑋𝑖 −𝑋) ∑𝑛 2 ̅ 2 ̅
𝑖=1(𝑋𝑖 +(𝑋 ) −2𝑋 𝑋𝑖 ) [ ∑𝑛 2 ̅ 2 ̅2
𝑖=1(𝑋𝑖 ) ] +𝑛(𝑋) −2𝑛(𝑋 )
By definition, 𝑆 2 = = =
𝑛 𝑛 𝑛
∑𝑛 2
𝑖=1 𝑖 )
(𝑋
= − (𝑋̅)2
𝑛
(since ∑𝑛𝑖=1(𝑋̅)2 = 𝑛(𝑋̅)2 and 𝑛𝑋̅ = ∑𝑛𝑖=1 𝑋𝑖 )
∑𝑛𝑖=1(𝑋𝑖 2 ) ∑𝑛𝑖=1(𝑋𝑖 2 )
2)
𝐸(𝑆 = 𝐸 { ̅ 2
− (𝑋) } = 𝐸 { } − 𝐸{(𝑋̅)2 }
𝑛 𝑛
1
= 𝐸(∑𝑛𝑖=1(𝑋𝑖 2 )) − {𝐸(𝑋̅̅̅)2 }
𝑛
(since 𝐸(𝑋 2 ) = 𝑉(𝑋) + 𝜇2 , 𝐸(𝑋̅ 2 ) = 𝑉(𝑋̅) + 𝜇2 )
1
= 𝐸(𝑛𝑋 2 ) − {𝐸(𝑋̅̅̅)2 }
𝑛
= 𝐸(𝑋 2 ) − {𝐸(𝑋̅)2 } = 𝑉(𝑋) + 𝜇2 - {𝑉(𝑋̅) + 𝜇2 }
𝜎2 𝜎2
= [(𝜎 2 + 𝜇2 )] − { + 𝜇2 } (since sample variance is )
𝑛 𝑛
𝜎2
= (𝜎 2 + 𝜇2 ) − { + 𝜇2 }
𝑛
𝜎2
2
𝑛−1
=𝜎 − = 𝜎2 ( )
𝑛 𝑛
∴ 𝐸(𝑆 2 ) ≠ 𝜎 2 i.e., sample variance 𝑆 2 is not an unbiased estimate of population
variance

3. Show that if 𝑋 is a random sample of size 𝑛 with pdf


1 −𝑥
𝑓(𝑥, 𝜃) = {𝜃 𝑒 0 < 𝑥 < ∞, 0<𝜃<∞
𝜃

0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
𝜃2
then 𝑋̅ is an unbiased estimate of 𝜃 and has variance .
𝑛
Solution:
∞ ∞
1 −𝑥
𝐸(𝑋) = 𝜇 = ∫ 𝑥 𝑓(𝑥, 𝜃) 𝑑𝑥 = ∫ 𝑥 𝑒 𝜃 𝑑𝑥
−∞ 0 𝜃
𝑥 𝑥 ∞
1 𝑒− 𝜃 𝑒− 𝜃
= {𝑥 − } =𝜃
𝜃 −1 1
𝜃 𝜃2 0

∞ ∞
2)
1 −𝑥
𝐸(𝑋 = ∫ 𝑥 𝑓(𝑥, 𝜃) 𝑑𝑥 = ∫ 𝑥 2
2
𝑒 𝜃 𝑑𝑥
−∞ 0 𝜃
𝑥 𝑥 𝑥 ∞
1 𝑒 𝑒− 𝜃 𝑒 −
𝜃

𝜃
{ 𝑥2
= − 2𝑥 + 2 } = 2𝜃 2
𝜃 −1 1 −1
𝜃 𝜃 2 𝜃3 0
∴ 𝑉(𝑋) = 𝜎 2 = 2𝜃 2 − 𝜃 2 = 𝜃 2
The random variable 𝑋 has mean 𝜃 and variance 𝜃 2
𝜃2 𝜎2
Hence, 𝐸(𝑋̅) = 𝜃 , 𝑉(𝑋̅) = and lim 𝑉(𝑋̅) = lim =0
𝑛 𝑛→∞ 𝑛→∞ 𝑛

4. Let 𝑋1 , 𝑋2 , … 𝑋𝑛 be samples taken from a normal distribution with 𝜇 = 0 and variance 𝜎 2


∑ 𝑥𝑖 2
= 𝜃, 0 < 𝜃 < ∞. Show that 𝑌 = is an unbiased and consistent estimate of 𝜃.
𝑛
Solution:
𝜃 𝜃
𝑋~𝑁(0, 𝜃) ⟹ 𝑋̅~𝑁 (0, ) i.e., 𝐸(𝑋̅) = 0 and 𝑉(𝑋̅) =
𝑛 𝑛
∑ 𝑥𝑖 2
Let 𝑌 =
𝑛
To prove: 𝐸(𝑌) = 𝜃 and 𝑉(𝑌) = 0 𝑎𝑠 𝑛 → ∞
∑ 𝑥𝑖 2 1 2
𝑛 𝐸(𝑋 2 )
𝐸(𝑌) = 𝐸 ( ) = 𝐸 (∑ 𝑥𝑖 ) = = 𝐸(𝑋 2 )
𝑛 𝑛 𝑛
𝐸(𝑌) = 𝐸(𝑋 2 ) = 𝑉(𝑋) + [ 𝐸(𝑋)]2 = 𝜃 which implies 𝑌 is an unbiased estimate of 𝜃.

2 𝑋− 𝜇 2 𝑋− 𝜇 2 𝑋− 0 2 𝑋2
We know that 𝑋 ∼ 𝑁( 𝜇, 𝜎 ) , 𝑍 = ∼ 𝑁( 0, 1) and 𝑍 = ( ) =( ) = ∼
𝜎 𝜎 √𝜃 𝜃
𝜒 2 (1)
𝑋2 𝑋2
∴ 𝐸 ( ) = 1 𝑎𝑛𝑑 𝑉 ( ) = 2
𝜃 𝜃
This implies 𝐸(𝑋 2 ) = 𝜃 , 𝑉(𝑋 2 ) = 2𝜃 2
∑ 𝑥𝑖 2 1 𝑛 𝑉(𝑋 2 ) 𝑉(𝑋 2 ) 2𝜃2
Consider 𝑉(𝑌) = 𝑉 (
𝑛
) = 𝑛2 𝑉(∑ 𝑥𝑖 2 ) = 𝑛2
=
𝑛
=
𝑛
2𝜃2
∴ lim 𝑉(𝑌) = lim = 0 which implies 𝑌 is a consistent estimate of 𝜃.
𝑛→∞ 𝑛→∞ 𝑛

5. Let 𝑌1 and 𝑌2 be two independent unbiased statistics for 𝜃. The variance of 𝑌1 is twice the
variance of 𝑌2 . Find the constants 𝑘1 and 𝑘2 such that 𝑌 = 𝑘1 𝑌1 + 𝑘2 𝑌2 is an unbiased
statistic for 𝜃 with smallest possible variance for such a linear combination.
Solution:
Given that 𝐸(𝑌1 ) = 𝐸(𝑌2 ) = 𝜃, 𝑉(𝑌1 ) = 2𝑉( 𝑌2 ) = 2𝜎 2
To find: 𝑘1 𝑎𝑛𝑑 𝑘2 such that 𝐸(𝑌) = 𝐸(𝑘1 𝑌1 + 𝑘2 𝑌2 ) = 𝜃
𝑘1 𝐸(𝑌1 ) + 𝑘2 𝐸(𝑌2 ) = 𝜃
𝑘1 𝜃 + 𝑘2 𝜃 = 𝜃
𝑘1 + 𝑘2 = 1 ⇒ 𝑘2 = 1 − 𝑘1
2 2
𝑉(𝑌) = 𝑉(𝑘1 𝑌1 + 𝑘2 𝑌2 ) = 𝑘1 2 𝑉(𝑌1 ) + 𝑘2 𝑉(𝑌2 ) = 𝜎 2 (2𝑘1 2 + 𝑘2 )
2
𝑉(𝑌) = 𝜎 2 (2𝑘1 2 + 𝑘2 ) = 𝜎 2 (2𝑘1 2 + (1 − 𝑘1 )2 )
𝑑𝑉(𝑌)
𝑉(𝑌) has minima if =0
𝑑𝑘1
𝑑𝑉(𝑌) 𝑑[𝜎 2 (2𝑘1 2 +(1−𝑘1 )2 )]
= =0
𝑑𝑘1 𝑑𝑘1
1 2
⇒ 4𝑘1 − 2(1 − 𝑘1 ) = 0 ⇒ 𝑘1 = and 𝑘2 =
3 3
6. Let 𝑋1 , 𝑋2 , … , 𝑋25 ; 𝑌1 , 𝑌2 , … , 𝑌25 be two independent random samples from the
𝑋̅
distribution 𝑁(3, 16) , 𝑁(4, 9) respectively. Evaluate P ( ̅ > 1)
𝑌
Solution:
Given that 𝑋 ~𝑁(3, 16), 𝑌~𝑁(4, 9).
16 9
Then, 𝑋̅ ~𝑁 (3, ) , 𝑌̅ ~𝑁(4, )
25 25
𝑋̅
Now > 1 ⇒ 𝑋̅ > 𝑌̅ ⇒ 𝑋̅ − 𝑌̅ > 0.
𝑌̅
16 9
Since 𝑋̅ − 𝑌̅ ~ 𝑁[ 3 × 1 + 4 × (−1), 12 × + (−1)2 × ]~𝑁(−1, 1)
25 25
We have, 𝑍 = 𝑋̅ − 𝑌̅ + 1 ~𝑁(0, 1)
𝑋̅
Consider 𝑃 ( ̅ > 1) = 𝑃(𝑋̅ − 𝑌̅ > 0)
𝑌
= 𝑃(𝑋̅ − 𝑌̅ + 1 > 1)
= 𝑃(𝑍 > 1)
= 1 − 𝜙(1) = 1 − 0.841 = 0.159

Interval estimation
Let 𝑋 be a random variable with some probability distribution, depending on an unknown
parameter 𝜃. An estimate of 𝜃 given by two magnitudes within which 𝜃 can lie is called an
interval estimate of the parameter 𝜃. The process of obtaining an interval estimate for 𝜃 is called
interval estimation.
Confidence interval
Let 𝜃 be an unknown parameter to be determined by a random sample 𝑋1 , 𝑋2 , 𝑋3 , … 𝑋𝑛 of size
𝑛. The confidence interval for the parameter θ is a random interval containing the parameter
with high probability say 1 − 𝛼 where 1 − 𝛼 is called the confidence coefficient.
NOTE:
Let 𝑋1 , 𝑋2 . . 𝑋𝑛 be a random sample of size 𝑛 from a normal distribution 𝑁(𝜇, 𝜎 2 )
∑𝑛
𝑖=1 𝑋𝑖 𝜎2
1. 𝑋̅ = ~ 𝑁(𝜇, ) for 𝜇, 𝜎 2 is known
𝑛 𝑛
𝑋̅−𝜇
2. 𝑇 = 𝑆 ~ 𝑇(𝑛 − 1) for 𝜇, 𝜎 2 is unknown
( )
√𝑛−1
∑𝑛 ̅ 2
𝑖=1(𝑋𝑖 −𝑋)
3. 𝑌 = ~𝜒 2 (𝑛) for 𝜎, 𝜇 is known
𝜎2
𝑛𝑆 2
4. 𝑌 = ~𝜒 2 (𝑛 − 1) for 𝜎, 𝜇 is unknown
𝜎2

Confidence Interval for mean

Confidence interval for mean


𝝈𝟐 is known: 𝝈𝟐 is unknown:
∑𝑛 𝑋 𝜎2 𝑋̅−𝜇
Consider 𝑋̅ = 𝑖=1 𝑖 ~ 𝑁(𝜇, ) Consider 𝑇 = 𝑆 ~𝑇(𝑛 − 1) i.e., t-
𝑛 𝑛
√𝑛−1
𝑋̅ − 𝜇
∴ 𝑍 = 𝜎 ~𝑁(0,1) distribution with 𝑛 − 1 degrees of freedom
√𝑛
To find 𝑎 such that 𝑃(−𝑎 < 𝑍 < 𝑎) = 1 −
𝛼 To find 𝑎 such that 𝑃(−𝑎 < 𝑇 < 𝑎) = 1 −
𝛼
𝑋̅ − 𝜇
𝑃 (−𝑎 < 𝜎 < 𝑎) = 1 − 𝛼 𝑋̅ − 𝜇
√𝑛 𝑃 (−𝑎 < < 𝑎) = 1 − 𝛼
𝑆
𝑎𝜎 𝑎𝜎
𝑃 (𝑋̅ − < 𝜇 < 𝑋̅ + ) = 1 − 𝛼 √𝑛 − 1
√𝑛 √𝑛 𝑎𝑆 𝑎𝑆
𝑎𝜎 𝑎𝜎 𝑃 (𝑋̅ − < 𝜇 < 𝑋̅ + )
𝜇 ∈ (𝑋̅ − , 𝑋̅ + ) √𝑛 − 1 √𝑛 − 1
√ 𝑛 √𝑛 =1−𝛼
𝑎𝑆 𝑎𝑆
𝜇 ∈ (𝑋̅ − , 𝑋̅ + )
√𝑛 − 1 √𝑛 − 1
Exercise
1. Let the observed value of 𝑋̅ of size 20 from a normal distribution with 𝜇 and 𝜎 2 = 80 be
81.2. Obtain 95% confidence interval for the mean 𝜇.
Solution:
80
Let 𝑋~ 𝑁(𝜇, 80) ⟹ 𝑋̅ ~ 𝑁 (𝜇, ) = 𝑁(𝜇, 4)
20
𝑋̅−𝜇 𝑋̅ −𝜇
∴𝑍= 𝜎 = ~𝑁(0,1)
2
√𝑛

𝑃(−𝑎 < 𝑍 < 𝑎) = 0.95


1.95
2 𝜙(𝑎) − 1 = 0.95 ⇒ 𝜙(𝑎) = = 0.975 ⇒ 𝑎 = 1.96
2
𝑎𝜎 𝑎𝜎
⇒ 𝜇 ∈ (𝑋̅ − , 𝑋̅ + ) = (81.2 − 1.96 × 2, 81.2 + 1.96 × 2)
√𝑛 √𝑛
⇒ 𝜇 ∈ (77.28, 85.12)

2. Let a random sample of size 17 from 𝑁(𝜇, 𝜎 2 ) yield 𝑋̅ = 4.7 and 𝑆 2 = 5.76. Determine
90% confidence interval for 𝜇.
Solution:
Given: 𝑛 = 17, 𝑋̅ = 4.7 and 𝑆 2 = 5.76
𝑋̅−𝜇 4(𝑋̅ −𝜇)
Let 𝑇 = 𝑆 = ~𝑇(17 − 1) ~𝑇(16)
√5.76
√𝑛−1

To find 𝑎 such that 𝑃(−𝑎 < 𝑇 < 𝑎) = 0.90


2𝜙(𝑎) − 1 = 0.90 ⇒ 𝑎 = 1.75

1.75 × √5.76 1.75 × √5.76


⟹ 𝜇 ∈ (4.7 − , 4.7 + ) ⇒ 𝜇 ∈ (3.65, 5.75)
√16 √16
Confidence interval for variance

Confidence interval for variance


𝝁 is known: 𝝁 is unknown:
∑𝑛 ̅ 2
𝑖=1(𝑋𝑖 −𝑋) 𝑛𝑆 2
Let 𝑌 = ~𝜒 2 (𝑛) Let 𝑌 = ~𝜒 2 (𝑛 − 1)
𝜎2 𝜎2
To find a and b such that 𝑃(𝑎 < 𝑌 < 𝑏) = To find 𝑎 and 𝑏 such that 𝑃(𝑎 < 𝑌 < 𝑏) =
1−𝛼 1−𝛼
𝛼 𝛼 𝛼 𝛼
i.e., 𝑃(𝑌 < 𝑎) = , 𝑃(𝑌 > 𝑏) = i.e., 𝑃(𝑌 < 𝑎) = , 𝑃(𝑌 > 𝑏) =
2 2 2 2

∑𝑛𝑖=1(𝑋𝑖 − 𝑋̅)2 ∴ 𝑃 (𝑎 <


𝑛𝑆 2
< 𝑏) = 1 − 𝛼
∴ 𝑃 (𝑎 < < 𝑏) = 1 − 𝛼 𝜎2
𝜎2
2
1 𝜎2 1
1 𝜎 1 𝑃( < < )=1−𝛼
𝑃( < < )=1−𝛼 𝑏 𝑛𝑆 2 𝑎
𝑏 ∑𝑛𝑖=1(𝑋𝑖 − 𝑋̅)2 𝑎
𝑛𝑆 2 2
𝑛𝑆 2
∑𝑛𝑖=1(𝑋𝑖 −𝑋 ̅ )2 ∑𝑛𝑖=1(𝑋𝑖 ̅ )2
−𝑋 𝑃( < 𝜎 < )=1−𝛼
𝑃( < 𝜎2 < ) 𝑏 𝑎
𝑏 𝑎
2
𝑛𝑆 2 𝑛𝑆 2
=1−𝛼 ⇒ 𝜎 ∈( , )
𝑏 𝑎
∑𝑛𝑖=1(𝑋𝑖 − 𝑋̅)2 ∑𝑛𝑖=1(𝑋𝑖 − 𝑋̅)2
2
⇒ 𝜎 ∈( , )
𝑏 𝑎
∑ 𝑋𝑖
where 𝜇 = = 𝑋̅
𝑛

Exercise
1. If 8.6, 7.9, 8.3, 6.4, 8.4, 9.8, 7.2, 7.8, 7.5 are the observed values of a random sample of
size 9 from a distribution 𝑁(8, 𝜎 2 ), construct 90% confidence interval for 𝜎 2 .
Solution:
Given: 𝜇 = 8, 𝑛 = 9
∑𝑛 ̅ 2
𝑖=1(𝑋𝑖 −𝑋)
𝑌=
𝜎2
1
= {0.62 + 0.12 + 0.32 + 1.62 + 0.42 + 1.82 + 0.82 + 0.22 + 0.52 }
𝜎2
7.35
= ~𝜒 2 (9)
𝜎2

To find a and b such that 𝑃(𝑎 < 𝑌 < 𝑏) = 1 − 𝛼 = 0.90


⇒ 𝛼 = 0.10
𝛼 0.10
𝑃(𝑌 < 𝑎) = = = 0.05 ⇒ 𝑎 = 3.33 ,
2 2
𝛼 0.10
𝑃(𝑌 > 𝑏) = = = 0.05
2 2
⇒ (𝑌 < 𝑏) = 1 − 0.05 = 0.95 ⇒ 𝑏 = 16.9
using chi square table for 9 degrees of freedom.
∑𝑛 ̅ 2 ∑𝑛
𝑖=1(𝑋𝑖 −𝑋 ) (𝑋 −𝑋̅)2 7.35 7.35
∴ 𝜎2 ∈ ( , 𝑖=1 𝑖 ) =( , ) = (0.43, 2.21)
𝑏 𝑎 16.9 3.33

2. A random sample of size 15 from a normal distribution 𝑁(𝜇, 𝜎 2 ) yields 𝑋̅ = 3.2 , 𝑆 2 =


4.24 . Determine a 90% confidence interval for 𝜎 2 .
Solution:
Given: 1 − 𝛼 = 0.9 ⇒ 𝛼 = 0.1
𝑛𝑆 2
𝑌 = 2 ~𝜒 2 (15 − 1) = 𝜒 2 (14)
𝜎
𝛼 0.10
∴ 𝑃(𝑌 < 𝑎) = = = 0.05 ⇒ 𝑎 = 6.57
2 2
𝛼 0.10
𝑃(𝑌 > 𝑏) = = = 0.05
2 2
⇒ (𝑌 < 𝑏) = 1 − 0.05 = 0.95 ⇒ 𝑏 = 23.7
using chi square table for 14 degrees of freedom.
15 × 4.24 15 × 4.24
∴ 𝜎2 ∈ ( , ) = (2.68, 9.68)
23.7 6.57

3. A random sample of size 9 from a normal distribution 𝑁(𝜇, 𝜎 2 ) yields 𝑆 2 = 7.63.


Determine a 95% confidence interval for 𝜎 2 .
(Ans: 𝜎 2 ∈ (3.924, 31.5))
4. A random sample of size 15 from a normal distribution 𝑁(𝜇, 𝜎 2 ) yields 𝑋̅ = 3.2 , 𝑆 2 =
4.24 . Determine a 95% confidence interval for 𝜇.
(Ans: 𝜇 ∈ (2.02, 4.38))
5. A random sample of size 25 from a normal distribution 𝑁(𝜇, 4) yields 𝑋̅ = 78.3 , 𝑆 2 =
4.24 . Determine a 99% confidence interval for 𝜇.
(Ans: 𝜇 ∈ (77.268, 79.332))
Maximum Likelihood Estimate for 𝜽 (MLE)
Let 𝑋1 , 𝑋2 , … , 𝑋𝑛 be a random sample from the random variable 𝑋 and let 𝑥1 , 𝑥2 , … , 𝑥𝑛 be
sample values. We denote likelihood function 𝐿 as the following function
𝐿(𝑋1 , 𝑋2 , … , 𝑋𝑛 ; 𝜃) = 𝑓(𝑥1 , 𝜃)𝑓(𝑥2 , 𝜃) … 𝑓(𝑥𝑛 , 𝜃)

Maximum Likelihood Estimate (MLE) of 𝜃, say 𝜃̂ based on a random sample 𝑋1 , 𝑋2 , … 𝑋𝑛 is that


value of 𝜃 that maximizes 𝐿(𝑋1 , 𝑋2 , … , 𝑋𝑛 ; 𝜃).

Exercise
1. Let 𝑋1 , 𝑋2 , … 𝑋𝑛 denote a random sample of size 𝑛 from a distribution having pdf
𝜃 𝑥 (1 − 𝜃)1−𝑥 , 0 ≤ 𝜃 ≤ 1
𝑓(𝑥, 𝜃) = {
0, 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
Find the MLE for 𝜃.
Solution:
𝐿(𝑋1 , 𝑋2 , … , 𝑋𝑛 ; 𝜃) = 𝑓(𝑥1 , 𝜃)𝑓(𝑥2 , θ) … 𝑓(𝑥𝑛 , 𝜃)
= 𝜃 𝑥1 (1 − 𝜃)1−𝑥1 . 𝜃 𝑥2 (1 − 𝜃)1−𝑥2 . . . 𝜃 𝑥𝑛 (1 − 𝜃)1−𝑥𝑛
𝑛 𝑛
= 𝜃 ∑𝑖=1 𝑥𝑖 (1 − 𝜃)𝑛−∑𝑖=1 𝑥𝑖
Taking log on both sides and then partially differentiating with respect to 𝜃,
𝜕 (log 𝐿) 1 1
= ∑𝑛𝑖=1 𝑥𝑖 × + (𝑛 − ∑𝑛𝑖=1 𝑥𝑖 ) (− )
𝜕𝜃 𝜃 1−𝜃

𝜕 (log 𝐿)
For maximum, =0
𝜕𝜃

On simplifying, we get MLE of 𝜃 = 𝜃̂ = 𝑋̅

2. Let 𝑋1 , 𝑋2 , … 𝑋𝑛 denote a random sample of size 𝑛 from a distribution having pdf


𝜃 𝑥 𝑒 −𝜃
𝑓(𝑥, 𝜃) = { 𝑥! , 0≤𝜃≤1
0, 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
Find MLE for 𝜃.
Solution:
𝐿(𝑋1 , 𝑋2 , … , 𝑋𝑛 ; 𝜃) = 𝑓(𝑥1 , 𝜃)𝑓(𝑥2 , θ) … 𝑓(𝑥𝑛 , 𝜃)
𝜃 𝑥1 𝑒 −𝜃 𝜃 𝑥2 𝑒 −𝜃 𝜃 𝑥𝑛 𝑒 −𝜃
= × × …×
𝑥1 ! 𝑥2 ! 𝑥𝑛 !
𝑛
= 𝜃 ∑𝑖=1 𝑥𝑖 𝑒 −𝑛𝜃
𝜕 (log 𝐿)
Taking log on both sides and = 0 gives MLE, 𝜃̂ = 𝑋̅
𝜕𝜃

3. Find the MLE of normal distribution 𝑁(𝜃1 , 𝜃2 ) where −∞ < 𝜃1 < ∞, 0 < 𝜃2 < ∞.
Solution:
(𝑥−𝜇)2
2 2) 1 −
If 𝑋~𝑁(𝜇, 𝜎 ), then 𝑓(𝑥, 𝜇, 𝜎 = 𝑒 2𝜎2
√2𝜋𝜎
(𝑥−𝜃1 )2
1 −
Therefore, 𝑓(𝑥, 𝜃1 , 𝜃2 ) = 𝑒 2𝜃2
√2𝜋𝜃2

𝐿(𝑋1 , 𝑋2 , … , 𝑋𝑛 ; 𝜃) = 𝑓(𝑥1 , 𝜃)𝑓(𝑥2 , θ) … 𝑓(𝑥𝑛 , 𝜃)

1 (𝑥 −𝜃 )2 1 (𝑥 −𝜃 )2 1 (𝑥 −𝜃 )2
− 1 1 − 2 1 − 𝑛 1
= 𝑒 2𝜃2 × 𝑒 2𝜃2 × ⋯× 𝑒 2𝜃2
√2𝜋𝜃2 √2𝜋𝜃2 √2𝜋𝜃2
1 1 𝑛
− ∑ (𝑥 −𝜃 )2
= 𝑛 𝑒 2 𝑖=1 𝑖 1
2𝜃
(2𝜋𝜃2 )2
Taking logarithm on both sides
𝑛
1 2 ∑𝑛𝑖=1(𝑥𝑖 − 𝜃1 )2
log 𝐿 = log ( ) − log 𝑒
2𝜋𝜃2 2𝜃2
𝑛 1 ∑𝑛𝑖=1(𝑥𝑖 − 𝜃1 )2
= log ( )−
2 2𝜋𝜃2 2𝜃2
𝑛 1 ∑𝑛𝑖=1(𝑥𝑖2 + 𝜃12 − 2𝑥𝑖 𝜃)
= log ( )−
2 2𝜋𝜃2 2𝜃2
Differentiating partially with respect to 𝜃1 , we get
𝜕(log 𝐿) 1
=− ∑𝑛𝑖=1(2𝜃1 − 2𝑥𝑖 )
𝜕𝜃1 2𝜃2

𝜕 (log 𝐿)
For maximum, =0
𝜕𝜃1
On simplifying, we get MLE of 𝜃1 = 𝜃̂1 = 𝑋̅
Differentiating partially with respect to 𝜃1 , we get
𝑛
𝜕(log 𝐿) 𝑛 1
=− + 2 [∑(𝑥𝑖 − 𝜃1 )2 ]
𝜕𝜃2 2𝜃2 2𝜃2
𝑖=1
𝜕 (log 𝐿)
For maximum, =0
𝜕𝜃2
𝑛 2
∑ (𝑥 −𝑋̅)
On simplifying, we get MLE of 𝜃2 = 𝜃̂2 = 𝑖=1 𝑖 = 𝑠2
𝑛

You might also like