MAST20005 Statistics Assignment 1
MAST20005 Statistics Assignment 1
MAST20005 Statistics Assignment 1
Question 1
1
The first and second moments of a Gamma distribution X = Γ(α, θ) are given by the mgf MX (t) = (1−θt)α :
(1) d 1
MX (0) = = αθ
dt (1 − θt)α
t=0
(2) d2 1
MX (0) = 2 = α(α + 1)θ2
α
dt (1 − θt)
t=0
αθ = 5.4
α(α + 1)θ2 = 32.63733
Which yields the solutions:
5.4
⇒ θ = α
α2 +α 32.63733
⇒ α2 = 5.42
1
⇒ 1+ α = 1.11925
1
⇒ α = 0.11925
⇒ α = 8.38574
⇒ θ = 0.64395
1
Question 2
(a)
1
Pn
Consider the estimator θ̂ = T̄ = n i=1 Ti . Now, since:
n
1 X
E[T̄ ] = E n Ti
i=1
1
= n E[T 1 + T2 + ...Tn ]
1
= n (E[T1 ] + E[T2 ] + ... + E[Tn ]) (since independent)
1
= n (nE[T ]) (since identically distributed)
= E[T ]
=θ (since T exponentially distributed with scale parameter θ)
(b)
The variance of θ̂ is:
V ar(θ̂) = V ar(T̄ )
= V ar( n1 (T1 + T2 + ... + Tn ))
1
= n2 (V ar(T1 ) + V ar(T2 ) + ... + V ar(Tn )) (since independent)
1
= n2 nV ar(T ) (since identically distributed)
1
= n V ar(T )
2
θ
= (since T exponentially distributed with scale parameter θ)
n
(c)
Let Z = min(T1 , T2 , ...Tn ), where Z ∼ Exp( nθ ) as stated in the problem.
Let θ̃ be the estimator given by:
θ̃ = nZ
Now, given that:
E[nZ] = nE[Z]
= n nθ (since Z exponentially distributed with parameter θ
n)
=θ
V ar(nZ) = n2 V ar(Z)
= n2 ( nθ )2 (since Z exponentially distributed with scale parameter θ
n)
= θ2
(d)
The estimator θ̂ has a variance of θ2 /n while the estimator θ̃ has a variance of θ2 . Hence, as n gets large, the variance
of θ̂ will decrease while the variance of θ̃ will remain unchanged.
∴ Since lower variance is a desirable property of an estimator, the θ̂ estimator should be preferred.
2
Question 3
(a)
Summary statistics:
Assignment 1 pic.png
The boxplot is centered around 1382 (mean) or 840 (median). It is right skewed with a single outlier at 6084.
(b)
The MLE for a Gamma distribution yields parameters X ∼ Γ(α = 0.897317, β = 0.000649), for which the pdf would
be:
βα α−1 −βx
fX (x) = Γ(α) x e
0.0006490.897317 0.897317−1 −0.000649x
= Γ(0.897317) x e
−0.000649x
0.00128784e
= 0≤x≤∞
x0.102683
The MLE for a Log-normal distribution yields parameters X ∼ lnN (µ = 6.579638, σ 2 = (1.178048)2 ), for which the
pdf would be:
1 (ln(x)−µ)2
fX (x) = √ e− 2σ 2
xσ 2π
(ln(x)−6.579638)2
1 − 2(1.178048)2
= √ e 0≤x≤∞
1.178048x 2π
The MLE for a Weibull distribution yields parameters X ∼ W (k = 0.894421, λ = 1301.860056), for which the pdf
would be:
k
fX (x) = ( λk )( λx )k−1 e−(x/λ)
0.894421
0.894421
= ( 1301.860056 x
)( 1301.860056 )0.894421−1 e−(x/1301.860056)
0.894421
0.00146491e−0.001638x
=
x0.105579
3
(c)
The log-likelihood values for the distributions above are:
Gamma −82.3
Log-normal −81.6
Weibull −82.2
The Log-normal model has the highest log-likelihood value, and hence gives the best fit.
Assignment 1 q3 gamma.png
Assignment 1 q3 lognormal.png
Assignment 1 q3 weibull.png
4
(d)
Histogram with pdf of lognormal model superimposed:
Assignment 1 q3 hist.png
5
Question 4
The general form for the lower endpoint of a 90% one sided confidence interval is derived from:
X̄ − µ
Pr √ ≤ t0.1 (n − 1) = 0.9
σ/ n
⇔ P r X̄ − t0.1 (n − 1) √σn ≤ µ = 0.9
Hence the lower end point of a 90% one-sided confidence interval for µ is given by:
l = x̄ − t0.1 (n − 1) · √s
n
0.61
= 35.4 − t0.1 (39) · √40
0.61
= 35.4 − 1.3036 · √ 40
= 35.2743
NOTE: I am assuming t0.1 (39) = 1.3036 as consistent the lecture notes, although when calculated in R, qt(0.1, df =
39) = −1.3036.
6
Question 5
(a)
The log-likelihood function is as follows (where log denotes the natural logarithm):
l(µ, λ) = logL(µ, λ)
n
Y
λ 1/2 −λ(xi − µ)2
= log ( 2πx 3) exp( )
i=1
i 2µ2 xi
n
X h
λ 1/2 −λ(xi − µ)2 i
= log ( 2πx3) exp( )
i=1
i 2µ2 xi
n h
X
1 λ λ(xi − µ)2 i
= 2 log( 2πx3i ) −
i=1
2µ2 xi
n h
X λ(xi − µ)2 i
= 1
2 log(λ) − 12 log(2πx3i ) −
i=1
2µ2 xi
n h
X x2i 2xi µ µ2 i
= 1
2 log(λ) − 12 log(2πx3i ) − λ( − + )
i=1
2µ2 xi 2µ2 xi 2µ2 xi
n h
X λxi λ λ i
= 1
2 log(λ) − 12 log(2πx3i ) − 2
+ −
i=1
2µ µ 2xi
n h
∂l(µ,λ)
X λxi λi
∂µ = −
i=1
µ3 µ2
n
λ X nλ
= xi − 2
µ3 i=1 µ
∂l(µ,λ)
And when ∂µ = 0:
n
λ X nλ
⇒ xi = 2
µ̂3 i=1 µ̂
Pn
xi
⇒ i=1 = µ̂
n
⇒ µ̂ = x̄
n h
∂l(µ,λ)
X 1 xi 1 1 i
∂λ = − 2+ −
i=1
2λ 2µ µ 2xi
n
n Xh xi 1 1 i
= + − 2+ −
2λ i=1 2µ µ 2xi
P Pn
And when equal to zero (letting µ = x̄ as per result above, noting that means i=1 throughout):
7
n X h xi 1 1 i
⇒ = − +
2λ̂ 2µ2 µ 2xi
n X h xi 2 1i
⇒ = 2
− +
λ̂ µ µ xi
n 1 X 2n X 1
⇒ = 2 xi − +
λ̂ µ µ xi
P P 1
1 1 xi 2 xi
⇒ = 2 − +
λ̂ µ n µ n
P 1
1 1 2 xi
⇒ = 2 x̄ − +
λ̂ µ µ n
P 1
1 1 2 xi
⇒ = 2 x̄ − + (since µ = x̄)
λ̂ x̄ x̄ n
P 1
1 1 2 xi
⇒ = − +
λ̂ x̄ x̄ n
P 1
1 xi 1
⇒ = −
λ̂ n x̄
P 1
1 [ xi ] − nx̄
⇒ =
λ̂ n
P 1
1 [ xi − x̄1 ]
⇒ =
λ̂ n
n
⇒ λ̂ = P 1
[ xi − x̄1 ]
n
Hence the MLE of λ is P 1 1
[ x − x̄ ]
i
(b)
Since nλ/λ̂ ≈ χ2 (n − 1), a 100(1 − α)% confidence interval can be constructed as follows:
(c)
The MLE for λ is:
n
λ̂ = P 1
[ xi − x̄1 ]
32
=P 1 1
[ xi − 21.26562 ]
= 7.234316
8
[χ2(α/2) (n − 1) · nλ̂ , χ2(1−α/2) (n − 1) · nλ̂ ]
= [χ20.025 (31) · 7.234316
32 , χ20.975 (31) · 7.234316
32 ]
7.234316 7.234316
= [17.53874 · 32 , 48.23189 · 32 ]
= [3.965025, 10.903898]
Question 6
(a)
Since T1 and T2 are unbiased estimators of θ:
(b)
Let α∗ be the value of α which minimizes the variance of T3 , that is:
α∗ = min V ar(T3 )
α
d
This can be found by solving dα V ar(T3 ) = 0 for α as follows:
d d h i
V ar(T3 ) = V ar(αT1 + (1 − α)T2 )
dα dα
d h 2 i
= α V ar(T1 ) + (1 − α)2 V ar(T2 ) (since T1 , T2 are independent)
dα
= 2αV ar(T1 ) − 2(1 − α)V ar(T2 )
Note that the second derivative test results in a necessarily positive value indicating that this extremum is a minimum.
V ar(T2 )
Hence α∗ = V ar(T1 )+V ar(T2 ) minimizes V ar(T3 ).
If V ar(T1 ) >> V ar(T2 ), then α∗ would be close to 0. The effect of this is that T3 would be the combination of
a large proportion of T2 (which has less variance), and and small proportion of T1 (which has greater variance).
9
(c)
In the case that T1 and T2 are dependent, the calculation can be done with Cov(T1 , T2 ) factored in:
d d h i
V ar(T3 ) = V ar(αT1 + (1 − α)T2 )
dα dα
d h 2 i
= α V ar(T1 ) + (1 − α)2 V ar(T2 ) + 2α(1 − α)Cov(T1 , T2 )
dα
d h 2 i
= α V ar(T1 ) + (1 − α)2 V ar(T2 ) + 2αCov(T1 , T2 ) − 2α2 Cov(T1 , T2 )
dα
= 2αV ar(T1 ) − 2(1 − α)V ar(T2 ) + 2Cov(T1 , T2 ) − 4αCov(T1 , T2 )
= 2αV ar(T1 ) − 2V ar(T2 ) + 2αV ar(T2 ) + 2Cov(T1 , T2 ) − 4αCov(T1 , T2 )
Note that the second derivative test results in a necessarily positive value indicating that this extremum is a minimum.
V ar(T2 )−Cov(T1 ,T2 )
Hence α∗ = (V ar(T1 )+V ar(T2 )−2Cov(T1 ,T2 )) minimizes V ar(T3 ).
The consequence of V ar(T1 ) >> V ar(T2 ) is similar as in (b), though complicated further depending on the cor-
relation between T1 and T2 .
Question 7
Not applicable
10