Ttest

Download as pdf or txt
Download as pdf or txt
You are on page 1of 16

t-Test Statistics

Overview of Statistical Tests


Assumption: Testing for Normality
The Students t-distribution
Inference about one mean (one sample t-test)
Inference about two means (two sample t-test)
Assumption: F-test for Variance
Students t-test
- For homogeneous variances
- For heterogeneous variances
Statistical Power
1

Overview of Statistical Tests


During the design of your experiment you must
specify what statistical procedures you will use.
You require at least 3 pieces of info:
Type of Variable
Number of Variables
Number of Samples
Then refer to end-papers of Sokal and Rohlf (1995)
-REVIEW2

Assumptions
Virtually every statistic, parametric or nonparametric,
has assumptions which must be met prior
to the application of the experimental design
and subsequent statistical analysis.
We will discuss specific assumptions associated with
individual tests as they come up.
Virtually all parametric statistics have an
assumption that the data come from a population
that follows a known distribution.
Most of the tests we will evaluate in this module
require a normal distribution.
3

Assumption: Testing for Normality


Review and Commentary:
DAgostino, R.B., A. Belanger, and R.B. DAgostino.
1990. A suggestion for using powerful and informative
tests of normality. The American Statistician 44: 316-321.
(See Course Web Page for PDF version.)
Most major normality tests have corresponding R code
available in either the base stats package or affiliated
package. We will review the options as we proceed.
4

Normality
There are 5 major tests used:
Shapiro-Wilk W test
Anderson-Darling test
Martinez-Iglewicz test
Kolmogorov-Smirnov test
DAgostino Omnibus test
NB: Power of all is weak if N < 10
5

Shapiro-Wilk W test
Developed by Shapiro and Wilk (1965).
One of the most powerful overall tests.
It is the ratio of two estimates of variance (actual
calculation is cumbersome; adversely affected by ties).
Test statistic is W; roughly a measure of the straightness
of the quantile-quantile plot.
The closer W is to 1, the more normal the sample is.
Available in R and most other major stats applications.

2.6
2.4
2.2

Sample Quantiles

2.8

3.0

Normal Q-Q Plot

2.0

Example in R
(Tutorial-3)

-2

-1

Theoretical Quantiles

Anderson-Darling test
Developed by Anderson and Darling (1954).
Very popular test.
Based on EDF (empirical distribution function)
percentile statistics.
Almost as powerful as Shapiro-Wilk W test.

Martinez-Iglewicz Test
Based on the median & robust estimator of
dispersion.
Very powerful test.
Works well with small sample sizes.
Particularly useful for symmetrically skewed samples.
A value close to 1.0 indicates normality.
Strongly recommended during EDA.
Not available in R.
9

Kolmogorov-Smirnov Test
Calculates expected normal distribution and compares
it with the observed distribution.
Uses cumulative distribution functions.
Based on the max difference between two distributions.
Poor for discrimination below N = 30.
Power to detect differences is low.
Historically popular.
Available in R and most other stats applications.

10

DAgostino et al. Tests


Based on coefficients of skewness (b1) and kurtosis (b2).
If normal, b1=1 and b2=3 (tests based on this).
Provides separate tests for skew and kurt:
- Skewness test requires N 8
- Kurtosis test best if N > 20
Provides combined Omnibus test of normality.
Available in R.
11

The t-Distribution
t-distribution is similar to Z-distribution
Note similarity:

Z=

y
y
vs. t =
/ N
S / N

The functional difference is between and S.


Virtually identical when N > 30.
Much like Z, the t-distribution can be used for inferences about .
One would use the t-statistic when is not known and
S is (the general case).

12

The t-Distribution
See Appendix
Statistical Table C
Standard
Normal (Z)
t, = 12
t, = 6

13

One Sample t-test


- Assumptions -

The data must be continuous.


The data must follow the normal probability
distribution.
The sample is a simple random sample
from its population.

14

One Sample t-test


t=

y
S / N

y t 2 , df SE y y t 2 , df SE y
2

df s
df s
2
2
2
/ 2, df
1/ 2, df
15

One Sample t-test


- Example Twelve (N = 12) rats are weighed before and after
being subjected to a regimen of forced exercise.
Each weight change (g) is the weight after exercise
minus the weight before:
1.7, 0.7, -0.4, -1.8, 0.2, 0.9, -1.2, -0.9, -1.8, -1.4, -1.8, -2.0

H0: = 0
HA: 0

16

One Sample t-test


- Example > W<-c(1.7, 0.7, -0.4, -1.8, 0.2, 0.9, -1.2,
-0.9, -1.8, -1.4, -1.8, -2.0)
> summary(W)
Min. 1st Qu.
-2.000 -1.800

Median
-1.050

Mean 3rd Qu.


-0.650
0.325

Max.
1.700

> hist(W, col=red)


> shapiro.test(W)
Shapiro-Wilk normality test
data: W
W = 0.8949, p-value = 0.1364
17

One-sample t-test using R


> W<-c(1.7, 0.7, -0.4, -1.8, 0.2, 0.9, -1.2, -0.9,
+ -1.8, -1.4, -1.8, -2.0)
> W
[1] 1.7 0.7 -0.4 -1.8 0.2 0.9 -1.2 -0.9 -1.8 -1.4
-1.8 -2.0
> t.test(W, mu=0)
One Sample t-test
data: W
t = -1.7981, df = 11, p-value = 0.09964
alternative hypothesis: true mean is not equal to 0
95 percent confidence interval:
-1.4456548 0.1456548
sample estimates:
mean of x
-0.65
18

One Sample t-test


For most statistical procedures, one will want to do a
post-hoc test (particularly in the case of failing to
reject H0) of the required sample size necessary to
test the hypothesis.
For example, how large of a sample size would be
needed to reject the null hypothesis of the onesample t-test we just did?
Sample size questions and related error rates are
best explored through a power analysis.
19

> power.t.test(n=15, delta=1.0, sd=1.2523, sig.level=0.05,


type="one.sample")
One-sample t test power calculation
n
delta
sd
sig.level
power
alternative

=
=
=
=
=
=

15
1
1.2523
0.05
0.8199317
two.sided

> power.t.test(n=20, delta=1.0, sd=1.2523, sig.level=0.05,


type="one.sample")
One-sample t test power calculation
n
delta
sd
sig.level
power
alternative

=
=
=
=
=
=

20
1
1.2523
0.05
0.9230059
two.sided

> power.t.test(n=25, delta=1.0, sd=1.2523, sig.level=0.05,


type="one.sample")
One-sample t test power calculation
n
delta
sd
sig.level
power
i

=
=
=
=
=

25
1
1.2523
0.05
0.9691447
i

20

Two Sample t-test


- Assumptions The data are continuous (not discrete).
The data follow the normal probability distribution.
The variances of the two populations are equal. (If not,
the Aspin-Welch Unequal-Variance test is used.)
The two samples are independent. There is no
relationship between the individuals in one sample as
compared to the other.
Both samples are simple random samples from their
respective populations.

21

Two Sample t-test


Determination of which two-sample t-test
to use is dependent upon first testing the
variance assumption:
Two Sample t-test for
Homogeneous Variances
Two-Sample t-test for
Heterogeneous Variances

22

Variance Ratio F-test


- Variance Assumption Must explicitly test for homogeneity of variance
Ho: S12 = S22
Ha: S12 S22
Requires the use of F-test which relies
on the F-distribution.
Fcalc = S2max / S2min
Get Ftable at N-1 df for each sample
If Fcalc < Ftable then fail to reject Ho.
23

Variance Ratio F-test


- Example Suppose you had the following sample data:
Sample-A
Sample-B

Sa2 = 16.27
Sb2 = 13.98

N = 12
N=8

Fcalc = 16.27/13.98 = 1.16


Ftable = 3.603 (df = 11,7)
Decision: Fcalc < Ftable therefore fail to reject Ho.
Conclusion: the variances are homogeneous.
24

Variance Ratio F-test


- WARNING Be careful!
The F-test for variance requires that the two samples are
drawn from normal populations
(i.e., must test normality assumption first).
If the two samples are not normally distributed,
do not use Variance Ratio F-test !
Use the Modified Levene Equal-Variance Test.
25

Modified Levene Equal-Variance Test


First, redefine all of the
variates as a function of
the difference with their
respective median.

z 1j =x j Med x
z 2j=y j Med y

Then perform a twosample ANOVA to get F


for redefined values.

Stronger test of homogeneity of variance assumption.


Not currently available in R, but code easily written and
executed.
26

Two-sample t-test
- for Homogeneous Variances Begin by calculating the mean & variance for
each of your two samples.
Then determine pooled variance Sp2:

2
p

N1

N2

i =1

j =1

yi yi 2 y j yj 2
N 11 N 21

Theoretical formula

N 1 1S 21 N 2 1 S 22
N 1 N 2 2

Machine formula
27

Two-sample t-test
- for Homogeneous Variances Determine the test statistic tcalc:

t=

y1 y2
2

df =N 1 N 2 2

S p Sp

N1 N 2

Go to t-table (Appendix) at the appropriate


and df to determine ttable
28

Two-sample t-test
- Example Suppose a microscopist has just identified
two potentially different types of cells based
upon differential staining.
She separates them out in to two groups
(amber cells and blue cells). She
suspects there may be a difference in
cell wall thickness (cwt) so she wishes to
test the hypothesis:
Ho: ACcwt = BCcwt
Ha: ACcwt BCcwt
29

Two-sample t-test
- Example Parameter
Mean
SS
N

AC-type
8.57
2.39
14

BC-type
8.40
2.74
18

Notes: She counts the number of cells in


one randomly chosen field of view. SS is
the sum of squares (theor. formula), or
numerator of the variance equation.
30

Two-sample t-test
- Example -

Sp =
t calc =

2.392.74
= 0.17
1317
8.578.40
= 1.13
0.17 0.17

14
18

df = 14182=30

Ho: ACcwt = BCcwt


Ha: ACcwt BCcwt
At = 0.05/2, df = 30
ttable = 2.042
tcalc < ttable
Therefore Fail to reject Ho.
Cell wall thickness
is similar btw 2 types.
31

Two-sample t-test using R


> B<-c(8.8, 8.4,7.9,8.7,9.1,9.6)
> G<-c(9.9,9.0,11.1,9.6,8.7,10.4,9.5)
> var.test(B,G)
F test to compare two variances
data: B and G
F = 0.5063, num df = 5, denom df = 6, p-value = 0.4722
alternative hypothesis: true ratio of variances is not
equal to 1
95 percent confidence interval:
0.0845636 3.5330199
sample estimates:
ratio of variances
0.50633

32

Two-sample t-test using R


> t.test(B,G, var.equal=TRUE)
Two Sample t-test
data: B and G
t = -2.4765, df = 11, p-value = 0.03076
alternative hypothesis: true difference in
means is not equal to 0
95 percent confidence interval:
-1.8752609 -0.1104534
sample estimates:
mean of x mean of y
8.750000 9.742857

33

Two-sample t-test
- for Heterogeneous Variances -

Q. Suppose we were able to meet the


normality assumption, but failed the
homogeneity of variance test. Can we
still perform a t-test?
A. Yes, but we but must calculate an adjusted
degrees of freedom (df).

34

Two-sample t-test
- Adjusted df for Heterogeneous Variances -

df

S1 S 2

N1 N2

S1
S2
N1
N2

N 1 1 N 2 1

Performs the t-test in


exactly the same
fashion as for
homogeneous
variances; but, you
must enter the table
at a different df. Note
that this can have a
big effect on
decision.
35

Two-sample t-test using R


- Heterogeneous Variance > Captive<-c(10,11,12,11,10,11,11)
> Wild<-c(9,8,11,12,10,13,11,10,12)
> var.test(Captive,Wild)
F test to compare two variances
data: Captive and Wild
F = 0.1905, num df = 6, denom df = 8, p-value = 0.05827
alternative hypothesis: true ratio of variances is not
equal to 1
95 percent confidence interval:
0.04094769 1.06659486
sample estimates:
ratio of variances
0.1904762

36

Two-sample t-test using R


- Heterogeneous Variance > t.test(Captive, Wild)
Welch Two Sample t-test
data: Captive and Wild
t = 0.3239, df = 11.48, p-value = 0.7518
alternative hypothesis: true difference in means
is not equal to 0
95 percent confidence interval:
-1.097243 1.478196
sample estimates:
mean of x mean of y
10.85714 10.66667

37

Matched-Pair t-test
It is not uncommon in biology to conduct an
experiment whereby each observation in a
treatment sample has a matched pair in a control
sample.
Thus, we have violated the assumption of
independence and can not do a standard
t-test.
The matched-pair t-test was developed to
address this type of experimental design.
38

Matched-Pair t-test
By definition, sample sizes must be equal.
Such designs arise when:
Same obs are exposed to 2 treatments over time.
Before and after experiments (temporally related).
Side-by-side experiments (spatially related).
Many early fertilizer studies used this design. One plot
received fertilizer, an adjacent plot did not. Plots were
replicated in a field and plant yield measured.
39

Matched-Pair t-test
The approach to this type of analysis is a bit
counter intuitive.
Even though there are two samples, you will
work with only one sample composed of:
STANDARD DIFFERENCES
and df = Nab - 1

40

Matched-Pair t-test
- Assumptions The data are continuous (not discrete).
The data, i.e., the differences for the matchedpairs, follow a normal probability distribution.
The sample of pairs is a simple random sample
from its population.

41

Matched-Pair t-test
N

yd =

2
d

S =

yd
d=1

d=1

d =1

y 2d y d / N
N 1
t =

yd
S d / N
42

Matched-Pair t-test using R


- Example > New.Fert<-c(2250,2410,2260,2200,2360,2320,2240,2300,2090)
> Old.Fert<-c(1920,2020,2060,1960,1960,2140,1980,1940,1790)
> hist(Dif.Fert)
> Dif.Fert<-(New.Fert-Old.Fert)
> shapiro.test(Dif.Fert)
Shapiro-Wilk normality
data: Dif.Fert
W = 0.9436,
p-value = 0.6202

150

250

200

300

400

Normal Q-Q Plot


Sample Quantiles

2.0
1.0
0.0

Frequency

3.0

Histogram of Dif.Fert

350

Dif.Fert

-1.5

-0.5

0.5

1.5

Theoretical Quantiles

43

Matched-Pair t-test using R


- One-tailed test > t.test(New.Fert, Old.Fert,
alternative=c("greater"), mu=250, paired=TRUE)
Paired t-test
data: New.Fert and Old.Fert
t = 1.6948, df = 8, p-value = 0.06428
alternative hypothesis: true difference in means is
greater than 250
95 percent confidence interval:
245.5710
Inf
sample estimates:
mean of the differences
295.5556

44

Statistical Power
Q. What if I do a t-test on a pair of
samples and fail to reject the null
hypothesis--does this mean that there is
no significant difference?
A. Maybe yes, maybe no.
Depends upon the POWER of your test
and experiment.

45

Power
Power is the probability of rejecting the
hypothesis that the means are equal when they
are in fact not equal.
Power is one minus the probability of Type-II
error ().
The power of the test depends upon the sample
size, the magnitudes of the variances, the alpha
level, and the actual difference between the two
population means.
46

Power
Usually you would only consider the power of a test
when you failed to reject the null hypothesis.
High power is desirable (0.7 to 1.0). High power means
that there is a high probability of rejecting the null
hypothesis when the null hypothesis is false.
This is a critical measure of precision in hypothesis
testing and needs to be considered with care.
More on Power in the next lecture.
47

Stay tuned. Coming Soon:


Responding to Assumption Violations, Power,
Two-sample Nonparametric Tests
48

You might also like