Why Do We Need Statistics? - P Values - T-Tests - Anova - Correlation33
Why Do We Need Statistics? - P Values - T-Tests - Anova - Correlation33
Why Do We Need Statistics? - P Values - T-Tests - Anova - Correlation33
• H4: A > B There is a positive relationship between A and B. Here, the >
suggests that the more B is involved, the better A. Alternate
2 types of statistics
• Descriptive Stats
– e.g., mean and standard deviation (S.D)
• Inferential statistics
– t-tests, ANOVAs and regression
Issues when making inferences
• So how do we know whether the effect
observed in our sample was genuine?
– We don’t
Time
x1 x 2
t
s x1 x2
2 2
s1 s2
Group 1 Group 2
s x1 x2
n1 n2
How do we apply this to fMRI
data analysis?
Time
Degrees of freedom
• = number of unconstrained data points
• Which in this case = number of data points
– 1.
• Factorial ANOVAs
– More than 1 factor
Acknowledgements
X X
– Covariance formula
( x x)( y
i i y)
cov( x, y ) i 1
n
n
– cf. variance formula i
( x x ) 2
S x2 i 1
n
but…
• the absolute value of cov(x,y) is also a function of the
standard deviations of x and y.
Describe correlation: Pearson
correlation coefficient (r)
• Equation cov( x, y)
rxy s = st dev of sample
sx s y
– r = -1 (max. negative correlation); r = 0 (no constant
relationship); r = 1 (max. positive correlation)
• Limitations:
5
0
0 1 2 3 4 5 6
= ŷ, predicted
= y i , observed
ε = residual
• F-statistic:
complicated
rearranging
sŷ2 r2 (n - 2)2
F(df ,df ) = =......=
ŷ er
ser2 1 – r2
• And it follows that:
r (n - 2) So all we need to
t(n-2) = know are r and n !
√1 – r2
Summary
• Correlation
• Regression
• Relevance to SPM
General Linear Model
• Linear regression is actually a form of the
General Linear Model where the
parameters are b, the slope of the line,
and a, the intercept.
y = bx + a +ε
• A General Linear Model is just any model
that describes the data in terms of a
straight line
One voxel: The GLM
Our aim: Solve equation for β – tells us how much BOLD signal is explained by X
b3
b4
b5
= b6 +
b7
b8
b9
Y = X × b + e
Multiple regression
• Multiple regression is used to determine the effect of a
number of independent variables, x1, x2, x3 etc., on a
single dependent variable, y
• The different x variables are combined in a linear way
and each has its own regression coefficient:
Thanks!