Materials SB: 1+3.3log Log (N)

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 10

MATERIALS SB

CHAPTER 1
❖ Descriptive data and Inferential Data
❖ Critical thinking
+ Conclusion from small samples
+ Conclusion from non-random samples
+ Conclusion from rare event
+ Poor survey methods
+ Post-hoc fallacy

CHAPTER 2
❖ Level of measurement
+ Nominal measurement
+ Ordinal measurement
+ Interval measurement
+ Ratio measurement
❖ Sampling method
+ Simple random sample
+ Systematic sample
+ Stratified sample Random sampling method
+ Cluster sample
+ Judgment sample
+ Convenience sample Non-random sampling method
+ Focus group

CHAPTER 3
❖ Stem and leaf
❖ Dot plot
❖ Sturges’ Rule: k ¿ 1+3.3 log log (n)

❖ Bin width
❖ Histogram and shape
CHAPTER 4: Descriptive Statistics
n
1
❖ Sample mean: x = ∑ x i
n i=1

❖ Geometric mean: G = √ x 1 x 2 … x n
n

❖ Range: R = x max −x min


xmin + x max
❖ Midrange =
2
❖ Sample standard deviation: s = √❑
σ s
❖ Coefficient of variation of population: CV = 100× μ ; CV of sample: CV = 100× x
xi −μ x −x
❖ Standardized variable of population: z i= ; of sample: z i= i
σ s
Q1+Q 2
❖ Midhinge =
2
n

❖ Sample correlation coefficient: r = ∑


i=1
(x i−x )( y i− y )

❖ Fences and Unusual Data Values
Inner fences Outer fences
Lower fence Q1 – 1.5(Q3 – Q1) Q1 – 3(Q3 – Q1)
Upper fence Q3 + 1.5(Q3 – Q1) Q3 + 3(Q3 – Q1)

CHAPTER 5: Probability
P( A) 1−P( A)
❖ Odds for A: ; Odds against A:
1−P( A) P( A)
❖ General Law of Addition: P ( A ∪ B ) =P ( A )+ P ( B )−P ( A ∩B )
P ( A ∩ B)
❖ Conditional probability: P ( B ) = P( B)

❖ Independence property: P ( A ∩ B )=P ( A ) . P(B)


P( A∨B). P(B)
❖ Bayes’ Theorem: P ( A ) = ' '
P (A∨B) . P ( B )+ P ( A∨B ). P( B )
n!
❖ Permutation: nPr =
(n−r )!
n!
❖ Combination: nCr =
r ! (n−r )!
CHAPTER 6: Discrete Probability Distributions
1
❖ Uniform PDF: P( X =x) = ; x=a ,a+ 1, … , b
b−a+1
n!
❖ Binomial PDF: P( X =x) = π x (1−π )n−x ; x=0 , 1 ,2 , … , n
x ! ( n−x ) !
λ x e− λ
❖ Poisson PDF: P( X =x) = ; x=0 , 1 ,2 , …
x!
❖ Bài toán vé số = (value win) x P(win) + (value lose) x P(lose)
C xs C n−x
N −s
❖ Hypergeometric: P( X =x) =
C nN

λ x e− λ
❖ Poisson approximation for binomial: λ=nπ ; P(X =x) = (when n ≥ 20 ; π ≤ 0.05)
x!
❖ Binomial approximation to the hypergeometric:
n s
when <0.05 → use n as sample size, π=
N N
❖ Geometric: P ( X=x )=π ( 1−π)x−1;
P ( X ≤ x )=1−(1−π ) x

CHAPTER 7: Continuous Probability Distributions


1
❖ Normal distribution: f ( x) =
σ √❑
1
❖ Standard normal distribution: f ( x) =
√❑
1
❖ Exponential Distribution PDF: f ( x) = λ e− λx ; Standard deviation =
λ
❖ Exponential Distribution CDP: P ( X ≤ x )=1−e− λx
❖ Normal Approximation to Binomial: μ=nπ ; σ =√ ❑ for nπ ≥ 10 and n(1−π) ≥ 10
❖ Normal Approximation to Poisson: μ= λ; σ =√❑ for λ ≥ 10
❖ Linear Transformation. Rule 1: μaX +b =a μ X + b (mean of a transformed variable)
❖ Linear Transformation. Rule 2: σ aX +b=a σ X (standard deviation of transformed
variable)
❖ Linear Transformation. Rule 3: μ X +Y =μ X + μY (mean of sum of tow random variables
X and Y)
❖ Linear Transformation. Rule 4: σ X + Y =√❑ (standard deviation of sum if X and Y are
independent)
❖ Covariance: σ X + Y =√❑ . If X and Y are correlated.

CHAPTER 8: Sampling Distributions and Estimation


Commonly Used Formulas in Sampling Distributions
x
❖ Sample proportion: p = n

σ
❖ Standard error of the sample mean: σ x =
√❑
σ
❖ Confidence interval for μ, known σ: x ± z α / 2
√❑
s
❖ Confidence interval for μ, unknown σ: x ± t α / 2 with d . f .=n−1
√❑
❖ Standard error of the sample proportion: σ p = √ ❑
❖ Confidence interval for π: p ± z α /2 √❑
2

❖ Sample size to estimate μ: n=( )
E

❖ Sample size to estimate π: n= ( )


z 2
E
π (1−π )

❖ Interpretation:
P - value Interpretation
P > 0,05 No evidence against H0
P < 0.05 Moderate evidence against H0
P < 0.01 Strong evidence against H0
P < 0.001 Very strong evidence against H0

CHAPTER 9: One – Sample Hypothesis Test


Commonly Used Formulas in One-Sample Hypothesis Tests

x−μ0 x−μ0
❖ Test statistic for a Mean (known σ): zcalc = = σ
σx
√❑
x−μ0
❖ Test statistic for a Mean (unknown σ): tcalc = s
√❑
p−π 0 p−π 0
❖ Test statistic for a Proportion: zcalc = =
σp √❑

❖ Test for one variance:


CHI-SQUARE DISTRIBUTION compares the sample variance with a benchmark

2 (n−1) s2 s2: sample variance


χ calc = 2 ;
σ σ2: population variance.

With n: sample size

CHAPTER 10: Two-sample Hypothesis Tests


Commonly Used Formulas in Two-Sample Hypothesis Tests
❖ Test Statistic for Zero Difference of Means:

❖ Confidence interval for μ1−μ 2: ( x 1−x 2 ) ±t α/ 2 √❑


*Note: for paired t test, n is number of pairs
n

❖ Paired t test: d = ∑ d i (mean of n differences)


i=1

❖ St. Dev of n differences: sd = √ ❑
d−μd
❖ Test statistic for paired samples: t calc = s d
√❑
❖ Degree of freedom: d . f .=n−1
❖ The ith paired difference is: Di= X 1 i – X 2 i
sd
❖ Confidence interval for μD: D ±t α /2
√❑

( p − p )−(π 1−π 2) ;
❖ Test statistic for equality of proportion: z calc = 1 2
√❑

x1 + x 2 x1 x2
p= ; p1= ; p2=
n1 + n2 n1 n2
❖ Confidence interval for the difference of two proportions, π 1−π 2:
( p1− p2 )± z α / 2. √ ❑
s 21
❖ Test statistic (two variances): F calc = , with df 1 =n1 – 1, df 2 =n2 – 1
s 22

1
❖ F-test: F R =Fdf 1 , df 2; FL =
F df 2 , df 1

CHAPTER 11: Analysis of Variance


❖ One-Factor ANOVA Table:

❖ Tukey’s test: Tukey is a two-tailed test for equality of paired means from groups
compare simultaneously
H0: μ j=μk
H1: μ j ≠ μk

Tcalc =
| y j− y k|
√❑
Tcalc > Tc,n-c -> Reject H0, Tc,n-c is a critical value. (pg. 449)
❖ Hartley’s Test:
H0: σ 21=σ 22 =…=σ 2c
H1: σ not equal
2
s max
Hcalc = 2 > Hcritical (or Hc,n/c-1) (pg. 451)
s min
CHAPTER 12: Simple Regression
Commonly Used Formulas in Simple Regression
n

❖ Sample correlation coefficient: r = ∑ (x i−x )( y i− y )


i=1

❖ Test statistic for zero correlation: tcalc = r √ ❑ with d . f .=n−2, tα/2
n

❖ Slope of fitted regression: b1 = ∑ (x i−x )( y i− y )


i=1

❖ Intercept of fitted regression: b 0= y−b1 x
n n
❖ Sum of squared residuals: SSE=∑ ( y i− ^y i ) =∑ ( y i−b0 −b1 xi )
2 2

i=1 i=1

❖ Coefficient of determination: R2 = 1 - ∑ ( y i− ^y i)2 = 1 - SSE = SSR


i=1 SST SST

❖ Standard error of the estimate: s = √ ❑ = √ ❑
s
❖ Standard error of the slope: sb = , with d . f .=n−2,
√❑ 1

b1−0
❖ T test for zero slope: tcalc = s b
1

❖ Confidence interval for true slope: b 1−t α/2 . s b ≤ β 1 ≤ b1 +t α/ 2 . s b ,


1 1

with d . f .=n−2
❖ Confidence interval for conditional mean of Y: ^y i ±t α/ 2 s. √ ❑
❖ Prediction interval for Y: ^y i ±t α / 2 s. √ ❑
❖ Excel Output:
Regression Statistics
R square SSR
R2 =
SST
Standard Error se = √ ❑

ANOVA

df SS MS F
Regression k (SSR) SSR MSR
MSR =
k MSE
Residual n-k-1 (SSE) SSE
MSE =
n−k−1
Total n-1 (SST)

Coefficient Standard Error T Stat


Intercept (b0) (sb0)
(b0)
Square feet (b1) (sb1) b1−β 1
(b1) sb
1

You might also like