Chapter 5
Chapter 5
Chapter 5
Inference
Contents
• One-Sample Hypothesis Tests • ANOVA
(mean and proportion)
• Hypothesis Development • Hypothesis
• Type I error and Type II error Development
• Z-statistic or t-statistic
• F-statistic
• One-tailed and two-tailed test
(critical value and p-value methods)
• Chi-Square Test
• Two-Sample Hypothesis Tests
• Hypothesis Development • Hypothesis
• Z-statistic, t-statistic Development
• Two-Sample test for means,
• ? Known
• Computing Expected
• ? Unknown, assumed equal Frequencies
variances
• ? Unknown, assumed unequal • Conducting the Chi-
variances
• Paired samples Square Test
Introduction
One of these propositions (called the null hypothesis, H 0) describes the existing
theory or a belief that is accepted as valid unless strong statistical evidence
exists to the contrary.
5. Applying the
decision rule to the
test statistic and
4. Collecting data and drawing a conclusion.
calculating a test
statistic
3. Determining a
decision rule on which
to base a conclusion
2. Selecting a level of
significance, which
defines the risk of
1. Identifying the drawing an incorrect
population parameter conclusion when the
of interest and assumed hypothesis is
formulating the actually true
hypotheses to test.
One-Sample Hypothesis
Tests
One-Sample Hypothesis Test
H0: population parameter ≥ constant vs H1: population parameter < constant
H0: population parameter ≤ constant vs H1: population parameter > constant
H0: population parameter = constant vs H1: population parameter ≠ constant
How to
formulate
hypothesis?
H0: 𝜇 ≥ 25
H1: 𝜇 < 25
Hypothesis Testing Results
The null hypothesis is actually true, and the test
correctly fails to reject it.
Two-tailed test:
(The sum of the area in the tails is the p-value)
p-value = p(Z < -|z0| or Z > |z0|)
= 2p(Z < |z0|)
1- 𝛼 1- 𝛼
The value 1 - 𝛽 is called the power of the test and represents the probability of
correctly rejecting the null hypothesis when it is indeed false, or P(rejecting H0|H0
is false). We would like the power of the test to be high (equivalently, we would
like the probability of a Type II error to be low) to allow us to make a valid
conclusion. The power of the test is sensitive to the sample size; small sample sizes
generally result in a low value of 1 - 𝛽. The power of the test can be increased by
taking larger samples, which enable us to detect small differences between the
sample statistics and population parameters with more accuracy.
Selecting the Test Statistic
• The next step is to collect sample data and use the data
to draw a conclusion.
• The decision to reject or fail to reject a null hypothesis is
based on computing a test statistic from the sample data.
• The test statistic used depends on the type of hypothesis
test.
• Different types of hypothesis tests use different test
statistics, and it is important to use the correct one.
• The proper test statistic often depends on certain
assumptions about the population—for example,
whether or not the standard deviation is known.
Which Test Statistic? Z-statistic
or t-statistic?
Example
Indicates that the sample mean of 21.91 is 1.05 standard errors below the
hypothesized mean of 25 minutes.
Conclusion: To Reject or Fail to
Reject H0
The conclusion to reject or fail to reject H0 is
based on comparing the value of the test
statistic to a “critical value at the chosen level
of significance, α.
• The critical value divides the sampling
distribution into two parts, a rejection region
and a non-rejection region.
• If the test statistic falls into the rejection region,
we reject the null hypothesis; otherwise, we fail
to reject it.
Rejection Region
Where 𝜋0 is the hypothesized value and ṕ is the sample proportion. Similar to the
test statistic for means, the z-test statistic shows the number of standard errors
that the sample proportion is from the hypothesized value. The sampling
distribution of this test statistic has a standard normal distribution.
Example: One Sample Test for
Proportions Refer to formula in the previous
For a lower-tailed test, the p-value would be computed by
the area to the left of the test statistic; that is:
NORM.S.DIST(z, TRUE)
• The test uses the average difference between the paired data and
the standard deviation of the differences similar to a one-sample
test.
• Excel has a Data Analysis tool, t-Test: Paired Two-Sample for
Means for conducting this type of test. In the dialog, you need to
enter only the variable ranges and hypothesized mean difference.
Example: Using the Paired Two-
Sample Test for Means
Test for Equality of Variances
• Understanding variation in business processes is
very important.
• E.g. does one location or group of employees show
higher variability than others?
• We can test for equality of variances between two
samples using a new type of test, the F-test.
• To use this test, we must assume that both samples
are drawn from normal populations. The
hypotheses we test are
Test for Equality of Variances
• To test these hypotheses, we collect samples of n1 observations
from population 1 and n2 observations from population 2.
• The test uses an F-test statistic, which is the ratio of the
variances of the two samples:
• A key marketing question is whether the proportion of males who prefer a particular brand is no
different from the proportion of females.
• If gender and brand preference are indeed independent, we would expect that about the same
proportion of the sample of female students would also prefer brand 1. If they are not
independent, then advertising should be targeted differently to males and females, whereas if
they are independent, it would not matter.
• We can test for independence by using a hypothesis test called the chi-square test for
independence. The chi-square test for independence tests the following hypotheses:
H0: the two categorical variables are independent
H1: the two categorical variables are dependent
Example: Independence and
Marketing Strategy cont.
Step 1: compute the expected frequency in each cell of the cross-tabulation if the two
variables are independent.
Example: Computing Expected
Frequency
Computing the Chi-Square
Statistic
• Step 2: Compute the chi-square statistic; the sum of
the squares of the differences between observed
frequency, fo, and expected frequency, fe, divided by
the expected frequency in each cell: