8614 PDF
8614 PDF
8614 PDF
B.ED (1.5-YEARS)
Faculty of Education
Early Childhood Education and Elementary Teacher
Education Department
ALLAMA IQBAL OPEN UNIVERSITY, ISLAMABAD
(Copyright © 2017 AIOU Islamabad)
All rights reserved. No part of this publication may be reproduced, stored in retrieval
system or transmitted in any form or by any means, electronic, mechanical, photocopying
recording, scanning or otherwise, except as permitted under AIOU copyright ACT.
Printer ..............................................
2
COURSE TEAM
Course Development:
3
ACKNOWLEDGEMENTS
This book reflects the efforts of many people. The Department of Early
Childhood Education and Elementary Teacher Education is grateful to all the
members of the course team for writing units and giving comments and feedback
to improve the materials.
It is worth mentioning that course coordinator has very skillfully mediated over
the whole process of course development, kept the course team engaged and
adhered to timeline. The composing and formatting of the scripts was very
laborious work requiring extend hours of sitting. Mr. Zahid Javed Bajwa has very
efficiently completed the task in short time.
Thanks to our Vice-Chancellor Prof. Dr. Shahid Siddiqui for his vision, support,
guidance and motivating work environment for all of us to focus on academic
excellence at university.
4
INTRODUCTION
Statistics is of vital importance in vast variety of fields. Particularly it is
invaluable for the field of research. In research and particularly in educational
research following questions cannot be answered without the use of proper
statistical techniques.
Owing to the importance, this course is included for prospective B Ed. graduates.
The very first unit of the course introduces, its characteristics, functions, its
importance and limitations and its application in educational research. Basic
overview of descriptive and inferential statistics, variables and its types, scientific
method and notation used in the subject is also given in this unit. Unit 2 explains
some basic concepts like variable, data, population sample. Unit 3 elaborate
graphical representation or exploratory data analysis techniques. Unit 4 highlights
some basic techniques of measures of dispersion like range, mean deviation,
variance and standard deviation, and measures of shape like skewness and
kurtosis. Measures of central tendency like mean, median and mode are described
in unit 5. Unit 6 deals with inferential statistics, its logic and importance in
educational research. Hypothesis testing, its logic, errors in hypothesis testing and
t-test and its types are also discussed in this unit. Correlation along with Pearson
and Spearman correlation method and regression and its types are discussed in
unit 7. Unit 8 deals with ANOVA, logic behind using ANOVA, F-distribution,
one-way ANOVA and multiple comparison procedures. Chi-square (χ2)
distribution, its uses and types are discussed in unit 9.
5
CONTENTS
Unit–1: Introduction to Statistics ................................................................. 1
1.1 Functions of Statistics ............................................................... 3
1.2 Characteristics of Statistics ....................................................... 4
1.3 Importance and Scope of Statistics .......................................... 5
1.4 Limitations of Statistics ............................................................. 6
1.5 Application of Statistics in Educational Research ..................... 6
1.6 Descriptive and Inferential Statistics ........................................ 7
1.7 Variable .................................................................................... 8
1.8 Level of Measurement .............................................................. 9
1.9 The Scientific Method ............................................................... 10
1.10 Statistical Notations .................................................................. 12
1.11 Self-Assessment Questions ....................................................... 13
6
3.5 Cumulative Frequency Polygon or Ogive................................. 38
3.6 Scatter Plot….. .......................................................................... 39
3.6.1 When to use Scatter Plot? ............................................. 39
3.7 Box Plot .................................................................................... 42
3.7.1 Types of Box Plot ......................................................... 42
3.8 Pie Chart.................................................................................... 42
3.8.1 How to Read a Pie Chart? ............................................. 43
3.8.2 When to Use the Pie Chart? .......................................... 43
3.8.3 DrawBacks of Pie-Charts .............................................. 43
3.9 Self-Assessment Questions ....................................................... 45
3.10 Activities ................................................................................... 45
3.11 Bibliography ............................................................................. 46
7
5.4.2 Demerits of Mode ......................................................... 59
5.5 Self-Assessment Questions ....................................................... 59
5.6 Activities .................................................................................... 59
5.7 Bibliography .............................................................................. 60
8
8.2.1 Interpretation of the F-Statistic ..................................... 85
8.3 One Way ANOVA (Logic and Procedure) ............................... 86
8.3.1 Assumptions Underlying the One Way ANOVA ......... 86
8.3.2 Logic Behind One Way ANOVA ................................. 87
8.3.3 Procedure for Using ANOVA ....................................... 87
8.4 Multiple Comparison Procedure ............................................... 88
8.5 Self-Assessment Questions ....................................................... 89
8.6 Activities .................................................................................. 89
8.7 Bibliography ............................................................................ 90
9
OBJECTIVES
10
UNIT-1
INTRODUCTION TO STATISTICS
Written By:
Aftab Ahmad
Reviewed By:
Dr. Rizwan Akram Rana
Introduction
Statistics is a broad subject with applications in vast variety of fields. The word
“statistics” is derived from the Latin word “Status”, which means a political state.
Statistics is a branch of knowledge that deals with facts and figures. The term statistics
refers to a set of methods and rules for organizing, summarizing, and interpreting
information. It is a way of getting information from data.
Statistics
Data Information
11
We can say that Statistics is a science of collecting, organizing, interpreting and reporting
data. It is a group of methods which are used for collecting, displaying, analyzing, and
drawing conclusions from the data.
In other words, statistics is a methodology which a researcher uses for collecting and
interpreting data and drawing conclusion from collected data (Anderson & Sclove, 1974;
Agresti & Finlay, 1997).
Objectives of Unit
After reading this unit the students will be able to:
12. demonstrate basic understanding of statistics.
13. know the characteristics of statistics.
14. explain the functions of statistics.
15. Enlist the characteristics of statistics.
16. tell the importance and limitations of statistics.
17. briefly explain the application of statistics in educational research.
18. distinguish between descriptive and inferential statistics.
19. describe variables and its types.
20. distinguish between the levels of measurement.
21. identify various statistical notations.
1.1 Functions of Statistics
Functions of Statistics are summarized under following headings.
i) To present facts in a definite form
Daily we encounter millions of pieces of information which are often vague,
indefinite and unclear. When such pieces of information undergo certain statistical
techniques and are represented in the form of tables or figures, they represent
things in a perspective which is easy to comprehend. For example, when we say
that some students out of 1000 who appeared for B. Ed examination were declared
successful. This statement is not giving as much information. But when we say that
900 students out of 1000 who appeared for B. Ed examination were declared
successful; and after using certain statistical techniques we conclude that “90% of
12
B. Ed. students were successful”; now the sentence becomes more clear and
meaningful.
13
Following are the characteristics of statistics.
i) Statistics consists of aggregate facts
The facts which can be studied in relation to time, place or frequency can be called
statistics. A single isolated and unconnected fact or figure is not statistics because
we cannot study it in relation to other facts and figures. Only aggregate of facts e.g.
academic achievement of the students, I.Q. of a group of students, weight of
students in a class, profit of a firm etc. are called statistics.
14
we will not be in a position to distinguish between necessary data and unnecessary
data or relevant data and irrelevant data.
15
We cannot apply statistical techniques to all type of phenomena. These techniques
can only be applied to the phenomena that are capable of being quantitatively
measured and numerically expressed. For example, the health, intelligence,
honesty, efficacy etc. cannot be quantitatively measured, and thus are unsuitable for
statistical study. In order to apply statistical techniques to these constructs, first we
will have to quantify them.
Statistics enters in the process right from the beginning of the research when whole plan
for the research, selection of design, population, sample, analysis tools and techniques
etc., is prepared.
16
Researchers use a variety of statistical procedures to organize and interpret data. These
procedures can be classified into two categories – Descriptive Statistics and Inferential
Statistics. The starting point for dealing with a collection of data is to organize, display, and
summarize it effectively. It is the major objective of descriptive statistics. Descriptive
Statistics, as the name implies, describes the data. Descriptive statistics consist of methods
for organizing and summarizing information. These are statistical procedures that are used
to organize, summarize, and simplify data. In these techniques raw scores are taken and
undergone some statistical techniques to obtain more manageable form. These techniques
allow the researcher to describe large amount of information or scores in a few indices such
as mean, median, standard deviation etc. When these indices are calculated for a sample,
they are called statistics; and when they are calculated from entire population, they are
called parameters (Fraenkel, Wallen, & Hyun, 2012). Descriptive statistics organizes scores
in the form of a table or a graph. It is especially useful when the researcher finds it
necessary to handle interrelationship among more than two variables.
Only summarizing and organizing data is not the whole purpose of a researcher. He often
wishes to make inferences about a population based on data he has obtained from a
sample. For this purpose, he uses inferential statistics. Inferential Statistics are techniques
that allow a researcher to study samples and then make generalizations about the
populations from which they are selected.
Population of a research study is typically too large and it is difficult for a researcher to
observe each individual. Therefore a sample is selected. By analyzing the results obtained
from a sample, a researcher hopes to make general conclusion about the population. One
problem with using sample is that a sample provides only limited information about the
population. To address this problem is the notion that the sample should be representative
of the population. That is, the general characteristics of the sample should be consistent
with the characteristics of the population.
1.7 Variable
A variable is something that is likely to vary or something that is subject to variation. We
can also say that a variable is a quantity that can assume any of a set of values. In other
words, we can say that a variable is a characteristic that varies from one person or thing
to another. It is a characteristic, number or quantity that increases or decreases over time
or takes different value in different situations; or in more precise words, it is a condition
or quality that can differ from one case to another. We often measure or count it.
A variable may also be called a data item. Examples of variables for human are height,
17
weight, age, number of siblings, business income and expenses, country of birth, capital
expenditure, marital status, eye color, gender, class grades, and vehicle type, etc.
On the other hand, variables such as time, height, and weight are not limited to a fixed set
of separate, indivisible categories. They are divisible in an infinite number of fractional
parts. Such variables are called continuous variables. For example, a researcher is
measuring the amount of time required to solve a particular mental arithmetic problem.
He can measure time in hours, minutes, seconds, or fractions of seconds
18
Variable
Categorical
Numeric
Discrete Nominal
Continuous Ordinal
.
1.8 Level of Measurement
There are two basic types of variables – quantitative and categorical. Each uses different
type of analysis and measurement, requiring the use of different type of measurement
scale. A scale of a variable gives certain structure to the variable and also defines the
meaning of the variable. There are four types of measurement scales: nominal, ordinal,
interval, and ratio.
Nominal Scale
A nominal scale is the simplest form of measurement researchers can use. The word
nominal means “having to do with names.” Measurements made on this scale involve
merely naming things. It consists of a set of categories that have different names. Such
measurements label and categorize observations but do not make quantitative distinctions
between them. For example, if we wish to know the sex of a person responding to the
questionnaire, we would measure it on nominal scale consisting of two categories (male
or female). A researcher observing the behavior of a group of infant monkeys might
categorize responses as playing, grooming, feeding, acting aggressively or showing
submissiveness. As the researcher merely gives names to each category so, this is a
nominal scale of measurement. The nominal scale consists of qualitative distinctions.
Although, a nominal scale consists of qualitative differences, yet it does not provide any
information about quantitative differences between individuals. Numerical values like 0
and 1 are merely used as code for nominal categories when entering data into computer
programs.
19
Ordinal Scale
In ordinal scale of measurement, the categories that make up the scale not only have
separate names but also are ranked in terms of magnitude. This scale consists of a set of
categories that are organized in an ordered sequence. For example, a manager of a
company is asked to rank employees in term of how well they perform their duties. The
collected data will tell us who the manager considers the best worker, the second best,
and so on. The data may reveal that the worker, who is ranked second, is viewed as doing
better work than the worker who is ranked third. However, we can get no information
about the amount that the workers differ in job performance, i.e. we cannot get the
answer of the question “How much better?” Thus, an ordinal scale provides us
information about the direction of difference between two measurements, but it does not
reveal the magnitude of the difference.
Interval Scale
An interval scale possesses all the characteristics of an ordinal scale, with additional
feature that the categories form a series of intervals that are exactly of the same size. This
additional information makes it possible to compute distances between values on an
interval scale. For example, on a ruler 1-inch interval is the same size at every location on
the ruler. Similarly 4-inch distance is exactly the same size no matter where it is
measured on the ruler. Similarly, the distance between the scores of 70 and 80 is
considered to be the same as the distance between scores of 80 and 90. For all practical
purposes these numbers can undergo arithmetic operations to be transformed into
meaningful results. Interval scale answers the question “How much better?” or “How
much is the difference?” But there is no intrinsic zero, or starting point. The zero point on
the interval scale does not indicate a total absence of what is being measured. For
example, 0o (zero degree) on the Celsius or Fahrenheit scale does not indicate no
temperature.
Ratio Scale
A ratio scale has all the characteristics of an interval scale but adds an absolute zero
point. It means on a ratio scale a value of zero indicates complete absence of the variable
being measured. Advantage of absolute zero is that a ratio of numbers on scale reflects
ratio of magnitude for the variable being measured. We can say that one measurement is
three times larger than another, or one score is only half as large as another. Thus, ratio
scale not only enables us to measure the difference between two individuals, but also to
describe the difference in terms of ratios.
Scientific method is a process for explaining the world we see. It is a process used to
validate observations while minimizing observer bias. This method is a series of steps
20
that lead to answers that accurately describe the things we observe. Its goal is to conduct
research in a fair, unbiased and repeatable manner.
Scientific method is a tool for: (a) forming and framing questions, (b) collecting
information to answer those questions, and (c) revising old and developing new questions.
The scientific method is not the only way, but the best-known way to discover how and
why the world works. It is not a formula. It is a process with a manner of sequential steps
designed to create an explainable outcome that increases our knowledge base. The
process is as follows:
i) Ask a question
Asking a question is the first step of scientific method. Good questions come from
careful observations. Our senses are a good source of observation. Sometime
certain instruments like a microscope or a telescope are also used. These
instruments extend the range of senses. During the observation many questions
come in the mind. These questions derive the scientific method.
21
1.10 Statistical Notations
Commonly used statistical notations are given in the following table.
22
34 ME Margin of error
35 DF or Df Degree of freedom
36 Q1 Lower/first quartile (25% of population are below this value)
37 Q2 Median/second quartile (50% of population are below this
value, also median of the sample)
38 Q3 Upper/third quartile (75% of population are below this value)
39 IQR Inter-quartile range (Q3 – Q1)
40 X~ Distribution of random variable X
41 N (µ,σ2) Normal distribution / Gaussian distribution
42 U (a, b) Uniform distribution (equal probability in range a, b)
43 gamma (c, λ) Gamma distribution
44 χ2 (k) Chi-square distribution
45 Bin (n, p) Binomial distribution
46 F (k1, k2) F distribution
47 Poisson (λ) Poisson distribution
1.12 Activities
1. Diagrammatically show how “data” becomes “information”.
2. Make a list of the questions that can be answered using statistics.
3. Make a list of the “functions of statistics”.
4. Think and write down any two characteristics not given in the unit.
5. Make a diagram to show the types of variables.
6. Draw a hierarchy of levels of measurement.
7. Make a list of the steps of scientific method.
23
1.13 Bibliography
Agresti, A. & Finlay, B. (1997). Statistical Methods for Social Sciences, (3rd Ed. ).
Prentice Hall.
Dietz, T., and Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley-Blackwell
Fraenkel, J. R., Wallen, N. E., & Hyun, H. H. (2012). How to Design and Evaluate in
Education. (8th Ed.) McGraw-Hill, New York
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
24
UNIT-2
Written By:
Aftab Ahmad
Reviewed By:
Dr. Rizwan Akram Rana
25
Introduction
In this unit you will study some basic concepts like variable, data, population, and
sample. Types of variable, types of data, types of population and types of sample are also
discussed. The purpose of this unit is to give an awareness of these commonly used
concepts.
Objectives
After reading this unit the students will be able to:
1. explain variable and its types.
2. explain data and its types.
3. explain population and its types.
4. explain sample and its types.
26
In research, change variables are referred to as independent variables while the outcome
variables are known as dependent variables. In cause effect relationship, there are some
unmeasured variables affecting the relationship. These are called extraneous variables.
The variables linking cause-effect relationship are called intervening variables. A brief
summary of above mentioned variables is given in the following table.
27
iii) Confounding Variable
A variable that has hidden effect on the experiment.
v) Dependent Variable
Outcome or response of an experiment. An independent variable has direct or
inverse effect upon dependent variable. In graph it is plotted on y-axis.
x) Ratio Variable
Similar to interval variable, but has meaningful zero.
28
ii) Collider Variable
A variable represented by a node on a causal graph that has paths pointing in as
well as out.
v) Dichotomous Variable
Another name for a binary variable.
x) Intervening variable
A variable that is used to explain the relationship between variables.
29
xv) Moderating variable
Changes the strength of an effect between independent and dependent variables.
For example, psychotherapy may reduce stress levels for women more than men,
so sex moderates the effect between psychotherapy and stress levels.
Data
The term “data” refers to the kind of information a researcher obtains to achieve
objectives of his research. All research processes start with collection of data, which
plays a significant role in the statistical analysis. This term is used in different contexts.
In general, it indicates facts or figures from which conclusions can be drawn. Or it is a
raw material from which information is obtained. Data are the actual pieces of
information that you collect through your study. In other words data can be defined as
collection of facts and details like text, figures, observations, symbols, or simply
description of things, event or entity gathered with a view of drawing inferences. It is a
raw fact which should be processed to get information
30
Primary data
As the name suggests, is one which is collected for the first time by the researcher
himself. Primary data is originated by the researcher for the first time for addressing his
research problem. It is also known as first hand raw data. The data can be collected using
various methods like survey, observations, physical testing, mailed questionnaire,
questionnaire filled and sent by enumerators, personal interviews, telephonic interviews,
focus groups discussion, case studies, etc.
Secondary data
Point towards the second hand information already collected and recorded by any other
person with a purpose not relating to current research problem. It is readily available
form of data and saves time and cast of the researcher. But as the data is gathered for the
purpose other than the problem under investigation, so the usefulness of the data may be
limited in a number of ways like relevance and accuracy. Also, the objectives and
methods adopted to collect data may not be suitable to the current situation. Therefore,
the researcher should be careful when using secondary data. Examples of secondary data
are censuses data, publications, internal records of the organizations, reports, books,
journal articles, websites etc.
31
Data
Primary Secondary
ii) The Accessible Population is also known as the study population. It is the
population to which a researcher can apply the conclusions of the study. This
population is a subset of the target population.
32
Sample
A sample is simply a subset or subgroup of population (Frey, Carl, & Gary, 2000).The
concept of sample arises from the inability of the researchers to test all the individuals in
a given population. Sampling is the process of selecting some individuals from the
accessible population, in a way that these individuals represent whole accessible
population. The sample should be representative in a sense that each individual should
represent the characteristics of the whole population (Lohr, 1999). The main function of
the sample is to allow the researchers to conduct the study to individuals from the
population so that the results of their study can be used to derive conclusions that will
apply to the entire population.
Probability sampling
Is a process that utilizes some form of random selection. In probability sampling, each
individual in chosen with a known probability. This type of sampling is also known as
random sampling or representative sampling; and depends on objective judgment.
Various types of probability are as under:
i) Simple Random sampling
In random sampling each member of the population has an equal chance of being
selected as subject. Each member is selected independently of the other member of
population. Many methods are used to proceed with random sampling. In a
commonly used method each member of the population is assigned a unique
number. All assigned numbers are placed in bowl and mixed thoroughly. The
researcher, then blind-folds and picks numbered tags from the bowl. All the
numbers picked are the subjects of the study. Another method is to use computer
for random selection from the population. For smaller population first method is
useful and for larger population computer-aided method is preferred.
33
ii) Systematic Random Sampling
In systematic random sampling, the researcher first randomly picks the first item or
the subject from the population. Then he selects each nth subject from the list. The
procedure involved in this sampling is easy and can be done manually. The sample
drawn using this procedure is representative unless certain characteristics of the
population are repeated for every nth member, which is highly risky.
Suppose a researcher has a population of 100 individuals and he needs 12 subjects.
He first picks his starting number 7. He then picks his interval 8. The members of
his sample will be individual 7, 15, 23, 31, 39, 47, 55, 63, 71, 79, 87, and 95
Advantages
This type of sampling is appropriate when the population has diversified social or ethnic
subgroups.
Disadvantages
While using this type of sampling, there is greater chance of overrepresentation of
subgroups in the sample.
34
or cluster of students such as classes, and then, select the sample of students from among
the clusters.
Advantages
This type of sampling is appropriate for larger population. It saves time and resources.
Disadvantages
In this type of sampling, there is a greater chance of selecting a sample that is not
representative of the whole population.
Non-Probability Sampling or Judgmental Sampling
This technique depends on subjective judgment. It is a process where probabilities cannot
be assigned to the individuals objectively. It means that in this technique samples are
gathered in a way does not give all individuals in the population equal chances of being
selected. Choose these methods could result in biased data or a limited ability to make
general inferences based on the findings. But there are also many situations in which
choosing this kind of sampling techniques is the best choice for a particular research
question or the stage of research.
35
iii) Snowball Sample
This type of sampling is appropriate when the members of the population are
difficult to locate, such as homeless industry workers, undocumented immigrants
etc. a snowball sample is one in which the researcher collects data on a few
members of the target population he or she can locate, then asks to locate those
individuals to provide information needed to locate other members of that
population whom they know. For example, if a researcher wants to interview
undocumented immigrants from Afghanistan, he might interview a few
undocumented individuals he knows or can locate, and would then rely on those
subjects to help locate more undocumented individuals. This process continues
until the researcher has all the interviews he needed, until all contacts have been
exhausted. This technique is useful when studying a sensitive topic that people
might not openly talk about, or if talking about the issue under investigation could
jeopardize their safety.
36
2.4 Activities
1. Suppose a scientist is conducting an experiment to test the what extant a vitamin
could extend a person’s life expectancy. Identify:
i) Independent Variable of the experiment.
ii) Dependent Variable of the experiment.
37
2.5 Bibliography
Bartz, A. E. (1981). Basic Statistical Concepts (2nd Ed.). Minnesota: Burgess Publishing
Company
Deitz, T., & Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley_-Blackwell
Gay, L. R., Mills, G. E., & Airasian, P. W. (2010). Educational Research: Competencies
for Analysis and Application, 10th Edition. Pearson, New York USA.
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
38
UNIT-3
STATISTICAL GRAPHICS /
EXPLORATORY DATA ANALYSIS
Written By:
Miss Sumbal Asghar
Reviewed By:
Dr. Rizwan Akram Rana
39
Introduction
Graphical representation of data is for the purpose of easier interpretation. Facts and
figures as such do not catch our attention unless they are presented in an interesting way.
Graphical representation of data is the most commonly used interesting modes of
presentation. The purpose of this unit is to make you familiar with this interesting mode
of presentation.
Objectives
After reading this unit, you will be able to explain:
1. Bar Chart
2. Pictograms
3. Histogram
4. Frequency Polygon or Ogive
5. Scatter Plot
6. Box Plot
7. Pie Chart
Data for a bar chart are entered in columns. Each numeric data value becomes a bar. The
chart is constructed such that lengths of the different bars are proportional to the size of
the category they represent. X-axis represents the different categories and has no scale;
the y-axis does have a scale and indicates the units of measurement, in case of vertical
bar charts, and vice versa in case of horizontal bar charts.
In the following figure result of first, second and third term of a student in the subjects of
English, Urdu, Mathematics and Pak-Studies.
40
120
100
80
1st Term
60
2ns Term
3rd Term
40
20
0
English Urdu Maths Pak Studies
Pak Studies
Maths
3rd Term
2nd Term
First Term
Urdu
English
0 20 40 60 80 100 120
Fig 2: Horizontal bar chart
3.1.1 Advantages and Disadvantages of Bar Charts
41
Following are the advantages of bar charts.
i) They show data category in a frequency distribution.
ii) They display relative numbers / proportions of multiple categories.
iii) They summarize a large amount of data in an easily interpretable manner.
iv) They make trends easier to highlight than tables do.
v) By bar charts estimates can be made quickly and accurately.
vi) They are easily accessible to everyone.
3.2 Pictograms
A pictogram is a graphical symbol that conveys its meaning through its pictorial
resemblance to a physical object. A pictogram may include a symbol plus graphic
elements such as border, back pattern, or color that is intended to covey specific
information s. we can also say that a pictogram is a kind of graph that uses pictures
instead of bars to represent data under analysis. A pictogram is also called “pictograph”,
or simply “picto”.
Pictograms form a part of our daily lives. They are used in transport, medication,
education, computers etc. they indicate, in iconic form, places, directions, actions or
constraints on actions in either the real world (a road, a town, etc) or in virtual world
(computer, internet etc.).
42
Following are the advantages of pictograms:
i) Pictograms can make warnings more eye-catching.
ii) They can serve as an “instant reminder” of a hazard or an established message.
iii) They may improve warning comprehension for those with visual or literacy
difficulties.
iv) They have the potential to be interpreted more accurately and more quickly than words.
v) They can be recognized and recalled far better than words.
vi) They can improve the legibility of warnings.
vii) They may be better when undertaking familiar routine tasks.
Example
The following table shows the number of laptops sold by a company for the months
January to March. Construct a pictograph for the table.
Month January February March
Number of laptops 25 15 20
Solution:
January
February
March
represents 5 laptops
Example
43
School Subject pictogram
Source: www.kids-pages.com
3.3 Histogram
A histogram is a type of graph that provides a visual interpretation of numerical data by
indicating the number of data points that lie within the range value. These range values
are called classes or bins.
A histogram looks similar to bar charts. Both are ways to display data set. The height of
the bar corresponds to the relative frequency of the amount of data in the class. The
higher the bar is, the greater the frequency of the data will bean vice versa. The main
difference between these graphs is the level of measurement of the data. Bar graphs are
used for data at nominal level of measurement. It measures the frequency of categorical
data. On the other hand histograms are used for data that is at least ordinal level of
measurement. As a common practice the bars of bar graph are rearranged in order for
decreasing height. However the bars of cannot be rearranged. They must be displayed in
order that the classes occur.
44
A bar graph presents actual counts against categories. The height of the bar indicates the
number of items in that category. A histogram displays the same categorical variables in
bins. While creating a histogram, you are actually creating a bar graph that shows how
many data points are there within the range (an interval), called a bin.
There are no hard and fast rules about how many bins there should be. But the rule of
thumb is 5-20 bins. Less than 5 bins will have little meaning and more than 20 bins, will
make data hard to read and interpret. Ideally 5-7 bins are enough.
ii) Bimodal
A bimodal shape, shown below, has two peaks. This shape may show that the data
has come from two different systems. Often in a single system, there may be two
modes in the data set.
45
iii) Skewed right
Some histograms will show a skewed distribution to the right, as shown below. A
distribution skewed to the right is said to be positively skewed. This kind of
distribution has a large number of occurrences in the lower value cells (left side)
and few in the upper value cells (right side). A skewed distribution can result when
data is gathered from a system with has a boundary such as zero. In other words, all
the collected data has values greater than zero.
46
v) Uniform
A uniform distribution, as shown below, provides little information about the
system. It may describe a distribution which has several modes (peaks). If your
histogram has this shape, check to see if several sources of variation have been
combined. If so, analyze them separately. If multiple sources of variation do not
seem to be the cause of this pattern, different groupings can be tried to see if a more
useful pattern results. This could be as simple as changing the starting and ending
points of the cells, or changing the number of cells. A uniform distribution often
means that the number of classes is too small.
vi) Random
A random distribution, as shown below, has no apparent pattern. Like the uniform
distribution, it may describe a distribution that has several modes (peaks). If your
histogram has this shape, check to see if several sources of variation have been
combined. If so, analyze them separately. If multiple sources of variation do not
seem to be the cause of this pattern, different groupings can be tried to see if a more
useful pattern results. This could be as simple as changing the starting and ending
points of the cells, or changing the number of cells. A random distribution often
means there are too many classes.
Source: http://www.pqsystems.com/qualityadvisor/DataAnalysisTools/histogram.php
47
3.4 Frequency Polygon
The frequency polygon is as graph that displays data by using lines that connect points
plotted for the frequencies at the midpoint of the classes. This graph is useful for
understanding the shape of distribution. They are good choice for displaying cumulative
frequency distribution.
There are two methods for of drawing a cumulative frequency curve or ogive.
i) The less than method
In this method a frequency distribution is prepared which gives the number of items
that are less than a certain size. It gives a series which is cumulatively upward.
ii) The greater than method
In this method a frequency distribution is prepared that gives the number of items
that exceed a certain size and gives a series which is cumulatively downward.
Example
Marks of 30 students of a class, obtained in a test out of 75, are given below: 42, 21, 50, 37,
38, 42, 49, 52, 38, 53, 57, 47, 29, 59, 61, 33, 17, 17, 39, 44, 42, 39, 14, 7, 27, 19, 54, 51.
Total 30
48
3.6 Scatter Plot
A scatter plot is used to plot data in XY- plane to show how much one variable or data set
is affected by another. It has points that show the relationship between two variables or
two sets of data. These points are sometimes called markers and position of these points
depends on the values in the columns sets on the XY axis. Scatter plot gives good visual
picture of the relationship or association between two variables or data sets, and aids to
interpretation of the correlation coefficient or regression model.
The relationship between two data sets or variables is called correlation. If the markers
are close together and make a straight line in the scatter plot, the two variables of data
sets have high correlation. If the markers are equally distributed in the scatter plot, the
correlation is low, or zero.
Correlation may be positive or negative. Correlation is positive when the values increase
together, i.e. if one value increases the other will also increase or if once value decreases
the other will also decrease. On the other hand, correlation is negative when one value
increases the other decreases, and vice versa.
49
Name of Student GPA
A 2.0
B 2.21
C 2.38
D 2.32
E 2.11
F 3.01
G 3.92
H 3.11
I 3.25
J 3.60
K 2.97
L 3.11
M 3.34
N 3.96
O 3.69
P 2.99
Q 2.94
R 3.41
S 3.90
Example
GPA
4.5
4
3.5
3
2.5
2
1.5
1
0.5 GPA
0
0 2 4 6 8 10 12 14 16 18 20
Example
50
Name of Student Achievement Motivation Anxiety
A 95 50 15
B 96 84 54
C 65 46 25
D 59 33 36
E 68 24 56
F 84 86 54
G 59 90 58
H 74 14 47
I 58 66 56
J 59 71 59
K 68 56 68
L 59 71 84
M 62 79 59
N 35 82 62
O 48 80 10
P 57 69 15
Q 96 64 59
R 58 86 67
S 86 90 68
120
100
80
Achievement
60 Motivation
Anxiety
40
20
0
0 2 4 6 8 10 12 14 16 18 20
51
The box plot is an exploratory graph. It is a standardized way of displaying the
distribution of data based on the five summary statistics: minimum, first quartile, median,
third quartile, and maximum. First and third quartile is called two hinges, first quartile is
the lower hinge and the third quartile is the upper hinge. Minimum and the maximum are
two whiskers. Minimum is the lower whisker and the maximum is the upper whisker. In
other words we can say that box plot visualizes five summary statistics: the median, two
hinges and two whiskers.
In the simplest box plot the central triangle spans the first quartile to the third quartile
(inter quartile range IQR). A segment inside the rectangle shows the median and whiskers
above and below the box show the locations of the minimum and maximum.
Box plot is useful for identifying outliers and for comparing distributions. In other words
we can say that box plot gives us information about the location and variation in the data
set. Particularly it helps us in detecting and illustrating location and variation changes
between different groups of data.
A box-plot can tell whether a data set is symmetric (when the median is in the center of
the box), but it can’t tell the shape of the symmetry the way a histogram can.
52
is easy to see which item has maximum frequency and which is not or which item is the
most popular and which is not. The main purpose of using a pie chart is to show part-
whole relationship. These charts are used for displaying data that are classified into
nominal or ordinal categories.
Scores
1st Term
2nd Term
3rd Term
3.8.1 How to Read a Pie Chart?
It is easy to read and interpret a pie-chart. Usually, a pie-chart has several bits of data,
and each is pictured on a pie-chart as a pie slice. Some data have larger slices than others.
So it is easy to decide which data have maximum frequency and which have minimum.
iii) Do you want to compare the parts to each other or the parts to the whole?
If the main purpose is to show part-whole relationship then pie chart is useful but if
the main purpose is to show part-part relationship then pie chart is useless and wise
to use another chart.
53
Column1
1st Qtr
2nd Qtr
3rd Qtr
4th Qtr
5th Qtr
6th Qtr
Source: https://eagereyes.org/techniques/pie-charts
54
3.9 Self Assessment Questions
Q. 1 What is a bar chart?
Q. 2 For what purpose bar carts are used?
Q. 3 What type of characteristics a pictogram should have to successfully
convey the meaning?
Q. 4 Write down the advantages and drawbacks of using pictograms.
Q. 5 What is a histogram?
Q. 6 Draw a bell-shaped histogram.
Q. 7 Write down the methods for drawing cumulative frequency polygon.
Q. 8 Write down the rationale for using scatter plot.
Q. 9 Write down any four questions that can be answered using scatter plot.
Q. 10 Write down the types of box plot.
Q. 11 What is a pie-chart?
Q. 12 Write down the criteria to determine whether pie-chart is a right choice.
3.10 Activities
1. Make a list of advantages and disadvantages of bar chart.
2. Make a list of advantages and disadvantages of pictogram.
3. Make a list of the situations that provide rationale to use scatter plot.
4. Make a pie chart that shows the drawback of pie chart.
55
3.11 Bibliography
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
https://eagereyes.org/techniques/pie-charts
http://www.pqsystems.com/qualityadvisor/DataAnalysisTools/histogram.php
56
UNIT-4
Written By:
Miss Sumbal Asghar
Reviewed By:
Dr. Rizwan Akram Rana
57
Introduction
Measures of central tendency estimate normal or central value of a dataset, while
measures of dispersion are important for describing the spread of the data, or its variation
around a central value. Two distinct samples may have same mean or median, but
completely different level of variability and vice versa. A proper description of a set of
data should include both of these characteristics. There are various methods that can be
used to measure the dispersion of a dataset. In this unit you will study range, quartiles,
quartile deviation, mean or average deviation, standard deviation and variance. Two
measures of shape and discussion about co-efficient of variation are also included in this
unit.
Objectives
After reading this unit, you will be able to:
1. tell the basic purpose of measure of central tendency.
2. define Range.
3. determine range of a given data.
4. write down the formulas for determining quartiles.
5. define mean or average deviation.
6. determine variance and standard deviation.
7. define normal curve.
8. explain skewness and kurtosis.
There is dispersion when there is dissimilarity among the data values. The greater the
dissimilarity, the greater the degree of dispersion will be.
58
Measure of dispersion enables us to compare two or more series with regards to their
variability. It is also looked as a means of determining uniformity or consistency. A high
degree would mean little consistency or uniformity whereas low degree of variation
would mean greater uniformity or consistency among the data set. Commonly used
measures of dispersion are range, quartile deviation, mean deviation, variance, and
standard deviation.
4.1.1 Range
The range is the simplest measure of spread and is the difference between the highest and
lowest scores in a data set. In other words we can say that range is the distance between
largest score and the smallest score in the distribution. We can calculate range as:
Range = Highest value of the data – Lowest value of the data
For example, if lowest and highest marks scored in a test are 22 and 95 respectively, then
Range = 95 – 22 = 73
The range is the easiest measure of dispersion, and is useful when you wish to evaluate
whole of a dataset. But it is not considered a good measure of dispersion as it does not
utilize the other information related to the spread. The outliers, either extreme low or
extreme high value, can considerably affect the range.
4.1.2 Quartiles
The values that divide the given set of data into four equal parts is called quartiles, and is
denoted by Q1, Q2, and Q3. Q1 is called the lower quartile and Q3 is called the upper
quartile. 25% of scores are less than Q1and 75% scores are less than Q3. Q2 is the median.
The formulas for the quartiles are:
Q1 = (N + )th Score
Q2 = 2 (N + )th = (N + ) th Score
Q3 = 3(N + 1) / 4th Score
In order to calculate quartile deviation from ungrouped data, following steps are used.
i) Arrange the test scores from highest to lowest
ii) Assign serial number to each score. The first serial number is assigned to the
lowest score.
59
iii) Determine first quartile (Q1) by using formula Q1 = . Use obtained value to
locate the serial number of the score that falls under Q1.
iv) Determine the third (Q3), by using formula Q3 = . Locate the serial number
corresponding to the obtained answer. Opposite to this number is the test score
corresponding to Q3.
v) Subtract the Q1 from Q3, and divide the difference by 2.
Or
Ϭ = √ Ʃ (X – X)2 / n
Ϭ is a Greek letter “Sigma”
4.1.6 Variance
The variance of a set of scores is denoted by σ2and is defined as
Ϭ2= Ʃ (X – X)2 / n
Where X is the mean, n is the number of data values and X stand for each of the scores,
and Ʃ means add up all the values.
60
distribution of data may naturally occur in several possible ways, with a number of
possibilities for standard deviation (which could be from 1 to infinity). A standard
normal curve has a mean of 0 and standard of 1. The larger the standard deviation, the
flatter the curve will be and vice versa. A standard normal distribution is given below.
a) Skewness
Skewness tells us about the amount and direction of the variation of the data set. It is a
measure of symmetry. A distribution or data set is symmetric if it looks the same to the left
and right of the central point. If bulk of data is at the left i.e. the peak is towards left and the
right tail is longer, we say that the distribution is skewed right or positively skewed.
On the other hand if the bulk of data is towards right or, in other words, the peak is
towards right and the left tail is longer, we say that the distribution is skewed left or
negatively skewed.If the skewness is equal to zero, the data are perfectly symmetrical.
But it is quiet unlikely in real world.
61
Source: Google Images
b) Kurtosis
Kurtosis is a parameter that describes the shape of variation. It is a measurement that tells
us how the graph of the set of data is peaked and how high the graph is around the mean.
In other words we can say that kurtosis measures the shape of the distribution, .i.e. the
fatness of the tails, it focuses on how returns are arranged around the mean. A positive
value means that too little data is in the tail and positive value means that too much data
is in the tail. This heaviness or the lightness in the tail means that data looks more peaked
of less peaked. Kurtosis is measured against the standard normal distribution. A standard
normal distribution has a kurtosis of 3.
Kurtosis has three types, mesokurtic, platykurtic, and leptokurtic. If the distribution has
kurtosis of zero, then the graph is nearly normal. This nearly normal distribution is called
mesokurtic. If the distribution has negative kurtosis, it is called platykurtic. An example
of platykurtic distribution is a uniform distribution, which has as much data in each tail as
it does in the peak. If the distribution has positive kurtosis, it is called leptokurtic. Such
distribution has bulk of data in the peak.
62
4.3 Co-Efficient of Variation
The coefficient of variation is another useful statistics for measuring dispersion of a data
set. The coefficient of variation is
C.V = (s / x ) × 100
The coefficient of variation is invariant with respect to the scale of the data. On the other
hand, standard deviation is not scale variant.
4.5 Activities
Take a cardboard. Cut it into 4x4 pieces, and:
i) Cut one piece into standard normal distribution shape and mention its name on it.
ii) Cut one piece into negatively skewed shape and mention its name on it.
iii) Cut one piece into positively skewed shape and mention its name on it.
iv) Cut one piece into no skewed shape and mention its name on it.
v) Cut one piece into mesokurtic shape and mention its name on it.
vi) Cut one piece into platykurtic shape and mention its name on it.
vii) Cut one piece into leptokurtic shape and mention its name on it.
63
4.6 Bibliography
Bartz, A. E. (1981). Basic Statistical Concepts (2nd Ed.). Minnesota: Burgess Publishing
Company
Deitz, T., & Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley_-Blackwell
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
64
UNIT-5
Written By:
Salman Khalil Chaudhary
Reviewed By:
Dr. Rizwan Akram Rana
65
Introduction
In this unit you will study three main measures of central tendency – the mean, median
and the mode. The main purpose of measures of central tendency is to identify the
location of the center of various distributions. This helps us to get a better idea as to
where the center of a distribution is located.
Merits and demerits of mean, median and mode are also discussed in the unit.
Objectives
After reading this unit, you will be able to:
1. write down the goals of measure of central tendency.
2. explain the characteristics of good measure of central tendency.
3. determine mean of a given set of data.
4. explain merits and demerits of mean.
5. define median.
6. explain procedures for determining median in case number of scores is even or odd.
7. explain merits and demerits of median.
8. calculate median of a given data.
9. define mode.
10. explain merits and demerits of mode.
11. calculate mode of a given data.
5.1 Introduction
Measures of central tendency (also referred as measures of center of central location) allow
us to summarize data with a single value. It is a typical score among a group of scores (the
midpoint). They give us an easy way to describe a set of data with a single number. This
single number represents a value or score that generally is in the middle of the dataset.
66
5.2 Mean
Mean is the most commonly used measure in educational research. It is appropriate for
describing ratio or interval data. It can also be used for both continuous and discrete
numeric data. It is the arithmetic average of the score. It is determined by adding up all
the scores and then by the sum by the total number of scores. Suppose we have scores,
40, 85, 94, 62, 76, 66, 90, 59, 68, and 84. In order to find the mean of these scores we
simply add all the scores, which comes to 724. Then divide this sum 10 (total number of
scores). We will get 72.4, which is the mean score.
Where Ʃ represents “Sum of”, X represents any raw score value, n represents total
number of scores.
We can also define mean as mean is the amount each individual would get if the total
(ƩX) were divided equally among all the individual members (n) in the distribution. In
some other words we can say that the mean is the balance point for the distribution.
To interpret the as the “balance point or the center value”, we can use the analogy of a
seesaw. Its mean lies right at the center where the fulcrum keeps the board perfectly
balanced. As the mean is based on every score or value of the dataset so it is influenced
by outliers and skewed distribution. Also it cannot be calculated for categorical data as
the values cannot be summed.
5.3 Median
Median is the middle value of rank order data. It divides the distribution in two halves (i.e. 50%
of scores or observations on either side of median value). It means that this value separates
higher half of the data set from the lower half. The goal of the median is to determine the
precise midpoint of the distribution. Median is appropriate for describing ordinal data.
67
5.3.1 Procedure for Determining Median
When the number of scores is odd, simply arrange the scores in order (from lower to higher
or from higher to lower). The median will be the middle score in the list. Consider the set of
scores 2, 5, 7, 10, 12. The score “7”lies in the middle of the scores, so it is median.
When there is an even number of scores in the distribution, arrange the scores in order
(from lower to higher or from higher to lower). The median will be the average of the
middle two score in the list. Consider the set of scores 4, 6, 9, 14 16, 20. The average of
the middle two scores 11.5 (i.e. 9+14/2 = 23/2 = 11.5) is the median of the distribution.
Median is less affected by outliers and skewed data and is usually preferred measure of
central tendency when the distribution is not symmetrical. The median cannot be
determined for categorical or nominal data.
5.4 Mode
The mode is the most frequently occurring score in the distribution. Consider following
data set.
25, 43, 39, 25, 82, 77, 25, 47.
The score 25 comes more frequently, so it is the mode. Sometimes there may be no single
mode if no one value appears more than any other. There may be one mode (uni-modal),
two modes (bi-model), three modes (tri-model), or more than three modes (multi-model).
Mode is useful when scores reflect a nominal scale of measurement. But along with mean
and median it can also be used for ordinal, interval or ratio data. It can be located
graphically by drawing histogram.
68
5.4.1 Merits of Mode
i) It is easy to understand and easy to calculate.
ii) It is not affected by extreme values.
iii) Even if the extreme values are not known mode can be calculated.
iv) It can be located just by inspection in many cases.
v) It can be located graphically.
vi) It is always present in the data.
vii) It is applicable for both quantitative and qualitative data.
viii) It is useful for methodological forecasts.
5.6 Activities
Discuss with your colleague and:
i) Try to make a list of merits and demerits of mean not given in the unit.
ii) Try to make a list of merits and demerits of median not given in the unit.
iii) Try to make a list of merits and demerits of mode not given in the unit.
69
5.7 Bibliography
Agresti, A. & Finlay, B. (1997). Statistical Methods for Social Sciences, (3rd Ed. ).
Prentice Hall.
Argyrous, G. (2012). Statistics for Research, with a guide to SPSS. India: SAGE
Publications.
Bartz, A. E. (1981). Basic Statistical Concepts (2nd Ed.). Minnesota: Burgess Publishing
Company.
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
70
UNIT-6
INFERENTIAL STATISTICS
Written By:
Salman Khalil Chaudhary
Reviewed By:
Dr. Rizwan Akram Rana
71
Introduction
Inferential statistics is of vital importance in educational research. It is used to make
inferences about the population on the bases of data obtained from the sample. It is also
used to make judgments of the probability that an observed difference among groups is a
dependable one or one that might have happened by chance in the study.
In this unit, you will study introduction, area, logic and importance of inferential
statistics. Hypothesis testing, logic and process of hypothesis testing and errors in
hypothesis are also discussed. In the last of the unit t-test, its types and general
assumptions regarding the use of t-test are discussed.
Objectives
After reading this unit, you will be able to:
1. explain the term “Inferential Statistics”.
2. explain the area of Inferential Statistics.
3. explain the logic of Inferential Statistics.
4. explain the Importance of Inferential Statistics in Educational Research.
5. tell, What Hypothesis Testing is.
6. explain the Logic of Hypothesis Testing.
7. explain the Uncertainty and errors in Hypothesis Testing.
8. explain t-test and its Types.
72
Descriptive statistics only gives us the central values, dispersion or the variability of the
data but inferential statistics leads us to take a decision about the whole population and in
the end to any conclusion. Inferential statistics allows us to use what we have learnt from
descriptive statistics. Inferential statistics enables us to infer from the data obtained the
sample what the population might think.
i) Estimating Parameter
This means taking a statistics from the sample data (e.g. the sample mean) and
saying something about population parameter (e.g. the population mean).
ii) Hypothesis testing
This is where a researcher can use sample data to answer research questions.
Inferential statistics deals with two or more than two variables. If in an analysis there are
two variables it is called bivariate analysis and if the variables are more than two it is
called multivariate analysis. A number of different types of inferential statistics are in
use. All of which depend of the type of variable i.e. nominal, ordinal, interval, and ratio.
Although the type of statistical analysis is different for these variables, yet the main
theme is the same we try to determine how one variable compare to another.
It should be noted that inferential statistics always talk in terms of probability. This can
be made highly reliable by designing right experimental conditions. The inferences are
always an estimate with a confidence interval. In some cases there is simply a rejection of
hypothesis.
Several models are available in inferential statistics that help in the process of data
analysis. A researcher should be careful while choosing any model. Because, choosing a
wrong model may give wrong conclusions.
73
Population of male Population of female
English language English language
students students
N = 1000 N = 1000
N N
Sample 1 Sample 2
n = 60 n = 60
The researcher wants to know whether the male population is different from female
population – that is, will the mean score of the male group on attitude scale is different
from the mean score of the female group? The researcher does not know the means of the
two populations. He only has mean scores of two samples on which he has to rely on to
provide information about the populations.
Now it comes in mind that is it reasonable to assume that each sample will give a fairly
accurate picture of the whole population? It certainly is possible, because each sample
was selected randomly from its population. On the other hand, the students in each
sample are only a small portion of their respective population. It is only rare that a sample
is absolutely identical to the population from which it is drawn, on given characteristics.
The data the researcher obtains from two samples depends on the individual students
selected to be in the sample. If another two samples were selected randomly their makeup
would differ from previously selected samples. Their mean on the attitude scale would be
different, and the researcher would end up with a different data set. How can the
researcher be sure that any particular selected sample is a true representative of its
population? Indeed he cannot. He needs some help to be sure that the sample is
representative of the population and the results obtained from the sample data be
generalized to whole population. Inferential statistics will help the researcher and allow
him to make judgment about data and make generalization about a population based on
the data obtained from the sample.
74
study. It helps enables researchers to infer properties of a population based on data
collected from a sample of individuals
Inferential statistics have larger value because these techniques offset problems
associated with data collection. For example, time-cost factor associated with collection
of data on the entire population may be prohibitive. The population may large and
difficult to manage. In this case inferential statistics can prove to be invaluable to
educational/social scientist.
Hypothesis testing is a statistical method that uses sample data to evaluate a hypothesis
about a population parameter (Gravetter & Wallnau, 2002).A hypothesis test is usually
used in context of a research study. Depending on the type of research and the type of
data, the details of the hypothesis test will change from on situation to another.
75
6.3.2 Four-Step Process for Hypothesis Testing
The process of hypothesis testing goes through following four steps.
i) Stating the Hypothesis
The process of hypothesis testing begins by stating a hypothesis about the unknown
population. Usually, a researcher states two opposing hypotheses. And both
hypotheses are stated in terms of population parameters.
The first and most important of two hypotheses is called null hypothesis. A null
hypothesis states that the treatment has no effect. In general, null hypothesis states
that there is no change, no effect, no difference – nothing happened. The null
hypothesis is denoted by the symbol Ho (H stands for hypothesis and 0 denotes that
this is zero effect).
The null hypothesis (Ho) states that in the general population there is no change, no
difference, or no relationship. In an experimental study, null hypothesis (Ho)
predicts that the independent variable (treatment) will have no effect on the
dependent variable for the population.
The second hypothesis is simply the opposite of null hypothesis and it is called the
scientific or alternative hypothesis. It is denoted by H1. This hypothesis states that
the treatment has an effect on the dependent variable.
76
6.3.3 Uncertainty and Error in Hypothesis Testing
Hypothesis testing is an inferential process. It means that it uses limited information
obtained from the sample to reach general conclusions about the population. As a sample
is a small subset of the population, it provides only limited or incomplete information
about the whole population. Yet hypothesis test uses information obtained from the
sample. In this situation, there is always the probability of reaching incorrect conclusion.
Generally two kinds of errors can be made.
i) Type I Errors
A type I error occurs when a researcher rejects a null hypothesis that is actually
true. It means that the researcher concludes that the treatment does have an effect
when in fact the treatment has no effect.
Type I error is not a stupid mistake in the sense that the researcher is overlooking
something that should be perfectly obvious. He is looking at the data obtained from
the sample that appear to show a clear treatment effect. The researcher then makes
a careful decision based on available information. He never knows whether a
hypothesis is true or false.
The consequences of a type I error can be very serious because the researcher has
rejected the null hypothesis and believed that the treatment had a real effect. it is
likely that the researcher will report or publish the research results. Other researchers
may try to build theories or develop other experiments based on false results.
The consequences of Type II error are not very serious. In case of Type II error the
research data do not show the results that the researcher had hoped to obtain. The
researcher can accept this outcome and conclude that the treatment either has no effect
or has a small effect that is not worth pursuing. Or the researcher can repeat the
experiment with some improvement and try to demonstrate that the treatment does
work. It is impossible to determine a single, exact probability value for a type II error.
Summarizing we can say that a hypothesis test always leads to one of two decisions.
i) The sample data provides sufficient evidence to reject the null hypothesis and the
researcher concludes that the treatment has an effect.
ii) The sample data do not provide enough evidence to reject the null hypothesis. The
researcher fails to reject the null hypothesis and concludes that the treatment does
not appear to have an effect.
77
In either case, there is a chance that the data are misleading and the decision is wrong.
The complete set of decision and outcome is shown in the following table.
Table: 6.1
Possible outcome of statistical decision
Actual Situation
No effect, Effect exists,
Ho true Ho false
6.4 T-Test
A t-test is a useful statistical technique used for comparing mean values of two data sets
obtained from two groups. The comparison tells us whether these data sets are different
from each other. It further tells us how significant the differences are and if these
differences could have happened by chance. The statistical significance of t-test indicates
whether or not the difference between the mean of two groups most likely reflects a real
difference in the population from which the groups are selected.
t-tests are used when there are two groups (male and female) or two sets of data (before and
after), and the researcher wishes to compare the mean score on some continuous variable.
78
Here at this level it is necessary to know some general assumptions regarding use of t-
test. The first assumption regarding t-test concerns the scale of measurement. It means
that it is assumed that the dependent variable is measured at interval or ratio scale. The
second assumption made is that of a simple random sample, that the data is collected
from a representative, randomly selected portion of the total population. The third
assumption is that the data, when plotted, results in a normal distribution i.e. in bell-
shaped distribution curve. The fourth assumption is that the observation that make up
data must independent of one another. That is, each observation or measurement must not
be influences by any other observation or measurement. The fifth assumption is that a
reasonably large sample size is used. A large sample size means that the distribution of
results should approach a normal bell-shaped curve. The final assumption is homogeneity
of variance. Variance will be homogeneous or equal when the standard deviation of
samples is approximately equal.
6.6 Activities
1. Suppose we exclude inferential statistics from our research. What will happen?
Write down a few lines.
2. You have scores of two different groups of students and you have to compare the
scores. Discuss with your colleague and select appropriate statistical test.
79
6.7 Bibliography
Fraenkel, J. R., Wallen, N. E., & Hyun, H. H. (2012). How to Design and Evaluate in
Education. (8th Ed.) McGraw-Hill, New York
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
Pallant, J. (2005). SPSS Survival Manual – A step by step guide to data analysis using
SPSS for Windows (Version 12). Australia: Allen & Unwin.
80
UNIT-7
INFERENTIAL STATISTICS:
CORRELATION AND REGRESSION
Written By:
Prof. Dr. Nasir Mahmood
Reviewed By:
Dr. Rizwan Akram Rana
81
Introduction
A correlation is a relationship between two variables. The purpose of using correlation in
research is to determine the degree to which a relationship exists between two or more
variables. Correlation is important in research because several hypotheses are stated in
terms of correlation or lack of correlation between two variables, so correlational studies
are directly related to such hypotheses.
Regression is used when the relationship includes a dependent variable and one or more
independent variables. It helps us understand which among the independent variables are
related to the dependent variable, and to explore the forms of these relationships.
Owing to the importance of correlation and regression in research, these are given in this
unit in detail.
Objectives
After reading this unit, you will be able to:
1. explain correlation.
2. explain where and why to use correlation.
3. explain what considerations should be kept in mind while interpreting correlation.
4. explain Pearson and Spearman correlation method.
5. explain the situations in which Spearman correlation can be used.
6. explain Regression.
7. explain why we use regression analysis.
8. explain types of regression.
9. explain p-value.
7.1 Correlation
Correlation is a statistical technique used to measure and describe relationship between two
variables. These variables are neither manipulated nor controlled, rather they simply are
observed as they naturally exist in the environment. Suppose a researcher is interested in
relationship between number of children in a family and IQ of the individual child. He would
take a group of students coming from different families. Then he simply observe or record the
number of children in a family and then measure IQ score of each individual student same
group. He will neither manipulate nor control any variable. Correlation requires two separate
scores for each individual (one score from each of two variables). These scores are normally
identified as X and Y and can be presented in a table or in a graph.
82
i) The Direction of the Relationship
The direction of the relationship can be classified into two basic categories:
positive and negative.
In a positive correlation both variables tend to change into same direction. When
variable X increases, the variable Y also increases. And if the variable X decreases,
the variable Y also decreases. In other words we can say that both variables are
directly proportional to each other.
In a negative correlation both variables do not tend to change into same direction.
They go in opposite direction of each other. When the variable X increases, the
variable Y decreases. And if the variable X decreases, the variable Y increases. In
other words we can say that both variables are indirectly proportional to each other.
83
To calculate the Pearson correlation r we use the formula
Two formulas (definitional and computational) are available to calculate the sum of
square of product. Both formulas are given in the following box.
1. The definitional formula is SP = ∑ (X – X ) (Y – Y )
2. The computational formula is
SS is sum of squares, SSx is the sum of squares of the variable X and SSy is the sum of
squares of variable Y. In the following lines different formulas are given to calculate SSx
and SSy. These formulas are categorized as definitional and computational. The
definitional formulas for sum of squares of variable X are:
SSx = ∑ (X − X) 2
It should be kept in mind that whichever formula one uses, it will yield similar result.
ii) Validity
Suppose a researcher develops a new test for measuring intelligence. It is necessary
that he should show that this new test valid and truly measures what it claims to
measure. One common technique for demonstrating validity is to use correlation.
84
If newly constructed test actually measures intelligence, then the scores on this test
should be related to other already established measures of intelligence – for
example standardized IQ tests, performance on learning tasks, problem-solving
ability, and so on. The newly constructed test can be correlated to each of these
measures to demonstrate that the new test is valid.
iii) Reliability
Apart from determining validity, correlations are also used to determine reliability.
A measurement procedure is reliable if it produces stable and consistent
measurement. It means a reliable measurement procedure will produce the same (or
nearly same) scores when the same individuals are measured under the same
conditions. One common way to evaluate reliability is to use correlations to
determine relationship between two sets of scores.
Now let us have a few words on interpreting correlation. For interpreting correlation
following consideration should be kept in mind.
i) Correlation simply describes a relationship between two variables. It does not
explain why two variables are related. That is why correlation cannot be interpreted
as a proof of cause and effect relationship between two variables.
ii) The value of the correlation cannot be affected by range of scores represented in
the data.
iii) One or two extreme data points, often called outliers, can have a dramatic effect on
the value of the correlation.
iv) When judging how good a relationship is, it is tempting to focus on the numerical
value of the correlation. For example, a correlation of + 5 is halfway between 0 and
1.00 and therefore appears to represent a moderate degree of relationship. Here it
should be noted that we cannot interpret correlation as a proportion. Although a
correlation of 1.00 means that there is a 100% perfectly predictable relationship
between variables X and Y; but a correlation of .5 does not mean that we can make
a prediction with 50% accuracy. The appropriate process of describing how
accurately one variable predicts the other is to square the correlation. Thus a
correlation of r = .5 provides r2 = .52 = .25, 25% accuracy. (The value r2 is called
coefficient of determination because it measures the proportion of variability in one
variable that can be determined from the relationship with the other variable).
85
7.3 The Spearman Correlation
The most commonly used measure of relationship is the Pearson correlation. It measures
the degree of linear relationship between two variables and is used with interval or ratio
data. However other measures of correlation have been developed for non-linear
relationship and for other type of data (or scale of measurement). One such measure is
the Spearman Correlation. The Spearman correlation is used in two situations.
i) The Spearman correlation is designed to measure the relationship between
variables measured on an ordinal scale of measurement.
ii) The Spearman correlation is used when the researcher wants to measure the
consistency of a relationship between the variables X and Y. In this case the
original scores are first converted into ranks, and then Spearman correlation is used
to measure the relationship for the ranks. Incidentally, when there is consistently
one-directional relationship between two variables, the relationship is said to be
monotonic. Thus, the Spearman correlation can be used to measure the degree of
monotonic relationship between two variables.
As the Pearson correlation measures the degree of linear relationship between two
variables, the spearman correlation measures the consistency of relationship. It can be
used as a valuable alternative of Pearson correlation even when the original raw scores
are on an interval or ratio scale. Generally Spearman correlation is computed by using
Pearson correlation formula, i.e.
rs
where D is the difference between X rank and Y rank for each individual. Again this
formula will yield the same result as Pearson correlation formula.
7.4 Regression
A correlation quantifies the degree and direction to which two variables are related. It
does not fit a line through the data points. It does not have to think about the cause and
effect. It does not natter which of the two variables is called dependent and which is
called independent.
On the other hand regression finds the best line that predicts dependent variables from the
independent variable. The decision of which variable is calls dependent and which calls
independent is an important matter in regression, as it will get a different best-fit line if
we exchange the two variables, i.e. dependent to independent and independent to
dependent. The line that best predicts independent variable from dependent variable will
not be the same as the line that predicts dependent variable from independent variable.
86
Let us start with the simple case of studying the relationship between two variables X and
Y. The variable Y is dependent variable and the variable X is the independent variable. We
are interested in seeing how various values of the independent variable X predict
corresponding values of dependent Y. This statistical technique is called regression
analysis. We can say that regression analysis is a technique that is used to model the
dependency of one dependent variable upon one independent variable. Merriam-Webster
online dictionary defines regression as a functional relationship between two or more
correlated variables that is often empirically determined from data and is used especially to
predict values of one variable when given variables of others. According to Gravetter &
Wallnua (2002), regression is a statistical technique for finding the best-fitting straight line
for a set of data is called regression, and the resulting straight line is called regression line.
These benefits help a researcher to estimate and evaluate the best set of variables to be
used for building productive models.
87
ii) Logistic Regression
Logistic regression is a statistical method for analyzing a dataset in which there are
one or more independent variables that determine an outcome. The outcome is
measured with the dichotomous (binary) variable. Like all regression analysis, the
logistic regression is a predictive analysis. It is used to describe and explain
relationship between one dependent binary variable and one or more nominal,
ordinal, interval or ratio level independent variables.
v) Ridge Regression
It is a technique for analyzing multiple regression data that suffer from
multicollinearity (independent variables are highly correlated). When
multicollinearity occurs, least squares estimates are unbiased, but their variances
are large so that they may be far from the true value. By adding the degree of bias
to the regression estimates, ridge regression reduces the standard errors.
7.5 P-Value
The p-value is the level of marginal significance within a statistical hypothesis test
representing the probability of occurrence of a given event. This value is used as an
alternative to rejection points to provide the smallest level of significance at which the
null hypothesis would be rejected.
88
In other words we can say that p-value is the calculated probability or the probability of
finding the observed or more extreme results when the null hypothesis is true. P-value is
also described in terms of rejecting null hypothesis when it is actually true.
A p-value is used in hypothesis testing to help researcher support or reject the null
hypothesis. It is evidence against the null hypothesis. The smaller p-value is the stronger
the evidence to reject the null hypothesis.
If the p-value gets lower (i.e. closer to 0% and farther away from 100), a researcher is
more inclined to reject the null hypothesis and accept the research hypothesis.
A relatively simple way to interpret p-value is to think of them as representing how likely
a result would occur by chance. For a calculated p-value of .01, we can say that the
observed outcomes would be expected to occur by chance only 1 in 100 times in repeated
tests on different samples of the population. Similarly a p-value of .05 would represent
the expected outcome to occur by chance only 5 times out of 100 times in repeated tests
and a p-value of .001 would represent the expected outcome to occur by chance only
once if the same treatment is repeated for 1000 times on different samples of the
population. In case of p-value .01, the researcher is 99% confident of getting similar
results if same test is repeated for 100 times. Similarly in case of p-value .05, the
researcher is 95% confident and in case of p-value .001, he is 999% confident of getting
similar results if same test is repeated for 100 times and 1000 times respectively.
7.7 Activities
1. Think and make a list of using correlation.
2. Enlist the consideration that you will keep in mind while using correlation.
3. Think and write primary objective of regression analysis.
89
7.8 Bibliography
Argyrous, G. (2012). Statistics for Research, with a guide to SPSS. India: SAGE
Publications.
Bartz, A. E. (1981). Basic Statistical Concepts (2nd Ed.). Minnesota: Burgess Publishing
Company
Deitz, T., & Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley_-Blackwell
Fraenkel, J. R., Wallen, N. E., & Hyun, H. H. (2012). How to Design and Evaluate in
Education. (8th Ed.) McGraw-Hill, New York
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
90
UNIT-8
Written By:
Prof. Dr. Nasir Mahmood
Reviewed By:
Dr. Rizwan Akram Rana
91
Introduction
Analysis of Variance (ANOVA) is a statistical procedure used to test the degree to which
two or more groups vary or differ in an experiment. This unit will give you an insight of
ANOVA, its logic, one-way ANOVA, its assumptions, logic and procedure. F-
distribution, interpretation of F-distribution and multiple procedures will also be
discussed.
Objectives
After reading this unit you will be able to:
1. explain what ANOVA is.
2. write down the logic behind using ANOVA.
3. explain what F-distribution is.
4. explain logic behind one-way ANOVA.
5. explain the assumptions underlying one way ANOVA.
6. explain multiple comparison procedures.
92
8.1.1 Logic of ANOVA
Let us take a hypothetical data given in the table.
Table 8.1
Hypothetical Data from an Experiment examining learning performance under three
Temperature condition
Treatment 1 Treatment 2 Treatment 3
50o 70o 90o
Sample 1 Sample 2 Sample 3
0 4 1
1 3 2
3 6 2
1 3 0
0 4 0
X=1 X=4 X=1
There are three separate samples, with n = 5 in each sample. The dependent variable is
the number of problems solved correctly
i) Between-Treatment Variance
Variance simply means difference and to calculate the variance is a process of
measuring how big the differences are for a set of numbers. The between-treatment
variance is measuring how much difference exists between the treatment
conditions. In addition to measuring differences between treatments, the overall
goal of ANOVA is to evaluate the differences between treatments. Specifically, the
purpose for the analysis is to distinguish is to distinguish between two alternative
explanations.
a) The differences between the treatments have been caused by the treatment effects.
b) The differences between the treatments are simply due to chance.
Thus, there are always two possible explanations for the variance (difference) that exists
between treatments
1) Treatment Effect: The differences are caused by the treatments. For the data in
table 8.1, the scores in sample 1 are obtained at room temperature of 50o and that of
93
sample 2 at 70o. It is possible that the difference between sample is caused by the
difference in room temperature.
2) Chance: The differences are simply due to chance. It there is no treatment effect,
even then we can expect some difference between samples. The chance differences
are unplanned and unpredictable differences that are not caused or explained by
any action of the researcher. Researchers commonly identify two primary sources
for chance differences.
Individual Differences
Each participant of the study has its own individual characteristics. Although
it is reasonable to expect that different subjects will produce different scores,
it is impossible to predict exactly what the difference will be.
Experimental Error
In any measurement there is a chance of some degree of error. Thus, if a
researcher measures the same individuals twice under same conditions, there is
greater possibility to obtain two different measurements. Often these differences
are unplanned and unpredictable, so they are considered to be by chance.
Total Variability
F=
The value obtained for F-ratio will help determine whether or not any treatment effects
exist. Consider above stated two possibilities.
1. When the treatment has no effect, then the difference between the treatments will
be entirely due to chance. In this case the numerator and the denominator of F
distribution are both measuring the same chance differences. Then F-ratio should
have a value equal to 1.00. In terms of formula’ we have
F=
=
= 1.00
The F-ratio equal to 1.00 indicates that the differences between treatments are about the
same as the difference expect by chance. So, when F-ratio is equal to 1.00, we will
conclude that there is no evidence to suggest that the treatment has any effect.
95
8.3 One Way ANOVA (Logic and Procedure)
The one way analysis of variance (ANOVA) is an extension of independent two-sample t-
test. It is a statistical technique by which we can test if three or more means are equal. It tests
if the value of a single variable differs significantly among three or more level of a factor. We
can also say that one way ANOVA is a procedure of testing hypothesis that K population
means are equal, where K ≥ 2. It compares the means of the samples or groups in order to
make inferences about the population means. Specifically, it tests the null hypothesis:
Ho : µ1 = µ2 = µ3 = ... = µk
If one way ANOVA yields statistically significant result, we accept the alternate
hypothesis (HA), which states that there are two group means that are statistically
significantly different from each other. Here it should be kept in mind that one way
ANOVA cannot tell which specific groups were statistically significantly different from
each other. To determine which specific groups are different from each other, a
researcher will have to use post hoc test.
As there is only one independent variable or factor in one way ANOVA so it is also
called single factor ANOVA. The independent variable has nominal levels or a few
ordinal levels. Also, there is only one dependent variable and hypotheses are formulated
about the means of the group on dependent variable. The dependent variable
differentiates individuals on some quantitative dimension.
One way ANOVA is considered a robust test against the assumption of normality and
tolerates the violation of this assumption. As regards the normality of grouped data, the
one way ANOVA can tolerate data that is normal (skewed or kurtotic distribution) with
96
only a small effect on I error rate. However, platykurtosis can have profound effect
when group sizes are small. This leaves a researcher with two options:
i) Transform data using various algorithms so that the shape of the distribution
becomes normally distributed. Or
ii) Choose nonparametric Kruskal-Wallis H Test which does not require the
assumption of normality. (This test is available is SPSS).
If assumption of homogeneity of variances has been violated then tow possible tests can
be run.
i) Welch test, or
ii) Brown and Forsythe test
Alternatively, Kruskal-Wallis H Test can also be used. All these tests are available in SPSS.
When we use a t-test once, there is a chance of Type I error. The magnitude of this error is
usually 5%. By running two tests on the same data we will have increased his chance of
making error to 10%. For the third administration, it will be 15%, and so on. These are
unacceptable errors. The number of t-tests needed to compare all possible means would be:
An ANOVA controls the chance for these errors so that the type I error remains at 5%
and a researcher can become more confident about the results.
97
SS Total
𝐺2
= X2
𝑁
Fig: 3 partitioning the total sum of square (SS Total) for the independent measure ANOVA
98
comparison procedures (MCPs) or sometimes, as post hoc (i.e. after the fact) tests. These
tests should be used regarded as an afterthought than a rigorous examination of pre-
specified hypotheses.
The various methods differ in how well they properly control the overall significance level
and in their relative power. Commonly used method sand their relative power is given below.
Bonferroni – It is extremely general and simple, but often not powerful.
Tucky’s – It is the best of all possible pair-wise comparisons when sample sizes are
unequal or confidence intervals are needed. It is also very good even with equal
sample sizes without confidence intervals.
Stepdown – It is the most powerful for all possible pair-wise comparisons when
sample sizes are equal.
Dunnett’s – It is suitable for comparing one sample to each of the others, but not
comparing the others to each other.
Hsu’s MCB – It compares each mean to the best of the other means.
Scheffè’s – It is suitable for unplanned contrasts among sets of means.
8.6 Activities
1. Suppose you have to see the difference between three groups. Discuss with your
colleague and select appropriate statistical test.
2. In your study, the treatment you used had no effect. What will be the F-ratio?
99
8.7 Bibliography
Deitz, T., & Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley_-Blackwell
Fraenkel, J. R., Wallen, N. E., & Hyun, H. H. (2012). How to Design and Evaluate in
Education. (8th Ed.) McGraw-Hill, New York
Pallant, J. (2005). SPSS Survival Manual – A Step by Step Guide to Data Analysis Using
SPSS for Windows (Version 12). Australia: Allen & Unwin.
100
UNIT-9
INFERENTIAL STATISTICS:
CHI-SQUARE(X2)
Written By:
Prof. Dr. Nasir Mahmood
Reviewed By:
Dr. Rizwan Akram Rana
101
Introduction
The chi-square (χ2) statistics is commonly used for testing relationship between
categorical variables. It is intended to test how likely it is that an observed difference is
due to chance. In most situations it can be used as a quick test of significance. In this unit
you will study this important technique in detail.
Objectives
After reading this unit you will be able to
1. Explain chi-square (x2) Distribution.
2. Describe uses of chi-square (x2) distribution.
3. Explain types of Chi-square (x2) distribution.
Let us consider a random sample taken from a normal distribution. The chi-square
distribution is the distribution of the sum of these random samples squared. The degrees
of freedom (say k) are equal to the number of samples being summed. For example, if 10
samples are taken from the normal distribution, then degree of freedom df = 10. Chi-
square distributions are always right skewed. The greater the degree of freedom, the more
the chi-square distribution looks like a normal distribution.
102
9.1.2 What is a Chi-Square Statistic?
A Chi-Square Statistic is one way to a relationship between two categorical (non-
numerical) variables. The Chi-Square Statistic is a is a single number that tells us how
much difference exists between the observed counts and the counts that one expects if
there is no relationship in the population.
There are two different types of chi-square tests, both involve categorical data. These are:
a) A chi-square goodness of fit test, and
b) A chi-square test of independence.
In the coming lines these tests will be dealt in some details.
In chi-square goodness of fit test sample data is divided into intervals. Then, the numbers
of points that fall into the intervals are compared with the expected numbers of points in
each interval. . The null hypothesis for the chi-square goodness of fit test is that the data
does not come from the specified distribution. The alternate hypothesis is that the data
comes from the specified distribution. The formula for chi-square goodness of fit test is:
χ2 = ∑
=∑
103
a) The chi-square test can only be used to put data into classes. If there is data that
have not been put into classes then it is necessary to make a frequency table of
histogram before performing the test.
b) It requires sufficient sample size in order for chi-square approximation to be valid.
For the chi-square goodness of fit test, the hypotheses take the form:
H0 : The data are not consistent with a specified distribution.
Ha : The data are consistent with a specified distribution.
The null hypothesis (H0) specifies the proportion of observations at each level of the
categorical variable. The alternative hypothesis (Ha) is that a least one of the specified
proportion is not true.
v) Reach a conclusion
We need a p-value less than the significance level, generally less than 5% (p < .05),
to reject the null hypothesis. It is suitable to write a sentence in the context of the
question, i.e. “the data appears to follow a normal distribution”
104
9.3 Chi-Square Independence Test
A chi-square (χ2) test of independence is the second important form of chi-square tests. It
is used to explore the relationship between two categorical variables. Each of these
variables can have two of more categories.
The chi-square test of independence being a non-parametric technique follow less strict
assumptions, there are some general assumptions which should be taken care of:
i) Random Sample - Sample should be selected using simple random sampling
method.
ii) Variables - Both variables under study should be categorical.
iii) Independent Observations – Each person or case should be counted only once and
none should appear in more than one category of group. The data from one subject
should not influence the data from another subject.
iv) If the data are displayed in a contingency table, the expected frequency count for
each cell of the table is at least 5.
Both the chi-square tests are sometime confused but they are quite different from each other.
The chi-square test for independence compares two sets of data to see if there is
relationship.
The chi-square goodness of fit test is to fit one categorical variable to a distribution.
105
9.4 Self-Assessment Questions
Q. 1 What is chi-square (χ2) distribution?
Q. 2 What are the uses of chi-square (χ2) distribution?
Q. 3 What is a chi-square (χ2) statistics?
Q. 4 What do you know about chi-square (χ2) goodness of fit test?
Q. 5 Write down the procedure for goodness of fit test.
Q. 6 When will you use chi-square (χ2) goodness of fit test?
Q. 7 Write down the basic framework of goodness of fit test.
Q. 8 What is chi-square (χ2) independence test?
9.5 Activities
1. Make a list of multiple comparison procedures.
2. Make a list of steps of using ANOVA.
106
9.6 Bibliography
Agresti, A. & Finlay, B. (1997). Statistical Methods for Social Sciences, (3rd Ed. ).
Prentice Hall.
Bartz, A. E. (1981). Basic Statistical Concepts (2nd Ed.). Minnesota: Burgess Publishing
Company
Deitz, T., & Kalof, L. (2009). Introduction to Social Statistics. UK: Wiley_-Blackwell
Fraenkel, J. R., Wallen, N. E., & Hyun, H. H. (2012). How to Design and Evaluate in
Education. (8th Ed.) McGraw-Hill, New York
Gay, L. R., Mills, G. E., & Airasian, P. W. (2010). Educational Research: Competencies
for Analysis and Application, 10th Edition. Pearson, New York USA.
Gravetter, F. J., & Wallnau, L. B. (2002). Essentials of Statistics for the Behavioral
Sciences (4th Ed.). Wadsworth, California, USA.
Pallant, J. (2005). SPSS Survival Manual – A step by step guide to data analysis using
SPSS for Windows (Version 12). Australia: Allen & Unwin.
107