What Are Examples of Variables in Research
What Are Examples of Variables in Research
What Are Examples of Variables in Research
IN RESEARCH?
OCTOBER 22, 2012 REGONIEL, PATRICK A. 35 COMMENTS
In the course of writing your thesis, one of the first terms that you encounter
is the word variable. Failure to understand the meaning and the usefulness
of variables in your study will prevent you from doing good research. What
then are variables and how do you use variables in your study? I explain the
concept below with examples of variables commonly used in research.
You may find it difficult to understand just what variables are in the context
of research especially those that deal with quantitative data analysis. This
initial difficulty about variables becomes much more confusing when you
encounter the phrases dependent variable and independent variable as
you go deeper in studying this important concept of research as well as
statistics.
Understanding what variables mean is crucial in writing your thesis
proposal because you will need these in constructing yourconceptual
framework and in analyzing the data that you have gathered. Therefore, it is
a must that you should be able to grasp thoroughly the meaning of variables
and ways on how to measure them. Yes, the variables should be measurable
so that you will be able to use your data for statistical analysis.
I will strengthen your understanding by providing examples of phenomena
and their corresponding variables below.
how complex phenomena can be broken down into manageable pieces for
better understanding and to subject the phenomena to research.
Phenomenon: Fishkill
Examples of variables related to fish kill:
1.
dissolved oxygen
2.
water salinity
3.
temperature
4.
age of fish
5.
presence or absence of parasites
6.
presence or absence of heavy metal
7.
stocking density
Phenomenon: Poor crop growth
heart rate. We know as the result of measurement research that the HR is a valid way
of predicting the change in energy sources.
The same would be true with a tennis test. The most valid way of assessing playing
ability would perhaps be to have skilled observers watch players and give ratings. A
tennis test that any coach could administer would be much more convenient.
However, in creating this test we would be wise to be sure that the results compared
similarly with the observers' ratings. Having found that concurrent validity exists we
could claim that our test was valid.
Predictive validity refers to the validity of a measurement to be used for the prediction
of future performance. The GRE is for example is used as a measure of predicting
future college success. Maybe health educators would like to predict the likelihood of
future drug dependency of elementary aged children. To attempt to find a valid
measure we would need to select several criteria then through the use of correlational
statistics examine the relationship. The major question (that statistics help answer) is
whether our measure has much predictive value.
If successful in finding measures that are proved to valid in terms of prediction, we
would like to think that our measure could be used elsewhere. What happens however
is that the validity tends to reduce when used with a different sample. This
phenomenon is know as shrinkage.
Construct validity is a concern when we attempt to measure something that is not
observable but which we attempt to infer. We do this all the time with concepts such
as intelligence, anxiety, arousal, learning, attitude, etc. To validate tests of these
variables we again need some type of comparison and usually in relation to an
observable behavior. For example, past assessments of a person's teaching
effectiveness have often been made by observers simply watching a teaching episode,
taking notes, and writing up a critique. Such assessments may or may not be valid.
They tend to be very subjective and often two observers will focus on different
aspects thereby producing sometimes contradictory evidence. More recently, lots of
assessment tools have been developed that direct observers to record specific,
observable behaviors, e.g. time spent organizing, time spent managing, # of feedback
statements, # of student names used. Based on these types of observations we attempt
to infer "teaching effectiveness" even though teaching effectiveness is not any clearly
observable behavior.
Measurement Reliability
Be sure you can distinguish between validity and reliability. Validity is whether or not
a measurement is really measuring the item of interest. In contrast, reliability focuses
that Sally was first in a quiz and Bill was second gives the rank order but not any
measurement of the difference between Sally and Bill's scores.
3. Interval - this measure provides both an order and shows the size of the difference
between scores. For example if we count the number of pull ups students can do we
can determine a rank order and also know that 2 pull ups is twice as many as 1.
4. Ratio - these scores have similar qualities to others but also have a true zero. Force,
time, and distance all have zero points.
Standard Scores
Often we want to compare scores on one measure to scores on a different measure.
This is impossible to do unless we can convert the two scores to a similar scale. In
other words you can't compare apples and oranges unless you convert the apple to the
orange - that sounds more confusing than I anticipated!
Anyway, there are ways to convert scores and the z score and T scale represent
commonly used standard scores.
Measuring Movement
As noted in the text, in PE especially we are often concerned with measuring
movement. The main point to appreciate is that it is important that we should always
be concerned about the validity and reliability of such measures.
Measuring Written Responses
Another common area of PEHL measurement. We always seem to be interested in
items such as attitudes, self-concept, anxiety, stress, motivation, communication and
so on.
The biggest problem faced in this area is defining the behavior we wish to measure
and then producing a valid and reliable measure. The main difficulty is that the quality
described by these words (attitude, self-concept, etc.) is rather intangible. In fact it
only exists to the extent that we define it. If we change our definition we have in
effect changed the quality itself!
It is for these reasons that I strongly advise graduate students to use preexisting
measures that have been proven to be somewhat valid and reliable whenever possible,
rather than attempt to design their own measurements.
Lots of energy has been devoted to research on attitudes and personality. Do athletes
have special personality characteristics in comparison to non-athletes? Does athletics
develop these characteristics or are certain personalities attracted to athletics? Are
there differences between sports? Do sports develop character or characters? You can
apply these and lots more examples to your own area of specialization.
Four commonly used measurement scales include the:
Likert Scale
Semantic Differential
Thurstone-type Scale
Rating Scales
The Likert Scale typically involves a 5-7 point scale on which subjects respond
according to levels of agreement. For example:
"I have enjoyed and learned a lot from participating in my graduate research methods
class."
|_________________|____________|_________|______________|
Strongly Agree Agree Undecided Disagree Strongly Disagree
This scale give a wider choice of expression than just yes/no
The Semantic Differential Scale uses bipolar adjectives (e.g. beautiful-ugly, skilledunskilled, supportive-critical), at the end of a 7 point scale. Subjects score 7 for the
most positive and 1 for the least positive.
In the Thurstone Scale subjects express agreement or disagreement with a written
statement. For example:
"PEHL 557 should be a 4 credit class."
These are harder to construct because they involved the use of judges in weighting
each statement for use in scoring.
Rating Scales are frequently used in research (e.g. Borg's Rating of Perceive Exertion,
ALT-PE scales, and many more). As pointed out in the text, when "experts" are
involved in ratings various types of inconsistencies sometimes emerge.
Leniency = overgenerous
Central tendency = tendency to grade everyone as average
Halo effect = use of prior knowledge about a subject can influence judgment
Proximity errors = is concerned with the location of the rating criteria on the rating
sheet
Observer bias = personal biases the judges may have
Observer expectation = based on a person's knowledge of the experimental
arrangements the rater may exhibit different expectations.
If possible it is better to create evaluation devices that reduce the need for subjective,
value judgments and increase objective measurements. This trend had occurred in
evaluating teaching effectiveness. For example, we now will count specific behaviors
exhibited by teachers rather than try to judge whether the behavior is good or bad. If I
told you that you said "um" forty times during your 5-minute presentation you would
probably conclude a need to improve communication without me having to say that I
think your communication skills rate a 3 on a 5-point rating scale.
Measuring Knowledge
Whenever we take tests or give our students tests we would like to believe that the
questions we are posing a valid measurements of their knowledge. Sometimes
subjects don't have the opportunity to express their concerns to the test creator.
Fortunately, it is possible for test creators to objectively evaluate the validity of their
own measurements.
Item difficulty is a way of assessing the value of a question. If everyone answers a
question correctly, the thought arises as to whether there is any point including the
question as a measure. Think about this...maybe we actually do want everyone to
answer the question correctly...or maybe we want to differentiate between the level of
knowledge of our students.
Anyway, as explained in the text we can calculate a difficulty index and learn that
many test makers will eliminate questions with a difficulty index below .10 or above .
90
Item discrimination is a way of learning how well our tests discriminate between high
achievers and low achievers. Many test makers strive for discrimination indexes of .20
or higher for each question.
Selecting the correct type from the different research methods can be a little
daunting, at first. There are so many factors to take into account and evaluate.
The research question, ethics, budget and time are all major considerations in anydesign.
This is before looking at the statistics required, and studying the preferred methods for the individual
scientific discipline.
Every experimental design must make compromises and generalizations, so the researcher must try
to minimize these, whilst remaining realistic.
For pure sciences, such as chemistry or astrophysics, experiments are quite easy to define and will,
usually, be strictly quantitative.
For biology, psychology and social sciences, there can be a huge variety of methods to choose from,
and a researcher will have to justify their choice. Whilst slightly arbitrary, the best way to look at the
various methods is in terms of strength.
In other fields of study, which do not always have the luxury of definable and quantifiable variables you need to use different research methods. These should attempt to fit all of the definitions of
repeatability or falsifiability, although this is not always feasible.
Case studies are often used as a pre-cursor to more rigorous methods, and avoid the problem of the
experiment environment affecting the behavior of an organism. Observational research methods are
useful when ethics are a problem.
Conclusion
In an ideal world, experimental research methods would be used for every type of research, fulfilling
all of the requirements of falsifiability and generalization.
However, ethics, time and budget are major factors, so any experimental design must make
compromises. As long as a researcher recognizes and evaluates flaws in the design when choosing
from different research methods, any of the scientific research methods are valid contributors to
scientific knowledge.
July18
Primary data: Data which are collected by researcher itself, said to be primary data. This data
collection method is more authoritative as it is not collected by third party. This data provide raw
form of information that can be tailored according to the need of researcher. As this data is collected
by researcher, this makes it more expensive and time consuming than secondary method (Boba,
2005).
This
includes:
Observation method: This method is concerned with behavior. In this method behavior is
recorded and observed systematically by researcher. Researcher can gather detailed information by
this but its also time consuming method. In this method, researcher does not change behavior but
record
it as
it
occurs.
This
method
is not
flexible
as
survey
method.
Questionnaire: This method used in survey. In this method, research related question format
distributed by mail or internet. By this method, researcher is able to collect data from wide
geographical area. This method is cost effective and easy to manage but also it is time consuming.
Questionnaire can be open-ended and close ended. In open ended questionnaire, alternative
responses not mention and in other alternative responses are provided to respondents. This method
requires
only
literate
respondents
so
it
creates
barriers
for
this.
Interview: In this data collection method, researcher collects data by communication with
respondents. It can be through personal meeting or vie telephone. Interview format depends upon
information quality and quantity of data that researcher requires for research. Researcher must clear
about purpose of research before designing the interview question. Each question relates to research
problem. Through this method, research able to access nonverbal behavior and gets immediate
feedback. On the other hand, only small number of respondents can access and also it is time
consuming
method.
Secondary data: Data which are not collected or gathered by researcher himself or herself, its
termed as secondary data. This type of data has previously collected by someone else for some other
purpose (Hodges and Videto, 2005). There are two benefit of this data collection method. This
method is less expensive and less time consuming. Through this method, data can be obtained easily
and quickly but its not authoritative. On the other hand, this data is not fit according to the need of
researcher as it is collected by third party for their own purposes. Books and periodicals,
Government sources, Regional Publications, commercial sources, media sources and selected
internet sites that provide financial data are some example of secondary data sources (Zikmund,
2009).