Reading Guide For Chapter 1 Through 3
Reading Guide For Chapter 1 Through 3
Reading Guide For Chapter 1 Through 3
Other psychology students may not want to work in a lab, but they do enjoy reading
about the structure of the brain, the behavior of dolphins or monkeys, the personalities
of their fellow students, or the behavior of children in a school setting. They are
interested in being consumers of research information—reading about research
so they can later apply it to their work, hobbies, relationships, or personal growth.
Why is the consumer role important (be sure your answer includes evidence-based
treatments)?
Even though they may produce some research during their undergraduate years, most
psychology majors do not eventually become full-time researchers. Regardless of the
career you choose, however, becoming a savvy consumer of information is
essential. You will need to develop the ability to read about research with
curiosity and a critical eye.
Think about how often you encounter news stories or look up information on the
Internet. Much of the time, the stories you read and the websites you visit will present
information based on data. For example, during an election year, Americans may come
across polling information in the media almost every day. Many online sources have
science sections that include stories on the latest research. Entire websites are dedicated
to psychology-related topics, such as treatments for autism, subliminal learning tapes, or
advice for married couples. While some of the research is accurate and useful, some of it
is dubious, and some is actually fake. How can you tell high-quality research information
when you see it? A course in research methods enables you to ask the
appropriate questions so you can evaluate information correctly.
Finally, being a smart consumer of research could be crucial to your future career.
Whether your goal is to be a social worker, a teacher, a sales representative, a human
resources professional, an entrepreneur, or a parent, you will need to know how to
interpret research data with a critical eye. Clinical psychologists, social workers,
and family therapists read research to know which therapies are the most effective. In
fact, obtaining a license in these helping professions requires knowing the
research behind evidence-based treatments—that is, therapies that are
supported by research. Teachers also use data to find out which methods work best.
And the business world depends on quantitative information: Research is used to predict
what sales will be like in the future or what consumers will buy.
2. How do the findings regarding the Scared Straight program illustrate the benefits of
being a good consumer of research?
The Scared Straight program is another commonsense story that turned out to be wrong.
As you read in Chapter 1, such programs propose that when teenagers susceptible to
criminal activity hear about the difficulties of prison from actual inmates, they will be
scared away from committing crimes in the future. It certainly makes sense that
impressionable young people would be frightened and deterred by such stories.
However, research has consistently found that Scared Straight programs are
ineffective; in fact, they sometimes even cause more crime. The intuitive appeal
of such programs is strong (which explains why some communities still invest in them),
but the research warns against them. One psychologist estimated that the widespread
use of the program in New Jersey might have “caused 6,500 kids to commit crimes they
otherwise would not have committed” (Wilson, 2011, p. 138). Faulty intuition can
even be harmful.
Define empiricism.
Answering psychological questions with direct, formal observations.
A hypothesis, or prediction, is stated in terms of the study design. It’s the specific
outcome the researcher will observe in a study if the theory is accurate.
Harlow’s hypothesis stated that the babies would spend more time on the cozy mother
than on the wire mother. Notably, one theory can lead to a large number of hypotheses
because a single study is not sufficient to test the entire theory—it is intended to test only
part of it. Most researchers test theories with a series of empirical studies, each designed
to test an individual hypothesis.
Data are a set of observations. (Harlow’s data were the amount of time the baby
monkeys stayed on each mother.) Depending on whether the data are consistent with
hypotheses based on a theory, the data may either support or challenge the theory. Data
that match the theory’s hypotheses strengthen the researcher’s confidence in the theory.
When the data do not match the theory’s hypotheses, however, those results indicate that
the theory needs to be revised or the research design needs to be improved.
Ideally, hypotheses are preregistered. That is, after the study is designed but
before collecting any data, the researcher states publicly what the study’s
outcome is expected to be (perhaps using a time-stamped Internet tool). It’s
unconvincing if a researcher collects the data first and then claims, “I knew that would
happen all along.”
What are the components of the theory-data cycle (hint: see figure 1.5)
This experiment sets up three possible outcomes. The contact comfort theory would be
supported if the babies spent most of their time clinging to the cloth mother. The
cupboard theory would be supported if the babies spent most of their time clinging to the
wire mother. Neither theory would be supported if monkeys divided their time equally
between the two mothers. In setting up this experiment, Harlow purposely created a
situation that might prove his theory wrong.
When Harlow put the baby monkeys in cages with the two mothers, the evidence in favor
of the contact comfort theory was overwhelming.
Harlow’s data showed that the little monkeys would cling to the cloth mother for 12–18
hours a day (Figure 1.4). When they were hungry, they would climb down, nurse from
the wire mother, and then go right back to the warm, cozy cloth mother. In short, Harlow
used the two theories to make specific predictions about how the monkeys would interact
with each mother. The data he recorded (how much time the monkeys spent on each
mother) supported only one of the theories. The theory-data cycle in action!
7. Define falsifiability.
A theory should lead to hypotheses that, when tested, could fail to support the
theory—in other words, falsifiability is a characteristic of good theories (Popper, 1963).
Read Figure 1.7 on page 15 and discuss how the tinfoil hat study does not allow for
falsifiability?
The tinfoil hat study is an example of a theory that is not falsifiable. Certain people might
wear a tinfoil hat, operating under the idea that the hat wards off government mental
surveillance. But like most conspiracy theories, this notion of remote government
mindreading is not falsifiable. If the government has been shown to read
people’s minds, the theory is supported. But if there is no physical evidence,
that also supports the theory because if the government does engage in such
surveillance, it wouldn’t leave a detectable trace of its secret operations.
How does the process of peer-review continue even after a study is published?
Other scientists can cite an article and do further work on the same subject. Moreover,
scientists who find flaws in the research (perhaps overlooked by the peer reviewers) can
publish letters, commentaries, or competing studies.
11. How does the “Mozart effect” example demonstrate how media coverage can
misrepresent science when they write for a popular audience.
In 1993, researcher Frances Rauscher found that when students heard Mozart music
played for 10 minutes, they performed better on a subsequent spatial intelligence test
when compared with students who had listened to silence or to a monotone speaking
voice (Rauscher et al., 1993). Rauscher said in a radio interview, “What we found was
that the students who had listened to the Mozart sonata scored significantly higher on
the spatial temporal task.” However, Rauscher added, “It’s very important to note that
we did not find effects for general intelligence . . . just for this one aspect of intelligence.
It’s a small gain and it doesn’t last very long” (Spiegel, 2010).
But despite the careful way the scientists described their results, the media that reported
on the story exaggerated its importance:
The headlines in the papers were less subtle than her findings: “Mozart makes
you smart” was the general idea. . . . But worse, says Rauscher, was that her very
modest finding started to be wildly distorted. “Generalizing these results to
children is one of the first things that went wrong. Somehow or another the myth
started exploding that children that listen to classical music from a young age will
do better on the SAT, they’ll score better on intelligence tests in general, and so
forth.” (Spiegel, 2010).
1. What is the Abstract section of an empirical journal article and what does it describe?
The abstract is a concise summary of the article. It briefly describes the study’s
hypotheses, method, and major results. When you are collecting articles for a
project, the abstracts can help you quickly decide whether each article describes the kind
of research you are looking for, or whether you should move on to the next article.
2. What does the Introduction section include and where can each of these be found in this
section?
The introduction is the first section of regular text, and the first paragraphs typically
explain the topic of the study. The middle paragraphs lay out the background
for the research. What theory is being tested? What have past studies found? Why is
the present study important? Pay special attention to the final paragraph, which states
the specific research questions, goals, or hypotheses for the current study.
3. What does the Method section explain in detail? An ideal Method section provides
enough detail to do what?
The Method section explains in detail how the researchers conducted their study.
It usually contains subsections such as Participants, Materials, Procedure, and
Apparatus. An ideal Method section gives enough detail that if you wanted to repeat
the study, you could do so without having to ask the authors any questions.
5. What does the Discussion section include and where can each of these components be
found?
The opening paragraph of the Discussion section generally summarizes the
study’s research question and methods and indicates how well the results of
the study supported the hypotheses. Next, the authors usually discuss the
study’s importance: Perhaps their hypothesis was new, or the method they used was a
creative and unusual way to test a familiar hypothesis, or the participants were unlike
others who had been studied before. In addition, the authors may discuss alternative
explanations for their data and pose interesting questions raised by the
research.
“Perseverance” (in young “The ability to push through How long a child will choose
children) when confronted with . . to engage in a slow-paced,
.obstacles” (White et al., boring activity that involves
2017) pressing a button when they
see a picture of cheese on a
screen and not pressing it
when they see a cat on the
screen.
4. What is the difference between a conceptual variable and the operational definition of a
variable?
Each variable in a study can be referred to in two ways. When researchers are discussing
their theories and when journalists write about their research, they use more abstract
names, called constructs or conceptual variables. When testing hypotheses with
empirical research, they create operational definitions of variables, also known as
operational variables, or operationalizations. To operationalize a concept of
interest means to turn it into a measured or manipulated variable.
How might the conceptual variables “level of eye contact,” “intelligence,” and “stress” be
operationalized by a researcher?
How long (in seconds) do participants maintain eye contact?
What is the score on a standard intelligence test?
What are participants’ ratings on a stress scale?
5. How many variables are there in a frequency claim? An association claim? A causal
claim?
A frequency claim describes a particular rate or degree of a single variable. An
association claim is about two variables, in which the value (level) of one variable is
said to vary systematically with the value of another variable. A causal claim argues that
a specific change in one variable is responsible for influencing the value of
another variable.
6. Which part of speech in a claim can help you differentiate between association and
causal claims?
Causal claims, however, go beyond a simple association between the two variables. They
use language suggesting that one variable causes the other — verbs such as cause,
enhance, affect, decrease, and change. In contrast, association claims use verbs such as
link, associate, correlate, predict, tie to, and be at risk for.
7. How are causal claims special, compared with the other two claim types?
Whereas an association claim merely notes a relationship between two variables, a
causal claim goes even further, arguing that one of the variables is
responsible for changing the other.
Construct How well has the How well has the How well has the
researcher measured researcher measured researcher measured
the variable in each of the two or manipulated the
question? variables in the variables in the
association? study?
Covariance The study’s results show that as A changes, B changes; e.g., high levels of
A go with high levels of B, and low levels of A go with low levels of B.
Temporal precedence The study’s method ensures that A comes first in time, before B.
Internal validity The study’s method ensures that there are no plausible alternative
explanations for the change in B; A is the only thing that changed.
Practice Activity: For each headline (a, b, and c) in Learning Actively #2 on p. 84, please do
the following:
1. Identify the type of claim being made
2. Based on the type of claim being made, which validities apply? (Hint: see table 3.6 for
quick reference on which validities will apply)
3. For construct validity applies, what questions would you ask? For example: How did they
operationalize “mood” and “focus”?
4. For external validity applies, what questions would you ask? For example: From what
populations did they sample “workaholics”?
1. Define population.
A population is the entire set of people or products in which you are
interested.
2. Define sample.
The sample is a smaller set, taken from that population.
3. What does the external validity of a study concern?
The external validity of a study concerns whether the sample used in the study is
adequate to represent the unstudied population.
4. Before researchers can decide whether a sample is biased, what do they have to specify?
Before researchers can decide whether a sample is biased or unbiased, they have to
specify a population to which they want to generalize: the population of
interest.
5. Samples are either biased or representative.
6. What is meant by a biased sample?
In a biased sample, also called an unrepresentative sample, some members of the
population of interest have a much higher probability than other members
of being included in the sample.
7. What is meant by an unbiased sample?
In an unbiased sample, also called a representative sample, all members of the
population have an equal chance of being included in the sample.
8. What are different ways to get a biased sample?
A sample could be biased in at least two ways: Researchers might study only those
they can contact conveniently or only those who volunteer to respond.
9. What is meant by a convenience sample?
Using a sample of people who are easy to contact and readily available to
participate.
10. What is meant by self-selection (NOT selection effects)?
A term used when a sample is known to contain only people who volunteer to
participate. Self-selection can cause serious problems for external validity.
11. What is meant by probability sampling? Does probability sampling help get an unbiased
or representative sample?
In probability sampling, also called random sampling, every member of the
population of interest has an equal and known chance of being selected for
the sample, regardless of whether they are convenient or motivated to volunteer.
Therefore, probability samples have excellent external validity and can generalize
to the population of interest.
12. What is the most basic form of probability sampling?
The most basic form of probability sampling is simple random sampling. To visualize
this process, imagine that the name of each member of the population of interest is
written on a plastic ball. The balls are rolled around in a bowl, and then a mechanism
spits out a number of balls equal to the size of the desired sample. The people whose
names are on the selected balls will make up the sample.
13. What is cluster sampling?
Cluster sampling is an option when people are already divided into arbitrary groups.
Clusters of participants within a population of interest are randomly selected, and then
all individuals in each selected cluster are used. If a researcher wanted to randomly
sample high school students in the state of Pennsylvania, for example, he could start with
a list of the 952 public high schools (clusters) in that state, randomly select 100 of those
high schools (clusters), and then include every student from each of those 100 schools in
the sample.
14. What is multi-stage sampling?
In multi-stage sampling, two random samples are selected: a random sample of clusters
and then a random sample of people within those clusters. In the high school example,
the researcher would start with a list of high schools (clusters) in the state and select a
random 100 of those schools. Then, instead of including all students at each school, the
researcher would select a random sample of students from each of the selected schools.
15. Distinguish between random sampling and random assignment.
In the context of research methods, it’s important not to confuse random sampling and
random assignment. With random sampling (probability sampling), researchers create a
sample using some random method, such as drawing names from a hat or using a
random-digit phone dialer, so that each member of the population has an equal chance
of being in the sample.
Random sampling enhances which validity?
Random sampling enhances external validity.
What is random assignment?
Random assignment is used only in experimental designs. When researchers want to
place participants into two different groups (such as a treatment group and a comparison
group), they usually assign them at random.
Which validity does it help enhance?
Random assignment enhances internal validity by helping ensure that the comparison
group and the treatment group have the same kinds of people in them, thereby
controlling for alternative explanations.
16. Samples obtained through random selection achieve excellent external validity, but
such samples can be difficult to obtain.
17. Are larger samples more representative?
For external validity, is a bigger sample always a better sample? The answer may surprise
you: not really. The idea that larger samples are more externally valid than smaller
samples is perhaps one of the most persistent misconceptions in a research methods
course.
■ Experiments (a study in which at least one variable is manipulated and another is measured)
are the only way to investigate causal relationships between variables. ■ In the “Taking Notes”
study, the manipulated variable was the notetaking method (laptop/longhand). In the
“Motivating Babies” study, the manipulated variable was the model that the babies
watched. In the “Taking Notes” study, the measured variable was the student performance
on the essay questions. In the “Motivating Babies” study, the measured variable was how
many times each baby pressed the inert button. ■ Independent variable (manipulated
variable) is called “independent” because the researcher has some independence in assigning
people to different levels (conditions) of this variable. An experiment must have at least one
manipulated (independent) variable. ■ Dependent variable (measured variable) is called the
“dependent” variable because how a participant acts on the measured variable depends on the
level of the independent variable. ■ Any variable that an experimenter holds constant on
purpose is called a control variable. Some of the control variables used in the Mueller and
Oppenheimer (2014) study were: People in both groups watched lectures in the same room and
had the same experimenter. They watched the same videos and answered the same questions
about them. Some of the control variables in the Leonard, et al. study (2017) were: The toys the
model was using. The model used the same cheerful, enthusiastic voice. The researchers also
kept constant how long the model demonstrated each toy (30 seconds), the gender of the model
(always female), the chair the infant sat in, and the cubicle where the experiment took place.
Control variables allow researchers to separate one potential cause from another and are
therefore important for establishing internal validity. ■ The three rules for causal claims are:
(1) Covariance. Do the results show that the causal variable is related to the outcome variable?
Are distinct levels of the independent variable associated with different levels of the dependent
variable. (2) Temporal precedence. Does the study design ensure that the causal variable
comes before the outcome variable in time? (3) Internal validity. Does the study design rule
out alternative explanations for the results? ■ The study on hugging dogs was flawed because
there was no comparison group. A control group is a level of an independent variable that
is intended to represent “no treatment” or a neutral condition. When a study has a control
group, the other level or levels of the independent variable are usually called the treatment
group(s). When the control group is exposed to an inert treatment such as a sugar pill, it is
called a placebo group. Not all experiments need a control group. However, all experiments
do need a comparison group. ■ In experiments, the experimenter can make sure that the
causal (independent) variable comes first. Internal validity is one of the most important
validities to interrogate when you encounter a causal claim. For a study to be internally valid,
the study must ensure that the causal variable, and not other factors, is responsible
for the change in the outcome variable. Confounds are possible alternative explanations.
The word confound means confuse. ■ When the kinds of participants in one level of the
independent variable are systematically different from those in the other, selection effects can
result. They can also happen when the experimenters let participants select which group they
want to be in. Well-designed experiments often use random assignment to avoid selection
effects. It is very unlikely that all the focused or fussy babies would have been clustered into the
same group in the baby study because assigning participants at random to different levels
of the independent variable (by flipping a coin, rolling a die, or using a random number
generator) controls for all sorts of potential selection effects. Random assignment creates
a situation in which the experimental groups will become virtually equal before the
independent variable is applied. ■ To create matched groups from a sample of 30, the
researchers would first measure the participants on a particular variable that might matter to
the dependent variable. Student achievement, operationalized by GPA, for instance, might
matter in the study of notetaking. The researchers would next match up participants in pairs,
starting with the two having the highest GPAs, and within that matched set, randomly assign
one of them to each of the two notetaking conditions. They would then take the pair with the
next-highest GPAs and within that set again assign randomly to the two groups. They would
continue this process until they reach the participants with the lowest GPAs and assign them at
random too. ■ In an independent-groups design (also called a between-subjects design
or between-groups design) separate groups of participants are placed into different levels of
the independent variable. Two basic forms of independent-groups designs are the
posttest-only design (participants are tested on the dependent variable once) and the
pretest/posttest design (participants are tested on the dependent variable both before and
after exposure to the independent variable). ■ In a within-groups design, or
within-subjects design, each person is presented with all levels of the independent variable.
A repeated-measures design is one in which participants are measured on a dependent variable
more than once, after exposure to each level of the independent variable. In a
concurrent-measures design, participants are exposed to all the levels of an independent
variable at roughly the same time, and a single attitudinal or behavioral preference is the
dependent variable. The main advantage of a within-groups design is that it ensures the
participants in the two groups will be equivalent. Another advantage is that it requires fewer
participants overall.
____________________________________________________
Three types of measures researchers must choose between when operationalizing measured
variables: self-report, observational, and physiological. ◼ The conceptual definition,
or construct, of a variable is the researcher’s definition of the variable in question at a
theoretical level. The operational definition of a measured variable represents a researcher’s
specific decision about how to measure or manipulate the conceptual variable. ◼ Ed Diener
conceptually defined happiness as “subjective well-being” (well-being from a person’s own
perspective) and he operationally defined it by asking people to respond to a five item
questionnaire using a 7-point scale. ◼ The general process researchers go through to study
conceptual variables is that they first state a definition of their construct and then they create an
operational definition. ◼ A self-report measure operationalizes a variable by recording
people’s answers to questions about themselves in a questionnaire or interview. ◼ An
observational measure, sometimes called a behavioral measure, operationalizes a variable
by recording observable behaviors or physical traces of behaviors. ◼ A physiological
measure operationalizes a variable by recording biological data, such as brain activity,
hormone levels, or heart rate. ◼ No one method of operationalization is the best. Researchers
normally expect all three measures to show similar patterns of results. ◼ Categorical (also
called nominal) variables are categories and quantitative (also called continuous) variables are
coded with meaningful numbers. ◼ Reliability refers to how consistent the results of a
measure are. Validity refers to whether the operationalization is measuring what it is supposed
to measure. ◼ Reliability can be assessed in three ways: test-retest reliability (a study
participant will get pretty much the same score each time they are measured with it),
interrater reliability (consistent scores are are obtained no matter who measures the
variable), and internal reliability, also called internal consistency (a study participant gives a
consistent pattern of answers no matter how the researchers phrase the question). ◼
Researchers know if their indirect operational definition of a construct is really measuring that
construct (they evaluate a measurement tool’s validity) by collecting a variety of data and
evaluating it in light of their theory about the construct. ◼ To demonstrate the difference
between reliability and validity, consider the example of head circumference as an
operationalization of intelligence. Head size measurements are usually reliable because
circumference is easy to measure. However, head circumference is not related to intelligence.
Therefore, the head circumference test may be reliable, but it is not valid as an intelligence test:
It does not adequately capture the construct of intelligence.
____________________________________________________
◼ Replicable, or reproducible (will they get the same results if they conduct the same study
again). A finding from a study must be replicable to be considered credible because otherwise it
could just be a fluke. ◼ Three major types of replications: direct replication (researchers repeat
an original study as closely as they can to see whether the effect is the same in the newly
collected data), conceptual replication (researchers explore the same research question but use
different procedures), and replication-plus-extension (researchers replicate their original
experiment and add variables to test additional questions). ◼ Replication studies can help
interrogate external validity (the degree to which a study’s results are generalizable to other
participants and other settings). ◼ When the author states, “It’s A population, not THE
population” she means that the population to which researchers want to generalize usually is not
the population of every living person. Instead, when researchers are generalizing from a sample
to a population, they will specify what the population of interest is. It might be all the college
students in Manitoba. It might be all U.S. soldiers. It might be all the third graders in a Beijing
elementary school. It might be a group of lab-reared rhesus monkeys. Researchers are at liberty
to specify what their population of interest is based on the theories they are testing. ◼ When the
author states, “External validity comes from HOW, not HOW MANY” she means that when you
are assessing the generalizability of a sample to a population, a randomly selected sample of 200
participants has external validity; a haphazardly selected sample of 2,000 participants does not.
■ Experiments (a study in which at least one V is manipulated and another is measured) are
the only way to investigate causal relationships between Vs. ■ Independent V (manipulated V)
the R has some independence in assigning people to different levels (conditions) of this V. An
experiment must have at least one IV. ■ Dependent V (measured V) how a P acts on the
measured V depends on the level of the IV. ■ Control Vs (any V that an experimenter holds
constant on purpose) allow Rs to separate one potential cause from another and are important
for establishing internal validity. ■ Three rules for causal claims are: (1) Covariance. Do the
results show that the causal V is related to the outcome V? Are distinct levels of the IV associated
with different levels of the DV? (2) Temporal precedence. Does the study design ensure that
the causal V comes before the outcome V in time? (3) Internal validity. Does the study design
rule out alternative explanations for the results?
■ Study on hugging dogs was flawed because there was no comparison group. ■ Control
group (a level of an IV that is intended to represent “no treatment” or a neutral condition).
When a study has a control group, the other level or levels of the IV are usually called the
treatment group(s). Placebo group (when the control group is exposed to an inert
treatment such as a sugar pill). Not all experiments need a control group. However, all
experiments do need a comparison group. ■ The experimenter can make sure that the causal
(independent) V comes first. Internal validity is one of the most important validities to
interrogate when you encounter a causal claim. For a study to be internally valid, the study must
ensure that the causal V, and not other factors, is responsible for the change in the
outcome V. Confounds (meaning confuse) are possible alternative explanations.
■ Selection effects (when the kinds of Ps in one level of the IV are systematically different
from those in the other). They can also happen when the experimenters let Ps select which group
they want to be in. Well-designed experiments often use random assignment to avoid selection
effects. It is very unlikely that all the focused or fussy babies would have been clustered into the
same group in the baby study because assigning Ps at random to different levels of the IV (by
flipping a coin, rolling a die, or using a random number generator) controls for all sorts of
potential selection effects. Random assignment creates a situation in which the
experimental groups will become virtually equal before the IV is applied. ■ To create
matched groups from a sample of 30, the Rs would first measure the Ps on a particular V that
might matter to the DV. The Rs would next match up Ps in pairs and within that matched set,
randomly assign one of them to each of the two conditions.
■ Within-groups AKA within-subjects design (each person is presented with all levels of the
IV) > (1) repeated-measures design (Ps are measured on a DV more than once, after
exposure to each level of the IV) and (2) concurrent-measures design (Ps are exposed to all
the levels of an IV at roughly the same time, and a single attitudinal or behavioral preference is
the DV). Advantages of a within-groups design: (1) it ensures the Ps in the two groups will be
equivalent and (2) it requires fewer Ps overall. Disadvantages of a within-groups design: (1)
order effect (when exposure to one level of IV changes P’s behavior in subsequent levels of IV: a)
Fatigue effect (deterioration in performance; P becomes tired, bored, distracted), b) Practice
effect (improvement in performance as result of repeated practice with task), c) Contrast effect
(response to second condition in experiment altered by going through first condition)) - To
address order effects, Rs can counterbalance the order of the conditions, (2) demand
characteristics (when Ps guess what the study is about and behave differently than they would
have without the guess), and (3) within-groups design might not be possible.
■ Replicable, or reproducible (will they get the same results if they conduct the same study
again). A finding from a study must be replicable to be considered credible because otherwise it
could just be a fluke. ■ Three major types of replications: direct replication (Rs repeat an original
study as closely as they can to see whether the effect is the same in the newly collected data),
conceptual replication (Rs explore the same research question but use different procedures),
and replication-plus-extension (Rs replicate their original experiment and add Vs to test
additional questions). ■ Replication studies can help interrogate external validity (the degree
to which a study’s results are generalizable to other Ps and other settings). ■ “It’s A population,
not THE population” (the population to which Rs want to generalize usually is not the
population of every living person. Instead, when Rs are generalizing from a sample to a
population, they will specify what the population of interest is). ■ “External validity comes from
HOW, not HOW MANY” (when you are assessing the generalizability of a sample to a
population, a randomly selected sample of 200 Ps has external validity; a haphazardly selected
sample of 2,000 Ps does not).
■ Survey types (phone/text, personal interviews, paper and pencil questionnaire, internet). ■
Survey question types (open-ended, forced choice AKA multiple choice, rating scale). ■
Potential threats to construct validity in survey research: question wording (leading questions:
questions that suggest an answer, double barreled questions: ask two questions in one but only
allow one response, double negatives), question order, response format issues, sensitive topics,
self-report measure).
GOOD LUCK!
1. No matter what type of claim researchers are investigating, they are obligated by law, by
morality, and by today’s social norms to treat the participants in their research
with kindness, respect, and fairness.
2. What was the time frame of the Tuskegee Syphilis study? In the late 1920s and early
1930s
3. What were the demographics of the participants (race and health status)? 600 black
men: about 400 were already infected with syphilis and about 200 were not.
4. Researchers wanted to study the effects of untreated syphilis on the men’s
health over the long term.
5. As part of the study, what did the participants experience, what were they kept from
doing? They received no treatment. They were subjected to spinal taps. They
were kept from joining the armed forces.
6. Why did the study finally end? 1972
7. Over the course of the study, what happened to the participants and their families? Over
the course of the study, many of the men got sicker, and dozens died. Several
men inadvertently infected their partners, in some cases causing congenital
syphilis in their children.
8. The ethical violations of the Tuskegee Study fall into what three categories? They were
not treated respectfully, the men in the study were harmed, and the
researchers targeted a disadvantaged social group.
9. In what year was the Belmont Report created? Started in 1976 (completed in 1979)
By the request of whom? the U.S. Congress
10. What are the three main principles for guiding ethical decision making included in the
Belmont Report? Respect for persons, beneficence, and justice.
11. What are the two provisions of the principle of respect for persons? Individuals
potentially involved in research should be treated as autonomous agents:
They should be free to make up their own minds about whether they wish to
participate in a research study. Some people have less autonomy, so they are
entitled to special protection when it comes to informed consent. What must
researchers do to apply this principle? Every participant is entitled to the
precaution of informed consent: each person learns about the research
project, considers its risks and benefits, and decides whether to participate.
Children, people with intellectual or developmental disabilities, and
prisoners should be treated with special consideration.
12. According to the principle of beneficence, researchers must take precautions to protect
against what and ensure what? To protect participants from harm and to ensure
their well-being.
13. What must researchers assess to apply this principle? The risks and benefits of the
study they plan to conduct.
14. How did the Tuskegee Syphilis Study fail to treat the participants in accordance with the
principle of beneficence? They harmed participants through risky and invasive
medical tests, they harmed the participants’ families by exposing them to
untreated syphilis, and they withheld benefits from the men in the study.
15. Today, are researchers able to withhold treatments that are known to be helpful? No
16. What’s the difference between an anonymous study and a confidential study? In an
anonymous study, researchers do not collect any potentially identifying
information. In a confidential study, researchers collect some identifying
information but prevent it from being disclosed.
17. What does the principle of justice call for? A fair balance between the kinds of
people who participate in research and the kinds of people who benefit from
it.
18. How did the Tuskegee Syphilis study violate this principle? Anybody can contract
syphilis and benefit from research on it, but the participants in the study
were all poor African American men.
19. When the principle of justice is applied, it means that researchers must consider what?
The extent to which the participants involved in a study are representative
of the kinds of people who would also benefit from its results.
20. The APA outlines five general principles for guiding individual aspects of ethical
behavior. These principles are intended to protect not only research participants but
who else? Students in psychology classes and clients of professional
therapists
21. What are the five general ethical principles of the APA ethical guidelines (2002)? Be able
to define all 5. (Hint: Three of the five are from the Belmont Report; only two principles
are new) Fidelity and responsibility: a clinical psychologist teaching in a
university may not serve as a therapist to one of his or her classroom
students, and psychologists must avoid sexual relationships with their
students or clients. Integrity: professors are obligated to teach accurately,
and therapists are required to stay current on the empirical evidence for
therapeutic techniques. Beneficence and nonmaleficence: Treat people in ways
that benefit them. Do not cause suffering. Conduct research that will benefit society.
Fidelity and responsibility: Establish relationships of trust; accept responsibility for
professional behavior (in research, teaching, and clinical practice). Integrity: Strive to
be accurate, truthful, and honest in one’s role as researcher, teacher, or practitioner.
Justice: Strive to treat all groups of people fairly. Sample research participants from the
same populations that will benefit from the research. Be aware of biases. Respect for
people’s rights and dignity: Recognize that people are autonomous agents.
Protect people’s rights, including the right to privacy, the right to give
consent for treatment or research, and the right to have participation
treated confidentially. Understand that some populations may be less able
to give autonomous consent, and take precautions against coercing such
people.
22. What does IRB stand for? Institutional Review Board
23. What is the IRB responsible for? Interpreting ethical principles and insuring that
research using human participants is conducted ethically.
24. Do all organizations that conduct research have to have an IRB? No. If an institution
uses federal money, it needs an IRB. Private research companies do not
need an IRB.
25. An IRB panel is composed of whom? (Number and backgrounds) Five or more
people. One must be a scientist, one must have academic interests outside
the sciences, one (or more) must be a community member that has no ties to
the institution. If prisoners are to be used in a study, one must be a prison
advocate.
26. Before conducting a study, researchers must do what? Fill out a detailed application
describing their study, its risks and benefits (to both participants and
society), its procedures for informed consent, and its provisions for
protecting people’s privacy. The IRB then reviews each application.
27. What must researchers include in their applications/proposals?
28. What type of research should IRBs not permit? Research that violates people’s
rights, research that poses unreasonable risk, or research that lacks a sound
rationale.
29. Ideally, IRBs should balance what? The welfare of research participants and the
researchers’ goal of contributing important knowledge to the field.
30. As part of informed consent, what is a researcher obligated to explain to potential
participants? To explain the study to potential participants in everyday
language and give them a chance to decide whether to participate.
31. When might informed consent not be necessary? If the study is not likely to cause
harm, involves a completely anonymous questionnaire, or takes place in an
educational setting, written informed consent might not be required.
Similarly, when the study involves naturalistic observation of participants in
low-risk public settings, such as a museum, classroom, or mall, people can
reasonably expect to be observed by others anyway.
32. Define deception. (Be sure your answer includes omission and commission) In some
cases, researchers withheld some details of the study from participants —
deception through omission. In other cases, researchers actively lied to
participants — deception through commission.
33. Is deception always considered unethical? No. If not, what must be considered in
determining whether or not it is ethical? What are the ethical costs and benefits of
doing the study with deception, compared to without. If deception is used, must
participants be debriefed? Yes.
34. What is included in a debriefing? The researchers describe the nature of the
deception and explain why it was necessary.
35. Do non-deceptive studies also provide a debriefing session? Sometimes. If so, what is
the purpose? At many universities, student participants in research receive a
written description of the study's goals and hypotheses, along with
references for further reading. The intention is to make participation in
research a worthwhile educational experience, so students can learn more
about the research process in general, understand how their participation
fits into the larger context of theory testing, and learn how their
participation might benefit others. In debriefing sessions, researchers might
also offer to share results with the participants.
36. What are far reaching consequences of falsified data? When people fabricate data,
they mislead others about the actual support for a theory. Fabricated data
might inspire other researchers to spend time (and often grant money)
following a false lead or to be more confident in theories than they should
be.
37. Why might researchers falsify data? In many universities, the reputations,
income, and promotions of professors are based on their publications and
their influence in the field. In high-pressure circumstances, the temptation
might be great to delete contradictory data or create supporting data. In
addition, countering Merton’s norm of disinterestedness, some researchers
become personally invested in their own hypotheses and believe that any
data that do not support their predictions must be inaccurate.
38. Research misconduct violates what two current goals of psychological science?
Openness and transparency.
39. How does open data support scientific progress? Open data upholds Merton’s norm
of communality, which states that science belongs to everyone. Open data
supports scientific progress because other scientists can replicate published
work and test their own novel hypotheses.
40. What does it mean for scientists to report their process transparently? Describing all
measured variables and statistical analyses in the Method section.
41. Define plagiarism. Representing the ideas or words of others as one’s own.
42. How does a writer avoid plagiarism? You must cite the sources of all ideas that
are not your own, to give appropriate credit to the original authors.
43. Should researchers self-plagiarize? No. Why/why not? When researchers publish
multiple articles in a line of research, they may end up recycling portions of
the Method section from previous work. But they should not repeat
sentences verbatim in the introduction or Discussion section. In a Results
section, recycling sentences is unacceptable because it means the article is
presenting previously published data as if it were new.
1. Carefully read the description of the study by Duke et al. (2011) on alcohol and
aggression. What is the study’s IV? The Duke et al. (2011) alcohol and aggression
study’s IV was level of alcohol (placebo group vs. drunk group). DV? The DV
was aggression as measured by intensity and duration of a shock delivered.
Is it a between-subjects or within-subjects design? This was a between-subjects
design. What are the study’s results? The results demonstrated that alcohol
causes people to behave aggressively.
2. Carefully read the description of the study by Strayer and Drews (2011) on the dangers of
distracted driving. What is the study’s IV? Cell phone use DV? Driving quality Is it a
between-subjects or within-subjects design? Within-subjects design What are the
study’s results? The results showed that when drivers were simply talking on
cell phones (not even texting or using apps), their reactions to road hazards
were 18% slower.
3. Adding an additional independent variable allows researchers to look for an interaction
effect (or interaction). What is an interaction? An interaction effect (or
interaction) is whether the effect of the original independent variable
depends on the level of another independent variable.
4. An interaction of two variables allows researchers to establish whether or
not “it depends.”
5. What is the mathematical way to describe an interaction between two independent
variables? Apply this to the ice cream, pancakes, hot, cold example. There is a
difference in differences.
6. Define factorial design. One in which there are two or more independent
variables.
7. What are independent variables referred to when conducting a factorial design? factors
8. Carefully review Strayer and Drews (2004) factorial design. What are the two
independent variables? cell phone use and driver age How many levels does each IV
have? Two Why is this design referred to as a 2 x 2 factorial design? How many cells are
in this study? There are two independent variables (two factors)—cell phone
use and age—and each one has two levels (driving while using a cell phone
or not; younger or older driver). This particular design is called a 2 × 2
(two-by-two) factorial design, meaning that two levels of one independent
variable are crossed with two levels of another independent variable. Since
2 × 2 = 4, there are four cells in this design.
9. What is meant by a participant variable? Age is an example of a participant
variable—a variable whose levels are selected (i.e., measured), not
manipulated.
10. For simplicity purposes, participant variables are often referred to as independent
variables/factors when conducting factorial designs. However, why aren’t participant
variables true independent variables? Because the levels are not manipulated,
variables such as age, gender, and ethnicity are not truly “independent”
variables. However, when they are studied in a factorial design, researchers
often call them independent variables for the sake of simplicity.
11. What is one reason mentioned in this section as to why researchers conduct studies with
factorial designs? To test whether an independent variable affects different
kinds of people, or people in different situations, in the same way.
12. Did the study by Strayer and Drews (2004) find the same relationship between cell
phone use and driving for young and old drivers? Yes Therefore, did the effects of cell
phone use on driving interact with, or depend on, age? No
13. Did Strayer and Drews (2004) find that the relationship between cell phone use and
driving generalized to young and old drivers? Yes
14. DeWall et al. (2010) tested whether the effect of alcohol intake on aggressive behavior
depends on body weight. Did they find an interaction between alcohol intake and body
weight? Yes In other words, did the effect of alcohol intake on aggressive behavior
depend on body weight? Yes
15. The process of using a factorial design to test limits is sometimes called testing for what?
testing for moderators
16. What is a moderator? A variable that changes the relationship between two
other variables. What does a moderator result in? An interaction The effect of one
independent variable depends on (is moderated by) the level of another independent
variable.
17. In the cell phone and age study by Strayer and Drew, did driver age moderate the impact
of cell phone use on braking onset time (driving performance)? No Did DeWall et al.
show that body weight moderates the effect of alcohol on aggression? Yes
18. The goal of most experiments in psychological science is to test hypotheses derived
from theories
19. Many theories make statements about how variables interact with one another.
20. The best way to study how variables interact is to combine them in a factorial
design and measure whether the results are consistent with the theory.
21. In a design with two independent variables, there will always be three results: 2 main
effects (one for each IV) and 1 interaction.
22. What is meant by a main effect? The overall effect of one independent variable
on the dependent variable, averaging over the levels of the other
independent variable. In other words, a main effect is a simple difference.
23. Is the main effect the most important effect in a study? No How might you think of a
main effect instead? The term main effect may be misleading because it seems
to suggest that it is the most important effect in a study. It is not. In fact,
when a study’s results show an interaction, the interaction itself is the most
important effect. Think of a main effect instead as an overall effect—the
overall effect of one independent variable at a time.
24. When results from a factorial design are plotted on a line graph and the lines are NOT
parallel (nonparallel), there may be what? An interaction This would have to be
confirmed by what? a statistical test
25. If the lines are parallel, there is probably NO what? interaction
26. Can you also estimate interactions from a bar graph? Yes
27. Interactions can be described using what phrases? “especially for,” “only for,” and
“depends on.”
28. When a study shows both a main effect and an interaction, what is almost always more
important? The interaction is almost always more important. How does the
study on alcohol, aggression, and body weight provide an example of this? Drunk men
that are heavy are the most aggressive of all.