Survey Research Methods (2017)
c European Survey Research Association
ISSN 1864-3361
http://www.surveymethods.org
Vol. 11, No. 1, pp. 45-61
doi:10.18148/srm/2017.v11i1.6304
Reducing speeding in web surveys by providing immediate feedback
Frederick G. Conrad
Roger Tourangeau
Institute for Social Research
Ann Arbor, USA
Westat, Inc.
Rockville, USA
Mick P. Couper
Chan Zhang
Institute for Social Research
Ann Arbor, USA
School of Journalism
Fudan University
Shanghai, China
It is well known that some survey respondents reduce the effort they invest in answering questions by taking mental shortcuts – survey satisficing. This is a concern because such shortcuts
can reduce the quality of responses and, potentially, the accuracy of survey estimates. This
article explores “speeding,” an extreme type of satisficing, which we define as answering so
quickly that respondents could not have given much, if any, thought to their answers. To reduce speeding among online respondents we implemented an interactive prompting technique.
When respondents answered faster than a minimal response time threshold, they received a
message encouraging them to answer carefully and take their time. Across six web survey
experiments, this prompting technique reduced speeding on subsequent questions compared to
a no prompt control. Prompting slowed response times whether the speeding that triggered the
prompt occurred early or late in the questionnaire, in the first or later waves of a longitudinal
survey, among respondents recruited from non-probability or probability panels, or whether
the prompt was delivered on only the first or on all speeding episodes. In addition to reducing speeding, the prompts increased response accuracy on simple arithmetic questions for a
key subgroup. Prompting also reduced later straightlining in one experiment, suggesting the
benefits may generalize to other types of mental shortcuts. Although the prompting could have
annoyed respondents, it was not accompanied by a noticeable increase in breakoffs. As an alternative technique, respondents in one experiment were asked to explicitly commit to responding
carefully. This global approach complemented the more local, interactive prompting technique
on several measures. Taken together, these results suggest that interactive interventions of this
sort may be useful for increasing respondents’ conscientiousness in online questionnaires, even
though these questionnaires are self-administered.
Keywords: web surveys; speeding; survey satisficing; interactivity; data quality
1 Introduction
survey respondents cut corners, it has been called “survey
satisficing” (e.g. Krosnick, 1991).
It is well established that some survey respondents take
shortcuts when answering survey questions, providing responses that are acceptable but not optimal (see, for example,
Cannell, Miller, & Oksenberg, 1981). This is reminiscent
of the more general phenomenon that Simon (1956) dubbed
“satisficing” in which people solve everyday problems using
incomplete information, producing solutions that are good
enough to largely achieve their goals but not as good as if
they used all the information available to them. Thus, when
Web survey respondents are prone to take certain shortcuts
more often than respondents in other modes. For example,
web survey respondents have been observed to produce more
“Don’t Know” responses, more missing data, and more nondifferentiation than face-to-face respondents (e.g. Heerwegh
& Loosveldt, 2008). This is likely related to the absence of an
interviewer in web surveys. When respondents participate in
survey interviews with a live interviewer, there is social pressure to invest effort in the task and avoid being seen as lazy.
In addition, the web (not just web surveys) seems to promote
superficial and hurried processing by users. For example,
Nielsen and Loranger (2006) report that web users are more
likely to scan rather than read text online relative to reading
text on paper. While this general picture does not bode well
Contact information: Frederick G. Conrad, Institute for Social
Research, 426 Thompson Street, Ann Arbor, MI 48104-2321 (
[email protected])
45
46
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
for the quality of web survey responses, there may be a silver lining. The interactive character of web questionnaires
potentially enables designers to promote more conscientious
and thoughtful responding by providing feedback when respondents seem to be taking shortcuts.
The current article focuses on “speeding,” i.e., answering so quickly that it is unlikely the question has been adequately processed. In this study, we define speeding as responding so quickly that it is unlikely respondents have read
the question, let alone given much thought to their answer.
Speeding is a phenomenon frequently observed in web surveys (e.g. Zhang & Conrad, 2014) but less so in intervieweradministered modes. As we see it, when respondents speed
they are likely to provide “throw-away” answers: they feel
compelled to enter something but not necessarily an accurate
response. Thus, when respondents speed, it is unlikely accuracy will be above what can be expected by chance.
We define speeding as an absolute rather than relative
phenomenon: response time below a psychologically based
threshold rather than responses that are simply faster than
other responses (as, for example, Greszki, Meyer, and
Schoen, 2015 and Malhotra, 2008 define speeding)1 . The
“speeding threshold” probably differs across people: A fastthinking respondent might well be able to answer carefully
in a period of time that would be too brief for a slower respondent to consider the question much at all. And a respondent with a ready-made, familiar answer would also be
much faster than one who is thinking about the issue for the
first time. But technically, it is hard to distinguish a legitimate fast response from a response that is fast because of
speeding. Therefore, in the experiments reported here we
use one speeding threshold for all respondents. As we will
demonstrate, the threshold we selected seems to discriminate
between more and less conscientious responses and between
more and less conscientious respondents.
Speeding in web surveys seems to be prevalent enough
to be of concern to survey researchers. It is identified as a
concern in the AAPOR task force report on online panels
(Baker et al., 2010, p. 32). Zhang and Conrad (2014) reported that respondents in a Dutch probability panel sped on
about 15 out of 54 questions on average and, if they sped at
all, were likely to speed throughout the questionnaire. In the
six experiments we report in this article, between 37 and 85
percent of respondents sped at least once in answering seven
critical questions in the control condition (in which there was
no intervention for speeding). While these speeding statistics
depend on exactly how speeding is defined, the phenomenon
is common by any definition of speeding. Thus, a method to
reduce speeding could help potentially improve the quality
of web survey data and increase confidence in the mode for
high stakes applications like social scientific and government
surveys. We explore one such method here – an interactive
prompting technique.
2 Current Experiments
We report six web survey experiments that explore the
effectiveness of an intervention triggered when respondents
answered faster than a fixed threshold (see Table 1). The
main question is whether the intervention reduced subsequent speeding. The intervention was an on-screen, textual
prompt indicating that speeding had been detected and encouraging the respondent to give adequate thought to his or
her answers. It was administered when respondents sped in
the course of providing numerical answers to seven questions
about autobiographical quantities (Experiments 1–3 and 5)
or simple mathematical problems (Experiment 4); we refer
to these seven items in each experiment as the “prompting
items.” Respondents were randomly assigned to either a control condition in which there was no prompting for speeding
on these items or an experimental condition in which respondents were prompted when they sped. In Experiments 1–3,
we manipulated the prompting “dose” so that speeders were
either prompted every time or just the first time they sped on
the prompting items.
Each experiment was embedded in a larger omnibus questionnaire. The types and content of items varied widely. We
were concerned only with performance on the seven prompting items and (in Experiments 1–4) two grid questions that
came after the prompting items; the remainder of each questionnaire was administered for purposes unrelated to the current study. Each questionnaire varied in length; because of
skip patterns, different respondents answered different numbers of questions within each survey. Table 1 presents completion times for the entire questionnaire as an indication of
overall effort. The table also includes key attributes about the
design of the six experiments.
Respondents were recruited from volunteer panels in Experiments 1–4 and from a probability panel in Experiment 5.
Our focus is on the effects of the speeding prompt, comparing performance of respondents in the experimental to control groups. The speeding prompt read: “You seem to have
responded very quickly. Please be sure you have given the
question sufficient thought to provide an accurate answer. Do
you want to go back and reconsider your answer?” We developed four hypotheses where we had clear directional predictions. In addition, we have three research questions for
which we do not have theoretically based predictions.
Because the prompts were tightly coupled with speeding
on a particular question, we expected the prompts to give respondents the impression that their behavior was being monitored, increasing their sense of accountability. To the extent
1
Related to this relative notion of speeding is the common practice of removing answers with the slowest and fastest response times
(e.g. Heerwegh, 2003). The general assumption is that these outliers
are of low quality and, in the case of very fast responses, are the
result of speeding.
47
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Table 1
Overview of Experiments.
Type of Questions
Mean
Complet.
Durationa
Sample
Size
Type of
Sample
Field Period
Prompting
Dose
1
2
3
4
Frequency, Grids
Frequency, Grids
Frequency, Grids
Numeracy, Grids
20.1
27.6
21.3
25.3
2463
2453
3046
2565
Volunteer
Volunteer
Volunteer
Volunteer
Aug–Sep, 2007
Apr–May, 2008
Sep 2008
Jun–Jul, 2010
Once, Every Time
Once, Every Time
Once, Every time
Every Time
5a
5b
Frequency
Frequency
42.9
47.4
929
913
Probability
Probability
Jul 2009
Aug 2009
Every Time
Every Time
Experiment
Distinguishing
Features
Replicat. of Exp. 1
Prompting early/late
Commitment crossed
with prompting
Longitudinal design
The seven frequency questions and seven numeracy questions enable measurement of speeding and its reduction with prompting. The
numeracy questions also enable measurement of response accuracy. Grids allow measurement of straightlining and its reduction with
prompting. Mean completion duration provides an estimate of the effort required to finish the full questionnaire, which varied in length
between experiments and, due to skip patterns, between respondents. Two prompting doses are used in experiments 1–3 to evaluate
how much the number of prompts matters. Measuring the effect of dose was not a goal of experiments 4-5b, so only the “larger” dose –
Every Time – was used.
a
in minutes
that speeding results from the absence of such accountability, prompting speeders in this way should reduce subsequent
speeding.
Hypothesis 1: Prompting will reduce speeding on subsequent questions.
When respondents are first prompted they might reason “I
am being monitored and do not want to be seen as lazy” and
so pay more careful attention to the task. Thus, it is possible
that the administration of a single prompt would be as effective at reducing speeding as prompting respondents whenever they were caught speeding. Alternatively, it’s possible
that some respondents would not be deterred from speeding
by a single prompt but would slow down when prompted
multiple times, because multiple prompts may emphasize the
importance of responding carefully. In extreme cases, some
respondents might be so focused on finishing quickly that
they will not slow down until the repeated prompts actually
delay completion. Because we do not have a directional prediction we ask the following research question:
Research Question 1: Does a single prompt, delivered the
first time a respondent is caught speeding, reduce
speeding as much as a prompts delivered each time the
respondent speeds?
It is likely that as respondents advance through any questionnaire more than a few items long, they become more fatigued. This might compromise their ability to control and
adjust how they produce their answers, i.e., it may be more
difficult for fatigued respondents to slow down after being
prompted than it is for less fatigued respondents. Thus,
prompting later in the questionnaire may be less effective
than prompting early. Alternatively, the fatigue created by
completing many questions may not affect respondents’ ability to control how they answer questions so that after being prompted they should be able to attend to the task more
carefully and work harder. In this case, prompts should be
equally effective early and late in the questionnaire. Because
there is reason to believe that prompting later in the questionnaire will moderate the effectiveness of prompting but also
that it will not, we ask:
Research Question 2: Will the effectiveness of prompts
vary throughout the questionnaire or remain constant,
even as speeding increases?
In Experiment 3, the questions on which speeders were
prompted appeared early in the questionnaire (Q4–10) for a
random subsample of about two thirds of respondents and
later (Q60–66) for the remaining third.
Because speeding is assumed to reflect minimal effort and,
as a result, low quality data, reducing speeding helps the survey enterprise to the extent that slower responses increase
data quality2 . We examine the relationship between reduced
speeding and data quality in Experiment 4 by measuring
speeding on questions for which the true value is known, enabling us to determine response accuracy. In particular we
administered simple probability or arithmetic questions designed to assess numerical literacy (“numeracy”), for which
there are clear right and wrong answers. While respondents
2
Of course very long response times can also indicate reduced
quality. Specifically, they might reflect difficulty for respondents
who are trying to answer in good faith (e.g. Conrad, Schober, &
Coiner, 2007; Draisma & Dijkstra, 2004; Yan & Tourangeau, 2008).
In such cases, simplifying and improving the offending questions
should lead to responses that are both faster and of improved quality.
48
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
might not be more accurate each time they answer after being
prompted – they might answer more carefully without necessarily answering correctly – if they slow down, their chances
of answering any given question accurately should increase.
Hypothesis 2: Prompting respondents when they speed will
increase response accuracy.
Other measures of data quality might potentially be informative. For example, prompting speeding might increase
the length of open-ended responses, reduce the prevalence of
primacy effects, and reduce the number of missing observations on subsequent questions. However, we focus here on
response accuracy, as it is generally the most direct measure
of response quality for factual questions (but see Zhang, 2013
and Zhang and Conrad, 2016 for evidence that prompting respondents for speeding can increase the length of open-ended
responses).
This raises the question of whether the effects of prompting are confined to the item on which a respondent has
been caught speeding or whether the effects can transfer to
other items in the questionnaire. There is some evidence
that speeding is related to other types of survey satisficing.
Malhotra (2008) observed more primacy among low education respondents who answered most quickly (fastest tercile). Similarly, Callegaro, Yang, Bhola, Dillman, and Chin
(2009) reported that “satisficers” (identified by exogenous
factors) answered faster than “optimizers” (who had more
reason to respond accurately), suggesting that the satisficers
were speeding or were at least less conscientious. Although
these authors define fast responding in relative terms, unlike our threshold-based definition, their findings suggest that
speeding and satisficing in general may have a common origin. If so, it is plausible that speeding and straightlining
(non-differentiation) are related. If that is the case, reducing speeding with interactive prompts may also reduce later
straightlining.
Hypothesis 3: Speeding prompts will reduce subsequent
straightlining.
To test this, we investigated whether there was less
straightlining in the prompting than control conditions for
several grid questions (three grid questions were presented in
Experiments 1 and 2, and two grid questions were presented
in Experiments 3 and 4; see Appendix C) that followed the
prompting items. If Hypothesis 3 is supported, this would
suggest that prompting can improve data quality beyond the
item on which the prompt was delivered.
It is possible that the prompts tested here are only effective when they are novel. More specifically, it may be that respondent’s prompted in more than one wave of a longitudinal
survey will dismiss the later prompts, having recognized in a
previous wave that there is little cost to speeding aside from
receiving the potentially annoying message. Respondents in
Experiment 5 were members of the Face-to-Face Recruited
Internet Survey Platform (FFRISP), an area probability sample representative of the US population (e.g., Villar, Malka,
& Krosnick, 2010). The FFRISP data for the current study
were collected in two waves (July and August, 2009) allowing us to test the effectiveness of the prompts across waves
of an online survey. We randomly assigned respondents to
a No Prompt control or Prompt Every Time condition; the
respondents remained in the same condition in both waves.
We asked them the same autobiographical quantity questions
that we asked respondents in Experiments 1–3. On the one
hand it is possible that prompting will become less effective
after the first wave of data collection because respondents
recall that there is little cost to being prompted. On the other
hand, prompting after the first wave may remind respondents
that the questionnaire software is tracking their actions.
Research Question 3: Is the effectiveness of prompting reduced over waves of a longitudinal survey?
To the extent prompting reduces speeding it may work by
shaming respondents into being more conscientious. An alternative method that appeals to respondents’ better nature
and is not linked to particular response behaviors is to simply ask respondents at the start of the questionnaire to provide high quality information. Charles Cannell and his associates (e.g., Cannell et al., 1981; Oksenberg, Vinokur, &
Cannell, 1979) pioneered an approach along these lines for
face-to-face interviews; they asked respondents to commit to
providing thoughtful responses by signing a printed commitment statement. Cannell and his colleagues generally found
that this improved data quality. However, the respondents
made a face-to-face commitment to a human interviewer so
it is possible that requesting a commitment from respondents
will not be as effective when it is made online.
We tested the impact of respondents’ commitment on their
speeding behavior by asking half of the respondents in Experiment 4 to commit to answering thoughtfully and carefully at the start of the questionnaire. Half of the respondents
were asked to explicitly commit by clicking an on-screen option; the other half of the respondents received a filler statement the same length as the commitment statement.
Hypothesis 4: Respondents who commit to conscientious
responding will speed less than uncommitted respondents.
We crossed prompting with commitment and randomly
assigned respondents to one of the four resulting groups. The
respondents in the commitment groups were presented the
following statement at the start of the questionnaire:
It is very important that you read each question carefully
and think about your answer before you give it. We rely on
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
our respondents being thoughtful and taking this task seriously.
Respondents were then asked to select a radio button to
register their commitment:
I commit to reading each question carefully and thinking
about my answer before I give it.
# Yes
# No, but I will participate anyway
The filler statement presented to the other respondents at
the start of the questionnaire was:
Thank you very much for participating in our study. We
are grateful that you are willing to contribute to our project
and hope you find it a stimulating and worthwhile experience.
No action was required by the respondent.
2.1
Experimental Materials
Respondents’ answers were timed (the difference between
the time the respondents pressed the Next button after answering the previous question and the time the respondent
pressed the Next button after answering the current question)
for seven questions requiring numerical answers. For respondents in the experimental conditions, the software prompted
them when a response time fell below a fixed threshold, either the first time or any time this happened depending on the
dosage condition. The seven items either concerned autobiographical quantities (e.g., “Overall, how many overnight trips
have you taken in the PAST 2 YEARS?”) or, in Experiment
4, simple arithmetic questions (e.g., “If the chance of getting a disease is 10%, how many people out of 100 would be
expected to get the disease: 1, 10 or 20?”). The arithmetic
items were adapted from items developed to assess numerical literacy (numeracy) (L. M. Schwartz, Woloshin, Black,
& Welch, 1997). For all of these items (see Appendices A
and B), an accurate answer required at least some thought so
speeding is likely to signal incomplete thinking at best and,
thus, poor data quality.
The prompt was triggered when respondents answered
more quickly than 350 milliseconds (msec) per word, which
means that the speeding threshold for a ten-word question
would have been 3500 msec (i.e., three and a half seconds).
The thinking behind this particular threshold is that if a respondent enters an answer faster than most people can read
the question (Carver, 1992), let alone think about the answer,
he or she should probably slow down and answer subsequent questions more carefully. But one could choose other
group-based thresholds, e.g., Taylor, Frankenpohl, and Pettee (1960) report a somewhat faster median reading speed of
about 215 msec per word for US college students, or individualized thresholds (e.g. Jackson & McClelland, 1979). Our
threshold is a coarse measure of speeding but the approach is
simple and inexpensive to implement; to the extent that intervening when response times fall below this threshold slows
49
respondents down and leads them to invest more effort in the
response task, it may be well-suited to production web surveys.
The first and third grid questions (Appendix C) each involved one reverse-worded statement so that straightlining,
if observed, would be particularlly likely to reflect low effort
responding. One of these concerned dietary practices and
the other two concerned the strength and extremity of the respondents’ opinions and the manner in which they carry out
different mental tasks.
3 Results
We begin by examining the prevalence and correlates of
speeding. Next, we examine the effect of prompts (single and multiple) on speeding, response accuracy and later
straightlining. Then, we examine the impact of prompting over two waves of data collection. We next evaluate
the impact of the prompts when respondents have committed to answering carefully and, finally, we examine a potential negative effect of prompting – whether prompts promote
breakoffs.
3.1
Prevalence and correlates of speeding
In all six experiments, there was a substantial amount
of speeding in the control condition (where there was no
prompting), although the exact amounts varied. The percentages of control respondents who sped (responded in less than
350 msec per word) at least one time in the six experiments
were 85%, 82%, 42%, 74%, 37%, and 53%. Many factors
differ between the experiments so the specific percent of respondents who sped in any one experiment is not by itself
meaningful. The point is that speeding prevalence can be
quite high, as in Experiments 1, 2 and 4 and, even where less
prevalent, it is far from absent.
Not only is speeding more common than one would like,
but the percentages of speeders who sped multiple times was
quite high, tailing off only for 6 and 7 speeding episodes
across the 7 prompting questions. This can be seen in Figure 1, which presents the percent of speeders by the number
of questions on which they sped in the control condition of
each experiment. The general pattern is that most speeders
sped just once or twice but between 11.5 and 19.0 percent of
the respondents sped four times, i.e., on more than half of the
prompting items, across the experiments. And in Experiment
4, 11.2 percent of the speeders sped on every question. Both
the number of respondents who ever sped and the number
of questions on which they sped argue for some measure to
reduce the prevalence of speeding.
Who are the speeders? Across all the experiments,
younger respondents (especially in the 18–34 year old age
group) sped on significantly more questions than did older
respondents, controlling for other demographic variables. In
50
0
% Speeders Speeding
10
20
30
40
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
1
2
3
4
5
6
Number of Questions on which Speeding Detected
Exp 1
Exp 4
Exp 2
Exp 5a
7
Exp 3
Exp 5b
Figure 1. Percent of speeders (respondents who sped at least
once) in the control condition, distributed over the number
of questions on which they sped, for each of the six experiments.
separate multivariate models for each experiment, controlling for gender, education, income, employment status, and
race and in Experiments 1–4 for web panel membership and
self-reported “Internet ability,” only age was consistently related to speeding.3 Yan and Tourangeau (2008) found that
age is generally related to response times for web survey
questions, with older respondents taking longer to answer
than younger ones. Thus, it is not surprising that older respondents were less likely to meet our threshold for speeding.
How much of the age effect reflects slowing from cognitive
aging (e.g., Park & Reuter-Lorenz, 2009; Park & Schwarz,
2000; Park et al., 1996; Salthouse, 1991, 1996) versus greater
conscientiousness among older respondents is hard to disentangle in the current data.
3.2
Impact of intervention on subsequent speeding
Our first hypothesis is that prompting respondents when
they speed will reduce subsequent speeding. To test this, we
compare speeding prevalence in the control and the prompting conditions. Although respondents who were prompted
were invited to revisit the question on which they had just
sped, we look only at the times for the initial response in order to promote comparability between the experimental and
control conditions. Similarly, we restrict the analysis to those
respondents who sped at least once during the seven prompting questions, whether in the control or prompting conditions. If respondents did not speed at all, they were treated
the same way (i.e., no prompt) regardless of the condition to
which they were assigned because their performance is not
germane to our evaluation of the prompting procedure. Because response times for the initial answer to the first question cannot be affected by prior prompting, we examine the
effect of prompting on speeding for questions 2–7.
We observed less speeding in the prompting than control
conditions in all six experiments, as can be seen in Table 2.
Respondents who sped at least once sped on fewer questions
in the experimental conditions than in the control condition.
The reduction ranges in size from .2 to .6 questions, that is,
reductions in speeding prevalence of 10% to 24%. Considering that the number of questions on which control respondents sped ranged from 2.0 to 3.1, these reductions due to
prompting are substantial.
Our first research question asked whether a single prompt
was as effective as multiple prompts. As can be seen in the
rightmost column of Table 2, the two prompting doses do
not differently affect the amount of speeding. This suggests
that a single prompt can be as effective as more prompts in
reducing the amount of speeding.
The analysis, thus far, examines the global effect of
prompting speeders. While this approach indicates that
prompting has the intended effect overall, it does not directly measure the effect of prompting on subsequent speeding question by question. The next analysis does just this
by comparing speeding on questions 2–7 among respondents
who were prompted for speeding on an earlier question to
speeding by control respondents on the same questions. We
conducted these analyses for each question individually for
each experiment. A summary is presented in Table 3. The
right-most column presents the number of questions (out of
the six) for which there was significantly less speeding after
prompting for prior speeding than without prompting (i.e. in
the control condition). By this analytic approach, prompting reduced subsequent speeding in all of the experiments;
the number of questions for which reductions were observed
ranged between 2 and 5 questions out of 6. A more detailed
presentation of the data appears in Appendix D.
3.3
Effectiveness of prompts throughout the questionnaire
Our second research question was whether the intervention was any less effective later than early in the questionnaire. There was in fact more speeding when the questions
appeared late in the questionnaire than early, F(1, 1157) =
7.12, p < .05, suggesting that speeding increased when respondents were fatigued. However, prompting appeared to
be equally effective early and late in the questionnaire, interaction of prompt condition and position in the questionnaire, F(2, 1155) < 1. Thus, it seems that when they were
prompted, respondents slowed down and answered more
conscientiously irrespective of fatigue.
3
The effects of age were significant in 4 out of 5 experiments
(treating 5a and 5b as one experiment), with p-values ranging from
0.024 to < .0001.
51
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Table 2
Mean number of speeding episodes and F values, by condition and experiment
Experiment
No
Prompt
Prompt
Prompt
Every Time
Once
3.1
3.1
2.0
2.7
2.5
2.6
2.8
2.8
1.8
2.1
1.9
2.1
1
2
3
4
5a
5b
2.9
2.9
1.9
-
ANOVA (omnibus)
ANOVA (Prompt Every
Time vs. Prompt Once)
F
df1, df2
p
F
df1, df2
p
6.47
5.57
2.64
50.00
9.12
7.46
2, 2038
2, 2045
2, 1158
1, 1860
1, 313
1, 453
0.0016
0.0039
0.0716
< 0.001
0.0027
0.0066
2.31
<1
1.91
-
1, 1329
1, 1326
1, 777
-
n.s.
n.s.
n.s.
-
Table 3
Percent of respondents who sped again after speeding on one or more previous question(s)
# of Qs for which
difference is significant
Condition
Experiment
1
2
3
4
5a
5b
a
3.4
No
Prompt
62.8
59.8
33.9
47.5
50.6
59.9
Prompt
Every Time
52.3
51.9
31.2
35.9
45.1
42.8
Prompt
Once
Prompta vs. No
Prompt (Control)
54.5
52.9
32.9
-
5 out of 6
4 out of 6
3 out of 6
6 out of 6
2 out of 6
4 out of 6
In Experiments 1–3, the two prompting conditions are collapsed.
Response Accuracy
We consider respondents to have been accurate if they correctly answered six or seven of the seven numeracy questions, and we based this on final answers in order to take
account of changes made after respondents returned to the
question, post-prompting, potentially changing their answers
(although speeders who changed their answers were not
more accurate than those who did not). Overall, prompting
did not affect accuracy: 22.6% percent of respondents were
accurate in the prompt condition while 22.0% were accurate
in the control condition, X 2 (1) = 0.10, ns.
The fact that prompting can lead to slower responses apparently does not guarantee that these responses will be more
accurate. But answering more slowly should increase the
chances of answering accurately, especially if the task is neither too easy nor too difficult. If response accuracy is already
near the ceiling without prompting then prompting will not
noticeably increase accuracy. Similarly, if response accuracy
is near the floor, at least some of that inaccuracy is more
likely due to task difficulty than to speeding; pushing respondents to try hard through prompting is unlikely to matter. But
if the task is moderately difficult, the effects of prompting
on accuracy should be more visible. The numeracy ques-
tions used in Experiment 4 involve basic mathematical skills
so we reasoned that respondents’ education level could affect the difficulty of the questions and therefore the impact
of prompting on response accuracy. More specifically, we
suspected that the most educated respondents (college graduate or higher) would find the numeracy questions easy and
so would be quite accurate overall, despite some speeding.
At the other extreme, the least educated respondents (high
school or less) might find the numeracy tasks so difficult that
increased effort would not increase accuracy. Respondents
with intermediate levels of education (some college or an
Associate’s degree), in contrast, could find the task to be of
modest difficulty and thus their accuracy should benefit most
from prompting.
This is largely what was observed (see Figure 2). The
percent of those with a high school degree or less who were
accurate was not affected by prompting: 5.9% with prompting, 8.9% without, X 2 = 1.09, n.s.. Similarly prompting did
not affect accuracy for those with at least a Bachelors Degree: 30.9% with prompting, 34.3% without, X 2 = 0.87, n.s..
However significantly more respondents with some college
or an Associate’s degree accurately answered the numeracy
items: 22.2% with prompting, 16.0%, without, X 2 = 4.21,
52
Some College or Assoc. Degr.
BA or More
40
% Straightlining
20
30
30
10
20
0
10
0
0
% Respondents Who Answered Accurately
40
High School or Less
50
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
No
Prompt
Prompt
Every Time
No
Prompt
Prompt
Every Time
No
Prompt
p = 0.04. We see this as at least modest support for the
prediction that prompting can improve response quality, particularly when the task is neither too difficult nor too easy.
For other questions and response tasks, more thought may
well improve accuracy for all respondents, but here the effects were concentrated in this broad group – about 40% of
all respondents – with intermediate levels of education.
3.5
Speeding and straightlining
In all four experiments, we observed a positive association between speeding and straightlining: the more respondents sped on the seven prompting items, the more likely
they were to straightline in later grid questions (see Figure
3, which presents the percentage of control respondents who
straightlined on at least one of the two grids by the number of
times they sped previously). This suggests that speeding and
straightlining may share an underlying tendency to minimize
effort. This is consistent with findings reported by Zhang and
Conrad (2014) and Greszki et al. (2015).
This raises the question of whether speeding prompts reduced later straightlining as predicted by Hypothesis 3? We
found mixed support for the hypothesis. In Experiment 4,
9.5% of respondents straightlined if they had been prompted
previously whereas 12.8% of control respondents straightlined, X 2 (1) = 4.94, p < 0.05. However the effect was not
significant in Experiments 1–3; it was in the predicted direction for one of the prompting conditions but in the opposite
direction in the other. There were no grids in Experiment 5.
It may be that the effects of speeding prompts have to be large
to be carried over to the subsequent grid questions; the effect
of prompting on speeding in Experiment 4 was as great as or
greater than the effects in the other experiments (see Tables
2 and 3).
There is evidence that speeding prompts can reduce
2
3
4
Speeding Incidences on Previous Questions
Exper 1
Exper 3
Prompt
Every Time
Figure 2. Percent accurate respondents at three levels of education
1
5
6 or 7
Exper 2
Exper 4
Figure 3. Proportion of respondents who straightlined at
least once by number of speeding incidents, Experiments 1–
4.
straightlining from a study by Zhang (2013), and Zhang and
Conrad (2016) who prompted both straightlining and speeding within grid questions and reported that either type of
prompt reduced both behaviors on subsequent grid questions
(see also Kunz & Fuchs, 2014a, 2014b). In addition, after
they were prompted for straightlining or speeding, respondents provided longer answers to an open-ended question.
These findings suggest that the act of prompting may be more
important than the particular behavior triggering the prompt
in reducing satisficing behaviors.
3.6
Longitudinal effectiveness of speeding prompts
Our third research question asks whether there is any reduction in the prompt’s effectiveness after the first wave of
a longitudinal survey. As indicated earlier, 37% and 53% of
the control respondents sped at least once in the two waves of
Experiment 5. Clearly, speeding is common among respondents in the representative sample used in the study, and so
is not restricted to volunteer web panels. Speeding was more
common in the second wave; this difference was significant
in a MANOVA comparing number of speeding episodes in
the seven prompting items across waves: F(1, 894) = 69.2,
p < 0.001. Some respondents may have answered faster
in the second wave because they recognized the questions
from the first wave and did not need to read them as carefully the second time. However, despite the difference in
speeding prevalence, prompting was equally effective in both
waves: in a MANOVA, prompting reliably reduced speeding episodes compared to no prompt (F(2, 893) = 3.07,
p = 0.05) and, most importantly, did not interact with the
wave F(1, 894) = 0.01, p = 0.91 (see Figure 4).
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Wave B
2.5
2
1.5
# of Speeding Episodes
3
Wave A
No
Prompt
Prompt
Every Time
No
Prompt
Prompt
Every Time
Figure 4. Impact of prompts on mean number of speeding
episodes in two waves of Experiment 5 with a probability
sample.
3.7
Commitment and speeding
Our fourth hypothesis is that by committing to careful participation, respondents will speed less. In order to test this,
we analyzed the speeding behavior of all respondents in the
commitment and control conditions in Experiment 4, not just
the speeders (as in the analyses of the prompt’s effectiveness
in this experiment), and we looked at the effect of commitment on all seven prompting items. We did this because (1)
almost all (99%) of those asked to commit did so, and (2) the
commitment or filler request was presented to all respondents
at the start of the questionnaire making it possible to assess
its effects from the first item onward in contrast to prompting
whose assessment required at least one prompt to be administered and so could not begin until the second item.
Commitment did, in fact, reduce speeding as predicted in
the fourth hypothesis. Respondents who were asked to commit to thoughtful and careful responding sped 2.0 times, on
average; respondents who were not asked to commit sped
on 2.4 questions, F(1, 2447) = 15.4, p < 0.001. Commitment and prompting both reduced speeding but their effects
seemed to be independent of each other. For comparability to
the commitment analyses, we look at the effects of prompting for all respondents and over all seven items: Respondents
in the prompt group sped on 1.98 questions in contrast to
respondents in the no prompt control group who sped 2.44
times on average (F(1, 2447) = 28.15, p < 0.001); prompting and commitment did not interact (F(2, 2447) = 0.11,
n.s.)
We also examined the effects of commitment on response
accuracy for the numeracy items. Commitment improved
accuracy (i.e., answering 6 or 7 questions correctly) for respondents with a college degree or higher (38.8% vs. 26.8%,
X 2 (1) = 11.15, p < 0.001) but not significantly for those with
53
a high school education or less (9.0% vs. 5.8% X 2 (1) = 1.23,
n.s.), or some college or an Associates degree, 18.8% vs.
19.2%, X 2 (1) = 0.02, n.s; the interaction of commitment
and education level was marginally significant, X 2 (2) = 5.23,
p = 0.08. This interaction contrasts with the interaction
of prompting and education where it was respondents with
intermediate levels of education (some college or an Associate’s degree) whose accuracy was improved by the intervention. When appropriately motivated (committed) it seems
that more educated respondents may shift into a higher mental gear that is not available to respondents with less education, and which they do not engage when they are simply
prompted.
Finally, commitment strengthened the effect of prompting in reducing later straightlining. As discussed before,
there was less straightlining in the prompt than control condition but straightlining was particularly rare when committed respondents were also in the prompt condition: 6.2%
of respondents straightlined in the Commitment-Prompting
condition compared to 12.4%, 11.9%, and 12.8% in the
No Prompt-Commitment, Prompt-No Commitment, and No
Prompt-No Commitment conditions, X 2 (3) = 9.67, p =
0.022. One possibility is that having committed to careful
performance, respondents were more sensitive to prompting
and thus maintained their increased conscientiousness when
completing subsequent items, including the grids.
In summary, asking respondents to commit to answering
with care led to fewer speeding episodes than did a neutral statement. The effect was independent of the effect of
prompting. Commitment also improved response accuracy
for respondents with at least a college degree. Given the independent effects of commitment and prompting on speeding, and their impact on different subgroups’ accuracy, joint
use of the two techniques would likely have additive benefit.
3.8
Breakoffs
Although it reduced speeding and improved response accuracy, the prompting intervention could have annoyed respondents, leading some to break off. Overall, the percentage of speeders who broke off in Experiments 1–4 was small,
ranging from none to 2.4%. (The Experiment 5 data set included only completed cases so we could not compute the
breakoff rate for that study.) Breakoff data are presented by
experimental condition in Table 4 (comparisons are evaluated with Fisher’s exact test because of the small number of
breakoffs).
Prompting respondents once did not significantly increase
breakoffs compared to the No Prompt control condition in
any of the experiments (rightmost column), suggesting that,
in these studies, an intervention does not discourage respondents from completing the task. Prompting every time also
did not reliably affect the number of breakoffs compared to
the no prompt control condition in Experiments 2, 3 and 4
54
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
Table 4
Percent and number of breakoffs, by condition and experiment
Breakoffs
Fisher’s Exact Test for Difference
No
Prompt
Prompt
Every Time
Prompt
Once
Experiment
%
n
%
n
%
1
2
3
4
0.0
0.1
1.0
0.3
0
1
4
3
1.1
0.6
1.7
0.3
7
4
7
3
0.2
0.7
2.4
-
but it did increase breakoffs slightly in Experiment 1 (second
column from right) where there were seven more breakoffs in
the prompt every time than no prompt conditions. Although
the difference in Experiment 1 is significant, the actual increase in breakoffs (parenthesized numbers in the table) is
small compared to the overall sample size. It could be that
prompting every time for a larger number of questions will
increase breakoffs for chronic speeders. Given that a single prompt reduced speeding as much as multiple prompts,
and there is no compelling evidence that a single prompt increases breakoffs, the one prompt design seems like the sensible approach.
4 Discussion and Conclusions
Speeding was common across the six web survey experiments we conducted: 40 to 80% of the control respondents
answered at least one question so quickly it is unlikely they
read the entire question and even less likely they thought
carefully about their answers. But, the lesson from these experiments is that, by incorporating interactive feedback into
online questionnaires, survey designers may be able to reduce the prevalence of speeding and thus improve the quality of responses. Prompting respondents immediately after
they were caught speeding reduced the prevalence of this
behavior in all of the experiments, whether the prompting
occurred early or late, whether in a single wave or across
two monthly waves. The reduction in speeding was associated with some evidence of improved response quality: the
prompts increased response accuracy on a set of seven numeracy items for the group of respondents with moderate
educational attainment and they reduced straightlining across
all respondents in the same experiment, although not consistently in the other experiments. The intervention had very
little impact on breakoffs, suggesting that its benefits substantially outweigh its costs. Moreover, prompting complemented the benefits of commitment: the two techniques improved response accuracy for different, non-overlapping subgroups, and when respondents had both committed to careful responding and were in the prompt condition they were
Every Time vs.
No Prompt
Once vs.
No Prompt
n
p
p
1
5
9
-
< 0.01
0.21
0.40
1.00
0.49
0.12
0.26
-
especially unlikely to straightline.
By exploiting the interactivity of web surveys to reduce
speeding, designers can potentially improve response quality beyond what is observed in other modes4 . For example, speeding is likely to be very common in paper questionnaires, although it is typically impossible for researchers to
detect or to intervene to reduce its prevalence. And in spoken interviews there is pressure to speak (respond) after more
than about a second of silence, suggesting that respondents
in theses modes may rush their answers to fill the silence (see
Schober et al., 2015). But designers of online questionnaires
can easily measure response times and deliver feedback accordingly.
Although we observed increased accuracy with slower
responses for a subset of respondents, we cannot say for
sure that speeding always leads to less accurate responses
or that slowing respondents necessarily increases their accuracy. Greszki et al. (2015) removed responses they flagged
as “too fast” and observed virtually no effect on response distributions. But they examined items in a single domain –
politics and elections. In our studies, the experimental items
required numerical responses that may call for memory- and
arithmetic-based processes. Taking more time could well
improve the accuracy of the answers and change the response distributions. More generally, the relationship between speeding and accuracy may depend on item type, difficulty, and domain. In the absence of clear information about
the quality of very fast answers to particular questions, it
seems preferable to reduce speeding rather than to adjust for
it analytically.
There may be ways to increase the benefits of interactive
prompting. For example, customizing the speeding thresholds could result in prompts for respondents whose response
times are above the generic threshold we used, but who are
4
Other examples of online, interactive prompts include (1) DeRouvray and Couper (2002) who prompted respondents after answering “Don’t know” to provide a substantive response, and (2) Holland and Christian (2009) who prompted respondents to provide
more content in open responses.
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
speeding nevertheless. The speeding threshold could be customized for individuals based, for example, on their reading
times, or for subgroups whose cognitive function or tendency
to speed is known a priori, such as younger versus older respondents. We see this as an important next step in this line
of research.
But the effectiveness of approaches like this, that try to
motivate respondents to exert more effort or at least avoid the
stigma of being caught taking shortcuts, have limits: hardcore satisficers seem to experience none of the negative psychological consequences that “maximizers” do when their
performance is sub-optimal (e.g. B. Schwartz et al., 2002).
Those who are strongly disposed to take shortcuts are likely
to be unaffected by prompting or any similar technique. This
suggests that knowing something about the characteristics of
respondents who are both likely to speed and likely to respond to a prompt could allow more targeted interventions.
Exploring this type of issue would also be an important extension of the current research.
A possible downside of interactive prompting is that the
sense of being monitored – and the likely increase in accountability – may create a sense of social presence (the feeling that someone is there) which may discourage candid responding about sensitive topics. Zhang (2013), and Zhang
and Conrad (2016) report that prompting reduced speeding
but increased socially desirable responding for two subsequent questions on sensitive topics (marijuana use and exercise frequency).
Online data collection affords survey designers many options to improve response quality that are not available to
designers of paper questionnaires, in particular interactive
techniques such as the prompting technique tested here. Yet,
in practice, web questionnaires do not exploit the interactive
capabilities of the mode very often. The results presented
here demonstrate that progress can be made with relatively
modest changes to current practice.
5 Acknowledgments
Replication materials are published on OpenICPSR (Conrad, Zhang, Tourangeau, & Couper, 2017) and – as usual –
as supplementary materials alongside this article on SRM’s
Website.
We thank Reg Baker, Courtney Kennedy, and Stephanie
Stern for advice and assistance. In addition, we thank
the National Institute of Child Health and Development
(NIH Grants R01 HD041386-04A1 and R01 HD04138601A1) and the National Science Foundation (NSF Grant
SES0106222) for financial support. Any opinions, findings
and conclusions or recommendations expressed in this paper are those of the authors and do not necessarily reflect
the views of the National Science Foundation or the National
Institutes of Health.
55
References
Baker, R., Blumberg, S. J., Brick, J. M., Couper, M. P., Courtright, M., Dennis, J. M., . . . Lavrakas, P. J. (2010).
Research synthesis: AAPOR report on online panels.
Public Opinion Quarterly, 74(4), 711–781.
Callegaro, M., Yang, Y., Bhola, D. S., Dillman, D. A., &
Chin, T. Y. (2009). Response latency as an indicator of optimizing in online questionnaires. Bulletin de
Méthodologie Sociologique, 103, 5–25.
Cannell, C. F., Miller, P. V., & Oksenberg, L. (1981).
Research on interviewing techniques. Sociological
Methodology, 12, 389–437.
Carver, R. P. (1992). Reading rate: theory, research, and practical implications. Journal of Reading, 36, 84–95.
Conrad, F. G., Schober, M. F., & Coiner, T. (2007). Bringing
features of human dialogue to web surveys. Applied
Cognitive Psychology, 21, 165–188.
Conrad, F. G., Zhang, C., Tourangeau, R., & Couper, M. P.
(2017). Reducing speeding in web surveys by providing immediate feedback [data set]. Ann Arbor:
Inter-university Consortium of Political and Social Research. doi:10.3886/E100431V1
DeRouvray, C. & Couper, M. P. (2002). Designing a strategy for reducing “no opinion” responses in web-based
surveys. Social Science Computer Review, 20, 3–9.
Draisma, S. & Dijkstra, W. (2004). Response latency and
(para) linguistic expressions as indicators of response
error. In S. Presser, J. M. Rothgeb, M. P. Couper, J. T.
Lessler, E. Martin, J. Martin, & E. Singer (Eds.), Methods for testing and evaluating survey questionnaires
(pp. 131–149). Hoboken, NJ: Wiley.
Greszki, R., Meyer, M., & Schoen, H. (2015). Exploring the
effects of removing “too fast” responses and respondents from web surveys. Public Opinion Quarterly,
79(2), 471–503.
Heerwegh, D. (2003). Explaining response latencies and
changing answers using client-side paradata from a
web survey. Social Science Computer Review, 21,
360–373.
Heerwegh, D. & Loosveldt, G. (2008). Face-to-face versus
web surveying in a high-internet-coverage population:
differences in response quality. Public Opinion Quarterly, 72(5), 836–846.
Holland, J. L. & Christian, L. M. (2009). The influence of
topic interest and interactive probing on responses to
open-ended questions in web surveys. Social Science
Computer Review, 27(2), 196–212.
Jackson, M. D. & McClelland, J. L. (1979). Processing determinants of reading speed. Journal of Experimental
Psychology: General, 108(2), 151.
Krosnick, J. A. (1991). Response strategies for coping with
the cognitive demands of attitude measures in surveys.
Applied cognitive psychology, 5(3), 213–236.
56
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
Kunz, T. & Fuchs, M. (2014a). Instant interactive feedback in
grid questions: insights from eye tracking data. Paper
presented at the sixth European Congress of Methodology, Utrecht, Netherlands.
Kunz, T. & Fuchs, M. (2014b). Instant interactive feedback
in grid questions: reminding web survey respondents
of speeding and nondifferentiation. Paper presented at
annual conference of the American Association for
Public Opinion Research, Anaheim, CA.
Malhotra, N. (2008). Completion time and response order effects in web surveys. Public Opinion Quarterly, 72(5),
914–934.
Nielsen, J. & Loranger, H. (2006). Prioritizing web usability.
Berkeley, CA: New Riders.
Oksenberg, L., Vinokur, A., & Cannell, C. F. (1979). Effects of commitment to being a good respondent on
interview performance. In C. F. Cannell, L. Oksenberg,
& J. M. Converse (Eds.), Experiments in interviewing
techniques (pp. 74–108). Ann Arbor: Survey Research
Center, University of Michigan.
Park, D. C. & Reuter-Lorenz, P. (2009). The adaptive brain:
aging and neurocognitive scaffolding. Annual Review
of Psychology, 60, 173–196.
Park, D. C. & Schwarz, N. (Eds.). (2000). Cognitive aging:
a primer. Philadelphia, PA: Psychology Press.
Park, D. C., Smith, A. D., Lautenschlager, G., Earles, J. L.,
Frieske, D., Zwahr, M., & Gaines, C. L. (1996). Mediators of long-term memory performance across the life
span. Psychology and aging, 11(4), 621.
Salthouse, T. A. (1991). Theoretical perspectives on cognitive aging. Hillsdale, NJ: Lawrence Erlbaum Associates.
Salthouse, T. A. (1996). The processing-speed theory of
adult age differences in cognition. Psychological review, 103(3), 403.
Schober, M. F., Conrad, F. G., Antoun, C., Ehlen, P., Fail, S.,
Hupp, A. L., . . . Zhang, C. (2015). Precision and disclosure in text and voice interviews on smartphones.
PLOS ONE, 10(6). doi:10.1371/journal.pone.0128337
Schwartz, B., Ward, A., Monterosso, J., Lyubomirsky, S.,
White, K., & Lehman, D. R. (2002). Maximizing versus satisficing: happiness is a matter of choice. Journal of Personality and Social Psychology, 83(5), 1178–
1197.
Schwartz, L. M., Woloshin, S., Black, W. C., & Welch, H. G.
(1997). The role of numeracy in understanding the
benefit of screening mammography. Annals of internal
medicine, 127(11), 966–972.
Simon, H. A. (1956). Rational choice and the structure of the
environment. Psychological review, 63(2), 129.
Taylor, S. E., Frankenpohl, H., & Pettee, J. L. (1960). Grade
level norms for the components of the fundamental
reading skills. EDL Research and Information Bulletin
#3. New York: EDL/McGraw Hill.
Villar, A., Malka, A., & Krosnick, J. (2010). Assessing the
accuracy of the face-to-face recruited internet survey platform: a comparison of behavioral and healthrelated findings from the FFRISP with those of major
national surveys. Paper presented at the annual conference of the American Association for Pubic Opinion
Research, Chicago, IL.
Yan, T. & Tourangeau, R. (2008). Fast times and easy questions: the effects of age, experience and question complexity on web survey response times. Applied Cognitive Psychology, 22(1), 51–68.
Zhang, C. (2013). Satisficing in web surveys: implications for
data quality and strategies for reduction. Unpublished
doctoral dissertation. University of Michigan, Ann Arbor, MI.
Zhang, C. & Conrad, F. G. (2014). Investigation of speeding in web surveys: tendency to speed and its association with straightlining. Survey Research Methods,
8(2), 127–135.
Zhang, C. & Conrad, F. G. (2016). Intervention as a strategy to reduce satisficing behaviors in web surveys: evidence from two experiments on how it works. Social
Science Computer Review. Online first. doi:10.1177/
0894439316683923
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Appendix A
Autobiographical Quantity Questions
1. Overall, how many overnight trips have you taken in
the PAST 2 YEARS?
2. In the past year, about how many calories did you consume in a typical day?
3. In an average month, how much money do you spend
on alcohol? (Please enter 0 if you spend nothing on
alcohol)
4. During the last 7 days, how many drinks of any kind
of alcoholic beverages did you drink?
5. During the past 30 days, how many times did you have
5 or more drinks of alcoholic beverage on one occasion?
6. On how many days in the past year did you stay in bed
because of illness or injury?
7. During the past 10 years, how many times did you get
any traffic tickets, including speeding tickets and parking tickets?
57
58
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
Appendix B
Numeracy Questions
Now for some questions on your perception of health related
risk . . .
1. Different sports vary in the risk of injury. Which of
the following numbers represents the biggest risk of
getting injured while playing sport?
2 1 in 200
2 1 in 10
2 1 in 50
2 1 in 100
2. Heart disease is the leading cause of death in the
United States. If Person A’s chance of getting a
particular type of heart disease is 1% in ten years, and
person B’s risk is half that of A’s, what is B’s risk?
2 2%
2 5%
2 0.2%
2 0.5%
3. Viruses cause some of the most familiar infectious
diseases, such as the common cold and the flu. If the
chance of getting a particular type of viral infection is
0.0005, about how many out of 10,000 are expected to
get infected?
2 0.5
25
2 50
2 500
4. Chronic diseases, such as heart disease, stroke, cancer,
diabetes, and arthritis, are among the most common,
costly, and preventable of all health problems in the
U.S. If the estimated prevalence of a particular type
of chronic disease is 1 out of 1,000, what percent of
people would get the disease?
2 1%
2 0.1%
2 0.01%
2 0.001%
5. If Vaccine A’s chance of causing serious side effects
is 0.1% and Vaccine B’s chance is double that of A’s,
what percent of people are not expected to experience
serious side effects after taking Vaccine B?
2 98%
2 99%
2 99.9%
2 99.8%
6. Vision loss is a public health problem in the U.S. If
Persons A’s chance of vision loss is 1 in 100 in twenty
years, and person B’s risk is double that of A’s, what
is B’s risk?
2 2 out of 200
2 1 out of 200
2 2 out of 50
2 None of the above
7. Iron deficiency is a condition resulting from too little
iron in the body. If the risk of iron deficiency in a
certain demographic group is 0.2%, how many people
out of 1,000 in that group would be expected to have
iron deficiency?
2 0.2
22
2 20
2 200
59
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Appendix C
Grid questions
1. Indicate how much you favor or oppose each of the following statements.
Strongly
oppose
Somewhat
oppose
Neither favor
nor oppose
Somwhat
favor
Strongly
favor
Avoiding “fast food?”
#
#
#
#
#
Maintaining a healthy diet?
#
#
#
#
#
Monitoring cholesterol levels
closely?
#
#
#
#
#
Emphasizing the taste of food rather
than its nutritional value?
#
#
#
#
#
Paying close attention to the nutritional information on food packaging?
#
#
#
#
#
Limiting the amount of red meat in
your diet?
#
#
#
#
#
Balancing one’s diet across the key
food groups?
#
#
#
#
#
2. (Experiment 1 and 2 only) The next few questions ask about the style or manner you use when carrying out different
mental tasks. Your answers to the questions should reflect the manner in which you typically engage in each of the tasks
mentioned. There are no right or wrong answers; we only ask that you provide honest and accurate answers.
Strongly
disagree
#
Moderately
disagree
#
Neither agree
nor disagree
#
Moderately
agree
#
Strongly
agree
#
When listening to someone describing their experiences, I try to mentally picture what was happening.
#
#
#
#
#
I do a lot of reading.
#
#
#
#
#
I find it helps to think in terms of
mental pictures when doing many
things.
#
#
#
#
#
I enjoy learning new words.
#
#
#
#
#
I enjoy work that requires the use of
words.
60
FREDERICK G. CONRAD, ROGER TOURANGEAU, MICK P. COUPER AND CHAN ZHANG
3. For each of the following statements, please rate to what extent it characterizes you.
Extremely
uncharacteristic
Somewhat
uncharacteristic
Somewhat
characteristic
Extremely
characteristic
I prefer to avoid taking extreme positions
#
#
#
#
I want to know exactly what is good
and bad about everything
#
#
#
#
If something does not affect me,
I do not usually determine if it is
good or bad
#
#
#
#
There are many things for which I
do not have a preference
#
#
#
#
I like to have strong opinions even
when I am not personally involved
#
#
#
#
I would rather have a strong opinion
than no opinion at all
#
#
#
#
I only form strong opinions when I
have to
#
#
#
#
61
REDUCING SPEEDING IN WEB SURVEYS BY PROVIDING IMMEDIATE FEEDBACK
Appendix D
Percent respondents speeding on each question after speeding on one or more earlier questions in each experiment.
No
Prompt
Prompt
Every
Time
Prompt
Once
%
%
%
Experiment 1
Q2
37.1
Q3
63.7
Q4
85.8
Q5
75.0
Q6
55.3
Q7
59.6
20.1
53.6
75.8
64.4
44.9
55.0
Experiment 2
Q2
24.2
Q3
64.8
Q4
83.9
Q5
72.0
Q6
47.1
Q7
66.5
Prompt vs. No
Prompt
Prompt Every
Time vs. Once
χ2
p
χ2
p
15.0
52.6
79.9
68.6
50.7
60.3
20.96
6.29
9.68
12.12
9.23
0.70
< 0.01
0.01
< 0.01
< 0.01
< 0.01
0.40
1.42
< 1.00
1.80
2.03
3.99
3.67
0.23
0.84
0.18
0.15
0.05
0.06
21.1
56.7
78.1
59.3
41.5
55.0
15.3
53.9
80.6
62.4
45.0
60.5
2.60
5.27
3.38
20.56
2.52
14.20
0.11
0.02
0.07
< 0.01
0.11
< 0.01
1.99
< 1.00
< 1.00
1.16
1.49
3.93
0.16
0.56
0.40
0.28
0.22
0.05
Experiment 3
Q2
0.7
Q3
43.9
Q4
40.3
Q5
58.8
Q6
17.1
Q7
42.5
2.0
49.7
39.5
49.1
13.7
33.4
2.1
55.9
36.8
52.3
14.5
36.0
Experiment 4
Q2
55.2
Q3
36.9
Q4
63.2
Q5
37.7
Q6
45.4
Q7
46.7
42.7
23.0
56.1
24.5
35.2
34.0
-
17.09
31.41
7.53
33.65
18.44
29.07
< 0.01
< 0.01
0.01
< 0.01
< 0.01
< 0.01
-
-
Experiment 5a
Q2
50
Q3
62.1
Q4
61.9
Q5
30.5
Q6
54.0
Q7
45.2
85.71
49.3
55.1
18.1
26.1
36.3
-
too few cases
1.34
0.25
0.55
0.46
3.72
0.05
15.52 < 0.01
1.76
0.19
-
-
Experiment 5b
Q2
100.0
Q3
58.3
Q4
74.1
Q5
27.3
Q6
44.1
Q7
55.5
70.6
42.4
48.2
21.5
31.1
43.2
-
too few cases
3.58
0.06
11.70 < 0.01
1.20
0.27
5.08
0.02
4.49
0.03
-
-
too few cases
3.17
0.08
0.33
0.56
5.06
0.02
1.63
0.20
6.13
0.01
too few cases
1.17 0.28
< 1.00 0.54
< 1.00 0.45
< 1.00 0.75
< 1.00 0.48