An Assessment of The Problem and Solutions: Cheating On Political Knowledge Questions in Online Surveys
An Assessment of The Problem and Solutions: Cheating On Political Knowledge Questions in Online Surveys
An Assessment of The Problem and Solutions: Cheating On Political Knowledge Questions in Online Surveys
858–887
SCOTT CLIFFORD*
JENNIFER JERIT
Scott Clifford is an assistant professor in the Department of Political Science at the University
of Houston, Houston, TX, USA. Jennifer Jerit is a professor in the Department of Political
Science at Stony Brook University, Stony Brook, NY, USA. The authors thank the following peo-
ple for helpful comments and suggestions on earlier versions of this paper: Craig Burnett, John
Bullock, Mona Kleinberg, Yanna Krupnikov, Thomas Leeper, Adam Seth Levine, and Hillary
Shulman. This work was supported by the College of Arts and Sciences at Stony Brook University
and the Division of Research at the University of Houston. *Address correspondence to Scott
Clifford, University of Houston, Department of Political Science, 3551 Cullen Boulevard Room
447, Houston, TX 77204-3011, USA; e-mail: sclifford@uh.edu.
doi:10.1093/poq/nfw030 Advance Access publication August 12, 2016
© The Author 2016. Published by Oxford University Press on behalf of the American Association for Public Opinion Research.
All rights reserved. For permissions, please e-mail: journals.permissions@oup.com
Cheating on Political Knowledge Questions 859
3. Gooch and Vavreck (2015) report low rates of cheating, but their study took place in a lab-like
setting, making it different than the typical online survey in which respondents participate at a
time and place of their own choosing.
Cheating on Political Knowledge Questions 861
4. Indeed, in the data we report below, student subjects had a tendency to place more importance
on being politically knowledgably than campus staff (p < .10).
862 Clifford and Jerit
5. Additionally, student subjects may be likely to offload the burden of remembering facts
because many of them came of age in the internet era (Kleinberg and Lau 2014).
6. Shulman and Boster (2014, 187) and Jensen and Thomsen (2014, 3353) both encourage future
researchers to examine how to reduce cheating on knowledge questions in online surveys.
Table 1. Summary of Data Sources
Sample size (control group) Approximate date of study Included experimental manipulation(s) Number of knowledge Qs
Student 1 84 (46) May 2014 Direct request 5
Student 2 845 (203) November 2014 Forgiving, Timer, Commitment 8
Student 3 271 (135) March 2015 Commitment 8
Student 4 66 (35) April 2015 Commitment 4
Campus staff 59 (30) May 2015 Commitment 4
YouGov 1,000 (354) December 2015 Timer, Commitment 8
CCES 1,000 (NA) November 2014 None 5
Cheating on Political Knowledge Questions
Note.—The data reported in Student 1, Student 4, and Campus Staff were from the second wave of two-wave studies. See appendix A for details on the admin-
istration of all studies.
863
Many people struggle to remember facts, even when they know them, and
so they get help remembering. When you were answering the factual knowl-
edge questions, did you get help from any other source, such as the internet
or another person? (Please be honest, this is for statistical purposes only.)
7. With the exception of MTurk 2 (N = 505), there was no explicit “Don’t Know” (DK) option
(see Miller and Orr [2008] for discussion).
8. The scales vary in terms of length and apparent difficulty, as judged by percent correct on the scale
(which ranges from 44 to 77 percent correct). There also were differences in the topics of the surveys
in which the knowledge scales were embedded. Any of these factors might influence the motivation
to cheat, but as we report below, we observe the expected patterns despite this variation. Moreover,
data from Study 4 suggest that scale differences do not seem to affect rates of self-reported cheating.
In that study, students were randomized to receive one of two knowledge batteries. One knowledge
scale was substantially more difficult, as evidenced by lower scores (p < .01, d = .34). Yet, respondents
receiving the more difficult scale were no more likely to report cheating (p = .56).
9. Sample sizes may be smaller than those reported in table 1, due to item non-response on the
cheating question.
Cheating on Political Knowledge Questions 865
MTurk 2
MTurk 3
MTurk 1
CCES*
YouGov
Student 4
Student 1
Student 3
Student 2
0 10 20 30 40 50
10. In later analyses, we examine whether MTurk subjects are more likely to underreport cheating
for fear of not being compensated.
866 Clifford and Jerit
11. Those studies include students, online panelists from the CCES, and MTurk workers.
12. The language was as follows: “Now we have a set of questions concerning various public
figures. We want to see how much information about them gets out to the public from television,
newspapers, and the like. [Please give your best guess and do NOT use outside sources like the
Internet to search for the correct answer.]”
Cheating on Political Knowledge Questions 867
reported cheating, while 18.4 percent reported cheating in the treatment condi-
tion, a difference that is not statistically significant (p = .54). Knowledge in the
treatment group is lower than in the control (3.7 vs. 3.9, respectively), but the
difference is not statistically significant or substantively large (p = .44, d = .17).
In our second student study (Student 2), we explored three different manipula-
tions: Timer, Commitment, and Forgiving Request. All respondents (including
Timer: “Please do NOT use outside sources like the Internet to search for
the correct answer. You will have 30 seconds to answer each question.”
The second treatment asked respondents whether they were willing to answer
the knowledge questions without getting help and requested a yes or no
response (“Commitment”; see Krosnick [1991] or Clifford and Jerit [2015]).
We expect this treatment to be effective because few people will reject a request
from an interviewer, and subsequently they will be motivated to maintain con-
sistency with their previous commitment (Cialdini et al. 1978; Cannell, Miller,
and Oskenbeurg 1981).
The third and final treatment was a stronger version of the Direct Request
from the Student 1 study. Like the Direct Request, the treatment instructed
subjects not to look up answers, but it also included additional forgiving lan-
guage intended to reduce social desirability pressures (“Forgiving Request”;
e.g., Duff et al. [2007]).
Forgiving Request: “Many people will know the answers but won’t be
able to remember them at the moment. However, it is essential that we
only measure what people are able to recall on their own. Thus, it is
important to us that you do NOT use outside sources like the Internet to
search for the correct answer.”
The comparisons from the Student 2 sample are shown in figure 2. On aver-
age, subjects in the control condition answered 5.1 questions out of 8 correctly,
13. Only 5 out of 222 subjects (2.3 percent) answered “no” to the question posed by the treatment.
All subjects are retained for analysis regardless of their response.
868 Clifford and Jerit
Self-Reported Cheating
Forgiving
Commitment
Timer
Control
Knowledge Scores
Forgiving
Commitment
Timer
Control
0 2 4 6 8
Forgiving
Commitment
Timer
Control
0 5 10 15 20 25
with 41 percent reporting cheating. Beginning with the Timer condition, sub-
jects were significantly less likely to report cheating (16.4 percent, p < .001)
and scored significantly lower on knowledge (4.3; p < .001, d = .54). One
would expect this treatment (by design) to reduce time spent on the knowledge
questions, and the results bear out this expectation (p < .001, d = .71).14
Turning to the Commitment condition, self-reported cheating was dramati-
cally lower than in the control condition (9.5 percent, p < .001). In addition,
knowledge scores (4.4; p < .001, d = .49) and time spent on the knowledge
questions were significantly lower as well (p = .015, d = .24). Finally, we
examine the Forgiving treatment, which was designed to be a stronger version
14. We use the log of the average time spent per question for all tests of reaction times.
Cheating on Political Knowledge Questions 869
of the Direct Request. Relative to the rate of cheating in the control condi-
tion (41.4 percent), the Forgiving treatment significantly decreased rates
of self-reported cheating (22.4 percent, p < .001), knowledge scores (4.5;
p < .001, d = .40), and time spent on the knowledge questions (p = .004, d = .29).
However, this manipulation was less effective than both the Commitment and
Timer conditions.
student studies, Timer and Commitment, in the adult sample collected through
YouGov.15 Recall from figure 1 that adult subjects were less likely than stu-
dents to report cheating. This pattern is consistent with the claim that online
panelists have weaker motives to cheat than student subjects. Yet, the weaker
motivation to cheat may make it more difficult to find a significant effect for
language designed to reduce cheating among adult subjects. In the control
We now turn to the question of whether our treatments are effective in MTurk
samples. Earlier, we argued that the low rate of self-reported cheating among
MTurk subjects (see figure 1) is a function of opportunity costs that countervail
the motivation to cheat (cf. Goodman, Cryder, Cheema [2013]). An alternative
15. We included an experiment involving the Commitment language in Student 4 and Campus
Staff (N = 66 and N = 59, respectively). Despite our recruiting efforts, these studies were smaller
than expected, which in turn reduced statistical power. In both studies, the magnitude of treatment
effects was consistent with our other studies, but the differences were not statistically significant.
Cheating on Political Knowledge Questions 871
16. We also have evidence from Student 1, which consisted of a two-wave panel in which stu-
dents answered political knowledge questions in a lab and then answered different knowledge
items in a follow-up survey online. The correlation between attention to politics and politi-
cal knowledge is significantly lower in the online wave, relative to the lab wave (see online
appendix C).
872 Clifford and Jerit
There was a parallel set of questions on the topic of health care reform.17 Once
again, we expected that political knowledge would be associated with listing
a larger number of considerations for both the thought-listing and attribution
items (Zaller 1992; Gomez and Wilson 2001). A coder who was blind to our
expectations and treatment assignment coded the number of distinct consid-
erations listed by respondents in each of the four questions (Krippendorf’s
17. The fixed choice questions read as follows: “Do you favor or oppose stricter gun control
laws?” and “Do you support or oppose the health care law passed by the President and Congress
in 2010?” See appendix D for question wording of the closed- and open-ended items.
18. The reported Krippendorf’s alpha comes from an intercoder reliability analysis based on a
random subset of the data (n = 60).
19. We exclude inattentive respondents because they are less likely to give reliable responses to
the SDE battery and are less responsive to experimental manipulations (Berinsky, Margolis, and
Sances 2014; see online appendix A for details).
Cheating on Political Knowledge Questions 873
condition, the average interest in taking a similar survey in the future was
4.08, roughly corresponding with the “very interested” response option.
Survey interest did not significantly differ in the Timer or Commitment con-
ditions (ps > .50), which is at odds with the claim that these interventions
upset respondents.
Additionally, we included an open-ended item asking if there was any-
Conclusion
Collectively, these studies demonstrate that respondents frequently look up
the answers to political knowledge questions in online surveys and provide
suggestive evidence that this behavior varies by sample. On MTurk, where
respondents have a financial incentive to finish quickly, cheating rates seem
relatively low. Among students, who face lower opportunity costs and may
have a greater desire to self-enhance on political knowledge, cheating behavior
is common. In this latter population, cheating rates are high enough to distort
levels of political knowledge. Finally, in national samples, self-reported cheat-
ing rates were moderate, though we still found that a cheating intervention
(i.e., the Commitment language) improved the validity of a knowledge scale.
Researchers and survey organizations have adopted practices such as time
limits and instructions to discourage cheating, but have provided little evi-
dence for their efficacy. According to our analyses, a direct request asking
20. Overall, 56 percent of respondents explicitly stated that there was nothing they disliked about
the survey, and this figure did not vary across conditions (p = .59).
874 Clifford and Jerit
people not to cheat was the least effective of the techniques we examined,
even when paired with forgiving language intended to reduce social desir-
ability pressures. Timers were more effective, but they may interfere with
honest subjects’ ability to answer the questions. As a result, we recommend
the commitment item—that is, asking respondents whether they are willing to
answer knowledge questions without help, and requiring a yes or no answer.
21. At present, researchers do not have a definitive way to identify cheating. The willingness to
admit to this behavior may be correlated with characteristics that are related to knowledge (e.g.,
engagement, interest). Although we believe a self-reported cheating question can be useful in
gauging the extent of this behavior in the aggregate, more caution is warranted when using this
item as a measure of cheating at the individual level.
Cheating on Political Knowledge Questions 875
Student Sample 2
A total of 845 subjects were recruited for an online survey from required intro-
ductory political science courses at the University of Houston during the fall
of 2014. Subjects were compensated with extra credit for completion of the
survey. The survey was approximately 30 minutes long, with political knowl-
edge questions coming near the end of the survey.
Student Sample 3
A total of 271 subjects were recruited for an online survey from required intro-
ductory political science courses at the University of Houston during the spring
of 2015. Subjects were compensated with extra credit for completion of the
survey. This survey was the second wave of a panel. The first wave (N = 370)
measured demographics and material for an unrelated study approximately
four weeks earlier. The attrition rate was 27 percent.
Student Sample 4
A total of 66 students were recruited from political science courses at Stony
Brook University during the spring of 2015. This survey was the second wave
of a panel study (first wave N = 154) focusing on an unrelated topic. The attri-
tion rate was 57 percent.
CCES 2014
A total of 1,000 respondents completed this survey as part of the Duke University
pre-election wave of the 2014 Cooperative Congressional Election Study
administered by YouGov/Polimetrix. Individuals are recruited into the YouGov/
Polimetrix Internet panel using targeted online advertisements, and the CCES
data are constructed from the thousands of people who have consented to be a
YouGov
A total of 1,000 respondents were recruited by YouGov during December
2015 using the process described above. YouGov calculated a participation
rate of 29.5 percent.
MTurk Sample 1
This convenience sample was collected in the spring of 2014 for an unre-
lated study. We recruited 500 respondents on an opt-in basis from Amazon’s
Mechanical Turk crowdsourcing platform. We restricted the posting to residents
of the United States who were at least 18 years of age and had at least a 97 percent
approval rate. The survey was advertised as a “short survey on political attitudes,”
and respondents were paid $0.60 for completing the survey. All respondents who
completed the survey were compensated and included in the analysis. Political
knowledge was measured at the end of the 5–10-minute survey.
MTurk Sample 2
This convenience sample was collected in the summer of 2014 for an unre-
lated study. We recruited 505 respondents on an opt-in basis from Amazon’s
Mechanical Turk crowdsourcing platform. We restricted the posting to resi-
dents of the United States who were at least 18 years of age and had at least a
97 percent approval rate. The survey was advertised as a study about “politi-
cal beliefs and attitudes,” and respondents were paid $2.00 for completing
the survey. All respondents who completed the survey were compensated and
included in the analysis. Political knowledge was measured at the end of the
30-minute survey.
MTurk Sample 3
This convenience sample was collected in the spring of 2015 for an unre-
lated study. We recruited 505 respondents on an opt-in basis from Amazon’s
Mechanical Turk crowdsourcing platform. We restricted the posting to
Cheating on Political Knowledge Questions 877
residents of the United States who were at least 18 years of age and had at
least a 95 percent approval rate. The survey was advertised as a “short politi-
cal survey,” and respondents were paid $0.50 for completing the survey. All
respondents who completed the survey were compensated and included in the
analysis. Political knowledge was measured at the end of the 10-minute survey.
Control
Now we have a set of questions concerning various public figures. We want to
see how much information about them gets out to the public from television,
newspapers, and the like.
Direct Request
Now we have a set of questions concerning various public figures. We want to
see how much information about them gets out to the public from television,
newspapers, and the like. Please give your best guess and do NOT use outside
sources like the Internet to search for the correct answer.
Timer
Now we have a set of questions concerning various political issues. We want
to see how much information about them gets out to the public from television,
newspapers, and the like. Please do NOT use outside sources like the Internet to
search for the correct answer. You will have 30 seconds to answer each question.
Commitment
Now we have a set of questions concerning various political issues. We want
to see how much information about them gets out to the public from television,
878 Clifford and Jerit
newspapers, and the like. It is important to us that you do NOT use outside
sources like the Internet to search for the correct answer. Will you answer the
following questions without help from outside sources?
•• Yes
•• No
1. For how many years is a United States Senator elected – that is, how
many years are there in one full term of office for a US Senator? 2; 4; 6; 8
2. What is Medicare? A program run by the US federal government to pay
for old people’s health care; A program run by state governments to pro-
Student 2
1. Do you happen to know the name of the current Speaker of the U.S.
House of Representatives? Is it...John Boehner; Nancy Pelosi; Harry
Reid; Mitch McConnell
2. What is the name of the Chair of the Federal Reserve Board? Janet
Yellen; Ben Bernanke; Kirsten Gillibrand; Eric Cantor
3. What is the current unemployment rate in this country? Is it…3.9; 5.9;
10.9; 15.9
4. On which of the following does the US federal government currently
spend the least? Foreign aid; Medicare; National defense; Social
Security
5. As far as you know, which foreign country holds the most US govern-
ment debt? China; Japan; Saudi Arabia; Canada
6. How much of a majority is required for the US Senate and US House
to override a presidential veto? One-half; Two-thirds; Three-fourths;
Three-fifths
7. Do you happen to know what job or political office is now held by
Ban Ki-Moon? Secretary of the US Department of Transportation;
Prime Minister of Japan; Secretary of the United Nations; United States
Ambassador to South Korea
8. In a recent address before the nation, President Obama outlined a four-
point plan for addressing the Islamic State. Which of the following is
not part of his plan? Humanitarian assistance to displaced civilians;
Airstrikes against Islamic State targets; Training Iraq national guard
units; Deployment of US ground combat units
880 Clifford and Jerit
Student 3
1. Do you happen to know what the federal minimum wage is today? Is it:
$5.25; $7.25; $10.50; $12.50
2. Who is the prime minister of Israel? Benjamin Netanyahu; Shimon
Peres; Ariel Sharon; Hosni Mubarak
Cheating on Political Knowledge Questions 881
3. How long is the term of office for a senator in the US Senate? 2 years;
4 years; 6 years; 8 years
4. Do you happen to know if the national inflation rate as reported by the
government is currently closer to…1%; 6%; 10%; 15%
CCES
MTurk 1
MTurk 2
4. What is ISIS? A Sunni extremist group seizing Syria and Iraq; A gov-
ernment program to monitor citizens; An energy company; A plan for
stabilizing Iraq; Don’t know
5. On which of the following does the US federal government currently
spend the least? Foreign aid; Medicare; National defense; Social
Security; Don’t know
YouGov
4. Do you support or oppose the health care law passed by the President
and Congress in 2010? Support; Oppose; Unsure
5. Still thinking about the question you just answered, what thoughts came
to mind while you were answering that question? Please try to list eve-
rything that came to mind. (open-ended)
6. For decades, experts have observed that the United States spends far
Supplementary Data
Supplementary data are freely available online at http://poq.oxfordjournals.org/.
References
Ansolabehere, Stephen, and Brian Schaffner. 2014. “Does Survey Mode Still Matter? Findings
from a 2010 Multi-Mode Comparison.” Political Analysis 22:285–303.
———. 2015. “Distractions: The Incidence and Consequences of Interruptions for Survey
Respondents.” Journal of Survey Statistics and Methodology 3:1–24.
Barabas, Jason, Jennifer Jerit, William Pollock, and Carlisle Rainey. 2015. “The Question(s) of
Political Knowledge.” American Political Science Review 108:840–55.
Berinsky, Adam J., Gregory A. Huber, and Gabriel S. Lenz. 2012. “Using Mechanical Turk as a
Subject Recruitment Tool for Experimental Research.” Political Analysis 20:351–68.
Berinsky, Adam J., Michele Margolis, and Michael Sances. 2014. “Separating the Shirkers from
the Workers? Making Sure Respondents Pay Attention on Internet Surveys.” American Journal
of Political Science 58:739–53.
Bobbio, Andrea, and Anna Maria Manganelli. 2011. “Measuring Social Desirability Responding:
A Short Version of Paulhus’ BIDR 6.” Testing, Psychometrics, Methodology 18:117–35.
Booth-Kewley, Stephanie, Gerald E. Larson, and Dina K. Miyoshi. 2007. “Social Desirability
Effects on Computerized and Paper-and-Pencil Questionnaires.” Computers in Human
Behavior 23:463–77.
Brown, Jonathan D. 2012. “Understanding the Better than Average Effect: Motives (Still) Matter.”
Personality and Social Psychology Bulletin 38:209–19.
Bullock, John G., Alan S. Gerber, Seth J. Hill, and Gregory A. Huber. 2015. “Partisan Bias in
Factual Beliefs about the Economy.” Quarterly Journal of Political Science 10:519–78.
Burnett, Craig, M. 2015. “Exploring the Difference in Performance on Knowledge by
Participants between Online and In-Person Modes.” Working Paper, University of North
Carolina–Wilmington.
Cannell, Charles F., Peter V. Miller, and Lois Oskenbeurg. 1981. “Research on Interviewing
Techniques.” Sociological Methodology 12:389–437.
Chandler, Jesse, Pam Mueller, and Gabriele Paolacci. 2014. “Nonnaiveté among Amazon
Mechanical Turk Workers: Consequences and Solutions for Behavioral Researchers.”
Behavioral Research 46:112–30.
Cialdini, Robert B., John T. Cacioppo, Rodney Bassett, and John A. Miller. 1978. “Low Ball
Procedure for Producing Compliance: Commitment Then Cost.” Journal of Personality and
Social Psychology 36:463–76.
886 Clifford and Jerit
Clifford, Scott, and Jennifer Jerit. 2014. “Is There a Cost to Convenience? An Experimental
Comparison of Data Quality in Laboratory and Online Studies.” Journal of Experimental
Political Science 1:120–31.
———. 2015. “Do Attempts to Improve Respondent Attention Increase Social Desirability
Bias?” Public Opinion Quarterly 79:790–802.
Delli Carpini, Michael X., and Scott Keeter. 1996. What Americans Know about Politics and
Why It Matters. New Haven, CT: Yale University Press.
Miller, Melissa K., and Shannon K. Orr. 2008. “Experimenting with a ‘Third Way’ in Political
Knowledge Estimation.” Public Opinion Quarterly 72:768–80.
Mondak, Jeffery J. 2001. “Developing Valid Knowledge Scales.” American Journal of Political
Science 45:224–38.
Mullinix, Kevin J., Thomas J. Leeper, James N. Druckman, and Jeremy Freese. 2015. “The
Generalizability of Survey Experiments.” Journal of Experimental Political Science
2:109–38.