Misinformation and Mass Audiences
()
About this ebook
Lies and inaccurate information are as old as humanity, but never before have they been so easy to spread. Each moment of every day, the Internet and broadcast media purvey misinformation, either deliberately or accidentally, to a mass audience on subjects ranging from politics to consumer goods to science and medicine, among many others. Because misinformation now has the potential to affect behavior on a massive scale, it is urgently important to understand how it works and what can be done to mitigate its harmful effects.
Misinformation and Mass Audiences brings together evidence and ideas from communication research, public health, psychology, political science, environmental studies, and information science to investigate what constitutes misinformation, how it spreads, and how best to counter it. The expert contributors cover such topics as whether and to what extent audiences consciously notice misinformation, the possibilities for audience deception, the ethics of satire in journalism and public affairs programming, the diffusion of rumors, the role of Internet search behavior, and the evolving efforts to counteract misinformation, such as fact-checking programs. The first comprehensive social science volume exploring the prevalence and consequences of, and remedies for, misinformation as a mass communication phenomenon, Misinformation and Mass Audiences will be a crucial resource for students and faculty researching misinformation, policymakers grappling with questions of regulation and prevention, and anyone concerned about this troubling, yet perhaps unavoidable, dimension of current media systems.
Related to Misinformation and Mass Audiences
Related ebooks
The Quest for Attention: Nonprofit Advocacy in a Social Media Age Rating: 0 out of 5 stars0 ratingsResist the Punitive State: Grassroots Struggles Across Welfare, Housing, Education and Prisons Rating: 0 out of 5 stars0 ratingsInvisible Punishment: The Collateral Consequences of Mass Imprisonment Rating: 4 out of 5 stars4/5Inside Story: How Narratives Drive Mass Harm Rating: 0 out of 5 stars0 ratingsActivist Media: Documenting Movements and Networked Solidarity Rating: 0 out of 5 stars0 ratingsBeyond Complicity: Why We Blame Each Other Instead of Systems Rating: 0 out of 5 stars0 ratingsPro Truth: A Practical Plan for Putting Truth Back Into Politics Rating: 0 out of 5 stars0 ratingsHomesick: Why Housing Is Unaffordable and How We Can Change It Rating: 0 out of 5 stars0 ratingsSoft Corruption: How Unethical Conduct Undermines Good Government and What To Do About It Rating: 0 out of 5 stars0 ratingsOne Nation Undecided: Clear Thinking about Five Hard Issues That Divide Us Rating: 3 out of 5 stars3/5Seen and Unseen: Technology, Social Media, and the Fight for Racial Justice Rating: 4 out of 5 stars4/5When Rape Goes Viral: Youth and Sexual Assault in the Digital Age Rating: 0 out of 5 stars0 ratingsPresumed Incompetent: The Intersections of Race and Class for Women in Academia Rating: 4 out of 5 stars4/5Delivering the People's Message: The Changing Politics of the Presidential Mandate Rating: 0 out of 5 stars0 ratingsWhat Matters Most: Conversations on the Art of Living Rating: 0 out of 5 stars0 ratings@ is for Activism: Dissent, Resistance and Rebellion in a Digital Culture Rating: 0 out of 5 stars0 ratingsDemocracy's Children: Intellectuals and the Rise of Cultural Politics Rating: 3 out of 5 stars3/5W.E.B. Du Bois: Revolutionary Across the Color Line Rating: 0 out of 5 stars0 ratingsWill the Last Reporter Please Turn Out the Lights: The Collapse of Journalism and What Can Be Done to Fix It Rating: 4 out of 5 stars4/5Awakening Democracy through Public Work: Pedagogies of Empowerment Rating: 0 out of 5 stars0 ratingsPenny Red: Notes from the New Age of Dissent Rating: 3 out of 5 stars3/5America Made Me a Black Man: A Memoir Rating: 3 out of 5 stars3/5The Activist's Media Handbook Rating: 4 out of 5 stars4/5Persistent Poverty: Voices From the Margins Rating: 0 out of 5 stars0 ratingsLiberal White Supremacy: How Progressives Silence Racial and Class Oppression Rating: 0 out of 5 stars0 ratingsAuthentic Patriotism: Restoring America's Founding Ideals Through Selfless Action Rating: 5 out of 5 stars5/5Journey into Social Activism: Qualitative Approaches Rating: 0 out of 5 stars0 ratingsSuburban Dreams: Imagining and Building the Good Life Rating: 0 out of 5 stars0 ratingsLincoln and Black Freedom: A Study in Presidential Leadership Rating: 4 out of 5 stars4/5How Elites Ate the Social Justice Movement Rating: 0 out of 5 stars0 ratings
Popular Culture & Media Studies For You
Everybody Lies: Big Data, New Data, and What the Internet Can Tell Us About Who We Really Are Rating: 4 out of 5 stars4/5All About Love: New Visions Rating: 4 out of 5 stars4/5Communion: The Female Search for Love Rating: 4 out of 5 stars4/5Betty Page Confidential: Featuring Never-Before Seen Photographs Rating: 4 out of 5 stars4/5Political Correctness Gone Mad? Rating: 4 out of 5 stars4/5Capitalist Realism: Is There No Alternative? Rating: 5 out of 5 stars5/5The Beauty Myth: How Images of Beauty Are Used Against Women Rating: 4 out of 5 stars4/5The Psychology of Totalitarianism Rating: 4 out of 5 stars4/5Goodbye Phone, Hello World: 60 Ways to Disconnect from Tech and Reconnect to Joy Rating: 4 out of 5 stars4/5Microdosing with Amanita Muscaria: Creativity, Healing, and Recovery with the Sacred Mushroom Rating: 5 out of 5 stars5/5Summary of The Paradox of Choice: by Barry Schwartz - Why More Is Less - A Comprehensive Summary Rating: 0 out of 5 stars0 ratingsThe Art of Libromancy: On Selling Books and Reading Books in the Twenty-first Century Rating: 5 out of 5 stars5/5Regarding the Pain of Others Rating: 4 out of 5 stars4/5The Story of Film Rating: 5 out of 5 stars5/5Burn Book: A Tech Love Story Rating: 4 out of 5 stars4/5Rage Becomes Her: The Power of Women's Anger Rating: 4 out of 5 stars4/5Understanding Media: The Extensions of Man Rating: 4 out of 5 stars4/5The Attention Merchants: The Epic Struggle to Get Inside Our Heads Rating: 0 out of 5 stars0 ratingsThe Truth: Sex, Love, Commitment, and the Puzzle of the Male Mind Rating: 4 out of 5 stars4/5Don't Trust Your Gut: Using Data to Get What You Really Want in LIfe Rating: 3 out of 5 stars3/5Ghosts of My Life: Writings on Depression, Hauntology and Lost Futures Rating: 4 out of 5 stars4/5Powerful Prose: How Textual Features Impact Readers Rating: 0 out of 5 stars0 ratingsSocial Theory after the Internet: Media, Technology, and Globalization Rating: 5 out of 5 stars5/5The Misinformation Age: How False Beliefs Spread Rating: 4 out of 5 stars4/5How Woke Won: The Elitist Movement That Threatens Democracy, Tolerance and Reason Rating: 3 out of 5 stars3/5Tokyo Travel Sketchbook: Kawaii Culture, Wabi Sabi Design, Female Samurais and Other Obsessions Rating: 4 out of 5 stars4/5
Reviews for Misinformation and Mass Audiences
0 ratings0 reviews
Book preview
Misinformation and Mass Audiences - Brian G. Southwell
Information, a series edited by Andrew Dillon
Misinformation and Mass Audiences
EDITED BY BRIAN G. SOUTHWELL, EMILY A. THORSON, AND LAURA SHEBLE
University of Texas Press
Austin
Copyright © 2018 by the University of Texas Press
All rights reserved
First edition, 2018
Requests for permission to reproduce material from this work should be sent to:
Permissions
University of Texas Press
P. O. Box 7819
Austin, TX 78713-7819
utpress.utexas.edu/rp-form
Library of Congress Cataloging-in-Publication Data
Names: Southwell, Brian G. (Brian Glen), 1974–, editor. | Thorson, Emily A., editor. | Sheble, Laura, editor.
Title: Misinformation and mass audiences / edited by Brian G. Southwell, Emily A. Thorson, and Laura Sheble.
Description: First edition. | Austin : University of Texas Press, 2018. | Series: Information | Includes bibliographical references and index.
Identifiers: LCCN 2017025953
ISBN 978-1-4773-1455-5 (cloth : alk. paper)
ISBN 978-1-4773-1456-2 (pbk. : alk. paper)
ISBN 978-1-4773-1457-9 (library e-book)
ISBN 978-1-4773-1458-6 (non-library e-book)
Subjects: LCSH: Mass media—Audiences. | Communication. | Common fallacies—Social aspects. | Deceptive advertising—Social aspects.
Classification: LCC P91.27 .M57 2018 | DDC 302.23—dc23
LC record available at https://lccn.loc.gov/2017025953
doi:10.7560/314555
Contents
Acknowledgments
Introduction: Misinformation among Mass Audiences as a Focus for Inquiry
BRIAN G. SOUTHWELL, EMILY A. THORSON, AND LAURA SHEBLE
PART I: Dimensions of Audience Awareness of Misinformation
ONE: Believing Things That Are Not True: A Cognitive Science Perspective on Misinformation
ELIZABETH J. MARSH AND BRENDA W. YANG
TWO: Awareness of Misinformation in Health-Related Advertising: A Narrative Review of the Literature
VANESSA BOUDEWYNS, BRIAN G. SOUTHWELL, KEVIN R. BETTS, CATHERINE SLOTA GUPTA, RYAN S. PAQUIN, AMIE C. O’DONOGHUE, AND NATASHA VAZQUEZ
THREE: The Importance of Measuring Knowledge in the Age of Misinformation and Challenges in the Tobacco Domain
JOSEPH N. CAPPELLA, YOTAM OPHIR, AND JAZMYNE SUTTON
FOUR: Measuring Perceptions of Shares of Groups
DOUGLAS J. AHLER AND GAURAV SOOD
FIVE: Dimensions of Visual Misinformation in the Emerging Media Landscape
JEFF HEMSLEY AND JAIME SNYDER
PART II: Theoretical Effects and Consequences of Misinformation
SIX: The Effects of False Information in News Stories
MELANIE C. GREEN AND JOHN K. DONAHUE
SEVEN: Can Satire and Irony Constitute Misinformation?
DANNAGAL G. YOUNG
EIGHT: Media and Political Misperceptions
BRIAN E. WEEKS
NINE: Misinformation and Science: Emergence, Diffusion, and Persistence
LAURA SHEBLE
TEN: Doing the Wrong Things for the Right Reasons: How Environmental Misinformation Affects Environmental Behavior
ALEXANDER MAKI, AMANDA R. CARRICO, AND MICHAEL P. VANDENBERGH
PART III: Solutions and Remedies for Misinformation
ELEVEN: Misinformation and Its Correction: Cognitive Mechanisms and Recommendations for Mass Communication
BRIONY SWIRE AND ULLRICH ECKER
TWELVE: How to Counteract Consumer Product Misinformation
GRAHAM BULLOCK
THIRTEEN: A History of Fact Checking in U.S. Politics and Election Contexts
SHANNON POULSEN AND DANNAGAL G. YOUNG
FOURTEEN: Comparing Approaches to Journalistic Fact Checking
EMILY A. THORSON
FIFTEEN: The Role of Middle-Level Gatekeepers in the Propagation and Longevity of Misinformation
JEFF HEMSLEY
SIXTEEN: Encouraging Information Search to Counteract Misinformation: Providing Balanced
Information about Vaccines
SAMANTHA KAPLAN
Conclusion: An Agenda for Misinformation Research
EMILY A. THORSON, LAURA SHEBLE, AND BRIAN G. SOUTHWELL
Contributors
Index
Acknowledgments
This project reflects the work of many authors, both those listed formally and others who helped behind the scenes. Collectively, we would like to thank the external reviewers as well as the University of Texas Press for their enthusiastic help in constructing this volume, especially Robert Devens, Sarah McGavick, and Lynne Chapman, and series editor Andrew Dillon at the University of Texas, who with a light hand nurtured this project as it grew. Many thanks also to freelance copyeditor Tana Silva, indexer Lisa Rivero, and proofreader Melissa McGee Tullos for their outstanding work.
In addition, Brian is grateful for the ongoing patience and support of his family, including Jessica, Gavin, and Ellerie. Emily would like to thank her family and coeditors. Laura thanks her family and friends for their encouragement and uncanny ability to know perfectly just when to share a joke, a walk, a sandwich—and especially their time. We would also like to thank all who have contributed to our understanding of misinformation. Finally, we thank those who are reading these pages for your interest in joining this conversation.
INTRODUCTION
Misinformation among Mass Audiences as a Focus for Inquiry
BRIAN G. SOUTHWELL, EMILY A. THORSON, AND LAURA SHEBLE
Misinformation—both deliberately promoted and accidentally shared—is a noteworthy and perhaps inevitable part of the world in which we live. People likely have lied to one another for roughly as long as verbal communication has existed. Deceiving others can offer an apparent opportunity to gain strategic advantage, to motivate others to action, or even to protect interpersonal bonds. Moreover, people inadvertently have been sharing inaccurate information with one another for thousands of years. We currently live in an era of mass audiences, however, which means that the potential for immediate and widespread misinformation effects now looms larger than in the past. Yet in those same patterns of mass communication and facilitated peer-to-peer information spread might also be the seeds of misinformation correction over time. Investigating the spread and effects of misinformation is as compelling now as it has ever been in human history.
As a focus for research and debate, the phenomenon of misinformation crosses disciplinary bounds; misinformation exists in and across topical domains, from science to politics to consumer advertising, and can quickly spread through a range of media, including television, radio, and pixels shared via social media on the Internet. Understanding misinformation requires working at many levels of analysis, from the policy infrastructure for mass media content to individual information processing ability. Thoroughly understanding and addressing the proliferation, effects, and correction of misinformation calls for a multidisciplinary approach.
In assembling this collection of essays, we have brought together evidence and ideas from communication research, public health, psychology, political science, environmental studies, information science, and other literatures to explore what constitutes misinformation, how it spreads, and how best to counter it. We consider various angles, such as the extent to which audiences consciously notice misinformation, possibilities for audience deception, the ethics of satire in journalism and public affairs programming, the diffusion of rumors, the role of Internet search behavior, and the development of efforts such as fact-checking programs.
Curation of a broad sweep of ideas related to misinformation admittedly risks creation of a smorgasbord that lacks useful argument and thematic coherence. In light of that risk, we have organized contributions into three types of essays—conceptualization and measurement of individual engagement with misinformation, discussion of potential effects, and exploration of remedies—and also have accepted some framing assumptions and boundaries as to what we can discuss in this book and what is best left for future exploration. The consequence of that is a book that simultaneously reminds us that misinformation will continue to pose a dilemma for mass media systems for the foreseeable future and that there are ways we can address concerns. Below, we argue that misinformation exists, that it is problematic in the context of mass audiences and societies for particular reasons, and that the essays we have selected contribute in different ways to thought in this arena, an arena we claim is a worthy one for future inquiry, policy making, and public discussion.
What Is Misinformation?
Misinformation is concerning because of its potential to unduly influence attitudes and behavior, leading people to think and act differently than they would if they were correctly informed. In other words, we worry that misinformation (or false information) might lead people to hold misperceptions (or false beliefs) and that these misperceptions, especially when they occur among mass audiences, may have downstream consequences for health, social harmony, and political life.
Here we are assuming the common existence of mass audiences as Price (1992) and Webster and Phalen (1996) have defined them: more than a dyad or family unit and large enough that members do not necessarily communicate with one another directly but rather are connected primarily through exposure to a common message or piece of information. The possibility that mass audiences would consider false information to be true and would act on that false information implies a serious social problem, in that a mass communication system employs technology to simultaneously and immediately reach and connect large, heterogeneous groups of geographically dispersed people but cannot be fully protected against repeated and widespread claims that are not true. Given this potential problem, we are interested in the diffusion, use, persistence, and refutation of false information on a mass scale.
As Stahl (2006) has noted, critical theorists have raised important concerns about the possibility of information having objective truth value, concerns that complicate any distinction of misinformation and information on the basis of the former being false and the latter being true. On this point, nonetheless, thinking by Habermas (1984, 1987) is useful, as he focuses on a speaker’s intent to deceive and distinguishes between misinformation and disinformation. Habermas views truth as only possible collectively among people as a product of consensus; one’s collegial participation in collective understanding also matters. Misinformation from such a perspective, then, is contentious information reflecting disagreement between people, whereas disinformation is more problematic, as it involves the deliberate alienation or disempowerment of other people. Lewandowsky, Stritzke, Freund, Oberauer, and Krueger (2013) have carried forward this sense of disinformation as intentionally incorrect information.
We respectfully reject a worldview in which no degree of consensus between people is possible. Moreover, misinformation can occur and mislead those exposed, at least relative to what they would have believed if exposed to more complete and accurate information. Habermas’s triad of concepts—information, misinformation, and disinformation—offers a path forward and useful category distinctions for our discussion. If we allow a claim acknowledged by consensus to hold truth value, we then can position misinformation as a category of claim for which there is at least substantial disagreement (or even consensus rejection) when judged as to truth value among the widest feasible range of observers. That approach would include disinformation as a special type of misinformation distinguished by the intent of the promoter. From an ethical perspective, many people will worry most about active promotion of disinformation. Nonetheless, we will use the word misinformation throughout this book to acknowledge that false information can mislead people even if unintentionally promoted or mistakenly endorsed as being true. We acknowledge that such an approach opens the door for certain claims to evolve from being information to becoming misinformation and vice versa over time as a function of a society’s changing consensus. For our purposes, though, we will focus on misinformation as being claims to which mass audiences are exposed that do not actually enjoy universal or near-universal consensus as being true at a particular moment in time.
Mass Audience Exposure to Misinformation as a Cause for Concern
As argued elsewhere (Southwell & Thorson, 2015), at least three observations related to misinformation in the contemporary mass media environment warrant the attention of researchers, policy makers, and laypeople. First, people who encounter misinformation tend to believe it, at least at first. Second, many types of misinformation can appear in content available to mass audiences without being blocked or censored in many media systems. Third, countering misinformation once it has enjoyed wide exposure can be a resource-intensive effort.
What happens when people initially encounter misinformation holds tremendous importance for estimating the potential for subsequent problems. Although individuals generally have considerable routine experience encountering information now considered to be false, the question of exactly how—and when—we should label information as true or false has garnered philosophical debate. As outlined later in this volume, the dilemma is neatly summarized by a contrast between how the philosophers Descartes and Spinoza described human information engagement centuries ago with conflicting predictions that only recently have been empirically tested in robust ways. Descartes argued that a person accepts or rejects information only after considering its truth or falsehood; Spinoza argued that people accept all encountered information (or misinformation) by default and then subsequently verify or reject it through a separate process. Empirical evidence in recent decades has supported Spinoza’s account (Asp & Tranel, 2012; Gilbert, Krull, & Malone, 1990; Gilbert, Tafarodi, & Malone, 1993): people appear to encode all new information as if it were true, even if only momentarily, and later tag the information as being either true or false, a pattern that seems consistent with the observation that skepticism physically resides in a different part of the brain than the resources used in perceiving and encoding. We also know people judge source credibility as a heuristic cue in determining message acceptability and will turn to others for confirmation of the truth value of a claim (Lewandowsky, Ecker, Seifert, Schwarz, & Cook, 2012; Southwell, 2013). If the people surrounding someone tend to initially believe misinformation, then it raises the specter of network reinforcement of misinformation, meaning that misinformation could become even more difficult to debunk over time as the false claim becomes held by more and more people.
What about our claim that misinformation often can appear in electronic or print media without being preemptively blocked? One might consider the nature of regulatory structures in countries such as the United States: regulatory agencies tend to focus on post hoc detection of broadcast information. Organizations such as the U.S. Federal Trade Commission, Federal Election Commission, and Food and Drug Administration offer considerable monitoring and notification functions, but these typically do not involve preemptive censoring. The Food and Drug Administration (FDA) oversees direct-to-consumer prescription drug advertising, for example, and has developed mechanisms such as the Bad Ad
program through which people can report advertising in apparent violation of FDA guidelines on the presentation of drug risks and benefits (O’Donoghue et al., 2015). Such programs, though laudable and useful, do not guarantee that false advertising never appears on the airwaves and, moreover, do not prevent false news stories from appearing. In addition, even misinformation that is successfully corrected can continue to affect attitudes (Thorson 2015).
Lastly, countering misinformation with new information requires effort not only to develop new content that is understandable but also to ensure adequate message exposure. As Hornik (2002) has argued, a communication campaign can succeed or fail, at least in part, as a function of exposure or lack thereof. A campaign to correct misinformation, even if rhetorically compelling, requires resources and planning to accomplish the necessary reach and frequency. For corrective information to be persuasive it needs to be comprehended by audiences, which requires either effort to frame messages in ways that are understandable or effort to educate and sensitize audiences to the possibility of misinformation. That audiences might not be aware of the potential for misinformation also suggests the potential utility of media-literacy efforts as early as elementary school. Even with journalists, pundits, and scholars pointing to the phenomenon of fake news
(Amarasingam, 2011), when scanning and processing information people often do not distinguish between demonstrably false stories and those based in fact.
As we consider mass audience vulnerability to misinformation, we should note how changes in electronic media have suggested to various scholars new possibilities for an individual actor to seed the world with falsehood. Social media offer the potential for a single person in a small town in South Dakota, for example, to reach people directly around the world with a video full of false information. Jenkins (2008) has noted the increased possibilities for user-produced content to be uploaded and shared with other individual information consumers. Such an emphasis on the potential for individual audience members’ agency relative to well-funded and centralized media institutions itself is not new, as media studies scholars have sought for decades to understand how people use media content for their own purposes in their daily lives (Brooker & Jermyn, 2003; Katz, Blumler, & Gurevitch, 1973–1974). Although these technology changes offer paths for problematic diffusion and dissemination of misinformation, evidence suggests that conventional media institutions still play a crucial role regarding message reach and immediacy in cases when information that first appears on a single page somewhere on the Internet subsequently diffuses widely (Weeks & Southwell, 2010). Although any one person can tell another a lie standing on a street corner, the dilemma of widespread misinformation among mass audiences should be understood as a problem of scope and scale not entirely appreciated from the vantage point of a single conversation.
An Overview of the Book
We have organized the chapters to address three important dimensions of misinformation that current and future scholars working in this arena will need to consider as we move forward to understand the phenomenon. First, we have asked a group of researchers to articulate what we know about conceptualizing and measuring audience awareness of misinformation. Second, we have gathered theoretical essays on the ways in which misinformation could have consequences for outcomes that matter to societal governance, individual decision making, and the everyday lives of audience members. Last, we approach the question of potential solutions and remedies for misinformation. Our particular focus on cognition, communication, and policy reflects concern for human engagement with misinformation and consequent potential for dysfunction in human societies. The human brain, interactions of people, and regulations and laws all shape problems and solutions in this arena and will continue to be relevant even as developing information technologies may automate some aspects of future detection, prevention, and correction of misinformation.
Our distinguished contributors for the first section of the book, on dimensions of audience awareness, cover a wide range of topical areas; as they do, some common themes surface about the importance of carefully defining variables and ensuring that measures are valid and reliable. Marsh and Yang approach this concern from a cognitive science perspective, highlighting human biases that favor misinformation acceptance to facilitate information-processing efforts. In this they set the stage for discussion of information encoding, forgetting, and retrieval in a manner that offers useful considerations for the rest of the chapters in the book. Boudewyns and colleagues consider consumer awareness in the specific context of health-related advertising. They find that although some researchers have investigated the prevalence and consequences of problematic claims in advertising, relatively little work to date sheds direct light on audience ability to detect deception. Cappella, Ophir, and Sutton argue for an approach to knowledge measurement that is relevant not only to the tobacco advertising research they have conducted but also to various situations in which knowing whether people hold specific factual claims in mind matters. They argue that many knowledge-based approaches to measuring audience exposure to misinformation are too narrow in scope and fail to take into account alternatives for measuring knowledge that would demonstrate manifestations of misinformation in relevant ways. Ahler and Sood also offer a measurement perspective, theirs grounded in thinking about political issues rather than health. Hemsley and Snyder extend our thinking to consider the possibility of visual misinformation, meaning the possibility for graphics and figures to lead people to misperceptions over and above verbal information. They argue that our current media environment offers numerous opportunities for visual deception.
As we turn to the theoretical consequences of misinformation, a group of experienced researchers reminds us that the ways in which we define and consider misinformation exposure and acceptance, as we do in the first part of the book, affect the range of possibilities for effects. Green and Donahue assess the potential effects of false information in news stories. They look at both immediate misinformation acceptance and persistence of such acceptance over time, and they usefully consider how what we know about audience transportation in the face of narratives helps account for such acceptance. Part of what is intriguing about Green and Donahue’s account is their consideration of news stories, a type of content in which misinformation sometimes can occur unintentionally. What about situations in which an author contorts or ignores some facts in an effort to shed light on truth? One can encounter such a situation with the use of satire and ironic humor. Young considers the possible effects of satire and irony and offers some answers to the question of whether satire and irony even should be categorized as misinformation. With the rise of political satire programs on channels such as Comedy Central and outlets such as the Onion in the United States and elsewhere, Young’s essay should be useful and timely for many years. Weeks focuses on the political realm in his essay on political misperceptions; he outlines factors such as the prevalence of ideologically oriented media content and the proliferation of social media tools as key considerations. Similarly, Maki, Carrico, and Vandenbergh look at misperceptions in the environmental arena. They specifically examine behavior with consequences for the environment, such as hand-washing and energy-saving practices, and they conclude that various fallacies sometimes have led to behavior at odds with recommended practice. Sheble helpfully reminds us that sometimes misperceptions about science stem not from explicit scientific fraud but by the diffusion of scientific information through social networks, a process that can result in the spread of misinterpretations over time.
The last section of the book addresses the difficult question of what, if anything, can be done to solve the problems caused by misinformation. Eliminating the existence of misinformation itself may be an impossible task in the media systems of relatively free and democratic societies, but nonetheless remedies may well exist to counter the effects of misinformation. Understanding whether such corrective efforts can work requires theory on the mechanisms of misinformation acceptance and rejection, for which we can draw on parts of section 1 as well as an artful review of mechanisms for corrective effect by Swire and Ecker. They note that many corrections are ineffective but also point out some reason for optimism. On a similar note, Thorson offers comparative evidence regarding the efficacy of various political campaign fact-checking approaches. She reports compelling reason for third-party fact-checking entities and journalists to get involved with the process, as she finds that adjudication by a fact-checking organization is more likely to offer corrective traction than simply allowing each side of a campaign to fact-check the other. Not all corrective efforts perform equally.
Corrective action is not limited to the political sphere. Researchers, professionals, and volunteers have attempted to develop remedies for misinformation for decades; we have even seen formal use of corrective remedies commissioned by agencies such as the FDA (Aikin et al., 2015). Our discussion would be incomplete without a look at the evolution of political campaign fact checking, and Poulsen and Young offer such an assessment with their discussion of the history of fact-checking initiatives related to political contests in the United States. At the same time, we can learn from the emerging literature on consumer product marketing as well. In this vein, Bullock reviews the consumer product label arena and concludes that regulations need to move beyond current practice to ideally combat fraudulent labeling claims. Strong remedies, however, also will face challenges from those who worry about censorship, as Kaplan notes in her review of efforts to provide balanced information about vaccine safety. The corrective-policy story is still being written, but our discussion nonetheless reveals some of the practical struggles that have been faced in balancing free speech concerns with worry about lying in stump speeches, political ads, or product promotion.
In the third section of the book, we again acknowledge the role that diffusion through networks can play in accounting for misinformation proliferation. A sociological approach not only highlights how information and misinformation move through communities but also might suggest yet another site for mitigation. In this regard, Hemsley’s chapter on what he calls middle-layer gatekeepers,
people who serve as network hubs as information spreads from elite outlets to audience members with few followers, suggests a twist on the prospect of unmitigated diffusion as he notes the power of such information hubs to filter or stop the flow of misinformation. Such an approach harks back to the opinion-leader concept articulated by Katz and Lazarsfeld (1955), offering not just a description of the flow of information but potential remedy for the spread of misperceptions.
Concluding Thoughts
The importance of misinformation as a concern has appeared to grow in recent years as technology affords quicker and more widespread dissemination of information and misinformation while various individual and societal constraints make the presentation and acceptance of misinformation common. At the same time, there is a growing interest in this dilemma from academics and policy makers, as evidenced by the contributions throughout this book. Moreover, technological change and human innovation have evolved in recent years to suggest some possibilities for immediate counterefforts that hold promise. One is real-time fact checking like the U.S. cable news channel CNN did for some presidential speeches in 2016; by immediately posting graphics noting claims that were not factually accurate, the organization leveraged technology to offer a potentially useful frame for viewers interested in the truth, a strategy recommended by Graves (2016). We live in a moment in which widespread misinformation is common. Yet, this is also a time in which many people are passionately developing potential solutions and remedies. The journey forward undoubtedly will be a long one. The way in which media systems have developed in many democratic societies inherently calls for both vulnerability to occasional misinformation and robust systems to detect and address it. Airtight preemptory systems, in fact, ultimately would be suffocating and unappealing. With this collection of essays, then, we hope to inform future efforts in what will be an ongoing struggle that stems from the nature of mass audience systems.
References
Aikin, K. J., Betts, K. R., O’Donoghue, A. C., Rupert, D. J., Lee, P. K., Amoozegar, J. B., & Southwell, B. G. (2015). Correction of overstatement and omission in direct-to-consumer prescription drug advertising. Journal of Communication, 65(4), 596–618.
Amarasingam, A. (2011). The Stewart/Colbert effect: Essays on the real impacts of fake news. Jefferson, NC: McFarland.
Asp, E. W., & Tranel, D. (2012). False tagging theory: Toward a unitary account of prefrontal cortex function. In D. T. Stuss & R. T. Knight (Eds.), Principles of frontal lobe function (pp. 383–416). New York, NY: Oxford University Press.
Brooker, W., & Jermyn, D. (Eds.). (2003). The audience studies reader. London, UK: Routledge.
Gilbert, D. T., Krull, D. S., & Malone, P. S. (1990). Unbelieving the unbelievable: Some problems in the rejection of false information. Journal of Personality and Social Psychology, 59, 601–613.
Gilbert, D., Tafarodi, R., & Malone, P. (1993). You can’t not believe everything you read. Journal of Personality and Social Psychology, 65(2), 221–233.
Graves, L. (2016). Deciding what’s true: The rise of political fact-checking in American journalism. New York, NY: Columbia University Press.
Habermas, J. (1984). The theory of communicative action: Vol. 1. Reason and the rationalization of society (T. McCarthy, Trans.). Boston, MA: Beacon Press.
Habermas, J. (1987). The theory of communicative action: Vol. 2. Lifeworld and system: A critique of functionalist reason (T. McCarthy, Trans.). Boston, MA: Beacon Press.
Hornik, R. C. (Ed.). (2002). Public health communication: Evidence for behavior change. Mahwah, NJ: Lawrence Erlbaum.
Jenkins, H. (2008). Convergence culture: Where old and new media collide. New York, NY: New York University Press.
Katz, E., Blumler, J. G., & Gurevitch, M. (1973–1974). Uses and gratifications research. Public Opinion Quarterly, 37(4), 509–523.
Katz, E., & Lazarsfeld, P. F. (1955). Personal influence: The part played by people in the flow of mass communications. Glencoe, IL: Free Press.
Lewandowsky, S., Ecker, U. K. H., Seifert, C. M., Schwarz, N., & Cook, J. (2012). Misinformation and its correction: Continued influence and successful debiasing. Psychological Science in the Public Interest, 13(3), 106–131.
Lewandowsky, S., Stritzke, W. G. K., Freund, A. M., Oberauer, K., & Krueger, J. I. (2013). Misinformation, disinformation, and violent conflict: From Iraq and the War on Terror
to future threats to peace. American Psychologist, 68(7), 487–501.
O’Donoghue, A. C., Boudewyns, V., Aikin, K. J., Geisen, E., Betts, K. R., & Southwell, B. G. (2015). Awareness of FDA’s Bad Ad program and education regarding pharmaceutical advertising: A national survey of prescribers in ambulatory care settings. Journal of Health Communication, 20(11), 1330–1336.
Price, V. (1992). Public opinion. Thousand Oaks, CA: Sage.
Southwell, B. G. (2013). Social networks and popular understanding of science and health: Sharing disparities. Baltimore, MD: Johns Hopkins University Press.
Southwell, B. G., & Thorson, E. A. (2015). The prevalence, consequence, and remedy of misinformation in mass media systems. Journal of Communication, 65(4), 589–595.
Stahl, B. C. (2006). On the difference or equality of information, misinformation, and disinformation: A critical research perspective. Informing Science Journal, 9, 83–96.
Thorson, E. (2015). Belief echoes: The persistent effects of corrected misinformation. Political Communication, 1–21.
Webster, J., & Phalen, P. F. (1996). The mass audience: Rediscovering the dominant model. New York, NY: Routledge.
Weeks, B., & Southwell, B. (2010). The symbiosis of news coverage and aggregate online search behavior: Obama, rumors, and presidential politics. Mass Communication and Society, 13(4), 341–360.
PART I
Dimensions of Audience Awareness of Misinformation
ONE
Believing Things That Are Not True
A Cognitive Science Perspective on Misinformation
ELIZABETH J. MARSH AND BRENDA W. YANG
One of these rather silly anecdotes is an actual news story: some newer Canadian hundred-dollar bills smell like maple syrup; England is considering issuing a coin featuring the pop band One Direction; the U.S. Treasury recently introduced Perry the Pyramid, a terrifying one-eyed mascot for the dollar. Choosing the real story was the task of a college student who called in to the Bluff the Listener
game on the National Public Radio program Wait Wait . . . Don’t Tell Me! (Danforth, 2013). He won the game by correctly selecting the true but rather obscure story about scented Canadian currency. How did the listener make this choice, given it was unlikely he had the relevant information in mind to make that decision?
In this chapter, we review the cognitive strategies and heuristics people use when deciding whether something is true. Our approach to understanding this issue is an experimental one, with the goal of isolating particular mechanisms that contribute to illusions of truth and the propagation of falsehoods. Many of the misconceptions covered in this volume are powerful precisely because they result from combinations of mental processes; there is not one simple trick to convincing people that Barack Obama was not born in the United States, that climate change is a hoax, or that other claims percolating through mass media are unsupported. Here, we consider how statements can be manipulated to seem more truthful than they are, why people unwittingly trust information from sources they initially knew to be unreliable, and how certain features of claims and arguments increase their persuasiveness; our objective in revealing these processes is to inform our overall understanding of misinformation in and out of the laboratory.
An Assumption of Truth
We begin with a very basic issue, namely, that even the most skeptical people have a bias toward accepting what they hear as true. This partly comes out of a need to have successful conversations; to do so, a person shapes what she says to be understood by those around her and assumes others are doing the same. Thus people expect what others say to be truthful, relevant, informative but not overly so, and clear (Grice, 1975). While these guidelines can be stretched or even broken—for instance, through deception, comedy, tall tales, and long academic lectures—they describe what people expect from others in everyday discourse; violations can be noticed in children as young as three years old (Eskritt, Whalen, & Lee, 2008). Our focus here is on the expectation that speakers are truthful. While this assumption is cognitively efficient compared to a strategy of evaluating each and every claim that comes one’s way, it can become problematic.
At issue is that people are often willing to be flexible in how they define truth.
We found evidence for this in a study in which Stanford undergraduates tracked instances when they shared their memories with others (Tversky & Marsh, 2000). Over the course of four weeks, students submitted more than a thousand reports of such conversations. Consistent with conversational norms not to tell more than the listener needs to know, more than a third of retellings were reported to be selective in some way. For example, a student sharing a story about an annoying dinner guest focused on the guest’s criticisms of the cooking, omitting how the guest also commented on how another one of my appetizers was tasty. . . . I mentioned everything except the positive comment [when sharing this story]. . . . I was already too annoyed to let it change my reaction.
Intriguingly, this speaker also labeled this retelling complete and accurate,
presumably because the story was consistent with the overall theme of the event. In this study, a third of retellings containing distortions like omissions and elaborations were considered accurate by the teller. As Neisser (1981) has noted, truth and accuracy are