Academia.eduAcademia.edu

Why Expertise Matters: A Response to the Challenges

2017, IEEE INTELLIGENT SYSTEMS

Our society depends on experts for mission-critical, complex technical guidance for high-stakes decision making because they can make decisions despite incomplete, incorrect, and contradictory information when established routines no longer apply. Experts are the people the team turns to when faced with difficult tasks.Despite this empirical base, we witness a number of challenges to the concept of expertise.

HUMAN-CENTERED COMPUTING Editor: Robert R. Hoffman, Institute for Human and Machine Cognition, [email protected] Why Expertise Matters: A Response to the Challenges Gary Klein, MacroCognition LLC Ben Shneiderman, University of Maryland Robert R. Hoffman and Kenneth M. Ford, Institute for Human and Machine Cognition We dedicate this article to our colleague Robert Wears, who tragically died in July 2017 just before we started to work on this article. O verwhelming scientific evidence demonstrates that experts’ judgments can be highly accurate and reliable. As defined in the scientific literature,1 experts • employ more effective strategies than others, and do so with less effort; • perceive meaning in patterns that others do not notice; • form rich mental models of situations to support sensemaking and anticipatory thinking; • have extensive and highly organized domain knowledge; and • are intrinsically motivated to work on hard problems that stretch their capabilities. Our society depends on experts for mission-critical, complex technical guidance for high-stakes decision making because they can make decisions despite incomplete, incorrect, and contradictory information when established routines no longer apply.2 Experts are the people the team turns to when faced with difficult tasks. Despite this empirical base, we witness a number of challenges to the concept of expertise. Tom Nichols’ The Death of Expertise presents a strong defense of expertise,3 a defense to which we are adding in this article. We address the attempts NOVEMBER/DECEMBER 2017 made by five communities to diminish the credibility and value of experts (see Figure 1). These challenges come from • decision researchers who show that simple linear models can outperform expert judgment; • heuristics and biases researchers who have claimed that experts are as biased as anyone else; • sociologists who see expertise as just a social attribution; • practice-oriented researchers seeking to replace professional judgments with data-based prescriptions and checklists; and • technophiles who believe that it is only a matter of time before artificial intelligence (AI) surpasses experts. Each of these communities has questioned the value of expertise, using different arguments, perspectives, and paradigms. Society needs experts, even though they are fallible. Although we are expert-advocates, eager to highlight the strengths of experts, we acknowledge that experts are not perfect and never will be. Our purpose is to correct the misleading claims and impressions being spread by the expertise-deniers. Then we hope to engage productively with each of these communities to improve human performance through better training, workflows, and technology. We begin with the challenge from the decision research community because this body of work can be traced back the furthest, to the mid-1960s, and echoes to this day. 1541-1672/17/$33.00 © 2017 IEEE Published by the IEEE Computer Society 67 Challenges to expertise Evidence-based practices Heuristics and biases Computer science Expertise Sociology Decision research Figure 1. The five communities that have challenged the concept of expertise. The Challenge from the Decision Research Community Research by some experimental psychologists shows that in judgment tasks, simple linear models will be more consistent in their performance than human judges. Examples are faculty ratings of graduate students versus a model based on grades and test scores, or physicians’ ratings of cancer biopsy results versus a model based on survival statistics: 4,5 There are some, but only a few, truly remarkable judges, whereas there are many so-called experts who are no better than complete novices … the picture of the expert painted in broad brush strokes by this research is relatively unflattering … whenever possible, human judges should be replaced by linear models.6 A few aspects of this literature are noteworthy: • The linear models are derived in the first place from the advice of experts about what the key variables are—the variables that experts themselves use in making judgments. 68 • The decision research tends to reduce expertise to single measures, such as judgment hit rate, ignoring more qualitative contributions to performance. • For many of the studies, it is not obvious that the particular judgment task that is presented to the participants is actually the same task that the experts routinely perform, and therefore might not be the task at which they are proficient. • For many of the studies, there is scant evidence that the participants who are called experts actually qualify for that designation, apart from their having had a certain numbers of years of experience. • Advocates of this view go beyond their empirical base by generalizing from studies using college students as judges to argue for the fallibility of all judges. • Although linear models are consistent, when they fail, they fail miserably. One problem involves “broken leg cues.” A linear model might do a decent job of predicting whether a given person is likely to go to the movies this weekend, but will fail because it is blind to the fact that the person in question just broke a leg.7 Experts will perform better than the model if they have information to which the model is insensitive. 8 Who are the experts? In some studies, the linear models were compared to college students, sometimes called “judges.” And even in studies in which the judges were professionals, perhaps the linear models should be compared to the best of the professionals rather than to the average. Even when the linear models outperformed the experts, it is a mistake to infer that the linear models got it right and the experts failed miserably. www.computer.org/intelligent The linear models had their greatest edge in domains6 and prediction tasks involving human activity (that is, clinical psychologists, psychiatrists, counselors, admissions officers, parole officers, bank loan officers, and so on). However, the linear models weren’t very accurate—it was just that the experts were even worse. In one often-cited study of cancer diagnosis, the linear model performed better than oncologists at predicting patient longevity, but a closer look shows that the model only accounted for 18 percent of the variance in the judgment data.9 The clearest conclusion from this and other studies on linear modeling is that some things are of intrinsic low predictability. Next is the challenge from the heuristics and biases (HB) community, which can be traced to the early 1970s. The Challenge from the Heuristics and Biases Community Led by Daniel Kahneman and Amos Tversky,10 the HB community has called into question assumptions about rationality by demonstrating that people fall prey to a wide variety of biases in judgment and probabilistic reckoning, and that even experts sometimes show these biases. This finding helped create a mindset that experts are not to be trusted. The proliferation of HB research in academic psychology departments has strengthened the impression that expert judgments are not accurate. The HB paradigm typically uses participants who are not experts (college students) and gives them artificial laboratory tasks that require little training and have little or no ecological validity. The tasks, conveniently enough, can be performed in a college class period. Bias effects found using this “paradigm of convenience” IEEE INTELLIGENT SYSTEMS diminish or disappear when researchers add context11 or when researchers have genuine experts engage in their familiar environments rather than work on artificial puzzles and probability-juggling tasks. Variations in the materials, instructions, procedures, or experimental design can cause bias effects to diminish or disappear.12 Although some studies have shown that bias can occur in expert reasoning,13 several studies show that bias effects are much smaller than those of the college students.14 There is mixed evidence for the claim that experts tend to be overconfident, and what evidence there is stems from narrow methods for measuring confidence. Experts such as weather forecasters and firefighters are careful to keep their judgments within their core specialty and to use experience and accurate feedback to attain reasonable levels of confidence in their judgments. Weather forecasters search for evidence that confirms a hypothesis; it would be irrational not to. On the other hand, weather forecasters deliberately and deliberatively look for evidence that their hypotheses might be wrong. The HB researchers’ antipathy toward experts opened the way for the emergence of the naturalistic decision-making movement,15 which regards heuristics as strengths acquired through experience, rather than weaknesses. Tversky and Kahneman were careful to state that, “In general these heuristics are quite useful, but sometimes they lead to severe and systematic errors.”16 However, the HB field usually ignores this caveat and emphasizes the downside of heuristics. We now come to the challenge from sociology, which began in the 1970s and emerged most forcefully in the 1980s. NOVEMBER/DECEMBER 2017 The Challenge from Sociology Sociological analysis of the consequences of occupational specialization has considered the value of professions to society.17 Given its close association to the concept of professions, the concept of expertise was also assessed from the sociological perspective, referred to as “science and technology studies.” Ethnographers, sociologists, and philosophers of science researched expertise in domains including astronomy, physics, and endocrinology.18–20 Their resonant paradigms have been referred to as “situated cognition,” “distributed cognition,” and the “sociology of scientific knowledge.”21–24 Some individuals have defined their paradigm, in part, as a reaction to cognitive psychology: If one relegates all of cognition to internal mental processes, then one is required to pack all the explanatory machinery of cognition into the individual mind as well, leading to misidentification of the boundaries of the cognitive system, and the over-attribution to the individual mind alone all of the processes that give rise to intelligent behavior. 25 Proponents of the situated cognition approach offer many good examples of why one should define “the cognitive system” as persons acting in coordination with a social group to conduct activities using tools and practices that have evolved within a culture.25 The core claim is that expertise and cognition reside in the interaction among the individual and the team, community, or organization. The strongest view is that expertise is a social attribution or role, a matter of prestige and authority. A moderate view is that individual cognition is an enabling condition for expertise, which just happens to be a www.computer.org/intelligent condition that is not of particular interest in a sociological analysis. One of the valuable aspects of this perspective is to sensitize us to the importance of external resources and community relationships for the acquisition, expression, and valuation of expertise. Thus, we respect these researchers and their contributions. The importance of context has been recognized in cognitive psychology for decades, 26 and in computer science as well.27 We agree completely that resources for cognition are in the world. We agree that teamwork and organizational issues are an important part of naturalistic decision making. Indeed, the notion of “macrocognition”28 refers to coordinating and maintaining common ground as primary functions. However, cognitive scientists are disappointed by any approach that takes the strong stance that expertise is merely a social attribution, a stance that discounts the importance and value of individual cognition, knowledge, and expertise.27 There is overwhelming empirical evidence that individual knowledge is crucial in expert reasoning and problem solving. If you plug experts and nonexperts into the same work settings you will find huge differences in the quality of the outputs of the groups/teams. The claims derived from a dismissive reaction to cognitiveindividualist views move the pendulum too far. Sociologists including Harry Collins29 and Harald Mieg30 have taken the balanced view, which we advocate, describing the importance of individual expertise along with social and contextual factors that can be essential for developing and maintaining expertise. We remain hopeful that over time, this balanced view will predominate. We now come to challenges that have emerged most recently. 69 The Challenge from the Evidence-Based Practices Community The evidence-based practices community argues that professionals need to find the best scientific evidence, derive prescriptive procedures for decisions, and adhere to these procedures rather than rely on their own judgments.31 This approach has been advocated in healthcare, where it is referred to as evidence-based medicine. This community argues against trusting experts because they rely on anecdotal practices and out-ofdate and ineffective remedies. This takeaway message seeks to replace reliance on experts with faith in defined scientific studies. Clearly, empirical evaluation studies have great value, but we do not believe that such studies deserve uncritical acceptance. Witness how the evidence seems to change every two years. Witness also the difficulty of sorting out the evidence base for a patient with multiple medical problems. Clinicians must consider the individual patient, who may differ from the criteria on which the evidence is based. We seek a balance between scientific evidence and broad experience.32 One way that evidence is compiled is through checklists. These are valuable safety tools to prevent decision makers from omitting important steps in a process, but they are not decision-support tools. We believe that reducing complex judgments to simple checklists often misses essential aspects of decision making.33 Checklists work for stable, welldefined tasks, and have to be carefully crafted with a manageable number of steps. If the checklist is sequential, each step must lead to a clear outcome that serves as the trigger for the next step. However, in complex and ambiguous situations, the antecedent conditions for each step are likely 70 to be murky; expert decision makers must determine when to initiate the next step or whether to initiate it at all. Although checklists can be helpful, it is risky to have individuals use checklists for complex tasks that depend on considerable tacit knowledge to judge when a step is appropriate, how to modify a step, and how to decide whether the checklist is working.34 Experts must decide what to do when the best practices conflict with their own judgments. They must revise plans that do not seem to be working. It is one thing to hold physicians to task for relying on ineffective remedies and ignoring scientific evidence that the procedures that they were once taught that have since been shown ineffective, but it is another thing to compel physicians to rely on scientific evidence by proceduralizing clinical judgment in a checklist and penalize them for not following the steps. Guidelines, rules, and checklists raise the floor by preventing silly errors—mistakes that even a firstyear medical student might recognize as an error. But they also lower the ceiling, making it easy to shift to an unthinking, uncritical mode that misses subtle warning signs and does not serve the needs of patients. Finally, we come to the challenge from within computer science itself. The Challenge from Computer Science This challenge has been presented on three fronts: AI, big data, and automation. It is claimed that these technologies are smarter and more reliable than any human. Since experts are the gold standard of performance, demonstrations of smart technology win big when they beat out an expert. AI successes have been widely publicized. IBM’s Deep Blue beat Garry Kasparov, the reigning chess www.computer.org/intelligent champion at the time. IBM’s Watson beat a panel of experts at the game of Jeopardy. AlphaGo trounced one of the most highly regarded Go masters. These achievements have been interpreted as showing that AI can outperform humans at any cognitively challenging task. But the successes involve games that are well-structured, with unambiguous referents and definitive correct answers. In contrast, most decision makers face wicked problems with unclear goals in ambiguous and dynamic situations. Roger Schank, an AI pioneer, stated flatly that “Watson is a fraud.”35 He objected to IBM’s claims that Watson could outthink humans and find insights within large datasets. Although Watson excels at keyword searches, it does not consider the context of the passages it is searching, and as a result is insensitive to underlying messages in the material. Schank’s position is that counting words is not the same as inferring insightful conclusions. Our experience is that AI developers have much greater appreciation for human expertise than the AI popularizers. A good example of the challenge to expertise comes from the weather forecasting domain. Articles with titles such as “All Hail the Computer!”36 promulgate the myth that if more memory and faster processing speeds could be thrown at the problem, the need for humans would evaporate. Starting in the late 1980s, as more computer models were introduced into operational forecasting, prognostications were made that computer models would outperform humans within the next 10 years—for example, “[The] human’s advantage over the computer may eventually be swamped by the vastly increased number crunching ability of the computer ... as the computer driven models will simply get bigger IEEE INTELLIGENT SYSTEMS and better.”37 Articles in the scientific literature as well as the popular press continue to present the stance of human versus machine, asking whether “machines are taking over.”36 This stance conveys a counterproductive attitude of competition in which the experts cannot beat the computers. A more productive approach would be to design technologies that enhance human performance. The evidence clearly shows that the expert weather forecaster adds value to the outputs of the computer models. Furthermore, “numerical prediction models do not produce a weather forecast. They produce a form of guidance that can help a human being decide upon a forecast of the weather.”38,39 Next, we turn to the denigration of expertise that has been expressed by advocates of big data analytics. Despite their widely publicized successes, a closer look often tells a different story. For instance, Google’s FluTrends project initially seemed successful at predicting flu outbreaks, but over time it misled public health planners.40 Advocates of big data claim that the algorithms can detect trends, spot problems, and generate inferences and insights; no human, no matter how expert, could possibly sift through all of the available sensor data; and no human can hope to interpret even a fraction of these data sources. These statements are all true. But the big data community wants to reduce our trust in domain experts so decision makers become comfortable using automated big data analyses. Here is a typical and dangerous claim: “The big target here isn’t advertising, though. It’s science … faced with massive data, this approach to science—hypothesize, model, test—is becoming obsolete … Petabytes allow us to say: Correlation is enough. We can stop looking for models.”41 NOVEMBER/DECEMBER 2017 A balanced view recognizes that big data analytics can identify patterns where none exist. Big data algorithms can follow historical trends but might miss departures from these trends, as in the broken leg cues, cues that have implications that are clear to experts but aren’t part of the algorithms. Further, experts can use expectancies to spot missing events that may be highly significant. In contrast, big data approaches, which crunch the signals received from a variety of sources, are unaware of the absence of data and events. Finally, we consider the challenge offered by proponents of automation. Some researchers in the automation community have promulgated the myth that more automation can obviate the need for humans, including experts. The enthusiasm for technologies is often extreme.42 Too many technologists believe that automation can compensate for human limitations and substitute for humans. They also believe the myth that tasks can be cleanly allocated to either the human or the machine. These misleading beliefs have been questioned by cognitive systems engineers for more than 35 years, yet the debunking has to be periodically refreshed in the minds of researchers and program managers.43 The misleading beliefs persist because of the promissory note that more automation means fewer people, fewer people means fewer errors, and (especially) fewer people means reduced costs.44 Nearly every funding program that calls for more automation is premised with the claim that the introduction of automation will entail a need for fewer expert operators at potentially lower cost to the organization. But the facts are in plain view: The introduction often requires more experts. Case studies 45 show that automation creates new kinds of cognitive work www.computer.org/intelligent for the operator, often at the wrong times. Automation often requires people to do more, to do it faster, or to do it in more complex ways. The explosion of features, options, and modes often creates new demands, new types of errors, and new paths toward failure. Ironically, as these facts became apparent, decision makers seek additional automation to compensate for the problems triggered by the automation.44 We see technology—AI, big data, and automation—continuing to improve, which will make computers ever more valuable tools. But in the spirit of human-centered computing, we define intelligent systems as human-machine systems that amplify and extend human abilities.46 The technology in such work systems is designed to improve human performance and accelerate the achievement of expertise. We hope that expertise-deniers can get past the mindset of trying to build systems to replace the experts and instead seek to build useful technologies that empower experts. If the challenges to expertise hold sway, the result might be degradation of the decision making and resilience of organizations and agencies. Once such organizations accept the expertise-deniers’ arguments, they may sideline domain experts in favor of statistical analysts and ever more automation. They are likely to divert funding from training programs that might produce experts into technology that makes decisions without human intervention or responsible action. Shifting cognitive work over to automation may deskill workers, erode the expertise that is crucial for adaptability, and lead to a downward spiral of diminishing expertise. Experts are certainly not perfect, so the challenges can be useful for increasing our understanding of the 71 boundary conditions of expertise. We do not want to return to an era where medicine was governed by anecdote rather than data—we think it essential to draw from evidence and from expertise. We appreciate the discoveries of the heuristics and biases researchers—the heuristics they have uncovered can have great value for fostering speculative thinking. We respect the judgment and decision research community—we want to take advantage of their efforts to improve the way we handle evidence and deploy our intuitions. We want to productively move forward with improved information technology—we want these tools to be designed to help us gain and enhance expertise. We value the social aspects of work settings—we want to design work settings and team arrangements that magnify expertise. Our hope is to encourage a balance that respects expertise while designing new ways to strengthen it. We regard the design of cognitive work systems as the design of humanmachine interdependencies, guided by the desire to make the machines comprehensible, predictable, and controllable. This course of action seems best suited to promote human welfare and enable greater achievements.47,48 Acknowledgments We thank Bonnie Dorr, Hal Daume, Jonathan Lazar, Jim Hendler, Mark Smith, and Jenny Preece for their comments on a draft of this article; and Jan Maarten Schraagen and Paul Ward for their helpful comments and suggestions and for their patience and encouragement. This essay was adapted from a longer and more indepth account, “The War on Experts,” that will appear in the Oxford Handbook of Expertise. 49 References 1. K.A. Ericsson et al., Cambridge Handbook of Expertise and Expert 72 Performance, 2nd ed., Cambridge Univ. Press, 2017. 2. B. Shneiderman and G. Klein, “Tools That Aid Expert Decision Making: Supporting Frontier Thinking, Social Engagement and Responsibility,” blog, Psychology Today, Mar. 2017; www .psychologytoday.com/blog/seeing-what -others-dont/201703/tools-aid-expert -decision-making-rather-degrade-it. 3. T. Nichols, The Death of Expertise, Oxford Univ. Press, 2017. 4. R. Dawes, “The Robust Beauty of Improper Linear Models,” American Psychologist, vol. 34, no. 7, 1979, pp. 571–582. 5. P.E. Meehl, “Seer Over Sign: The First Good Example,” J. Experimental Research in Personality, vol. 1, no. 1, 1965, pp. 27–32. 6. R. Hastie and R. Dawes, Rational Choice in an Uncertain World, Sage Publications, 2001. 7. K. Salzinger, “Clinical, Statistical, and Broken-Leg Predictions,” Behavior and Philosophy, vol. 33, 2005, pp. 91–99. 8. R. Johnston, Analytic Culture in the U.S. Intelligence Community: An Ethnographic Study, Center for the Study of Intelligence, Washington, DC, 2005. 9. H.J. Einhorn and R.M. Hogarth, “Judging Probable Cause,” Psychological Bull., vol. 99, no. 1, 1978, pp. 3–19. 10. D. Kahneman and A. Tversky, “Prospect Theory: An Analysis of Decision under Risk, Econometrica, vol. 47, no. 2, 1979, pp. 263–291. 11. D.W. Cheng et al., “Pragmatic Versus Syntactic Approaches to Training Deductive Reasoning,” Cognitive Psychology, vol. 18, no. 3, 1986, pp. 293–328. 12. R. Hertwig and G. Gigerenzer, “The ‘Conjunction Fallacy’ Revisited: How Intelligent Inferences Look like Reasoning Errors,” J. Behavioral Decision Making, vol. 12, no. 2, 1999, pp. 27–305. 13. B. Fischhoff, “Eliciting Knowledge for Analytical Representation,” IEEE Trans. Systems, Man, and Cybernetics, vol. 19, no. 3, 1989, pp. 448–461. www.computer.org/intelligent 14. M.D. Shields, I. Solomon, and W.S. Waller, “Effects of Alternative Sample Space Representations on the Accuracy of Auditors’ Uncertainty Judgments,” Accounting, Organizations, and Society, vol. 12, no. 4, 1987, pp. 375–385. 15. G. Klein, R. Calderwood, and A. ClintonCirocco, “Rapid Decision Making on the Fire Ground,” Proc. Human Factors and Ergonomics Soc. Ann. Meeting, vol. 30, no. 6, 1986, pp. 576–580. 16. A. Tversky and D. Kahneman, “Judgment under Uncertainty: Heuristics and Biases,” Science, vol. 185, Sept. 1974, pp. 1124–1131. 17. J. Evetts, “Professionalism: Value and Ideology,” Current Sociology, vol. 61, no. 5–6, 2013, pp. 778–779. 18. H.M. Collins, Changing Order. Replication and Induction in Scientific Practice, 2nd ed., Univ. of Chicago Press, 1992. 19. B. Latour and S. Woolgar, Laboratory Life. The Social Construction of Scientific Facts, Sage Publications, 1979. 20. M. Lynch, Scientific Practice and Ordinary Action, Cambridge Univ. Press, 1993. 21. K.D. Knorr-Cetina, The Manufacture of Knowledge, Pergamon Press, 1981. 22. J. Lave, “Situating Learning in Communities of Practice,” Perspectives on Socially Shared Cognition, L.B. Resnick, J.M. Levine, and S.D. Teasley, eds., American Psychological Assoc., 1993, pp. 63–82. 23. L. Suchman, Plans and Situated Actions: The Problem of Human-Machine Communication, Cambridge Univ. Press, 1987. 24. E. Wenger, Communities of Practice: Learning, Meaning, & Identity, Cambridge Univ. Press, 1998. 25. M.S. Weldon, “Remembering as a Social Process,” The Psychology of Learning and Motivation, vol. 40, no. 1, 2000, pp. 67–120. 26. G.A. Miller, “Dismembering Cognition,” One Hundred Years of Psychological Research in America, IEEE INTELLIGENT SYSTEMS Johns Hopkins Univ. Press, 1986, pp. 277–298. 27. N.M. Agnew, K.M. Ford, and P.J. Hayes, “Expertise in Context: Personally Constructed, Socially Selected and Reality-Relevant?” Int’l J. Expert Systems, vol. 7, no. 1, 1994, pp. 65–88. 28. G. Klein et al., “Macrocognition,” IEEE Intelligent Systems, vol. 18, no. 3, 2003, pp. 81–85. 29. H.M. Collins, “A Sociological/ Philosophical Perspective on Expertise: The Acquisition of Expertise Through Socialization,” Cambridge Handbook of Expertise and Expert Performance, 2nd ed., K.A. Ericsson et al., Cambridge Univ. Press, 2017. 30. H.A. Mieg, “Social and Sociological Factors in the Development of Expertise,” Cambridge Handbook of Expertise and Expert Performance, K.A. Ericsson et al., Cambridge Univ. Press, 2006, pp. 743–760. 31. A.R. Roberts and K.R. Yeager, eds., Evidence-Based Practice Manual: Research and Outcome Measures in Health and Human Services, Oxford Univ. Press, 2004. 32. E. Barends, D.M. Rousseau, and R.B. Briner, Evidence-Based Management: The Basic Principles, Center for Evidence-Based Management, Amsterdam, 2014. 33. R.L. Wears and G. Klein, “The Rush from Judgment,” Annals of Emergency Medicine, forthcoming, 2017. 34. D.E. Klein et al., “Can We Trust Best Practices? Six Cognitive Challenges of Evidence-Based Approaches,” J. Cognitive Eng. and Decision Making, vol. 10, no. 3, 2016, pp. 244–254. 35. R. Schank, “The Fraudulent Claims Made by IBM about Watson and AI,” 2015; www.rogerschank.com /fraudulent-claims-made-by-IBM -about-Watson-and-AI. 36. R.A. Kerr, “Weather Forecasts Slowly Clearing Up,” Science, vol. 38, no. 388, 2012, pp. 734–737. NOVEMBER/DECEMBER 2017 37. P.S. Targett, “Predicting the Future of the Meteorologist: A Forecaster’s View,” Bull. Australian Meteorological and Oceanographic Soc., vol. 7, no. 1, 1994, pp. 46–52. 38. H.E. Brooks, C.A. Doswell, and R.A. Maddox, “On the Use of Mesoscale and Cloud-Scale Models in Operational Forecasting,” Weather and Forecasting, vol. 7, Mar. 1992, pp. 120–132. 39. R.R. Hoffman et al., Minding the Weather: How Expert Forecasters Think, MIT Press, 2017. 40. D. Lazer et al., “The Parable of Google Flu: Traps in the Big Data Analysis,” Science, vol. 343, 14 Mar. 2014, pp. 1203–1205. 41. C. Anderson, “The End of Theory: The Big Data Deluge Makes the Scientific Method Obsolete,” Wired, 23 June 2008; www.wired.com/2008/06 /pb-theory. 42. E. Brynjolfsson and A. McAfee, The Second Machine Age: Work, Progress, and Prosperity in a Time of Brilliant Technologies, W.W. Norton & Co., 2014. 43. J.M. Bradshaw et al., “The Seven Deadly Myths of ‘Autonomous Systems’,” IEEE Intelligent Systems, vol. 28, no. 3, 2013, pp. 54–61. 44. “Technical Assessment: Autonomy,” report from the Office of the Assistant Secretary of Defense for Research and Engineering, Office of Technical Intelligence, US Department of Defense, 2015. 45. R.R. Hoffman, T.M. Cullen, and J.K. Hawley, “Rhetoric and Reality of Autonomous Weapons: Getting a Grip on the Myths and Costs of Automation,” Bull. Atomic Scientists, vol. 72, no. 4, 2016; doi:10.1080/00963402.2016.1194619. 46. J.M. Johnson et al., “Beyond Cooperative Robotics: The Central Role of Interdependence in Coactive Design,” IEEE Intelligent Systems, vol. 26, no. 3, 2011, pp. 81–88. 47. M. Johnson et al., “Seven Cardinal Virtues of Human-Machine Teamwork,” IEEE Intelligent Systems, vol. 29, no. 6, 2014, pp. 74–79. www.computer.org/intelligent 48. G. Klein et al., “Ten Challenges for Making Automation a ‘Team Player’ in Joint Human-Agent Activity,” IEEE Intelligent Systems, vol. 19, no. 6, 2004, pp. 91–95. 49. P. Ward et al., eds., Oxford Handbook of Expertise, Oxford Univ. Press, forthcoming. Gary Klein is senior scientist at MacroCognition LLC. His research interests include naturalistic decision making. Klein received his PhD in experimental psychology from the University of Pittsburgh. Contact him at [email protected]. Ben Shneiderman is distinguished university professor in the Department of Computer Science at the University of Maryland. His research interests include human-computer interaction, user experience design, and information visualization. Shneiderman has a PhD in computer science from SUNY-Stony Brook. Contact him at [email protected]. Robert R. Hoffman is a senior research scientist at the Institute for Human and Machine Cognition. His research interests include macrocognition and complex cognitive systems. Hoffman has PhD in experimental psychology from the University of Cincinnati. He is a fellow of the Association for Psychological Science and the Human Factors and Ergonomics Society and a senior member of IEEE. Contact him at rhoffman@ ihmc.us. Kenneth M. Ford is director of the Insti- tute for Human and Machine Cognition. His research interests include artificial intelligence and human-centered computing. Ford has a PhD in computer science from Tulane University. Contact him at kford@ ihmc.us Read your subscriptions through the myCS publications portal at http://mycs.computer.org 73