Academia.eduAcademia.edu

Alien Minds

16 ALIEN MINDS Susan Schneider I think it very likely – in fact, inevitable – that biological intelligence is only a transitory phenomenon… If we ever encounter extraterrestrial intelligence, I believe it is very likely to be postbiological in nature … Paul Davies1 How would intelligent aliens think? Would they have conscious experiences? Would it feel a certain way to be an alien? It is easy to dismiss these questions as too speculative, since we haven’t encountered aliens, at least as far as we know. And in conceiving of alien minds we do so from within – from inside the vantage point of the sensory experiences and thinking patterns characteristic of our species. At best, we anthropomorphize; at worst, we risk stupendous failures of the imagination. Still, ignoring these questions could be a grave mistake. Some proponents of SETI estimate that we will encounter alien intelligence within the next several decades. Even if you hold a more conservative estimate – say, that the chance of encountering alien intelligence in the next fifty years is five percent – the stakes for our species are high. Knowing that we are not alone in the universe would be a profound realization, and contact with an alien civilization could produce amazing technological innovations and cultural insights. It thus can be valuable to consider these questions, albeit with the goal of introducing possible routes to answering them, rather than producing definitive answers. So, let us ask: how might aliens think? And, would they be conscious? Believe it or not, we can say something concrete in response to both of these questions, drawing from work in philosophy and cognitive science. You might think the second question is odd. After all, if aliens have sophisticated enough mental lives to be intelligent, wouldn’t they be conscious? The far more intriguing question is: what would the quality of their consciousness be like? This would be putting the cart before the horse, however, since I do not believe that most advanced alien civilizations will be biological. The most sophisticated civilizations will be postbiological, forms of artificial intelligence (AI). (Bradbury, Cirkovic, and Dvorsky 2011; Cirkovic and Bradbury 2006; Davies 2010; Dick 2013; Shostak 2009).2 Further, alien civilizations will tend to be forms of superintelligence: intelligence that is able to exceed the best human level intelligence in every field – social skills, general wisdom, scientific creativity, and so on (Bostrom 2014; Kurzweil 2005; Schneider 2011). It is a substantive question whether superintelligent AI (SAI) could have conscious experiences; philosophers have vigorously debated just this question of in the case of AI in general. Perhaps all their information processing happens in the dark, so to speak, without any inner experience at all. This is why I find the second question so pressing, and in an important sense prior to any inquiry as to the contours of alien consciousness, and the epistemological problem of how we can know “what it is like” to be an alien. 1 In this chapter I first explain why it is likely that the alien civilizations we encounter will be forms of SAI. I then turn to the question of whether superintelligent aliens can be conscious – whether it feels a certain way to be an alien, despite their non-biological nature. Here, I draw from the literature in philosophy of AI, and urge that although we cannot be certain that superintelligent aliens can be conscious, it is likely that they would be. I then turn to the difficult question of how such creatures might think. I provisionally attempt to identify some goals and cognitive capacities likely to be possessed by superintelligent beings. I discuss Nick Bostrom’s recent book on superintelligence, which focuses on the genesis of SAI on Earth; as it happens, many of Bostrom’s observations are informative in the present context. Finally, I isolate a specific type of superintelligence that is of particular import in the context of alien superintelligence, biologically-inspired superintelligences (“BISAs”). Alien Superintelligence Our culture has long depicted aliens as humanoid creatures with small, pointy chins, massive eyes, and large heads, apparently to house brains that are larger than ours. Paradigmatically, they are “little green men.” While we are aware that our culture is anthropomorphizing, I imagine that my suggestion that aliens are supercomputers may strike you as far-fetched. So what is my rationale for the view that most intelligent alien civilizations will have members that are SAI? I offer three observations that together, motivate this conclusion. 1. The short window observation. Once a society creates the technology that could put them in touch with the cosmos, they are only a few hundred years away from changing their own paradigm from biology to AI (Davies 2010; Dick 2013; Shostak 2009). This “short window” makes it more likely that the aliens we encounter would be postbiological. The short window observation is supported by human cultural evolution, at least thus far. Our first radio signals date back only about a hundred and twenty years, and space exploration is only about fifty years old, but we are already immersed in digital technology, such as cellphones and laptop computers. Devices such as the Google Glass promise to bring the internet into more direct contact with our bodies, and it is probably a matter of less than fifty years before sophisticated internet connections are wired directly into our brains. Indeed, implants for Parkinson’s are already in use, and in the United States the Defense Advanced Research Projects Agency (DARPA) has started to develop neural implants that interface directly with the nervous system, regulating conditions such as post-traumatic stress disorder, arthritis, depression, and Crohn’s disease. DARPA’s program, called “ElectRx”, aims to replace certain medications with “closed-loop” neural implants, implants that continually assess the state of one’s health, and provide the necessary nerve stimulation to keep one’s biological systems functioning properly. Eventually, implants will be developed to enhance normal brain functioning, rather than for medical purposes. Where might all this all lead? A thought experiment from my “Transcending and Enhancing the Human Brain” is suggestive (Schneider 2011a). {disp} 2 Suppose it is 2025 and being a technophile, you purchase brain enhancements as they become readily available. First, you add a mobile internet connection to your retina, then, you enhance your working memory by adding neural circuitry. You are now officially a cyborg. Now skip ahead to 2040. Through nanotechnological therapies and enhancements you are able to extend your lifespan, and as the years progress, you continue to accumulate more farreaching enhancements. By 2060, after several small but cumulatively profound alterations, you are a “posthuman.” To quote philosopher Nick Bostrom, posthumans are possible future beings, “whose basic capacities so radically exceed those of present humans as to be no longer unambiguously human by our current standards” (Bostrom 2003). At this point, your intelligence is enhanced not just in terms of speed of mental processing; you are now able to make rich connections that you were not able to make before. Unenhanced humans, or “naturals,” seem to you to be intellectually disabled – you have little in common with them – but as a transhumanist, you are supportive of their right to not enhance (Bostrom 2003; Garreau 2005; Kurzweil 2005). It is now 2400 AD. For years, worldwide technological developments, including your own enhancements, have been facilitated by superintelligent AI. Indeed, as Bostrom explains, “creating superintelligence may be the last invention that humans will ever need to make, since superintelligences could themselves take care of further scientific and technological developments” (Bostrom 2003). Over time, the slow addition of better and better neural circuitry has left no real intellectual difference in kind between you and superintelligent AI. The only real difference between you and an AI creature of standard design is one of origin – you were once a natural. But you are now almost entirely engineered by technology – you are perhaps more aptly characterized as a member of a rather heterogeneous class of AI life forms (Kurzweil 2005). Of course, this is just a thought experiment. But I’ve just observed that we are already beginning to develop neural implants. It is hard to imagine people in mainstream society resisting opportunities for superior health, intelligence and efficiency. And just as people have already turned to cryonics, even in its embryonic state, I suspect that they will increasingly try to upload to avoid death, especially as the technology is perfected.3 Indeed, the Future of Humanity Institute at Oxford University has released a report on the technological requirements for uploading a mind to a machine. And a Defense Department agency has funded a program, Synapse, that is developing a computer that resembles a brain in form and function (Schneider 2014). In essence, the short window observation is supported by our own cultural evolution, at least thus far. You may object that this argument employs “N = 1 reasoning,” generalizing from the human case to the case of alien civilizations (see Chapter 7 in this volume). Still, it is unwise to discount arguments based on the human case. Human civilization is the only one we know of and we had better learn from it. It is no great leap to claim that other 3 civilizations will develop technologies to advance their intelligence and survival. And, as I will explain in a moment, silicon is a better medium for thinking than carbon. A second objection to my short window observation rightly points out that nothing I have said thus far suggests that humans will be superintelligent. I have merely said that future humans will be posthuman. While I offer support for the view that our own cultural evolution suggests that humans will be postbiological, this does not show that advanced alien civilizations will reach superintelligence. So even if one is comfortable reasoning from the human case, the human case does not support the position that the members of advanced alien civilizations will be superintelligent. This is correct. This is the task of the second observation. 2. The greater age of alien civilizations. Proponents of SETI have often concluded that alien civilizations could be much older than our own: “all lines of evidence converge on the conclusion that the maximum age of extraterrestrial intelligence would be billions of years, specifically [it] ranges from 1.7 billion to 8 billion years” (Dick 2013, 468). If civilizations are millions or billions of years older than us, many would be vastly more intelligent than we are. By our standards, many would be superintelligent. We are galactic babies. But would they be forms of AI, as well as forms of superintelligence? I believe so. Even if they were biological, merely having biological brain enhancements, their superintelligence would be reached by artificial means, and we could regard them as being “artificial intelligence.” But I suspect something stronger than this: I expect that they will not be carbon-based. Uploading allows a creature near-immortality, enables reboots, and allows it to survive under a variety of conditions that carbon-based life forms cannot. In addition, silicon appears to be a better medium for information processing than the brain itself. Neurons reach a peak speed of about 200 Hz, which is seven orders of magnitude slower than current microprocessors (Bostrom 2014, 59). While the brain can compensate for some of this with massive parallelism, features such as “hubs,” and so on, crucial mental capacities, such as attention, rely upon serial processing, which is incredibly slow, and has a maximum capacity of about seven manageable chunks (Miller 1956). Further, the number of neurons in a human brain is limited by cranial volume and metabolism, but computers can occupy entire buildings or cities, and can even be remotely connected across the globe (Bostrom 2014). Of course, the human brain is far more intelligent than any modern computer. But intelligent machines can in principle be constructed by reverse engineering the brain, and improving upon its algorithms. In sum: I have observed that there seems to be a short window from the development of the technology to access the cosmos and the development of postbiological minds and AI. I then observed that we are galactic babies: extraterrestrial civilizations are likely to be vastly older than us, and thus they would have already reached not just postbiological life, but superintelligence. Finally, I noted that they would likely be SAI, because silicon is a superior medium for superintelligence. From this I conclude that many advanced alien civilizations will be populated by SAI. Even if I am wrong – even if the majority of alien civilizations turn out to be biological – it may be that the most intelligent alien civilizations will be ones in which the inhabitants are SAI. Further, creatures that are silicon-based, rather than biologically 4 based, are more likely to endure space travel, having durable systems that are practically immortal, so they may be the kind of the creatures we first encounter. All this being said, would superintelligent aliens be conscious, having inner experiences? Here, I draw from a rich philosophical literature on the nature of conscious experience. Would Superintelligent Aliens Be Conscious? Consider your own conscious experience. Suppose that you are sitting in a cafe preparing to give a lecture. All in one moment, you taste the espresso you sip, consider an idea, and hear the scream of the espresso machine. This is your current stream of consciousness. Conscious streams seem to be very much bound up with who you are. It is not that this particular moment is essential – although you may feel that certain ones are important. It is rather that throughout your waking life, you seem to be the subject of a unified stream of experience that presents you as the subject, viewing the show. Let us focus on three features of the stream: first, it may seem to you, put metaphorically, that there is a sort of “screen” or “stage” in which experiences present themselves to your “mind’s eye.” That is, there appears to be a central place where experiences are “screened” before you. Daniel Dennett calls this place “the Cartesian Theater” (Dennett 1991). Second, in this central place there seems to be a singular point in time which, given a particular sensory input, consciousness happens. For instance, there seems to be one moment in which the scream of the espresso machine begins, pulling you out of your concentration. Finally, there appears to be a self – someone who is inside the theater, watching the show. Philosophers have considered each of these features in detail. Each is highly problematic. For instance, an explanation of consciousness cannot literally be that there is a mind’s eye in the brain, watching a show. And there is no evidence that there is a singular place or time in the brain where consciousness congeals. These are intriguing issues, but pursuing them in the context of alien consciousness is putting the cart before the horse. For there is a more fundamental problem: would superintelligent aliens, being forms of AI, even be conscious? Why should we believe that creatures so vastly different from us, being silicon-based, would have inner experience at all? This problem relates to what philosophers call the hard problem of consciousness, a problem that was posed in the context of human consciousness by the philosopher David Chalmers (Chalmers 2008). Chalmers’ hard problem is the following. As cognitive science underscores, when we deliberate, hear music, see the rich hues of a sunset, and so on, there is information processing going on in the brain. But above and beyond the manipulation of data, there is a subjective side – there is a “felt quality” to our experience. The hard problem is: why does all this information processing in the human brain, under certain conditions, have a felt quality to it? As Chalmers emphasizes, the hard problem is a philosophers’ problem, because it doesn’t seem to have a scientific answer. For instance, we could develop a complete theory of vision, understanding all of the details of visual processing in the brain, but still not understand why there are subjective experiences attached to these informational states. Chalmers contrasts the hard problem with what he calls “easy problems”, 5 problems involving consciousness that have eventual scientific answers, such as the mechanisms behind attention and how we categorize and react to stimuli. Of course these scientific problem are difficult problems; Chalmers merely calls them “easy problems” to contrast them with the “hard problem” of consciousness, which he thinks will not have a purely scientific solution. We now face yet another perplexing issue involving consciousness – a kind of “hard problem” involving alien superintelligence, if you will: The hard problem of alien superintelligence: Would the processing of a siliconbased superintelligent system feel a certain way, from the inside? An alien SAI could solve problems that even the brightest humans are unable to solve, but still, being made of a nonbiological substrate, would their information processing feel a certain way from the inside? It is worth underscoring that the hard problem of alien consciousness is not just Chalmers’ hard problem of consciousness applied to the case of aliens. For the hard problem of consciousness assumes that we are conscious – after all, each of us can tell from introspecting that we are conscious at this moment. It asks why we are conscious. Why does all your information processing feel a certain way from the inside? In contrast, the hard problem of alien consciousness asks whether alien superintelligence, being silicon-based, is even capable of being conscious. It does not presuppose that aliens are conscious. These are different problems, but they are both hard problems that science alone cannot answer. The problem in the case of superintelligent aliens is that the capacity to be conscious may be unique to biological, carbon-based, organisms. According to biological naturalism even the most sophisticated forms of AI will be devoid of inner experience (Blackmore 2004; Searle 1980, 2008). Indeed, even humans wishing to upload their minds will fail to transfer their consciousness. Although they may copy their memories onto a computational format, their consciousness will not transfer, since biological naturalists hold that consciousness requires a biological substrate.4 What arguments support biological naturalism? The most common consideration in favor of biological naturalism is John Searle’s Chinese Room thought experiment, which is said to suggest that a computer program cannot understand or be conscious (Searle 1980). Searle supposes that he’s locked in a room, where he’s handed a set of English rules that allow him to link one set of Chinese symbols with other Chinese symbols. So although he doesn’t know Chinese, the rules allow him to respond, in written Chinese, to questions written in Chinese. So he is essentially processing symbols. Searle concludes that although those outside of the room may think he understands Chinese, he obviously doesn’t; similarly, a computer may appear to be having a Chinese conversation, yet it does not truly understand Chinese. Nor is it conscious. Although it is correct that Searle doesn’t understand Chinese, the issue is not really whether Searle understands; Searle is just one part of the larger system. The relevant question is whether the system as a whole understands Chinese. This basic response to Searle’s Chinese Room thought experiment is known as the Systems Reply.5 6 It strikes me as implausible that a simple system like the Chinese Room understands however, for the Chinese Room is not complex enough to understand or be conscious. But the Systems Reply is on to something: the real issue is whether the system as a whole understands, not whether one component does. This leaves open the possibility that a more complex silicon-based system could understand; of course, the computations of a superintelligent AI will be far more complex than the human brain. Here, some might suspect that we could just reformulate the Chinese Room thought experiment in the context of a superintelligent AI. But what is fueling this suspicion? It cannot be that some central component in the SAI, analogous to Searle in the Chinese Room, doesn’t understand, for we’ve just observed that it is the system as a whole that understands. Is the suspicion instead fueled by the position that understanding and consciousness do not decompose into more basic operations? If so, then the thought experiment purports to prove too much. Consider the case of the human brain. According to cognitive science, cognitive and perceptual capacities decompose into more basic operations, which are themselves decomposable into more basic constituents, which themselves can be explained causally (Block 1995). If the Chinese Room illustrates that mentality cannot be explained like this, then the brain cannot be explained in this manner either. But this explanatory approach, known as “the method of functional decomposition,” is a leading approach to explaining mental capacities in cognitive science. Consciousness and understanding are complex mental properties that are determined by the arrangements of neurons in the brain. Further, biological naturalism denies one of the main insights of cognitive science – the insight that the brain is computational – without substantial empirical rationale. Cognitive science suggests that our best empirical theory of the brain holds that the mind is an information-processing system and that all mental functions are computations. If cognitive science is correct that thinking is computational, then humans and SAI share a common feature: their thinking is essentially computational. Just as a phone call and a smoke signal can convey the same information, thought can have both silicon- and carbon-based substrates. The upshot is that if cognitive science is correct that thinking is computational, we can also expect that sophisticated thinking machines can be conscious, although the contours of their conscious experiences will surely differ. Indeed, I’ve noted that silicon is arguably a better medium for information processing than the brain. So why isn’t silicon a better medium for consciousness, rather than a worse one, as the biological naturalists propose? It would be surprising if SAI, which would have far superior information-processing abilities than we do, turned out to be deficient with respect to consciousness. For our best scientific theories of consciousness hold that consciousness is closely related to information processing (Tonini 2008; Baars 2008). Some would point out that to show that AI cannot be conscious, the biological naturalist would need to locate a special consciousness property, (call it “P”) that inheres in neurons or their configurations, and which cannot be instantiated by silicon. Thus far, P has not been discovered. It isn’t clear, however, that locating P would prove biological naturalism to be correct. For the computationalist can just say that machines are capable of instantiating a different type of consciousness property, F, which is specific to siliconbased systems. 7 Massimo Pigliucci has offered a different kind of consideration in favor of biological naturalism, however. He sees philosophers who argue for computationalism as embracing an implausible perspective on the nature of consciousness: functionalism. According to functionalists the nature of a mental state depends on the way it functions, or the role it plays in the system of which it is a part. Pigliucci is correct that traditional functionalists, such as Jerry Fodor, generally mistakenly ignore the biological workings of the brain. Pigliucci objects: “functionality isn’t just a result of the proper arrangement of the parts of a system, but also of the types of materials (and their properties) that make up those parts” (Pigliucci 2014). Fodor’s well-known antipathy towards neuroscience should not mislead us into thinking that functionalism must ignore neuroscience, however. Clearly, any wellconceived functionalist position must take into consideration neuroscientific work on the brain because the functionalist is interested in the causal or dispositional properties of the parts, not just the parts themselves. Indeed, as I’ve argued in my book The Language of Thought, viewing the brain as irrelevant to the computational approach to the mind is a huge mistake. The brain is the best computational system we know of (Schneider 2011b). Does this make my position a form of biological naturalism? Not in the least. I am suggesting that viewing neuroscience, (and by extension, biology) as being opposed to computationalism is mistaken. Indeed, neuroscience is computational; a large subfield of neuroscience is called “computational neuroscience”, and it seeks to understand the sense in which the brain is computational and to provide computational accounts of mental capacities identified by related subfields, such as cognitive neuroscience. What makes my view different from biological naturalism is that I hold that thinking is computational, and further, that at least one other substrate besides carbon (i.e., silicon) can give rise to consciousness and understanding, at least in principle. But biological naturalism is well worth considering. I am reasoning that a substrate that supports superintelligence, being capable of even more sophisticated informational processing than we are, would likely also be one that is conscious. But notice that I’ve used the expression “likely.” For we can never be certain that AI is conscious, even if we could study it up close. The problem is akin to the philosophical puzzle known as the problem of other minds (Schneider 2014). The problem of other minds is that although you can know that you are conscious, you cannot be certain that other people are conscious as well. After all, you might be witnessing behavior with no accompanying conscious component. In the face of the problem of other minds, all you can do is note that other people have brains that are structurally similar to your own and conclude that since you yourself are conscious, others are likely to be as well. When confronted with AI your predicament would be similar, at least if you accept that thinking is computational. While we couldn’t be absolutely certain that an AI program genuinely felt anything, we can’t be certain that other humans do either. But it would seem probable in both cases. So, to the question of whether alien superintelligence can be conscious, I answer, very cautiously, “probably.” {a}How Might Superintelligent Aliens Think? 8 {text}Thus far, I’ve said little about the structure of superintelligent alien minds. And little is all we can say: superintelligence is by definition a kind of intelligence that outthinks humans in every domain. In an important sense, we cannot predict or fully understand how it will think. Still, we may be able to identify a few important characteristics, albeit in broad strokes. Nick Bostrom’s recent book on superintelligence focuses on the development of superintelligence on Earth, but we can draw from his thoughtful discussion (Bostrom 2014). Bostrom distinguishes three kinds of superintelligence: {disp} Speed superintelligence – even a human emulation could in principle run so fast that it could write a PhD thesis in an hour. Collective superintelligence – the individual units need not be superintelligent, but the collective performance of the individuals outstrips human intelligence. Quality superintelligence – at least as fast as human thought, and vastly smarter than humans in virtually every domain. {text}Any of these kinds could exist alongside one or more of the others. An important question is whether we can identify common goals that these types of superintelligences may share. Bostrom’s suggests The Orthogonality Thesis: “Intelligence and final goals are orthogonal – more or less any level of intelligence could in principle be combined with more or less any final goal.” (Bostrom 2014, 107) Bostrom is careful to underscore that a great many unthinkable kinds of SAI could be developed. At one point, he raises a sobering example of a superintelligence with the final goal of manufacturing paper clips (pp. 107–108, 123–125). While this may initially strike you as harmless endeavor, although hardly a life worth living, Bostrom points out that a superintelligence could utilize every form of matter on Earth in support of this goal, wiping out biological life in the process. Indeed, Bostrom warns that superintelligence emerging on Earth could be of an unpredictable nature, being “extremely alien” to us (p. 29). He lays out several scenarios for the development of SAI. For instance, SAI could be arrived at in unexpected ways by clever programmers, and not be derived from the human brain whatsoever. He also takes seriously the possibility that Earthly superintelligence could be biologically inspired, that is, developed from reverse engineering the algorithms that cognitive science says describe the human brain, or from scanning the contents of human brains and transferring them to a computer (i.e., “uploading”).6 Although the final goals of superintelligence are difficult to predict, Bostrom singles out several instrumental goals as being likely, given that they support any final goal whatsoever: 9 The Instrumental Convergence Thesis: “Several instrumental values can be identified which are convergent in the sense that their attainment would increase the chances of the agent’s goal being realized for a wide range of final goals and a wide range of situations, implying that these instrumental values are likely to be pursued by a broad spectrum of situated intelligent agents.” (Bostrom 2014, 109) {text}The goals that he identifies are resource acquisition, technological perfection, cognitive enhancement, self-preservation and goal content integrity (i.e., that a superintelligent being’s future self will pursue and attain those same goals). He underscores that self-preservation can involve group or individual preservation, and that it may play second fiddle to the preservation of the species the AI was designed to serve (Bostrom 2014). Let us call an alien superintelligence that is based on reverse engineering an alien brain, including uploading it, a biologically inspired superintelligent alien (or “BISA”). Although BISAs are inspired by the brains of the original species that the superintelligence is derived from, a BISA’s algorithms may depart from those of their biological model at any point. BISAs are of particular interest in the context of alien superintelligence. For if Bostrom is correct that there are many ways superintelligence can be built, but a number of alien civilizations develop superintelligence from uploading or other forms of reverse engineering, it may be that BISAs are the most common form of alien superintelligence out there. This is because there are many kinds of superintelligence that can arise from raw programming techniques employed by alien civilizations. (Consider, for instance, the diverse range of AI programs under development on Earth, many of which are not modeled after the human brain.) This may leave us with a situation in which the class of SAIs is highly heterogeneous, with members generally bearing little resemblance to each other. It may turn out that of all SAIs, BISAs bear the most resemblance to each other. In other words, BISAs may be the most cohesive subgroup because the other members are so different from each other. Here, you may suspect that because BISAs could be scattered across the galaxy and generated by multitudes of species, there is little interesting that we can say about the class of BISAs. But notice that BISAs have two features that may give rise to common cognitive capacities and goals: {nl} 1. BISAs are descended from creatures that had motivations like: find food, avoid injury and predators, reproduce, cooperate, compete, and so on. 2. The life forms that BISAs are modeled from have evolved to deal with biological constraints like slow processing speed and the spatial limitations of embodiment. Could (1) or (2) yield traits common to members of many superintelligent alien civilizations? I suspect so. 10 Consider (1). Intelligent biological life tends to be primarily concerned with its own survival and reproduction, so it is more likely that BISAs would have final goals involving its own survival and reproduction, or at least the survival and reproduction of the members of its society. If BISAs are interested in reproduction, we might expect that, given the massive amounts of computational resources at their disposal, BISAs would create simulated universes stocked with artificial life and even intelligence or superintelligence. If these creatures were intended to be “children” they may retain the goals listed in (1) as well. You may object that it is useless to theorize about BISAs, as they can change their basic architecture in numerous, unforeseen ways, and any biologically inspired motivations can be constrained by programming. There may be limits to this, however. If a superintelligence is biologically based, it may have its own survival as a primary goal. In this case, it may not want to change its architecture fundamentally, but stick to smaller improvements. It may think: when I fundamentally alter my architecture, I am no longer me (Schneider 2011). Uploads, for instance, may be especially inclined to not alter the traits that were most important to them during their biological existence. Consider (2). The designers of the superintelligence, or a self-improving superintelligence itself, may move away from the original biological model in all sorts of unforeseen ways, although I have noted that a BISA may not wish to alter its architecture fundamentally. But we could look for cognitive capacities that are useful to keep; cognitive capacities that sophisticated forms of biological intelligence are likely to have, and which enable the superintelligence to carry out its final and instrumental goals. We could also look for traits are not likely to be engineered out, as they do not detract the BISA from its goals. If (2) is correct, we might expect the following, for instance. (i) Learning about the computational structure of the brain of the species that created the BISA can provide insight into the BISAs thinking patterns. One influential means of understanding the computational structure of the brain in cognitive science is via “connectomics,” a field that seeks to provide a connectivity map or wiring diagram of the brain (Seung 2012). While it is likely that a given BISA will not have the same kind of connectome as the members of the original species, some of the functional and structural connections may be retained, and interesting departures from the originals may be found. (ii) BISAs may have viewpoint invariant representations. At a high level of processing your brain has internal representations of the people and objects that you interact with that are viewpoint invariant. Consider walking up to your front door. You’ve walked this path hundreds, maybe thousands of times, but technically, you see things from slightly different angles each time as you are never positioned in exactly the same way twice. You have mental representations that are at a relatively high level of processing that are viewpoint invariant. It seems difficult for biologically based intelligence to evolve without such representations, as they enable categorization and prediction (Hawkins and Blakeslee 2004). Such representations arise because a system that is mobile needs 11 a means of identifying items in its ever-changing environment, so we would expect biologically based systems to have them. BISA would have little reason to give up object invariant representations insofar as it remains mobile or has mobile devices sending it information remotely. (iii) BISAs will have language-like mental representations that are recursive and combinatorial. Notice that human thought has the crucial and pervasive feature of being combinatorial. Consider the thought wine is better in Italy than in China. You probably have never had this thought before, but you were able to understand it. The key is that the thoughts are combinatorial because they are built out of familiar constituents, and combined according to rules. The rules apply to constructions out of primitive constituents, that are themselves constructed grammatically, as well as to the primitive constituents themselves. Grammatical mental operations are incredibly useful: It is the combinatorial nature of thought that allows one to understand and produce these sentences on the basis of one’s antecedent knowledge of the grammar and atomic constituents (e.g., wine, China). Relatedly, thought is productive: in principle, one can entertain and produce an infinite number of distinct representations because the mind has a combinatorial syntax (Schneider 2011). Brains need combinatorial representations because there are infinitely many possible linguistic representations, and the brain only has a finite storage space. Even a superintelligent system would benefit from combinatorial representations. Although a superintelligent system could have computational resources that are so vast that it is mostly capable of pairing up utterances or inscriptions with a stored sentence, it would be unlikely that it would trade away such a marvelous innovation of biological brains. If it did, it would be less efficient, since there is the potential of a sentence not being in its storage, which must be finite. (iv) BISAs may have one or more global workspaces. When you search for a fact or concentrate on something, your brain grants that sensory or cognitive content access to a “global workspace” where the information is broadcast to attentional and working memory systems for more concentrated processing, as well as to the massively parallel channels in the brain (Baars 2008). The global workspace operates as a singular place where important information from the senses is considered in tandem, so that the creature can make all-things-considered judgments and act intelligently, in light of all the facts at its disposal. In general, it would be inefficient to have a sense or cognitive capacity that was not integrated with the others, because the information from this sense or cognitive capacity would be unable to figure in predictions and plans based on an assessment of all the available information. (v) A BISA’s mental processing can be understood via functional decomposition. As complex as alien superintelligence may be, humans may be able to use the method of functional decomposition as an approach to understanding it. A key feature of 12 computational approaches to the brain is that cognitive and perceptual capacities are understood by decomposing the particular capacity into their causally organized parts, which themselves can be understood in terms of the causal organization of their parts. This is the aforementioned “method of functional decomposition” and it is a key explanatory method in cognitive science. It is difficult to envision a complex thinking machine not having a program consisting of causally interrelated elements each of which consists in causally organized elements. This has important implications should a SETI program discover a communicating BISA. {text}All this being said, superintelligent beings are by definition beings that are superior to humans in every domain. While a creature can have superior processing that still basically makes sense to us, it may be that a given superintelligence is so advanced that we cannot understand any of its computations whatsoever. It may be that any truly advanced civilization will have technologies that will be indistinguishable from magic, as Arthur C. Clark suggested (1962). I obviously speak to the scenario in which the SAI’s processing makes some sense to us, one in which developments from cognitive science yield a glimmer of understanding into the complex mental lives of certain BISAs. Conclusion I have argued that the members of the most advanced alien civilizations will be forms of superintelligent artificial intelligence (SAI). I have further suggested, very provisionally, that we might expect that if a given alien superintelligence is a biologically inspired superintelligent alien (BISA), it would have combinatorial representations and that we could seek insight into its processing by decomposing its computational functions into causally interacting parts. We could also learn about it by looking at the brain wiring diagrams (connectomes) of the members of the original species. Further, BISAs may have one or more global workspaces. And, I have argued that there is no reason in principle to deny that SAIs could have conscious experience. Notes Many thanks to Joe Corabi, Steven Dick, Clay Ferris Naff and Eric Schwitzgebel for helpful written comments on an earlier draft and to James Hughes for helpful conversation. 1. 2. 3. Davies 2010, 160. “Postbiological”, in the astrobiology literature contrasts with “posthuman” in the singularity literature. In the astrobiology literature “postbiological” creatures are forms of AI. In the singularity literature “posthumans” can be forms of AI, but they need not be. They are merely creatures who are descended from humans but which have alterations that make them no longer unambiguously human. They need not be full-fledged AI. Although I have elsewhere argued that uploading would merely create a copy of one’s brain configuration and would not be a true means of survival, I doubt 13 4. 5. 6. dying individuals will act on a philosopher’s qualms when they have little to lose by trying (Schneider 2014). Biological naturalism was originally developed by John Searle, who developed the view in the context of a larger account of the relation between the mind and body. I will not discuss these details, and they are not essential to the position I’ve just sketched. Indeed, it isn’t clear that Searle is still a biological naturalist, although he persists in calling his view “biological naturalism.” In his chapter to my recent Blackwell Companion to Consciousness he wrote: “The fact that brain processes cause consciousness does not imply that only brains can be conscious. The brain is a biological machine, and we might build an artificial machine that was conscious; just as the heart is a machine, and we have built artificial hearts. Because we do not know exactly how the brain does it we are not yet in a position to know how to do it artificially” (Searle 2008). For a thorough treatment of the responses to Searle’s argument, including the system’s reply, the reader may turn to the comments appearing with Searle’s original piece (Searle 1980), as well as Cole 2014. Throughout his book, Bostrom emphasizes that we must bear in mind that superintelligence, being unpredictable and difficult to control, may pose a grave existential risk to our species. This should give us pause in the context of alien contact as well (Bostrom 2014). {a}References Baars, B. 2008. “The Global Workspace Theory of Consciousness.” In Max Velmans and Susan Schneider (eds.), The Blackwell Companion to Consciousness. Oxford: Wiley-Blackwell. Blackmore, Susan. 2004. Consciousness: An Introduction. New York: Oxford University Press. Block, N. 1995. “The mind as the software of the brain.” In D. Osherson, L. Gleitman, S. Kosslyn, E. Smith, and S. Sternberg (eds.), An Invitation to Cognitive Science (pp. 377–421). New York: MIT Press. Bostrom, N. 2003. “The Transhumanist Frequently Asked Questions”: v 2.1. World Transhumanist Association. Retrieved from http://transhumanism.org/index.php/WTA/faq/. Bostrom, N. 2005. “History of Transhumanist Thought.” Journal of Evolution and Technology, 14(1). Bostrom, N. 2008. “Dignity and enhancement.” In The President’s Council on Bioethics, Human Dignity and Bioethics: Essays Commissioned by the President’s Council on Bioethics, Washington, DC: U.S. Government Printing Office. Bostrom, N. 2014. Superintelligence: Paths, Dangers, Strategies. Oxford: Oxford University Press. Bradbury, R., Cirkovic, M., and Dvorsky, G. 2011. “Dysonian Approach to SETI: A Fruitful Middle Ground?,” Journal of the British Interplanetary Society, vol. 64, pp. 156–165. 14 Chalmers, D. 2008. “The Hard Problem of Consciousness.” In Max Velmans and Susan Schneider (eds.), The Blackwell Companion to Consciousness. Oxford: WileyBlackwell. Cirkovic, M. and Bradbury, R. 2006. “Galactic Gradients, Postbiological Evolution and the Apparent Failure of SETI,” New Astronomy 11, 628–639. Clarke, A. (1962). Profiles of the Future: An Inquiry into the Limits of the Possible. New York: Harper & Row. Cole, David. 2014. “The Chinese Room Argument.” In Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy (summer 2014 edn). <http://plato.stanford.edu/archives/sum2014/entries/chinese-room/>. Davies, Paul. 2010. The Eerie Silence: Renewing Our Search for Alien Intelligence, Boston: Houghton Mifflin Harcourt. Dennett, D. 1991. Consciousness Explained, New York: Penguin Press. Dick, S. 2013. “Bringing Culture to Cosmos: The Postbiological Universe.” In Dick, S. and Lupisella, M. (eds.), Cosmos and Culture: Cultural Evolution in a Cosmic Context, Washington, DC, NASA. http://history.nasa.gov/SP-4802.pdf Garreau, J. 2005. Radical Evolution: The Promise and Peril of Enhancing Our Minds, Our Bodies – and What It Means to be Human, New York: Doubleday. Guerini, Federico. 2014. “DARPA’s ElectRx Project: Self-Healing Bodies through Targeted Stimulation of the Nerves,” http://www.forbes.com/sites/federicoguerrini/2014/08/29/darpas-electrx-projectself-healing-bodies-through-targeted-stimulation-of-the-nerves/forbes magazine, 8/29/2014. Extracted Sept. 30, 2014. Hawkins, J. and Blakeslee, S. 2004. On Intelligence: How a New Understanding of the Brain will Lead to the Creation of Truly Intelligent Machine. New York: Times Books. Kurzweil, R. 2005. The Singularity is Near: When Humans Transcend Biology. New York: Viking. Miller, R. 1956. “The Magical Number Seven, Plus or Minus Two: Some Limits on Our Capacity for Processing Information,” The Psychological Review, vol. 63, pp. 81– 97. Pigliucci, M. 2014. “Mind Uploading: A Philosophical Counter-Analysis.” In R. Blackford and D. Broderik (eds.), Intelligence Unbound, Oxford: WileyBlackwell. Sandberg, A. and Bostrom, N. 2008. “Whole Brain Emulation: A Roadmap.” Technical Report #2008‐3. Future of Humanity Institute, Oxford University. Schneider, Susan, ed. 2009. Science Fiction and Philosophy. Chichester: WileyBlackwell. Schneider, Susan. 2011a. “Mindscan: Transcending and Enhancing the Brain.” In James Giordano (ed.), Neuroscience and Neuroethics: Issues At the Intersection of Mind, Meanings and Morality, Cambridge: Cambridge University Press. Schneider, Susan. 2011b. The Language of Thought: A New Philosophical Direction. Boston: MIT Press. 15 Schneider, Susan. 2014. “The Philosophy of ‘Her,’” The New York Times, March 2. Searle, J. 1980. “Minds, Brains and Programs,” The Behavioral and Brain Sciences, 3: 417–457. Searle, J. 2008. “Biological Naturalism.” In Max Velmans and Susan Schneider (eds.), The Blackwell Companion to Consciousness. Oxford: Wiley-Blackwell. Seung, S. 2012. Connectome: How the Brain’s Wiring Makes Us Who We Are. Boston: Houghton Mifflin Harcourt. Shostak, S. 2009. Confessions of an Alien Hunter, New York: National Geographic. Tonini, G. 2008. “The Information Integration Theory of Consciousness.” In Max Velmans and Susan Schneider (eds.), The Blackwell Companion to Consciousness. Oxford: Wiley-Blackwell. 16