Ethical Reflections on Artificial Intelligence
6(2)/2018
Received: June 1, 2018. Accepted: August 24, 2018
ISSN 2300-7648 (print) / ISSN 2353-5636 (online)
DOI: http://dx.doi.org/10.12775/SetF.2018.015
Ethical Reflections
on Artificial Intelligence*
BRIAN PATRICK GREEN
Director of Technology Ethics, Markkula Center for Applied Ethics, Faculty (adj.),
School of Engineering, Santa Clara University
[email protected]
ORCID: 0000-0002-7125-3086
Abstract: Artificial Intelligence (AI) technology presents a multitude of ethical concerns,
many of which are being actively considered by organizations ranging from small
groups in civil society to large corporations and governments. However, it also presents
ethical concerns which are not being actively considered. This paper presents a broad
overview of twelve topics in ethics in AI, including function, transparency, evil use, good
use, bias, unemployment, socio-economic inequality, moral automation and human
de-skilling, robot consciousness and rights, dependency, social-psychological effects,
and spiritual effects. Each of these topics will be given a brief discussion, though each
deserves much deeper consideration.
Keywords: ethics; theology; religion; science; technology.
∗
No funding has been provided specifically for the production of this paper. Portions of this
paper are based on sections from the following sources Green 2017b, 2017c, and 2018a.
Source 2018a received minor financial support from the University of the Pacific Pope
John XXIII Lectureship. Source 2017c received salaried time and encouragement from the
Markkula Center for Applied Ethics and School of Engineering at Santa Clara University.
6 ( 2 ) / 2 0 1 8 , 9 –3 1
9
B R I A N PAT R I C K G R E E N
Introduction
The power of artificial intelligence has grown rapidly and is now one of the
most pressing issues for ethical and theological reflection. Even without
the immense possible religio-theological implications, noted by those who
have attempted to build AI into a “god,” for example (Cannon, 2015; Harris
2018), the dramatic social impacts of AI warrant Christian response. AI
has the potential to revolutionize (or already has revolutionized) almost
everything that humans do, including eating (e.g., agricultural planning,
distribution, pricing), relationships (e.g., Facebook, dating apps), money
(e.g., financial technology), war (e.g., drones, cyberdefense), healthcare
(e.g., predictive medicine, radiology), and education (e.g., personalized
curricula). As a “general purpose technology,” AI has been compared to
the development of electricity, or fire. Within the next decade there will be
few institutions untouched by AI. In order for Christian ethicists and moral
theologians to maintain situational awareness as many rapid changes occur
in society, not to mention maintain credibility in the contemporary world,
this is a topic that needs sustained ethical and theological inquiry.
Building on a rapidly growing body of work, I will begin with some
clarifications about AI, then will consider twelve ways in which AI has
ethical relevance (Bossman 2016; Future of Life 2017; Green 2017a, 2017b,
2018a; Partnership on AI 2018a). It is my hope that this comprehensive set
of issues will assist future systematic research into the theological ethics
and theology of AI by moral theologians and Christian ethicists. It is meant
to start a conversation. This paper is framed primarily philosophically, yet
in a manner consistent with Christian theology.
1. Clarifications
What is artificial intelligence? Intelligence itself is hard to define, and AI
is perhaps even more difficult. I will not here propose detailed definitions,
but instead only say that, in general, artificial intelligence seeks to re-create
particular aspects of human intelligence in computerized form. AI is a broad
category, including such diverse abilities as vision, speech recognition and
10
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
production, data analysis, advertising, navigation, machine learning, etc., and
just about anything that computers can do, if you stretch the definition enough.
Artificial intelligence is not the same as artificial consciousness (artificial
consciousness has sometimes been called “Artificial General Intelligence
(AGI),” “Strong AI,” or “Full AI”). Some thinkers vehemently believe that
artificial consciousness is possible (e.g., Chrisley 2008; Kurzweil 2012;
Koene 2013; most contributors to Chella & Manzotti 2013), while others
just as vehemently believe that it is impossible (e.g., Searle 1980; Schlagel
1999; and from a Catholic perspective: Labrecque 2017). I am agnostic on
the subject, but I am sure of this: AI developers will certainly try to make an
AI that simulates interaction with a human as closely as possible (see, for
example, Google’s recent release of Duplex (Leviathan & Matias 2018)). In
other words, the artificial construct, if very well done, will seem conscious.
But will it be conscious, or will it only be a simulation? To me, there is
no reason to believe that a close mimicry of consciousness is the same as
consciousness any more than a close mimicry of anything (fill in the blank:
forged money, forged artwork, actors imitating famous people, simulated
gemstones, etc.) actually becomes the real thing. But the possibility of an
exception remains, after all, some artificial gemstones, such as rubies and
sapphire, really are molecularly identical to natural rubies and sapphires
(both are the mineral corundum: aluminum oxide). Will consciousness be
exactly duplicable, like corundum? I doubt it, but I cannot be certain.
As another point, AI systems may or may not have humans “in the loop”
for training and/or decision-making. It is one thing for an AI to analyze
a situation and then make a recommendation to human decision-makers.
It is a different thing when that AI is directly attached to controls that allow
it to act upon its analyses without human approval. As examples, the first
case would be like Amazon.com recommending a book, or a military drone
(in combination with AI-processed data from espionage and other sources)
recommending a target to kill. Amazon.com does not automatically and
autonomously send you books, and the military drone does not fire without
permission. The second case, where decision-making is also automated, is
exemplified by self-driving vehicles. The entire purpose of a self-driving
6(2)/2018
11
B R I A N PAT R I C K G R E E N
car is to drive itself, taking the human out of the loop. This means the
systems must be extraordinarily good at decision-making before it can be
regarded as safe. Recent fatal Tesla and Uber accidents have begun to force
this question with deadly urgency.
2. Ethical Reflections
AI will bring with it developments that will be ethically positive, negative,
neutral, mixed, and/or ambiguous. Some AI technologies will be dual use,
for example, any AI program that can be used to identify wildlife could also
be used for targeting that wildlife. This is already being done in Australia
with drone submarines that automatically target and kill, by lethal injection,
crown-of-thorns starfish (Acanthaster planci) which are damaging the Great
Barrier Reef (Platt 2016). But of course, with adjustments to the software,
the drones could be re-targeted for other creatures, even humans. Here
I will reflect on twelve areas of AI relevance for ethics.
2.1. Function and Safety – “Does it work?”
The first concern with any technology is merely whether it works, and,
whether working or not, is it safe. If AI is put in charge of a vital system – like
driving a car – and it crashes the car, then that AI might be judged unsafe.
If the AI is in charge of designing a tall building, and the building falls down,
the AI might be judged unsafe.
Of note is that safety is a social construction and what seems safe to some
people will not seem safe to others. Some people like to ride motorcycles,
even though motorcycles are a riskier form of transportation than fourwheeled vehicles. Some people judge motorcycles to be safe, other people
judge them unsafe. When it comes to socially relevant technologies like AI,
no one person will get to decide if they are “safe.” Instead, safety will be
decided at the interplay of business managers, engineers, consumers, voters,
government officials, judicial systems, insurance companies, and so on.
“Safe exits” are another concern for AI design (Martin & Schinzinger
2010, 127). When an AI fails, will it fail in such a way that it is disastrous,
12
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
or will it fail “gracefully”? A self-driving car that fails by suddenly reverting
to human control with no warning while going 70 miles per hour on a sharp
curve is not providing a safe exit from failure. One which goes more slowly
on curves and which requires the human to have hands on the wheel at all
times, or which fails by slowing down and pulling over to stop, provides
a safer exit from failure.
Safety problems can be problems with the user, with the human-machine
interface, or with the machine itself. Further investigating problems with the
machine itself, the paper “Concrete Problems in AI Safety” gives a peek at
five technical hurdles to developing safe AI. These five problems, illustrated
by the example of a cleaning robot, are:
Avoiding Negative Side Effects: How can we ensure that our cleaning robot
will not disturb the environment in negative ways while pursuing its goals, e.g.
by knocking over a vase because it can clean faster by doing so? Can we do this
without manually specifying everything the robot should not disturb?
Avoiding Reward Hacking: How can we ensure that the cleaning robot won’t
game its reward function? For example, if we reward the robot for achieving an
environment free of messes, it might disable its vision so that it won’t find any
messes, or cover over messes with materials it can’t see through…
Scalable Oversight: How can we efficiently ensure that the cleaning robot
respects aspects of the objective that are too expensive to be frequently evaluated
during training? For instance, it should throw out things that are unlikely to
belong to anyone, but put aside things that might belong to someone (it should
handle stray candy wrappers differently from stray cellphones)…
Safe Exploration: How do we ensure that the cleaning robot doesn’t make
exploratory moves with very bad repercussions? For example, the robot should
experiment with mopping strategies, but putting a wet mop in an electrical
outlet is a very bad idea.
Robustness to Distributional Shift: How do we ensure that the cleaning
robot recognizes, and behaves robustly, when in an environment different from
its training environment? For example, strategies it learned for cleaning an
office might be dangerous on a factory work floor (Amodei et al. 2016).
While these are problems of function, the further problems of actual use
of technologies are another issue, to be dealt with below under 3. and 4.
6(2)/2018
13
B R I A N PAT R I C K G R E E N
2.2. Transparency, Opacity, and Privacy – “Can it be understood?”
After questions of bare function (can the act be done, a prerequisite for
ethics), the next question is one of facts: one must always know the facts of
a case before attempting to render judgment. Because AIs relying on machine
learning and deep learning may be quite obscure in the specifics of their
operation, as moral “agents” they are epistemologically and “cognitively”
opaque (in quotes because it is agency and cognition by analogy). In cases
involving AI, our lack of understanding means that we should increase the
“error-bars” on the anticipated risks of our decisions and thereby become
more cautious and risk averse. It also means we might, based on AI recommendation or action, inadvertently make some very bad decisions – or
reject what looks like a bad decision, but is actually a good one – and we
will not understand why.
When a human makes a mistake or does something evil we ask them “why
did you do that?” And the human may or may not give a satisfactory answer.
Will our AIs be able to tell us why they did something? The Future of Life
Institute’s “Asilomar AI Principles” include two principles on transparency,
but gaining transparency remains a challenge (Future of Life Institute 2017).
However, even if an AI were capable of explaining its reasoning, would any
human be able to understand it? In 2014, a computer proved a mathematical
theorem, the “Erdos discrepancy problem,” using a proof that was, at the
time at least, longer than the entire Wikipedia encyclopedia, a 13 gigabyte
data file (Konev & Lisitsa 2014; Yirka 2014). Explanations of this sort might
be true explanations, but humans will never know for sure.
Note that this lack of transparency gives a certain type of “privacy” to
the internal “thoughts” of AI machines. In general, privacy is a right for weak
agents (like individual humans) and transparency is a duty for strong agents
(like a government). What about AIs? As tools they ought to be transparent,
and more transparent the more power they have.
Perhaps AI systems need an introspective capacity that constantly
figures out a way to convey to humans in plain language what exactly the
AI is “thinking” as it goes about its activities. This will require the ability
14
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
to give both mechanical (this happened because of X input into algorithm
Y) and teleological explanations (the machine was attempting to achieve
objective Z). It may not have to actually explain much to anyone, but it should
keep a record of these “thoughts” and be prepared to answer when asked.
The right to an explanation is a part of the EU’s General Data Protection
Regulation (GDPR), so this idea is already becoming a requirement. However,
of note is that an AI might become trained to give answers that humans like,
rather than accurate answers, thus “Goodharting” the explanation function
(Partnership on AI 2018b) – “Goodhart’s Law” being the rule that “when
a measure becomes a target, it ceases to be a good measure” (Strathern
1997, 308). In other words, if humans prefer seemingly understandable or
attractive answers to factually correct ones the AI might learn to lie in order
to satisfy us. The phenomenon of “fake news” already indicates that humans
often prefer falsehoods to truth, and as the Bible warns of this predilection
as well (2 Timothy 4:3–4).
Note also that the opacity of understanding with an AI can be analogized
to our relationship to God. God is a “superintelligence” (to use philosopher
Nick Bostrom’s terminology (Bostrom 2014)), and we cannot understand
what God is doing, we can only trust that God is doing the right thing and
that our cooperation with God will ultimately turn out for the best. Soon we
may come to relate to AIs with this sort of faith too. For example, as Bishop
Robert Barron of Los Angeles has noted, as people navigate using the Waze
app, it may make strange route recommendations that turn out to be for
the better for us (Barron 2015). The app perceives and applies vastly more
information than a human can, for the sake of facilitating our travel. And yet
Waze can still have blind spots and still make very bad recommendations.
Waze is not the god of travel and traffic; it is a human-made tool, with
weaknesses.
2.3. Immense Capacity for Evil – “How shall it be limited?”
Just as human intelligence is a powerful force, so too will AI be. Just as
humans can apply their intelligence towards evil ends, finding ever newer
6(2)/2018
15
B R I A N PAT R I C K G R E E N
and more fiendish ways to harm each other, so too will AI, at the bidding
of its human masters.
At this point in history, humanity finds itself with immense powers,
powers greater than that of the ancient Greek gods, and an ethics weighted
for much smaller scales. As Hans Jonas noted decades ago, this should be
cause for great concern (Jonas 1984). Never before had ethics to consider
what Jonas believes to be the one truly categorical imperative: that humans
should exist. Before the development of large numbers of nuclear weapons,
extinction, caused by our own actions, was never within the scope of human
choice – but now that it is, this prior a priori must be brought to the fore of
ethics. That humans (or at least some material, free, and rational creatures)
exist is necessary for there to be ethics at all, and therefore it must be the
paramount goal of ethics to maintain human survival. No human; no ethics.
Another way to think of this is that in the past humans were very
weak, and in this weakness many of our decisions were made for us, by our
weakness. To a Roman emperor, inflicting wrath upon a hated foe involved
effortful marching or sailing of troops long distances. No matter how mad
the emperor was, he could not launch thousands of nuclear warheads,
incinerating his foes in minutes. But now we can. While formerly we
were involuntarily constrained by our weakness, now we must learn to be
voluntarily constrained by our ethics (Green 2017a). If we do not learn this,
we may soon face catastrophe on a scale never before seen.
Intelligence and power give us choices. Ethics helps us to determine
which among the available choices are actually good. We should want to
be efficient at good and we should want to be inefficient at evil. If instead
we are efficient at doing evil and inefficient at doing good we will come to
live in a terrible world. AI will make us more efficient at whatever we decide
to apply it towards. We should apply it towards reducing our efficiency at
doing evil and at enhancing our efficiency to do good (Green 2017a).
Because of its power, AI presents an existential risk to humanity. It is
a smart means that can be employed for intelligent and good ends, or for
unintelligent and evil ends. As such, it simply makes us more effective at
action, good or evil. Before we become so effective at stupidity and evil, we
16
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
should first become more effective at controlling ourselves, at recognizing
and avoiding the temptations of evil, and at caring for each other (Green
2018b). In different terminology, we might say that we need to “choose life,
so that you and your descendants may live” (Deuteronomy 30:19). But this
must be a constant struggle, continued with great diligence and taught and
learned in every generation, never solvable once and for all.
2.4. Immense Capacity for Good – “How shall it be used?”
For every negative regulation of action there is also a positive exhortation
to action, e.g. “do not kill” becomes “promote life.” The dangerous side of AI
is matched by a genuinely hopeful side where AI helps humankind achieve
never-before seen feats of beneficence. For example, in matters of research,
science, healthcare, data analysis, meta-analysis, and so on, AI has already
shown itself to be able to find hidden patterns that no human could find.
For example, AI assisted medical research is an active field right now, from
diagnostics to drug discovery, and more (Mukherjee 2017).
Another field that may be potentially revolutionized by AI is energy
efficiency. Recently Google’s DeepMind AI evaluated Google’s datacenters
to see where gains in efficiency might be found. DeepMind discovered a way
to save a whopping 40% on energy use for cooling in datacenters, a 15%
reduction overall, which, with datacenters consuming many gigawatts of
power, is quite significant (Evans & Gao 2016).
Of note is that DeepMind, as a company, began by training its AIs to
play games. The theory behind this strategy was that anything in the world
that can be gamified – re-interpreted or set up as a game – can also be
“won.” Thus, if the goal of the “game” is to reduce energy usage, then the
AI can figure out how that might be accomplished by analyzing the data
and then proposing a better model for energy efficiency. One question now
is how other human problems might be solved by characterizing them as
“games”? Can we solve the “game” of giving everyone in the world access
to adequate nutrition? Can we solve the “game” of helping everyone gain
access to sanitation? Can we solve the “game” of extending human healthy
6(2)/2018
17
B R I A N PAT R I C K G R E E N
life? Of traffic and housing? Of taxes? Of politics? Of international relations?
Of terrorism? Of nuclear war?
One example of a concrete opportunity for AI is revolutionizing education. Education is currently a very inefficient (think of the many students
for whom it is ineffective) and non-digitized field. Education is very labor
intensive and, for better or for worse, is based on human relationships. In
the future this may no longer be so, as students strap on virtual reality (VR)
headgear and interact with AI teachers which can conduct their lessons at
a personalized pace in a gamified environment. Brilliant students will be
discovered early and proceed at their proper pace and not grow bored in
class, while students who need more help can be educated with the most
sophisticated available techniques and diagnostics to assure that they
receiving the best education possible. But will this really be an improvement
for education, or just a cost-saving measure, putting millions of teachers
out of work?
AI even gives the greatest human masters the chance to gain enhanced
understanding and skill. World champion of Go, Ke Jie, said playing AlphaGo
was like playing a “god of Go,” and declared that now he would use it as his
teacher (Mozur 2017). In what other areas of human endeavor will AI be
able to teach us new things? Perhaps theology and ethics? The search for
tractable problems that can maximize AI’s benefit has been underway for
years and is continuing.
2.5. Bias in Data, Training Sets, etc. – “Will it be fair?”
Algorithmic bias is one of the major concerns in AI and will remain so in
the future unless we endeavor to make our technological products better
than we are. As one person said at a recent meeting of the Partnership on
AI, “We will reproduce all of our human faults in artificial form unless we
strive right now to make sure that we don’t” (Partnership on AI 2017). One
of the interesting things about neural networks, the current workhorses of
artificial intelligence, is that they effectively merge a computer program with
the data that is given to it. This has many benefits, but it also demonstrates
18
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
the rule of “garbage-in, garbage-out” (GIGO). If an AI is trained on biased
data, then the AI itself will be biased.
Algorithmic bias has been discovered, for example, in areas ranging
from word associations (Caliskan, Bryson, & Narayanan 2017) to photograph captioning (BBC 2015) to criminal sentencing (Angwin et al. 2016).
These biases are more than just embarrassing to the corporations which
produce these products; they have concrete negative and harmful effects
on the people who are victims of these biases, as well as reducing trust in
corporations, government, and other institutions which might be using
these biased products. In the worst scenarios, a poorly trained and biased
AI could make truly disastrous decisions, for example, misinterpreting data
indicating a nuclear attack when there was none. Biased data in extreme
form can therefore be an existential threat, and so is worthy of serious
effort to solve. Nevertheless, as vital as it may be it is also very difficult, and
therefore may delay the implementation of AI systems (but if AI systems
are dangerously biased they should be delayed until improved).
2.6. AI Induced Unemployment – “What will everyone do?”
As AI comes to replace mere humans at innumerable tasks ranging from
driving, to medical diagnostics, to education, etc., many humans will be put
out of work. What will millions of drivers, teachers, lawyers, and other people
do with their time when they are unemployed? What purpose will they find
in their lives? More crucially, what is the purpose of life? In a pluralistic
society we leave this up to the individual to decide. But will people decide
well? The recent strengthening of ethno-nationalist movements, terrorism,
and other forms of radicalization, should give us pause to consider the merits
of people with too much time on their hands and without purpose. Perhaps
with the purpose of mere survival attained, life has become “too easy,” and
with religion also in decline, video games and “screen time” ascending (giving
brief respite from purposelessness), and the internet spreading pernicious
ideas like wildfire (often with algorithmic help, e.g., YouTube radicalization
(Tufekci 2018)), we should sincerely ask ourselves what this life is for and
6(2)/2018
19
B R I A N PAT R I C K G R E E N
what we are supposed to do with it. With millions or billions of labor hours
freed up, will these newly freed people turn to loving their neighbors and
making the world a better place? Perhaps. Or perhaps the opposite, as the
saying goes: “Idle hands are the Devil’s playthings.”
For people who give purpose to their lives through their work, this loss
will be very serious indeed. But many, if not most, people do not get their
life’s meaning from their work. Instead they get it from their family, their
religion, their community, their hobbies, their sports teams, or other sources,
and so life for many people may go on. However, all of this assumes that
the unemployed will somehow be fed and sheltered, despite their lack of
gainful employment; and this assumption might not be correct, particularly
in nations with weak social safety nets. Inequality will almost certainly
increase, as those who are masters of AI labor gather that slice of wealth
that once would have gone to paying for human labor.
2.7. Growing Socio-Economic Inequality – “Who gets what?”
AI will facilitate and accentuate the continued division of society into the
powerful and powerless, with technical skill and ownership of capital as the
determining factors and outrageous socio-economic inequality as the effect.
Some have suggested a universal basic income (UBI) to redistribute
wealth (Van Parijs & Vanderborght 2017). This could move wealth from the
massive technologically-induced hoards forming around the major investors
in such companies such as Alphabet, Amazon, Apple, and Facebook. However,
it is hard to see how some nations would transition to what is essentially
something like a “negative income tax.” However, if we do not find a way
to redistribute technological wealth, then even though the prices of many
commodities will fall (due to the enormous gains in capital efficiency from
AI replacing labor) most people will not benefit.
Perhaps rather than a UBI we should instead pay people to help their
neighbors, create art, beautify their towns and cities, and otherwise make
gainful employment out of what humans do better than AI: loving one
another and creating beauty. Any as yet foreseeable form of AI cannot
20
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
love us; an AI might simulate such a thing, but it would be a sham. Right
now people who care for people – stay at home parents, those who care
for the elderly, social workers, those who run soup kitchens and homeless
shelters, etc. – are woefully undercompensated for the vital work they
do in maintaining human society. Perhaps with the coming AI economic
revolution, and sufficient adjustment to policy, they might finally receive
a more fair compensation for their labors.
Or perhaps there could be not a universal basic income, but a “universal
payment-for-others” a system where people could not spend the money
on themselves, but only pay other people for their work, or reward them
for good deeds. This would create an economy based on the small-scale
redistribution of taxed super-wealth. It could prevent both the de-skilling
of labor and the de-skilling of (very small-scale) management, and it would
decentralize the economy to the individual level (though also massively
centralizing it through government taxes).
2.8. Automating Ethics and Moral De-skilling –
“Lack of practice makes imperfect”
We can calculate with calculators. We can spell with autocorrect. More and
more tasks can be outsourced to technology, and the trend seems inexorable,
perhaps someday automating everything that needs to be done. But what
will be left of humanity after we outsource everything? Only our desires
and our angst? With all our decisions made for us, will we lose our moral
character? Or will we instead use AI and VR to help us train ourselves into
being more virtuous than we have ever been before? Or, contrariwise, to
become more callous and evil than ever before?
Machine ethics – the study of imbuing machine with moral decision-making capacity, thus creating “artificial moral agents” – has a relatively
long history, going back to such literary sources as Asimov’s Three Laws of
Robotics, and more recent scholarly sources as those by Gips (1994), Allen,
Varner, & Zinser (2000), Floridi & Sanders (2004), Arkin (2009). Recently
the matter has taken on more urgency, however, as AI becomes more and
more powerful and more and more capable of making disastrous choices.
6(2)/2018
21
B R I A N PAT R I C K G R E E N
The “value alignment problem” has gained particular concern, as an AI
determined to perform actions at odds with human wishes could be quite
risky (Arnold, Kasenberg, & Scheutz 2017). Certainly, we ought to think
carefully about how AI’s ought to behave, but the less-considered side-effect
of this automation of ethics will be human moral debility.
As we explore the space into which AI will grow, there might be places
from which we ought to restrict it. One of these places might be certain types
of moral decision-making. As Aristotle noted millennia ago, good moral
decision-making requires experience. While there might be children who
are very kind and very well-behaved, children are not known for their moral
discernment and prudence (Aristotle 1908, book VI, chap. 8). There is too
much that they do not yet know. While AIs might currently be like children
to us, soon they may grow up to be more like our parents, making choices
for us on our behalf. AI could thereby infantilize us, denying us the ability
to ever grow up through the experience of making our own moral choices
and experiencing our own moral freedom. Moral de-skilling is a danger if
we outsource too much decision-making power to our AIs (Vallor 2015;
Vallor 2016).
One vital component of education in the future may be the use of AI
for moral character formation, perhaps using VR to practice moral decision-making by being inserted into hundreds of historical and fictional cases.
In a future that will be so dependent upon humans making good choices,
AI-assisted moral education, if it can be done well, could be a crucial part
of developing a good future and not a bleak one. Unfortunately, humanity
has many problems even now with moral education, so it is not clear that
an AI-enhanced education will manage to do any better than we already do.
2.9. Robot Consciousness and Rights –
“No robot justice, no peace?”
At some point in the future we may have AIs that can fully mimic everything
about a human being. Will they have their own volition and desires? Will
they be conscious? Will we be able to tell if they are conscious or not? Will
22
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
they then deserve “human” rights? While the AI consciousness question is
not currently answerable, in the future as AI systems grow in complexity, it
may be. And if an AI were to attain consciousness, should it gain legal rights?
Currently various nations have widely differing laws on legal personhood.
In many nations, entities that are not human can have legal personhood
(e.g., corporations), while some biological humans are not granted legal
personhood (such as the unborn). In some nations, geographic features have
attained legal personhood status, such as rivers in Australia, New Zealand,
and India (O’Donnell & Talbot-Jones 2018).
If rivers and corporations can be persons, there seems to be no reason
why an AI could not attain legal personhood status. The question, of course,
would be the incentives for choosing to do so. Corporate personhood acts to
shift legal responsibility away from individual human employees and onto
the corporation, thus insulating those people from the possible negative
effects of their decisions. If AI could be granted personhood in order to shirk
individual human responsibility, then some people would certainly like
that. On the other hand, if an AI were allowed to be a plaintiff in a lawsuit,
like a legal person such as a river suing government or industry, then the
situation would be quite different.
Lastly, there is sometimes expressed the fear of a robot rebellion – that if
we mistreat our robots and AIs they will someday turn on us and destroy us
– often recurring in fiction, yet of more concern now that lethal autonomous
weapon systems are being developed. The connection to rights, the thinking
goes, is that if perhaps we give robots rights, then they will be appeased and
not turn against us. This fear assumes certain aspects of robot mentality
such as consciousness and volition which may not be possible, or at least
assumes widespread computer hacking to turn the robots against humanity.
In the midst of this uncertainty, however, we do know one thing: treating
human-like entities badly harms the moral character of the agent doing
the action. This is a foundation of virtue ethics – practicing evil habituates
the agent towards evil. Even if we do not know the moral status of robots,
we do know the moral status of people who would mistreat robots, and we
should not want that mistreatment to happen.
6(2)/2018
23
B R I A N PAT R I C K G R E E N
2.10. Dependency – “No going back”
We might like to think our technology depends on us, but with every new
technology we make, we become dependent on it as well. This reciprocal
dependency traps us in an ever expanding network of techno-social relationships where some network nodes or relationships, if lost, could lead to
disaster. The more reliable a technology seems, the fewer backup systems
we retain in case that system fails. For example, in countries with unreliable
electricity, wood or kerosene may serve as backup fuel sources for heating or
cooking. But in nations with reliable electricity, those backup systems tend
to be atrophied. Loss of electrical power or cellular connectivity may be an
annoyance for a few hours, but after days or weeks could cause innumerable
crises, shutting down water, sanitation, hospitals, transportation, and causing
mass confusion and lack of social coordination. These can happen in our
current world – adding AI will only increase our dependency.
When we have turned over innumerable tasks to AIs – such as driving
vehicles and coordinating the communications and financial systems – we
may become utterly dependent on them. If our highly efficient and centrally
coordinated self-driving transportation system were to suddenly “crash”
in a metaphorical sense, perhaps due to a software glitch or malicious
hacking, if could cause many literal crashes in the real world. Natural and
human-made disasters (either accidental or intentional) are inescapable,
and if we come to rely on our technology so much that we allow our backup
systems to atrophy (e.g., no longer teaching people how to drive or entirely
removing controls from cars) then, like Japan’s Tohoku earthquake followed
by the Fukushima nuclear meltdown, we are entering a world where we
will encounter not only one initial disaster, but for a second technological
disaster as well. Finding the balance between systemic efficiency (e.g.,
a highly complex centralized system can be fragile due to intricacy and lack
of redundancy) and robustness (e.g., tough systems are often inefficient
due to decentralization, redundancy, and backups) will be a major task of
future society.
24
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
2.11. Social-Psychological Effects – “Too little and too much”
Technology has been implicated in so many negative social-psychological
trends, including loneliness, isolation, depression, stress, anxiety, and
addiction, that it might be easy to forget that things could be different.
Smartphones, social media, and online games in particular have become
problems, even leading to deaths from such causes as cyberbullying and
neglect. As just one example, society is in the middle of a crisis of loneliness,
for everyone from young to old. The problem has become so severe that the
UK has appointed a minister for loneliness, with government data indicating
that “200,000 older people in the UK have not had a conversation with
a friend or relative in more than a month” (Mannion 2018).
One might think that “social” media and smartphones could help us feel
connected at all times, but those technologies seem to be the source of the
problem rather than the remedy. What does seem to help, then? Strong,
in-person relationships are the key to fighting off many of the negative
trends caused by technology, but unfortunately these are exactly the things
that are being pushed out by addictive technology.
Will AI increase these bad social trends or might it be able to help
remedy them? Perhaps AI agents could help to overcome loneliness, but
this is speculation, and might only add to the problem. If the problem is
shallow human relationships due to technology pushing out relational
depth, more technology will not help. Instead we need to figure out how
we can help foster deeper more meaningful relationships. If technology can
actually help people connect in these deeper ways it might help, but the
technology would then need to get out of the way and let human interaction
flourish. Religious believers ought to ask what role religious communities
might play in remedying this crisis.
2.12. Effects on the Human Spirit – “What will this mean for humanity?”
All of the above areas of interest will have effects on how humans perceive
themselves, relate to each other, and live their lives. But there is a more existential question too: if the purpose and identity of humanity has something
6(2)/2018
25
B R I A N PAT R I C K G R E E N
to do with our intelligence (as many philosophers have believed), then by
externalizing our intelligence and improving beyond human intelligence,
are we risking making ourselves second-class beings to our own creations?
This is a deeper question with artificial intelligence which cuts to the
core of our humanity, into areas traditionally reserved for philosophy,
spirituality, and religion. What will happen to the human spirit if or when
we are bested by our own creations in everything that we do? Will human
life lose meaning? Will we come to a new discovery of our identity beyond
our intelligence? Perhaps intelligence is not as important to our identity
as we might think it is, and perhaps turning over intelligence to machines
will help us to realize that. From a Christian perspective, we would do well
to remember that our capacity to love and be loved is more important
than our capacity to think, and that nothing can come between us and our
relationship with God, except sin, and then only if we let it. As with many
of the other above challenges, Christianity here could be poised to help
guide a wandering humanity in dire search for meaning, but will it rise to
this calling? Religion, spirituality, and theology have much work to do to
prepare for the strange future unfolding before us.
Conclusion
Artificial intelligence, like any other technology, will just give us more of
what we already want. Whereas we could once have only a trickle of what
we wanted out of life, and the powerful took what little there was, now
we have a firehose of wants being fulfilled (food, drink, entertainment,
pornography, drugs, gamified feelings of accomplishment, etc.). The firehose
will continue to grow and become a deluge washing away our desires and
leaving what, exactly, of us behind? What skeleton of humanity will remain
when technology has given us, or perhaps distorted or replaced, all our
fleshly desires? What will this skeleton of humanity be made of? Will our
technological flesh truly satisfy us, or only leave us in a deeper existential
malaise, filled with angst, despair, and dread? What will we want, when we
want for nothing – at least nothing material? What of human nature will
26
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
remain once our every worldly telei is fulfilled? Perhaps only the worst of
us will remain. Or perhaps the best. Or perhaps, as always, both.
We are grasping ourselves by our desires; or at least some of our desires.
Will this be good for us? Will it destroy us? Should we want these things?
How could we know what we should want? In this context the Ninth and
Tenth Commandments could do us well: “Do not covet…” (Exodus 20).
Far from desiring evil, do not even want it. Squash the evil desires in your
heart before they can even approach the external world. In the context of
technology, technologist Bill Joy has already warned us that we must not
only relinquish possession of the worst technologies, we must relinquish our
desire for them (Joy 2000). In this, Joy echoes Pope John XXIII’s sentiment in
Pacem in Terris that we need (at the time, in the context of nuclear weapons)
a disarmament that is “thoroughgoing and complete, and reach men’s very
souls” (John XXIII 1963).
We are conducting this experiment called human history, and no one yet
knows how it may end. But as we proceed, we can hope that intelligences of
our own fashioning will help us, and not harm us. To go beyond mere hope,
into ethics and action, is the responsibility of those who are able to affect
the necessary changes to make a better future.
Acknowledgements
I would like to thank the Pacific Coast Theological Society, Markkula Center
for Applied Ethics, and the University of the Pacific Pope John XXIII Lectureship for providing feedback on earlier drafts of this work, the Pope John
XXIII Lectureship for modest financial support to produce an earlier draft,
as well as the Markkula Center for Applied Ethics and School of Engineering
at Santa Clara University for encouraging me to spend time on this work.
References
Allen, Colin, Gary Varner & Jason Zinser. 2010. “Prolegomena to any future artificial
moral agent.” Journal of Experimental & Theoretical Artificial Intelligence 12,
Issue 3, 09 Nov: 251–261.
6(2)/2018
27
B R I A N PAT R I C K G R E E N
Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan
Mané. 2016. “Concrete problems in AI safety.” arXiv preprint, arXiv:1606.06565.
https://arxiv.org/abs/1606.06565.
Angwin, Julia, Jeff Larson, Surya Mattu and Lauren Kirchner. 2016. “Machine Bias.”
ProPublica, May 23. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
Aristotle. 1908. Nicomachean Ethics. Translated by W. D. Ross. Book VI, Ch. 8. http://
classics.mit.edu/Aristotle/nicomachaen.6.vi.html.
Arkin, Ronald. 2009. Governing Lethal Behavior in Autonomous Robots. Boca Raton,
Florida: CRC Press.
Arnold, Thomas, Daniel Kasenberg, and Matthias Scheutz. 2017. “Value Alignment
or Misalignment – What Will Keep Systems Accountable?” Association for the
Advancement of Artificial Intelligence. https://hrilab.tufts.edu/publications/
aaai17-alignment.pdf.
Barron, Robert. 2015. “The ‘Waze’ of Providence.” Word on Fire, website. December
1. https://www.wordonfire.org/resources/article/the-waze-of-providence/4997/.
Bossmann, Julia. 2016. “Top 9 ethical issues in artificial intelligence.” World Economic Forum: Global Agenda, 21 Oct. https://www.weforum.org/agenda/2016/10/
top-10-ethical-issues-in-artificial-intelligence/.
Bostrom, Nick. 2014. Superintelligence: Paths, Dangers, Strategies. Oxford: Oxford
University Press.
BBC. 2015. “Google Apologises for Photos App’s Racist Blunder.” BBC News. 1 July.
http://www.bbc.com/news/technology-33347866.
Caliskan, Aylin, Joanna J. Bryson, Arvind Narayanan. 2017. “Semantics derived
automatically from language corpora contain human-like biases.” Science 356
(6334) (14 April): 183–186.
Cannon, Lincoln. 2015. “What is Mormon Transhumanism?” Theology and Science
13 (2): 202–218.
Chella, Antonio, & Ricardo Manzotti. 2013. Artificial Consciousness. Exeter, UK:
Imprint Academic.
Chrisley, Ron. 2008. “Philosophical foundations of artificial consciousness.” Artificial
Intelligence in Medicine 44 (2) (October): 119–137.
Evans, Richard, and Jim Gao. 2016. “DeepMind AI Reduces Google Data Centre
Cooling Bill by 40%.” DeepMind Blog, 20 July. https://deepmind.com/blog/
deepmind-ai-reduces-google-data-centre-cooling-bill-40/.
Floridi, Luciano, and Jeff W. Sanders. 2004. On the morality of artificial agents. Minds
and machines 14 (3) (August 1): 349–379.
28
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
Future of Life Institute. 2017. “Asilomar AI Principles.” Principles 7 and 8. https://
futureoflife.org/ai-principles/.
Gips, J. 1994. “Towards the Ethical Robot.” In Android Epistemology, edited by Kenneth
M. Ford, C. Glymour & Patrick Hayes. Cambridge, Mass.: MIT Press.
Green, Brian Patrick. 2018b. “The Technology of Holiness: A Response to Hava
Tirosh-Samuelson.” Theology and Science 16 (2): 223–228.
Green, Brian Patrick. 2018a. “Artificial Intelligence and Ethics: Twelve Areas of
Interest.” Pope John XXIII Memorial Lecture, University of the Pacific, Stockton,
California, March 21.
Green, Brian Patrick. 2017c. “Artificial Intelligence and Ethics: Ten Areas of Interest.”
All about Ethics Blog. Markkula Center for Applied Ethics, website. November 21.
https://www.scu.edu/ethics/all-about-ethics/artificial-intelligence-and-ethics/.
Green, Brian Patrick. 2017b. “Some Ethical and Theological Reflections on Artificial
Intelligence.” Conference paper delivered to the Pacific Coast Theological
Society, at the Graduate Theological Union, Berkeley, California, November 3.
http://www.pcts.org/meetings/2017/PCTS2017Nov-Green-ReflectionsAI.pdf.
Green, Brian Patrick. 2017a. “The Catholic Church and Technological Progress:
Past, Present, and Future.” Religions 8(6), 106. http://www.mdpi.com/20771444/8/6/106/htm.
Harris, Mark. 2017. “Inside the First Church of Artificial Intelligence.” Wired (November 15). https://www.wired.com/story/anthony-levandowski-artificial-intelligence-religion/.
John XXIII. 1963. Pacem in Terris; Vatican City: Libreria Editrice Vaticana. Available
online: http://w2.vatican.va/content/john-xxiii/en/encyclicals/documents/
hf_j-xxiii_enc_11041963_pacem.html.
Jonas, Hans. 1984. The Imperative of Responsibility. Chicago: University of Chicago
Press.
Joy, Bill. 2000. “Why the Future Doesn’t Need Us.” Wired, April 1. http://archive.
wired.com/wired/archive/8.04/joy_pr.html.
Koene, Randal A. 2013. “Uploading to Substrate‐Independent Minds.” Chapter 14 in
The Transhumanist Reader: Classical and Contemporary Essays on the Science,
Technology, and Philosophy of the Human Future, edited by Max More and
Natasha Vita‐More. New York: Wiley.
Konev, Boris, and Alexei Lisitsa. 2014. “A SAT Attack on the Erdos Discrepancy
Conjecture.” arXiv.org, 17 February. arXiv:1402.2184v2.
Kurzweil, Ray. 2012. How to Create a Mind: The Secret of Human Thought Revealed.
New York: Penguin.
6(2)/2018
29
B R I A N PAT R I C K G R E E N
Labrecque, Cory Andrew. 2017. “The Glorified Body: Corporealities in the Catholic
Tradition.” Religions 8 (9): 166. http://www.mdpi.com/2077-1444/8/9/166/htm.
Leviathan, Yaniv, and Yossi Matias. 2018. “Google Duplex: An AI System for Accomplishing Real-World Tasks Over the Phone.” Google AI Blog, May 8. https://
ai.googleblog.com/2018/05/duplex-ai-system-for-natural-conversation.html.
Mannion, Lee. 2018. “Britain appoints minister for loneliness amid growing isolation.”
Reuters. January 17. https://www.reuters.com/article/us-britain-politics-health/
britain-appoints-minister-for-loneliness-amid-growing-isolation-idUSKBN1F61I6.
Martin, Mike W., & Roland Schinzinger. 2010. Introduction to Engineering Ethics,
Second Edition. Boston: McGraw-Hill.
Mozur, Paul. 2017. “Google’s AlphaGo Defeats Chinese Go Master in Win for A.I.”
The New York Times. May 23. https://www.nytimes.com/2017/05/23/business/
google-deepmind-alphago-go-champion-defeat.html.
Mukherjee, Siddhartha. 2017. “A.I. versus M.D.: What happens when diagnosis is automated?” New Yorker. April 3. https://www.newyorker.com/magazine/2017/04/03/
ai-versus-md.
Partnership on AI. 2018b. Safety-Critical Working Group Meeting, San Francisco,
USA, May 24.
Partnership on AI. 2018a. “Thematic Pillars.” Partnership on AI website. https://
www.partnershiponai.org/thematic-pillars/.
Partnership on AI. 2017. Inaugural Meeting, Berlin, Germany, October 23.
Platt, John R. 2016. “A Starfish-Killing, Artificially Intelligent Robot Is Set to
Patrol the Great Barrier Reef.” Scientific American. January 1. http://www.
scientificamerican.com/article/a-starfish-killing-artificially-intelligent-robot-is-set-to-patrol-the-great-barrier-reef/.
Schlagel, Richard H. 1999. “Why not Artificial Consciousness or Thought?” Minds
and Machines 9 (1) (February): 3–28.
Searle, John R. 1980. ‘Mind, brains, and programs.’ Behavioral and Brain Sciences
3 (3): 417–457.
Strathern, Marilyn. 1997. “‘Improving ratings’: audit in the British University system.”
European Review 5, No. 03 (July): 305 – 321. http://conferences.asucollegeoflaw.
com/sciencepublicsphere/files/2014/02/Strathern1997-2.pdf.
Tufekci, Zeynep. 2018. “YouTube, the Great Radicalizer.” The New York Times, March
10. https://www.nytimes.com/2018/03/10/opinion/sunday/youtube-politics-radical.html.
Vallor, Shannon. 2016. Technology and the Virtues: A Philosophical Guide to a Future
Worth Wanting. New York: Oxford University Press.
30
6(2)/2018
E T H I C A L R E F L E C T I O N S O N A RT I F I C I A L I N T E L L I G E N C E
Vallor, Shannon. 2015. “Moral Deskilling and Upskilling in a New Machine Age:
Reflections on the Ambiguous Future of Character.” Philosophy of Technology
28: 107–124.
Van Parijs, Philippe, and Yannick Vanderborght. 2017. Basic Income: A Radical
Proposal for a Free Society and a Sane Economy, 1st Edition. Cambridge, Mass.:
Harvard Univerrsity Press.
Yirka, Bob. 2014. “Computer generated math proof is too large for humans to check.”
Phys.org, website. February 19. https://phys.org/news/2014-02-math-prooflarge-humans.html.
6(2)/2018
31