Rationality and Ethics in Artificial Intelligence
Rationality and Ethics in Artificial Intelligence
Rationality and Ethics in Artificial Intelligence
in Artificial Intelligence
Rationality and Ethics
in Artificial Intelligence
Edited by
All rights for this book reserved. No part of this book may be reproduced,
stored in a retrieval system, or transmitted, in any form or by any means,
electronic, mechanical, photocopying, recording or otherwise, without
the prior permission of the copyright owner.
Chapter I ..................................................................................................... 1
A Brief History of Computer Ethics and how it is Connected to AI Ethics
Mariana Todorova
Chapter II .................................................................................................. 10
Ethical Clashes in the Prospects for Autonomous Vehicles
Silviya Serafimova
Chapter IV ................................................................................................ 52
The AI-Run Economy: Some Ethical Issues
Anton Gerunov
Chapter V ................................................................................................. 71
Scratch My Back & I Will Scratch Yours: Putting Game Theory and
Artificial Intelligence on a Converging Path beyond Computational
Capabilities
Boris Gurov
Chapter VI ................................................................................................ 93
Artificial Intelligence in Defence
Todor Dimitrov
MARIANA TODOROVA
In his article “A very short history of computer ethics”,1 the author Terrell
Bynum defines computer ethics as a scientific field that emerged in the first
years of the outbreak of World War II, beginning with Massachusetts
Institute of Technology professor Norbert Wiener. He defined the new field
while participating in the development of an anti-aircraft weapon, the
purpose of which was to intercept and track an enemy aircraft, then calculate
its probable trajectory and inform the other parts of the weapon to activate
the projectile. The emerging challenges for engineers, according to Bynum,
led to the creation of a new branch of science by Wiener and his colleagues,
which they called cybernetics, the science of information feedback systems.
It was cybernetics combined with the digital computers created at the time
that motivated the inventor to draw several important ethical conclusions.
In 1950, he (Wiener) published the first book of its kind on computer ethics
(though he nowhere explicitly calls his reasoning that way): The Human Use
of Human Beings,2 where he spoke of the benefits and risks of automation
and the use of computers. The text sounds like a come true (self-fulfilling
prediction), as Wiener predicts that computers will enhance human
capabilities, free people from repetitive manual labor, but also allow for
processes of dehumanization and subordination of the human species. The
Wiener adds that the invention of digital computers will lead to a second
industrial revolution, which will have multi-layered dimensions, will unfold
for decades and will lead to radical changes. For these reasons, he explicitly
warns in the chapter, “Someone Communication Machines and Their
Future,” as well as throughout the book, that workers must adapt to changes
in their jobs. Governments need to draft new laws and regulations.
Industries and businesses need to create new policies and practices.
Professional organizations need to prepare new codes for their members.
Sociologists and psychologists need to study new phenomena, and
philosophers need to rethink and redefine outdated social and ethical
concepts.
Walter Manner was the researcher who formalized the term “computer
ethics”, defining it as part of the applied (ethics) in his work: Starter Kit on
Teaching Computer Ethics3 in 1976. He devoted his subsequent work to
efforts to emancipate this title as a separate scientific field. The intention to
strictly distinguish computer ethics from fundamental ethical issues is
implicit.
James Moore, also known by his law of the same name, also dedicated an
article4 on this issue. He believes that the ambiguity surrounding computer
ethics arises because there is a political vacuum over how to use computer
technology. Through computers, we acquire new abilities that provide us
with new opportunities and choices for action. Very often, according to him,
there are no political measures for such situations, and if there are any, they
are inadequate. For Moore, a central task in computer ethics is to determine
what we need to do in specific computer-related situations, such as
formulating policies and action guides. Computer ethics, according to him,
must take into account both individual and social rights and policies.
Therefore, it identifies four areas of computer ethics: 1) identifying the
computer-generated policy vacuum; 2) clarification of conceptual
ambiguities; 3) formulation of policies for the use of computer technologies
4) ethical justifications.
Moore correctly outlines the steps that should be taken to fill the ethical and
then legal and regulatory gaps, but fails to point out that this task is too
ambitious to be executed. Philosophers, anthropologists, psychologists and
neuroscientists must take the lead in such a task, but they must work
alongside representatives of labor and social systems, education, medicine,
security and military technology. That is, with experts from all fields who
will be influenced by computer technology and artificial intelligence.
4 Moor, James H. “What Is Computer Ethics?” In Computers and Ethics, ed. Terrell
Ward Bynum (Basil Blackwell, 1985), 266 – 275.
5 Deborah Johnson, “Computer Ethics”, Prentice-Hall, reprinted by Metaphilosophy,
She associates it with the times of the print media, arguing that Bentham
and Kant developed their ethical theories in response to this discovery, and
believes that the same will be repeated with computer ethics, which must
respond to the computer-induced revolution. According to her, the nature of
the expected phenomenon (computer revolution) is such that the ethics of
the future will have a global character, in the sense that it will also address
the integrity of human actions and relationships. Because computers know
no boundaries and operate globally, they will set the stage for a new global
universal ethic that must be universal to all human beings. The author adds
that all local ethics (considering both individual areas and cultural and
regional features (of Asia, Africa, Europe, America, etc.) may grow into a
global ethic, inheriting computer ethics in information era.
Johnson (1985) also talks about global ethics, but uses a different meaning.
For her, it will belong to new kinds of extensive moral problems. Inherited
and contemporary ethical theories will continue to be fundamental, and this
will not lead to a revolution in ethics. With the help of Terrell Bynum, two
opposite concepts of computer ethics are revealed to us. On the one hand,
there is the thesis of Wiener, Maner and Gorniyak-Kosikovska about a
revolution in ethics and an obligation of humanity to rethink its very
foundations, as well as of human life. On the other hand, Johnson's more
conservative view is presented, which defends the position that ethics will
remain “untouched” and that these are the same old ethical issues in a new
reading, which in turn will make computer ethics meaningless as a separate
part.
In the debate thus outlined, we can agree in part with statements from both
theses. Ethics that address information and computer technology, as well as
artificial intelligence, will be global and universal, as responses to the
28, 2019.
6 Chapter I
The concept behind Tallinn is that software does not need to be programmed
to destroy humanity, but can “decide” so along the course of its existence.
As we have already noted, this might be the product of a small error,
software bug, etc. The scientist also refers to the example of Bostrom, who
reveals that artificial intelligence could decide that the atoms in the human
body are a good raw material and can be used in another way as a resource.
Objections to such arguments come from the Technology Guild, which says
it is too early to seek a solution against hostility. They recommend shifting
the focus to current problems, such as the fact that most of the algorithms
were created by white men and that this fact has given rise to the biases that
accompany them.
Tallinn, on the other hand, believes that even if the power button is masked
and not of interest to artificial intelligence, there is still no solution to the
problem of potential threat, as artificial intelligence may have secretly
replicated itself hundreds of thousands of times on the web. For these
reasons, researchers and practitioners are united around the idea of artificial
intelligence being taught to recognize and study human values. That is,
according to Tallinn, he must learn to evaluate people outside the canons of
strict logic. For example, that we often say one thing but think another, that
we enter conflicts or think differently when we are drunk, and so on. a state
of irrationality (ibid.).
A Brief History of Computer Ethics and how it is Connected to AI Ethics 7
Tallinn takes as its best formula the statement of the Cambridge philosopher
Hugh Price, who defines that artificial intelligence in ethical and cognitive
aspects should be like a “superman”. Other questions arise - if we do not
want artificial intelligence to dominate us, then should we surpass it. And
these questions again inevitably lead us to the presence of consciousness
and free will in artificial intelligence.
The formula that Hugh Price, Boris Grozdanoff and other scientists offer is
correct, but much work is needed to precede it. Today, we are witnessing a
resurgent wave of neoconservatism, which sharply criticizes liberal theories
such as multiculturalism, globalism, etc. In parallel with these processes, we
can predict a resurgence of nationalism, hardening of the concepts of
“nation states” and “identities”. This context would certainly prevent
attempts to seek a universally valid formula for human ethics that could
eventually be applied as a matrix for the training of general artificial
intelligence. Cultural diversity and different human civilizational norms do
not share the same views on the categories of “good” and “bad”, human
rights, etc. their definition, which mostly fit world-class institutions such as
the UN.
References
Bynum, Terrell Ward. “A Very Short History of Computer Ethics”.
Accessed July 8, 2022.
https://web.archive.org/web/20080418122849/http:/www.southernct.edu/o
rganizations/rccs/resources/research/introduction/bynum_shrt_hist.html.
Górniak-Kocikowska, Krystyna. “The Computer Revolution and the
Problem of Global Ethics.” In Global Information Ethics, edited by
Terrell Ward Bynum and Simon Rogerson, 177–190. (Guildford:
Opragen Publications, 1996).
Grozdanoff, Boris D. “Prospects for a Computational Approach to
Savulescu’s Artificial Moral Advisor.” ȿɬɢɱɟɫɤɢ ɢɡɫɥɟɞɜɚɧɢɹ. ɛɪ. 5.,
A Brief History of Computer Ethics and how it is Connected to AI Ethics 9
SILVIYA SERAFIMOVA
Introduction
Clarifications
The moral issues regarding the use of autonomous vehicles (AVs) are not a
new phenomenon in the ethical discourse.1 Some of them date back to the
concerns about the so-called trolley dilemma, introduced by Philippa Foot
in 1967. The dilemma is a thought experiment according to which a
fictitious onlooker can choose to save five people in danger of being hit by
a trolley, by diverting the trolley to kill just one person.2
At first sight, the trolley dilemma looks like a utilitarian moral dilemma,
which is based upon calculating the maximization of well-being for more
representatives at the expense of the suffering of the few. If that were the
case, there would be no dilemma whatsoever. The solution would be one to
switch the trolley so that the five people can be saved. However, such a
decision is an act utilitarian decision par excellence.
1 For the challenges in relating the trolley cases to the ethics of AVs, see Geoff
Keeling,”The Ethics of Automated Vehicles,” (PhD thesis, University of Bristol,
2020), 45-68.
2 Philippa Foot, “The Problem of Abortion and the Doctrine of the Double Effect,”
that could guide the machines’ behavior”.9 The largest project of how one
can “train” AI morality in such situations is so-called Moral Machine
experiment. It is an online experimental platform that is designed to
examine the moral dilemmas faced by AVs. It collects data of millions of
moral decisions for the purposes of training machine-learning algorithms.10
Structure
The main objective of this paper is to demonstrate why finding some
potential solutions to the moral design problem and Moral Machine
experiment requires one to recognize the challenges in building AVs as
moral rather than purely computational challenges. For the purposes of
exemplifying the latter, I tackle the benefits and disadvantages of two types
of AVs projects, viz. the Rawlsian collision algorithm, which displays a
The moral design problem inherits some of the moral concerns about the
trolley dilemma, when examined from a utilitarian perspective. That is why
one should look to adopt another approach. An illuminative example of such
an approach is found in Rawls’ theory of justice. It is elaborated upon by
Leben into so-called Rawlsian collision algorithm.
Vehicles,” 4.
15 John Rawls, Theory of Justice (Cambridge, Mass.: The Belknap Press of Harvard
Vehicles.”, 2.
17 Rawls, Theory of Justice, 137. Keeling, ”Against Leben’s Rawlsian Collision
within the Rawlsian collision algorithm, the fatal injury is given priority in
making the corresponding decision.
The strive to expand the scope of the maximin rule by introducing the
leximin rule requires a reevaluation of Rawls’ idea of life project. The
leximin rule makes room for comparing the second-lowest survival
probabilities on the remaining alternatives, but does not shed light upon
whether it might be “more just” for the worst-off person or people to die
than to suffer debilitating injuries. In turn, this specification necessitates one
to reconsider Rawls’ original position. One should also analyze how the veil
of ignorance should be elaborated upon so that the graduation of justice can
meet the requirements of an internally differentiated group of worst-off
people.
As Keeling points out, Rawls does not assume the application of the
maximin rule as universally valid.24 He clearly denies its application as a
general principle of rational decisions in the case of risk and uncertainty.
This clarification puts in question the extrapolation of the maximin rule to
that of the leximin, when survival probabilities are at stake. One of the
reasons is that when such probabilities are tested, the parties reaching an
agreement should not be indifferent to their own and others’ life projects. A
specification that contradicts the requirements set by the veil of ignorance.
Vehicles,” 5.
26 Keeling, ”Against Leben’s Rawlsian Collision Algorithm for Autonomous
Vehicles,” 5.
Ethical Clashes in the Prospects for Autonomous Vehicles 17
Scenario 1
The AV can swerve left or right. If the AV swerves left, there is a 0% chance
that its passenger will sustain a fatal injury and a 100% chance that its
passenger will sustain a lifelong debilitating injury. If the AV swerves right,
there is a 1% chance that its passenger will sustain a fatal injury and a 99%
chance that its passenger will remain unharmed.
Vehicles,” 10.
29 Keeling, ”Against Leben’s Rawlsian Collision Algorithm for Autonomous
Vehicles,” 10-11.
18 Chapter II
Scenario 2
Keeling’s main concern about the second scenario is that the maximin rule
gives “undue weight to the moral claims of the worst-off”.30
The AV can swerve left or right. If the AV swerves left, there is a 100%
chance that its passenger will die, and twenty nearby pedestrians will be
unharmed. If the driverless car swerves right, there is a 99% chance that its
passenger will die, and a 100% chance that twenty nearby pedestrians will
receive lifelong debilitating injuries.
Vehicles,” 11.
31 Keeling, ”Against Leben’s Rawlsian Collision Algorithm for Autonomous
Vehicles,” 11.
32 Leben, “A Rawlsian Algorithm for Autonomous Vehicles,” 114; Keeling,
Vehicles,” 12.
34 Keeling, ”Against Leben’s Rawlsian Collision Algorithm for Autonomous
Vehicles,” 12.
Ethical Clashes in the Prospects for Autonomous Vehicles 19
That is why I would argue that the problems with the second scenario do not
derive from “the undue weight” to the moral claims of the worst-off people,
but rather from the fact that the due weight is not relevantly graduated in
moral terms. The lack of such a graduation affects the way in which the
group of the worst-off people is determined.
Scenario 3
Keeling points out that there is a scenario which includes an algorithm that
assigns a higher survival probability to the worst-off people than Leben’s
algorithm. This is the greatest equal chance algorithm.35
The AV can swerve left or swerve right. If the AV swerves left, there is a 0%
chance that Anne will survive, and a 70% chance that Bob will survive. If
the AV swerves right, there is a 1% chance that Bob will survive, and a 60%
chance that Anne will survive.
Vehicles,” 12.
36 Keeling, ”Against Leben’s Rawlsian Collision Algorithm for Autonomous
Vehicles,” 12.
20 Chapter II
that the affected parties receive the greatest equal survival probabilities”,37 the
following problem occurs. Precising the survival probability of 32.6%, which
is certainly greater than 1%, is only the first step in resolving the moral
dilemma. The precision of survival probabilities in programming the greatest
equal chances does not trigger unquestionable moral consequences for the
affected parties. For instance, computing the greatest equal chances for
survival does not shed light upon the case when Anne, who could be a mother
of three kids, or when Benn, who could be a researcher able to find a cure for
cancer, should be sacrificed.38
Vehicles,” 12.
38
Having argued that there are some collisions in which the greatest equal chances
algorithm is preferred to Leben’s algorithm” (Keeling, ”Against Leben’s Rawlsian
Collision Algorithm for Autonomous Vehicles,” 13), Keeling elaborates upon his
view saying that the greatest equal chances algorithm is “not great” either (Keeling,
“The Ethics of Automated Vehicles,” 108). The reason is that it depends upon the
ties between the affected parties, as well as assuming that each person has an equal
moral claim to be saved (Keeling, “The Ethics of Automated Vehicles,” 108). These
conditions are not satisfied in AV collisions.
39 Frank, Chrysochou, Mitkidis and Ariely, “Human Decision-Making Biases in the
optimal objectivity by applying the veil of ignorance. The idea is that the
self-preserving intentions should be reduced to a minimum so that the group
of the worst-off people can be determined in the most objective manner.
If the pedestrians in the Rawlsian collision algorithm are placed in the worst-
off group due to the objectivity of the high norm violation, does it mean that
we can deprive them of the right to survive? Furthermore, do we have the
Elaborating upon this approach, I argue that Rawls’ original position should
be modified. If the parties agree to sacrifice themselves, in case they are
norm violators, denying the survival probability by default is the only just
decision. Certainly, such a line of thought reaches a dead end not only with
respect to Rawls’ theory of justice.
On the other hand, if the parties agree to sacrifice the pedestrians because
they believe that the pedestrians are guilty, such an agreement hardly can be
called moral at all. It questions the moral design problem by compromising
the initial status of the pedestrians. The moral gist of the dilemma is who
has the moral right to decide on behalf of others for their own sake so that
one can avoid the implications of moral arbitrariness.
43Jean-François Bonnefon, Azim Shariff and Iyad Rahwan, “The Social Dilemma of
Autonomous Vehicles,” Science, No. 352 (June 2016): 1573-1576.
44 Azim Shariff, Jean-François Bonnefon and Iyad Rahwan, “Psychological Roadblocks
45 David Rand, “Cooperation, Fast and Slow: Meta-Analytic Evidence for a Theory
of Social Heuristics and Self-Interested Deliberation,” Psychological Science, 27,
No. 9 (September 2016): 1192-1206. Frank, Chrysochou, Mitkidis and Ariely,
“Human Decision-Making Biases in the Moral Dilemmas of Autonomous
Vehicles,” 1.
46 Bonnefon,Shariff and Rahwan, “The Social Dilemma of Autonomous Vehicles,”
1575.
24 Chapter II
the resources in question and secondly, that every single time when
cognitive resources are mobilized in making decisions, these decisions are
necessarily utilitarian.
The findings support the hypothesis that people employ a utilitarian moral
doctrine in a deliberate decision-making mode, while people in an intuitive
decision-making mode rely upon the more accessible deontological
doctrine.49 This means that when the participants in the thought experiment
react spontaneously, they decide to save the pedestrian. As a reason for that
Frank et al. point out the role of a culturally embedded bias, namely, that
US citizens are taught that pedestrians on public roads may not be hurt by
drivers.50 In turn, when the participants carefully think about the situation,
they prefer to sacrifice the pedestrian for the sake of the two passengers
since thus they can maximize the well-being of the majority. However, the
general trend, as demonstrated by Study 3, is that the prevalent choice of the
participants remains in favor of the single pedestrian. In this context, Frank
et al. ask the question of what degree of utility trade-off would be necessary
In Study 4, one more control variable is added, namely, that of age. Its role
is exemplified by the specification that one of the passengers is a child
sitting in the back sit. Frank et al. clarified that this variable concerns the
value-of-life heuristics, which is typical for Western cultures. The life of a
younger person is valued over that of an older person.52 Surprisingly,
participants’ decisions are almost identical to these in Study 3–there is no
significant increase in the likelihood of sacrificing the pedestrian. However,
the difference becomes obvious, when comparing the passenger and the
pedestrian perspectives. The likelihood of people’s intuitive decisions to
sacrifice the pedestrian in the passenger perspective is four times higher than
in the pedestrian perspective condition.53 The investigation shows that
people are less protective of the child in the pedestrian condition than in the
passenger and control conditions.54
In this context, I would argue that the main problem with the ethical
explanatory instrumentarium concerns firstly, the reasons behind relating
deliberate decision-making to utilitarian decisions and, consequently,
intuitive decision-making to deontological decisions. Secondly, one should
keep in mind the specification that intuitivism and deontology are mutually
exclusive in moral terms. That is why I raise a hypothesis that the biggest
complications regarding people’s biased moral preferences derive from
narrowing the role of emotions to intuitive and deontological decision-
Moral utilitarianism does not exclude the role of moral feelings. The latter
can play not only a negative role, as is in the case with hedonism, but also a
positive one. Moral feelings such as empathy and sympathy contribute to
the maximization of the collective well-being. In turn, the role of emotions
is explicitly neglected in deontological ethical projects such as Kant’s
ethics.
Judging by the results of Studies 3 and 4, I argue that that there are two
questionable points in favoring the role of deontological ethical reasons,
when one shows a high likelihood to avoid the sacrifice of the passenger.
Firstly, emotions, which make the Self identify with the passenger as other,
triggering the self-projective bias in Self’s favor, can hardly be called a
reason for a deontological ethical stance. In other words, the result seems
deontologically acceptable due to the fact that one survival probability is
ascribed a higher moral value. However, the motivation for that, which is
intuitively triggered by the self-preservation, could be egoistic, and nothing
to do with deontological ethics.
55
Frank, Chrysochou, Mitkidis and Ariely, “Human Decision-Making Biases in the
Moral Dilemmas of Autonomous Vehicles,” 7.
Ethical Clashes in the Prospects for Autonomous Vehicles 27
The simplest explanation is that the more respondents think, the more they
are inclined to think altruistically. The respondents are supposed to be more
zealous of sacrificing themselves for the sake of preserving the survival
probability for the rest. Such an explanation, however, raises some
methodological concerns. Certainly, the act of self-sacrifice is emotionally
entangled, which means that it cannot be examined as a use of more
cognitive resources.58 It concerns the participants’ self-projection ability to
put themselves into others’ shoes and feel empathy for them by evaluating
their value-of-life heuristics.
in the contractualist mode, it is used for delineating the group of the worst-
off people.59
Some of Frank et al.’s findings also support the thesis that sacrificing one
person for the sake of saving many is not necessarily a moral utilitarian
outcome. This is well-demonstrated by the increase in the number of control
variables in Study 7. The results illustrate that the possession of a driver’s
license, higher education, knowledge about AVs etc. increases the
likelihood of sacrificing the pedestrian.60 The attitudes towards sacrificing
a single person for the sake of saving many will influence the significant
increase in the likelihood of sacrificing the pedestrian.61 However, this is
not necessarily the main reason for such a sacrifice. Otherwise, it would
have meant that highly-educated people, with a high economic standard, are
more inclined to promote utilitarian moral values. Drawing such a
conclusion is problematic in many respects. For instance, the respondents
may simply identify “more easily”62 with the passengers than with the
pedestrian due to their own background of being active car users.
59 This interpretation supports Keeling’s concerns that Bonnefon et al.’s research can
solve the moral design problem by using empirical methods (Keeling, “The Ethics
of Automated Vehicles”, 35).
60 Frank, Chrysochou, Mitkidis and Ariely, “Human Decision-Making Biases in the
Conclusion
By examining the moral challenges in building AVs, I argue that the pitfalls
deriving from the utilitarian implications of the one-versus-many case gain
new strength. For the purposes of demonstrating the origin of some
problems with the design of a universal moral code for AVs, I have
examined two types of projects. The first one is a contractualist project,
namely, Rawlsian collision algorithm formulated by Leben, while the
second project is underlined by the findings of some utilitarian and
deontological AV thought experiments, as conducted by Frank et al.
Concerning the elaboration of the original position and the maximin rule, I
argue that tackling their reception is of crucial importance for understanding
the problematic aspects of Rawlsian collision algorithm. As such a
30 Chapter II
questionable issue, I point out the lack of some knowledge about value-of-
life heuristics in the original position, which is initially denied by both
Rawls and Leben. However, such heuristics affects the application of the
leximin rule. Specifically, if one does not know whether the worst-off
person prefers to die rather than suffering debilitating injuries, one cannot
make a just decision for that other. The impact of the blurred boundaries
between the values of survival probabilities and survival upon the leximin
rule shows that the latter might be no longer considered as just. That is why
I argue that elaborating upon the potential moral efficiency of Rawlsian
collision algorithm requires Rawls’ idea of the veil of ignorance to be
modified. It should address the enrichment of knowledge about the life
projects of all parties.
Having analyzed the three examples, which Keeling sets as a test for
Leben’s collision algorithm, I clarify why they face some limitations in
justifying the values of survival probabilities. Regarding the first scenario,
I point out the complexity of the logical distinction between the preferences
to survival as strict preferences and those to non-fatal injuries as weak
preferences. However, providing a logical distinction is insufficient for cases
when a rational preference to non-fatal injuries can be considered as a strict
preference in moral terms.
The third scenario, which represents Keeling’s own suggestion for the
greatest equal chance algorithm, contributes to increasing the survival rate,
but does not provide moral criteria of differentiation for cases such as that
of one-versus-one.
If one examines the impact of the pedestrian’s social norm violation upon
the personal perspective of the observer within Rawlsian collision
algorithm, there are two significant problems, at least. If the deciding parties
agree to sacrifice themselves, in case they are violators, they should initially
deny the survival probability for themselves as the only just decision.
On the other hand, if the parties agree to sacrifice the pedestrians because
they believe that the pedestrians are guilty, such an agreement cannot be
called moral at all. The moral gist of the dilemma is who has the moral right
to decide on behalf of others for their own sake so that one can avoid the
implications of moral arbitrariness.
Regardless of the fact that the potential solutions to the trolley dilemma go
beyond the utilitarian explanatory framework, utilitarian decision-making
plays a crucial role in understanding the challenges in building a universal
moral code for AVs. While tackling the implications of the decision-making
in question, one should keep in mind that it is determined as such due to the
limited methods of experimental ethics, which may encourage the
recognition of some ungrounded ethical generalizations.
The analysis of Frank et al.’s findings shows that even if there are no formal
reasons to reject the statement that the high likelihood of sacrificing the
passengers is driven by deontological ethical arguments, the emotion of self-
32 Chapter II
I draw the conclusion that the relevant search for a universal moral code for
AVs requires the reconsideration of the triplet of emotions–intuitive
decision-making–deontological decision-making and that of cognitive
knowledge–deliberate decision-making–utilitarian decision-making. Such a
clarification reveals why the sacrifice of one person for the sake of many is
not necessarily a moral utilitarian outcome, although it can meet the formal
utilitarian requirements of solving the one-versus-many case.
References
Bonnefon, Jean-François, Azim Shariff, and Iyad Rahwan. “The Social
Dilemma of Autonomous Vehicles.” Science, No. 352 (2016): 1573-
1576. https://doi. org/ 10.1126/science.aaf2654.
Foot, Philippa. “The Problem of Abortion and the Doctrine of the Double
Effect.” Oxford Review, No. 5 (1967): 5-15.
Frank, Darius-Aurel, Polymeros Chrysochou, Panagiotis Mitkidis, and Dan
Ariely. “Human Decision-Making Biases in the Moral Dilemmas of
Autonomous Vehicles.” Sci Rep, No. 9, 13080 (2019).
https://doi.org/10.1038/s41598-019-49411-7.
Goodall, Noah. “Ethical Decision Making during Automated Vehicle
Crashes.” Transp. Res. Record J., Transp. Res. Board 2424, No. 1
(2014): 58-65. https://doi.org/10.3141/2424-07.
Greene, Joshua D., Silviya A. Morelli, Kelly Lowenberg, Leigh E. Nystrom,
and Jonathan D. Cohen. “Cognitive Load Selectively Interferes with
Utilitarian Moral Judgment.” Cognition, 107, No. 3 (2008): 1144-1154.
https://doi.org/10.1016/j.cognition.2007.11.004.
Keeling, Geoff. “Commentary: Using Virtual Reality to Assess Ethical
Decisions in Road Traffic Scenarios: Applicability of Value-of-Life-
Based Models and Influences of Time Pressure.” Front. Behav.
Neurosci, 11, 247 (2017). https://doi.org/10.3389/fnbeh.2017.00247.
PMID: 29311864; PMCID: PMC5733039.
Keeling, Geoff. 2018. “Against Leben’s Rawlsian Collision Algorithm for
Autonomous Vehicles.” In Philosophy and Theory of Artificial Intelligence
2017 (Studies in Applied Philosophy, Epistemology and Rational Ethics.
44), edited by Vincent C. Müller, 259-272. (Leeds: Springer 2018).
https://doi.org/10.1007/978-3-319-96448-5_29.
Keeling, Geoff. “The Ethics of Automated Vehicles.” PhD thesis.,
University of Bristol, 2020.
Leben, Derek. “A Rawlsian Algorithm for Autonomous Vehicles.” Ethics
Inf Technol 19, No. 2 (2017): 107-115. https://doi.org/10.1007/s10676-
017-9419-3.
34 Chapter II
IVA GEORGIEVA
Introduction
The ethical consequences of the work done in the field of Artificial
Intelligence (AI) in the current pandemic world differ extensively from
those that were discussed in the previous times before the outbreak of
COVID-19. The focus on the nature of the events happening worldwide due
to the outbreak of the virus led to various ripple effects on the notions
directed toward science and technology. It also, however, provokes the
occurrence of extreme phenomena such as conspiracy theories, fake news,
anti-vaccination movements, deep fakes, and similar ones, just to name a
few. It is now more than ever important to restore the trust in science and
deploy the means of technology in service of people’s health and wellbeing
and to prove the ethical foundations in any actions taken toward serving
these goals with the available and new technological endeavors. However,
exactly at this point, science, technology, and AI in particular might seem
even more threatening as one of the strongest traumatic results of the
pandemic is the one that questions the ability of humankind to survive and
undermines the innate ability to believe that we are not so existentially
vulnerable and it is important not to surrender to ideas that we might be a
subject to extinction.
Even though technology is not directly held responsible for the current state
of affairs, the theories about the origin of the pandemic, as well as fears of
malicious use of various technological means cause increased distrust in the
36 Chapter III
for COVIDဨ19,” Dermatologic Therapy (May 2020). Tim Robbins et al., “COVID-
19: A New Digital Dawn?” Digital Health (April 2020). Abu Sufian et al., “Insights
of Artificial Intelligence to Stop Spread of Covid-19.” In Big Data Analytics and
Artificial Intelligence against COVID-19: Innovation Vision and Approach, eds.
Aboul-Ella Hassainen, Nilanjan Dey and Sally Elghamrawy. (Springer, Cham:
Springer, 2020), 177-190.
8 Babylon Health, accessed June 12, 2022, https://www.babylonhealth.com.
9 John McGreevy et al., “Clinical, Legal, and Ethical Aspects of Artificial
answers to the already discussed questions such as what is the artificial agent
that finds updated meaning in the new context of the pandemic, the issue of
heightened cyber-security threats in the midst of a worldwide situation
challenged with the sense of authenticity, should change the way these
perceptions are analyzed by specialists and accepted by society.11
Steps in the Right Direction,” Harvard Data Science Review (April 2020).
14 Paul M. Leonardi, “COVIDဨ19 and the New Technologies of Organizing: Digital
The “infodemic” that also resulted exponentially from the specifics of the
pandemic could be addressed with the means of AI. This term refers to the
widespread misinformation that circulates through social media and creates
divisions between groups of people, nations and cultures and consequently
deepens the negative effects of the pandemic. Issues related to the ethical
and social acceptability of tracking technologies are just one side of the
irrational fears about the technological aids, on the other side there is a real
during an Epidemic Like the Novel Coronavirus,” Healthcare, vol. 8, No. 2 (October
2020): 154.
18 Sara Gerke, Timo Minssen and Glenn Cohen., “Ethical and Legal Challenges of
danger of privacy breaches, online hoaxes and malicious software that might
significantly endanger the users while providing untrue information.19
Detecting these will again require technological means and possibly relying
on AI, and that is why trust becomes such a twofold phenomenon now – a
problem and a solution, and even more complex than that.20 Concerns about
the realization the such methods, which help mitigating the problems
resulting from technology require technology itself, are real and create
necessity for a separate response in relation to the current state of affairs of
ethics of technology and the increasingly leading role of AI in its application
in the context of the pandemic.
19 Harbind Kharod and Israel Simmons, “How to Fight an Infodemic: The Four
Pillars of Infodemic Management,” Journal of Medical Internet Research 22, No. 6
(2020): e21820.
20 Muhammad Mustafa Kamal, “The Triple-edged Sword of COVID-19:
IEEE Technology and Society Magazine 37, No. 4 (December 2018): 20-25.
Ethical Challenges to Artificial Intelligence in the Context 41
of Pandemic and Afterwards
developed virtual partner that serves its function, being it the care of an
elderly person who lacks contact with relatives, or a helper of kids who need
to learn from a different source than their teachers or overly busy parents in
the new era of virtual meetings and online education, seems like a refreshing
and promising alternative.23 Nevertheless, rethinking the way things function
became necessity and redesigning the means of achieving our goals is
necessary to answer this challenge and be prepared for the future.
It is without a doubt that digital health assistants and medical chatbots, e.g.
from sophisticated ones meant for visually impaired persons to offer more
freedom and better life to more common ones such as the period tracking
apps that also remind of birth control pills to women, they can all provide
significant changes in the facilitation of everyday tasks, especially when
people are overwhelmed with the response to more daunting necessities
such as prevention and treatment of a viral infection. Of course, the most
common application of such smart advisors will be the detection and
solution of various symptoms, related or not to the pandemic, that might
challenge the integrity of people’s lifestyles and threaten their health.24
Other areas of life that suffer strongly from the pandemic are more subtle,
namely the psychological changes resulting in decreased motivation and
creativity 25that can be addressed again with the help of AI and VR in areas
that are inevitably changed due to the pandemic, such as tourism for
example.26 The possibility of offering gamification in addition to the
intelligent assistants turns the interest toward serious games that again can
23 Hui Luan et al., “Challenges and Future Directions of Big Data and Artificial
Intelligence in Education,” Frontiers in Psychology 11 (October 2020).
24 Jingwen Zhang et al., “Artificial Intelligence Chatbot Behavior Change Model for
connect AI and VR and offer a new platform for treating patients with
different kinds of suffering (both purely physical and subtly psychological) and
serve as a training tool for education, prevention and building of resilience
in the context of the increasingly challenging contemporary life.27
From IoT solutions29 that are needed more than ever in the everyday life
now to other specific for this time issues such as the analysis of anti-
vaccination movements,30 the detection of misinformation and panic-
induced/inducing information online,31 as well as more global problems
2021).
29 Musa Ndiaye et al., “IoT in the Wake of COVID-19: A Survey on Contributions,
on Facebook and Twitter toward COVID-19 Vaccines in the United Kingdom and
the United States: Observational Study,” Journal of Medical Internet Research 23,
No. 4 (April 2021): e26627.
31 Jim Samuel et al., “Covid-19 Public Sentiment Insights and Machine Learning for
such as the “infodemic” turning into a “digital pandemic”,32 these tasks can
all be handled by AI. It is a matter of collective responsibility33 to also
realize that there are the previously mentioned here long-term effects and
the general double-edged sword effect of technology, along with more
subtle ones that actually affect populations on a very deep level, e.g., the
problem with digital inequalities 34that accumulate and create a pandemic
of separate pandemic-related issues.
Intelligence.” In Imagery and Visual Literacy: Selected Readings from the Annual
Conference of the International Visual Literacy Association (Tempre, Arizona,
October 12–16, 1994).
41 Rana Saeed Al-Maroof et al., “Fear from COVID-19 and Technology Adoption:
Conclusion
The aftermath of the pandemic is still unpredictable, but the new world after
it will and at the same time will not be so globalized or that much depending
on technology. Divisions and contradictions become stronger, however the
need for smarter solutions also emerges above previous needs. It is a matter
of devoted scientific work to determine which is important to stay and hold
and which is no longer serving, but it is undoubtedly sure that the
preservation of what is humane and what is ensuring the wellbeing of
humankind remains the greater goal of them all.
of the Line of Life and Its Connection to a Healthier Self, “ Behavioral Sciences 9,
no. 11 (November: 2019): 111.
46 Chapter III
References
Al-Maroof, Rana Saeed, Said A Salloum, Aboul Ella Hassanien, and Khaled
Shaalan. “Fear from COVID-19 and Technology Adoption: The Impact
of Google Meet during Coronavirus Pandemic.” Interactive Learning
Environments (2020): 1-16.
https://www.tandfonline.com/doi/full/10.1080/10494820.2020.1830121.
Arnold, Mark, and Ian Kerridge. “Accelerating the De-personalization of
Medicine: The Ethical Toxicities of COVID-19.” Journal of Bioethical
Inquiry 17, No. 4 (2020): 815-821.
https://link.springer.com/article/10.1007/s11673-020-10026-7.
Babylon Health. Accessed June 12, 2022. https://www.babylonhealth.com.
Bailenson, Jeremy N. “Nonverbal Overload: A Theoretical Argument for
the Causes of Zoom Fatigue.” Technology, Mind, and Behavior” 2, No.
1 (2021). https://tmb.apaopen.org/pub/nonverbal-overload/release/2
Battineni, Gopi, Nalini Chintalapudi, and Francesco Amenta. “AI Chatbot
Design during an Epidemic Like the Novel Coronavirus.” Healthcare,
vol. 8, No. 2 (2020): 154. https://doi.org/10.3390/healthcare8020154.
Bhaskar, Sonu, Sian Bradley, Sateesh Sakhamuri, Sebastian Moguilner,
Vijay Kumar Chattu, Shawna Pandya, Starr Schroeder et al. “Designing
Futuristic Telemedicine Using Artificial Intelligence and Robotics in the
COVID-19 Era.” Frontiers in Public Health 8 (2020): 708.
https://org.doi/10.3389/fpubh.2020.556789.
Bragazzi, Nicola Luigi, Haijiang Dai, Giovanni Damiani, Masoud
Behzadifar, Mariano Martini, and Jianhong Wu. “How Big Data and
Artificial Intelligence Can Help Better Manage the COVID-19
Pandemic.” International Journal of Environmental Research and
Public Health 17, No. 9 (2020): 3176.
https://pubmed.ncbi.nlm.nih.gov/32370204/.
Cave, Stephen, Jess Whittlestone, Rune Nyrup, and Rafael A. Calvo. “Using
AI Ethically to Tackle Covid-19.” BMJ 372 (2021).
https:/doi:org/https://doi.org/10.1136/bmj.n364.
ûosiü, Krešimir, Siniša Popoviü, Marko Šarlija, and Ivan Kesedžiü. “Impact
of Human Disasters and COVID-19 Pandemic on Mental Health:
Potential of Digital Psychiatry.” Psychiatria Danubina 32, No. 1 (2020):
25-31. https://pubmed.ncbi.nlm.nih.gov/32303026/
Ethical Challenges to Artificial Intelligence in the Context 47
of Pandemic and Afterwards
Luan, Hui, Peter Geczy, Hollis Lai, Janice Gobert, Stephen J.H. Yang,
Hiroaki Ogata, Jacky Baltes et al. “Challenges and Future Directions of
Big Data and Artificial Intelligence in Education.” Frontiers in
Psychology 11 (2020). https://pubmed.ncbi.nlm.nih.gov/33192896/.
Luck, Michael, and Ruth Aylett. “Applying Artificial Intelligence to Virtual
Reality: Intelligent Virtual Environments.” Applied Artificial
Intelligence 14, No. 1 (2000): 3-32.
https://doi.org/10.1080/088395100117142.
Marabelli, Marco, Emmanuelle Vaast, and Lydia Li. “Preventing the Digital
Scars of COVID.” European Journal of Information Systems (2021).
https://dx.doi.org/10.2139/ssrn.3755050.
McGreevey, John D., C. William Hanson, and Ross Koppel. “Clinical,
Legal, and Ethical Aspects of Artificial Intelligence–assisted
Conversational Agents in Health Care.” Jama 324, No. 6 (2020): 552-
553. https://pubmed.ncbi.nlm.nih.gov/32706386/.
McLellan, Hilary. 1995. “Magical Stories: Blending Virtual Reality and
Artificial Intelligence.” In Imagery and Visual Literacy: Selected
Readings from the Annual Conference of the International Visual
Literacy Association (Tempre, Arizona, October 12–16, 1994).
Miller, Seumas, and Marcus Smith. “Ethics, Public Health and Technology
Responses to COVIDဨ19.” Bioethics Vo. 35, No. 4 (2021): 366-371.
https://doi.org/10.1111/bioe.12856.
Muthuppalaniappan, Menaka, and Kerrie Stevenson. “Healthcare Cyber-
Attacks and the COVID-19 Pandemic: An Urgent Threat to Global
Health.” International Journal for Quality in Health Care” 33, No. 1
(2021). https://pubmed.ncbi.nlm.nih.gov/33351134/.
Ndiaye, Musa, Stephen S Oyewobi, Adnan M. Abu-Mahfouz, Gerhard P.
Hancke, Anish M, Kurien, and Karim Djouani. “IoT in the Wake of
COVID-19: A Survey on Contributions, Challenges and Evolution.”
IEEE Access 8 (2020): 186821-186839.
https://ieeexplore.ieee.org/document/9220109.
Petroviü, Vladimir M. “Artificial Intelligence and Virtual Worlds–Toward
Human-level AI Agents.” IEEE Access 6 (2018): 39976-39988.
https://ieeexplore.ieee.org/document/8410872.
Puaschunder, Julia M. “The Future of Artificial Intelligence in International
Healthcare: An Index.” In Proceedings of the 17th International RAIS
50 Chapter III
ANTON GERUNOV
Introduction
The increased automation of human activity essentially means that a
plethora of activities that used to be performed by sapient and conscious
human beings are now gradually transferred to machines, or more precisely,
to artificial intelligence agents. The most visible examples of such automation
are phenomena like self-driving cars and autonomous weapons systems. It
is, therefore, little surprise that the ethical issues connected to those specific
applications are gaining increasing prominence. Those concerns are clearly
exemplified in the journal Nature’s specific ethics of AI invited
commentaries. There Stuart Russell1 strongly warns of the Lethal
Autonomous Weapons Systems (LAWS), while Veloso2 urges to find
complementarities and synergies between humans and AI.
1 Stuart Russell, “Take a Stand on AI Weapons,” Nature, 521 (7553) (May 2015):
415-416.
2 Veloso, Manuela, “Embrace a Robot-Human World,” Nature, 521 (7553) (May
2015): 416-418.
The AI-Run Economy 53
quite obvious at this point3 but the key question remains as to what would
be the implications of marginalizing or even almost completely eliminating
human actors from a majority of economic choices. Apart from the obvious
individual consequences, such a move is likely to generate systemic
consequences that can be aptly explored via means of comprehensive
simulations. Our research aims to outline such an approach by constructing
two very simple model economies–a human-run and an AI-run one and
comparing and contrasting their performance. Based on this simulation
experiment, we further analyze some ethical issues that will arise as
business automation and AI-driven decision-making become ubiquitous.
Literature Review
The increasing abilities to process potentially sensitive information together
with the explosion of data availability and the rapidly decreasing costs of
computing have enabled unprecedented mass data processing with
potentially large implications for humans. Such data processing – either on
its own, or as enabling AI creates a plethora of ethical issues that have
increased in complexity over the past decades. In the era before autonomous
decision-making by machines (so-called domain AI) ethical issues revolved
mainly around data and included issues such as the privacy, accuracy,
property, and accessibility of information.4 A key issue here was how
personal information is protected in such a way so as not to cause harm to
individuals, and what its overall governance (and thus power) structure is.
This early thinking on data ethics presupposes human agency in the use of
information and so the key question is who possesses the data and has the
right to profit from it. A natural early question of data ethics is about the
locus of ownership and the power asymmetries surrounding it. This is then
still a human-centric version of ethics.
3 Hsinchun Chen, Roger H. Chiang and Veda C. Storey, “Business Intelligence and
Analytics: From Big Data to Big Impact,” MIS Quarterly, 36(4) (December 2012):
1165-1188.
4 Mason, Richard O. “Four Ethical Issues of the Information Age.” In Computer
A note of clarification is in order here. When using the term “AI” the usual
connotation entails domain AI, where a decision-making algorithm can
perform domain-specific activities, replacing or augmenting the human
agent. The broader term “Artificial General Intelligence”, or AGI, has come
to describe the “real” all-encompassing multi-domain intelligence that will
make machines both sentient and sapient.6 The former is already a fact, the
latter is in the research phase and will likely need several decades before it
is fully developed.
5 Joanna Bryson and Alan F.T. Winfield, “Standardizing Ethical Design for Artificial
Research in this domain continues activity.9 Another major spur for the
ethics discussions was the development and introduction of self-driving cars
and their potential deleterious impact on human beings.10 In this context
Etzioni and Etzioni11 underline that there are two major ways to introduce
ethical aspects to AI-driven decision-makers: top-down and bottom-up.
Top-down approaches include “teaching” a system of ethics to the machine,
while bottom-up approaches involve learning from a large number of actual
human decisions in such a way that the AI can reconstruct a system of ethics.
Etzioni and Etzioni12 argue that both approaches are unfeasible and probably
unnecessary as the locus of ethics needs to remain in humans either through
their preferences or through legislated norms.
Wrong (Oxford: Oxford University Press, 2008). Patrick Lin, Keith Abney and
George A. Bekey, eds. Robot Ethics. The Ethical and Social Implications of Robotics
(Cambridge, Massachusetts. London, England: The MIT Press, 2012).
10 Jean-François Bonnefon, Azim Shariff and Iyad Rahwan, “The Social Dilemma
Torresen15 underlines the important point that even under strict rules for AI,
some resultant harm may be difficult to predict or avoid, thus further
Simulation Results
Two simulations are run to better compare and contrast two types of
economies. The former is a realistic human-driven economy, where a large
number of heterogeneous human agents with bounded rationality make
sometimes suboptimal decisions. Those agents are modeled in a
behaviorally realistic fashion by drawing major insights from the fruitful
research in the field of behavioral economics. While this economy does
exhibit a tendency towards full resource utilization, it is hampered by
imperfections in markets and individual decisions (inertia, habit formation,
imperfect markets, suboptimal decisions, etc.)
16 Jordi Gali, Monetary Policy, Inflation and the Business Cycle (Princeton:
Princeton University Press, 2008). Carl Walsh, Monetary Theory and Policy (US:
MIT Press, 2003).
58 Chapter IV
To close the system, we postulate a process for the interest rates, following
a standard Taylor rule.17 Essentially, this rule states that the Central Bank
sets the economy-wide interest rates taking reference to the current prices
(ʌt) in order to meet its inflation targets, to the current input (yt) in order to
avoid excessive harm to the real economy, and is bound by previous interest
(rt-1) rates both through decisions and through efforts to implement interest
rate smoothing. This equation has no clear behavioral component and can
be implemented by both human actors (as it is currently the case) or by
algorithms (in a possible future development):
Despite being quite simple, those three key equations present the mainstay
of the modeled economy and provide for relatively realistic overall
dynamics.
17
John B. Taylor, Macroeconomic Policy in a World Economy: From Econometric
Design to Practical Operation (New York: W.W. Norton, 1993).
The AI-Run Economy 59
This system of equations (4) through (9) aims to realistically model uniquely
human behavior and allow us to see how it influences observed economic
dynamics. This is done in exactly the same fashion for inflation as it is
shown for output in eq. (4)-(9). In short, here we model economic
expectations and decisions as dependent on irrational human impulses and
show how this irrationality spills into economy-wide inertia and sub-
optimality.
19Daniel Kahneman, Thinking, Fast and Slow (New York: Farrar, Straus and
Giroux, 2011).
The AI-Run Economy 61
We hypothesize that the AI is not bound by irrational habits and will have
no incentive to switch away from the optimal decision-making. The fact that
we are dealing with a single unified decision-maker also means that there is
no need for an aggregation mechanism for disparate individual expectations.
Due to the long-term trend of the economy to drift toward its potential
production, the rational expectations for output and inflation would be set
to zero plus a random shock due to unforeseen circumstances.
Simulation Results
To simulate the two model economies, we use standard numeric values for
the parameters as per the literature.20 Those are presented in Table IV.1 and
can be shown to generate realistic economic dynamics.21 Each economy is
then simulated over 10,000 periods and its overall statistical properties are
shown and investigated. Detailed results for inflation and output are
presented in Table IV.2.
Both simulated economies are designed in such a way that they tend to
converge to full utilization of production factors (or potential output). Thus,
it is not surprising that both of them have a mean output gap that is not
significantly different from zero. This is also true for the corresponding
20
Gali, Monetary Policy, Inflation and the Business Cycle, 58.
21Anton Gerunov, Macroeconomic Modeling: Modern Approaches (Sofia: Sofia
University Press, 2015).
62 Chapter IV
aggregate of prices. Over the long run, both human and AI-run economies
tend to stabilize in the absence of large external shock.
However, the key factor to note here is that the AI economy produces very
different dynamics in this process. The human-run economy has both
pronounced peaks and deep recessions in growth, with a total range over
15% between the highest and the lowest points. This is the typical economic
cycle of boom and bust. In contrast, the AI-run economy has an extremely
limited range of the gap–only 5.1%. This is also reflected in the standard
deviations–the human-run economy features standard deviations for output
that are three times as large as those generated by the simulated rational
decision-maker. The proportions are similar when it comes to the price
dynamics. The range of the human-run economy is about 2.5 times larger
than that for the AI one, and the standard deviations are two times greater.
Those results can be easily followed in Figure IV.1.
In short, the AI-run economy features much lower volatility, greater short-
run stability, and has virtually eliminated violent economic crises and
possible hyperinflation risks. We will use those economic simulations as
starting points for investigating possible ethical issues in case such a
transformation takes place.
Ethical Issues
Leveraging the Boström and Yudkowski’s framework,22 we investigate the
ethical aspects of transforming the economy from a decentralized boundedly
rational system, run by humans into a centralized rational system, run by
AI. The following aspects are evaluated toward this end: responsibility,
transparency, auditability, incorruptibility, predictability, avoidance of harm.
The responsibility of the system refers to its ability to make fair and
responsible decisions in pursuit of a beneficial goal. Depending on how the
AI is implemented, we may observe different facets. In case the AI decision-
maker is programmed top-down with some pre-defined objective functions
that are to be optimized, the AI can only be expected to take limited
responsibility for its narrowly defined objectives. If the AI is instead trained
in a bottom-up fashion, it should have a holistic understanding of its
responsibility. At any rate, the ultimate rationale for the AI decisions and
actions needs to be clear and will likely converge to some type of
maximization of individual and group welfare. The end deliberation of
whether the AI is responsible may need to be taken after extensive analysis
is conducted.
The auditability of the AI is yet another issue at hand. With the automation
of decisions and under current best practices and extant trends for logging,
this seems to be a relatively minor issue. It is highly likely that a large part
(if not all) of the actions and decisions of the AI are to be meticulously
documented and available for further investigation. A possible concern here
may be tampering with those logs, but some emerging technologies such as
the blockchain can prevent that as well. Overall, auditability seems to be of
limited concern.
The AI-Run Economy 65
Discussion
The two simulated economies presented here serve as useful vantage points
for applying a set of ethical criteria toward a hypothetical AI economic
decision-maker. It seems that algorithmic responsibility and transparency
will crucially depend on the specific implementation choices. In the
example presented, both conditions were satisfied as responsibility was
ensured through clear optimization behavior, while transparency stemmed
from a nearly defined set of equations used by the AI. In a more realistic
setting that leverages complex machine learning algorithms such as deep
neural networks, those two properties are far from guaranteed. The
auditability property of AI seems to be the least concern of all since
extensive logging allows for a clear audit trail enabling ex-post forensics (if
needed).
algorithm is both transparent and auditable, then the human agent will find
it much easier to ascertain whether the AI has been corrupted or is
malfunctioning. Predictability also relies heavily on transparency, but it
seems that outcomes will be easier to predict while the process used to reach
them will not be. This is particularly true if the AI is trained using
sophisticated black-box methods. Finally, the possibility of inflicting
unforeseen and possibly undetectable damage is quite high in the case of
economic decision-making by AI. Given the high complexity of the
environment and since there are no discernible human losses, a possible
suboptimal, malicious, or malfunctioning AI may remain undetected for an
extended period of time. This holds particularly true for intangible damages,
such as incentive misalignment and decreasing innovative and creative
potential.
It is clear that while some of the ethical requirements align neatly with
economic imperatives (e.g. responsibility and effectiveness), others diverge
quite dramatically (e.g. transparency and efficiency). Thus, the automation
of economic and business decision-making will pose substantive moral and
ethical questions and will necessitate significant tradeoffs between the
desirable properties of the AI. We venture to propose that in order to achieve
an optimal balance between objective constraints, large-scale and high-risk
Artificial Intelligence algorithms need to go through a formal assessment
and approval procedure. This may be in the form of an AI Impact
Assessment that is conducted to answer the complex and interconnected
ethical questions posed by complex AI algorithms. This Impact Assessment
needs to balance the economic and business needs of producers and
consumers against the safety and ethical concerns of broader society. While
such an assessment does not need to be conducted by a state authority and
can be relegated to alternative providers, its results may have to achieve a
minimum amount of consensus among stakeholders before the evaluated AI
is put into production. Such a rigorous multi-stakeholder approach will
enable humankind to ensure that AI will end up creating more benefits than
inflicting unsuspected harm.
68 Chapter IV
Conclusion
This paper focused its attention on an area of AI decision-making that is still
not substantially researched–the ethical issues stemming from the
automation of economic and business decisions. To this end, we have
presented and contrasted two model economies and shown how an AI-
driven economy dramatically differs from the current one. Its new properties
raise a set of ethical questions that were formally addressed by using the
framework topic, as summarized by Boström and Yudkowski.24 While some
of the issues necessitate concrete implementation to be fully resolved, we
reached some preliminary conclusion that outline that a number of difficult
tradeoffs may need to be made when AIs are comprehensively put into
production and service. Our proposal is to define a rigorous assessment
process with large-scale stakeholder involvement that ensures the beneficial
utilization of the upcoming Artificial Intelligence algorithms.
References
Bonnefon, Jean-François, Azim Shariff, and Iyad Rahwan. “The Social
Dilemma of Autonomous Vehicles.” Science, No. 352 (2016): 1573-
1576. https://doi. org/ 10.1126/science.aaf2654.
Boström, Nick, and Eliezer Yudkowsky. “The Ethics of Artificial
Intelligence.” In The Cambridge Handbook of Artificial Intelligence,
edited by Keith Frankish and William M. Ramsey, 316-334. (New York:
Cambridge University Press, 2014).
Bryson, Joanna, and Alan F.T. Winfield. “Standardizing Ethical Design for
Artificial Intelligence and Autonomous Systems.” Computer, 50 (5)
(2017): 116-119. http://doi.org/10.1109/MC.2017.154.
Chen, Hsinchun, Roger H. Chiang, and Veda C. Storey. “Business
Intelligence and Analytics: From Big Data to Big Impact.” MIS
Quarterly, 36(4) (2012): 1165-1188.
Etzioni, Amitai, and Oren Etzioni. “Incorporating Ethics into Artificial
Intelligence.” The Journal of Ethics, 21(4) (2017): 403-418.
https://doi.org/10.1007/s10892-017-9252-2.
Gali, Jordi. 2008. Monetary Policy, Inflation and the Business Cycle.
Princeton: Princeton University Press.
Gerunov, Anton. 2015. Macroeconomic Modeling: Modern Approaches.
Sofia: Sofia University Press.
Gerunov, Anton. “Experimental Study of Economic Expectations.” In
Yearbook of St Kliment Ohridski University of Sofia, 14(1), 103-130
(Sofia: Sofia University Press, Faculty of Economics and Business
Administration, 2016).
Kahneman, Daniel. 2011. Thinking, Fast and Slow. New York: Farrar,
Straus and Giroux.
Lin, Patrick, Keith Abney, and George A. Bekey, eds. 2012. Robot Ethics.
The Ethical and Social Implications of Robotics. Cambridge,
Massachusetts. London, England: The MIT Press.
Mason, Richard O. “Four Ethical Issues of the Information Age.” In
Computer Ethics, edited by John Weckert, 41-48. (London: Routledge,
2017).
Rogozea, Liliana. “Towards Ethical Aspects on Artificial Intelligence.” In
Proceedings of the 8th WSEAS International Conference on Artificial
Intelligence, Knowledge Engineering and Data Bases, 507-512. (World
Scientific and Engineering Academy and Society (WSEAS), 2009).
Russell, Stuart. “Take a Stand on AI Weapons.” Nature, 521 (7553) (2015):
415-416.
Sidhu, Ikhlaq, and Paris Deletraz. “Effect of Comfort Zone on
Entrepreneurship Potential, Innovation Culture, and Career Satisfaction.”
In 122nd ASEE Annual Conference and Exposition. Making Value of
Society, 1770-1783 (Seattle, WA, June 14-17, 2015).
Taylor, John B. 1993. Macroeconomic Policy in a World Economy: From
Econometric Design to Practical Operation. New York: W.W. Norton.
Torresen, Jim. “A Review of Future and Ethical Perspectives of Robotics
and AI.” Frontiers in Robotics and AI, 4, 75 (2018).
https://doi/org/10.3389/frobt.2017.00075.
Veloso, Manuela. “Embrace a Robot-Human World.” Nature, 521 (7553)
(2015): 416-418.
70 Chapter IV
SCRATCH MY BACK
& I WILL SCRATCH YOURS:
PUTTING GAME THEORY AND ARTIFICIAL
INTELLIGENCE ON A CONVERGING PATH
BEYOND COMPUTATIONAL CAPABILITIES
BORIS GUROV
It is very tempting to declare that game theory was once a very promising
research field, which reached its analytical limits and provided some good
insight into decision making under uncertainty, and gave some theoretical
awareness of the complex links between autonomous and free individual
decision-making and the harmonious functioning of the collective, or the
group. The problem with such a position lies in the fact that game theory is
the sole analytical tool available to us for bridging the gap between
72 Chapter V
What is the best possible approach for a fleet of military ships to engage an
enemy fleet depending on its goals (destroy the enemy, inflict heavy
damage, hold them off, limit its own losses)?1 How do ships coordinate
themselves and how do they cooperate with each other? Is it better that there
is a centralized top-down decision-making mechanism in place, or the
overall performance will gain from different ships in the fleet benefitting
from some form of autonomy?
1 The history of Naval warfare testifies for the very heterogeneous goals of fleet
commanders in different contexts. Of course, classical game theory is unable to
explore the interdependency of such heterogeneous logics. Generally, it will assume
the goal to be full-scale maximum damage to the opponent type of interaction.
Scratch My Back & I Will Scratch Yours 73
potential benefits are obvious. Last, but not least, exploring game-
theoretical models via AI has the potential, if it goes beyond the usual
omission of theoretical discussions, characteristic of the field (a tendency,
which was rather present even in the 1980s and 1990s in the domain of
evolutionary GT), to reconceptualize our understanding and analytical
procedures of rationality. Bluntly said, engineers should start consulting
more often the literature of economists.
2 Emile Borel,”La théorie des jeux et les équations à noyau symétrique gauche.” In
Comptes rendus de l’Académie des sciences, Vol. 173 (Paris: Académie des
Sciences Morales et Politiques), 1304–1308.
3 There are two separate translations of the original text from German (“Zur Theorie
for GT of the impact that Von Neuman had on its research program by subordinating
74 Chapter V
axiomatically the rationality of players to the solution of the game, which in turn
becomes defining for the definition of the rationality of players. Cf. Schmidt,
Christian “From the ‘Standards of Behaving’ to the ‘Theory of Social Situations’. A
Contribution of Game Theory to the Understanding of Institutions.” In Knowledge,
Social Institutions, and the Division of Labor, eds. Pier Luigi Porta, Roberto
Scazzieri and Andrew Skinner (Cheltenham: Edward Elgar Publishing, 2001), 153-
168.
Scratch My Back & I Will Scratch Yours 75
5 It is the Rand Corporation who decided to finance the pioneers Flood and Dresher
to test “Nash solutions” for non-cooperative games to support the US nuclear policy
and doctrine. Cf Marvin Flood, Some Experimental Games. Research Memorandum
RM–789 (Santa Monica, Cal.: Rand Corporation, 1952). Flood and Dresher are the
first who have formalized the PD. Often it is the name of Albert Tucker, which is
mistakenly associated with the primacy as to putting the PD in the centre of interest
of GT. It is somewhat ironic that Tucker, who is a mathematician and who was the
Ph.D. advisor of non-other than John Nash (popularized by actor Russell Crowe in
the biographical drama film A Beautiful Mind) is credited with putting together the
dramatic literary narrative that accompanies the model.
6
Rapoport, Anatol. “Prisoner’s Dilemma-Recollections and Observations.” In Game
Theory as a Theory of Conflict Resolution, ed. Anatol Rapoport (Dordrecht: D.
Reidel Publishing Company, 1974), 17-34.
76 Chapter V
Prize winner in 2015) represents the outcome of the game in which the strategy of
each player is the best possible response to the strategy chosen by the other player.
It is also called “no-regret situation” given the fact that after the game is played no
ulterior modification of any player’s strategy could provide him/her with a better
payoff.
9 In contemporary economics a situation is Pareto-optimal if there is no other social
Press, 1960).
Scratch My Back & I Will Scratch Yours 77
Publishing, 1969).
15 Cyril Hedoin, “A Framework for Community-based Salience: Common Knowledge,
18When doing the preliminary research for this far-fetched idea I stumbled onto a
piece by Elena Nisioty, “In need of evolution: game theory and AI” MAY 12,
2018/#MACHINE LEARNING [“In Need of Evolution: Game Theory and AI”,
accessed June 1, 2021, https://www.freecodecamp.org/news/game-theory-and-ai-
where-it-all-started-and-where-it-should-all-stop-82f7bd53a3b4/ “MAY 12, 2018/
#MACHINE LEARNING], which was a factor to pursue in such direction and where
very similar ideas, concerning the opportunities for the development of supra-
individual neural networks are put to the front.
Scratch My Back & I Will Scratch Yours 79
19 As Vilfredo Pareto put it already in 1900: “From the point of view of choices, the
Homo oeconomicus becomes a machine to make these choices, which given the
circumstance, always makes the same choice”, cited by Bruni and Guala. Cf. Bruni,
Luigino and Francesco Gualo. “Pareto’s Theory of Choice from the Cours to the
Traittato. Utility, Idealization and Concrete Deductive Method.” In Pareto
aujourd’hui, coll. Sociologies, ed. Bouvier Alban. (Paris: P.U.F., 1999), 118.
20 Except from experimental psychologists who used GT models (mostly PD) and
22 John M. Smith and George R. Price, “The Logic of Animal Conflict,” Nature 246
(1973): 15–18.
Scratch My Back & I Will Scratch Yours 81
23 The economic agent must execute the optimal choice in any theoretically
Value of Changing Utility,” The American Economic Review, Vol. 74, No. 1 (March
1974): 30-42. John H. Holland, Adaptation in Natural and Artificial Systems
(Michigan: The University of Michigan Press, 1975). Richard H. Day and Theodore
Groves, eds. Adaptive Economic Models (New York: Academic Press, 1975).
Patrick D. Nolan, “External Selection and Adaptive Change: Alternative Models of
Sociocultural Evolution,” Sociological Theory, Vol. 2, (1984): 117-139. Robert
Axelrod, The Complexity of Cooperation: Agent-Based Models of Competition and
Collaboration (Princeton, N.J.: Princeton University Press, 1997).
25 Richard M. Cyert and Morris DeGroot, “Bayesian Analysis and Duopoly Theory,”
The Journal of Political Economy Vol. 78, No. 5 (September–October 1970): 1168-
1184. Richard M. Cyert and Morris DeGroot, “An Analysis of Cooperation and
Learning in a Duopoly Context,” The American Economic Review, Vol. 63, No. 1
(March 1973): 24-37. Richard M. Cyert and Morris DeGroot, “Rational
Expectations and Bayesian Analysis,” The Journal of Political Economy, Vol. 82,
No. 3 (May–June 1974): 521-536. Robert Denhardt and Philip Jeffers, “Social
Learning and Economic Behavior: The Process of Economic Socialization,” The
American Journal of Economics and Sociology, Vol. 30, No 2 (April: 1971): 113-
125. Day and Groves, Adaptive Economic Models. John Harsanyi, “Bayesian Theory
and Utilitarian Ethics,” The American Economic Review. Papers and Proceedings
of the Ninetieth Meeting of the American Economic Association, Vol. 68, No. 2 (May
1978): 223-228.
26
The GA, very roughly, is functioning by safeguarding interactional information
that procured relative success to the strategy and discarding information that didn’t
for determining better strategic behavior in future interactions (Cf. Holland,
Adaptation in Natural and Artificial Systems. John H. Holland, “Adaptive
Algorithms for Discovering and Using General Patterns in Growing Knowledge
Bases,” International Journal of Policy Analysis and Information Systems, No. 4
(1980): 245-268. John H. Holland, “Genetic Algorithms,” Scientific American, 267,
No. 7 (July 1992): 44-50. David. E. Goldberg, Genetic Algorithms in Search,
Optimization, and Machine Learning (Reading, Mass.: Addison-Wesley Pub. Co.,
1989). Rick L. Riolo, “The Amateur Scientist: Survival of the Fittest Bits,” Scientific
84 Chapter V
The EGT even produced a best-seller and a star author. Robert Axelrod’s
The Evolution of Cooperation (1983)27 became one of the most read,
translated (20 + languages), and quoted scientific books. It is an excellent
starting point for everybody interested in the problemɚtic. It showed in a
rather convincing manner that from the moment the interaction in a PD has
a future (iterated games) and there is room for a decision-making autonomy
outside of the box (strategical heterogeneity), mutual general defection is no
longer a pole of universal attraction. Nevertheless, a warning regarding
Axelrod’s conclusions is of order. The hasty supposition of Axelrod that the
simple reciprocal imitation of the strategy of the other player (the TIT-FOR-
TAT strategy)28 is the best possible strategy, and that, gradually it becomes
an evolutionary stable outcome on the level of the population (that the
population will evolve towards an all TFT composition and no one will have
the incentive to deviate from that collective equilibrium) is flawed. Kenneth
Binmore, one of the leading game-theorists during the last 40 years, went
so far as to devote a whole chapter to understanding the impact of Axelrod’s
work. He labeled the whole interest generated by the work of Axelrod and
the hasty acceptance of his conclusions the “TIT FOR TAT bubble”.29
Binmore is critical of the impact of Axelrod’s work and especially of his
method of merging easily the borders between classical GT, CAS and quasi-
experimentation, theory construction, and the empiric reality. The term is
also an important reminder of the effects that taking methodological
shortcuts has in the domain of the use of AI: by accepting Axelrod’s
conclusions as firm theoretical foundations and proven decision-making
patterns number of researchers in the field started integrating his theory of
cooperative behavior in their own experiments. This conclusion is valid
even to this day and in our understanding, is one of the determining factors
for grounding the use of AI in the optic of individual learning/adaptation
and the definition of the “work” of agents as understanding, mapping, and
American, Vol. 267 (1) (1992): 114-117. This process is coupled with the
maintenance of a general strategical line, which is rectified via the input of new
interactional information.
27 Robert Axelrod, The Evolution of Cooperation (New York: Basic Books, 1984).
28 Cooperate when the other party cooperates, defect to defection and never start an
interaction by a defection.
29 Kenneth Binmore, Playing Fair: Game Theory and the Social Contract, Vol. 1
30 Peter Stone and Manuela Veloso, “Multiagent Systems: A Survey from a Machine
Learning Perspective,” Autonomous Robots, 8(3) (2000): 345–383.
31 Littman, Michael. “Markov Games as a Framework for Multi-agent Reinforcement
Dilemma Using Conditional Joint Action Learning,” Autonomous Agents and Multi-
Agent Systems 15(1) (2007): 91–108.
34 Konstantinos Giannakis et al., “Quantum Conditional Strategies and Automata for
Prisoners’Dilemmata under the EWL Scheme,” Appl. Sci. 9(13) (2019): 2635.
86 Chapter V
35 Kenneth J. Arrow, Social Choice and Individual Values (New York: Wiley, 1951).
36
Radiyna B. Braithwaite, Theory of Games as a Tool for the Moral Philosopher
(Cambridge: Cambridge University Press, 1955).
37 Garry W. Runciman and Amartya K. Sen, “Games, Justice and the General Will,”
Mind, New Series, Vol. 74, No. 296 (October 1965): 554-562.
38 Amartya K. Sen, Ethique et Economie (Paris: Presses Universitaires de France,
2001).
39 John. Rawls, “Justice as Fairness,” Philosophical Review, Vol. 67 (April 1958):
As you might already have suspected the author of this article is not an
expert in AI (despite a Ph.D. involving CAS), but rather in decision-making.
The present text is an effort to persuade specialists in AI to consider the
possibility that even though GT proved to be an extremely fruitful lab field
88, No. 1, (1979): 3-38. David Gauthier, Morals by Agreement (Oxford: Clarendon,
1986).
42 Gregory Kavka, Hobbesian Moral and Political Theory (Princeton: Princeton
France, 1977).
88 Chapter V
References
Arrow, Kenneth J. 1951. Social Choice and Individual Values. New York:
Wiley.
Axelrod, Robert. 1984. The Evolution of Cooperation. New York: Basic
Books.
Axelrod, Robert. 1997. The Complexity of Cooperation: Agent-Based
Models of Competition and Collaboration. Princeton, N.J.: Princeton
University Press.
Axelrod, Robert, and Michael D. Cohen. “Coping with Complexity: The
Adaptive Value of Changing Utility.” The American Economic Review,
Vol. 74, No. 1 (1974): 30-42.
TODOR DIMITROV
Introduction
Today, geopolitical competition is global and the race of technological
adoption is extremely important. One of the main issues is the acceptance,
integration and use of new technology in society. From Artificial
Intelligence to quantum and everything in between, governments are in a
race to leverage these technologies at scale and speed. The first adopter
advantage for emerging disruptive technology could not be more prevalent
in the world of geopolitics and deterrence. It is quite possible that the nations
that win this race may be those with the most flexible bureaucracy rather
than those with the best technology.
The topic of AI has been widely discussed in the last few decades, but the
world has evolved, and nowadays there is a rapid technological
development in all areas of our society. Some of them constantly involve
new applications in the military and security affairs developing different
systems, based on AI. It is quite important to pay attention to the issues of
implementation of that new technology for defense capabilities because they
could impact the entire future world. The emerging and disrupting
technologies are expected to have a tremendous impact on the military
affairs and not exclusively for creating killing robots. These technologies
even do not have to be used only in lethal systems. According to the
specialists the main issues for defense sector that AI technologies could
improve are management, operational speed, precision, logistic, procurement,
command and control, communications, intelligence, situational awareness,
electronic warfare dominance, cybersecurity, sensors, integration and
interoperability, different types of analysis and voice recognition algorithms,
military medicine, etc. Also, it is critical that other factors such as ethics,
legality, privacy and human control be considered.
October 2, 2021,
https://www.nationaldefensemagazine.org/articles/2021/2/10/federal-ai-spending-
to-top-$6-billion.
5 Sebastian-Gabriel Popescu, “Artificial Intelligence in Naval Operations” (paper
in each above aspect of narrow AI. This further expands the possibility for
discrepancies between human designers and the behavior of the final
product. This tremendous change will affect the balance of power between
states by increasing the relative strength of the attack, challenging the ability
of human commanders to exercise control after the start of hostilities and
behaving in ways that are in mutual conflict, go beyond human boundaries
and are difficult to predict. The problems could be exacerbated when it
comes to escalating military conflict, deterrence, or violence in AI use.
These aspects have so far been measured only by human psychology and,
eventually, in the future, a novel discipline, “machine psychology,” will
attempt a similar analysis of decision making, justified and performed by
machines.6
Tactical AI, as we have seen, will have a significant impact on the strategy,
even if it uses only the limited technologies available today. This type of
narrow AI is much more “basic” than the scheme outlined before the AGI.
It lacks the degree of cognitive flexibility to easily switch between tasks and
adapt quickly to new circumstances. It can be combined with another
domain-specific AI to create complex knowledge, as in the intelligent poker
machine or similar to the group “swarm intelligence”. In contrast to this
model, the AGI would be an integrated package capable of managing the
individual discrete tasks of composite AI and many other elements. In its
military form, this type of AI could integrate the activities of multiple nets
into a distributed, swarm-like system. It could concentrate power at all
levels of intensity. Of course, this AGI will retain the basic qualities of
narrow AI, such as pattern recognition, memory, and speed, which are easily
transformed into military power. Additionally, it is assumed that the overall
AI will be able to generate distinctive motivation that is not provided by its
designers. Combined with speed and the ability to escalate your efforts
sharply, this makes it dangerous and unpredictable.
The AGI will have various attributes: its nets will maneuver at speed, and it
will concentrate its efforts quickly, outrunning any adversary known to date
in military science. Additionally, it will be able to coordinate efforts across
its full range of capabilities, the main impact of which would be the
management of existing parameters for escalation, violence and deterrence.
When considering the military aspects of AI, much attention is paid to the
synergy between humans and machine. With the increasing speed of
processes, people will not have the ability to stay in the cycle due to the lack
of speed. In such cases, it is good that they can at least remain in the
management chain and be able to intervene and stop actions that are already
under way, especially escalation actions.
The AGI does not need to be aware of what it is doing in a military conflict.
He qualifies as a flexible and adaptive intelligence, able to navigate in a
complex environment and achieve its goals. These qualities are especially
useful when it comes to combat. The integration of different types of
weapon systems with AGI will allow the machine to coordinate not only at
the tactical level, but also at the operational and possibly strategic level.
While tactical AI decisions may have operational and strategic effects in
certain situations, combat AGI will deliberately modulate its behavior to
prevail at higher levels of combat operations.
7 Dominic Johnson and Niall MacKay, “Fight the Power: Lanchester’s Laws of
Combat in Human Evolution,” Journal Evolution and Human Behavior, Vol.36, No.
2 (March 2015): 152-163.
Artificial Intelligence in Defence 101
take into account the fact that there are limitations, both in terms of available
technology and at the conceptual level, where the strategy, at this stage,
remains undoubtedly psychological.
All the world's smart munitions and network sensors cannot determine how
effective an operation in Afghanistan against insurgents and the radical
Taliban would be. At this stage in the development of military affairs,
unequivocally, every automated technique needed a man for selecting the
target or carry out some other decisive intervention. On the one hand, the
development of AI can be seen as a logical continuation of this ongoing
revolution in information processing. In this way, AI could be seen as
another tool for processing information about people. Claims for military
revolutions should be analyzed with skepticism because often the conditions
set for their introduction are too insignificant. Examples in this regard are
concepts that have come and gone at high speed in recent years, such as
hybrid warfare, the indirect approach, the comprehensive approach, the
fourth-generation war, effect-based operations, and much more. There is a
certain possibility that the revolution in strategy, as a consequence of AI,
will pass quickly and unnoticed. But it is important to stress that AI is more
than just another means of processing information. The friction of war and
luck will remain key elements of the strategy, but AI has enormous potential
to bring about a radical change in strategic issues. By making strategic
decisions on behalf of people, and even more so by performing them on
their own, based on calculations that are not within human limits, intelligent
machines can have far-reaching consequences. This could lead to the
transformation of the societies that rule them, to a change in the relative
power between them and to the evolution of the overall nature of future
conflicts. Although the current technologies have made great changes, AI is
much more radical because it changes the very basis of strategic decisions.
Machines do not use the cognitive heuristics that inform human decisions,
much less the conscious reflection we use to imagine the future or to predict
what the adversary may aim for. In such a situation, when the intelligent
machine wants to act at all costs to achieve the goals set by man, it is very
likely that part of the communication will be “lost in translation”.
Additionally, in the near future, AI actions will remain largely at the tactical
level, but the development of technology and competitive relationships will
Artificial Intelligence in Defence 103
growing in line with increased funding. This, in turn, leads to easier and
broader implementation of AI in the defense sector.
Battle Platforms
Armed forces from around the world are increasingly using AI in weapons
and other systems based on land, air, sea and space platforms. They perform
a variety of tasks: surveillance and reconnaissance, barrier patrol for defense
and maintenance of space/base, protection of forces, countering the mine
threat, anti-submarine warfare, reconnaissance, hydrographic surveys,
communication, information and navigation services, cargo delivery,
information operations, performing high-precision synchronized strikes and
others.
Unmanned aerial and land vehicles, surface platforms and submarines with
integrated AI can patrol large areas, identify potential threats and
automatically transmit information about these threats to the response
forces. The use of remote systems increases the security of the bases and
guarded objects, as well as the safety and efficiency of the personnel in a
combat situation or during a transition.
Data Processing
AI is particularly useful for quickly and efficiently processing large
databases in retrieving valuable information. Intelligent machines can assist
in the collection and analysis of information from different data sets, as well
as in the retrieval and processing of information from various sources. This
advanced analysis enables military personnel to recognize patterns and
derive correlations, significantly reducing the timing of individual stages.
For collection, the explosion of data that is occurring because of smart
devices, the Internet of Things, and human internet activity is a tremendous
source of potential information. This information would be impossible for
humans to manually process and understand, but AI tools can help analyze
connections between data, flag suspicious activity, spot trends, fuse
disparate elements of data, map networks, and predict future behavior.
Force Management
Starting with combat information and control systems, the defense programs
of a number of countries expand their scope in the planning and
implementing of AI in the command and control systems. These systems
are used by platforms on land, sea, air, space, as well as in the cyber domain.
The use of these types of systems leads to improved interaction between the
individual components and the functioning of the military system as a
whole. Simultaneously, they require significantly more limited maintenance
and minimal human intervention. AI will facilitate the management and
configuration of autonomous and high-speed platforms and weapons in the
implementation of joint attacks and automatic control. A typical example in
this regard is the use of swarm tactics, in which a large number of platforms
As the pace of battle accelerates and the volume and speed of information
eclipse the ability of human warfighters, AI will become increasingly
important for command and control. Autonomous systems that have been
delegated authority for certain actions can react at machine speed at the
battlefield’s edge without waiting for human approval. AI can also help
commanders process information faster, allowing them to better understand a
rapidly changing battlespace. Through automation, commanders can then
relay their orders to their forces – human or machine – faster and more
precisely. AI systems can also aid the military in a range of non-combat
support functions. One use of AI will be to help defense leaders better
understand their own forces. By analyzing large amounts of data, AI
systems may be able to predict stress on the force in various components:
when equipment requires maintenance; when programs are likely to face
cost overruns or schedule delays; and when service members are likely to
suffer degraded performance or physical or psychological injuries. Overall,
AI has tremendous potential to help defense leaders improve the readiness
of their own forces by assembling and fusing data and performing predictive
analysis so that problems can be addressed before they become critical. AI
is also is ripe for transforming traditional business processes within military
and other government organizations.13
Logistics
AI systems could play a crucial role in military logistics. The efficient
supply and transportation of goods, ammunition, weapons and personnel is
an essential factor for success in modern operations. The purpose of
providing what is needed, where and when you need it, can be accomplished
with the help of intelligent machines that synchronize the request, the
13 Singh Manvendra, “Is Predictive Analytics the Future? How AI Will Take It to
database and the chain provider with an automated design of systems for
loading and sending items.
Targeting
Some AI systems are developed with the task of increasing the accuracy in
defining and prioritizing targets or target systems, as well as to determine
the adequate impact on them in a complex combat environment. They allow
the Defence Forces to gain an in-depth understanding of potential areas of
work by analyzing reports, documents, information channels and other
forms of unstructured information. In addition, the presence of AI in target
recognition systems improves the ability of these systems to identify the
position of their objects.
The machine learning of these types of systems is also used to detect, track
and investigate targets from the data received from the sensors. For
example, DARPA's Target Recognition and Adaptation in Contested
Environments (TRACE) program uses machine learning techniques to
14 Harald Bauer and Peter Breuer, “Smartening up with Artificial Intelligence (AI)
— What’s in It for Germany and Its Industrial Sector?”, accessed April 01, 2017,
https://www.mckinsey.com/~/media/mckinsey/industries/
semiconductors/our%20insights/smartening%20up%20with%20artificial%20intelli
gence/smartening-up-with-artificial-intelligence.ashx.
110 Chapter VI
Electronic Warfare
AI systems could be used for electromagnetic spectrum dominance. They
have ability to generate novel methods of jamming and communications
through self-play, like AlphaGo Zero improving its game by playing itself.
For example, one AI system could try to send signals through a contested
electromagnetic environment while another system attempts to jam the
signal. Through these adversarial approaches, both systems could learn and
improve. In 2014 an US government agency DARPA held a Spectrum
Challenge with human players competing to send radio signals in a
contested environment. DARPA is now using machine learning to aid in
15 Ke Wang and Gong Zhang, “SAR Target Recognition via Meta-Learning and
Amortized Variational Inference”, accessed October 21, 2020,
https://www.mdpi.com/1424-8220/20/20/5966/pdf.
Artificial Intelligence in Defence 111
radio spectrum allocation, but this concept could also be applied to jamming
and creating jam-resistant signals.16
Cybersecurity
Defense systems are often vulnerable to cyberattacks, which can lead to loss
of classified information and damage to combat systems. This is why
significant efforts are being made to systems equipped with AI to
autonomously protect networks, computers, programs and data from any
kind of unauthorized access. In addition, supported AI systems for web
security can record the pattern of cyberattacks and develop counterattack
tools to deal with them.
Conclusion
In conclusion, the following key remarks can be made:
4. Maintaining the status quo fuels a race in AI armaments and increases the
pressure on leading nations.
Expanding the role of AI can increase the benefits of the military system.
Not only does the integration of greater AI autonomy reduce casualty rates
of personnel, but such systems can adopt riskier tactics; target with greater
accuracy; and operate with greater endurance, range, and speed while
retaining a greater level of flexibility and mobility.
There is a need for regulation of the legal and ethical aspects that AI in the
defense sector will inevitably pose in the near future. In the process of
developing international humanitarian law, it is essential that states ensure
human control in decision making for the use of autonomous lethal weapons
systems.
Our new tools can make us smarter and can enable us to better understand
the military battlefield, our world and ourselves. Deep Blue didn’t
understand chess, or even know it was playing chess, but it played it very
well. We may not comprehend all the rules our machines invent, but we will
benefit from them nonetheless. Synergy between the human mind and AI
has tremendous potential to increase our defense capabilities and at this
moment, everything depends on us. The thought from the President of the
Future of Life Institute Max Tegmark, is exactly in that direction:
“Everything we love about civilization is a product of intelligence, so
amplifying our human intelligence with artificial intelligence has the
potential of helping civilization flourish like never before–as long as we
manage to keep the technology beneficial.”17
References
Barnaby, Frank. 1987. The Automated Battlefield: New Technology in
Modern Warfare. Oxford: Oxford University Press.
Bauer, Harald, and Peter Breuer. “Smartening up with Artificial Intelligence
(AI)—What’s in It for Germany and Its Industrial Sector?”. Accessed
April 01, 2017.
https://www.mckinsey.com/~/media/mckinsey/industries/semiconducto
rs/our%20insights/smartening%20up%20with%20artificial%20intellig
ence/smartening-up-with-artificial-intelligence.ashx.
DARPA. “The Radio Frequency Spectrum + Machine Learning=A New
Wave in Radio Technology”. Accessed August 11, 2017.
https://www.darpa.mil/news-events/2017-08-11a.
DARPA. “Selects Performers to Advance Unmanned Underwater Vehicle
Project”. Accessed February 05, 2021. https://www.darpa.mil/news-
events/2021-02-05a
Ford, Martin. 2019. Rise of the Robots: Technology and the Threat of a
Jobless Future. New York: Basic Books.
Future of Life Institute. “Benefits and Risks of Artificial Intelligence”.
Accessed February 12, 2022.
https://futureoflife.org/background/benefits-risks-of-artificial-
intelligence/.
Harper, Jon. “Federal AI Spending to Top $6 Billion.” National Defense
Magazine, October 2, 2021.
https://www.nationaldefensemagazine.org/articles/2021/2/10/federal-
ai-spending-to-top-$6-billion.
Johnson, Dominic, and Niall MacKay. “Fight the Power: Lanchester’s Laws
of Combat in Human Evolution.” Journal Evolution and Human
Behavior, Vol.36, No. 2 (2015): 152-163.
NIKOLA SOTIROV
Introduction
The problem of predicting processes has been arising in frequency in recent
times due to many companies attempting to digitally transform and describe
the process traces within their systems optimally. The purpose of this project
is to tackle the problem by creation and comparison of different machine
learning techniques.1 As event logging has been part of the majority of the
industry sectors for many years, enough data can be collected and used to
create and test mathematical models, which will be trained to perform
pattern recognition, event and time of event predictors.2
Methodology
The project was developed and implemented by a team of five researchers.
The workflow of the project was segregated into the following subdivisions:
1 “Your First Steps into the World of Process Mining”, Software AG, accessed May
25, 2021,
https://www.softwareag.com/en_corporate/resources/asset/ebook/business-process-
transformation/first-steps-processmining.html?utm_source=bing&utm_medium=
cpc&utm_campaign=bt_aris_process-mining&utm_region=hq&utm_subcampaign
=stg1&utm_content=ebook_your-first-steps-into-process-mining_
stage1&msclkid=1a8e9abfd556137ca255f8c70261ec49.
2 Software AG, “Your First Steps into the World of Process Mining”.
Process Mining with Machine Learning 117
The dataset used in the project is the process traces collection from 2012 in
the BPI challenge. The columns extracted and used from the dataset are the
following: case (a number, which indicates which trace the event belongs
to), event (the name of the process), startTime (the timestamp on which the
process occurred), completeTime (the timestamp of completion of the
event).3
The dataset has the following columns in it: lifecycle: transition, concept:
name, time: timestamp, case: REG DATE, case: concept: name, case:
AMOUNT REQ. Each step will be explained in detail in this section. The
emphasis of the exploratory analysis was to visualize and understand the
overall process flow of the dataset and how much a process matters (via
looking at the transitions from process to process).
Figure 2 denotes the types of processes that occur in the system, and the
respective transitions from and to each other.
The more highlighted the transitions are in Figure 2, the larger their numbers
of occurrences are. The goal is to create and use algorithms that learn the
patterns of transitions between the processes and forecast where the next
transition will lead towards and at what timestamp that particular transition
will occur in the future with reasonably small errors.
As to the event prediction, two models were used and evaluated a random
forest classifier, and a neural network classifier. For each classifier, the
confusion matrices were plotted and compared. Accuracy has been the used
as the primary metric for both models. See section “Results”.
The time until the next event prediction was implemented by using a neural
network with a singular regression output neuron. R2 score and relative
absolute errors were used to evaluate the performance. See section
“Results”.
In this project, such a model is utilized to classify what the next event is
based on the previous event as input. Figure 4 showcases an example
random forest and its generalized workflow.8
https://www.researchgate.net/figure/Workflow-random-forest-classifier_
fig3_343627992.
7 “Workflow of the Random Forest Classifier”
8
Workflow of the Random Forest Classifier”.
9 “Artificial Neural Networks”, accessed June 4, 2022, https://iq.opengenus.org/
artificial-neural-networks/.
Process Mining with Machine Learning 121
Figure 5 represents a neural network with four hidden layers, one input and
one output layer.
A neural network was used to predict the next event, given an arbitrary state
of the system. The network has four layers, each with the following
sequence of neuron numbers in each layer: (128, 64, 48, 24) and respective
activation functions: (’relu’, ’relu’, ’relu’, ’softmax’). The optimizer for
adjusting the learning rate used is ADAM (Adaptive Moment Estimation).
The loss function used is Binary Cross-entropy.
The following features were used as inputs for the neural network: concept:
name, previous event, lifecycle: transition, pp event, ppp event, p lifecycle:
transition, where previous event, pp event, ppp event, p lifecycle: transition
are the engineered features from the dataset.
previous event The event that occurred before the current event.
pp event The event type of the process that occurred before the
previous event.
ppp event The event type of the process that occurred before the
pp event.
p lifecycle: The state in which the previous event was at that
transition point of time in its life cycle.
The input data are the data frame of those columns, converted to a matrix
(or tensor), which is then input to the model. The output layer [See Section
“Neural network for next process prediction”] contains 24 neurons, and the
activation is softmax, meaning that each neuron has a value between 0 and
1. The neuron where the value is the largest indicates which position in the
one-hot encoding of the event type contains a ’1’. Therefore, this is chosen
as the predicted event type of the next process.
The data used as inputs for the time of event prediction model consist of the
following columns in the data set: concept: name, previous event, lifecycle:
transition, pp event, p lifecycle: transition, weekday. The subset of these
features that contains only the engineered ones is previous event, pp event,
p lifecycle: transition, weekday. Analogously, as in the case of the model in
Section “Preprocessing for the neural network classifier of the next event”,
the input data frame is converted to a matrix tensor that is then processed by
the network. The last layer’s neuron contains the timestamp value of the
result. That is then compared to the expected result, and the mean squared
error is calculated in order to minimize the error.11
Random Forest
The random forest is trained by preprocessing the data from the default data
set and then passing it to the model. Since the model cannot see the meaning
behind a string enumeration value, the training becomes much more
difficult, and inefficient. The training occurs in the following steps:
1. Separate the event columns into separate binary columns, where the
name of the column is the name of the concrete event, and the value
is whether the event is that or not by placing ’1’ and ’0’ respectively.
2. Drop the rows where the values are null.
3. Divide the set of dependent and independent variables - next event
and event, lifecycle: transition, respectively.
4. Train the model with the data.
The training step is implemented with the default method fit in the sklearn
library. The fitting is done by the bagging method, which means that the
dataset is fed to the forest in a random manner each time.
Process Mining with Machine Learning 125
Inputting the training data as batches of tensors to the input layer, and
producing the results in the last layer, where each neuron contains the results
of the previous hidden layers, passed through a softmax activation. This
means that every neuron in the output layer contains a probability of being
active relatively to the others. So the highest probability neuron becomes
’1’, and the rest - ’0’. This layer of neurons linearized in to a 1-dimensional
vector, corresponds to the one-hot encoding of the event type that is
predicted. The binary cross-entropy function is used to calculate the
deviation from the actual result, produced by the network (aka the predicted
event) and the expected predicted event.
Knowing the error from the forward propagation phase, the neural network
can adjust the weights (the connections between the layers), such that the
error is minimized. The learning rate is variable as this model utilizes the
126 Chapter VII
ADAM algorithm for momentum gain of the learning rate. As soon as the
values of the weights are recalculated, the training procedure repeats until a
reasonably small error is reached.
In the forward propagation, the final layer contains one neuron with a linear
activation, hence making it appropriate to use a MSE (Mean Squared Error)
activation function.
Then the error is backwards propagated in the same manner, the only
difference being the adjustment of weights, since the learning rate
adaptation is now not achieved using the ADAM algorithm, but FTRL as it
reduces the loss much faster. In this case, FTRL does not introduce issues
regarding stability since the model is relatively shallow, and each layer
contains a relatively large enough number of neurons.
Results
This section explains the results of each model’s performance and opens a
discussion as to which model behaves best by comparing the metrics.
Random Forest
To describe the results of the random forest, a confusion matrix was
generated and plotted to describe the different metrics used to validate the
classifier. From Figure 8, one can conclude that for some of the events the
random forest generates a relatively high number of true positives, however,
for some the true positives are low in comparison to the false negatives and
the false positives.
Process Mining with Machine Learning 127
The method for deriving results for the neural network classifier was
identical to the random forest - a confusion matrix was generated. From
Figure 9, it can be seen that there is a much higher percentage of true
positives in the neural network predictions than in that of the random forest.
The neural network predictor of the next event type is better in comparison
to the random forest, as seen from 9 and by having compared the accuracy
of both models.
As the predictions of the regressor are not binary, but rather continuous, one
cannot generate a confusion matrix to visualize the accuracy and precision
of the model. Another method was used to visualize the model performance.
From Figure 10, one can infer that the predictions somewhat align with the
real timedeltas with some deviations, especially in the 0.5 to 2 seconds
prediction interval. That is due to the fact that the model is biased. However,
the predictions stabilize at the 3 to 5-second interval. The plot was done by
implementing a logarithmic scale in order to fit all the bars in one plot. The
calculated R2 score of the neural network timedelta predictor was 0.64. This
means that the model is relatively accurate (predicting correctly in more than
60 percent of the cases), but definitely not as precise as the event type predictor.
Conclusion
In conclusion, the results indicate that for classifying the next event type,
the neural network prevails in performance, and in the context of prediction
of the time until the next event occurrence, a reasonable R2 score was
reached, even though the deep learning model was shallow.
Discussion
Although reasonable results were achieved, the models could further be
improved to fit the data better with increased accuracy and precision. For
instance, the deep learning model that classifies the next event type can be
improved by adding at least one LSTM layer that would capture the
temporal aspect of the data. As to the timedelta predictor, more layers could
be added with various activations. The number of epochs for both models
could be increased, which could mean improvements in the results.
References
“Artificial Neural Networks”. Accessed June 4, 2022.
https://iq.opengenus.org/ artificial-neural-networks/.
“Nikola Sotirov. Github Project Repository”. Accessed June 24, 2021.
https://github.com/ NickSot/process%5C_mining.
Software AG. “Your First Steps into the World of Process Mining”.
Accessed May 25, 2021. https://www.softwareag.com/en_corporate/
resources/asset/ebook/business-process-transformation/first-steps
processmining.html?utm_source=bing&utm_medium=cpc&utm_camp
aign=bt_aris_process-mining&utm_region=hq&utm_subcampaign=stg
1&utm_content=ebook_your-first-steps-into-process-mining_stage1
&msclkid=1a8e9abfd556137ca255f8c70261ec49.
“Workflow of the Random Forest Classifier”. Accessed May 18, 2022.
https://www.researchgate.net/figure/Workflow-random-forest-
classifier_ fig3_343627992.
CHAPTER VIII
ALEXANDER LAZAROV
Introduction
A philosophical discussion of how to behave when communicating with
autonomous, Advanced Artificial Intelligence (AAI) objects requires
consideration of the following questions:
Answers to the above questions will offer a discourse for both theoretical
analysis and future AI development. This paper will conclude that the
progress of AI’s analytic and synthetic capabilities—in some perspectives,
already close to human mental faculties—involve applied philosophy with
a never previously experienced framework and range. Therefore, in
designing the next generation AI—one intended to correspond to social
expectations—thinkers of all philosophical traditions can no longer draw a
strict borderline separating their wisdom from what science and engineering
have already demonstrated.
132 Chapter VIII
Artificial Intelligence, August 31, 1955,” AI Magazine 27, No. 4 (Winter 2006): 12.
5
Alex Wissner-Gross, “A New Equation for Intelligence.” TEDx-Beacon Street
Talk. November 2013, accessed July 3, 2022,
https://www.ted.com/talks/alex_wissner_gross_a_new_equation_for_intelligence.
6 Stewart J. Russell and Peter Norvig, Artificial Intelligence: A Modern Approach,
8 Sylvain Duranton, “How Humans and AI Can Work Together to Create Better
binary opposites to sustain a picture and figure out its meaning no matter if
considering physical entities or abstract social issues like good and evil.
9 Gottfried Leibniz, The Monadology. Trans. Robert Latta (Oxford: Clarendon Press,
1898).
136 Chapter VIII
Each distinguishable state is a bit of data. The binary symbols “1” and
“0” have just two states. They encode either a positive or negative
answer to any assessment: “1” signifies “yes” while “0” signifies “no”
in reply to a specific question.
Each distinguishable state is a bit of data. The binary symbols “1” and
“0” have just two states. They encode either a positive or negative
answer to any assessment: “1” signifies “yes” while “0” signifies “no”
in reply to a specific question.
Significantly, a data bit is an elementary particle that can have only one of
two possible values and thus, it is one which can be used in digital
processing. I am not claiming that the same process happens in the human
brain. However, I do not exclude such an option. There is a similarity
because humans, and animals with higher nervous systems, and AI, have
the capability to memorize large quantities of data, as well as to generate
numerous links and references among these data. The higher the intelligent
body’s memory and data processing capability, the greater the number of
produced data links and references.
Most often, data, and information do not refer to a single bit, but to
digital bit configurations (patterns) that emerge due to simultaneous
asking and answering large-scale sets of various questions. Thus, the
highest intelligence approaches difficult to understand and complex
issues.
All intelligent bodies can extract the data from any of its carriers, and
further deal with it by modeling.
10 Luciano Floridi, “What the Near Future of Artificial Intelligence Could Be,”
Philosophy & Technology 32 (March 2019): 19,
https://link.springer.com/article/10.1007/s13347-019-00345-y.
138 Chapter VIII
Single digital bits as well as sets of bits move from sensors to brains and
artificial processing units and undergo grouping, configuration, and
structuring, which results in the generation of new digital constructions.
These results are conserved in memory in a hierarchical architecture.
Unlike the polar “yes” or “no” of a single data bit, a targeted set of
questions can lead to an unlimited variety of digital patterns. They
signify much more complex answers that can shift the pure positive or
negative statements to deep multi-layered conclusions. To illustrate,
applying digital bit calculation any computer-drawing program can
suggest a color palette that is richer in nuances than that which the
human eye can discern.
Importantly, the higher the level of intelligence, the more frequent the
inquiries and recognition of the environment via pattern applications. The
intelligent actor’s capacity, and experience to examine and learn predefine
the range and deepness of what he or it could achieve via an investigative
analytic approach. In other words, as Michelangelo has claimed, every block
of stone has a statue inside it, and it is the task of the sculptor to discover it.
At a later stage, the conserved bits and patterns can be recalled from memory
whenever needed and used for new analytic or synthetic processing. This in
turn leads to partial or total restructuring, reorganization, and rearrangement
as well as placing them in a set with new incoming data flows.
Simultaneously, depending on its memory and processing capacity, an
intelligent system may record a great number of both old and new digital
constructions (codes) combined with various links and co-relationships.
Importantly, these codes present a large variety of facts and dynamic events,
but also synthesize data such as conclusions, assessments, judgments, and
others that may affect future activities.
found. Once the intelligent body recognizes that the new data acquisition
from the physical and social environment is enough and appropriate, the
new data search stops. Then, the data bits rearrangement, reconfiguration,
and reconstruction occur combined with re-judgement and re-assessment of
the new product’s applicative potential. The generative process ends when
a representation/model is built on what is expected to happen in case the
plan is put to practice. Thus, the in-form is created.
To illustrate steps one to three: most people know that Frankfurt, Paris, and
London are among the largest civil aviation centers in Europe. This is data
they have probably captured in an act of non-focused memorization.
Imagine that, at a moment, someone intends to fly from Sydney to Sofia and
he discovers that there are no direct flights between those cities. Then, he
recalls which are the biggest European airports and decides to check for
connecting flights. This is a targeted additional data search to fill in an
evident gap. Once a connection is found, the waiting time between flights is
assessed as non-risky short or not inconveniently long, and the ticket price
is judged as acceptable, the plan is recognized as ready to apply. The in-
form becomes available, and significantly, it belongs only to its author. AI
follows a similar path. Once the in-form is produced, the intelligent
procedure has two options:
First, the intelligent body decides to keep the in-form in privacy. The
human or the machine puts the plan in memory and never presents it to
others. If this option is chosen, the intelligent procedure ends here.
However, this happens more rarely than the second option – step 4.
may happen through speech, writing, gesture, drawing, etc. The per-form
mode may be either direct or encrypted, and data identification and
acquisition may occur immediately or in the future.
At first glance, it seems illogical to think about the in-form reaching others
in advance of its author willingness to present it. Significantly, there is a
tendency to develop AI to read human’s thoughts, even when one has no
intention to publish them. In a philosophical discourse, this is a critical
aspect of the human-to-AI interface that may change human subjectivity.
However, I want to underscore two features that can appear and characterize
any in-form performance. Demonstrating the in-form by performing it never
brings considerable direct changes of state within the physical and social
environment. Those may only happen later if the intelligent procedure has
provoked some activity. Simultaneously, without exception, the per-form
requires another intelligent body’s involvement. Any type of in-form
sharing between two or more intelligent actors is what contemporary
philosophy accepts as virtual communication.
Research on the in-form, per-form, and trans-form triad must also consider
the de-form phenomenon. Deformations occur once the in-form is
presented, no matter of whether the last appears as performance or
transformation. In cases of per-form, the de-form is expressed as some
misunderstanding between what the in-form author made accessible, and
what its counterpart has understood during the communication process.
There are many reasons that cause de-forms, but their analysis falls beyond
this paper’s scope. In cases of trans-form production, the de-form is the
difference between the info-product creator’s intended goal and what was
successfully put into practice. One can imagine various reasons that cause
such deviations, ranging from incorrect planning of activities to some
general impossibilities due to physical or social circumstances.
To better understand this issue, one should know the three essential
characteristics of big data:
I will not enter this discussion of timing, although it is impressive that the
focus on AGI emergence falls on when rather than whether. However, from
my perspective, the crucial question is also how it will arise. Most texts,
dedicated to AGI issues focus either on its differences with human
intelligence (often claiming that machines just simulate intelligence), or
draw a picture in which suddenly, a new revolutionary AGI performing
automation is invented. Perhaps, thinkers are inspired by publications about
the Deep Mind algorithms involved in demonstrations with the Japanese
board game Go.12 Arguably, this is the hardest human intellectual game.
Initially, the AlphaGo deep learning algorithm won the World Championship
in Go by studying the entire human history of the game. Subsequently,
AlphaZero achieved brighter success having been provided with just the
game rules and having played for six weeks against itself. It studied its own
experience, entirely ignoring the human practice history. Logically, as this
AI managed learning Go in a short period, it was expected later to run into
new spheres. Thus, radically new AGI expectations emerged. However,
there are no publications to confirm success in such a trend.
Although not positioning myself on the list of the AGI revolution path
followers, I do not exclude it. Simultaneously, I can also draw two distinct
scenarios for the narrow AI’s development:
At the same time, a second creative option also seems possible. A new
type of narrow AI could be developed, which could coordinate the
https://deepmind.com/research/case-studies/alphago-the-story-so-far .
146 Chapter VIII
I recognize that these paths for narrow AI evolution are easier to design
compared to AGI and thus they could emerge more rapidly. I call this
eventual technology Advanced AI (AAI). In my opinion, it will be the
apparent narrow AI next generation and as it will bring considerable new
circumstances, due to having innovative intelligent bodies in play, our
contemporary society should get prepared to meet it in short terms.
Therefore, I will focus first on what seems common to us and them and what
immediately appears is the intelligence procedure, which both humans and
machines employ. Here, the issues to highlight are the following:
not change. Moreover, in any event, the intelligent data input to in-form
output process will suffer delays in relation to the examined changes of
state because of the necessary period to process the acquired data and
then determine follow-up moves on behalf of the intelligent actor. In
many cases, the shorter the intelligent processing period is, the fewer the
number of prediction errors. In this trend, it seems that machines will
overpower humans.
First, we must expect AAI to focus on shaping its future together with
us. Therefore, we shall have to assess all ongoing physical and social
processes based on a prediction of what the innovative automation is
about to demonstrate. Simultaneously, AAI will do the same about us.
The above list of issues may not cover all the variations typical for human-
to-AI and AI-to-AI communications if compared to our human-to-human
practices. However, I see them as a comprehensive background that sets out
a framework for future research and analysis. I believe, many other issues
will emerge. However, regarding human-to-human communication, many
thinkers note that effective communication is less about talking and more
about listening to reach a common understanding. I will not discuss if this
is the best behavior among humans, but I think that it would work perfectly
in our communication with advanced and general intelligence computers. I
do not claim that AAI is smarter than we are—there are many arguments
against such a statement—but without doubt, AAI will be smart enough and
completely foreign to us. Thus, at least at the first stage, many of its
conclusions will seem surprising both as to content and analytic perspective.
However, many of them will be correct.
Approaching the Advanced Artificial Intelligence 149
Conclusion
Today, our civilization is at the dawn of a new era. We are about to meet an
unknown innovative autonomous intelligent actor—AAI operating
machines that surpass us in some areas of our mental faculty. Moreover,
AAI memorizes its deep learning investigation results for future application.
Therefore, in the future, when analyzing a particular event or process, a
group of similar machines all running similar hardware and software
solutions, will generate diverse in-forms depending on what they have
explored in the past. Although this is not an expression of human
subjectivity, no doubt, this is a step toward AI individualization. Therefore,
in addition to our human-to-human face-to-face communication traditions,
we must develop a new human-to-AI and vice-versa interface that should
not follow a pure subjective-objective path but should recognize the
following new circumstances:
Advanced AI systems can act in our favor, but also as our competitors
and opponents.
References
AlphaGo. Accessed July 1, 2022. https://deepmind.com/research/case-
studies/alphago-the-story-so-far.
Duranton, Sylvain. How Humans and AI Can Work Together to Create
Better Businesses. Accessed May 7, 2020.
https://www.ted.com/talks/sylvain_duranton_how_humans_and_ai_ca
n_work_together_to_create_better_businesses?language=en.
Floridi, Luciano. “What the Near Future of Artificial Intelligence Could
Be.” Philosophy & Technology 32 (2019): 1–15.
https://link.springer.com/article/10.1007/s13347-019-00345-y.
Floridi, Luciano, and J. W. Sanders. “On the Morality of Artificial Agents.”
Minds and Machines 14, No. 3 (2004): 349–379.
Hanson, David. “Expanding the Design Domain of Humanoid Robot.”
Paper presented at the ICCS Cognitive Science Conference, Special
Session on Android Science, Vancouver, USA, 2016.
Leibniz, Gottfried. The Monadology. Translated by Robert Latta. Oxford:
Clarendon Press, 1898.
McCarthy, John, Marvin L. Minsky, Nathaniel Rochester, and Claude E.
Shannon. “A Proposal for the Dartmouth Summer Research Project on
Artificial Intelligence, August 31, 1955.” AI Magazine 27, No. 4 (2006):
12–14.
Approaching the Advanced Artificial Intelligence 151
Pareira, Luis Moniz. “AI & Machine Ethics and Its Role in Society.”
Keynote speech at the International Conference of Artificial Intelligence
and Information, Porto, Portugal, Dec. 5–7 2017.
Russell, Stewart J., and Peter Norvig. 2016. Artificial Intelligence: A
Modern Approach, 4th edition. London: Pearson Education Ltd.
Turing, Alan. “Computing Machinery and Intelligence.” Mind: A Quarterly
Review of Psychology and Philosophy (October: 1950): 433–460.
Wissner-Gross, Alex. “A New Equation for Intelligence.” Accessed
September 7, 2016.
https://www.ted.com/talks/alex_wissner_gross_a_new_equation_for_i
ntelligence .
CHAPTER IX
BORIS GROZDANOFF
Abstract
Deep Reinforcement Learning (DRL) is at the forefront of modern artificial
general intelligence (AGI) research and formalizes AI tasks in terms of
agent, environment, state, action, policy and reward and also harnesses the
function approximation power of the other leading AI instrument, the
artificial neural networks (ANNs). One of the main challenges contemporary
efforts to build artificial general intelligence (AGI) systems face, in a broad
Turing-test framed setting, is the artificial emulation of key components of
human language-based reasoning, like meaning of singular terms and
expressions but also semantically enabled logical reasoning. Here I envision
a high-level AGI system, RAISON, that could be driven by a DRL
architecture. I suggest that we can use Frege’s influential distinction
between sense and reference in order to emulate linguistic meaning as found
in real world human practice. For that purpose, I propose that a semantic
graph space (SGS) can be trained on available NL datasets to deliver
artificial semantic ability. I also suggest that the formal abilities of syntax
manipulation and rules-based logical reasoning can harness the expressive
power of SGS and thus enable knowledge-based «educational» training of
an emerging general AI model.
Introduction
In May 2022, the lead research scientist at Google’s DeepMind, arguably
the world leading company in the development of Artificial General
Intelligence systems, Nando De Freitas, provocatively claimed that
The Structure of Artificial Rationality 153
“It’s all about scale now! The Game is Over! It’s about making these models
bigger, safer, compute efficient, faster at sampling, smarter memory, more
modalities, INNOVATIVE DATA, on/offline. Solving these scaling
challenges is what will deliver AGI. Research focused on these problems, eg
S4 for greater memory, is needed. Philosophy about symbols isn’t. Symbols
are tools in the world and big nets have no issue creating them and
manipulating them”. 1
1https://twitter.com/NandoDF/status/1525397036325019649
2 For a brief summary of the discussion see Sparkes, Matthew article in “Is
DeepMind's Gato AI really a human-level intelligence breakthrough?” in New
Scientist, 19.05.2022. https://www.newscientist.com/article/2320823-is-deepminds-
gato-ai-really-a-human-level-intelligence-breakthrough/
154 Chapter IX
complements it is the one that while the above can deliver AGI, «philosophy
of symbols» cannot. I will discuss both theses in more depth, but two points
deserve mentioning before that: first, the term «more modalities» in De
Freitas' claim and second, the apparent derogatory formulation of
«philosophy of symbols». Under modalities De Freitas clearly has in mind
modalities of multi-modal systems like GATO,3 where each modality has a
distinct function and nature: playing a certain game, leading a conversation,
image recognition and the like.
3 For the GATO system see DeepMind’s paper: Scott Reed et al. (2022) “A
Generalist Agent” in arXiv:2205.06175v2 [cs.AI].
4
Puppe, F. (1993) “Characterization and History of Expert Systems” In: Systematic
Introduction to Expert Systems. Springer, Berlin, Heidelberg.
https://doi.org/10.1007/978-3-642-77971-8_1
The Structure of Artificial Rationality 155
Marcus5 and LeCun6 were among the first to dispute De Freitas, on different
grounds and in different directions. LeCun disagrees about AGI on a
fundamental level and argues that AGI is not and could not be well
formulated, let alone achieved technically. The next best thing, according to
him, is HLAI or human-like AI and our actual progress is only to be
expected in the direction of HLAI, not AGI. The main challenges before
HLAI are the lack of generalized self-supervised learning, no learning
paradigm for machines to learn how “the world works”. I agree with him
that it is best first to attempt HLAI, but I also believe, unlike LeCun, that
HLAI can lead to AGI. The argument behind this is fairly trivial: if a true
HLAI is artificially emulated, it would be trained on virtually all human
knowledge and there is no reason whatsoever to think that it would not learn
superhuman abilities in knowledge but also human-like rationality. From
DRL systems like AlphaGo, we already know that AI can find better
“moves” than humans have managed to for millennia. There is no obvious
reason why the DRL model that manages to reach HLAI would not do the
same with rational human abilities.
Marcus, following his long defended positions, that step on the shoulders of
AI giants, in his own words, like MacCarthy, Minsky and Simon, among
others, directly argued against the thesis (3). Following upon his long-
standing view that understanding the mind even at a high level is a necessary
prerequisite for success in AGI and the impressive results of artificial neural
networks (including in deep reinforcement learning models with neural
networks for value and policy learning functions) could not alone achieve
this goal, but should also be complemented with what he calls, following a
science-of-alt-intelligence?s=r
6 Matsuoa, LeCun et Al. (2022) “Deep learning, reinforcement learning, and world
«Alt Intelligence isn’t about building machines that solve problems in ways
that have to do with human intelligence. It’s about using massive amounts
of data – often derived from human behavior – as a substitute for
intelligence. Right now, the predominant strand of work within Alt
Intelligence is the idea of scaling. The notion that the bigger the system, the
closer we come to true intelligence, maybe even consciousness.» 8
Marcus argues that not just De Freitas provocative statement that AI in its
HLAI form is being reduced to the above alternative intelligence, but also
that our best and most successful systems, like the recent GATO and
Flamingo,9 express nothing else but alternatively intelligent capabilities and
not at all human-like intelligent capabilities, least of all, generally intelligent
capabilities. The core of his argument can be unpacked in the following
form:
concerns the very structure of the AI model: the kind of networks chosen,
the type of their integration, the learning algorithm, the activation function,
etc. If a model like GATO or subsequent more successful and more
powerful multimodal «generalizing» agents pretend human-like general
intelligence they need to prove that the intelligent functionality of the model
mirrors intelligence abilities of humans. Short of that, I am afraid that even
Markus' alternative intelligence term would be too beneficial for such
systems, since our only scientific criteria for intelligence are our criteria for
human intelligence. Systems pretending general human-like intelligence
would not only have failed to prove that they are human-like intelligent,
they would have failed to prove that they are intelligent at all and thus they
would be at best, only considered «alternative».
Here I suggest that the leading ANN and DRL technologies are not to be
opposed with syntactic (but also semantic and reasoning AI models) but,
on the opposite, their power can be harnessed to drive such models in order
to learn human-like abilities of syntax proficiency, semantic proficiency,
reasoning proficiency and knowledge proficiency or all the kinds we find
in intelligent human practice based on natural language (NL).
The AI model would also need to reason causally, among other things. It
would need to understand the usage of the human terms cause and effect and
to learn to use them in the same way. And at the end, it would need to unite
all those abilities in a coherent unity, such as we find in a human's conscious
rational mind. Only then would we have grounds to assess it as a human-
like artificial intelligence, if not in structure, at least in function. And that
would be considered a success by the interdisciplinary community that
slowly erects the complex theoretical foundation of human intelligence,
which could only serve as a basis for its artificial counterpart, unlike the
black-box model. The alternative would be to formulate virtually ex novo a
novel, non-human intelligence and to the best of my knowledge, science
knows much less about such intelligences than it does about the human one.
The oldest and, arguably, the richest scientific probe into the nature and
structure of human intelligence, as comprised by understanding, reasoning
and self-awareness, is the western civilization's endeavor of philosophy that
has developed immensely complex and critically evolved debates on the
problems of semantics, reasoning, rationality and consciousness. This paper
follows in the steps of this tradition, which, I believe, can be immensely
useful for the practical purposes of arriving at a functioning general artificial
intelligence. This approach, to structure an AGI model, based on
philosophical doctrines, is surely unusual and highly unpopular in today's
exclusively computer science based projects, which loosely take
mathematical models of human neurons (like the ANNs) and models of
human-environment interaction (like the reinforcement learning model) as
sufficient to emulate the yet «unsolved» riddle of human intelligence. At
the end of the day, however, all approaches would reach the same judge and
the same field of assessment: the one that works first or best would
undoubtedly be considered a valuable if not the valuable approach. Thus,
the philosophy, mostly in its analytic line, faces an opportunity that is rarely
if ever present, to prove success in a practically measurable setting, such as
the AI and AGI race.
160 Chapter IX
Symbolism Unpacked
The generally quoted «symbolism», however, both in pure mathematics and
in logic, but also in written natural language, always comes in two parts:
syntax and semantics.11 Computation, in its forms of, say, functional
analysis, linear algebra or tensor calculus, is impossible without syntax and
equally impossible without the semantics that interprets expressions in this
syntax as well formed, as true or false, but also as meaningful. If the last
seems like too hard a requirement for a mere purely mathematical
expression of the form f(x)=y let us compare it with an alternative ()fyx=,
which contains the exact same symbols and the same number of them.
Where the two «expressions» differ syntactically is only their order. The
«symbols», to use De Freitas preferred term, are the same. They are just not
ordered in the same way in the two symbol strings. And yet the difference
is highly non-trivial, for their meaning, and for their ability to be interpreted
in truth-value. For the first is a well-formed formula of a function and as
such is mathematical, whereas the latter is not; but is a meaningless and
uninterpretable string of the same symbols not even capable to be
approached with truth or falsity. And computations, such as those we use in
ANN and DRL models, can only use the former kind, whereas the latter is
certainly of no use not just for computer science but also for the whole of
mathematics; symbolic gibberish can only figure with some valuable
significance in science fiction or in a modernist poem.
What is of key importance here, is that the syntax and its interpretation are
inseparably intertwined, even if we decide to take a slight historical detour
on Hilbert's formalism school of thought. From Hilbert, we know that any
choice of syntax, of symbols and order assigned to them, could be, of course,
arbitrary.12 But once chosen, it ceases to be arbitrary and begins to follow
formal rules and principles and to require ultimate rigor in its writing in
order to be semantically interpretable, that is, to be formally functional and
thus significant and meaningful. The semantic interpretation of any
11 Van Fraassen, Bas C. (1971) Formal Semantics and Logic, The Macmillan
Paul Benacerraf and Hilary Putnam, pp. 183 – 201, Cambridge University Press.
The Structure of Artificial Rationality 161
13 Schrittwieser, J., Antonoglou, I., Hubert, T. et al. (2020) “Mastering Atari, Go,
chess and shogi by planning with a learned model” in Nature 588, 604–609.
The Structure of Artificial Rationality 163
restrained way how the human brain functions. What is characteristic of the
DL technology is that it is able to be fed with large volumes of previously
unstructured data or largely unstructured data, and then process it via a
dynamically structured architecture of neural networks, with input layer,
which receives the data, hidden (or deep) layers, that process the data sent
from the input layer and an output layer that classifies or predicts the desired
output. DL is able to classify and suggest predictions with enormous
accuracy and precision.
14MDP analizes tasks in terms of agent, environment, observations and actions, for
an in-depth discussion see Olivier Sigaud and Olivier Buffet (Eds.) (2010) Markov
Decision Processes in Artificial Intelligence, Wiley.
164 Chapter IX
15The RAISON name of the model is devised to accent on the English word reason,
as one of the key components of the suggested model of human-like intelligence, the
other being understanding, and contains the AI abbreviation for obvious purposes.
The Structure of Artificial Rationality 165
The RL and neural network modules have been successfully integrated with
two additional AI methods: self-play (SP) and tabula rasa (TR) training,
that is, training with no previously provided data sets. The potential of DRL
systems integrating SP and TR (DRL-SPTR) was first demonstrated by
DeepMind’s algorithms AlphaGo Zero, AlphaZero and most recently
MuZero and LaMDA. What is truly remarkable in MuZero is the novel
ability of the system to learn the rules of the game by itself and then to
develop the ability of superhuman performance in the recognized game.
The syntax ability (SYA) should consist of syntax structures for recognition
and classification and it would need to be integrable with the semantic and
the reasoning systems. To illustrate, if we want the syntax ability to be
trained in 2nd order predicate logic, it would need to recognize 2nd order well
formed formulae and differentiate them from non-wwfs in the same logic.
This ability, albeit formal, is essential for human-like NL abilities and would
The Structure of Artificial Rationality 167
The reasoning ability (RA) will represent human’s real life practice where
we reason, for our concrete purposes and along the main thesis of analytic
tradition, within the natural language.16 Generally, this is the ability to
operate with syntactically well-formed linguistic expressions, to interpret
them semantically in truth-value, in sense and in reference, and on the basis
of these to introduce and discharge connectives for well-formed formulae
(atomic and molecular) as well as to concatenate them into larger sets of
expressions. The RA system would learn to construct ordered chains of
expressions following the logical forms found in human practice: the logical
laws, such as modus ponens and modus tollens and many others. Most of
all, the system would acquire the ability to reason syllogistically, to draw
logically valid conclusions from available premises and to assess logical
arguments as valid, invalid, sound or not sound.
16Central thesis of analytic philosophy accepts that all reasoning occurs within NL
– for a discussion see Martinich, A. P. and Sosa, David (eds.). 2001a: Analytic
Philosophy: An Anthology, Blackwell Publishers, and Martinich, A. P. and Sosa,
David (eds.). 2001b: A Companion to Analytic Philosophy, Blackwell Publishers.
168 Chapter IX
problems of knowledge: its nature, its kinds, criteria for, etc. It is my assumption
that analytic theories of knowledge are both the most promising ones and most
suitable for the practical purposes of building AGI models. For a high-level
The Structure of Artificial Rationality 169
In this way, I hope it is clear that only a @WK trained SSR system can
become artificially rational and truly artificially intelligent. The SSRR
system (syntax, semantics, reasoning and rationally trained system or SR2)
would be an integrated AI system that exhibits human-like abilities to
comprehend expressions, reasons using them and which knows, like
humans, via @WK “educational” training in AI training datasets, like the
WiKi Corpus, facts about the real world. But most of all, it would be able
to form sound logical arguments and communicate via them. The AI ability
in SR2 would represent a successful operation of human-like intelligent
abilities in the actual world. Such ability is arguably the best candidate to be
a decisive criterion to judge the intelligence of humans - it could not be mere
communication between humans, as presented in the classical Turing test
setting, but it should be the evolutionary environment of successful
functioning, survival, adaptation and evolution of rational humans in the
real world. A fully functional artificial agent, like RAISON, acting in this
environment without being identified by humans as an artificial vs. human
agent would be a much stronger version of the Turing test than the original
one.
The data for the DRL system training would need to allow its successful
training and this is a non-trivial task. It would need to be mathematically
encoded as n-dimensional tensors, as popular ML systems of language
encoding of the word2vec 23kind would lack expressive power and
sufficiently powerful integrability with semantic spaces like SGS. Without
these properties, however, traditional encodings would be unable to serve
their AGI purpose; the necessary mathematical expressive power can be
delivered only by general n-dimensional tensors and thus the word-level NL
encoding would be a word2tensor (or w2t) encoding. The necessary
encoding for the DRL-SPTR system would need, on the one hand, to
provide the specific encoding for each system (SYA, SA, RA and RAA):
syntax2tensor (SY2T), semantics2tensor (SE2T), logic2tensor (L2T) and
actualknowledge2tensor (@2T), but on the other, the dimensionality and the
23 Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean (2013) “Efficient
Estimation of Word Representations in Vector Space”, in
https://arxiv.org/pdf/1301.3781.pdf
The Structure of Artificial Rationality 171
The SY2T would encode the syntax well formed formulae of words and
expressions. The SE2T would encode the graph content of terms and
expressions that are the results of the three semantic interpretations, in sense,
in reference and in truth-value, that comprise the semantic interpretation in
meaning. The L2T would assemble semantically interpreted SY2T formulae
into argument premises according to the rules of introduction and rules of
discharge of logical connectives. The @2T would encode the interpretation
of expressions with respect to actual world knowledge.
The semantic space SGS that would enable the SA system but that would
also host the @WK training data would best be formalized as a highly
expressive graph space with operanda, syntactic and semantic, as nodes and
operations as edges, where the 2nd order logical functionality would be
introduced by allowing edge to edge connections that would represent
predication of predicates, but would also allow for the natural formal
expression of universal and existential quantifiers. Expressions of all kinds
in the SR2 system, linguistic, syntactic, and semantically interpreted, would
be graphs and graph-chains that would be attributed a number of indices,
which would express the different systems’ properties and roles. In SGS, we
introduce a syntactic index for well-formedness, possessed by graph g and
not, say, graph e, a semantic index that expresses sense and another semantic
index, that represents reference (pointing to the graphs that express the
reference and negating a number of graphs that oppose this reference
interpretations), logical index for a graph that is a valid premise in a valid
argument A and a logical index for a graph that is a premise in a sound
argument S. Sound arguments thus would be ordered sets of connected
graphs in SGS and sound reasoning would be an ordered set of graphs where
every graph represents either a true premise or a true conclusion.
In the suggested DRL HL-AGI model RAISON, all of the above abilities
would be abilities of the agent who would operate in an environment
segmented in syntactic layer, semantic layers, reasoning layers and actual
world data trained or educated layer. The model needs to learn a high-level
structure that emulates the function. The elements of RAISON are
172 Chapter IX
In the case of the syntax ability we need first to identify what would be the
syntax environment of the agent, what would be the states of the
environment and what be the actions of the agent in it. The very agent would
be a syntax agent and his agency role in the general DRL model of the HL
rational agent would effectively be a one of a syntactic avatar which is
integrated with the general agent and mediates between it and the syntactic
environment. The environment would be the complete space of available
ordered syntactic elements, like well-formed formulae and non-well formed
formulae. However, since the syntactic function is inseparable in the
integrated functionality of the AGI-DRL agent, and since in real life
scenario humans never face a syntactic only environment, but face a
naturally linguistic environment of expressions, in order for the syntactic
agent to contribute to the DRL agent in a human-like way he would need to
24 See David Silver, Thomas Hubert et al. (2017) “Mastering Chess and Shogi by
Self-Play with a General Reinforcement Learning Algorithm” in
https://arxiv.org/pdf/1712.01815 and Schrittwieser, J., Antonoglou, I., Hubert, T. et
al. (2020) “Mastering Atari, Go, chess and shogi by planning with a learned model”
in Nature 588, 604–609.
The Structure of Artificial Rationality 173
face the same linguistic environment as the one the DRL faces. Because its
task is only syntactic, upon observation of a fully linguistic environment
(FLE – an environment of all available NL expressions that follow syntactic
rules, receive semantic interpretations and participate in reasoning
algorithms), that is, an environment of expressions such as the one humans
face in real-time dialogues, speeches, or readouts, the syntactic agent would
need to extract only syntactic features and structures from it.
25 Ji He, Jianshu Chen, Xiaodong He, Jianfeng Gao, Lihong Li, Li Deng and Mari
The syntax ability would be modeled by a deep neural network that would
take as an input naturally linguistic expressions and would learn to
recognize its logically formalized form solely on the basis of its syntax. In
the DRRN model of He and al. they use two separate DNNs, one for the
states of the text environment and one for the text actions, due to the
expected significant difference between the two inputs in terms of
complexity and length. In our syntactic model, we do not need to account
for such differences and we can therefore use one DNN that would take both
states and actions as inputs. Our initial choice for syntax would be 2nd order
predicate logic, which provides immense expressivity, sufficient for
formalizing scientific expressions but also, and more importantly, it mirrors
human linguistic practice more naturally than other logical formalisms
because it allows quantification and predication over predicates. Again, the
The Structure of Artificial Rationality 175
rigor of the system would be available to the DRL model, but on the surface,
in its mirroring human behavior, expressivity would need to dominate over
rigor.
The environment in our setting would be the initially unordered set of all
available text expressions to the agent. The DRL system would use a deep
neural network with an input layer that would take as input data the state of
the environment, which the agent would “observe”. In our case, this would
be the equivalent to the “move” of the computer in this “game”: in a Turing
set dialogue, the computer would utter an expression and that would be the
move of the machine in a gamified setting. The move would be a novel and
last expression, which appears in the environment and which would change
its state. The agent would take the move within the general state of the
environment as an input and would employ the neural network to output a
tensor of “move” probabilities for his actions and a scalar value for the
expected “outcome” from the new state of the environment which would
include its action.
The success of the agent response would be the adequacy between the
agents’ response and the sequence of the observed state-response pairs. This
adequacy would be optimized for syntax adequacy, semantic adequacy and
logical adequacy, by the corresponding layers of the model.
27 For my stance on Frege’s theory of sense and reference I am very much indebted
to my doctoral supervisor and mentor, Nenad Miscevic, as well as to Katalin Farkas
and Howard Robinson from the CEU, and, later, to Gabriel Uzquiano and Ofra
Magidor from the University of Oxford. I am also indebted to the late Sir Michael
Dummett, whom I had the chance to listen to on a number of seminars at Oxford just
a year before his passing.
178 Chapter IX
The semantic agent would interact with a semantic environment, where the
states of the environment and the agent-actions are semantic. In order for
the semantic environment to be as human-like as practically possible, it
would need to provide recognition of semantic truth and semantic meaning.
Human linguistic practice is devoid of truth determinacy and meaning
determinacy because we almost never find in real life human practice truth-
rigorous expressions and meaning-rigorous expressions. In Frege’s theory
of sense and reference, meaning of linguistic expressions can be analyzed
in terms of sense and reference and this dynamic duality exhibits the
potential of expressing a significant volume of real life human language
interactions while having the potential for actual rigor in truth and meaning.
These tasks are key to human-like reasoning via language and the first acts
that build up the human understanding ability. In this case, the semantic
observation of the text-state by the semantic agent would need to understand
the meaning of the terms and then to understand the meaning of the intended
statement of ST. The understanding of the terms would consist in identifying
in our semantic graph-modeled space (SGS) the concepts of the terms and
28Gottlob Frege (1948) “Sense and Reference” in The Philosophical Review, Vol.
57, No. 3 pp. 209-230 (22 pages), Duke University Press.
The Structure of Artificial Rationality 179
the connections with other concepts in other expressions that convey best
the meaning they have in ST. Thus, the meaning of every term in SGS would
be the node of the term (the node for “Hesperus” or the edge of “is”, as
operandae are modeled on graph nodes and operatii are modeled on graph
edges) and the graphs, which employ it in other expressions. The totality of
actual expressions with the node-term would represent the totality of the
human practice of the term. The sub-space of the usage of the term in typical
contexts, which form a certain meaning of the term (any term can have
numerous meanings, some of the very distinct and remote from others), are
represented by the connected graphs of the expressions, where graphs are
expressions in the form aRb where a and b, as operandae are nodes and R,
as an operatio, is an edge.
The SGS graph space would represent the semantic source available to the
environment, in our DRL model, but also to the general RL agent. In order
for any semantic agent to be able to perform a semantic interpretation, it
would need access to the SGS.
On the basis of these expressions, the agent can retrieve the actual meaning
of the term in human practice (we assume that SGS has been trained on
actual human practice). Graph selection can serve to form the sense of the
term and a ST or the sense of an expression. It can also serve to fix the
reference of the expressions. After the semantic interpretations in sense and
reference, the agent can perform the semantic interpretation in truth-value.
180 Chapter IX
Frege defines the sense as the content of the cognitive result upon
understanding the linguistic expression by the human agent.29 The sense can
be exchanged with a fellow conversant in the same language or translated
into a different language. The sense determines the reference (the object or
state of affairs) to which the linguistic expression applies in a desirably
unique way. To use Frege’s own illustration with the identity statements,
which have a simple yet immensely significant structure of two names or
descriptions being connected by a single linguistic identity operator like “is”
or “are”: if we consider the two different expressions “Venus is the morning
star” (Vm) and “Venus is Venus” (VV) we observe that they differ
significantly in their informative content even if they actually refer to the
same object, the planet Venus. This and many similar cases raise a puzzle
about meaning, which requires an explanation by a successful theory of
meaning. Frege’s explanation distinguishes between two elements of the
meaning of the expressions. The first one is the sense of the meaning, which,
in this puzzle, carries the load of the difference in informational content,
evident from the different syntactic structure of the expressions, which
differ in syntactic content and order. The second one is the reference of the
expression, which in the puzzle happens to be one and the same object. The
identity in the meanings of both expressions is explained by the identity in
reference, while the difference in the meaning of the expressions is
explained by their different senses.
29 Ibid.
The Structure of Artificial Rationality 181
that sometimes captures it, as in expressions like “John loves Jane” and
sometimes fails, as in “The car’s engine released a lot of phlogiston” where
there is no real object “phlogiston”, referred to by the term. In mathematical
cases like “5+7=12” or “5+7=10”, which mathematicians do find non-
trivially distinct, in terms of truth and meaning, the inquiry into the referents
of purely mathematical expressions continues the heated debates even
today, as philosophers of mathematics clash their preferred views within the
frames of Platonism, structuralism, formalism, fictionalism and other
influential stances.
We can observe that the sense can be formalized as a recipe for an operation
and thus as the initial step of a semantic algorithm, where in a finite number
Semantic Algorithm
1. Intention of the speaker
2. NL encoding of intention following a syntactic selection of symbols
and an order that holds between them with a semantic load of the
terms
3. Formation of expression
4. Formation of intended reference
5. Utterance
6. Acquisition by a listener (or the computer)
7. Decoding of the symbolic structure
8. Semantic interpretation in meaning: decoding the acquired sense
9. Semantic interpretation in reference: attempting to comprehend the
intended reference
10. Act of referring after (9)
11. Semantic interpretation in truth value: the listener is able to assess
truth of the expression on the basis of (8) and (9)
12. End of communication cycle (1 - 11)
We see that in the SA, the steps contain both operandae and operatii. This
structure makes it quantifiable for the purposes of the DRL model.
and the integrability with syntax, reasoning and rationality necessary for the
“education”.
Atoms of Reasoning
The first and minimal syntactic act of reasoning is, formally and
syntactically, the predicate attribution to a bearer of a predicate, a subject,
or, in 2nd order predicate logic, a predicate as well. This predication is also
the first and minimal expression. The syntactic consideration of a single
syntactic element, of any kind (subject, predicate, relation or quantifier) is
not sufficient to be accepted as the first and minimal syntactic act of
reasoning because it receives a good syntacticity or a syntactic well-
formedness only at the general level of predicate or relation attribution to an
x that is not itself the predicate or the relation and yet can accept it and carry
it. The availability of a single element allows for the first act of
quantification, since we can define both existential and universal quantifiers
over it and we cannot do it without the availability of an element. In order
to be able to form the quantification “There is an x” we need the x to be
present, for otherwise the expression could only take the form “There is”,
denoted syntactically by the existential quantifier symbol and this is not a
well-formed expression in logical formalism, which also, for that very
reason, fails to be able to carry any meaningful semantic load. The
184 Chapter IX
introduction of quantifiers thus becomes first possible only after our syntax
provides us with an x the quantifiers can range over. After the quantification,
the element, however, is not a single element anymore: we have two
elements, a quantifier and an x, connected in a syntactically proper way
which allows their concatenation to be regarded as the formal structure of
an expression. The expression represents a formal act of quantification,
which, when considered as a schema, has the same formally general form
as the schema of predication of a predicate to a subject.
setting. For example, the singular term “John” does not form a meaningful
expression and cannot be regarded as expressing something beyond itself.
By itself, it does not provide us with the possibility to quantify over it and
to predicate it: we have no quantifiers and no predicate terms available. So
no introductions of logical operators can be performed on “John” unless we,
before that, make available a second term, like “sings”. The availability of
two terms allows all of the above that can give the rise of a syntactically
well-formed expression “John sings” that expresses something beyond what
the singular terms express: it expresses an order that holds between them.
This order is encoded by Frege’s sense and fixes the references of the
singular terms and the expression.
The connection between the reference of the name John and the reference
of the act of singing in real life furnishes the state of affairs, which should
provide the reference of the expression. Our initial expression “John sings”
gives no information as to what other things John does or is: we have no
information in the sense that he is a man or a dog, etc., but in a context we
can infer those from the context and see that the utterer implicitly means
that John is a man. Once the sense portion of the meaning of the expression
186 Chapter IX
We can now, armed with our sense, enter the domain of referential
interpretation, say the set of observational data that we have, and investigate
it as to the reference state of affairs of the expression. We should first look
for the existence of the referents of the singular terms, John and the singing.
If we find them in the data, we should investigate if an order holds between
them and if there is one if the order is the one described in the sense. Thus,
we can find in the data both John and singing, but we can also happen to
find Jane and, say, the only order that we can find, is an order that holds
between the singing and Jane and not John. In this case, we would have
found the referents of the singular terms, John and the singing, but not the
presence of the sense prescribed order to hold between them. Our semantic
interpretation of the expression would be that referentially John does not
sing and consequently, the expression, while well formed semantically, does
not express the true state of affairs that actually obtains. This would
represent the semantic interpretation of the expression in truth-value, where
we have approached and examined the expression with respect to its truth
or falsity. Note that being meaningful does not at all lead automatically to
truth, but allows for an interpretation in truth-value.
Other practices are also relevant and inform the domain of numbers, like
foundations of mathematics and mathematical epistemology. From them,
among others, we know that numbers, even if being all the time related to
the real life states of affairs, as in expressions like “John sang two songs”,
do not by themselves inhabit the physical world but only the abstract,
causally inert, outside physical space and time domain of abstract objects.
Abstract mathematical objects, like numbers and functions, can be applied
to non-abstract objects, via a non-trivial process of application;31 non-trivial
because it combines two distinct domains with very different rules of
meaning construction, meaning examination and most of all – reference
examination. The main epistemological method in the empirically
accessible domain, such as the physical world with (some, and far not all,
as quantum mechanics revealed to use) its states of affairs, is empirical
observation and measurement. This is the main epistemic tool of natural
sciences 32 and common sense knowledge acquisition, the empirical method.
Whereas in abstract domains such as the ones of pure, unapplied
mathematics, logic and metalogic, we have very different conditions of truth
and meaning, which are impenetrable for the empirical method 33. Thus, if
we want to be able to interpret purely mathematical and logical expression,
in meaning and in truth value, and especially, with respect to their delivering
31 Perhaps the best place to see a still very much up to date discussion of the
as Penelope Maddy’s. For their views see: The Collected Works of John Stuart Mill,
edited by John M. Robson (Toronto: University of Toronto Press, London:
Routledge and Kegan Paul, 1963–91); Philip Kitcher (1980) “Arithmetic for the
millian” in Philosophical Studies 37 (3):215 - 236 (1980); Penelope Maddy (1992)
Realism in Mathematics, Clarendon Press; James Robert Brown (2008) Philosophy
of Mathematics: A Contemporary Introduction to the World of Proofs and Pictures,
Routledge.
188 Chapter IX
knowledge, we need a different tool than the one we use in the empirical
domain. We cannot observe numbers (perhaps only in a neo-Platonic
epistemology of mathematics or in a neo-formalist kind of epistemology of
mathematical knowledge as the one influentially proposed by Charles
Parsons34) the way we observe birds and cars, but we nevertheless know that
humanity has an enormous body of truths about them that are considered by
science firm truths that bring about knowledge to those who are lucky
enough to gain access to them, if not even firmer than the ones of physics.
1. Formation of sense
2. Formation of reference
3. Execution of sense – via uttering, communicating or formalizing
4. Execution of reference – act of intended reference, act of fixing
reference (by conversant)
(4) obtains when the sense and the reference have been well-formulated and
an interpretation of both has been executed. To use a loose analogy from
quantum mechanics, we can approach the wff sense and reference as a
quantum object that remains in a certain state while unobserved, that is,
while not semantically interpreted. In QM, any observation of a quantum
object collapses its wave function and thus changes the parameters of its
physical state, and correspondingly, values that figure in its mathematical
description. In our “quantum sentence” case, we can assign a structure of
quantification that would have one set of values while the sense of the
Any interpretation of the sense changes the formal uninterpreted state. For
example, the semantic load of the interpreter is anything but trivial: if the
sense of the term Venus as Hesperus is the only one used by the formulator,
and Venus as Phosphorus is the only one available at the interpretor, we will
end up in a miscommunication situation, where same string of symbols (for
Venus) is intended as Hesperus by the formulator and interpreted as
Phosphorus by the interpretor. Thus, the identity of syntax would not be
sufficient in order to convey the intended sense to a user different from the
formulator and this would prevent the communication with a conversant.
And we need not even agree that the main function of language is the
communicative function.35 We should only agree that the communicative
function of language is non-trivial for its emergence, existence and usage.
The first and minimal logical act of reasoning is the first logical operation
over a logical operandum. This could be, on an intra-expression level, the
attribution of a predicate to a subject (“John is tall”), the quantification over
subjects (“there is an x”), or quantification over predicates (“for all x” where
x is a predicate). But on the inter-expression level, the logical acts of
reasoning emerge via the usage of logical operators between expressions
Edward Arnold.
190 Chapter IX
(“P & Q” where P and Q are expressions). Those become available only
when we first have the expressions, syntactically well formed and
semantically interpreted. Once we have a number of available expressions,
like the text-actions in the DRL model, we can use the rules of introduction
and discharge of logical operators to establish order onto them. Thus, if we
have P, Q and R, we can introduce conjunctions to form “P&Q” or “P&R”
or “P&R”, etc. We can further develop logical complexity by using the
forms of logical laws (modus ponens, modus tollens, etc.) and attempt to
find if one of their forms holds for our available expressions, given their
syntactic well-formedness, and their semantic interpretation.
as modus ponens and modus tollens and others. Most of all, the system
would acquire the ability to reason syllogistically, to draw logically valid
conclusions and to assess logical arguments as valid, invalid, sound or
unsound.
The observed text-states from SGS, ST, can now, after a full and successful
semantic interpretation in meaning, be fed to the reasoning ability of our
DRL agent. Each term, formalizable syntactically can be retrieved from
SGS and presented to the reasoning engine for logical operations. Thus, if
we have as a ST the expression “Hesperus is Phosphorus” we can send it to
the reasoning engine, along with all relevant definitions for the terms and
connected expressions and we can attempt to reason about it. For example,
the agent already has the results of the semantic interpretations and knows
that the referent r of both terms is actually one and the same, the planet
Venus. But he also knows that the statement uses two distinct meanings,
which even if they happen to refer to the same r are not identical. The graph
structures for the 1st sense and the 2nd sense do not coincide in SGS. The
agent can compile tentative sets of expressions, connect them according to
logical rules and see whether a law can deliver a conclusion from some
assembled lists.
Methods
The RAISON model is a deep reinforcement learning model, which would
develop the suggested abilities for syntax recognition, semantic
interpretations, logical reasoning and rationality. These abilities are very
distinct, but in order RAISON to function they need to be seamlessly
integrated. In the DRL setting, the abilities will be modeled as possessed by
the RL agent, which develops them and perfects them in the course of its
interaction with an environment. For the AGI purposes the environment
would possess a complex structure and would have two layers: the deep
layer,36 which possesses the syntactic, the semantic and the reasoning
structures, modeled by the semantic graph space (SGS), and the surface
36 Not to be confused with the “deep” notion from the deep neural networks (DNNs).
192 Chapter IX
layer, which is the portion of the environment the agents observes and with
which it interacts.
The agent would learn a strategy to find the most adequate text response-
actions, modeled by a policy. The policy in DRL is best found through deep
neural networks and especially so in cases, like the present one, where the
space of environment states or the space of agents’ actions is immense and
practically unknowable. The general deep neural networks (NG) of the
RAISON model are functions ƒș(s), where the learnable parameters are
denoted by ș and the observed by the agent state of the environment is
denoted by s; ƒș(s) takes natural linguistic text data as input and outputs the
probability of a text-response given state s: pa = Pr (a|s) and a scalar value
v, which assesses the quality of the adequacy of the response. Each ability
in the suggested architecture would have a dedicated DRL agent whose
expertise would be integrated in NG.
The syntactic ability would have a DRL agent ASY which, which interacts
with the environment but whose DNN, NSY, learns to recognize syntactic
formulae.
The semantic ability would have a DRL agent ASE, which interacts with the
environment but whose DNN, NSE, learns the human-like semantic
interpretations via dedicated deep neural networks. The semantic
interpretations to be learned by NSE are the above formulated interpretations
in sense, reference and truth-value. Each semantic interpretation would have
its own dedicated neural network, NSS, for the sense interpretation, NSR for
the reference interpretation and NST for the truth-value interpretation. Once
each semantic neural network learns its ability to perform adequate
interpretations in sense, in reference and in truth-value, these outputs would
be fed to the RL agent in order to choose an action, given the state of the
environment. He would grow to learn a strategy that would endow it with
the ability to “beat” the machine, that is, to deliver the most adequate
semantic text-response to the observed state of the environment, or the
“move” of the environment, in game terms.
Thus, the reasoning ability would have a DRL agent AR, which interacts
with the environment but whose DNN, NR, learns to reason like humans.
The Structure of Artificial Rationality 193
Thus, the rational ability would have the same DRL agent ARA, which
interacts with the environment, with the key difference that his DNN, NRA,
has been trained on immense volumes of actual world data which training
would render him immensely “educated”. NRA would take as input the
training data and it would output the semantic graph space, SGS, that would
receive upon training artificial education of real world knowledge. This
process would be the artificial analog of an artificial education of the agents,
that would draw on the trained SGS for real world facts.
The distinction between a reasoning agent and a rational agent would be that
the rational one is educated and knows many things, whereas the reasoning
agent only needs a minimal amount of training, which would make it able
to reason like humans. The rational agent would be a functional cognitive
agent in a possession of human-like knowledge.
human thought. Again, since they are explicit and rigorously defined the
training would be performed by supervised learning. The last task of
training would be actual world knowledge training, or artificial “education”.
Due to the actuality of knowledge, the technique of training would again be
supervised learning.
The tabula rasa, approach, familiar from Aristotle’s epistemology and used
to great success by DeepMind in their DRL agents, if at all available, would
be present only within the agent-environment interaction, where the agent
would learn in a semi-tabula rasa fashion the environment of the conversant,
which would consist of the ad hoc naturally linguistic communication free
of any previous load but the choice of the environment. Those can be
performed artificially, emulating, say, a human Turing kind jury, by using
quantum randomness37 to determine the choices for the topic and the
directions of the conversation with the agent.
The RAISON model can be trained on available NLP datasets, like (as of
2022) UCI’s Spambase, Enron dataset, Recommender Systems dataset,
Dictionaries for Movies and Finance, Sentiment 140, Multi-Domain
Sentiment dataset, Legal Case Reports Dataset, but mostly on the Wiki QA
Corpus and the Jeopardy datasets. The syntax training would be supervised
and the syntax rules would be explicitly trained, as well as the reasoning
rules.
The main challenge is the semantic graph space model SGS, the main layer
of RAISON, which would directly benefit from the NLP datasets training.
The semantic structures, that will emerge upon training the SGS space on
available big volume NL datasets would establish a significant number of
subject-predicate structures, predicate-predicate structures, the subject-
relations structures, predicate-relation structures as well as linguistic forms,
found in real life human linguistic practice. Those structures are available
in the datasets and would be embedded in the SGS space upon the training
that would seek the functionality, devised for the semantic module of
RAISON.
The suggested AGI model has the main justification for its proviso structure
in the rich and critically evolved tradition of analytic philosophy, which is
reflected in the syntax, semantics, reasoning and relational abilities to be
learned. Due to the immense development and training work, the model
would be truly in a position to be assessed both as an AI potential and as
prospects for improvement only when developed in code and after a certain
preliminary but still significant training on large volumes of data.
196 Chapter IX
Bibliography
Alayrac, J.B., Donahue, J. et al. (2002) “Tackling multiple tasks with a
single visual language model” in https://arxiv.org/abs/2204.14198
Brown, J.B. (2008) Philosophy of Mathematics: A Contemporary
Introduction to the World of Proofs and Pictures, Routledge.
Halliday, M.A.K. (1973) Explorations in the Functions of Language.
London: Edward Arnold.
Etchemendy, J. (1999): The Concept of Logical Consequence. Stanford:
CSLI Publications.
van Fraassen, Bas C. (1971) Formal Semantics and Logic, The Macmillan
Company, New York.
Frege, Gottlob (1948) “Sense and Reference” in The Philosophical Review
Vol. 57, No. 3 pp. 209-230 (22 pages), Duke University Press.
Gettier, E. (1963). “Is Justified True Belief Knowledge?”. Analysis. 23 (6):
pp. 121–123.
Grozdanoff, B. (2014) A Priori revisability in Science, Cambridge Scholars
Publishing. Newcastle upon Tyne.
He, J. Jianshu Chen, Xiaodong He, Jianfeng Gao, Lihong Li, Li Deng and
Mari Ostendorf (2016) Deep Reinforcement Learning with a Natural
Language Action Space in: arXiv:1511.04636v5 [cs.AI]
https://doi.org/10.48550/arXiv.1511.04636 , last accessed 17.06.2022.
Hilbert, David (1984) “On the infinite” in Philosophy of Mathematics,
Edited by Paul Benacerraf and Hilary Putnam, pp. 183 – 201, Cambridge
University Press
Kitcher, Ph. (1980) “Arithmetic for the millian” in Philosophical Studies 37
(3):215 – 236.
Linnebo, Øystein (2008) “Compositionality and Frege’s Context Principle”
in Philosophy.
Lycan, W. G. (2006). “On the Gettier Problem Problem.” In Epistemology
Futures, (ed.) S. Hetherington, Oxford University Press.
Maddy, P. (1992) Realism in Mathematics, Clarendon Press.
Marcus, G. F. (2001) The Algebraic Mind: Integrating Connectionism and
Cognitive Science, Cambridge, MA: MIT Press.
The Structure of Artificial Rationality 197
DISCRIMINATOR OF WELL-FORMED
FORMULAE AS A COMPONENT
OF ARTIFICIAL HUMAN RATIONALITY
DIMITAR POPOV
best debatable and only when certain caveats are applied 2before training.
Neural-symbolic computing has been an active area of research for many
years seeking to bring together robust learning in artificial neural networks
with human-like reasoning capabilities and self-explainability via symbolic
computational frameworks. The central aspect of such symbolic frameworks
is that intelligence results from the manipulation of abstract compositional
representations whose elements stand for objects and relations.3 In this
paper, I argue the foundation of a state-of-the art computational framework,
which relies on logical formalization and operation for reasoning and neural
networks for knowledge embedding. By knowledge embedding, we mean
the elicitation of features of a particular real-world object or concept in a
given domain, those features are then encoded inside the connections of the
NN, representing real world justifiable true beliefs. The utilization of this
encoding is transformed into a set of relational rules that collectively
represent the captured knowledge. The end goal of achieving systems with
better robustness, as it is believed to be one of the basic trends of Human
Rationality (HR), constitutes what Gary Marcus defines as:
22 Yann LeCun, Yoshua Bengio and Geoffrey Hinton, “Deep Learning,” Nature,
521(7553) (May: 2015): 436-444. Irina Higgins Loïc Matthey, Acra Pal et al., “Beta-
VAE: Learning Basic Visual Concepts with a Constrained Variational Framework,”
Proceedings of ICLR (2017), accessed May 3, 2022,
https://www.semanticscholar.org/paper/beta-VAE%3A-Learning-Basic-Visual-
Concepts-with-a-Higgins-
Matthey/a90226c41b79f8b06007609f39f82757073641e2. Irina Higgins, Nicolas
Sonnerat, Loïc Matthey et al., “Scan: Learning Abstract Hierarchical Compositional
Visual Concepts,” Proceedings of ICLR (2018), accessed March 23, 2021, arXiv
preprint arXiv:1707.03389. Marta Garnelo, Kai Arulkumaran and Murray
Shanahan, “Towards Deep Symbolic Reinforcement Learning”, accessed May 25,
2022, arXiv preprint arXiv:1609.05518. Brendan M. Lake and Marco Baroni,
“Generalization without Systematicity: On the Compositional Skills of Sequence-
to-Sequence Recurrent Networks,” accessed June 14, 2022, arXiv preprint
arXiv:1711.00350. Gary Marcus, “Deep Learning: A Critical Appraisal,” accessed
June 29, 2022, arXiv preprint arXiv:1801.00631. Matthew Botvinick, David Barrett
and Peter Battaglia, “Building Machines that Learn and Think for Themselves:
Commentary on Lake et al.,” Behavioral and Brain Sciences (2017), accessed July
2, 2022, arXiv preprint arXiv:1711.08378.
3
Marta Garnelo and Murray Shanahan, “Reconciling Deep Learning with Symbolic
Artificial Intelligence: Representing Objects and Relations,” Current Opinion in
Behavioral Sciences, Vol. 29 (October: 2019): 17-23.
Discriminator of Well-Formed Formulae as a Component of Artificial 201
Human Rationality
Contemporary Uncertainty
Integration between learning and reasoning is one of the key challenges in
artificial intelligence and Machine Learning (ML) today. Where in humans
the ability to learn and reason, from and to the surrounding environment is
categorized as a central preordain feature of the natural intelligence,
engineering the same two interrelated concepts inside programmable
hardware has spurred a plethoric creation of novelty approaches in
Computer Science.
4 Gary Marcus, “The Next Decade in AI: Four Steps towards Robust Artificial
Intelligence,”
accessed February 12, 2022, arXiv:2002.06177.
5 Daniel Kahneman, Thinking, Fast and Slow (New York: Farrar, Straus and Giroux,
2011).
66Jaakko Hintikka, “Philosophy of Logic”, Encyclopædia Britannica, accessed
7 Tom Mitchell and Hill McGraw, Machine Learning (New York: McGraw Hill,
1997).
Discriminator of Well-Formed Formulae as a Component of Artificial 203
Human Rationality
forward, the signal is first scaled by the weights of the edges, wi,j , where i
denotes the position of the neuron in the previous layer and j denotes the
position of the neuron in the layer the signal is been propagated to, for
example w3,5 is the weight between the third and the fifth neurons between
two consecutive layers. The signal is then fed to the first hidden layer. Inside
each neuron in the hidden layer, two operations take place, summations of
all weighted signals from a previous layer and applying an activation
function, if a provided threshold is exceeded, the neuron is activated and
propagation through this neuron is enabled, or we are saying that the neuron
fires. Then, at time tk to tk+1, the input of a neuron is calculated as:
Where bi is called a bias and is an arbitrary value predefined for each neuron,
it could also be omitted. The activation function could vary, depending on
the problem at hand, in our case we are using the Softmax activation
function:
Here, the exponent of a given neuron is divided by the sum of all exponents
of the rest of the neuron. The Softmax function returns the probability of a
given example to be in a class. After each cycle, an optimization technique
is applied to adjust the weights, if the output value of the Softmax functions
in the output neurons differs from those of the pre-labeled training
examples. This optimization is performed by calculating the Gradient
Descent of each neuron weight in relation to the difference between the
current output value and the desired value, for every single neuron in the
entire network. The difference between the sum of the outputs of the last
layer and the pre-labeled training examples is computed by a math function
called a loss function, which could take various mathematical forms. That
difference is then used to start an automatic differentiation algorithm called
204 Chapter X
Where,
Neuro-Symbolic Approach
Neuro-Symbolic Artificial Intelligence can be defined as a subfield of
Artificial Intelligence (AI) that combines neural and symbolic approaches.
By neural, we mean approaches based on artificial neural networks –
sometimes called connectionist or sub-symbolic approaches – and in
particular this includes deep learning, which has provided very significant
breakthrough results in the recent decade and is fueling the current general
interest in AI. By symbolic we mean approaches that rely on the explicit
representation of knowledge using formal languages – including formal
logic – and the manipulation of language items (“ symbols”) by algorithms
to achieve a goal. Mostly, neuro-symbolic AI utilizes formal logics, as
studied in the Knowledge Representation and Reasoning subfield of AI, but
the lines blur, and tasks such as general term rewriting or planning, that may
not be framed explicitly in formal logic, bear significant similarities, and
should reasonably be included.
9Steven Pinker and Alan Prince, “On Language and Connectionism: Analysis of a
Parallel Distributed Processing Model of Language Acquisition,” Cognition, 28(1-
2) (March: 1988): 73–193.
Discriminator of Well-Formed Formulae as a Component of Artificial 207
Human Rationality
11
David Silver, Aja Huang, Chris Maddison et al., “Mastering the Game of Go with
Deep Neural Networks and Tree Search,” Nature 529 (January: 2016): 484-489.
210 Chapter X
We follow the syntactical rules for creating logical expressions for First-
Order Logic to restrict the formalism. Such formalism, although having its
limitations, for example, cannot define finiteness or countability, is
considered powerful enough to represent the natural language statements
sufficiently well. The allowed symbols are restricted over a signature, here
Discriminator of Well-Formed Formulae as a Component of Artificial 213
Human Rationality
we are considering all letters from the English alphabet as well as the
following operational symbols.
Ɣ Implication operator :
In the last row, the last three symbols are making, respectively, the NOR,
NAND and XOR Boolean operators; these particular symbol’s annotations
are purely our own choice. We will refer to this list as the operational
symbol list. We are also using the following symbols for existential (exists,
does-not-exist) and universal quantifiers:
Every symbol described above is used to construct the following sets, when
we create a syntactically valid expression:
5. Set of terms T.
Is the union .
Training Data
The training data will be synthesized from a custom code that will first
create the mentioned collection of sets . Second, the
sets (A, M) will be written into two text files, one for the atomic formulas
and one for the molecular. In these files, each member of the set will be
written in a separate line as the string that we want to pass to the input layer
of our RNN. One additional text file will also be added, any line of this file
will be a string of jumbled random characters in ascii, this will be our noise.
As a whole, we have three labels for classification Noise, Atomic,
Molecular. The training set contains the following records described in the
table below.
Training set
Label Records
Noise 18
Atomic 32617
Molecular 300000
In our experiment, the atomic set is used to generate the molecular set, this
is a tricky part for the network to discover, since the information encoded
into the molecular set is also encoded in the atomic set. The noise represents
a particularly blunt difference between the form of a syntactically correct
expression and random strings, that is why the network, should need very
few of these examples to train itself.
Discriminator’s Architecture
As we pointed out, the discriminator would be a many-to-one recurrent
neural network. Our first layer will have a size of 100 input neurons, this is
exactly to cover all printable characters, digits, punctuation and brackets in
ascii encoding. The hidden layer is constructed out of 128 fully connected
neurons, which are connected to the three output neurons. In practice,
because of the specifics of PyTorch, what we end up is a combination of
216 Chapter X
The training is performed along the course of 100 000 cycles, at each cycle
a random training sample is selected from any of (N, A), and is fed to the
100 input neurons, for each character in the line. The selected learning rate
is . We are selecting such a miniscule learning parameter
because we want the model to learn the subtle difference which members of
A are used for in the creation of the members of M.
The result from the training can be seen in Figure 8. This is the so-called
confusion matrix, which represents the average normalized prediction of the
model for each 1000 cycles. We could see that the model has achieved great
performance, during the training phase, as the yellow color in the diagonal
corresponds to the normalized degree of prediction in this case 1.0, meaning
extremely close to 100%.
Testing
In order to test our model, we have generated a new logical expression using
anew construction of the sets (S, F, Fs, T, A, M). The table below shows the
record distribution of the test set and the test results.
Test set
Label Records
Noise 3 274 584
Atomic 2 976 230
Molecular 3 000 000
Conclusion
This work has introduced a theoretical compositional framework for neural-
symbolic integration that utilizes functional hybrid as a leading design
paradigm. The main contribution of this paper is the elaborated schema of
compositional design for an artificial agent with human-like level of
rationality. One module of this framework, a sub-symbolic discriminator for
valid logical syntax, is constructed in this article. The results from training
and test cycles have achieved over 99.13% accuracy. In future work, a better
discriminator of valid logical arguments, semantic parser from natural
language to logical expression reasoning and semantic engines will be built
to complete the neuro-symbolic framework proposed in this article.
References
Botvinick, Matthew, David Barrett, Peter Battaglia, Nando de Freitas,
Darshan Kumaran, Joel Leibo, , Timothy Lillicrap et al.. “Building
Machines to Learn and Think for Themselves: Commentary on Lake et
al.” Behavioral and Brain Sciences (2017). arXiv preprint
arXiv:1711.08378.
Cranmer, Miles, Alvaro Sanchez-Gonzalez, Peter Battaglia, Riu Xu, Kyle
Cranmer, David Spergel, and Shirley Ho. “Discovering Symbolic
Models from Deep Learning with Inductive Biases”. NeurlPS 2020.
Accessed July 3, 2022. arXiv:2006.11287.
Encyclopedia Britannica. “CYC”. Accessed June 6, 2022.
https://www.britannica.com/topic/CYC.
220 Chapter X
Marcus, Gary. “The Next Decade in AI: Four Steps toward Robust Artificial
Intelligence”. Accessed February 12, 2021.
arXiv:2002.06177.
Marcus, Gary. “Deep Learning: A Critical Appraisal”. Accessed June 29,
2022. arXiv preprint arXiv:1801.00631.
McCloskey, Michael, and Neal J. Cohen. “Catastrophic Interference in
Connectionist Networks: The Sequential Learning Problem.”
Psychology of Learning and Motivation, Vol. 24 (1989): 109–165.
https://doi.org/10.1016/S0079- 7421(08)60536-8.
Miranda, Lester J. “Understanding Softmax and the Negative Log-
likelihood”. Accessed July 3, 2022.
https://ljvmiranda921.github.io/notebook/2017/08/13/softmax-and-the-
negative-log-likelihood/.
Mitchell, Tom, and Hill McGraw. 1997. Machine Learning. New York:
McGraw Hill.
Pinker, Steven, and Alan Prince. “On Language and Connectionism:
Analysis of a Parallel Distributed Processing Model of Language
Acquisition.” Cognition, 28(1-2) (1988): 73–193.
Ratcliff, Roger. “Connectionist Models of Recognition Memory: Constraints
Imposed by Learning and Forgetting Functions.” Psychological Review,
97 No. 2 (1990): 285–308.
https://doi.org/10.1037/0033-295x.97.2.285.
Rumelhart, David, Geoffrey Hinton, and Ronald Williams. “Learning
Representations by Backpropagating Errors.” Nature 323 (1986): 533-
536. https://doi.org/10.1038/323533a0.
Silver, David, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre,
George van den Driessche, Julian Schrittwiesser et al. “Mastering the
Game of Go with Deep Neural Networks and Tree Search.” Nature 529
(2016): 484-489. https://www.nature.com/articles/nature16961.