Journal of Physics: Conference Series
PAPER • OPEN ACCESS
Conflict resolution via emerging technologies?
To cite this article: Chika Yinka-Banjo et al 2019 J. Phys.: Conf. Ser. 1235 012022
View the article online for updates and enhancements.
This content was downloaded from IP address 207.90.39.16 on 24/07/2019 at 13:54
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
Conflict resolution via emerging technologies?
Chika Yinka-Banjo1 , Ogban-Asuquo Ugot1
Sanjay Misra2,3, Adewole Adewumi2,
Robertas Damasevicius4 and Rytis Maskeliunas4
1
University of Lagos, Lagos, Nigeria
Covenant University Nigeria, 3Atilim University, Turkey
4
Kaunas University of Technology, Kanus, Lithuania
2
Abstract. This paper presents a review of the current techniques and approaches adopted in conflict
resolution in Multi-Agent Systems (MAS). The review highlights the strength and weaknesses, and thus,
their success in fostering cooperation and collaboration in multi-agent systems. We survey alternative
approaches to conflict resolution that rely on emerging technologies such as deep learning. From the survey, we discuss the benefits of using these emerging technologies in the conflict resolution process.
1. Introduction
An agent is a computational entity with the ability to perceive and act upon its environment, through sensors and
effectors. Intelligent (rational) agents strive to achieve a goal by choosing the optimal path to a solution, one that
maximizes a performance measure. The intelligent agent maximizes its performance measure by making rational
decisions [20]. In a single-agent environment, trade-offs between decisions are made without considering how
these decisions affect any other agent [20]. This is not the case when solving problems in a multiagent system.
Conflict may occur when the optimality of an individual agent’s decision is limited by decisions of other agents
within the same MAS. Conflict resolution therefore, is the essential process of managing conflict within multiagent environments in order to restore cooperation and collaboration. Current approaches to conflict resolution in
MAS include negotiation and arbitration. Negotiation has been the focus of central interest in multiagent systems,
as it has been in the social sciences [14]. This is because negotiation provides a powerful approach for dealing
with inter-agent dependencies at run-time [5]. Negotiation can be used to resolve a wide range of MAS conflicts.
When agents in conflict cannot achieve an agreement arbitration strategy are adopted. In arbitration, an arbitrage
agent is entrusted with the responsibility of coordinating conflict [5]. We shall consider these approaches and
others in detail.
Section 2 of the paper presents a literature review of multiagent systems and section 3 discusses the current
approaches to conflict resolution in multiagent systems. In section 4, a review of relevant modern approaches to
conflict resolution in multiagent systems is presented, followed by a discussion of the relevance of these emerging
techniques in section 5.
2. Literature Review
2.1. Multiagent Systems (MAS)
Multiagent systems can be viewed as a subset of Distributed Artificial Intelligence (DAI) [2]. In MAS, the main
focus is how agents organize their knowledge about their environment and activities and analyze logically, the
process of coordination [6]. Agents are generally classified into passive and active agents; the former has no goals
while the latter has clearly defined goals and it sets out to achieve those goals [2].
•
•
•
•
•
The agents in MAS have several important characteristics [2];
Agents are intelligent and autonomous
Centralized control
Decentralized data
Asynchronous processing and computation
Applications of multiagent systems can be found in industrial and commercial environments [5, 7]. Some applications include;
Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution
of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.
Published under licence by IOP Publishing Ltd
1
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
•
•
•
Modelling and optimization of transportation systems.
Ecommerce and their markets, where agents act as buyers and sellers, and can thus purchase and sell goods
on behalf of their users.
Real-time monitoring and management of telecommunication networks, where agents are responsible, e.g.,
for call forwarding and multiplexing and transmission.
2.2. MAS Architecture and Environment
The environment of an agent might be open or closed, deterministic or stochastic and it might be single or multiagent [16, 15]. Although situations exist where an agent may operate by itself in a single agent environment, the
increase in the Internet of Things is making such situations rare, therefore agents usually interact with at least one
other agent [17]. The focus of this subsection is on systems with multiple agents. Characteristics of a Multiagent
environment include [2, 18];
• Multiagent environments, specify interaction protocols and communication protocols.
• Environment is usually open and decentralized.
• Agents in a multiagent environment are usually autonomous and distributed, and may be adversarial or cooperative.
Fig 1: MAS environment
2.3. Deep Learning
Deep Learning is a subset of Machine Learning which focuses on an area of algorithms which was inspired by
our understanding of how the brain works in order to obtain knowledge. It’s also referred to as Deep Structured
Learning or Hierarchical Learning. Deep Learning builds upon the idea of Artificial Neural Networks and scales
it up to be able to consume large amounts of data by deepening (adding more layers) the networks [3]. By having
a large number of layers, a deep learning model has the capability of extracting features from raw data and "learn"
about those features little-by-little in each layer, building up to the higher-level knowledge of the data. This technique is called Hierarchical Feature Learning, and it allows such systems to automatically learn complex features
through multiple levels of abstraction with minimal human intervention [3]. Implementing deep learning can be
outlined in the following steps;
•
•
•
•
raw data → s is fed to input layer
Data is transformed by hidden layer
Output layer returns target scalars Q (s; →)
Using back-propagation, train network on labelled data
2
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
Fig 5: Deep Learning Neural Network with 3 hidden layers
One of the most distinct characteristics of Deep Learning – and one that made it quite popular and practical – is
that it scales well, that is, the more data given to it, the better it performs [3]. Unlike many older machine learning
algorithms which has a higher bound to the amount of data they can ingest – often called a "plateau in performance" – Deep Learning models has no such limitations (theoretically) and they may be able to go beyond human
comprehension, which is evident with the modern deep learning-based image processing systems being able to
outperform humans [3].
3. Conflict Resolution strategies in MAS- State of the art
One can approach the research of conflict resolution from three perspectives: system autonomy [10, 4], adversarial
domains [10] and cooperative multi-agent systems [6]. Conflict resolution of cooperative multi-agent systems can
be classified into: distributed systems [8, 9], model description [19], and applications [12, 13]. From the literature,
it is obvious that the most commonly used techniques for conflict resolution in non-cooperative MAS are Negotiation and Arbitration [3].
3.1. Other Conflict Resolution Strategies
In [3], the authors propose the ConfRSSM (Conflict Resolution Strategy Method) in the domain of Learning
Management System (LMS). The researchers demonstrate enhancing agents’ interactions and cooperation can
greatly benefit from classifying conflicts. Researchers in [7] presented a hybrid approach as a solution for issues
related to resolving conflicts in MAS. The approach is well suited to avoidance, prevention and detection of
conflicts. Majorly, the conflicts arise due to fuzzy tasks, indistinguishable sub goals and unknown associated
constraints [1].
3.2. Problems with Current Conflict Resolution Strategies
The current research on resolving conflict fails to address some concerns such as;
1.
2.
3.
4.
Attention is not given to the conflicting agents and their confidence level
Attention is not given to the number of groups of conflicting agents and the number of issues the conflict
is centered around
Some conflicts are less important and can be ignored, there is no means of identify the importance of a
conflict
Cost and time efficiency are important factors when choosing a resolution strategy, there are no rules to
selecting amongst resolution strategies [6].
4. MAS Learning within the Context of Conflict Resolution
This section discusses the role of learning in conflict resolution and it provides an entry point for discussing the
potential of emerging deep learning techniques in conflict resolution.
4.1 Robot Path Planning and Collision Avoidance
Agents with shared resources within a multiagent system, can benefit from prioritization [2]. Researchers in [3]
explore the use of genetic algorithms to assign priority dynamically. The performance of a team of agents needs
to be improved without affecting the individual agent’s performance. The study implements a decoupled heuristic
approach where a high-level planner agent is trained to reduce conflict by assigning priority. This planner agent
is introduced only after individual agents have learned to optimize their performance. Specially designed for
3
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
Partially Observable Markov Decision Process (POMDP) environments, the approach is demonstrated on a problem in 3D aircraft path planning.
4.2. Network Management Systems
There is a reasonable amount of effort that has been dedicated to developing real-time traffic management systems
[4]. By training the system on high resolution simulation models, agents are able to achieve state estimation and
short-term prediction. State estimation is useful where real time predictions have to be made and short-term predictions provide fast predictions [5]. One will still need to augment online adjustment modules together with
several consistency checking techniques could be integrated with the simulation model and periodically activated
in order to maintain the model consistency. The study in [5] presents a multi-agent learning methodology for
consistency checking and online calibration of real-time traffic network simulation models. The approach strives
to adjust the agent’s performance enabling them to learn based on their own performance and percept sequence.
The performance of the methodology is examined using real-world data. The results show that the methodology
is promising as an efficient mechanism for maintaining model estimation consistency.
4.3. Reinforcement Learning for Resolving Demand Capacity imbalance
The work in [9] proposes and investigates the use of collaborative reinforcement learning methods for resolving
demand-capacity imbalances during pre-tactical Air Traffic Management. The work recognizes the importance
of online real-time data and leverage on this by building data-driven techniques for predicting correlated aircraft
trajectories and, as such, respond to a need to handle airplane collision, a problem identified in contemporary
research and practice in air-traffic management. The simulations, designed based on real-world data, confirm the
electiveness of their [9] methods in resolving the demand-capacity problem, even in extremely hard scenarios.
4.4. Reinforcement Learning for Autonomous Vehicles
In [7], the authors investigate the application of Reinforcement Learning to two well-known decision dilemmas,
namely Newcomb’s Problem and Prisoner’s Dilemma. These problems are exemplary for dilemmas that autonomous agents are faced with when interacting with humans [6]. Furthermore, that argue that a Newcomb-like
formulation is more adequate in the human-machine interaction case and demonstrate empirically that the unmodified Reinforcement Learning algorithms end up with the well-known maximum expected utility solution.
4.5. Deep Reinforcement Learning in Conflict Resolution
Recent advances in combining deep learning and Reinforcement Learning have shown a promising path for designing new control agents that can learn optimal policies for challenging control tasks [9, 13]. These new methods address the main limitations of conventional Reinforcement Learning methods such as customized feature
engineering and small action/state space dimension requirements. In [6], the authors leverage one of the state-ofthe-art Reinforcement Learning methods, known as Trust Region Policy Optimization, to tackle intersection management for autonomous vehicles. They [6] show that using this method, they can perform fine-grained acceleration control of autonomous vehicles in a grid street plan to achieve a global design objective and avoid unnecessary conflict.
4.6. Deep Q-learning in Conflict Resolution
Cooperation and competition are rising behaviors seen in versatile multiagent systems. Within the work in [9] the
researchers think about how participation and competition are new in independent agents that utilize reinforcement learning prepared utilizing camera sensor information for state representation. The deep Q-Learning system
is actualized in multiagent situations and the interaction between two learning agents is explored within the wellknown video game Pong. By manipulating the remunerate rules of Pong, the authors are able to illustrate how
competitive and collaborative behavior develop. The conflict resolution procedures recommended focusses on
ways to extend the motivation to collaboration and it guides the movement from competitive to collaborative
conduct. Finally, by playing against another versatile agent, rather than against a hard-wired agent, the authors
accomplish more vigorous techniques. The work [11] appears that Deep Q-Networks can serve as a valuable
device for consideration when dealing with multiagent systems.
5. Discussion
Multiagent deep reinforcement learning has a considerably large literature on conflict resolution in the multiagent
environment [9, 11]. However, it is important to note that most of the research and experiments have been conducted in either simple grid worlds or with agents already equipped with abstract and high-level sensory perception. Despite this one can still observe the importance of learning in MAS.
4
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
Deep Learning and Deep reinforcement learning in particular, serve more as a prevention of conflict rather than
a conflict resolution strategy [12, 10]. At first glance this might seem contradictory to the goal of the study which
is conflict resolution via emerging technology. From our study of Deep Learning applied in MAS, one can see
that when an agent possesses a learning element, the instances of conflict with other agents within the MAS
reduces [6, 5]. Therefore, our proposition, based on the study is that conflict resolution strategies are required less
if the agents learn about the multiagent environment using a deep reinforcement learning model. In any case, if
conflict occurs, agents with a model about the MAS, obtained from learning, find it easier to resolve conflict [12,
13]. Future work on conflict resolution in MAS should adopt learning strategies to enhance the performance of
the agents within the multiagent system. Such enhancements in turn can increase the efficiency of current conflict
resolution strategies.
Acknowledgment
The authors gratefully acknowledge the financial support of African Institute for mathematical sciences (AIMS)
Alumni small research grant (AASRG), the Organization for Women in Science for the Developing World
(OWSD), and L’oreal-Unesco for Women in Science.
References
[1] Aaron R, Enrique M, Saul P and Enrique S 2012 Conflict resolution in multiagent systems: Balancing optimality and learning speed, Eleventh Mexican International Conference on Artifitia l Intelligence.
[2] Albrecht S and Stone P Multiagent Learning: Foundations and Recent Trends, viewed on 26 November
2017, from http://www.cs.utexas.edu/~larg/ijcai17_tutorial/multiagent_learning.pdf
[3] Albrecht S, Crandall J and Ramamoorthy S 2016 Belief and truth in hypothesised behaviours. Artificial
Intelligence, 235:63–94.
[4] Alicia Y, and Ghusoon S 2017 A Conflict Resolution Strategy Selection Method (ConfRSSM) in MultiAgent Systems, International Journal of Advanced Computer Science and Applications (IJACSA), Vol
8, Issue 5, pp. 398-404.
[5] Banerjee B and Peng J 2004 Performance bounded reinforcement learning in strategic interactions. In Proceedings of the 19th AAAI Conference on Artificial Intelligence, pages 2–7.
[6] Banerjee D and Sen S 2007 Reaching pareto-optimality in prisoner’s dilemma using conditional joint action
learning. Autonomous Agents and Multi-Agent Systems, 15(1):91–108.
[7] Barber K, Goel A and Martin C 2000 “Dynamic Adaptive Autonomy in Multi-agent Systems,” Jour. of
Experimental and Theoretical AI, 12(2): 129-147.
[8] Barber K, Han D and Liu T 2000 Coordinating Distributed Design Making Using Reusable Interaction
Specifications
[9] Barber K, Liu T and Ramaswamy S 2001 “Conflict Detection During Plan Integration for Multi-Agent Systems,” IEEE Trans. on Systems, Man, and Cybernetics, 31(4): 616-628.
[10] Barber K, Liu T, Goel A and Martin C 1999 “Conflict Representation and Classification in a Domain-Independent Conflict Management Framework,” Proc. of the Third Inter.Conference on Autonomous Agents
1999, pp. 346-347.
[11] Barrett S and Stone P 2015 Cooperating with unknown teammates in complex domains: A robot soccer case
study of ad hoc teamwork. In Proceedings of the 29th AAAI Conference on Artificial Intelligence, pages
2010–2016.
[12] Bond A and Gasser L 2002 An analysis of problems and research in DAI. In A.H. Bond and L. Gasser,
editors, Readings in Distributed Artificial Intelligence, pages 3-35. Morgan Kaufmann, San Mateo, CA.
[13] Bowling M. and Veloso M.. Multiagent learning using a variable learning rate. Artificial Intelligence,
136(2):215–250.
[14] Brandolese A, Brun A and Portioli-Staudacher A A multi-agent approach for the capacity allocation problem, viewed on 18th November 2017 from https://doi.org/10.1016/S0925-5273(00)00004-9
[15] Busoniu L, Babuska R and Bart D 2008 A comprehensive survey of multiagent reinforce-ment learning.
IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 2(38):156–
172.
[16] Caruana R and Niculescu-Mizil A 2006 An Empirical Comparison of Supervised Learning Algorithms, Proc. 23rd Int'l Conf. Machine Learning (ICML '06), pp. 161-168.
[17] Chakraborty D and Stone P 2014 Multiagent learning in the presence of memory-bounded agents. Autonomous Agents and Multi-Agent Systems, 28(2):182–213.
5
The 3rd International Conference on Computing and Applied Informatics 2018
IOP Publishing
IOP Conf. Series: Journal of Physics: Conf. Series 1235 (2019) 012022 doi:10.1088/1742-6596/1235/1/012022
[18] Chen K, Lindsay A, Robinson P and Abbass H 2009 A hierarchical conflict resolution method for multiagent path planning. IEEE - Evolutionary Computation, pages 1169–1176.
[19] Christoforos M, Valts B and Ross K 2017 Socially Competent Navigation Planning by Deep Learning of
Multi-Agent Path Topologies., viewed on 29 November 2017 from http://bit.ly/2BFQM4e.
[20] Huang C, Ceroni A and Nof S 2000 Agility of Networked Enterprises - Parallelism, Error Recovery and
Conflict Resolution Computers in Industry, 42(2-3): 275-287.
6