AreviewofHopfieldNNforsolvingMPP PDF
AreviewofHopfieldNNforsolvingMPP PDF
AreviewofHopfieldNNforsolvingMPP PDF
Invited Review
Department of Industrial Engineering & Engineering Management, National Tsing Hua University, Hsinchu 30013, Taiwan, ROC
Graduate Institute of Management Sciences, Tamkang University, Tamsui, Taipei 25137, Taiwan, ROC
a r t i c l e
i n f o
Article history:
Received 1 October 2007
Accepted 4 November 2008
Available online xxxx
Keywords:
Hopeld neural networks
Energy function
Mathematical programming
penalty function
Lagrange multiplier
Primal and dual functions
a b s t r a c t
The Hopeld neural network (HNN) is one major neural network (NN) for solving optimization or mathematical programming (MP) problems. The major advantage of HNN is in its structure can be realized on
an electronic circuit, possibly on a VLSI (very large-scale integration) circuit, for an on-line solver with a
parallel-distributed process. The structure of HNN utilizes three common methods, penalty functions,
Lagrange multipliers, and primal and dual methods to construct an energy function. When the function
reaches a steady state, an approximate solution of the problem is obtained. Under the classes of these
methods, we further organize HNNs by three types of MP problems: linear, non-linear, and mixed-integer. The essentials of each method are also discussed in details. Some remarks for utilizing HNN and difculties are then addressed for the benet of successive investigations. Finally, conclusions are drawn
and directions for future study are provided.
2008 Elsevier B.V. All rights reserved.
1. Introduction
In the past two decades, articial neural networks or neural networks (NNs) have been widely developed in solving mathematical
programming (MP) or optimization problems. Among various NNs,
a couple of networks have been utilized for optimization (Looi,
1992; Kumar, 2005), such as Hopeld neural netwoksHNNs) (Hopeld, 1982), self-organizing feature maps (Kohonen, 1982), and
Boltzmann machines (Ackley et al., 1985). As HNNs and general
NN structure have been developed for almost all classes of optimization problems, it is worth reviewing their inside techniques, benets and drawbacks, and future perspectives so that the
presentation will enrich the theories and applications of HNNs in
the future.
Research studies on neurons and their networks for information
processing have drawn much attention by scholars in the elds of
physics, electronics, and electricity. They have tried to formulate
the problems mimicking the behavior of biological NNs to abstract
principles of computation employed by the brain. Hopeld and
Tank made a momentous contribution for neural computation by
solving a traveling salesman problem (TSP) and a linear programming (LP) problem (Hopeld, 1982, 1984; Hopeld and Tank,
1985; and Tank and Hopeld, 1986). Although Shirazi and Yih
* Corresponding author. Tel.: +886 3 574 2653; fax: +886 3 572 2204.
E-mail addresses: [email protected] (U.-P. Wen), [email protected]
(H.-S. Shih).
1
Tel.: +886 2 8631 3221; fax: +886 2 8631 3214.
0377-2217/$ - see front matter 2008 Elsevier B.V. All rights reserved.
doi:10.1016/j.ejor.2008.11.002
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
2
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
primal and dual method for solving linear and quadric programming problems. Nonetheless, the studies on NN on optimization
have appeared in diversied journals. Interested readers can refer
to Computers and Operations Research (No.34, Vol. 19, 1992), Neurocomputing (No. 13, Vol. 8, 1995), and European Journal of Operational Research (No. 2, Vol. 93, 1996). Some complete
bibliography of this literature can be taken from books and papers
(Klimasauskas, 1989; Looi, 1992; Sharda, 1994; Widrow et al.,
1994; Wilson and Sharda, 1992).
Because there is no systematic review among plenty of articles
on HNNs for optimization up until now, we intend to classify the
essences of these HNNs for optimization as a guideline of future
studies. Their special structures such as penalty function methods,
Lagrange multiplier function methods, and primal and dual methods are also illustrated. In particular, energy functions of the networks characterize the behavior of the networks and support the
directions to search out solutions for real-time optimization. The
functions are the key to bridging the HNNs and the constraints
optimization problems. On account of this signicant importance,
we review the literature based on three technical categories: penalty functions, Lagrange functions, and primal and dual functions.
The contents are organized in the following. Section 2 discusses
general functions and basic aspects of HNNs. The penalty function
methods, Lagrange multiplier methods, and primal and dual functions are overviewed in Sections 35, respectively. The uses of
HNNs for LP, NLP, and mixed-integer linear programming (MILP)
are presented by each method. In Section 6, remarks are made
for illustrating the possible difculties when utilizing HNNs, and nally, conclusions and directions for future studies are given.
2. General function and basic aspects for HNNs
Neural networks have been characterized in various ways
according to many relevant features (Basheer and Hajmeer, 2000;
Sima and Orponen, 2001). In general, NNs have two kinds of structures (Zurada, 1992). Both have to be congured such that the
application of a set of inputs produces (either direct or via a relaxation process) the desired set of outputs, and various methods exist
to set out the strengths of the connections. One way is to place the
weights explicitly, using a priori knowledge. The other way is to
train the neural network by feeding it teaching patterns and letting
it change its weights according to some learning rule. HNNs belong
to non-training model. In particular, HNNs are considered as feedback (or recurrent) without learning from a pattern set (Zurada,
1992). The HNNs consist of a set of neurons and a corresponding
set of unit delays, forming a multiple loop feedback system. Each
feedback loops is equal to a neuron. Basically, the output of each
neuron is feedback, via a unit delay element, to each of the other
neurons in the system.
HNNs are not affected by additional inputs, and each neuron is
fully connected with the remaining neurons by weights and always
updates the connection weights. The associative memory or content-addressable memory is a memory organization that accesses
memory by its contents and it always searches for the closest
matching from all stored prototypes (Du and Swamy, 2006). This
is analogous to the storage of information in an associative memory as biological neurons (Kohonen, 1982). In other words, the process of association and information retrieval is simulated by the
dynamical behavior of a highly interconnected system of non-linear circuit elements with collective computational abilities (Hopeld, 1982).
Due to the above characteristics, HNNs are easily realized for
their capability of parallel computation as a result of being a fully
connected property. The networks use electric circuits to simulate
the actions of biological neurons, and the basic model can be
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
3
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
(u1 )
(u2 )
(un )
v2
vn
v1
- j
x1 (k)
x2 (k)
( )
wi 1
wi 2
yj (k+1)
vj (k+1)
win
xn (k)
yj (k)
Z -1
EDHNN
n X
n
n
X
1X
xi wij xj
xi Ii
2 i1 j1
i1
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
4
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
ECHNN
n X
n
m
n Z xi
X
X
1X
1
xi wij xj
xi I i
g 1
i xi dx;
2 i1 j1
R
i
0
i1
i1
min f x c T x;
s:t: Ax P b g j x P bj ; j 1; 2; . . . ; m
and x P 0;
LPF
n1
mn
where x; c 2 R
; A2R
max gy b y;
m1
, and b 2 R
s:t: A y 6 c fi y 6 ci ; i 1; 2; . . . ; n
and y P 0;
DLPF
ETH x f x
m
n
X
X
g j x2
x2i =2sRii ;
j1
i1
T
and g g
1 ; g 2 ; . . . ; g m is a penalty vector when gj(x) < bj. Here,
the penalty parameter s must be sufciently large to lead the last
term to be neglected. However, this assumption might make their
model unreliable for solving the LP problem (Kennedy and Chua,
1988).
To improve Tank and Hopelds NN, Kennedy and Chua (1988)
proposed a kind of NN structure with an inexact penalty function,
where the energy function is
EKC x f x
m
s X
g x2 :
2 j1 j
m
X
ERV x; a f x a
minf0; g j x bj g;
j1
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
5
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
min f x;
s:t:
hi x 0 for i 1; 2; . . . ; l;
NLP
g j x 6 0 for j 1; 2; . . . ; m;
which is simpler and more intuitive for solving convex NLP problems. They used a constraint set instead of penalty parameters to
overcome the difculty of selection. Their work is valuable for
tackling NLP problems.
min
x 2 X;
s:t:
f x; v ;
hi x; v 0;
i 1; 2; . . . ; q;
g j x; v 6 0;
j 1; 2; . . . ; p;
ml 6 xl 6 M l ;
v k 2 f0; 1g;
MILP
l 1; 2; . . . ; n;
k 1; 2; . . . ; r;
where f, g1, . . . , gp, h1, . . . , hq are functions dened on Enr, x is a vector of n components x1, . . . , xn which satisfy boundaries between ml
and Ml, and v is the 01 integer vector of components, v1, . . . , vr.
Since, the problem with integer variables is difcult to solve
by an approximate technique, its development is rather slow.
Watta and Hassoun (1996) used a coupled gradient network approach for solving a MILP problem. Their energy function
EWH(x, v) is constructed by summing the objective function
f(x, v) and the penalty function P(x, v) to the constraints. The dynamic update is carried out by the steepest gradient descent
with a xed increment.
We can in general separate the constraints into inequality equations, equality equations, and bounded variables constraints as
P(x, v) = G(x, v) + H(x, v) + V(v). The rst penalty term of the rightP
hand side is Gx; v pj1 #g j x; v , where #(n) = n2 for all n > 0,
and #(n) = 0 for all n 6 0. If the solution is unsatisfactory with the
inequality constraints, then the penalty will be effective. The secP
2
ond term Hx; v qi1 hi x; v enforces the equality constraint.
Pr
The last term Vv k1 v k 1 v k is to keep the 01 variables
feasible. Note that the larger value of penalty functions is 0.5 and
falls to zero as it saturates at vi = 0 or vi = 1. They thus recommend
an energy function as
EWH x; v A f x; v B
p
X
#g j x; v C
j1
r
X
q
X
hi x; v
i1
v k 1 v k :
k1
Table 1
Penalty function methods for optimization problems.
Proposed method
Problem
type
Activation function
Penalty
Learning rate
Initial state
LP
Sigmoid function
N/A
LP and
NLP
LP and
NLP
LP and
NLP
MILP
Sigmoid function
C is a positive
diagonal matrix
C is a positive
diagonal matrix
0 < li < 1
N/A
Rodrguez-Vzquez et al.
(1988, 1990)
Maa and Shanblatt (1992)
Watta and Hassoun
(1996)
Aourid and Kaminska
(1996)
Chong et al. (1999)
Silva et al. (2005)
Effati and Baymain (2005)
xi(k + 1) = max{xi(k), 0}
Sigmoid function
N/A
Feasible variables
Sigmoid function
A, B, C, D > 0
MILP
Sigmoid function
N/A
LP
NLP
NLP
Sigmoid function
Sigmoid function
N/A
N/A
N/A
N/A
Feasible variables
Feasible variables
Feasible variables
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
6
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
method for solving LP problems through HNNs. Their energy functions by Lagrange and augment Lagrange multipliers are listed as
EL x; k f x
m
X
kg j x
j1
and
EaL x; k; K f x
m
X
kg j x
j1
m
1X
Kjg j xj2 ;
2 j1
1
K
ES x; k cT x kT Ax b kT k Ax bT Ax b:
2
2
Table 2
Lagrange multiplier related methods for optimization problems.
Proposed method
Problem
type
Activation function
Penalty
Learning rate
Initial state
LP and
NLP
NLP
N/A
0<c<5
N/A
N/A
N/A
qt 1qa0 t or
NLP
MINLP
Sigmoid function
A hard-limit function for discrete
variables, and a sigmoid function
for continuous variables
k>0
bj > 0 and ri is a scaling
factor known as the slope in
the continuous neuron i
q(t) = q0 expat q0
is large, and a is a positive for
Lagrange multiplier vector; l > 0
for variables
N/A
0 < lc, lk < 1
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
7
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
faL x; k; K f x
m
X
k maxf0; g j xg
j1
m
X
K
maxf0; g j xg2 ;
2
j1
10
where k and K are the vectors of Lagrange multipliers and the penalty parameters with 2 Rm1 , respectively. A compact form can be
expressed as (Ham and Kostanic, 2001):
faL x; k; K f x
m
X
Sj kg j x K=2g j x2 ;
11
j1
min
s:t:
ron cij and Idij for discrete neuron dij, respectively. The matrix T
is given a standard interconnection among all neurons, continuous
and discrete. Here, Tdij?ckm is the connection from the discrete neuron dij to the continuous neuron ckm. However, a new form of
interconnection, matrix W, between neuron pairs is introduced.
For example, Wij?km is the connection from neuron pair cij and
dij to neuron pair ckm and dkm. Their energy function of the augmented HNN is then proposed as
Ewalsh
1 XX
T ckm!cij V cij V ckm
2 i;j k;m
1 XX
T dkm!dij V dij V dkm
2 i;j k;m
X
i;j
XX
i;j
!
k;m
Icij V cij
i;j
!
1 XX
Idij V dij
W km!ij V dij V cij V dkm V ckm :
2 i;j k;m
12
It is proven that Eq. (12) should be (local) minimized for both the
continuous variables output Vcij and the discrete variables output
Vdij. Since the activation function of the discrete neuron is a hardlimit function, it will easily drop in a local optimum. If the discrete
variables are selected to be zero, then the binary variables are not
modied for later iterations. Moreover, the continuous activation
function takes advantage of a general sigmoid function tanh (output = (einput einput)/(einput + einput)), whose output is between 1 and
1. Thus, the continuous variables do not stably converge to the
real variables which are greater than 1 or less than 1.
Dillon and OMalley (2002) proposed another augmented HNN to
solve MINLP problems. Their structure requires many variables and
parameters, and it will adds much computational burden. Hence, the
proposed network may be inappropriate for applications.
4.4. Summary
After surveying the above developments, we nd that the
parameter setting for Lagrange multipliers and penalties as well
as the learning rates have major inuences on the convergence
and stability of HNNs. Lagrange multiplier and augmented Lagrange multiplier methods require more variables and parameters
than do the penalty function methods. This increases the computational complexity and is difcult to control, but its structure can
provide a precise solution. Furthermore, we also collect the characteristics of this category of approach for NNs in terms of problem
type, activation function, penalty, learning rate, and initial state
as shown in Table 2, so as to understand in what way the key parts
are involved.
f x; v
hi x; v 0;
xl P 0;
i 1; 2; . . . ; q;
l 1; 2; . . . ; n;
v k 2 f0; 1g;
MIN LP
k 1; 2; . . . ; r;
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
8
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
eters. They suggested that the equilibrium point is the same as the
exact solution when simultaneously solving the primal problem
and its dual problem. Hence, they dened an energy function for
solving LP problems as Exw(x, y):
EXW x; y
1 T
1
1
1
T
c x b y2 xT x jxj yT y jyj kAT x
2
2
2
2
1 T
2
T
T
bk2 A y cA y c jA y cj:
13
2
min
n X
n
X
i1
n
X
cij xij ;
n
X
s:t:
j1
8j 1; 2; . . . ; n;
xij 1;
i1
min
n
n
X
X
s:t:
8
if i 1
>
< 1;
xik
xli 0;
if i1 & in
>
:
k1;ki
l1;li
1; if i n;
n
X
n
X
max yn y1 ;
s:t:
max
ui
i1
max zn ;
s:t:
Dual SPP2
8i; j 1; 2; . . . ; n:
zj zi 6 cij ; ij;
#2
n
X
xX
Ew98 t; xt
i1
ki
n1 X
X
cij expt=sxij t;
16
i1 j2;ji
v j;
s:t: ui v j 6 cij
8i; j 1; 2; . . . n:
j1
DualAP
Here, cij and xij are the respective cost and decision variables, associated with assigning entity i to position j. Here, xij = 1 if entity i is
assigned to position j, and ui and vi are dened as the dual decision
variables. The energy function of the primal assignment problem is
Ew97 t; xt
Dual SPP1
8i; j 1; 2; . . . ; n:
yj yi 6 cij ; ij;
AP
n
X
SPP
ij; i; j 1; 2; . . . ; n:
j1
n
X
cij xij
i1 j1;ji
"
8i 1; 2; . . . ; n;
xij 1;
8
"
n
n
X
x <X
2 : i1
#2
xij t 1
j1
a exp
"
n
n
X
X
j1
n X
n
t X
i1
xij t 1
i1
cij xij t;
n X
n
xX
i1
Ew98d t; zt
#2 9
=
n X
xX
i1
ji
17
;
14
j1
where x, a, and s are positive constants, and a exp(t/s) is a temperature parameter whose role is to balance the effect of cost minimization and constraint satisfaction (Wang, 1993). On the other
hand, the form of the dual AP energy function is formulated as
Ew97d t; ut; v t
j1
b expt=s
n
X
ui t v i t;
15
i1
where b > 0, g() is a non-negative and non-decreasing activation function given by w(s) = 0, if s 6 0 and w(s) > 0, if s > 0.
min
f q x
1 T
x Ax ax;
2
s:t: Dx b and x P 0;
QP
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
max
1
T
f q y b y xT Ax;
2
s:t: DT y Ax a 6 0;
DQP
Dx b;
DT y Ax a 6 0;
xDT y Ax a 0:
QP-DOP
ECHNN
TSP
A=2
XXX
X
C=2
ji
XX
X
V Xi V Xj B=2
XXX
i
!2
V Xi n
D=2
dXY V Xi V Y;i1 V Y;i1 :
V Xi V Yi
XY
XX
X
YX
18
Here, A, B, C, and D are positive. The four terms in Eq. (18) are
dened as: (i) term 1 is zero if and only if each city row X
contains no more than one 1, i.e., the rest of the entries being
zero; (ii) term 2 is zero if and only if each position in tour
column contains no more than one 1, i.e., the rest of the entries being zero; (iii) term 3 is zero if and only if there are
n entries of one in the entire matrix; and (iv) term 4 describes
lower values corresponding to shorter tours. The rst three
terms describe the feasibility requirements and result in a valid
tour by zero (Hopeld and Tank, 1985). The last term represents the objective function of TSP. In addition, the four parameters A, B, C, and D must be chosen by trial and error for a
particular problem. Various results illustrate different strategies
for the actual implementation of HNNs, such as a sigmoid function with the value in the range (0, 1). Eq. (18) can be used to
derive the following expression for the strength of connection
between pairs of neurons Xi and Yj. Since the time for the system to converge is an arbitrary unit, the value of the time constant of the ampliers in the system, s, can be set to 1.0, and
gain u0 is a scalar factor.
Although HNN is the most popular model of NNs for optimization, it suffers from some drawbacks: (i) the minimal solution does
not guarantee global optimization; (ii) the articial variables A, B, C
and D are not easily implemented; and (iii) the energy function
may pose many unnecessary local solutions which are difcult to
avoid (Takahashi, 1998). Until now, many researchers have tried
to improve the drawbacks of HNNs, and we briey show offer
attention so that future studies could follow.
Wilson and Pawley (1988) rst focused on the effectiveness of
computation for the HNN dealing with TSPs. They examined the
original model of Hopeld and Tank (1985) through six aspects.
First, the parameters, i.e., penalty parameters, city sizes, and gain
u0, are rather sensitive to the operating points, as changing by as
little as 10% of their original values can prevent the system from
converging. Second, the input current is at best in randomization.
Third, the distance setting (for the cities) must not be the same
to make reorganization easily. Fourth, the xed city positions in
a tour can reduce computational complexity, and the quality of
the paths obtained is improved signicantly. Fifth the algorithm
of Durbin and Willshaw (1987) provides global information on
its initialization. Their idea introduces a bias into the initial values
to reect the cities located on opposite sides of the unit square that
are likely to be on opposite sides of a tour. Finally, the weights of an
inter-neural connection must be within a range of values, instead
of the same value initially, in order to avoid degeneracy in the
hardware implementation. After the modications, the HNNs can
reliably solve small-size TSPs (Wilson and Pawley, 1988).
According to our survey, continuous HNNs have four deciencies that need to be improved for solving large-size problems
(Takahashi, 1998). In the following four sections, we examine
these issues and try to provide some directions for future
research.
6.1. Parameter setting and the initial state
How to set the parameters and choose an initial state are the
two challenges in utilizing HNNs for beginners. We rst list these
two parts in our remarks.
6.1.1. How to set parameters
HNN could lead to an infeasible solution nally if the parameters are inadequately set for computation. In fact, its energy function often causes infeasible solutions in solving TSPs (Kennedy and
Chua, 1988). The investigation on the inter-relationship among the
parameters shows that TSPs do not have a scaling property and
only a small range of parameter combinations result in the stable
solution (Kennedy and Chua, 1988). In addition, HNNs commonly
take advantages of gradient and Lagrange multiplier methods for
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
10
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
Table 3
Primal and dual methods for optimization problems.
Proposed method
Problem type
Activation function
Penalty
Learning
rate
Initial state
LP
N/A
N/A
N/A
LP and NLP
N/A
N/A
N/A
Feasible
variables
Feasible
variables
Assignment problem
(network ows)
N/A
Feasible
variables
N/A
Feasible
variables
Feasible
variables
Any initial
state
Wang (1997)
Wang (1998)
LP
N/A
LP and NLP
N/A
N/A
g, g1, g2 > 0
N/A
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
(Sexton et al., 1998, 1999). This direction can help HNNs to solve
large-scale problems in the future.
6.4. Computational efciency
A noteworthy feature of the HNNs is their ability for massive
parallel computation (da Silva et al., 2006; Hopeld and Tank,
1985; Li, 1996; Lin et al., 2000; Pavlik, 1992; Wang, 1993). However, their complex structure for parallel computation may lead
to difculties in neuron updating (Looi, 1992; Smith, 1999). Therefore, how to design an effective parallel process is an important issue in the future.
Along with the trend of simulation procedure for solving problems on digital computers, Mida-Casermeiro et al. (2001) proposed a new concept for updating units in a discrete HNN for
solving TSPs. Two neuron updating rules, the one-path updating
rule and the two-path updating rule, which are similar to two-optimal and three-optimal tour improvement heuristics are respectively supplied (Croes, 1958). The computational results with city
numbers from 51 to 532 are demonstrated, and a direction to improve the computational efciency is also discussed. Munehisa
et al. (2001) presented another HNN structure for updating the
weights for a number of neurons. Four test problems from the
TSP library are examined by the proposed and show a rather good
performance (Reinelt, 1991). However, larger problems take up
signicant computational time by soft computing.
Hardware computing, on the other hand, is much faster than
soft computing and is the noteworthy advantage of HNNs. It needs
more elements to be implemented in the hardware for large-sized
problems. Moreover, the development of VLSI circuits has improved so tremendously that implementation will not be a problem later if the material problem can be overcome.
The HNN structure causes a logistic difculty (solutions
with several local minima) in computing (Gee and Prager,
1995). For this topic, Takahashi (1999) provided two structures
of the specic problems which are related to the formulation.
One is how to map a real world problem onto a discrete
HNN problem (Takahashi, 1999), and the other is how to organize a mathematical formulation to a continuous HNN problem
(Takahashi, 1999). The rst problem gives hope for future
applications of HNNs if the transformation can be interpreted
adequately. Although many researchers have been successful
in illustrating their HNNs to solve real problems such as
assignment problems, scheduling problems, shortest path problems, TSPs, and vehicle routing problems (Arajo et al., 2001;
Nygard et al., 1990; Gee and Prager, 1995; Smith et al.,
1996; Wang, 1996), more and more applications are expected
in the future. The second problem is difcult to be conquered
since it still lacks a general network, guaranteeing an optimal
solution for a general MP problem. Our review collects and
organizes a series of methods for solving the special cases of
MP problems. Some connections among these cases and discussed drawbacks might be helpful for blossoming more general cases in the future.
We have posted four issues in dealing with the use of HNNs for
optimization. These contents are practical when utilizing the networks, but we only discuss how to formulate the problems and
then solve the problems through mathematical software on a digital computer. In this paper we have not implemented HNNs by
hardware. This might be another story for later research.
7. Conclusions and future works
After reviewing the HNNs for optimization in the area of operations research, we can see that the networks solve various MP
problems. After respectively classifying the techniques of NNs,
11
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
12
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
(1) Stable equilibrium:If the Lyapunov function V is locally positive denite and the time derivative of the Lyapunov func_
tion is locally negative semi-denite, where Vx
6 0;
8x 2 S for some neighborhood space S, then the equilibrium
is proven to be stable.
(2) Locally asymptotically stable equilibrium:If the Lyapunov
function V is locally positive denite and the time derivative
of the Lyapunov function is locally negative denite, where
_
Vx
6 0; 8x 2 S n f0g for some neighborhood space S, then
the equilibrium is proven to be locally asymptotically stable.
(3) Globally asymptotically stable equilibrium:If the Lyapunov
function V is globally positive denite and the time derivative of the Lyapunov function is globally negative denite,
_
where Vx
6 0 8x 2 Rn n f0g, then the equilibrium is proven
to be globally asymptotically stable.
In many problems of the HNN, the energy function can be considered as a Lyapunov function. However, the existence of a Lyapunov function is sufcient but not necessary for stability. The
Lyapunov function V(x) provides the mathematical basis for
the global stability analysis of the non-linear dynamical system.
The global stability analysis is much more powerful in its conclusion than local stability analysis that is, every globally stable
system is also locally stable, but not vice versa.
References
Abe, S., 1993. Global convergence and suppression of spurious states of the Hopeld
neural networks. IEEE Transactions on Circuits Systems 40, 246257.
Ackley, D.H., Hinton, G.E., Sejnowski, T.J., 1985. A learning algorithm for Boltzmann
machines. Cognitive Science 9, 147169.
Aiyer, S.V.B., Niranjan, M., Fallside, F., 1990. A theoretical investigation into the
performance of the Hopeld model. IEEE Transactions on Neural Networks 1,
204215.
Aourid, M., Kaminska, B., 1996. Minimization of the 01 linear programming
problem under linear constraints by using neural networks: Synthesis and
analysis. IEEE Transactions on Circuits and Systems I: Fundamental Theory
and Applications 43, 421425.
Arajo, F., Rebeiro, B., Rodrigues, L., 2001. A neural network for shortest path
computation. IEEE Transactions on Neural Network 12, 10671073.
Basheer, I.A., Hajmeer, M., 2000. Articial neural networks: Fundamentals,
computing, design, and application. Journal of Microbiological Methods 43, 3
31.
Bazaraa, M.S., Jarvis, J.J., Sherali, H.D., 2005. Programming and Network Flows, third
ed. John-Wiley, New York.
Bazaraa, M.S., Sherali, H.D., Shetty, C.M., 2006. Nonlinear Programming Theory and
Algorithms, third ed. John-Wiley, New York.
Bouzerdoum, A., Pattison, T.R., 1993. Neural network for quadratic optimization
with bound constraints. IEEE Transactions on Neural Networks 4, 293304.
Bruck, J., San, J., 1988. A study on neural networks. International Journal of
Intelligent Systems 3, 5975.
Burke, L.I., Ignizio, J.P., 1992. Neural network and operations research: An overview.
Computers and Operations Research 19, 179189.
Chen, L., Aihara, K., 1995. Chaotic simulated annealing by a neural network model
with transient chaos. Neural Networks 8, 915930.
_
Chong, E.K.P., Hui, S., Zak,
S.H., 1999. An analysis of a class of neural networks for
solving linear programming problems. IEEE Transactions on Automatic Control
44, 19952006.
Cichocki, A., Unbehauen, R., 1993. Neural Networks for Optimization and Signal
Processing. John Wiley, New York.
Croes, G.A., 1958. A method for solving traveling salesman problems. Operations
Research 6, 791812.
da Silva, I.N., Amaral, W.C., Arruda, L.V.R., 2006. Neural approach for solving several
types of optimization problems. Journal of Optimization Theory and
Applications 128, 563580.
Dillon, J.D., OMalley, M.J., 2002. A Lagrangian augmented Hopeld network for
mixed integer non-linear programming problems. Neurocomputing 42, 323
330.
Du, K.L., Swamy, M.N.S., 2006. Neural Networks in a Softcomputing Framework.
Springer, London, UK.
Durbin, R., Willshaw, D., 1987. An analogue approach to the traveling salesman
problem using an elastic net method. Nature 326, 689691.
Effati, S., Baymain, M., 2005. A new nonlinear neural network for solving convex
nonlinear programming problems. Applied Mathematics and Computation 168,
13701379.
El-Bouri, A., Balakrishnan, S., Popplewell, N., 2005. A neural network to enhance
local search in the permutation owshop. Computers and Industrial
Engineering 49, 182196.
Gee, A.H., Prager, R.W., 1995. Limitations of neural networks for solving travelings
salesman problems. IEEE Transactions on Neural Networks 6, 280282.
Gill, P.E., Murray, W., Wright, M.H., 1981. Practical Optimization. Academic, London.
Gong, D., Gen, M., Yamazaki, G., Xu, W., 1997. Lagrangian ANN for convex
programming with linear constraints. Computers and Industrial Engineering
32, 429443.
Hagan, M.T., Demuth, H.B., Beale, M., 1996. Neural Network Design. PWS Publishing
Co, Massachusetts.
Ham, F.M., Kostanic, I., 2001. Principles of Neurocomputing for Science and
Engineering. McGraw-Hill, New York.
Hasam, G.-O., Nezam, M.-A., 2006. An efcient simplied neural network for solving
linear and quadratic programming problems. Applied Mathematics and
Computation 175, 452464.
Haykin, S., 1999. Neural Networks: A Comprehensive Foundation, second ed.
Prentice-Hall, New Jersey.
Hopeld, J.J., 1982. Neural networks and physical systems with emergent collective
computational abilities. Proceedings of the National Academy of Sciences 79,
25542558.
Hopeld, J.J., 1984. Neurons with graded response have collective computational
properties like those of two-state neurons. Proceedings of the National
Academy of Sciences 81, 30883092.
Hopeld, J.J., 1988. Articial neural networks. IEEE Circuits and Devices Magazine
September, 310.
Hopeld, J.J., Tank, D.W., 1985. Neural computation of decisions in optimization
problems. Biological Cybernetics 52, 141152.
Kamgar-Parsi, Behzad, Kamgar-Parsi, Behrooz, 1990. On problem solving with
Hopeld neural networks. Biological Cybernet 62, 415423.
Kennedy, M.P., Chua, L.O., 1988. Neural networks for nonlinear programming. IEEE
Transportation Circuits System 35, 554562.
Klimasauskas, C.C., 1989. Neural networks: A new technology for information
processing. Data Base 20, 2123.
Kohonen, T., 1982. Self-organized formation of topologically correct feature maps.
Biological Cybernetics 43, 5969.
Kumar, S., 2005. Neural Networks: A Classroom Approach. International Edition,
McGraw-Hill, Singapore.
Kurita, N., Funahashi, K., 1996. On the Hopeld neural networks and mean eld
theory. Neural Networks 9, 15311540.
Lan, K.M., Wen, U.P., Shih, H.S., Lee, E.S., 2007. A hybrid neural network approach to
bilevel programming. Applied Mathematics Letters 20 (8), 880884.
Li, S.Z., 1996. Improving convergence and solution quality of Hopeld-type neural
networks with augmented Lagrange multipliers. IEEE Transactions on Neural
Networks 7, 15071516.
Lin, C.L., Lai, C.C., Huang, T.H., 2000. A neural network for linear matrix inequality
problems. IEEE Transactions on Neural Networks 11, 10781092.
Looi, C.K., 1992. Neural network methods in combinatorial optimization. Computers
and Operations Research 19 (3/4), 191208.
Maa, C.Y., Shanblatt, M.A., 1992. Linear and quadratic programming neural network
analysis. IEEE Transaction on Neural Networks 3, 380394.
Malek, A., Tari, A., 2005. Primaldual solution for the linear programming problems
using neural networks. Applied Mathematics and Computation 167, 198211.
Martn-Valdivia, M., Ruiz-Seplveda, A., Triguero-Ruiz, F., 2000. Improving local
minima of Hopeld networks with augmented Lagrange multipliers for large
scale TSPs. Neural Networks 13, 283285.
Mida-Casermeiro, E., Galn-Marn, G., Muoz-Perz, J., 2001. An efcient
multivalued Hopeld network for the traveling salesman problem. Neural
Processing Letters 14, 203216.
Munehisa, T., Kobayashi, M., Yamazaki, H., 2001. Cooperative updating in the
Hopeld model. IEEE Transactions on Neural Networks 12, 12431251.
Nygard, K.E., Jueli, P., Kadaba, N., 1990. Neural networks for selecting vehicle
routing heuristic. ORSA Journal of Computing 2, 353364.
Papageorgiou, G., Likas, A., Stafylopatis, A., 1998. Improved exploration in Hopeld
network state-space through parameter perturbation driven by simulated
annealing. European Journal of Operations Research 108, 283292.
Pavlik, P., 1992. Parallel Hopeld machine. Neurocomputing 4, 8991.
Peng, M.K., Gupta, N.K., Armitage, A.F., 1996. An investigation into the improvement
of local minima of the Hopeld network. Neural Networks 9, 12411253.
Qu, H., Yi, Z., Tang, H., 2006. Improving local minima of columnar
competitive model for TSPs. IEEE Transactions on Circuit and Systems I
53, 13531362.
Reinelt, G., 1991. TSPLIB A traveling salesman problem library. ORSA Journal on
Computing 3, 376384.
Ricanek, K., Lebby, G.L., Haywood, K., 1999. Hopeld like networks for pattern
recognition with application to face recognition. In: International Joint
Conference on Neural Network, vol. 5, pp. 32653269.
nguez-Castro, R., Rueda, A., Huertas, J.L., SnchezRodrguez-Vzquez, A., Dom
Sinencio, E., 1988. Switched-capacitor neural networks for linear programming.
Electronics Letters 24, 496498.
nguez-Castro, R., Rueda, A., Huertas, J.L., SnchezRodrguez-Vzquez, A., Dom
Sinencio, E., 1990. Nonlinear switched-capacitor neural networks for
optimization problems. IEEE Transactions on Circuits Systems 37, 384397.
Sexton, R.S., Aldaee, B., Dorsey, R.E., Johnson, J.D., 1998. Global optimization for
articial neural networks: A tabu search application. European Journal of
Operational Research 106, 570584.
Sexton, R.S., Dorsey, R.E., Johnson, J.D., 1999. Optimization of neural networks: A
comparative analysis of the genetic algorithm and simulated annealing.
European Journal of Operational Research 114, 589601.
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002
ARTICLE IN PRESS
U.-P. Wen et al. / European Journal of Operational Research xxx (2008) xxxxxx
Sharda, R., 1994. Neural networks for the MS/OR analyst: An application
bibliography. Interfaces 24, 116130.
Shih, H.S., Wen, U.P., Lee, E.S., Lan, K.M., Hsiao, H.C., 2004. A neural network
approach to multiobjective and multilevel programming problems. Computers
and Mathematics with Applications 48, 95108.
Shirazi, B., Yih, S., 1989. Critical analysis of applying Hopeld neural net model to
optimization problems. In: IEEE International Conference on Systems, Man and
Cybernetics, vol. 1, 1417 November, pp. 210215.
Silva, I.N., Amaral, W.C., Arruda, L.V., 2005. Design and analysis of an efcient neural
network model for solving nonlinear optimization problems. International
Journal of Systems Science 36, 833843.
Sima, J., Orponen, P., 2001. A Computational Taxonomy and Survey of Neural
Network
Models.
Available
from:
<http://citeseer.ist.psu.edu/
sima01computational.html>.
Smith, K.A., 1999. Neural networks for combinatorial optimization: A review on
more than a decade of research. INFORMS Journal on Computing 11, 15
34.
Smith, K.A., Gupta, J.N.D., 2000. Neural networks in business: Techniques and
applications for the operations research. Computers and Operations Research
27, 10231044.
Smith, K., Palaniswami, M., Krishnamoorthy, M., 1996. Traditional heuristic versus
Hopeld neural network approaches to a car sequencing problem. European
Journal of Operational Research 93, 300316.
Smith, K., Palaniswami, M., Krishnamoorthy, M., 1998. Neural techniques for
combinatorial optimization with applications. IEEE Transactions on Neural
Networks 9, 13011318.
Takahashi, Y., 1998. Mathematical improvement of the Hopeld model for TSP
feasible solutions by synapse dynamic systems. IEEE Transactions on Systems,
Man, and Cybernetics, Part B 28, 906919.
Takahashi, Y., 1999. A neural network theory for constrained optimization.
Neurocomputing 24, 117161.
Talavn, P.M., Yez, J., 2002. Parameter setting of the Hopeld network applied to
TSP. Neural Networks 15, 363373.
Tan, K.C., Tang, H., Ge, S.S., 2005. On parameter settings of Hopeld networks
applied to traveling salesman problems. IEEE Transactions on Circuits and
Systems I 52, 9941002.
Tank, D.W., Hopeld, J.J., 1986. Simple neural optimization networks: An A/D
converter, signal decision network and a linear programming circuit. IEEE
Transactions on Circuits Systems 33, 533541.
Walsh, M.P., Flynn, M.E., OMalley, M.J., 1998. Augmented Hopeld network for
mixed integer programming. IEEE Transactions on Neural Networks 10, 456
458.
Wang, J., 1992. Analogue neural network for solving the assignment problem.
Electronics Letters 28, 10471050.
Wang, J., 1993. Analysis and design of a recurrent neural network for linear
programming. IEEE Transactions on Circuits and Systems I: Fundamental
Theory and Applications 40, 613618.
13
Wang, J., 1996. A recurrent neural network for solving the shortest path problem.
IEEE Transactions on Circuits and Systems I: Fundamental Theory and
Applications 43, 482486.
Wang, J., 1997. Primal and dual assignment networks. IEEE Transactions on Neural
Networks 8, 784790.
Wang, J., 1998. Primal and dual neural networks for shortest-path routing. IEEE
Transactions on Systems, Man, and Cybernetics Part A: Systems and Humans
28, 864869.
Watta, P.B., Hassoun, M.H., 1996. A coupled gradient network approach for static
and temporal mixed-integer optimization. IEEE Transactions on Neural
Networks 7, 578593.
Widrow, B., Rumelhart, D.E., Lehr, M.A., 1994. Neural networks: Applications in
industry, business and science. Communication of the ACM 37, 93105.
Wilson, G.V., Pawley, G.S., 1988. On the stability of the traveling salesman problem
algorithm of Hopeld and Tank. Biological Cybernetics 58, 6370.
Wilson, R.L., Sharda, R., 1992. Neural networks. OR/MS Today 19, 3642.
Wong, B.K., Lai, V.S., Lam, J., 2000. A bibliography of neural networks in business:
Techniques and applications research: 19941998. Computers and Operations
Research 27, 10451076.
Wu, A.I., Tam, P.K.S., 1999. A neural network methodology and strategy of quadratic
optimization. Neural Computing and Applications 8, 283289.
Wu, X.Y., Xia, Y.S., Li, J., Chen, W.K., 1996. A high-performance neural network for
solving linear and quadratic programming problems. IEEE Transactions on
Neural Network 7, 643651.
Xia, Y., 1996. A new neural network for solving linear programming and its
application. IEEE Transactions on Neural Networks 7, 525529.
Xia, Y., Wang, J., 1995. Neural network for solving linear programming problem
with bound variables. IEEE Transactions on Neural Networks 6, 515519.
Xia, Y., Wang, J., 1998. A general methodology for designing globally convergent
optimization neural networks. IEEE Transactions on Neural Networks 9, 1331
1334.
Xia, Y., Feng, G., Wang, J., 2005. A primaldual neural network for online resolving
constrained kinematic redundancy in robot motion control. IEEE Transactions
on Systems Man and Cybernetics Part B Cybernetics 35, 5464.
Yalcinoz, T., Short, M.J., 1997. Large-scale economic dispatch using an improved
Hopeld neural network. IEE Proceedings Generation, Transmission and
Distribution 144, 181185.
Zak, S.H., Upatising, V., Hui, S., 1995. Solving linear programming problems with
neural networks: A comparative study. IEEE Transactions on Neural Networks 6,
94104. The MatheWorks: <http://www.mathworks.com/>.
Zhang, X.S., 2000. Neural Networks in Optimization. Kluwer Academic, London.
Zhang, S., Constantinides, A.G., 1992. Lagrange programming neural networks. IEEE
Transactions on Circuits Systems 39, 441452.
Zhang, H.C., Huang, S.H., 1995. Applications of neural networks in manufacturing: A
state-of-the-art survey. International Journal of Production Research 33, 705
728.
Zurada, J.M., 1992. Introduction to Articial Neural Systems. West, New York.
Please cite this article in press as: Wen, U.-P. et al., A review of Hopeld neural networks for solving mathematical ..., European Journal of
Operational Research (2008), doi:10.1016/j.ejor.2008.11.002