neural network to solve traveling salesman problem
DESCRIPTION
Neural Network to solve Traveling Salesman Problem. Amit Goyal 01005009 Koustubh Vachhani 01005021 Ankur Jain 01D05007. Roadmap. Hopfield Neural Network Solving TSP using Hopfield Network Modification of Hopfield Neural Network - PowerPoint PPT PresentationTRANSCRIPT
Neural Network to solve Traveling Salesman Problem
Amit Goyal 01005009 Koustubh Vachhani 01005021Ankur Jain 01D05007
Roadmap Hopfield Neural Network
Solving TSP using Hopfield Network
Modification of Hopfield Neural Network
Solving TSP using Concurrent Neural Network
Comparison between Neural Network and SOM for solving TSP
Background
Neural Networks Computing device composed of
processing elements called neurons Processing power comes from
interconnection between neurons Various models are Hopfield, Back
propagation, Perceptron, Kohonen Net etc
Associative memory
Associative memory Produces for any input pattern a
similar stored pattern Retrieval by part of data Noisy input can be also recognized
Original Degraded Reconstruction
Hopfield Network Recurrent network
Feedback from output to input
Fully connected Every neuron connected
to every other neuron
Hopfield Network
Symmetric connections Connection weights from unit i to
unit j and from unit j to unit i are identical for all i and j
No self connection, so weight matrix is 0-diagonal and symmetric
Logic levels are +1 and -1
Computation
For any neuron i, at an instant t input is
Σj = 1 to n, j≠i wij σj(t)
σj(t) is the activation of the jth neuron
Threshold function θ = 0
Activation σi(t+1)=sgn(Σj=1 to n, j≠i wijσj(t))
where
Sgn(x) = +1 x>0Sgn(x) = -1 x<0
Modes of operation Synchronous
All neurons are updated simultaneously
Asynchronous Simple : Only one unit is randomly selected
at each step
General : Neurons update themselves independently and randomly based on probability distribution over time.
Stability
Issue of stability arises since there is a feedback in Hopfield network
May lead to fixed point, limit cycle or chaos Fixed point : unique point attractor Limit cycles : state space repeats
itself in periodic cycles Chaotic : aperiodic strange attractor
Procedure
Store and stabilize the vector which has to be part of memory.
Find the value of weight wij, for all i, j such that :<σ1, σ2, σ3 …… σN> is stable in Hopfield
Network of N neurons.
Weight learning Weight learning is given by
wij = 1/(N-1) σi σj
1/(N-1) is Normalizing factor σi σj derives from Hebb’s rule
If two connected neurons are ON then weight of the connection is such that mutual excitation is sustained.
Similarly, if two neurons inhibit each other then the connection should sustain the mutual inhibition.
Multiple Vectors If multiple vectors need to be
stored in memory like<σ1
1, σ21, σ3
1 …… σN1>
<σ12, σ2
2, σ32 …… σN
2>……………………………….<σ1
p, σ2p, σ3
p …… σNp>
Then the weight are given by:wij = 1/(N-1) Σm=1 to pσi
m σj
m
Energy
Energy is associated with the state of the system.
Some patterns need to be made stable this corresponds to minimum energy state of the system.
Energy function
Energy at state σ’ = <σ1, σ2, σ3 ……
σN> E(σ’) = -½ Σi Σj≠i wij σiσj
Let the pth neuron change its state from σp
initial to σp
final so Einitial = -½ Σj≠p wpj σp
initial σj + T Efinal = -½ Σj≠p wpj σp
final σj + T ΔE = Efinal – Einitial
T is independent of σp
Continued…ΔE = - ½ (σp
final - σpinitial ) Σj≠p wpj σj
i.e. ΔE = -½ Δσp Σj≠p wpj σj
Thus: ΔE = -½ Δσp x (netinputp)
If p changes from +1 to -1 then Δσp is negative and netinputp is negative and vice versa.
So, ΔE is always negative. Thus energy always decreases when neuron changes state.
Applications of Hopfield Nets
Hopfield nets are applied for Optimization problems.
Optimization problems maximize or minimize a function.
In Hopfield Network the energy gets minimized.
Traveling Salesman Problem
Given a set of cities and the distances between them, determine the shortest closed path passing through all the cities exactly once.
Traveling Salesman Problem One of the classic and highly researched
problem in the field of computer science.
Decision problem “Is there a tour with length less than k" is NP - Complete
Optimization problem “What is the shortest tour?” is NP - Hard
Hopfield Net for TSP N cities are
represented by an N X N matrix of neurons
Each row has exactly one 1
Each column has exactly one 1
Matrix has exactly N 1’s
σkj = 1 if city k is in position j
σkj = 0 otherwise
Hopfield Net for TSP
For each element of the matrix take a neuron and fully connect the assembly with symmetric weights
Finding a suitable energy function E
Determination of Energy Function E function for TSP has four components
satisfying four constraints
Each city can have no more than oneposition i.e. each row can have no morethan one activated neuron
E1= A/2 Σk Σi Σj≠i σki σkj A - Constant
Energy Function (Contd..)
Each position contains no more than one city i.e. each column contains no more than one activated neuron
E2= B/2 Σj Σk Σr≠k σkj σrj B - constant
Energy Function (Contd..)
There are exactly N entries in the output matrix i.e. there are N 1’s in the output matrix
E3= C/2 (n - ΣkΣi σki)2 C - constant
Energy Function (cont..) Fourth term incorporates the
requirement of the shortest path
E4= D/2 ΣkΣr≠kΣj dkr σkj(σr(j+1) + σr(j-1))
where dkr is the distance between city-k and city-r
Etotal = E1 + E2 + E3 + E4
Energy Function (cont..) Energy equation is also given by
E= -½ΣkiΣrj w(ki)(rj) σki σrj
σki – City k at position i
σrj – City r at position j Output function σki
σki = ½ ( 1 + tanh(uki/u0))
u0 is a constant
uki is the net input
Weight Value
Comparing above equations with the energy equation obtained previously
W(ki)(rj) = -A δkr(1 – δrj) - Bδij(1 – δkr) –C –Ddkr(δj(i+1) + δj(i-1))
Kronecker Symbol : δkr
δkr = 1 when k = r δkr = 0 when k ≠ r
Observation
Choice of constants A,B,C and D that provide a good solution vary between Always obtain legitimate loops (D is
small relative to A, B and C)
Giving heavier weights to the distances (D is large relative to A, B and C)
Observation (cont..)
Local minima Energy function full of dips, valleys
and local minima Speed
Fast due to rapid computational capacity of network
Concurrent Neural Network
Proposed by N. Toomarian in 1988
It requires N(log(N)) neurons to compute TSP of N cities.
It also has a much higher probability to reach a valid tour.
Objective Function
Aim is to minimize the distance between city k at position i and city r at position i+1
Ei = Σk≠rΣrΣi δkiδr(i+1) dkr
Where δ is the Kronecers Symbol
Cont …
Ei = 1/N2 Σk≠rΣrΣi dkr Πi= 1 to ln(N) [1 + (2עi – 1) σki] [1
+ (2µi – 1) σri]
Where (2µi – 1) = (2עi – 1) [1 – Πj= 1 to i-1 עi ]
Also to ensure that 2 cities don’t occupy same position
Eerror = Σk≠rΣr δkr
Solution
Eerror will have a value 0 for any valid tour.
So we have a constrained optimization problem to solve.
E = Ei + λ Eerror
λ is the Lagrange multiplier to be calculated form the solution.
Minimization of energy function
Minimizing Energy function which is in terms of σki
Algorithm is an iterative procedure which is usually used for minimization of quadratic functions
The iteration steps are carried out in the direction of steepest decent with respect to the energy function E
Minimization of energy function
Differentiating the energy
dUki/dt = - δE/ δσki = - δEi/ δσki - λδEerror/ δσki
dλ/dt = ± δE/ δλ = ± Eerror
σki = tanh(αUki) , α – const.
Implementation Initial Input Matrix and the value of λ is
randomly selected and specified
At each iteration, new value of σki and λ is calculated in the direction of steepest descent of energy function
Iterations will stop either when convergence is achieved or when the number of iterations exceeds a user specified number
Comparison – Hopfield vs Concurrent NN
Converges faster than Hopfield Network
Probability to achieve valid tour is higher than Hopfield Network
Hopfield doesn’t have systematic way to determine the constant terms.
Comparison – SOM and Concurrent NN
Data set consists of 52 cities in Germany and its subset of 15 cities.
Both algorithms were run for 80 times on 15 city data set.
52 city dataset could be analyzed only using SOM while Concurrent Neural Net failed to analyze this dataset.
Result Concurrent neural network always
converged and never missed any city, where as SOM is capable of missing cities.
Concurrent Neural Network is very erratic in behavior , whereas SOM has higher reliability to detect every link in smallest path.
Overall Concurrent Neural Network performed poorly as compared to SOM.
Shortest path generatedConcurrent Neural Network (2127 km) Self Organizing Maps (1311km)
Behavior in terms of probability
Concurrent Neural Network Self Organizing Maps
Conclusion Hopfield Network can also be used
for optimization problems. Concurrent Neural Network
performs better than Hopfield network and uses less neurons.
Concurrent and Hopfield Neural Network are less efficient than SOM for solving TSP.
References N. K. Bose and P. Liang, ”Neural Network
Fundamentals with Graphs, Algorithms and Applications”, Tata McGraw Hill Publication, 1996
P. D. Wasserman, “Neural computing: theory and practice”, Van Nostrand Reinhold Co., 1989
N. Toomarian, “A Concurrent Neural Network algorithm for the Traveling Salesman Problem”, ACM Proceedings of the third conference on Hypercube concurrent computers and applications, pp. 1483-1490, 1988.
References R. Reilly, “Neural Network approach to solving the
Traveling Salesman Problem”, Journals of Computer Science in Colleges, pp. 41-61,October 2003
Wolfram Research inc., “Tutorial on Neural Networks”, http://documents.wolfram.com/applications/neuralnetworks/NeuralNetworkTheory/2.7.0.html, 2004
Prof. P. Bhattacharyya, “Introduction to computing with Neural Nets”,http://www.cse.iitb.ac.in/~pb/Teaching.html.
NP-complete NP-hard
When a decision version of a combinatorial optimization problem is proved to belong to the class of NP-complete problems, which includes well-known problems such as satisfiability,traveling salesman, the bin packing problem, etc., then the optimization version is NP-hard.
NP-complete NP-hard
“Is there a tour with length less than k" is NP-complete:
It is easy to determine if a proposed certificate has length less than k
The optimization problem : "what is the shortest tour?", is NP-
hard Since there is no easy way to determine if a certificate is the shortest.
Path lengths
Concurrent Neural Network Self Organizing Maps