recurrent networks

39
Jun 15, 2022 Rudolf Mak TU/e Computer Science 2L490 Hopfield 1 Recurrent Networks A recurrent network is characterized by The connection graph of the network has cycles, i.e. the output of a neuron can influence its input There are no natural input and output nodes Initially each neuron has a given input state Neurons change state using some update rule The network evolves until some stable situation is reached The resulting state is the output of the network

Upload: wayne-kirk

Post on 31-Dec-2015

43 views

Category:

Documents


2 download

DESCRIPTION

Recurrent Networks. A recurrent network is characterized by The connection graph of the network has cycles, i.e. the output of a neuron can influence its input There are no natural input and output nodes Initially each neuron has a given input state - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 1

Recurrent Networks

A recurrent network is characterized by • The connection graph of the network has cycles,

i.e. the output of a neuron can influence its input• There are no natural input and output nodes• Initially each neuron has a given input state• Neurons change state using some update rule• The network evolves until some stable situation

is reached• The resulting state is the output of the network

Page 2: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 2

Pattern Recognition

Recurrent networks can be used for patternrecognition in the following way:

• The stable states represent the patterns to be recognized• The initial state is a noisy or otherwise mutilated version of one of the patterns• The recognition process consists of the network evolving from its initial state to a stable state

Page 3: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 3

Pattern Recognition Example

Page 4: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 4

Pattern Recognition Example (cntd)

Noisy image Recognized

pattern

Page 5: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 5

Bipolar Data Encoding

• In bipolar encoding firing of a neuron is repre-sented by the value 1, and non-firing by the value –1

• In bipolar encoding the transfer function of the neurons is the sign function sgn

• A bipolar vector x of dimension n satisfies the equations

– sgn(x ) = x– xTx = n

Page 6: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 6

Binary versus Bipolar Encoding

The number of orthogonal vector pairs is much

larger in case of bipolar encoding. In an n-

dimensional vector space:• For binary encoding

• For bipolar encoding

Page 7: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 7

Hopfield Networks

A recurrent network is a Hopfield network when

• The neurons have discrete output (for

convenience we use bipolar encoding)

• Each neuron has a threshold

• Each pair of neurons is connected by a weighted

connection. The weight matrix is symmetric and

has a zero diagonal (no connection from a

neuron to itself)

Page 8: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 8

Network states

If a Hopfield network has n neurons, then the state of the network at time t is the vector x(t) 2 {-1, 1}n with components x i (t) that describe the states of the individual neurons.

Time is discrete, so t 2 N

The state of the network is updated using a so-called update rule. (Not) firing of a neuron at time t+1 will depend on the sign of the total input at time t

Page 9: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 9

Update Strategies

• In a sequential network only one neuron at a time is allowed to change its state. In the asyn-chronous update rule this neuron is randomly selected.

• In a parallel network several neurons are allowed to change their state simultaneously. – Limited parallelism: only neurons that are not

connected can change their state simultaneously– Unlimited parallelism: also connected neurons may

change their state simultaneously– Full parallelism: all neurons change their state simul-

taneously

Page 10: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 10

Asynchronous Update

Page 11: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 11

Asynchronous Neighborhood

Because wkk = 0 , it follows that for every pair of neighboring states x* 2 Na(x)

The asynchronous neighborhood of a state

x is defined as the set of states

Page 12: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 12

Synchronous Update

This update rule corresponds to full

parallelism

Page 13: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 13

Sign-assumption

In order for both update rules to be applica-ble, we assume that for all neurons i

Because the number of states is finite, it is

always possible to adjust the thresholds such that the above assumption holds.

Page 14: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 14

Stable States

A state x is called a stable state, when

For both the synchronous and the asyn-chronous update rule we have:

a state is a stable state if and only if the update rule does not lead to a different state.

Page 15: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 15

Cyclic behavior in asymmetric RNN

-11

-1-1

1

1

1 1

1

-1

-1

-11

1

1

Page 16: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 16

Basins of Attraction

state space

initial state

stable state

Page 17: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 17

Consensus and Energy

The consensus C(x) of a state x of a

Hopfield network with weight matrix W and bias vector b is defined as

The energy E(x) of a Hopfield network in state x is defined as

Page 18: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 18

Consensus difference

For any pair of vectors x and x* we have

Page 19: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 19

Asynchronous Convergence

If in an asynchronous step the state of the network changes from x to x-2xkek, then the consensus increases.

Since there are only a finite number of states, the consensus serves as a variant function that shows that a Hopfield network evolves to a stable state, when the asynchronous update rule is used.

Page 20: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 20

Stable States and Local maxima

A state x is a local maximum of the consensus

function when

Theorem:

A state x is a local maximum of the consensus

function if and only if it is a stable state.

Page 21: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 21

Stable equals local maximum

Page 22: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 22

Modified Consensus

The modified consensus of a state x of a Hopfield network with weight matrix W and bias vector b is defined as

Let x , x*, and x** be successive states obtained

with the synchronous update rule. Then

Page 23: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 23

Synchronous Convergence

Suppose that x, x*, and x** are successive states

obtained with the synchronous update rule. Then

Hence a Hopfield network that evolves using the synchronous update rule will arrive either in a stable state or in a cycle of length 2.

Page 24: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 24

Storage of a Single Pattern

How does one determine the weights of a

Hopfield network given a set of desired sta-

ble states?

First we consider the case of a single stable state. Let x be an arbitrary vector. Choos-ing weight matrix W and bias vector b as

makes x a stable state.

Page 25: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 25

Proof of Stability

Page 26: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 26

Example

Page 27: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 27

State encoding

0 1 2 3 4 5 6 7 8 9 10

11

12

13

14

15

x1 -1 1 -1 1 -1 1 -1 1 -1 1 -1 1 -1 1 -1 1

x2 -1 -1 1 1 -1 -1 1 1 -1 -1 1 1 -1 -1 1 1

x3 -1 -1 -1 -1 1 1 1 1 -1 -1 -1 -1 1 1 1 1

x4 -1 -1 -1 -1 -1 -1 -1 -1 1 1 1 1 1 1 1 1

0 1 2 3 4 5 6 7 8 9 10

11

12

13

14

15

4z1 1 1 -1 -1 -1 -1 -3 -3 -1 1 -1 1 -1 1 -1 1

4z2 1 -1 1 -1 3 1 3 1 -1 -1 1 1 -1 -1 1 1

4z3 1 -1 3 1 1 -1 3 1 -1 -1 -1 -1 1 1 1 1

4z4 1 3 -1 1 -1 1 -3 -1 1 1 1 1 1 1 -3 1

Page 28: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 28

Finite state machine for async update

Page 29: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 29

Weights for Multiple Patterns

Let {x(p) j 1 · p · P } be a set of patterns, and

let W(p) be the weight matrix corresponding to pattern number p.

Choose the weight matrix W and the bias vector b for a Hopfield network that must recognize all P patterns as

Question: Is x(p) indeed a stable state?

Page 30: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 30

Remarks

• It is not guaranteed that a Hopfield network with weight matrix as defined on the previous slide indeed has the patterns as it stable states

• The disturbance caused by other patterns is called crosstalk. The closer the patterns are, the larger the crosstalk is

• This raises the question how many patterns there can be stored in a network before crosstalk gets the overhand

Page 31: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 31

Input of neuron i in state x(p)

Page 32: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 32

Crosstalk

The crosstalk term is defined by

Neuron i is stable when , because

Page 33: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 33

Spurious States

Besides the desired stable states the network can have additional undesired (spurious) stable states

• If x is stable and b = 0, then –x is also stable.

• Some combinations of an odd number of stable states can be stable.

• Moreover there can be more complicated additional stable states (spin glass states) that bare no relation to the desired states.

Page 34: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 34

Storage Capacity

Question: How many stable states P can

be stored in a network of size n ?

Answer: That depends on the probability of

instability one is willing to accept. Experi-

mentally P ¼ 0.15n has been found (by

Hopfield) to be a reasonable value.

Page 35: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 35

Probabilistic analysis 1

Assume that all components of the patterns are

random variables with equal probability of being

1 and -1

Then it can be shown that has ap-

proximately the standard normal distribu-

tion N(0, 1).

Page 36: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 36

Probabilistic Analysis 2

From these assumptions it follows that

Application of the central limit theorem yields

Page 37: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 37

Standard Normal Distribution

The shaded area under the

bell-shaped curve gives the

probability

Pr[y ¸ 1.5]

Page 38: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 38

Probability of Instability

0.05 1.645 0.370

0.01 2.326 0.185

0.005 2.576 0.151

0.001 3.090 0.105

Page 39: Recurrent Networks

Apr 19, 2023 Rudolf Mak TU/e Computer Science 2L490 Hopfield 39

Topics Not Treated

• Reduction of crosstalk for correlated patterns

• Stability analysis for correlated patterns

• Methods to eliminate spurious states

• Continuous Hopfield models

• Different associative memories

– Binary Associative Memory (Kosko)

– Brain State in a Box (Kawamoto, Anderson)