# Hopfield network

## Usage as Wiki2Reveal Presentations

This page can be displayed as Wiki2Reveal slides. Single sections are regarded as slides and modifications in this article will immediately affect the next Wiki2Reveal preview of the slides.

## Introduction to Network

A Hopfield network (or Ising model of a neural network or Ising–Lenz–Little model) is a form of recurrent artificial neural network and a type of spin glass system popularised by John Hopfield in 1982[1] as described earlier by Little in 1974[2] based on Ernst Ising's work with Wilhelm Lenz on the Ising model.[3] Hopfield networks serve as content-addressable ("associative") memory systems with binary threshold nodes, or with continuous variables.[4] Hopfield networks also provide a model for understanding human memory.[5][6]

• Explore the areas of application in Digital Learning Environment where Hopfield networks can be applied for adaption of the learning environment to the requirements and constraints of the learner.

## Structure

The units in Hopfield nets are binary threshold units, i.e. the units only take on two different values for their states, and the value is determined by whether or not the unit's input exceeds its threshold ${\displaystyle U_{i}}$ .

### Discrete Hopfield Nets

Discrete Hopfield nets describe relationships between binary (firing or not-firing) neurons with the index ${\displaystyle 1,2,\ldots ,i,j,\ldots ,N}$ .[1]  At a certain time, the state of the neural net is described by a vector ${\displaystyle V}$ , which records which neurons are firing in a binary word of ${\displaystyle N}$  bits.


### Weighted Graphs and Interactions

The interactions ${\displaystyle w_{ij}}$  between neurons have units that usually take on values of 1 or −1, and this convention will be used throughout this article. However, other literature might use units that take values of 0 and 1.

### Learning Algorithm - Hebbian Learning Rule

These interactions are "learned" via Hebb's law of association, such that, for a certain state ${\displaystyle V^{s}}$

${\displaystyle w_{ij}=V_{i}^{s}V_{j}^{s}}$

but ${\displaystyle w_{ii}=0}$ .

(Note that the Hebbian learning rule takes the form ${\displaystyle w_{ij}=(2V_{i}^{s}-1)(2V_{j}^{s}-1)}$  when the units assume values in ${\displaystyle \{0,1\}}$ .)

### Pattern Recognition with Hopfield Nets

Once the network is trained, ${\displaystyle w_{ij}}$  no longer evolve. If a new state of neurons ${\displaystyle V^{s'}}$  is introduced to the neural network, the net acts on neurons such that

• ${\displaystyle V_{i}^{s'}\rightarrow 1}$  if ${\displaystyle \sum _{j}w_{ij}V_{j}^{s'}>U_{i}}$
• ${\displaystyle V_{i}^{s'}\rightarrow -1}$  if ${\displaystyle \sum _{j}w_{ij}V_{j}^{s'}

where ${\displaystyle U_{i}}$  is the threshold value of the i'th neuron (often taken to be 0).[7] In this way, Hopfield networks have the ability to "remember" states stored in the interaction matrix, because if a new state ${\displaystyle V^{s'}}$  is subjected to the interaction matrix, each neuron will change until it matches the original state ${\displaystyle V^{s}}$  (see the Updates section below).

### Network Restrictions - Properties of the directed Graph

The connections in a Hopfield net typically have the following restrictions:

• ${\displaystyle w_{ii}=0,\forall i}$  (no unit has a connection with itself)
• ${\displaystyle w_{ij}=w_{ji},\forall i,j}$  (connections are symmetric)

### Energy function - Convergence of Network

The constraint that weights are symmetric guarantees that the energy function decreases monotonically while following the activation rules.[8] A network with asymmetric weights may exhibit some periodic or chaotic behaviour; however, Hopfield found that this behavior is confined to relatively small parts of the phase space and does not impair the network's ability to act as a content-addressable associative memory system.

### Continuous Activity Values of Neurons =

Hopfield also modeled neural nets for continuous values, in which the electric output of each neuron is not binary but some value between 0 and 1.[4] He found that this type of network was also able to store and reproduce memorized states.

### Weighted Graphs/Nets

Notice that every pair of units i and j in a Hopfield network has a connection that is described by the connectivity weight ${\displaystyle w_{ij}}$ . In this sense, the Hopfield network can be formally described as a complete undirected graph ${\displaystyle G=\langle V,f\rangle }$ , where ${\displaystyle V}$  is a set of McCulloch–Pitts neurons and ${\displaystyle f:V^{2}\rightarrow \mathbb {R} }$  is a function that links pairs of units to a real value, the connectivity weight.

## Updating

Updating one unit (node in the graph simulating the artificial neuron) in the Hopfield network is performed using the following rule:

${\displaystyle s_{i}\leftarrow \left\{{\begin{array}{ll}+1&{\text{if }}\sum _{j}{w_{ij}s_{j}}\geq \theta _{i},\\-1&{\text{otherwise.}}\end{array}}\right.}$

where:

• ${\displaystyle w_{ij}}$  is the strength of the connection weight from unit j to unit i (the weight of the connection).
• ${\displaystyle s_{i}}$  is the state of unit i.
• ${\displaystyle \theta _{i}}$  is the threshold of unit i.

### Synchronoous/Asynchronous Update

Updates in the Hopfield network can be performed in two different ways:

• Asynchronous: Only one unit is updated at a time. This unit can be picked at random, or a pre-defined order can be imposed from the very beginning.
• Synchronous: All units are updated at the same time. This requires a central clock to the system in order to maintain synchronization. This method is viewed by some as less realistic, based on an absence of observed global clock influencing analogous biological or physical systems of interest.

### Neurons "attract or repel each other" in state space

The weight between two units has a powerful impact upon the values of the neurons. Consider the connection weight ${\displaystyle w_{ij}}$  between two neurons i and j. If ${\displaystyle w_{ij}>0}$ , the updating rule implies that:

• when ${\displaystyle s_{j}=1}$ , the contribution of j in the weighted sum is positive. Thus, ${\displaystyle s_{i}}$  is pulled by j towards its value ${\displaystyle s_{i}=1}$
• when ${\displaystyle s_{j}=-1}$ , the contribution of j in the weighted sum is negative. Then again, ${\displaystyle s_{i}}$  is pushed by j towards its value ${\displaystyle s_{i}=-1}$

Thus, the values of neurons i and j will converge if the weight between them is positive. Similarly, they will diverge if the weight is negative.

## Working principles of discrete and continuous Hopfield networks

Bruck shed light on the behavior of a neuron in the discrete Hopfield network when proving its convergence in his paper in 1990.[9] A subsequent paper[10] further investigated the behavior of any neuron in both discrete-time and continuous-time Hopfield networks when the corresponding energy function is minimized during an optimization process. Bruck shows[9] that neuron j changes its state if and only if it further decreases the following biased pseudo-cut. The discrete Hopfield network minimizes the following biased pseudo-cut[10] for the synaptic weight matrix of the Hopfield net.

${\displaystyle J_{pseudo-cut}(k)=\sum _{i\in C_{1}(k)}\sum _{j\in C_{2}(k)}w_{ij}+\sum _{j\in C_{1}(k)}{\theta _{j}}}$

where ${\displaystyle C_{1}(k)}$  and ${\displaystyle C_{2}(k)}$  represents the set of neurons which are −1 and +1, respectively, at time ${\displaystyle k}$ . For further details, see the recent paper.[10]

The discrete-time Hopfield Network always minimizes exactly the following pseudo-cut[9][10]

${\displaystyle U(k)=\sum _{i=1}^{N}\sum _{j=1}^{N}w_{ij}(s_{i}(k)-s_{j}(k))^{2}+2\sum _{j=1}^{N}\theta _{j}s_{j}(k)}$

The continuous-time Hopfield network always minimizes an upper bound to the following weighted cut[10]

${\displaystyle V(t)=\sum _{i=1}^{N}\sum _{j=1}^{N}w_{ij}(f(s_{i}(t))-f(s_{j}(t))^{2}+2\sum _{j=1}^{N}\theta _{j}f(s_{j}(t))}$

where ${\displaystyle f(\cdot )}$  is a zero-centered sigmoid function.

The complex Hopfield network, on the other hand, generally tends to minimize the so-called shadow-cut of the complex weight matrix of the net.[11]

## Energy

Hopfield nets have a scalar value associated with each state of the network, referred to as the "energy", E, of the network, where:

${\displaystyle E=-{\frac {1}{2}}\sum _{i,j}w_{ij}s_{i}s_{j}+\sum _{i}\theta _{i}s_{i}}$

This quantity is called "energy" because it either decreases or stays the same upon network units being updated. Furthermore, under repeated updating the network will eventually converge to a state which is a local minimum in the energy function (which is considered to be a Lyapunov function).[1] Thus, if a state is a local minimum in the energy function it is a stable state for the network. Note that this energy function belongs to a general class of models in physics under the name of Ising models; these in turn are a special case of Markov networks, since the associated probability measure, the Gibbs measure, has the Markov property.

## Hopfield network in optimization

Hopfield and Tank presented the Hopfield network application in solving the classical traveling-salesman problem in 1985.[12] Since then, the Hopfield network has been widely used for optimization. The idea of using the Hopfield network in optimization problems is straightforward: If a constrained/unconstrained cost function can be written in the form of the Hopfield energy function E, then there exists a Hopfield network whose equilibrium points represent solutions to the constrained/unconstrained optimization problem. Minimizing the Hopfield energy function both minimizes the objective function and satisfies the constraints also as the constraints are “embedded” into the synaptic weights of the network. Although including the optimization constraints into the synaptic weights in the best possible way is a challenging task, indeed many various difficult optimization problems with constraints in different disciplines have been converted to the Hopfield energy function: Associative memory systems, Analog-to-Digital conversion, job-shop scheduling problem, quadratic assignment and other related NP-complete problems, channel allocation problem in wireless networks, mobile ad-hoc network routing problem, image restoration, system identification, combinatorial optimization, etc, just to name a few. Further details can be found in e.g. the paper.[10]

## Initialization and running

Initialization of the Hopfield networks is done by setting the values of the units to the desired start pattern. Repeated updates are then performed until the network converges to an attractor pattern. Convergence is generally assured, as Hopfield proved that the attractors of this nonlinear dynamical system are stable, not periodic or chaotic as in some other systems. Therefore, in the context of Hopfield networks, an attractor pattern is a final stable state, a pattern that cannot change any value within it under updating.

## Training

Training a Hopfield net involves lowering the energy of states that the net should "remember". This allows the net to serve as a content addressable memory system, that is to say, the network will converge to a "remembered" state if it is given only part of the state. The net can be used to recover from a distorted input to the trained state that is most similar to that input. This is called associative memory because it recovers memories on the basis of similarity. For example, if we train a Hopfield net with five units so that the state (1, −1, 1, −1, 1) is an energy minimum, and we give the network the state (1, −1, −1, −1, 1) it will converge to (1, −1, 1, −1, 1). Thus, the network is properly trained when the energy of states which the network should remember are local minima. Note that, in contrast to Perceptron training, the thresholds of the neurons are never updated.

### Learning rules

There are various different learning rules that can be used to store information in the memory of the Hopfield network. It is desirable for a learning rule to have both of the following two properties:

• Local: A learning rule is local if each weight is updated using information available to neurons on either side of the connection that is associated with that particular weight.
• Incremental: New patterns can be learned without using information from the old patterns that have been also used for training. That is, when a new pattern is used for training, the new values for the weights only depend on the old values and on the new pattern.[13]

These properties are desirable, since a learning rule satisfying them is more biologically plausible. For example, since the human brain is always learning new concepts, one can reason that human learning is incremental. A learning system that was not incremental would generally be trained only once, with a huge batch of training data.

### Hebbian learning rule for Hopfield networks

The Hebbian Theory was introduced by Donald Hebb in 1949, in order to explain "associative learning", in which simultaneous activation of neuron cells leads to pronounced increases in synaptic strength between those cells.[14] It is often summarized as "Neurons that fire together, wire together. Neurons that fire out of sync, fail to link".

The Hebbian rule is both local and incremental. For the Hopfield networks, it is implemented in the following manner, when learning ${\displaystyle n}$  binary patterns:

${\displaystyle w_{ij}={\frac {1}{n}}\sum _{\mu =1}^{n}\epsilon _{i}^{\mu }\epsilon _{j}^{\mu }}$

where ${\displaystyle \epsilon _{i}^{\mu }}$  represents bit i from pattern ${\displaystyle \mu }$ .

If the bits corresponding to neurons i and j are equal in pattern ${\displaystyle \mu }$ , then the product ${\displaystyle \epsilon _{i}^{\mu }\epsilon _{j}^{\mu }}$  will be positive. This would, in turn, have a positive effect on the weight ${\displaystyle w_{ij}}$  and the values of i and j will tend to become equal. The opposite happens if the bits corresponding to neurons i and j are different.

### Storkey learning rule

This rule was introduced by Amos Storkey in 1997 and is both local and incremental. Storkey also showed that a Hopfield network trained using this rule has a greater capacity than a corresponding network trained using the Hebbian rule.[15] The weight matrix of an attractor neural network[clarification needed] is said to follow the Storkey learning rule if it obeys:

${\displaystyle w_{ij}^{\nu }=w_{ij}^{\nu -1}+{\frac {1}{n}}\epsilon _{i}^{\nu }\epsilon _{j}^{\nu }-{\frac {1}{n}}\epsilon _{i}^{\nu }h_{ji}^{\nu }-{\frac {1}{n}}\epsilon _{j}^{\nu }h_{ij}^{\nu }}$

where ${\displaystyle h_{ij}^{\nu }=\sum _{k=1~:~i\neq k\neq j}^{n}w_{ik}^{\nu -1}\epsilon _{k}^{\nu }}$  is a form of local field[13] at neuron i.

This learning rule is local, since the synapses take into account only neurons at their sides. The rule makes use of more information from the patterns and weights than the generalized Hebbian rule, due to the effect of the local field.

## Spurious patterns

Patterns that the network uses for training (called retrieval states) become attractors of the system. Repeated updates would eventually lead to convergence to one of the retrieval states. However, sometimes the network will converge to spurious patterns (different from the training patterns).[16] The energy in these spurious patterns is also a local minimum. For each stored pattern x, the negation -x is also a spurious pattern.

A spurious state can also be a linear combination of an odd number of retrieval states. For example, when using 3 patterns ${\displaystyle \mu _{1},\mu _{2},\mu _{3}}$ , one can get the following spurious state:

${\displaystyle \epsilon _{i}^{\rm {mix}}=\pm \operatorname {sgn}(\pm \epsilon _{i}^{\mu _{1}}\pm \epsilon _{i}^{\mu _{2}}\pm \epsilon _{i}^{\mu _{3}})}$

Spurious patterns that have an even number of states cannot exist, since they might sum up to zero[16]

## Capacity

The Network capacity of the Hopfield network model is determined by neuron amounts and connections within a given network. Therefore, the number of memories that are able to be stored is dependent on neurons and connections. Furthermore, it was shown that the recall accuracy between vectors and nodes was 0.138 (approximately 138 vectors can be recalled from storage for every 1000 nodes) (Hertz et al., 1991). Therefore, it is evident that many mistakes will occur if one tries to store a large number of vectors. When the Hopfield model does not recall the right pattern, it is possible that an intrusion has taken place, since semantically related items tend to confuse the individual, and recollection of the wrong pattern occurs. Therefore, the Hopfield network model is shown to confuse one stored item with that of another upon retrieval. Perfect recalls and high capacity, >0.14, can be loaded in the network by Storkey learning method; ETAM,[17][18] ETAM experiments also in.[19] Ulterior models inspired by the Hopfield network were later devised to raise the storage limit and reduce the retrieval error rate, with some being capable of one-shot learning.[20]

The storage capacity can be given as ${\displaystyle C\cong {\frac {n}{2\log _{2}n}}}$  where ${\displaystyle n}$  is the number of neurons in the net

## Human memory

The Hopfield model accounts for associative memory through the incorporation of memory vectors. Memory vectors can be slightly used, and this would spark the retrieval of the most similar vector in the network. However, we will find out that due to this process, intrusions can occur. In associative memory for the Hopfield network, there are two types of operations: auto-association and hetero-association. The first being when a vector is associated with itself, and the latter being when two different vectors are associated in storage. Furthermore, both types of operations are possible to store within a single memory matrix, but only if that given representation matrix is not one or the other of the operations, but rather the combination (auto-associative and hetero-associative) of the two. It is important to note that Hopfield's network model utilizes the same learning rule as Hebb's (1949) learning rule, which basically tried to show that learning occurs as a result of the strengthening of the weights by when activity is occurring.

Rizzuto and Kahana (2001) were able to show that the neural network model can account for repetition on recall accuracy by incorporating a probabilistic-learning algorithm. During the retrieval process, no learning occurs. As a result, the weights of the network remain fixed, showing that the model is able to switch from a learning stage to a recall stage. By adding contextual drift they were able to show the rapid forgetting that occurs in a Hopfield model during a cued-recall task. The entire network contributes to the change in the activation of any single node.

McCulloch and Pitts' (1943) dynamical rule, which describes the behavior of neurons, does so in a way that shows how the activations of multiple neurons map onto the activation of a new neuron's firing rate, and how the weights of the neurons strengthen the synaptic connections between the new activated neuron (and those that activated it). Hopfield would use McCulloch–Pitts's dynamical rule in order to show how retrieval is possible in the Hopfield network. However, it is important to note that Hopfield would do so in a repetitious fashion. Hopfield would use a nonlinear activation function, instead of using a linear function. This would therefore create the Hopfield dynamical rule and with this, Hopfield was able to show that with the nonlinear activation function, the dynamical rule will always modify the values of the state vector in the direction of one of the stored patterns.

## Dense associative memory or modern Hopfield network

Hopfield networks[1][4] are recurrent neural networks with dynamical trajectories converging to fixed point attractor states and described by an energy function. The state of each model neuron ${\textstyle i}$  is defined by a time-dependent variable ${\displaystyle V_{i}}$ , which can be chosen to be either discrete or continuous. A complete model describes the mathematics of how the future state of activity of each neuron depends on the known present or previous activity of all the neurons.

In the original Hopfield model of associative memory,[1] the variables were binary, and the dynamics were described by a one-at-a-time update of the state of the neurons. An energy function quadratic in the ${\displaystyle V_{i}}$  was defined, and the dynamics consisted of changing the activity of each single neuron ${\displaystyle i}$  only if doing so would lower the total energy of the system. This same idea was extended to the case of ${\displaystyle V_{i}}$  being a continuous variable representing the output of neuron ${\displaystyle i}$ , and ${\displaystyle V_{i}}$  being a monotonic function of an input current. The dynamics became expressed as a set of first-order differential equations for which the "energy" of the system always decreased.[4] The energy in the continuous case has one term which is quadratic in the ${\displaystyle V_{i}}$  (as in the binary model), and a second term which depends on the gain function (neuron's activation function). While having many desirable properties of associative memory, both of these classical systems suffer from a small memory storage capacity, which scales linearly with the number of input features.[1]

Dense Associative Memories[21] (also known as the modern Hopfield networks[22]) are generalizations of the classical Hopfield Networks that break the linear scaling relationship between the number of input features and the number of stored memories. This is achieved by introducing stronger non-linearities (either in the energy function or neurons’ activation functions) leading to super-linear[21] (even an exponential[23]) memory storage capacity as a function of the number of feature neurons. The network still requires a sufficient number of hidden neurons.[24]

The key theoretical idea behind the modern Hopfield networks is to use an energy function and an update rule that is more sharply peaked around the stored memories in the space of neuron’s configurations compared to the classical Hopfield Network.[21]

### Discrete variables

A simple example[21] of the modern Hopfield network can be written in terms of binary variables ${\displaystyle V_{i}}$  that represent the active ${\displaystyle V_{i}=+1}$  and inactive ${\displaystyle V_{i}=-1}$  state of the model neuron ${\displaystyle i}$ .${\displaystyle E=-\sum \limits _{\mu =1}^{N_{\text{mem}}}F{\Big (}\sum \limits _{i=1}^{N_{f}}\xi _{\mu i}V_{i}{\Big )}}$ In this formula the weights ${\textstyle \xi _{\mu i}}$  represent the matrix of memory vectors (index ${\displaystyle \mu =1...N_{\text{mem}}}$  enumerates different memories, and index ${\displaystyle i=1...N_{f}}$  enumerates the content of each memory corresponding to the ${\displaystyle i}$ -th feature neuron), and the function ${\displaystyle F(x)}$  is a rapidly growing non-linear function. The update rule for individual neurons (in the asynchronous case) can be written in the following form ${\displaystyle V_{i}^{(t+1)}=Sign{\bigg [}\sum \limits _{\mu =1}^{N_{\text{mem}}}{\bigg (}F{\Big (}\xi _{\mu i}+\sum \limits _{j\neq i}\xi _{\mu j}V_{j}^{(t)}{\Big )}-F{\Big (}-\xi _{\mu i}+\sum \limits _{j\neq i}\xi _{\mu j}V_{j}^{(t)}{\Big )}{\bigg )}{\bigg ]}}$ which states that in order to calculate the updated state of the ${\textstyle i}$ -th neuron the network compares two energies: the energy of the network with the ${\displaystyle i}$ -th neuron in the ON state and the energy of the network with the ${\displaystyle i}$ -th neuron in the OFF state, given the states of the remaining neuron. The updated state of the ${\displaystyle i}$ -th neuron selects the state that has the lowest of the two energies.[21]

In the limiting case when the non-linear energy function is quadratic ${\displaystyle F(x)=x^{2}}$  these equations reduce to the familiar energy function and the update rule for the classical binary Hopfield Network.[1]

The memory storage capacity of these networks can be calculated for random binary patterns. For the power energy function ${\displaystyle F(x)=x^{n}}$  the maximal number of memories that can be stored and retrieved from this network without errors is given by[21]${\displaystyle N_{\text{mem}}^{max}\approx {\frac {1}{2(2n-3)!!}}{\frac {N_{f}^{n-1}}{\ln(N_{f})}}}$ For an exponential energy function ${\textstyle F(x)=e^{x}}$  the memory storage capacity is exponential in the number of feature neurons[23]${\displaystyle N_{\text{mem}}^{max}\approx 2^{N_{f}/2}}$

### Continuous variables

Modern Hopfield networks or dense associative memories can be best understood in continuous variables and continuous time.[22][24] Consider the network architecture, shown in Fig.1, and the equations for neuron's states evolution[24]

${\displaystyle {\begin{cases}\tau _{f}{\frac {dx_{i}}{dt}}=\sum \limits _{\mu =1}^{N_{h}}\xi _{i\mu }f_{\mu }-x_{i}+I_{i}\\\tau _{h}{\frac {dh_{\mu }}{dt}}=\sum \limits _{i=1}^{N_{f}}\xi _{\mu i}g_{i}-h_{\mu }\end{cases}}}$

(1)

where the currents of the feature neurons are denoted by ${\textstyle x_{i}}$ , and the currents of the memory neurons are denoted by ${\displaystyle h_{\mu }}$  (${\displaystyle h}$  stands for hidden neurons). There are no synaptic connections among the feature neurons or the memory neurons. A matrix ${\displaystyle \xi _{\mu i}}$  denotes the strength of synapses from a feature neuron ${\displaystyle i}$  to the memory neuron ${\displaystyle \mu }$ . The synapses are assumed to be symmetric, so that the same value characterizes a different physical synapse from the memory neuron ${\displaystyle \mu }$  to the feature neuron ${\displaystyle i}$ . The outputs of the memory neurons and the feature neurons are denoted by ${\displaystyle f_{\mu }}$  and ${\displaystyle g_{i}}$ , which are non-linear functions of the corresponding currents. In general these outputs can depend on the currents of all the neurons in that layer so that ${\displaystyle f_{\mu }=f(\{h_{\mu }\})}$  and ${\textstyle g_{i}=g(\{x_{i}\})}$ . It is convenient to define these activation function as derivatives of the Lagrangian functions for the two groups of neurons

${\displaystyle f_{\mu }={\frac {\partial L_{h}}{\partial h_{\mu }}},\ \ \ \ {\text{and}}\ \ \ \ g_{i}={\frac {\partial L_{v}}{\partial x_{i}}}}$

(2)

This way the specific form of the equations for neuron's states is completely defined once the Lagrangian functions are specified. Finally, the time constants for the two groups of neurons are denoted by ${\displaystyle \tau _{f}}$  and ${\displaystyle \tau _{h}}$ , ${\displaystyle I_{i}}$  is the input current to the network that can be driven by the presented data.

General systems of non-linear differential equations can have many complicated behaviors that can depend on the choice of the non-linearities and the initial conditions. For Hopfield Networks, however, this is not the case - the dynamical trajectories always converge to a fixed point attractor state. This property is achieved because these equations are specifically engineered so that they have an underlying energy function[24]

${\displaystyle E(t)={\Big [}\sum \limits _{i=1}^{N_{f}}(x_{i}-I_{i})g_{i}-L_{x}{\Big ]}+{\Big [}\sum \limits _{\mu =1}^{N_{h}}h_{\mu }f_{\mu }-L_{h}{\Big ]}-\sum \limits _{\mu ,i}f_{\mu }\xi _{\mu i}g_{i}}$

(3)

The terms grouped into square brackets represent a Legendre transform of the Lagrangian function with respect to the states of the neurons. If the Hessian matrices of the Lagrangian functions are positive semi-definite, the energy function is guaranteed to decrease on the dynamical trajectory[24]

${\displaystyle {\frac {dE(t)}{dt}}=-\tau _{f}\sum \limits _{i,j=1}^{N_{f}}{\frac {dx_{i}}{dt}}{\frac {\partial ^{2}L_{x}}{\partial x_{i}\partial x_{j}}}{\frac {dx_{j}}{dt}}-\tau _{h}\sum \limits _{\mu ,\nu =1}^{N_{h}}{\frac {dh_{\mu }}{dt}}{\frac {\partial ^{2}L_{h}}{\partial h_{\mu }\partial h_{\nu }}}{\frac {dh_{\nu }}{dt}}\leq 0}$

(4)

This property makes it possible to prove that the system of dynamical equations describing temporal evolution of neurons' activities will eventually reach a fixed point attractor state.

In certain situations one can assume that the dynamics of hidden neurons equilibrates at a much faster time scale compared to the feature neurons, ${\textstyle \tau _{h}\ll \tau _{f}}$ . In this case the steady state solution of the second equation in the system (1) can be used to express the currents of the hidden units through the outputs of the feature neurons. This makes it possible to reduce the general theory (1) to an effective theory for feature neurons only. The resulting effective update rules and the energies for various common choices of the Lagrangian functions are shown in Fig.2. In the case of log-sum-exponential Lagrangian function the update rule (if applied once) for the states of the feature neurons is the attention mechanism[22] commonly used in many modern AI systems (see Ref.[24] for the derivation of this result from the continuous time formulation).

### Relationship to classical Hopfield network with continuous variables

Classical formulation of continuous Hopfield Networks[4] can be understood[24] as a special limiting case of the modern Hopfield networks with one hidden layer. Continuous Hopfield Networks for neurons with graded response are typically described[4] by the dynamical equations

${\displaystyle \tau _{f}{\frac {dx_{i}}{dt}}=\sum \limits _{j=1}^{N_{f}}T_{ij}V_{j}-x_{i}+I_{i}}$

(5)

and the energy function

${\displaystyle E=-{\frac {1}{2}}\sum \limits _{i,j=1}^{N_{f}}T_{ij}V_{i}V_{j}-\sum \limits _{i=1}^{N_{f}}V_{i}I_{i}+\sum \limits _{i=1}^{N_{f}}\int \limits ^{V_{i}}g^{-1}(z)\,dz}$

(6)

where ${\textstyle V_{i}=g(x_{i})}$ , and ${\displaystyle g^{-1}(z)}$  is the inverse of the activation function ${\displaystyle g(x)}$ . This model is a special limit of the class of models that is called models A,[24] with the following choice of the Lagrangian functions

${\displaystyle L_{v}=\sum \limits _{i=1}^{N_{f}}\int \limits ^{x_{i}}g(x)dx,\ \ \ \ \ {\text{and}}\ \ \ \ \ L_{h}={\frac {1}{2}}\sum \limits _{\mu =1}^{N_{h}}h_{\mu }^{2}}$

(7)

that, according to the definition (2), leads to the activation functions

${\displaystyle V_{i}=g(x_{i}),\ \ \ \ \ {\text{and}}\ \ \ \ \ f_{\mu }=h_{\mu }}$

(8)

If we integrate out the hidden neurons the system of equations (1) reduces to the equations on the feature neurons (5) with ${\displaystyle T_{ij}=\sum \limits _{\mu =1}^{N_{h}}\xi _{\mu i}\xi _{\mu j}}$ , and the general expression for the energy (3) reduces to the effective energy

${\displaystyle E=-{\frac {1}{2}}\sum \limits _{i,j=1}^{N_{f}}T_{ij}V_{i}V_{j}-\sum \limits _{i=1}^{N_{f}}V_{i}I_{i}+\sum \limits _{i=1}^{N_{f}}{\Big (}x_{i}V_{i}-\int \limits ^{x_{i}}g(x)dx{\Big )}}$

(9)

While the first two terms in equation (6) are the same as those in equation (9), the third terms look superficially different. In equation (9) it is a Legendre transform of the Lagrangian for the feature neurons, while in (6) the third term is an integral of the inverse activation function. Nevertheless, these two expressions are in fact equivalent, since the derivatives of a function and its Legendre transform are inverse functions of each other. The easiest way to see that these two terms are equal explicitly is to differentiate each one with respect to ${\displaystyle x_{i}}$ . The results of these differentiations for both expressions are equal to ${\displaystyle x_{i}g(x_{i})'}$ . Thus, the two expressions are equal up to an additive constant. This completes the proof[24] that the classical Hopfield Network with continuous states[4] is a special limiting case of the modern Hopfield network (1) with energy (3).

### General formulation of the modern Hopfield network

Biological neural networks have a large degree of heterogeneity in terms of different cell types. This section describes a mathematical model of a fully connected modern Hopfield network assuming the extreme degree of heterogeneity: every single neuron is different.[25] Specifically, an energy function and the corresponding dynamical equations are described assuming that each neuron has its own activation function and kinetic time scale. The network is assumed to be fully connected, so that every neuron is connected to every other neuron using a symmetric matrix of weights ${\displaystyle W_{IJ}}$ , indices ${\displaystyle I}$  and ${\displaystyle J}$  enumerate different neurons in the network, see Fig.3. The easiest way to mathematically formulate this problem is to define the architecture through a Lagrangian function ${\displaystyle L(\{x_{I}\})}$  that depends on the activities of all the neurons in the network. The activation function for each neuron is defined as a partial derivative of the Lagrangian with respect to that neuron's activity

${\displaystyle g_{I}={\frac {\partial L}{\partial x_{I}}}}$

(10)

From the biological perspective one can think about ${\displaystyle g_{I}}$  as an axonal output of the neuron ${\displaystyle I}$ . In the simplest case, when the Lagrangian is additive for different neurons, this definition results in the activation that is a non-linear function of that neuron's activity. For non-additive Lagrangians this activation function can depend on the activities of a group of neurons. For instance, it can contain contrastive (softmax) or divisive normalization. The dynamical equations describing temporal evolution of a given neuron are given by[25]

${\displaystyle \tau _{I}{\frac {dx_{I}}{dt}}=\sum \limits _{J=1}^{N}W_{IJ}g_{J}-x_{I}}$

(11)

This equation belongs to the class of models called firing rate models in neuroscience. Each neuron ${\displaystyle I}$  collects the axonal outputs ${\displaystyle g_{J}}$  from all the neurons, weights them with the synaptic coefficients ${\displaystyle W_{IJ}}$  and produces its own time-dependent activity ${\displaystyle x_{I}}$ . The temporal evolution has a time constant ${\displaystyle \tau _{I}}$ , which in general can be different for every neuron. This network has a global energy function[25]

${\displaystyle E=\sum \limits _{I=1}^{N}x_{I}g_{I}-L-{\frac {1}{2}}\sum \limits _{I,J=1}^{N}g_{I}W_{IJ}g_{J}}$

(12)

where the first two terms represent the Legendre transform of the Lagrangian function with respect to the neurons' currents ${\displaystyle x_{I}}$ . The temporal derivative of this energy function can be computed on the dynamical trajectories leading to (see [25] for details)

${\displaystyle {\frac {dE}{dt}}=-\sum \limits _{I,K=1}^{N}{\frac {dx_{I}}{dt}}M_{IK}{\frac {dx_{K}}{dt}}\leq 0,\ \ \ \ {\text{where}}\ \ \ \ M_{IK}=\tau _{I}{\frac {\partial ^{2}L}{\partial x_{I}\partial x_{K}}}}$

(13)

The last inequality sign holds provided that the matrix ${\displaystyle M_{IK}}$  (or its symmetric part) is positive semi-definite. If, in addition to this, the energy function is bounded from below the non-linear dynamical equations are guaranteed to converge to a fixed point attractor state. The advantage of formulating this network in terms of the Lagrangian functions is that it makes it possible to easily experiment with different choices of the activation functions and different architectural arrangements of neurons. For all those flexible choices the conditions of convergence are determined by the properties of the matrix ${\displaystyle M_{IJ}}$  and the existence of the lower bound on the energy function.

### Hierarchical associative memory network

The neurons can be organized in layers so that every neuron in a given layer has the same activation function and the same dynamic time scale. If we assume that there are no horizontal connections between the neurons within the layer (lateral connections) and there are no skip-layer connections, the general fully connected network (11), (12) reduces to the architecture shown in Fig.4. It has ${\displaystyle N_{\text{layer}}}$  layers of recurrently connected neurons with the states described by continuous variables ${\displaystyle x_{i}^{A}}$  and the activation functions ${\displaystyle g_{i}^{A}}$ , index ${\displaystyle A}$  enumerates the layers of the network, and index ${\displaystyle i}$  enumerates individual neurons in that layer. The activation functions can depend on the activities of all the neurons in the layer. Every layer can have a different number of neurons ${\displaystyle N_{A}}$ . These neurons are recurrently connected with the neurons in the preceding and the subsequent layers. The matrices of weights that connect neurons in layers ${\displaystyle A}$  and ${\displaystyle B}$  are denoted by ${\displaystyle \xi _{ij}^{(A,B)}}$  (the order of the upper indices for weights is the same as the order of the lower indices, in the example above this means that the index ${\displaystyle i}$  enumerates neurons in the layer ${\displaystyle A}$ , and index ${\displaystyle j}$  enumerates neurons in the layer ${\displaystyle B}$ ). The feedforward weights and the feedback weights are equal. The dynamical equations for the neurons' states can be written as[25]

${\displaystyle \tau _{A}{\frac {dx_{i}^{A}}{dt}}=\sum \limits _{j=1}^{N_{A-1}}\xi _{ij}^{(A,A-1)}g_{j}^{A-1}+\sum \limits _{j=1}^{N_{A+1}}\xi _{ij}^{(A,A+1)}g_{j}^{A+1}-x_{i}^{A}}$

(14)

with boundary conditions

${\displaystyle g_{i}^{0}=0,\ \ \ \ \ {\text{and}}\ \ \ \ \ g_{i}^{N_{\text{layer}}+1}=0}$

(15)

The main difference of these equations from the conventional feedforward networks is the presence of the second term, which is responsible for the feedback from higher layers. These top-down signals help neurons in lower layers to decide on their response to the presented stimuli. Following the general recipe it is convenient to introduce a Lagrangian function ${\displaystyle L^{A}(\{x_{i}^{A}\})}$  for the ${\displaystyle A}$ -th hidden layer, which depends on the activities of all the neurons in that layer.[25] The activation functions in that layer can be defined as partial derivatives of the Lagrangian

${\displaystyle g_{i}^{A}={\frac {\partial L^{A}}{\partial x_{i}^{A}}}}$

(16)

With these definitions the energy (Lyapunov) function is given by[25]

${\displaystyle E=\sum \limits _{A=1}^{N_{\text{layer}}}{\Big [}\sum \limits _{i=1}^{N_{A}}x_{i}^{A}g_{i}^{A}-L^{A}{\Big ]}-\sum \limits _{A=1}^{N_{\text{layer}}-1}\sum \limits _{i=1}^{N_{A+1}}\sum \limits _{j=1}^{N_{A}}g_{i}^{A+1}\xi _{ij}^{(A+1,A)}g_{j}^{A}}$

(17)

If the Lagrangian functions, or equivalently the activation functions, are chosen in such a way that the Hessians for each layer are positive semi-definite and the overall energy is bounded from below, this system is guaranteed to converge to a fixed point attractor state. The temporal derivative of this energy function is given by[25]

${\displaystyle {\frac {dE}{dt}}=-\sum \limits _{A=1}^{N_{\text{layer}}}\tau _{A}\sum \limits _{i,j=1}^{N_{A}}{\frac {dx_{j}^{A}}{dt}}{\frac {\partial ^{2}L^{A}}{\partial x_{j}^{A}\partial x_{i}^{A}}}{\frac {dx_{i}^{A}}{dt}}\leq 0}$

(18)

Thus, the hierarchical layered network is indeed an attractor network with the global energy function. This network is described by a hierarchical set of synaptic weights that can be learned for each specific problem.

## References

1. Hopfield, J. J. (1982). "Neural networks and physical systems with emergent collective computational abilities". Proceedings of the National Academy of Sciences 79 (8): 2554–2558. doi:10.1073/pnas.79.8.2554. PMID 6953413. PMC 346238. //www.ncbi.nlm.nih.gov/pmc/articles/PMC346238/.
2. Little, W. A. (1974). "The Existence of Persistent States in the Brain". Mathematical Biosciences 19 (1–2): 101–120. doi:10.1016/0025-5564(74)90031-5.
3. Brush, Stephen G. (1967). "History of the Lenz-Ising Model". Reviews of Modern Physics 39 (4): 883–893. doi:10.1103/RevModPhys.39.883.
4. Hopfield, J. J. (1984). "Neurons with graded response have collective computational properties like those of two-state neurons". Proceedings of the National Academy of Sciences 81 (10): 3088–3092. doi:10.1073/pnas.81.10.3088. PMID 6587342. PMC 345226. //www.ncbi.nlm.nih.gov/pmc/articles/PMC345226/.
5. Amit, D.J. (1992). Modeling Brain Function: The World of Attractor Neural Networks. Cambridge University Press. ISBN 978-0-521-42124-9.
6. Rolls, Edmund T. (2016). Cerebral Cortex: Principles of Operation. Oxford University Press. ISBN 978-0-19-878485-2.
7. Hopfield, J. J. (1982). "Neural networks and physical systems with emergent collective computational abilities". Proceedings of the National Academy of Sciences 79 (8): 2554–2558. doi:10.1073/pnas.79.8.2554. PMID 6953413. PMC 346238. //www.ncbi.nlm.nih.gov/pmc/articles/PMC346238/.
8. MacKay, David J. C. (2003). "42. Hopfield Networks". Information Theory, Inference and Learning Algorithms. Cambridge University Press. p. 508. ISBN 978-0521642989. "This convergence proof depends crucially on the fact that the Hopfield network's connections are symmetric. It also depends on the updates being made asynchronously."
9. Bruck, J. (October 1990). "On the convergence properties of the Hopfield model". Proc. IEEE 78 (10): 1579–85. doi:10.1109/5.58341.
10. Uykan, Z. (September 2020). "On the Working Principle of the Hopfield Neural Networks and its Equivalence to the GADIA in Optimization". IEEE Transactions on Neural Networks and Learning Systems 31 (9): 3294–3304. doi:10.1109/TNNLS.2019.2940920. PMID 31603804.
11. Uykan, Z. (March 2021). "Shadow-Cuts Minimization/Maximization and Complex Hopfield Neural Networks". IEEE Transactions on Neural Networks and Learning Systems 32 (3): 1096–1109. doi:10.1109/TNNLS.2020.2980237. PMID 32310787.
12. Hopfield, J.J.; Tank, D.W. (1985). "Neural computation of decisions in optimization problems". Biological Cybernetics 52 (3): 141–6. doi:10.1007/BF00339943. PMID 4027280.
13. Storkey, A.J.; Valabregue, R. (1999). "The basins of attraction of a new Hopfield learning rule". Neural Networks 12 (6): 869–876. doi:10.1016/S0893-6080(99)00038-6. PMID 12662662.
14. Hebb 1949
15. Storkey, Amos (1997). "Increasing the capacity of a Hopfield network without sacrificing functionality". Artificial Neural Networks – ICANN'97. Lecture Notes in Computer Science. 1327. Springer. pp. 451–6. doi:10.1007/BFb0020196. ISBN 978-3-540-69620-9.
16. Hertz 1991
17. Liou, C.-Y.; Lin, S.-L. (2006). "Finite memory loading in hairy neurons". Natural Computing 5 (1): 15–42. doi:10.1007/s11047-004-5490-x.
18. Liou, C.-Y.; Yuan, S.-K. (1999). "Error Tolerant Associative Memory". Biological Cybernetics 81 (4): 331–342. doi:10.1007/s004220050566. PMID 10541936.
19. Yuan, S.-K. (June 1997). Expanding basins of attraction of the associative memory (Master thesis). National Taiwan University. 991010725609704786.
20. ABOUDIB, Ala; GRIPON, Vincent; JIANG, Xiaoran (2014). "A study of retrieval algorithms of sparse messages in networks of neural cliques". COGNITIVE 2014 : The 6th International Conference on Advanced Cognitive Technologies and Applications. pp. 140–6. Bibcode: 2013arXiv1308.4506A.
21. Krotov, Dmitry; Hopfield, John (2016). "Dense Associative Memory for Pattern Recognition". Neural Information Processing Systems 29: 1172–1180.
22. Ramsauer, Hubert et al. (2021). "Hopfield Networks is All You Need". International Conference on Learning Representations.
23. Mete, Demircigil et al. (2017). "On a model of associative memory with huge storage capacity.". Journal of Statistical Physics 168 (2): 288–299. doi:10.1007/s10955-017-1806-y.
24. Krotov, Dmitry; Hopfield, John (2021). "Large associative memory problem in neurobiology and machine learning". International Conference on Learning Representations.
25. Krotov, Dmitry (2021). "Hierarchical Associative Memory". arXiv:2107.06446 [cs.NE].