Attractor network

Last updated

An attractor network is a type of recurrent dynamical network, that evolves toward a stable pattern over time. Nodes in the attractor network converge toward a pattern that may either be fixed-point (a single state), cyclic (with regularly recurring states), chaotic (locally but not globally unstable) or random (stochastic). [1] Attractor networks have largely been used in computational neuroscience to model neuronal processes such as associative memory [2] and motor behavior, as well as in biologically inspired methods of machine learning. An attractor network contains a set of n nodes, which can be represented as vectors in a d-dimensional space where n>d. Over time, the network state tends toward one of a set of predefined states on a d-manifold; these are the attractors.

Contents

Overview

In attractor networks, an attractor (or attracting set) is a closed subset of states A toward which the system of nodes evolves. A stationary attractor is a state or sets of states where the global dynamics of the network stabilize. Cyclic attractors evolve the network toward a set of states in a limit cycle, which is repeatedly traversed. Chaotic attractors are non-repeating bounded attractors that are continuously traversed.

The network state space is the set of all possible node states. The attractor space is the set of nodes on the attractor. Attractor networks are initialized based on the input pattern. The dimensionality of the input pattern may differ from the dimensionality of the network nodes. The trajectory of the network consists of the set of states along the evolution path as the network converges toward the attractor state. The basin of attraction is the set of states that results in movement towards a certain attractor. [1]

Types

Various types of attractors may be used to model different types of network dynamics. While fixed-point attractor networks are the most common (originating from Hopfield networks [3] ), other types of networks are also examined.

Fixed point attractors

The fixed point attractor naturally follows from the Hopfield network. Conventionally, fixed points in this model represent encoded memories. These models have been used to explain associative memory, classification, and pattern completion. Hopfield nets contain an underlying energy function [4] that allow the network to asymptotically approach a stationary state. One class of point attractor network is initialized with an input, after which the input is removed and the network moves toward a stable state. Another class of attractor network features predefined weights that are probed by different types of input. If this stable state is different during and after the input, it serves as a model of associative memory. However, if the states during and after input do not differ, the network can be used for pattern completion.

Other stationary attractors

Line attractors and plane attractors are used in the study of oculomotor control. These line attractors, or neural integrators, describe eye position in response to stimuli. Ring attractors have been used to model rodent head direction.

Cyclic attractors

Cyclic attractors are instrumental in modelling central pattern generators, neurons that govern oscillatory activity in animals such as chewing, walking, and breathing.

Chaotic attractors

Chaotic attractors (also called strange attractors) have been hypothesized to reflect patterns in odor recognition. While chaotic attractors have the benefit of more quickly converging upon limit cycles, there is yet no experimental evidence to support this theory. [5]

Continuous attractors

Neighboring stable states (fix points) of continuous attractors (also called continuous attractor neural networks) code for neighboring values of a continuous variable such as head direction or actual position in space.

Ring attractors

A subtype of continuous attractors with a particular topology of the neurons (ring for 1-dimensional and torus or twisted torus for 2-dimensional networks). The observed activity of grid cells is successfully explained by assuming the presence of ring attractors in the medial entorhinal cortex. [6] Recently, it has been proposed that similar ring attractors are present in the lateral portion of the entorhinal cortex and their role extends to registering new episodic memories. [7]

Implementations

Attractor networks have mainly been implemented as memory models using fixed-point attractors. However, they have been largely impractical for computational purposes because of difficulties in designing the attractor landscape and network wiring, resulting in spurious attractors and poorly conditioned basins of attraction. Furthermore, training on attractor networks is generally computationally expensive, compared to other methods such as k-nearest neighbor classifiers. [8] However, their role in general understanding of different biological functions, such as, locomotor function, memory, decision-making, to name a few, makes them more attractive as biologically realistic models.

Hopfield networks

Hopfield attractor networks are an early implementation of attractor networks with associative memory. These recurrent networks are initialized by the input, and tend toward a fixed-point attractor. The update function in discrete time is , where is a vector of nodes in the network and is a symmetric matrix describing their connectivity. The continuous time update is .

Bidirectional networks are similar to Hopfield networks, with the special case that the matrix is a block matrix. [4]

Localist attractor networks

Zemel and Mozer (2001) [8] proposed a method to reduce the number of spurious attractors that arise from the encoding of multiple attractors by each connection in the network. Localist attractor networks encode knowledge locally by implementing an expectation-maximization algorithm on a mixture-of-gaussians representing the attractors, to minimize the free energy in the network and converge only the most relevant attractor. This results in the following update equations:

  1. Determine the activity of attractors:
  2. Determine the next state of the network:
  3. Determine the attractor width through network:

( denotes basin strength, denotes the center of the basin. denotes input to the net.)

The network is then re-observed, and the above steps repeat until convergence. The model also reflects two biologically relevant concepts. The change in models stimulus priming by allowing quicker convergence toward a recently visited attractor. Furthermore, the summed activity of attractors allows a gang effect that causes two nearby attractors to mutually reinforce the other's basin.

Reconsolidation attractor networks

Siegelmann (2008) [9] generalized the localist attractor network model to include the tuning of attractors themselves. This algorithm uses the EM method above, with the following modifications: (1) early termination of the algorithm when the attractor's activity is most distributed, or when high entropy suggests a need for additional memories, and (2) the ability to update the attractors themselves: , where is the step size parameter of the change of . This model reflects memory reconsolidation in animals, and shows some of the same dynamics as those found in memory experiments.

Further developments in attractor networks, such as kernel based attractor networks, [10] have improved the computational feasibility of attractor networks as a learning algorithm, while maintaining the high-level flexibility to perform pattern completion on complex compositional structures.

Related Research Articles

Artificial neural network Computational model used in machine learning, based on connected, hierarchical functions

Artificial neural networks (ANNs), usually simply called neural networks (NNs), are computing systems vaguely inspired by the biological neural networks that constitute animal brains.

Perceptron

In machine learning, the perceptron is an algorithm for supervised learning of binary classifiers. A binary classifier is a function which can decide whether or not an input, represented by a vector of numbers, belongs to some specific class. It is a type of linear classifier, i.e. a classification algorithm that makes its predictions based on a linear predictor function combining a set of weights with the feature vector.

Unsupervised learning Machine learning technique

Unsupervised learning (UL) is a type of algorithm that learns patterns from untagged data. The hope is that, through mimicry, the machine is forced to build a compact internal representation of its world and then generate imaginative content. In contrast to supervised learning (SL) where data is tagged by a human, e.g. as "car" or "fish" etc, UL exhibits self-organization that captures patterns as neuronal predilections or probability densities. The other levels in the supervision spectrum are reinforcement learning where the machine is given only a numerical performance score as its guidance, and semi-supervised learning where a smaller portion of the data is tagged. Two broad methods in UL are Neural Networks and Probabilistic Methods.

Hebbian theory is a neuroscientific theory claiming that an increase in synaptic efficacy arises from a presynaptic cell's repeated and persistent stimulation of a postsynaptic cell. It is an attempt to explain synaptic plasticity, the adaptation of brain neurons during the learning process. It was introduced by Donald Hebb in his 1949 book The Organization of Behavior. The theory is also called Hebb's rule, Hebb's postulate, and cell assembly theory. Hebb states it as follows:

Let us assume that the persistence or repetition of a reverberatory activity tends to induce lasting cellular changes that add to its stability. ... When an axon of cell A is near enough to excite a cell B and repea cells firing B, is increased.

Boltzmann machine

A Boltzmann machine is a type of stochastic recurrent neural network. It is a Markov random field. It was translated from statistical physics for use in cognitive science. The Boltzmann machine is based on a stochastic spin-glass model with an external field, i.e., a Sherrington–Kirkpatrick model that is a stochastic Ising Model and applied to machine learning.

A Hopfield network is a form of recurrent artificial neural network and a type of spin glass system popularised by John Hopfield in 1982 as described earlier by Little in 1974 based on Ernst Ising's work with Wilhelm Lenz on the Ising model. Hopfield networks serve as content-addressable ("associative") memory systems with binary threshold nodes. Hopfield networks also provide a model for understanding human memory.

Recurrent neural network Computational model used in machine learning

A recurrent neural network (RNN) is a class of artificial neural networks where connections between nodes form a directed graph along a temporal sequence. This allows it to exhibit temporal dynamic behavior. Derived from feedforward neural networks, RNNs can use their internal state (memory) to process variable length sequences of inputs. This makes them applicable to tasks such as unsegmented, connected handwriting recognition or speech recognition.

Feedforward neural network Type of artificial neural network

A feedforward neural network is an artificial neural network wherein connections between the nodes do not form a cycle. As such, it is different from its descendant: recurrent neural networks.

Neural gas is an artificial neural network, inspired by the self-organizing map and introduced in 1991 by Thomas Martinetz and Klaus Schulten. The neural gas is a simple algorithm for finding optimal data representations based on feature vectors. The algorithm was coined "neural gas" because of the dynamics of the feature vectors during the adaptation process, which distribute themselves like a gas within the data space. It is applied where data compression or vector quantization is an issue, for example speech recognition, image processing or pattern recognition. As a robustly converging alternative to the k-means clustering it is also used for cluster analysis.

Quantum neural network Quantum Mechanics in Nueral Network

Quantum neural networks are computational neural network models which are based on the principles of quantum mechanics. The first ideas on quantum neural computation were published independently in 1995 by Subhash Kak and Ron Chrisley, engaging with the theory of quantum mind, which posits that quantum effects play a role in cognitive function. However, typical research in quantum neural networks involves combining classical artificial neural network models with the advantages of quantum information in order to develop more efficient algorithms. One important motivation for these investigations is the difficulty to train classical neural networks, especially in big data applications. The hope is that features of quantum computing such as quantum parallelism or the effects of interference and entanglement can be used as resources. Since the technological implementation of a quantum computer is still in a premature stage, such quantum neural network models are mostly theoretical proposals that await their full implementation in physical experiments.

The softmax function, also known as softargmax or normalized exponential function, is a generalization of the logistic function to multiple dimensions. It is used in multinomial logistic regression and is often used as the last activation function of a neural network to normalize the output of a network to a probability distribution over predicted output classes, based on Luce's choice axiom.

Autoassociative memory, also known as auto-association memory or an autoassociation network, is any type of memory that is able to retrieve a piece of data from only a tiny sample of itself. They are very effective in de-noising or removing interference from the input and can be used to determine whether the given input is “known” or “unknown”.

Bidirectional associative memory (BAM) is a type of recurrent neural network. BAM was introduced by Bart Kosko in 1988. There are two types of associative memory, auto-associative and hetero-associative. BAM is hetero-associative, meaning given a pattern it can return another pattern which is potentially of a different size. It is similar to the Hopfield network in that they are both forms of associative memory. However, Hopfield nets return patterns of the same size.

Competitive learning is a form of unsupervised learning in artificial neural networks, in which nodes compete for the right to respond to a subset of the input data. A variant of Hebbian learning, competitive learning works by increasing the specialization of each node in the network. It is well suited to finding clusters within data.

There are many types of artificial neural networks (ANN).

Rulkov map

The Rulkov map is a two-dimensional iterated map used to model a biological neuron. It was proposed by Nikolai F. Rulkov in 2001. The use of this map to study neural networks has computational advantages because the map is easier to iterate than a continuous dynamical system. This saves memory and simplifies the computation of large neural networks.

Extreme learning machine Type of artificial neural network

Extreme learning machines are feedforward neural networks for classification, regression, clustering, sparse approximation, compression and feature learning with a single layer or multiple layers of hidden nodes, where the parameters of hidden nodes need not be tuned. These hidden nodes can be randomly assigned and never updated, or can be inherited from their ancestors without being changed. In most cases, the output weights of hidden nodes are usually learned in a single step, which essentially amounts to learning a linear model. The name "extreme learning machine" (ELM) was given to such models by its main inventor Guang-Bin Huang.

Fusion adaptive resonance theory is a generalization of self-organizing neural networks known as Adaptive Resonance Theory for learning recognition categories across multiple pattern channels.

Transformer (machine learning model) Machine learning algorithm used for natural language processing

A transformer is a deep learning model that adopts the mechanism of attention, weighing the influence of different parts of the input data. It is used primarily in the field of natural language processing (NLP), but increasingly in computer vision (CV) as well.

The spike response model (SRM) is a spiking neuron model in which spikes are generated by either a deterministic or a stochastic threshold process. In the SRM, the membrane voltage V is described as a linear sum of the postsynaptic potentials (PSPs) caused by spike arrivals to which the effects of refractoriness and adaptation are added. The threshold is either fixed or dynamic. In the latter case it increases after each spike. The SRM is flexible enough to account for a variety of neuronal firing pattern in response to step current input. The SRM has also been used in the theory of computation to quantify the capacity of spiking neural networks; and in the neurosciences to predict the subthreshold voltage and the firing times of cortical neurons during stimulation with a time-dependent current stimulation. The name Spike Response Model points to the property that the two important filters and of the model can be interpreted as the response of the membrane potential to an incoming spike and to an outgoing spike. The SRM has been formulated in continuous time and in discrete time. The SRM can be viewed as a generalized linear model (GLM) or as an a generalized integrate-and-fire model with adaptation.

References

  1. 1 2
    • Amit, D. J. (1989). Modeling brain function: The world of attractor neural networks. New York, NY: Cambridge University Press.
  2. 1 2
  3. McNaughton BL, Battaglia FP, Jensen O, Moser EI, Moser MB (August 2006). "Path integration and the neural basis of the "cognitive map"". Nat. Rev. Neurosci. 7 (8): 663–678. doi:10.1038/nrn1932. PMID   16858394. S2CID   16928213.
  4. Kovács KA (September 2020). "Episodic Memories: How do the Hippocampus and the Entorhinal Ring Attractors Cooperate to Create Them?". Frontiers in Systems Neuroscience. 14: 68. doi: 10.3389/fnsys.2020.559186 . PMC   7511719 . PMID   33013334.
  5. 1 2