Anti-Hebbian learning

Last updated

In neuroethology and the study of learning, anti-Hebbian learning describes a particular class of learning rule by which synaptic plasticity can be controlled. These rules are based on a reversal of Hebb's postulate, and therefore can be simplistically understood as dictating reduction of the strength of synaptic connectivity between neurons following a scenario in which a neuron directly contributes to production of an action potential in another neuron.

Contents

Evidence from neuroethology

Neuroethological study has provided strong evidence for the existence of a system which adheres to an anti-Hebbian learning rule. Research on the mormyrid electric fish has demonstrated that the electrosensory lateral-line lobe (ELL) receives sensory input from knollenorgans (electroreceptive sensory organs) which utilize a self-generated electrical discharge (called an EOD; electric organ discharge) to extract information from the environment about objects in close proximity to the fish.

In addition to information from sensory receptors, the ELL receives a signal from the area of the brain responsible for initiating the electrical discharges, known as the EOD command nucleus. This efference copy diverges, transmitted through two separate pathways, before the signals converge along with electrosensory input on Purkinje-like Medium Ganglion cells in the ELL. These cells receive information through extensive apical dendritic projections from parallel fibers that signal the transmission of an order to release an EOD. These cells also receive information from neurons conveying electrosensory information.

Important to anti-Hebbian learning, the synapses between the parallel fibers and the apical dendrites of Medium Ganglion cells show a specific pattern of synaptic plasticity. Should activation of the dendrites by parallel fibers occur in a short time period preceding the initiation of a dendritic broad spike (an action potential which travels through the dendrites), the strength of the connection between the neurons at these synapses will be reduced. Activation by the parallel fibers in all other circumstances – including activation significantly preceding as well as any activation following the broad spike – will result in the strengthening of the synapse.

Significance

Since the neurons of the ELL receive both a corollary discharge (another term for an efference copy) of the motor output commands sent to the EOD, and afferent input from the electrosensory receptors, the animal is able to eliminate predictable inputs produced by its own motor output. The system is able to filter the expected input from the EOD, while signals which are unexpected, arriving at odd intervals with regard to the motor command are effectively strengthened by the learning rule. This allows the extraction of information about objects which cause an alteration in the flow of the electric field around the fish, highlighting changes while discarding uninformative sensory inputs.

The adaptation of these synapses, though, will only increase the strength of a synaptic connection until the resulting excitation aids in activation of a broad-spike wave. As a result, if changes in external environment are consistent, the connections between the neurons previously described will reach a level at which excitation, similar to the initial state, is once again held at a threshold, so that slight changes in the incoming sensory information will result in contribution to broad-spike initiation. In this manner, the organism is able to learn to ignore redundant sensory information in the environment. The eventual desensitization to these consistencies is essential to prevent excessive noise from masking important sensory information. Numerous potential causes which could result in a consistent alteration in the reception of EOD signals include: growth, changes in water conductance (salinity), low water levels (where the shallow bottom of the body of water would interfere with electrical currents), and possibly injuries.

Predicted application

Synaptic plasticity operating under the control of an anti-Hebbian learning rule is thought to occur in the cerebellum. Understanding the operation of neural learning could provide valuable insights for the treatment of cerebellar-related disorders. The knowledge could also serve a significant function in the computer-based collection of data, repeatedly adjusting to redundant inputs while emphasizing the appearance of alterations.

Example

The following algorithm was proposed by (Péter Földiák, 1990). [1] In this algorithm, there are input neurons and output neurons . Each input neuron is connected to an output neuron by a weight . Also, the output neurons are connected to each other by weights in a symmetric weight matrix . The neural activation of the output neurons is defined as a fixed point of the following stable state equation:where is the logistic activation function, and is the activation threshold of the -th output neuron.

This can be more succinctly written in matrix notation as .

The algorithm is as follows:

Here, , , , and are hyperparameters. can be interpreted as the target probability for each output neuron to be activated. If a certain input pattern is very common, the anti-Hebbian term forces its output to be very sparse. In this way, the anti-Hebbian term automatically generates efficient sparse codes, analogous to an emergent Huffman code.

See also

Related Research Articles

<span class="mw-page-title-main">Artificial neuron</span> Mathematical function conceived as a crude model

An artificial neuron is a mathematical function conceived as a model of a biological neuron in a neural network. Artificial neuron is the elementary unit of an artificial neural network. The artificial neuron is a function that receives one or more inputs, applies weights to these inputs, and sums them to produce an output.

<span class="mw-page-title-main">Long-term potentiation</span> Persistent strengthening of synapses based on recent patterns of activity

In neuroscience, long-term potentiation (LTP) is a persistent strengthening of synapses based on recent patterns of activity. These are patterns of synaptic activity that produce a long-lasting increase in signal transmission between two neurons. The opposite of LTP is long-term depression, which produces a long-lasting decrease in synaptic strength.

Hebbian theory is a neuropsychological theory claiming that an increase in synaptic efficacy arises from a presynaptic cell's repeated and persistent stimulation of a postsynaptic cell. It is an attempt to explain synaptic plasticity, the adaptation of brain neurons during the learning process. It was introduced by Donald Hebb in his 1949 book The Organization of Behavior. The theory is also called Hebb's rule, Hebb's postulate, and cell assembly theory. Hebb states it as follows:

Let us assume that the persistence or repetition of a reverberatory activity tends to induce lasting cellular changes that add to its stability. ... When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.

<span class="mw-page-title-main">Synaptic plasticity</span> Ability of a synapse to strengthen or weaken over time according to its activity

In neuroscience, synaptic plasticity is the ability of synapses to strengthen or weaken over time, in response to increases or decreases in their activity. Since memories are postulated to be represented by vastly interconnected neural circuits in the brain, synaptic plasticity is one of the important neurochemical foundations of learning and memory.

Spike-timing-dependent plasticity (STDP) is a biological process that adjusts the strength of connections between neurons in the brain. The process adjusts the connection strengths based on the relative timing of a particular neuron's output and input action potentials. The STDP process partially explains the activity-dependent development of nervous systems, especially with regard to long-term potentiation and long-term depression.

A Hopfield network is a form of recurrent neural network, or a spin glass system, that can serve as a content-addressable memory. The Hopfield network, named for John Hopfield, consists of a single layer of neurons, where each neuron is connected to every other neuron except itself. These connections are bidirectional and symmetric, meaning the weight of the connection from neuron i to neuron j is the same as the weight from neuron j to neuron i. Patterns are associatively recalled by fixing certain inputs, and dynamically evolve the network to minimize an energy function, towards local energy minimum states that correspond to stored patterns. Patterns are associatively learned by a Hebbian learning algorithm.

In machine learning, backpropagation is a gradient estimation method commonly used for training neural networks to compute the network parameter updates.

<span class="mw-page-title-main">Neural circuit</span> Network or circuit of neurons

A neural circuit is a population of neurons interconnected by synapses to carry out a specific function when activated. Multiple neural circuits interconnect with one another to form large scale brain networks.

In neuroscience, homeostatic plasticity refers to the capacity of neurons to regulate their own excitability relative to network activity. The term homeostatic plasticity derives from two opposing concepts: 'homeostatic' and plasticity, thus homeostatic plasticity means "staying the same through change". In the nervous system, neurons must be able to evolve with the development of their constantly changing environment while simultaneously staying the same amidst this change. This stability is important for neurons to maintain their activity and functionality to prevent neurons from carcinogenesis. At the same time, neurons need to have flexibility to adapt to changes and make connections to cope with the ever-changing environment of a developing nervous system.

Oja's learning rule, or simply Oja's rule, named after Finnish computer scientist Erkki Oja, is a model of how neurons in the brain or in artificial neural networks change connection strength, or learn, over time. It is a modification of the standard Hebb's Rule that, through multiplicative normalization, solves all stability problems and generates an algorithm for principal components analysis. This is a computational form of an effect which is believed to happen in biological neurons.

Neural cryptography is a branch of cryptography dedicated to analyzing the application of stochastic algorithms, especially artificial neural network algorithms, for use in encryption and cryptanalysis.

Bienenstock–Cooper–Munro (BCM) theory, BCM synaptic modification, or the BCM rule, named after Elie Bienenstock, Leon Cooper, and Paul Munro, is a physical theory of learning in the visual cortex developed in 1981. The BCM model proposes a sliding threshold for long-term potentiation (LTP) or long-term depression (LTD) induction, and states that synaptic plasticity is stabilized by a dynamic adaptation of the time-averaged postsynaptic activity. According to the BCM model, when a pre-synaptic neuron fires, the post-synaptic neurons will tend to undergo LTP if it is in a high-activity state, or LTD if it is in a lower-activity state. This theory is often used to explain how cortical neurons can undergo both LTP or LTD depending on different conditioning stimulus protocols applied to pre-synaptic neurons.

The generalized Hebbian algorithm, also known in the literature as Sanger's rule, is a linear feedforward neural network for unsupervised learning with applications primarily in principal components analysis. First defined in 1989, it is similar to Oja's rule in its formulation and stability, except it can be applied to networks with multiple outputs. The name originates because of the similarity between the algorithm and a hypothesis made by Donald Hebb about the way in which synaptic strengths in the brain are modified in response to experience, i.e., that changes are proportional to the correlation between the firing of pre- and post-synaptic neurons.

In neuroscience and computer science, synaptic weight refers to the strength or amplitude of a connection between two nodes, corresponding in biology to the amount of influence the firing of one neuron has on another. The term is typically used in artificial and biological neural network research.

Activity-dependent plasticity is a form of functional and structural neuroplasticity that arises from the use of cognitive functions and personal experience; hence, it is the biological basis for learning and the formation of new memories. Activity-dependent plasticity is a form of neuroplasticity that arises from intrinsic or endogenous activity, as opposed to forms of neuroplasticity that arise from extrinsic or exogenous factors, such as electrical brain stimulation- or drug-induced neuroplasticity. The brain's ability to remodel itself forms the basis of the brain's capacity to retain memories, improve motor function, and enhance comprehension and speech amongst other things. It is this trait to retain and form memories that is associated with neural plasticity and therefore many of the functions individuals perform on a daily basis. This plasticity occurs as a result of changes in gene expression which are triggered by signaling cascades that are activated by various signaling molecules during increased neuronal activity.

Competitive learning is a form of unsupervised learning in artificial neural networks, in which nodes compete for the right to respond to a subset of the input data. A variant of Hebbian learning, competitive learning works by increasing the specialization of each node in the network. It is well suited to finding clusters within data.

A Bayesian Confidence Propagation Neural Network (BCPNN) is an artificial neural network inspired by Bayes' theorem, which regards neural computation and processing as probabilistic inference. Neural unit activations represent probability ("confidence") in the presence of input features or categories, synaptic weights are based on estimated correlations and the spread of activation corresponds to calculating posterior probabilities. It was originally proposed by Anders Lansner and Örjan Ekeberg at KTH Royal Institute of Technology. This probabilistic neural network model can also be run in generative mode to produce spontaneous activations and temporal sequences.

<span class="mw-page-title-main">Heterosynaptic plasticity</span>

Synaptic plasticity refers to a chemical synapse's ability to undergo changes in strength. Synaptic plasticity is typically input-specific, meaning that the activity in a particular neuron alters the efficacy of a synaptic connection between that neuron and its target. However, in the case of heterosynaptic plasticity, the activity of a particular neuron leads to input unspecific changes in the strength of synaptic connections from other unactivated neurons. A number of distinct forms of heterosynaptic plasticity have been found in a variety of brain regions and organisms. These different forms of heterosynaptic plasticity contribute to a variety of neural processes including associative learning, the development of neural circuits, and homeostasis of synaptic input.

An artificial neural network (ANN) combines biological principles with advanced statistics to solve problems in domains such as pattern recognition and game-play. ANNs adopt the basic model of neuron analogues connected to each other in a variety of ways.

Modern Hopfield networks are generalizations of the classical Hopfield networks that break the linear scaling relationship between the number of input features and the number of stored memories. This is achieved by introducing stronger non-linearities leading to super-linear memory storage capacity as a function of the number of feature neurons. The network still requires a sufficient number of hidden neurons.

References

  1. Földiák, P. (December 1990). "Forming sparse representations by local anti-Hebbian learning". Biological Cybernetics. 64 (2): 165–170. doi:10.1007/BF02331346. ISSN   0340-1200.