Delay reduction hypothesis

Last updated

In operant conditioning, the delay reduction hypothesis (DRH; also known as delay reduction theory) is a quantitative description of how choice among concurrently available chained schedules of reinforcement is allocated. The hypothesis states that the greater improvement in temporal proximity to reinforcement (delay reduction) correlated with the onset of a stimulus, the more effectively that stimulus will function as a conditional reinforcer. [1]

Contents

The hypothesis was originally formulated to describe choice behaviour among concurrently available chained schedules of reinforcement; [2] however, the basic principle of delay reduction as the basis for determining a stimulus’ conditionally reinforcing function can be applied more generally to other research areas. [1] [3] [4]

A variety of empirical data corroborate and are consistent with DRH and it represents one of the most substantiated accounts of conditional reinforcement to date. [5]

Application to Concurrent Chain Schedules

Given two concurrently available chained schedules of reinforcement, and represent the number of responses made during alternative A and B’s initial link stimulus.

and represent the average duration of each choice’s respective terminal link. is the average duration to terminal reinforcement from the onset of either initial link stimulus.

The expression represents the delay reduction on a given alternative.

Extensions to the Original Model

Squires and Fantino (1971)

The original formulation by Fantino predicted that choices with equivalent terminal link durations would produce equal allocation of responding (e.g., 0.5 across two choices) regardless the duration of the initial links. [2] Squires and Fantino (1971) proposed including the rate of terminal reinforcement on each choice alternative. [6]

The rate of terminal reinforcement is where is the average duration of an initial link and is the number of terminal reinforcements obtained during a single entry to a terminal link. A critical prediction of this formulation is that matching is obtained when the terminal links are equal durations.

See also

Related Research Articles

<span class="mw-page-title-main">B. F. Skinner</span> American psychologist and social philosopher (1904–1990)

Burrhus Frederic Skinner was an American psychologist, behaviorist, inventor, and social philosopher. Considered the father of Behaviorism, he was the Edgar Pierce Professor of Psychology at Harvard University from 1958 until his retirement in 1974.

In electrochemistry, the Nernst equation is a chemical thermodynamical relationship that permits the calculation of the reduction potential of a reaction from the standard electrode potential, absolute temperature, the number of electrons involved in the redox reaction, and activities of the chemical species undergoing reduction and oxidation respectively. It was named after Walther Nernst, a German physical chemist who formulated the equation.

Operant conditioning, also called instrumental conditioning, is a learning process where voluntary behaviors are modified by association with the addition of reward or aversive stimuli. The frequency or duration of the behavior may increase through reinforcement or decrease through punishment or extinction.

<span class="mw-page-title-main">Operant conditioning chamber</span> Laboratory apparatus used to study animal behavior

An operant conditioning chamber is a laboratory apparatus used to study animal behavior. The operant conditioning chamber was created by B. F. Skinner while he was a graduate student at Harvard University. The chamber can be used to study both operant conditioning and classical conditioning.

Classical conditioning is a behavioral procedure in which a biologically potent physiological stimulus is paired with a neutral stimulus. The term classical conditioning refers to the process of an automatic, conditioned response that is paired with a specific stimulus.

<span class="mw-page-title-main">Reinforcement</span> Consequence affecting an organisms future behavior

In behavioral psychology, reinforcement is any consequence that increases the likelihood of an organism's future behavior whenever that behavior is preceded by a particular antecedent stimulus. For example, a rat can be trained to push a lever to receive food whenever a light is turned on. In this example, the light is the antecedent stimulus, the lever pushing is the behavior, and the food is the reinforcement. Likewise, a student that receives attention and praise when answering a teacher's question will be more likely to answer future questions in class. The teacher's question is the antecedent, the student's response is the behavior, and the praise and attention are the reinforcements.

The experimental analysis of behavior is a science that studies the behavior of individuals across a variety of species. A key early scientist was B. F. Skinner who discovered operant behavior, reinforcers, secondary reinforcers, contingencies of reinforcement, stimulus control, shaping, intermittent schedules, discrimination, and generalization. A central method was the examination of functional relations between environment and behavior, as opposed to hypothetico-deductive learning theory that had grown up in the comparative psychology of the 1920–1950 period. Skinner's approach was characterized by observation of measurable behavior which could be predicted and controlled. It owed its early success to the effectiveness of Skinner's procedures of operant conditioning, both in the laboratory and in behavior therapy.

Behaviorism is a systematic approach to understand the behavior of humans and other animals. It assumes that behavior is either a reflex evoked by the pairing of certain antecedent stimuli in the environment, or a consequence of that individual's history, including especially reinforcement and punishment contingencies, together with the individual's current motivational state and controlling stimuli. Although behaviorists generally accept the important role of heredity in determining behavior, they focus primarily on environmental events. The cognitive revolution of the late 20th century largely replaced behaviorism as an explanatory theory with cognitive psychology, which unlike behaviorism examines internal mental states.

<span class="mw-page-title-main">Conditional independence</span> Probability theory concept

In probability theory, conditional independence describes situations wherein an observation is irrelevant or redundant when evaluating the certainty of a hypothesis. Conditional independence is usually formulated in terms of conditional probability, as a special case where the probability of the hypothesis given the uninformative observation is equal to the probability without. If is the hypothesis, and and are observations, conditional independence can be stated as an equality:

In economics, hyperbolic discounting is a time-inconsistent model of delay discounting. It is one of the cornerstones of behavioral economics and its brain-basis is actively being studied by neuroeconomics researchers.

Extinction is a behavioral phenomenon observed in both operantly conditioned and classically conditioned behavior, which manifests itself by fading of non-reinforced conditioned response over time. When operant behavior that has been previously reinforced no longer produces reinforcing consequences the behavior gradually stops occurring. In classical conditioning, when a conditioned stimulus is presented alone, so that it no longer predicts the coming of the unconditioned stimulus, conditioned responding gradually stops. For example, after Pavlov's dog was conditioned to salivate at the sound of a metronome, it eventually stopped salivating to the metronome after the metronome had been sounded repeatedly but no food came. Many anxiety disorders such as post traumatic stress disorder are believed to reflect, at least in part, a failure to extinguish conditioned fear.

In operant conditioning, the matching law is a quantitative relationship that holds between the relative rates of response and the relative rates of reinforcement in concurrent schedules of reinforcement. For example, if two response alternatives A and B are offered to an organism, the ratio of response rates to A and B equals the ratio of reinforcements yielded by each response. This law applies fairly well when non-human subjects are exposed to concurrent variable interval schedules ; its applicability in other situations is less clear, depending on the assumptions made and the details of the experimental situation. The generality of applicability of the matching law is subject of current debate.

Behavioral momentum is a theory in quantitative analysis of behavior and is a behavioral metaphor based on physical momentum. It describes the general relation between resistance to change and the rate of reinforcement obtained in a given situation.

In behavioral psychology, stimulus control is a phenomenon in operant conditioning that occurs when an organism behaves in one way in the presence of a given stimulus and another way in its absence. A stimulus that modifies behavior in this manner is either a discriminative stimulus or stimulus delta. For example, the presence of a stop sign at a traffic intersection alerts the driver to stop driving and increases the probability that braking behavior occurs. Stimulus control does not force behavior to occur, as it is a direct result of historical reinforcement contingencies, as opposed to reflexive behavior elicited through classical conditioning.

Two-alternative forced choice (2AFC) is a method for measuring the sensitivity of a person or animal to some particular sensory input, stimulus, through that observer's pattern of choices and response times to two versions of the sensory input. For example, to determine a person's sensitivity to dim light, the observer would be presented with a series of trials in which a dim light was randomly either in the top or bottom of the display. After each trial, the observer responds "top" or "bottom". The observer is not allowed to say "I do not know", or "I am not sure", or "I did not see anything". In that sense the observer's choice is forced between the two alternatives.

Beliefs depend on the available information. This idea is formalized in probability theory by conditioning. Conditional probabilities, conditional expectations, and conditional probability distributions are treated on three levels: discrete probabilities, probability density functions, and measure theory. Conditioning leads to a non-random result if the condition is completely specified; otherwise, if the condition is left random, the result of conditioning is also random.

Murray Sidman was an American behavioral scientist, best known for Sidman Avoidance, also called "free-operant avoidance", in which an organism learns to avoid an aversive stimulus by engaging in a targeted response in the absence of stimuli indicating whether the aversive stimulus is forthcoming. Sidman's explanation of free-operant avoidance is an alternative to the Miller-Mowrer two-process theory of avoidance.

<span class="mw-page-title-main">Conditional probability</span> Probability of an event occurring, given that another event has already occurred

In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This particular method relies on event A occurring with some sort of relationship with another event B. In this situation, the event A can be analyzed by a conditional probability with respect to B. If the event of interest is A and the event B is known or assumed to have occurred, "the conditional probability of A given B", or "the probability of A under the condition B", is usually written as P(A|B) or occasionally PB(A). This can also be understood as the fraction of probability B that intersects with A, or the ratio of the probabilities of both events happening to the "given" one happening (how many times A occurs rather than not assuming B has occurred): .

James "Jim" A. Dinsmoor was an American experimental psychologist who published work in the field of the experimental analysis of behavior.


Human contingency learning (HCL) is the observation that people tend to acquire knowledge based on whichever outcome has the highest probability of occurring from particular stimuli. In other words, individuals gather associations between a certain behaviour and a specific consequence. It is a form of learning for many organisms.

References

  1. 1 2 Fantino, E. (1977). Conditioned reinforcement: Choice and information. In W. K. Honig & J. E. R. Staddon (Eds.), Handbook of operant behavior (pp. 313–339). Prentice-Hall
  2. 1 2 Fantino, E. (1969). Choice and rate of reinforcement. Journal of the Experimental Analysis of Behavior, 12 (5), 723–730. https://doi.org/10.1901/jeab.1969.12-723
  3. Fantino, E. (2012). Optimal and non-optimal behavior across species. Comparative Cognition & Behavior Reviews, 7, 44-54. https://doi.org/10.3819/ccbr.2012.70003
  4. Shahan, T. A., & Cunningham, P. (2015). Conditioned reinforcement and information theory reconsidered. Journal of the Experimental Analysis of Behavior, 103 (2), 405–418. https://doi.org/10.1002/jeab.142
  5. Williams B. A. (1994). Conditioned reinforcement: Neglected or outmoded explanatory construct?. Psychonomic Bulletin & Review, 1(4), 457–475. https://doi.org/10.3758/BF03210950
  6. Squires, N., & Fantino, E. (1971). A model for choice in simple concurrent and concurrent-chains schedules. Journal of the Experimental Analysis of Behavior, 15 (1), 27–38. https://doi.org/10.1901/jeab.1971.15-27