Operant conditioning chamber

Last updated

Skinner box Skinner box scheme 01.svg
Skinner box

An operant conditioning chamber (also known as a Skinner box) is a laboratory apparatus used to study animal behavior. The operant conditioning chamber was created by B. F. Skinner while he was a graduate student at Harvard University. The chamber can be used to study both operant conditioning and classical conditioning. [1] [2]

Contents

Skinner created the operant conditioning chamber as a variation of the puzzle box originally created by Edward Thorndike. [3] While Skinner's early studies were done using rats, he later moved on to study pigeons. [4] [5] The operant conditioning chamber may be used to observe or manipulate behaviour. An animal is placed in the box where it must learn to activate levers or respond to light or sound stimuli for reward. The reward may be food or the removal of noxious stimuli such as a loud alarm. The chamber is used to test specific hypotheses in a controlled setting.

Name

Students using a Skinner box UNMSM PsiExperimental 1998 2.jpg
Students using a Skinner box

Skinner was noted to have expressed his distaste for becoming an eponym. [6] It is believed that Clark Hull and his Yale students coined the expression "Skinner box". Skinner stated he did not use the term himself, and went so far as to ask Howard Hunt to use "lever box" instead of "Skinner box" in a published document. [7]

History

Original puzzle box designed by Edward Thorndike Original "Puzzle Box" Apparatus Design.png
Original puzzle box designed by Edward Thorndike

In 1898, American psychologist, Edward Thorndike proposed the 'law of effect', which formed the basis of operant conditioning. [8] Thorndike conducted experiments to discover how cats learn new behaviors. His work involved monitoring cats as they attempted to escape from puzzle boxes. The puzzle box trapped the animals until they moved a lever or performed an action which triggered their release. [9] Thorndike ran several trials and recorded the time it took for them to perform the actions necessary to escape. He discovered that the cats seemed to learn from a trial-and-error process rather than insightful inspections of their environment. The animals learned that their actions led to an effect, and the type of effect influenced whether the behavior would be repeated. Thorndike's 'law of effect' contained the core elements of what would become known as operant conditioning. B. F. Skinner expanded upon Thorndike's existing work. [9] Skinner theorized that if a behavior is followed by a reward, that behavior is more likely to be repeated, but added that if it is followed by some sort of punishment, it is less likely to be repeated. He introduced the word reinforcement into Thorndike's law of effect. [10] Through his experiments, Skinner discovered the law of operant learning which included extinction, punishment and generalization. [10]

Skinner designed the operant conditioning chamber to allow for specific hypothesis testing and behavioural observation. He wanted to create a way to observe animals in a more controlled setting as observation of behaviour in nature can be unpredictable. [2]

Purpose

A rat presses a button in an operant conditioning chamber. Augustin Lignier rat selfie 1 (cropped).jpg
A rat presses a button in an operant conditioning chamber.

An operant conditioning chamber allows researchers to study animal behaviour and response to conditioning. They do this by teaching an animal to perform certain actions (like pressing a lever) in response to specific stimuli. When the correct action is performed the animal receives positive reinforcement in the form of food or other reward. In some cases, the chamber may deliver positive punishment to discourage incorrect responses. For example, researchers have tested certain invertebrates' reaction to operant conditioning using a "heat box". [11] This box has two walls used for manipulation; one wall can undergo temperature change while the other cannot. As soon as the invertebrate crosses over to the side which can undergo temperature change, the researcher will increase the temperature. Eventually, the invertebrate will be conditioned to stay on the side that does not undergo a temperature change. After conditioning, even when the temperature is turned to its lowest setting, the invertebrate will avoid that side of the box. [11]

Skinner's pigeon studies involved a series of levers. When the lever was pressed, the pigeon would receive a food reward. [5] This was made more complex as researchers studied animal learning behaviours. A pigeon would be placed in the conditioning chamber and another one would be placed in an adjacent box separated by a plexiglass wall. The pigeon in the chamber would learn to press the lever to receive food as the other pigeon watched. The pigeons would then be switched, and researchers would observe them for signs of cultural learning.

Structure

On the left are two mechanisms including two levers and light signals. There is a light source and speaker above the box and an electrified floor at the bottom. Skinner box photo 02.jpg
On the left are two mechanisms including two levers and light signals. There is a light source and speaker above the box and an electrified floor at the bottom.

The outside shell of an operant conditioning chamber is a large box big enough to easily accommodate the animal being used as a subject. Commonly used animals include rodents (usually lab rats), pigeons, and primates. The chamber is often sound-proof and light-proof to avoid distracting stimuli.

Operant conditioning chambers have at least one response mechanism that can automatically detect the occurrence of a behavioral response or action (i.e., pecking, pressing, pushing, etc.). This may be a lever or series of lights which the animal will respond to in the presence of stimulus. Typical mechanisms for primates and rats are response levers; if the subject presses the lever, the opposite end closes a switch that is monitored by a computer or other programmed device. [12] Typical mechanisms for pigeons and other birds are response keys with a switch that closes if the bird pecks at the key with sufficient force. [5] The other minimal requirement of an operant conditioning chamber is that it has a means of delivering a primary reinforcer such as a food reward.

A pigeon offering the correct response to stimuli is rewarded with food pellets Operant Conditioning Involves Choice.png
A pigeon offering the correct response to stimuli is rewarded with food pellets

A simple configuration, such as one response mechanism and one feeder, may be used to investigate a variety of psychological phenomena. Modern operant conditioning chambers may have multiple mechanisms, such as several response levers, two or more feeders, and a variety of devices capable of generating different stimuli including lights, sounds, music, figures, and drawings. Some configurations use an LCD panel for the computer generation of a variety of visual stimuli or a set of LED lights to create patterns they wish to be replicated. [13]

Some operant conditioning chambers can also have electrified nets or floors so that shocks can be given to the animals as a positive punishment or lights of different colors that give information about when the food is available as a positive reinforcement. [14]

Research impact

Operant conditioning chambers have become common in a variety of research disciplines especially in animal learning. The chambers design allows for easy monitoring of the animal and provides a space to manipulate certain behaviours. This controlled environment may allow for research and experimentation which cannot be performed in the field.

There are a variety of applications for operant conditioning. For instance, shaping the behavior of a child is influenced by the compliments, comments, approval, and disapproval of one's behavior. [15] An important factor of operant conditioning is its ability to explain learning in real-life situations. From an early age, parents nurture their children's behavior by using reward and praise following an achievement (crawling or taking a first step) which reinforces such behavior. When a child misbehaves, punishment in the form of verbal discouragement or the removal of privileges are used to discourage them from repeating their actions.

Skinner's studies on animals and their behaviour laid the framework needed for similar studies on human subjects. Based on his work, developmental psychologists were able to study the effect of positive and negative reinforcement. Skinner found that the environment influenced behaviour and when that environment is manipulated, behaviour will change. From this, developmental psychologists proposed theories on operant learning in children. This research was applied to education and the treatment of illness in young children. [10] Skinner's theory of operant conditioning played a key role in helping psychologists understand how behavior is learned. It explains why reinforcement can be used so effectively in the learning process, and how schedules of reinforcement can affect the outcome of conditioning.

Commercial applications

Slot machines, online games, and dating apps are examples where sophisticated operant schedules of reinforcement are used to reinforce certain behaviors. [16] [17] [18] [19]

Gamification, the technique of using game design elements in non-game contexts, has also been described as using operant conditioning and other behaviorist techniques to encourage desired user behaviors. [20]

See also

Related Research Articles

<span class="mw-page-title-main">B. F. Skinner</span> American psychologist and social philosopher (1904–1990)

Burrhus Frederic Skinner was an American psychologist, behaviorist, inventor, and social philosopher. He was the Edgar Pierce Professor of Psychology at Harvard University from 1958 until his retirement in 1974.

<span class="mw-page-title-main">Edward Thorndike</span> American psychologist (1874–1949)

Edward Lee Thorndike was an American psychologist who spent nearly his entire career at Teachers College, Columbia University. His work on comparative psychology and the learning process led to the theory of connectionism and helped lay the scientific foundation for educational psychology. He also worked on solving industrial problems, such as employee exams and testing.

Operant conditioning, also called instrumental conditioning, is a learning process where voluntary behaviors are modified by association with the addition of reward or aversive stimuli. The frequency or duration of the behavior may increase through reinforcement or decrease through punishment or extinction.

<span class="mw-page-title-main">Reinforcement</span> Consequence affecting an organisms future behavior

In behavioral psychology, reinforcement refers to consequences that increases the likelihood of an organism's future behavior, typically in the presence of a particular antecedent stimulus. For example, a rat can be trained to push a lever to receive food whenever a light is turned on. In this example, the light is the antecedent stimulus, the lever pushing is the operant behavior, and the food is the reinforcer. Likewise, a student that receives attention and praise when answering a teacher's question will be more likely to answer future questions in class. The teacher's question is the antecedent, the student's response is the behavior, and the praise and attention are the reinforcements.

Radical behaviorism is a "philosophy of the science of behavior" developed by B. F. Skinner. It refers to the philosophy behind behavior analysis, and is to be distinguished from methodological behaviorism—which has an intense emphasis on observable behaviors—by its inclusion of thinking, feeling, and other private events in the analysis of human and animal psychology. The research in behavior analysis is called the experimental analysis of behavior and the application of the field is called applied behavior analysis (ABA), which was originally termed "behavior modification."

The experimental analysis of behavior is a science that studies the behavior of individuals across a variety of species. A key early scientist was B. F. Skinner who discovered operant behavior, reinforcers, secondary reinforcers, contingencies of reinforcement, stimulus control, shaping, intermittent schedules, discrimination, and generalization. A central method was the examination of functional relations between environment and behavior, as opposed to hypothetico-deductive learning theory that had grown up in the comparative psychology of the 1920–1950 period. Skinner's approach was characterized by observation of measurable behavior which could be predicted and controlled. It owed its early success to the effectiveness of Skinner's procedures of operant conditioning, both in the laboratory and in behavior therapy.

<span class="mw-page-title-main">Animal cognition</span> Intelligence of non-human animals

Animal cognition encompasses the mental capacities of non-human animals including insect cognition. The study of animal conditioning and learning used in this field was developed from comparative psychology. It has also been strongly influenced by research in ethology, behavioral ecology, and evolutionary psychology; the alternative name cognitive ethology is sometimes used. Many behaviors associated with the term animal intelligence are also subsumed within animal cognition.

Behaviorism is a systematic approach to understand the behavior of humans and other animals. It assumes that behavior is either a reflex evoked by the pairing of certain antecedent stimuli in the environment, or a consequence of that individual's history, including especially reinforcement and punishment contingencies, together with the individual's current motivational state and controlling stimuli. Although behaviorists generally accept the important role of heredity in determining behavior, they focus primarily on environmental events. The cognitive revolution of the late 20th century largely replaced behaviorism as an explanatory theory with cognitive psychology, which unlike behaviorism examines internal mental states.

The law of effect is a psychology principle advanced by Edward Thorndike in 1898 on the matter of behavioral conditioning which states that "responses that produce a satisfying effect in a particular situation become more likely to occur again in that situation, and responses that produce a discomforting effect become less likely to occur again in that situation."

Behaviour therapy or behavioural psychotherapy is a broad term referring to clinical psychotherapy that uses techniques derived from behaviourism and/or cognitive psychology. It looks at specific, learned behaviours and how the environment, or other people's mental states, influences those behaviours, and consists of techniques based on behaviorism's theory of learning: respondent or operant conditioning. Behaviourists who practice these techniques are either behaviour analysts or cognitive-behavioural therapists. They tend to look for treatment outcomes that are objectively measurable. Behaviour therapy does not involve one specific method, but it has a wide range of techniques that can be used to treat a person's psychological problems.

Shaping is a conditioning paradigm used primarily in the experimental analysis of behavior. The method used is differential reinforcement of successive approximations. It was introduced by B. F. Skinner with pigeons and extended to dogs, dolphins, humans and other species. In shaping, the form of an existing response is gradually changed across successive trials towards a desired target behavior by reinforcing exact segments of behavior. Skinner's explanation of shaping was this:

We first give the bird food when it turns slightly in the direction of the spot from any part of the cage. This increases the frequency of such behavior. We then withhold reinforcement until a slight movement is made toward the spot. This again alters the general distribution of behavior without producing a new unit. We continue by reinforcing positions successively closer to the spot, then by reinforcing only when the head is moved slightly forward, and finally only when the beak actually makes contact with the spot. ... The original probability of the response in its final form is very low; in some cases it may even be zero. In this way we can build complicated operants which would never appear in the repertoire of the organism otherwise. By reinforcing a series of successive approximations, we bring a rare response to a very high probability in a short time. ... The total act of turning toward the spot from any point in the box, walking toward it, raising the head, and striking the spot may seem to be a functionally coherent unit of behavior; but it is constructed by a continual process of differential reinforcement from undifferentiated behavior, just as the sculptor shapes his figure from a lump of clay.

Instinctive drift, alternately known as instinctual drift, is the tendency of an animal to revert to unconscious and automatic behaviour that interferes with learned behaviour from operant conditioning. Instinctive drift was coined by Keller and Marian Breland, former students of B.F. Skinner at the University of Minnesota, describing the phenomenon as "a clear and utter failure of conditioning theory." B.F. Skinner was an American psychologist and father of operant conditioning, which is a learning strategy that teaches the performance of an action either through reinforcement or punishment. It is through the association of the behaviour and the reward or consequence that follows that depicts whether an animal will maintain a behaviour, or if it will become extinct. Instinctive drift is a phenomenon where such conditioning erodes and an animal reverts to its natural behaviour.

In operant conditioning, punishment is any change in a human or animal's surroundings which, occurring after a given behavior or response, reduces the likelihood of that behavior occurring again in the future. As with reinforcement, it is the behavior, not the human/animal, that is punished. Whether a change is or is not punishing is determined by its effect on the rate that the behavior occurs. This is called motivating operations (MO), because they alter the effectiveness of a stimulus. MO can be categorized in abolishing operations, decrease the effectiveness of the stimuli and establishing, increase the effectiveness of the stimuli. For example, a painful stimulus which would act as a punisher for most people may actually reinforce some behaviors of masochistic individuals.

Comparative cognition is the comparative study of the mechanisms and origins of cognition in various species, and is sometimes seen as more general than, or similar to, comparative psychology. From a biological point of view, work is being done on the brains of fruit flies that should yield techniques precise enough to allow an understanding of the workings of the human brain on a scale appreciative of individual groups of neurons rather than the more regional scale previously used. Similarly, gene activity in the human brain is better understood through examination of the brains of mice by the Seattle-based Allen Institute for Brain Science, yielding the freely available Allen Brain Atlas. This type of study is related to comparative cognition, but better classified as one of comparative genomics. Increasing emphasis in psychology and ethology on the biological aspects of perception and behavior is bridging the gap between genomics and behavioral analysis.

In behavioral psychology, stimulus control is a phenomenon in operant conditioning that occurs when an organism behaves in one way in the presence of a given stimulus and another way in its absence. A stimulus that modifies behavior in this manner is either a discriminative stimulus or stimulus delta. For example, the presence of a stop sign at a traffic intersection alerts the driver to stop driving and increases the probability that braking behavior occurs. Stimulus control does not force behavior to occur, as it is a direct result of historical reinforcement contingencies, as opposed to reflexive behavior elicited through classical conditioning.

Discrimination learning is defined in psychology as the ability to respond differently to different stimuli. This type of learning is used in studies regarding operant and classical conditioning. Operant conditioning involves the modification of a behavior by means of reinforcement or punishment. In this way, a discriminative stimulus will act as an indicator to when a behavior will persist and when it will not. Classical conditioning involves learning through association when two stimuli are paired together repeatedly. This conditioning demonstrates discrimination through specific micro-instances of reinforcement and non-reinforcement. This phenomenon is considered to be more advanced than learning styles such as generalization and yet simultaneously acts as a basic unit to learning as a whole. The complex and fundamental nature of discrimination learning allows for psychologists and researchers to perform more in-depth research that supports psychological advancements. Research on the basic principles underlying this learning style has their roots in neuropsychology sub-processes.

Self-administration is, in its medical sense, the process of a subject administering a pharmacological substance to themself. A clinical example of this is the subcutaneous "self-injection" of insulin by a diabetic patient.

<span class="mw-page-title-main">Marine mammal training</span>

Marine mammal training encompasses both the training and caretaking of various marine mammals, including dolphins, orcas, sea lions, and walruses. This discipline involves teaching these animals specific behaviors for different purposes, such as participation in shows, scientific research, military operations, or simply for their health and enrichment. Caretaking elements include ensuring the animals' well-being through proper diet, habitat maintenance, and health monitoring.


Human contingency learning (HCL) is the observation that people tend to acquire knowledge based on whichever outcome has the highest probability of occurring from particular stimuli. In other words, individuals gather associations between a certain behaviour and a specific consequence. It is a form of learning for many organisms.

Association in psychology refers to a mental connection between concepts, events, or mental states that usually stems from specific experiences. Associations are seen throughout several schools of thought in psychology including behaviorism, associationism, psychoanalysis, social psychology, and structuralism. The idea stems from Plato and Aristotle, especially with regard to the succession of memories, and it was carried on by philosophers such as John Locke, David Hume, David Hartley, and James Mill. It finds its place in modern psychology in such areas as memory, learning, and the study of neural pathways.

References

  1. Carlson NR (2009). Psychology-the science of behavior. U.S: Pearson Education Canada; 4th edition. p.  207. ISBN   978-0-205-64524-4.
  2. 1 2 Krebs JR (1983). "Animal behaviour. From Skinner box to the field". Nature. 304 (5922): 117. Bibcode:1983Natur.304..117K. doi: 10.1038/304117a0 . PMID   6866102. S2CID   5360836.
  3. Schacter DL, Gilbert DT, Wegner DM, Nock MK (January 2, 2014). "B. F. Skinner: The Role of Reinforcement and Punishment". Psychology (3rd ed.). Macmillan. pp. 278–80. ISBN   978-1-4641-5528-4.
  4. Kazdin A (2000). Encyclopedia of Psychology, Vol. 5. American Psychological Association.
  5. 1 2 3 Sakagami T, Lattal KA (May 2016). "The Other Shoe: An Early Operant Conditioning Chamber for Pigeons". The Behavior Analyst. 39 (1): 25–39. doi:10.1007/s40614-016-0055-8. PMC   4883506 . PMID   27606188.
  6. Skinner BF (1959). Cumulative record (1999 ed.). Cambridge, MA: B.F. Skinner Foundation. p. 620.
  7. Skinner BF (1983). A Matter of Consequences. New York, NY: Alfred A. Knopf, Inc. pp. 116, 164.
  8. Gray P (2007). Psychology. New York: Worth Publishers. pp. 108–109.
  9. 1 2 "Edward Thorndike – Law of Effect | Simply Psychology". www.simplypsychology.org. Retrieved November 14, 2021.
  10. 1 2 3 Schlinger H (January 17, 2021). "The Impact of B. F. Skinner's Science of Operant Learning on Early Childhood Research, Theory, Treatment, and Care". Early Child Development and Care. 191 (7–8): 1089–1106. doi:10.1080/03004430.2020.1855155. S2CID   234206521 via Routledge.
  11. 1 2 Brembs B (December 2003). "Operant conditioning in invertebrates" (PDF). Current Opinion in Neurobiology. 13 (6): 710–717. doi:10.1016/j.conb.2003.10.002. PMID   14662373. S2CID   2385291.
  12. Fernández-Lamo I, Delgado-García JM, Gruart A (March 2018). "When and Where Learning is Taking Place: Multisynaptic Changes in Strength During Different Behaviors Related to the Acquisition of an Operant Conditioning Task by Behaving Rats". Cerebral Cortex. 28 (3): 1011–1023. doi: 10.1093/cercor/bhx011 . PMID   28199479.
  13. Jackson K, Hackenberg TD (July 1996). "Token reinforcement, choice, and self-control in pigeons". Journal of the Experimental Analysis of Behavior. 66 (1): 29–49. doi:10.1901/jeab.1996.66-29. PMC   1284552 . PMID   8755699.
  14. Craighead, W. Edward; Nemeroff, Charles B., eds. (2004). The Concise Corsini Encyclopedia of Psychology and Behavioral Science 3rd ed. Hoboken, New Jersey: John Wiley & Sons, Inc. p. 803. ISBN   0-471-22036-1.
  15. Shrestha P (November 17, 2017). "Operant Conditioning". Psychestudy. Retrieved November 14, 2021.
  16. Hopson, J. (April 2001). "Behavioral game design". Gamasutra. Retrieved April 27, 2019.
  17. Coon D (2005). Psychology: A modular approach to mind and behavior. Thomson Wadsworth. pp. 278–279. ISBN   0-534-60593-1.
  18. "The science behind those apps you can't stop using". Australian Financial Review. October 7, 2016. Retrieved January 23, 2024.
  19. "The scientists who make apps addictive". The Economist. ISSN   0013-0613 . Retrieved January 23, 2024.
  20. Thompson A (May 6, 2015). "Slot machines perfected addictive gaming. Now, tech wants their tricks". The Verge .