Connectionist Learning with Adaptive Rule Induction On-line (CLARION) is a computational cognitive architecture that has been used to simulate many domains and tasks in cognitive psychology and social psychology, as well as implementing intelligent systems in artificial intelligence applications. An important feature of CLARION is the distinction between implicit and explicit processes and focusing on capturing the interaction between these two types of processes. The system was created by the research group led by Ron Sun.
CLARION is an integrative cognitive architecture, it is used to explain and simulate cognitive-psychological phenomena, which could potentially lead to an unified explanation of psychological phenomena. There are three layers to the CLARION theory, the first layer is the core theory of mind. The main theories consists of a number of distinct subsystems, which are the essential structures of CLARION, with a dual representational structure in each subsystem (implicit versus explicit representations [1] ). Its subsystems include the action-centered subsystem, the non-action-centered subsystem, the motivational subsystem, and the meta-cognitive subsystem. The second layer consists of the computational models that implements the basic theory; it is more detailed than the first level theory but is still general. The third layer consists of the specific implemented models and simulations of the psychological processes or phenomena. The models of this layer arise from the basic theory and the general computational models.
The distinction between implicit and explicit processes is fundamental to the Clarion cognitive architecture. [2] This distinction is primarily motivated by evidence supporting implicit memory and implicit learning. Clarion captures the implicit-explicit distinction independently from the distinction between procedural memory and declarative memory. To capture the implicit-explicit distinction, Clarion postulates two parallel and interacting representational systems capturing implicit an explicit knowledge respectively. Explicit knowledge is associated with localist representation and implicit knowledge with distributed representation.
Explicit knowledge resides in the top level of the architecture, whereas implicit knowledge resides in the bottom level. [2] [3] In both levels, the basic representational units are connectionist nodes, and the two levels differ with respect to the type of encoding. In the top level, knowledge is encoded using localist chunk nodes whereas, in the bottom level, knowledge is encoded in a distributed manner through collections of (micro)feature nodes. Knowledge may be encoded redundantly between the two levels and may be processed in parallel within the two levels. In the top level, information processing involves passing activations among chunk nodes by means of rules and, in the bottom level, information processing involves propagating (micro)feature activations through artificial neural networks. Top-down and bottom-up information flows are enabled by links between the two levels. Such links are established by Clarion chunks, each of which consists of a single chunk node, a collection of (micro)feature nodes, and links between the chunk node and the (micro)feature nodes. In this way a single chunk of knowledge may be expressed in both explicit (i.e., localist) and implicit (i.e., distributed) form, though such dual expression is not always required.
The dual representational structure allows implicit and explicit processes to communicate and, potentially, to encode content redundantly. As a result, Clarion theory can account for various phenomena, such as speed-up effects in learning, verbalization-related performance gains, performance gains in transfer tasks, and the ability to perform similarity-based reasoning, in terms of synergistic interaction between implicit and explicit processes. [2] [4] [5] [6] These interactions involve both the flow of activations within the architecture (e.g., similarity-based reasoning is supported by spreading activation among chunks through shared (micro)features) as well as bottom-up, top-down and parallel learning processes. In bottom-up learning, associations among (micro)features in the bottom level are extracted and encoded as explicit rules. In top-down learning, rules in the top level guide the development of implicit associations in the bottom level. Additionally, learning may be carried out in parallel, touching both implicit and explicit processes simultaneously. Through these learning processes knowledge may be encoded redundantly or in complementary fashion, as dictated by agent history. Synergy effects arise, in part, from the interaction of these learning processes. Another important mechanism for explaining synergy effects is the combination and relative balance of signals from different levels of the architecture. For instance, in one Clarion-based modeling study, it has been proposed that an anxiety-driven imbalance in the relative contributions of implicit versus explicit processes may be the mechanism responsible for performance degradation under pressure. [7]
The Clarion cognitive architecture consists of four subsystems.
The role of the action-centered subsystem is to control both external and internal actions. The implicit layer is made of neural networks called Action Neural Networks, while the explicit layer is made up of action rules. There can be synergy between the two layers, for example learning a skill can be expedited when the agent has to make explicit rules for the procedure at hand. It has been argued that implicit knowledge alone cannot optimize as well as the combination of both explicit and implicit.
The role of the non-action-centered subsystem is to maintain general knowledge. The implicit layer is made of Associative Neural Networks, while the bottom layer is associative rules. Knowledge is further divided into semantic and episodic, where semantic is generalized knowledge, and episodic is knowledge applicable to more specific situations. It is also important to note since there is an implicit layer, that not all declarative knowledge has to be explicit.
The role of the motivational subsystem is to provide underlying motivations for perception, action, and cognition. The motivational system in CLARION is made up of drives on the bottom level, and each drive can have varying strengths. There are low level drives, and also high level drives aimed at keeping an agent sustained, purposeful, focused, and adaptive. The explicit layer of the motivational system is composed of goals. explicit goals are used because they are more stable than implicit motivational states. The CLARION framework views that human motivational processes are highly complex and can't be represented through just explicit representation.
Examples of some low level drives include:
Examples of some high level drives include:
There is also a possibility for derived drives (usually from trying to satisfy primary drives) that can be created by either conditioning, or through external instructions. Each drive needed will have a proportional strength, opportunity will also be taken into account
The role of the meta-cognitive subsystem is to monitor, direct, and modify the operations of all the other subsystems. Actions in the meta-cognitive subsystem include: setting goals for the action-centred subsystem, setting parameters for the action and non-action subsystems, and changing an ongoing process in both the action and non-action subsystems.
Learning can be represented with both explicit and implicit knowledge individually while also representing bottom-up and top-down learning. Learning with implicit knowledge is represented through Q-learning, while learning with just explicit knowledge is represented with one-shot learning such as hypothesis testing. Bottom-up learning [8] is represented through a neural network propagating up to the explicit layer through the Rule-Extraction-Refinement algorithm (RER), while top-down learning can be represented through a variety of ways.
To compare with a few other cognitive architectures: [2]
CLARION has been used to account for a variety of psychological data, [9] [2] such as the serial reaction time task, the artificial grammar learning task, the process control task, a categorical inference task, an alphabetical arithmetic task, and the Tower of Hanoi task. The serial reaction time and process control tasks are typical implicit learning tasks (mainly involving implicit reactive routines), while the Tower of Hanoi and alphabetic arithmetic are high-level cognitive skill acquisition tasks (with a significant presence of explicit processes). In addition, extensive work has been done on a complex minefield navigation task, which involves complex sequential decision-making. Work on organizational decision tasks and other social simulation tasks, [10] as well as meta-cognitive tasks, has also been initiated.
Other applications of the cognitive architecture include simulation of creativity [11] and addressing the computational basis of consciousness (or artificial consciousness). [12]
Cognitive science is the interdisciplinary, scientific study of the mind and its processes. It examines the nature, the tasks, and the functions of cognition. Mental faculties of concern to cognitive scientists include language, perception, memory, attention, reasoning, and emotion; to understand these faculties, cognitive scientists borrow from fields such as linguistics, psychology, artificial intelligence, philosophy, neuroscience, and anthropology. The typical analysis of cognitive science spans many levels of organization, from learning and decision to logic and planning; from neural circuitry to modular brain organization. One of the fundamental concepts of cognitive science is that "thinking can best be understood in terms of representational structures in the mind and computational procedures that operate on those structures."
Artificial consciousness, also known as machine consciousness, synthetic consciousness, or digital consciousness, is the consciousness hypothesized to be possible in artificial intelligence. It is also the corresponding field of study, which draws insights from philosophy of mind, philosophy of artificial intelligence, cognitive science and neuroscience. The same terminology can be used with the term "sentience" instead of "consciousness" when specifically designating phenomenal consciousness.
Connectionism is the name of an approach to the study of human mental processes and cognition that utilizes mathematical models known as connectionist networks or artificial neural networks. Connectionism has had many 'waves' since its beginnings.
A cognitive model is a representation of one or more cognitive processes in humans or other animals for the purposes of comprehension and prediction. There are many types of cognitive models, and they can range from box-and-arrow diagrams to a set of equations to software programs that interact with the same tools that humans use to complete tasks. In terms of information processing, cognitive modeling is modeling of human perception, reasoning, memory and action.
Semantic memory refers to general world knowledge that humans have accumulated throughout their lives. This general knowledge is intertwined in experience and dependent on culture. New concepts are learned by applying knowledge learned from things in the past.
ACT-R is a cognitive architecture mainly developed by John Robert Anderson and Christian Lebiere at Carnegie Mellon University. Like any cognitive architecture, ACT-R aims to define the basic and irreducible cognitive and perceptual operations that enable the human mind. In theory, each task that humans can perform should consist of a series of these discrete operations.
In cognitive psychology, chunking is a process by which small individual pieces of a set of information are bound together to create a meaningful whole later on in memory. The chunks, by which the information is grouped, are meant to improve short-term retention of the material, thus bypassing the limited capacity of working memory and allowing the working memory to be more efficient. A chunk is a collection of basic units that are strongly associated with one another, and have been grouped together and stored in a person's memory. These chunks can be retrieved easily due to their coherent grouping. It is believed that individuals create higher-order cognitive representations of the items within the chunk. The items are more easily remembered as a group than as the individual items themselves. These chunks can be highly subjective because they rely on an individual's perceptions and past experiences, which are linked to the information set. The size of the chunks generally ranges from two to six items but often differs based on language and culture.
A cognitive architecture refers to both a theory about the structure of the human mind and to a computational instantiation of such a theory used in the fields of artificial intelligence (AI) and computational cognitive science. These formalized models can be used to further refine comprehensive theories of cognition and serve as the frameworks for useful artificial intelligence programs. Successful cognitive architectures include ACT-R and SOAR. The research on cognitive architectures as software instantiation of cognitive theories was initiated by Allen Newell in 1990.
The Levels of Processing model, created by Fergus I. M. Craik and Robert S. Lockhart in 1972, describes memory recall of stimuli as a function of the depth of mental processing. More analysis produce more elaborate and stronger memory than lower levels of processing. Depth of processing falls on a shallow to deep continuum. Shallow processing leads to a fragile memory trace that is susceptible to rapid decay. Conversely, deep processing results in a more durable memory trace. There are three levels of processing in this model. Structural processing, or visual, is when we remember only the physical quality of the word. Phonemic processing includes remembering the word by the way it sounds. Lastly, we have semantic processing in which we encode the meaning of the word with another word that is similar or has similar meaning. Once the word is perceived, the brain allows for a deeper processing.
Computational cognition is the study of the computational basis of learning and inference by mathematical modeling, computer simulation, and behavioral experiments. In psychology, it is an approach which develops computational models based on experimental results. It seeks to understand the basis behind the human method of processing of information. Early on computational cognitive scientists sought to bring back and create a scientific form of Brentano's psychology.
In psychology, a dual process theory provides an account of how thought can arise in two different ways, or as a result of two different processes. Often, the two processes consist of an implicit (automatic), unconscious process and an explicit (controlled), conscious process. Verbalized explicit processes or attitudes and actions may change with persuasion or education; though implicit process or attitudes usually take a long amount of time to change with the forming of new habits. Dual process theories can be found in social, personality, cognitive, and clinical psychology. It has also been linked with economics via prospect theory and behavioral economics, and increasingly in sociology through cultural analysis.
Artificial grammar learning (AGL) is a paradigm of study within cognitive psychology and linguistics. Its goal is to investigate the processes that underlie human language learning by testing subjects' ability to learn a made-up grammar in a laboratory setting. It was developed to evaluate the processes of human language learning but has also been utilized to study implicit learning in a more general sense. The area of interest is typically the subjects' ability to detect patterns and statistical regularities during a training phase and then use their new knowledge of those patterns in a testing phase. The testing phase can either use the symbols or sounds used in the training phase or transfer the patterns to another set of symbols or sounds as surface structure.
In cognitive psychology, sequence learning is inherent to human ability because it is an integrated part of conscious and nonconscious learning as well as activities. Sequences of information or sequences of actions are used in various everyday tasks: "from sequencing sounds in speech, to sequencing movements in typing or playing instruments, to sequencing actions in driving an automobile." Sequence learning can be used to study skill acquisition and in studies of various groups ranging from neuropsychological patients to infants. According to Ritter and Nerb, “The order in which material is presented can strongly influence what is learned, how fast performance increases, and sometimes even whether the material is learned at all.” Sequence learning, more known and understood as a form of explicit learning, is now also being studied as a form of implicit learning as well as other forms of learning. Sequence learning can also be referred to as sequential behavior, behavior sequencing, and serial order in behavior.
Implicit cognition refers to cognitive processes that occur outside conscious awareness or conscious control. This includes domains such as learning, perception, or memory which may influence a person's behavior without their conscious awareness of those influences.
CHREST is a symbolic cognitive architecture based on the concepts of limited attention, limited short-term memories, and chunking. The architecture takes into low-level aspects of cognition such as reference perception, long and short-term memory stores, and methodology of problem-solving and high-level aspects such as the use of strategies. Learning, which is essential in the architecture, is modelled as the development of a network of nodes (chunks) which are connected in various ways. This can be contrasted with Soar and ACT-R, two other cognitive architectures, which use productions for representing knowledge. CHREST has often been used to model learning using large corpora of stimuli representative of the domain, such as chess games for the simulation of chess expertise or child-directed speech for the simulation of children's development of language. In this respect, the simulations carried out with CHREST have a flavour closer to those carried out with connectionist models than with traditional symbolic models.
Psi-theory, developed by Dietrich Dörner at the University of Bamberg, is a systemic psychological theory covering human action regulation, intention selection and emotion. It models the human mind as an information processing agent, controlled by a set of basic physiological, social and cognitive drives. Perceptual and cognitive processing are directed and modulated by these drives, which allow the autonomous establishment and pursuit of goals in an open environment.
Ron Sun is a cognitive scientist who has made significant contributions to computational psychology and other areas of cognitive science and artificial intelligence. He is currently professor of cognitive sciences at Rensselaer Polytechnic Institute, and formerly the James C. Dowell Professor of Engineering and Professor of Computer Science at University of Missouri. He received his Ph.D. in 1992 from Brandeis University.
Indirect memory tests assess the retention of information without direct reference to the source of information. Participants are given tasks designed to elicit knowledge that was acquired incidentally or unconsciously and is evident when performance shows greater inclination towards items initially presented than new items. Performance on indirect tests may reflect contributions of implicit memory, the effects of priming, a preference to respond to previously experienced stimuli over novel stimuli. Types of indirect memory tests include the implicit association test, the lexical decision task, the word stem completion task, artificial grammar learning, word fragment completion, and the serial reaction time task.
In psychology, implicit memory is one of the two main types of long-term human memory. It is acquired and used unconsciously, and can affect thoughts and behaviours. One of its most common forms is procedural memory, which allows people to perform certain tasks without conscious awareness of these previous experiences; for example, remembering how to tie one's shoes or ride a bicycle without consciously thinking about those activities.
Fusion adaptive resonance theory (fusion ART) is a generalization of self-organizing neural networks known as the original Adaptive Resonance Theory models for learning recognition categories across multiple pattern channels. There is a separate stream of work on fusion ARTMAP, that extends fuzzy ARTMAP consisting of two fuzzy ART modules connected by an inter-ART map field to an extended architecture consisting of multiple ART modules.