CLARION (cognitive architecture)

Last updated

Connectionist Learning with Adaptive Rule Induction On-line (CLARION) is a computational cognitive architecture that has been used to simulate many domains and tasks in cognitive psychology and social psychology, as well as implementing intelligent systems in artificial intelligence applications. An important feature of CLARION is the distinction between implicit and explicit processes and focusing on capturing the interaction between these two types of processes. The system was created by the research group led by Ron Sun.

Contents

Clarion Framework Clarion Cognitive Architecture.jpg
Clarion Framework

Overview

CLARION is an integrative cognitive architecture, it is used to explain and simulate cognitive-psychological phenomena, which could potentially lead to an unified explanation of psychological phenomena. There are three layers to the CLARION theory, the first layer is the core theory of mind. The main theories consists of a number of distinct subsystems, which are the essential structures of CLARION, with a dual representational structure in each subsystem (implicit versus explicit representations; Sun et al., 2005). Its subsystems include the action-centered subsystem, the non-action-centered subsystem, the motivational subsystem, and the meta-cognitive subsystem. The second layer consists of the computational models that implements the basic theory, it is more detailed than the first level theory but is still general. The third layer consists of the specific implemented models and simulations of the psychological processes or phenomena. The models of this layer arise from the basic theory and the general computational models.

Dual Representational Structure

The distinction between implicit and explicit processes is fundamental to the Clarion cognitive architecture. [1] This distinction is primarily motivated by evidence supporting implicit memory and implicit learning. Clarion captures the implicit-explicit distinction independently from the distinction between procedural memory and declarative memory. To capture the implicit-explicit distinction, Clarion postulates two parallel and interacting representational systems capturing implicit an explicit knowledge respectively. Explicit knowledge is associated with localist representation and implicit knowledge with distributed representation.

Explicit knowledge resides in the top level of the architecture, whereas implicit knowledge resides in the bottom level. [1] [2] In both levels, the basic representational units are connectionist nodes, and the two levels differ with respect to the type of encoding. In the top level, knowledge is encoded using localist chunk nodes whereas, in the bottom level, knowledge is encoded in a distributed manner through collections of (micro)feature nodes. Knowledge may be encoded redundantly between the two levels and may be processed in parallel within the two levels. In the top level, information processing involves passing activations among chunk nodes by means of rules and, in the bottom level, information processing involves propagating (micro)feature activations through artificial neural networks. Top-down and bottom-up information flows are enabled by links between the two levels. Such links are established by Clarion chunks, each of which consists of a single chunk node, a collection of (micro)feature nodes, and links between the chunk node and the (micro)feature nodes. In this way a single chunk of knowledge may be expressed in both explicit (i.e., localist) and implicit (i.e., distributed) form, though such dual expression is not always required.

The dual representational structure allows implicit and explicit processes to communicate and, potentially, to encode content redundantly. As a result, Clarion theory can account for various phenomena, such as speed-up effects in learning, verbalization-related performance gains, performance gains in transfer tasks, and the ability to perform similarity-based reasoning, in terms of synergistic interaction between implicit and explicit processes. [1] [3] [4] [5] These interactions involve both the flow of activations within the architecture (e.g., similarity-based reasoning is supported by spreading activation among chunks through shared (micro)features) as well as bottom-up, top-down and parallel learning processes. In bottom-up learning, associations among (micro)features in the bottom level are extracted and encoded as explicit rules. In top-down learning, rules in the top level guide the development of implicit associations in the bottom level. Additionally, learning may be carried out in parallel, touching both implicit and explicit processes simultaneously. Through these learning processes knowledge may be encoded redundantly or in complementary fashion, as dictated by agent history. Synergy effects arise, in part, from the interaction of these learning processes. Another important mechanism for explaining synergy effects is the combination and relative balance of signals from different levels of the architecture. For instance, in one Clarion-based modeling study, it has been proposed that an anxiety-driven imbalance in the relative contributions of implicit versus explicit processes may be the mechanism responsible for performance degradation under pressure. [6]

Subsystems

The Clarion cognitive architecture consists of four subsystems.

Action-centered subsystem

The role of the action-centered subsystem is to control both external and internal actions. The implicit layer is made of neural networks called Action Neural Networks, while the explicit layer is made up of action rules. There can be synergy between the two layers, for example learning a skill can be expedited when the agent has to make explicit rules for the procedure at hand. It has been argued that implicit knowledge alone cannot optimize as well as the combination of both explicit and implicit.

Non-action-centered subsystem

The role of the non-action-centered subsystem is to maintain general knowledge. The implicit layer is made of Associative Neural Networks, while the bottom layer is associative rules. Knowledge is further divided into semantic and episodic, where semantic is generalized knowledge, and episodic is knowledge applicable to more specific situations. It is also important to note since there is an implicit layer, that not all declarative knowledge has to be explicit.

Motivational subsystem

The role of the motivational subsystem is to provide underlying motivations for perception, action, and cognition. The motivational system in CLARION is made up of drives on the bottom level, and each drive can have varying strengths. There are low level drives, and also high level drives aimed at keeping an agent sustained, purposeful, focused, and adaptive. The explicit layer of the motivational system is composed of goals. explicit goals are used because they are more stable than implicit motivational states. the CLARION framework views that human motivational processes are highly complex and can't be represented through just explicit representation.

Examples of some low level drives include:

  • food
  • water
  • reproduction
  • avoiding unpleasant stimuli (not mutually exclusive of other low level drives, but separate for the possibility of more specific stimuli)

Examples of some high level drives include:

  • Affiliation and belongingness
  • Recognition and achievement
  • Dominance and power
  • Fairness

There is also a possibility for derived drives (usually from trying to satisfy primary drives) that can be created by either conditioning, or through external instructions. each drive needed will have a proportional strength, opportunity will also be taken into account

Meta-cognitive subsystem

The role of the meta-cognitive subsystem is to monitor, direct, and modify the operations of all the other subsystems. Actions in the meta-cognitive subsystem include: setting goals for the action-centred subsystem, setting parameters for the action and non-action subsystems, and changing an ongoing process in both the action and non-action subsystems.

Learning

Learning can be represented with both explicit and implicit knowledge individually while also representing bottom-up and top-down learning. Learning with implicit knowledge is represented through Q-learning, while learning with just explicit knowledge is represented with one-shot learning such as hypothesis testing. Bottom-up learning (Sun et al., 2001) is represented through a neural network propagating up to the explicit layer through the Rule-Extraction-Refinement algorithm (RER), while top-down learning can be represented through a variety of ways.

Comparison with other cognitive architectures

To compare with a few other cognitive architectures (Sun, 2016):

Theoretical applications

CLARION has been used to account for a variety of psychological data (Sun, 2002, 2016), such as the serial reaction time task, the artificial grammar learning task, the process control task, a categorical inference task, an alphabetical arithmetic task, and the Tower of Hanoi task. The serial reaction time and process control tasks are typical implicit learning tasks (mainly involving implicit reactive routines), while the Tower of Hanoi and alphabetic arithmetic are high-level cognitive skill acquisition tasks (with a significant presence of explicit processes). In addition, extensive work has been done on a complex minefield navigation task, which involves complex sequential decision-making. Work on organizational decision tasks and other social simulation tasks (e.g., Naveh and Sun, 2006), as well as meta-cognitive tasks, has also been initiated.

Other applications of the cognitive architecture include simulation of creativity (Helie and Sun, 2010) and addressing the computational basis of consciousness (or artificial consciousness) (Coward and Sun, 2004).

Related Research Articles

Artificial consciousness (AC), also known as machine consciousness (MC) or synthetic consciousness, is a field related to artificial intelligence and cognitive robotics. The aim of the theory of artificial consciousness is to "Define that which would have to be synthesized were consciousness to be found in an engineered artifact".

<span class="mw-page-title-main">Connectionism</span> Cognitive science approach

Connectionism refers to both an approach in the field of cognitive science that hopes to explain mental phenomena using artificial neural networks (ANN) and to a wide range of techniques and algorithms using ANNs in the context of artificial intelligence to build more intelligent machines. Connectionism presents a cognitive theory based on simultaneously occurring, distributed signal activity via connections that can be represented numerically, where learning occurs by modifying connection strengths based on experience.

A cognitive model is an approximation of one or more cognitive processes in humans or other animals for the purposes of comprehension and prediction. There are many types of cognitive models, and they can range from box-and-arrow diagrams to a set of equations to software programs that interact with the same tools that humans use to complete tasks.

Semantic memory refers to general world knowledge that humans have accumulated throughout their lives. This general knowledge is intertwined in experience and dependent on culture. We can learn about new concepts by applying our knowledge learned from things in the past.

ACT-R is a cognitive architecture mainly developed by John Robert Anderson and Christian Lebiere at Carnegie Mellon University. Like any cognitive architecture, ACT-R aims to define the basic and irreducible cognitive and perceptual operations that enable the human mind. In theory, each task that humans can perform should consist of a series of these discrete operations.

In cognitive psychology, chunking is a process by which individual pieces of an information set are bound together into a meaningful whole. The chunks, by which the information is grouped, are meant to improve short-term retention of the material, thus bypassing the limited capacity of working memory and allowing the working memory to be more efficient. A chunk is a collection of basic units that have been grouped together and stored in a person's memory. These chunks can be retrieved easily due to their coherent grouping. It is believed that individuals create higher-order cognitive representations of the items within the chunk. The items are more easily remembered as a group than as the individual items themselves. These chunks can be highly subjective because they rely on an individual's perceptions and past experiences, which are linked to the information set. The size of the chunks generally ranges from two to six items but often differs based on language and culture.

A cognitive architecture refers to both a theory about the structure of the human mind and to a computational instantiation of such a theory used in the fields of artificial intelligence (AI) and computational cognitive science. The formalized models can be used to further refine a comprehensive theory of cognition and as a useful artificial intelligence program. Successful cognitive architectures include ACT-R and SOAR. The research on cognitive architectures as software instantiation of cognitive theories was initiated by Allen Newell in 1990.

The Levels of Processing model, created by Fergus I. M. Craik and Robert S. Lockhart in 1972, describes memory recall of stimuli as a function of the depth of mental processing. Deeper levels of analysis produce more elaborate, longer-lasting, and stronger memory traces than shallow levels of analysis. Depth of processing falls on a shallow to deep continuum. Shallow processing leads to a fragile memory trace that is susceptible to rapid decay. Conversely, deep processing results in a more durable memory trace. There are three levels of processing in this model. Structural processing, or visual, is when we remember only the physical quality of the word E.g how the word is spelled and how letters look. Phonemic processing includes remembering the word by the way it sounds. E.G the word tall rhymes with fall. Lastly, we have semantic processing in which we encode the meaning of the word with another word that is similar of has similar meaning. Once the word is perceived, the brain allows for a deeper processing.

Computational cognition is the study of the computational basis of learning and inference by mathematical modeling, computer simulation, and behavioral experiments. In psychology, it is an approach which develops computational models based on experimental results. It seeks to understand the basis behind the human method of processing of information. Early on computational cognitive scientists sought to bring back and create a scientific form of Brentano's psychology.

In psychology, a dual process theory provides an account of how thought can arise in two different ways, or as a result of two different processes. Often, the two processes consist of an implicit (automatic), unconscious process and an explicit (controlled), conscious process. Verbalized explicit processes or attitudes and actions may change with persuasion or education; though implicit process or attitudes usually take a long amount of time to change with the forming of new habits. Dual process theories can be found in social, personality, cognitive, and clinical psychology. It has also been linked with economics via prospect theory and behavioral economics, and increasingly in sociology through cultural analysis.

Artificial grammar learning (AGL) is a paradigm of study within cognitive psychology and linguistics. Its goal is to investigate the processes that underlie human language learning by testing subjects' ability to learn a made-up grammar in a laboratory setting. It was developed to evaluate the processes of human language learning but has also been utilized to study implicit learning in a more general sense. The area of interest is typically the subjects' ability to detect patterns and statistical regularities during a training phase and then use their new knowledge of those patterns in a testing phase. The testing phase can either use the symbols or sounds used in the training phase or transfer the patterns to another set of symbols or sounds as surface structure.

Implicit cognition refers to unconscious influences such as knowledge, perception, or memory, that influence a person's behavior, even though they themselves have no conscious awareness whatsoever of those influences.

In mental memory, storage is one of three fundamental stages along with encoding and retrieval. Memory is the process of storing and recalling information that was previously acquired. Storing refers to the process of placing newly acquired information into memory, which is modified in the brain for easier storage. Encoding this information makes the process of retrieval easier for the brain where it can be recalled and brought into conscious thinking. Modern memory psychology differentiates between the two distinct types of memory storage: short-term memory and long-term memory. Several models of memory have been proposed over the past century, some of them suggesting different relationships between short- and long-term memory to account for different ways of storing memory.

CHREST is a symbolic cognitive architecture based on the concepts of limited attention, limited short-term memories, and chunking. The architecture takes into low-level aspects of cognition such as reference perception, long and short term memory stores, and methodology of problem-solving and high-level aspects such as the use of strategies. Learning, which is essential in the architecture, is modelled as the development of a network of nodes (chunks) which are connected in various ways. This can be contrasted with Soar and ACT-R, two other cognitive architectures, which use productions for representing knowledge. CHREST has often been used to model learning using large corpora of stimuli representative of the domain, such as chess games for the simulation of chess expertise or child-directed speech for the simulation of children's development of language. In this respect, the simulations carried out with CHREST have a flavour closer to those carried out with connectionist models than with traditional symbolic models.

Psi-theory, developed by Dietrich Dörner at the University of Bamberg, is a systemic psychological theory covering human action regulation, intention selection and emotion. It models the human mind as an information processing agent, controlled by a set of basic physiological, social and cognitive drives. Perceptual and cognitive processing are directed and modulated by these drives, which allow the autonomous establishment and pursuit of goals in an open environment.

Ron Sun is a cognitive scientist who made significant contributions to computational psychology and other areas of cognitive science and artificial intelligence. He is currently professor of cognitive sciences at Rensselaer Polytechnic Institute, and formerly the James C. Dowell Professor of Engineering and Professor of Computer Science at University of Missouri. He received his Ph.D. in 1992 from Brandeis University.

Indirect memory tests assess the retention of information without direct reference to the source of information. Participants are given tasks designed to elicit knowledge that was acquired incidentally or unconsciously and is evident when performance shows greater inclination towards items initially presented than new items. Performance on indirect tests may reflect contributions of implicit memory, the effects of priming, a preference to respond to previously experienced stimuli over novel stimuli. Types of indirect memory tests include the implicit association test, the lexical decision task, the word stem completion task, artificial grammar learning, word fragment completion, and the serial reaction time task.

In psychology, implicit memory is one of the two main types of long-term human memory. It is acquired and used unconsciously, and can affect thoughts and behaviours. One of its most common forms is procedural memory, which allows people to perform certain tasks without conscious awareness of these previous experiences; for example, remembering how to tie one's shoes or ride a bicycle without consciously thinking about those activities.

The LIDA cognitive architecture is an integrated artificial cognitive system that attempts to model a broad spectrum of cognition in biological systems, from low-level perception/action to high-level reasoning. Developed primarily by Stan Franklin and colleagues at the University of Memphis, the LIDA architecture is empirically grounded in cognitive science and cognitive neuroscience. In addition to providing hypotheses to guide further research, the architecture can support control structures for software agents and robots. Providing plausible explanations for many cognitive processes, the LIDA conceptual model is also intended as a tool with which to think about how minds work.

Fusion adaptive resonance theory (fusion ART) is a generalization of self-organizing neural networks known as the original Adaptive Resonance Theory models for learning recognition categories (or cognitive codes) across multiple pattern channels. There is a separate stream of work on fusion ARTMAP, that extends fuzzy ARTMAP consisting of two fuzzy ART modules connected by an inter-ART map field to an extended architecture consisting of multiple ART modules.

References

  1. 1 2 3 Sun, Ron (2016-03-01). Anatomy of the Mind: Exploring Psychological Mechanisms and Processes with the Clarion Cognitive Architecture. Oxford University Press. doi:10.1093/acprof:oso/9780199794553.001.0001. ISBN   978-0-19-979455-3.
  2. Sun, Ron (2001-09-01). Duality of the Mind : A Bottom-up Approach Toward Cognition. Psychology Press. doi:10.4324/9781410604378. ISBN   978-1-4106-0437-8.
  3. Sun, Ron; Merrill, Edward; Peterson, Todd (2001). "From implicit skills to explicit knowledge: a bottom-up model of skill learning". Cognitive Science. 25 (2): 203–244. doi: 10.1207/s15516709cog2502_2 . ISSN   1551-6709.
  4. Sun, Ron; Zhang, Xi (2006-06-01). "Accounting for a variety of reasoning data within a cognitive architecture". Journal of Experimental & Theoretical Artificial Intelligence. 18 (2): 169–191. doi:10.1080/09528130600557713. ISSN   0952-813X. S2CID   29729283.
  5. Domangue, Thomas J.; Mathews, Robert C.; Sun, Ron; Roussel, Lewis G.; Guidry, Claire E. (2004). "Effects of Model-Based and Memory-Based Processing on Speed and Accuracy of Grammar String Generation". Journal of Experimental Psychology: Learning, Memory, and Cognition. 30 (5): 1002–1011. doi:10.1037/0278-7393.30.5.1002. ISSN   1939-1285. PMID   15355132.
  6. Wilson, Nicholas R.; Sun, Ron; Mathews, Robert C. (2009-07-01). "A motivationally-based simulation of performance degradation under pressure". Neural Networks. 22 (5–6): 502–508. doi:10.1016/j.neunet.2009.06.022. ISSN   0893-6080. PMID   19608380.

Coward, L.A. & Sun, R. (2004). Criteria for an effective theory of consciousness and some preliminary attempts. Consciousness and Cognition, 13, 268-301.

Helie, H. and Sun, R. (2010). Incubation, insight, and creative problem solving: A unified theory and a connectionist model. Psychological Review, 117, 994-1024.

Naveh, I. & Sun, R. (2006). A cognitively based simulation of academic science. Computational and Mathematical Organization Theory, 12, 313-337.

Sun, R. (2002). Duality of the Mind: A Bottom-up Approach Toward Cognition. Mahwah, NJ: Lawrence Erlbaum Associates.

Sun, R. (2016). Anatomy of the Mind: Exploring Psychological Mechanisms and Processes with the Clarion Cognitive Architecture. Oxford University Press, New York.

Sun, R. (2003). A Tutorial on CLARION 5.0. Technical Report, Cognitive Science Department, Rensselaer Polytechnic Institute.

Sun, R., Merrill, E., & Peterson, T. (2001). From implicit skills to explicit knowledge: A bottom-up model of skill learning. Cognitive Science, 25, 203-244. https://web.archive.org/web/20191218033659/http://www.cogsci.rpi.edu/~rsun/

Sun, R., Slusarz, P., & Terry, C. (2005). The interaction of the explicit and the implicit in skill learning: A dual-process approach. Psychological Review, 112, 159-192. https://web.archive.org/web/20191218033659/http://www.cogsci.rpi.edu/~rsun/

Sun, R. & Zhang, X. (2006). Accounting for a variety of reasoning data within a cognitive architecture. Journal of Experimental and Theoretical Artificial Intelligence , 18, 169-191.