Belief bias is the tendency to judge the strength of arguments based on the plausibility of their conclusion rather than how strongly they support that conclusion. [1] A person is more likely to accept an argument that supports a conclusion that aligns with their values, beliefs and prior knowledge, while rejecting counter arguments to the conclusion. [2] Belief bias is an extremely common and therefore significant form of error; we can easily be blinded by our beliefs and reach the wrong conclusion. Belief bias has been found to influence various reasoning tasks, including conditional reasoning, [3] relation reasoning [4] and transitive reasoning. [5]
A syllogism is a kind of logical argument in which one proposition (the conclusion) is inferred from two or more others (the premises ) of a specific form. The classical example of a valid syllogism is:
An example of an invalid syllogism is:
Typically, a majority of test subjects in studies incorrectly identify this syllogism as one in which the conclusion follows from the premises. [1] It might be true in the real world that a) girls study and b) this is because they are ambitious. However, this argument is a fallacy, because the conclusion is not supported by its premises. The validity of an argument is independent from the truth of its conclusion: there are valid arguments for false conclusions and invalid arguments for true conclusions. Hence, it is an error to judge the validity of an argument from the plausibility of its conclusion. This is the reasoning error known as belief bias. [1]
When a person gives a response that is determined by the believability of the conclusion rather than logical validity, this is referred to as belief bias only when a syllogism is used. This phenomenon is so closely related to syllogistic reasoning that, when it does occur, in areas such as Wason's selection task or the THOG problem, it is called "memory cueing" or the "effects of content". [2]
Many researchers in thinking and reasoning have provided evidence for a dual-process cognitive approach to reasoning, judgment and decision making. They argue that these two mental processes (system 1 and system 2) engage in a constant battle for control over our brain to reason and make decisions. System 1 can be described as an automatic response system [6] characterised by "unconscious", [7] "intuitive" [8] and "rapid" [6] evaluation; whereas system 2 is said to be a controlled response system, [6] characterised by “conscious”, [7] “analytic” [8] and “slow” [6] evaluation; some researchers even claimed to have found a link between general intelligence and the effectiveness of decision making. [9] [10] It is important to note that the dual-process cognitive theory is different from the two minds hypothesis. Research done by Jonathan St B. T. Evans in 2007 provided evidence for the view that System 1, which serves as a quick heuristic processor, fights for control over System 2's slower analytical approach. [11] In the experiment, participants were asked to evaluate syllogisms that have valid arguments with unconvincing conclusions; valid arguments with convincing conclusions; invalid arguments with unconvincing conclusions; invalid arguments with convincing conclusions. The results show that when the conclusion is believable, people blindly accept invalid conclusions more so than invalid arguments are accepted.
Various studies have proved that the time period for which a subject is allowed to think when evaluating arguments is related to the tendency for belief bias to take place. In a study done by Evans and Holmes in 2005, [12] they recruited two different groups of people to answer a series of reasoning questions. One group of people were given only two seconds to answer the questions; whereas the other group of people were allowed to use as much time as they would like to answer the questions. The result obtained was that a higher percentage of incorrect answers were found in the time pressured group than the other; they concluded that this was a result of a shift in logical to belief-biased way of thinking.
The nature of the content presented can also affect belief bias of an individual as shown by a study done by Goel & Vartanian in 2011. [13] In their experiment, 34 participants were presented with a syllogism upon each trial. Each trial were either neutral or carried some degree of negative content. Negative content involved in the experiment were politically incorrect social norm violations, such as the statement “Some wars are not unjustified, Some wars involve raping of women, therefore, Some raping of women is not unjustified”. For syllogisms where the content was neutral, the results were consistent with studies of belief bias; however, for syllogisms with negative emotional content, participants were more likely to reason logically on invalid syllogisms with believable conclusions instead of automatically judging them to be valid. In other words, the effect of belief bias is reduced when the content presented is of negative emotion. According to Goel and Vartanian, this is because negative emotions prompts us to reason in more careful and in more detail. This argument was supported by the observation that for questions with negative emotions, the reaction time taken was significantly longer than that of questions that a neutral.
In an experiment done by Evans, Newstead, Allen & Pollard in 1994, [14] where subjects were given detailed instructions which lack specific reference to the notion of logical necessity when answering questions, it was shown that a larger proportion of answers actually rejected invalid arguments with convincing conclusions as opposed to when no further instructions were given when the subjects were asked to answer the questions. The results of the experiments reflects that when elaborated instructions were given to subjects to reason logically, the effects of belief bias is decreased.
In a series of experiments by Evans, Barston and Pollard (1983), [15] participants were presented with evaluation task paradigms, containing two premises and a conclusion. The participants were asked to make an evaluation of logical validity. The subjects, however, exhibited a belief bias, evidenced by their tendency to reject valid arguments with unbelievable conclusions, and endorse invalid arguments with believable conclusions. Instead of following directions and assessing logical validity, the subjects based their assessments on personal beliefs.
Consequently, these results demonstrated a greater acceptance of more believable (80%), than unbelievable (33%) conclusions. Participants also illustrated evidence of logical competences and the results determined an increase in acceptance of valid (73%) than invalid (41%). Additionally, there's a small difference between believable and valid (89%) in comparison to unbelievable and invalid (56%) (Evans, Barston & Pollard, 1983; Morley, Evans & Handley, 2004). [15] [16]
It has been argued that using more realistic content in syllogisms can facilitate more normative performance from participants. It has been suggested that the use of more abstract, artificial content will also have a biasing effect on performance. Therefore, more research is required to fully understand how and why belief bias occurs and if there are certain mechanisms that are responsible for such things. [17] There is also evidence of clear individual differences in normative responding that are predicted by the response times of participants. [18]
A 1989 study by Markovits and Nantel gave participants four reasoning tasks. The results indicated “a significant belief-bias effect” that existed “independently of the subjects' abstract reasoning ability.” [19]
A 2010 study by Donna Torrens examined differences in belief bias among individuals. Torrens found that “the extent of an individual's belief bias effect was unrelated to a number of measures of reasoning competence” but was, instead, related to that person's ability “to generate alternative representations of premises: the more alternatives a person generated, the less likely they were to show a belief bias effect." [20]
In a 2010 study, Chad Dube and Caren M. Rotello of the University of Massachusetts and Evan Heit of the University of California, Merced, showed that “the belief bias effect is simply a response bias effect.” [21]
In a 2012 study, Adrian P. Banks of the University of Surrey explained that “belief bias is caused by the believability of a conclusion in working memory which influences its activation level, determining its likelihood of retrieval and therefore its effect on the reasoning process.” [22]
Michelle Colleen and Elizabeth Hilscher of the University of Toronto showed in 2014 that belief bias can be affected by the difficulty level and placement of the syllogism in question. [23]
A cognitive bias is a systematic pattern of deviation from norm or rationality in judgment. Individuals create their own "subjective reality" from their perception of the input. An individual's construction of reality, not the objective input, may dictate their behavior in the world. Thus, cognitive biases may sometimes lead to perceptual distortion, inaccurate judgment, illogical interpretation, and irrationality.
A syllogism is a kind of logical argument that applies deductive reasoning to arrive at a conclusion based on two propositions that are asserted or assumed to be true.
A fallacy, also known as paralogia in modern psychology, is the use of invalid or otherwise faulty reasoning in the construction of an argument that may appear to be well-reasoned if unnoticed. The term was introduced in the Western intellectual tradition by the Aristotelian De Sophisticis Elenchis.
Confirmation bias is the tendency to search for, interpret, favor, and recall information in a way that confirms or supports one's prior beliefs or values. People display this bias when they select information that supports their views, ignoring contrary information, or when they interpret ambiguous evidence as supporting their existing attitudes. The effect is strongest for desired outcomes, for emotionally charged issues, and for deeply entrenched beliefs. Confirmation bias cannot be eliminated, but it can be managed, for example, by education and training in critical thinking skills.
Deductive reasoning is the mental process of drawing deductive inferences. An inference is deductively valid if its conclusion follows logically from its premises, i.e. it is impossible for the premises to be true and the conclusion to be false.
Inferences are steps in reasoning, moving from premises to logical consequences; etymologically, the word infer means to "carry forward". Inference is theoretically traditionally divided into deduction and induction, a distinction that in Europe dates at least to Aristotle. Deduction is inference deriving logical conclusions from premises known or assumed to be true, with the laws of valid inference being studied in logic. Induction is inference from particular evidence to a universal conclusion. A third type of inference is sometimes distinguished, notably by Charles Sanders Peirce, contradistinguishing abduction from induction.
Moral reasoning is the study of how people think about right and wrong and how they acquire and apply moral rules. It is a subdiscipline of moral psychology that overlaps with moral philosophy, and is the foundation of descriptive ethics.
In logic and philosophy, a formal fallacy, deductive fallacy, logical fallacy or non sequitur is a pattern of reasoning rendered invalid by a flaw in its logical structure that can neatly be expressed in a standard logic system, for example propositional logic. It is defined as a deductive argument that is invalid. The argument itself could have true premises, but still have a false conclusion. Thus, a formal fallacy is a fallacy where deduction goes wrong, and is no longer a logical process. This may not affect the truth of the conclusion, since validity and truth are separate in formal logic.
In psychology, a dual process theory provides an account of how thought can arise in two different ways, or as a result of two different processes. Often, the two processes consist of an implicit (automatic), unconscious process and an explicit (controlled), conscious process. Verbalized explicit processes or attitudes and actions may change with persuasion or education; though implicit process or attitudes usually take a long amount of time to change with the forming of new habits. Dual process theories can be found in social, personality, cognitive, and clinical psychology. It has also been linked with economics via prospect theory and behavioral economics, and increasingly in sociology through cultural analysis.
The need for cognition (NFC), in psychology, is a personality variable reflecting the extent to which individuals are inclined towards effortful cognitive activities.
Hot cognition is a hypothesis on motivated reasoning in which a person's thinking is influenced by their emotional state. Put simply, hot cognition is cognition coloured by emotion. Hot cognition contrasts with cold cognition, which implies cognitive processing of information that is independent of emotional involvement. Hot cognition is proposed to be associated with cognitive and physiological arousal, in which a person is more responsive to environmental factors. As it is automatic, rapid and led by emotion, hot cognition may consequently cause biased decision making. Hot cognition may arise, with varying degrees of strength, in politics, religion, and other sociopolitical contexts because of moral issues, which are inevitably tied to emotion. Hot cognition was initially proposed in 1963 by Robert P. Abelson. The idea became popular in the 1960s and the 1970s.
An argument is a series of sentences, statements or propositions some of which are called premises and one is the conclusion. The purpose of an argument is to give reasons for one's conclusion via justification, explanation, and/or persuasion.
The psychology of reasoning is the study of how people reason, often broadly defined as the process of drawing conclusions to inform how people solve problems and make decisions. It overlaps with psychology, philosophy, linguistics, cognitive science, artificial intelligence, logic, and probability theory.
Cognitive inertia is the tendency for a particular orientation in how an individual thinks about an issue, belief, or strategy to resist change. In clinical and neuroscientific literature, it is often defined as a lack of motivation to generate distinct cognitive processes needed to attend to a problem or issue. The physics term inertia is to emphasize the rigidity and resistance to change in the method of cognitive processing that has been in use for a significant amount of time. Commonly confused with belief perseverance, cognitive inertia is the perseverance of how one interprets information, not the perseverance of the belief itself.
Motivated reasoning is a cognitive and social response in which individuals, consciously or unconsciously, allow emotion-loaded motivational biases to affect how new information is perceived. Individuals tend to favor evidence that coincides with their current beliefs and reject new information that contradicts them, despite contrary evidence.
The rhyme-as-reason effect, or Eaton–Rosen phenomenon, is a cognitive bias whereupon a saying or aphorism is judged as more accurate or truthful when it is rewritten to rhyme.
Logic-based therapy (LBT) is a proposed modality of philosophical counseling developed by philosopher Elliot D. Cohen beginning in the mid-1980s. It is a philosophical variant of rational emotive behavior therapy (REBT), which was developed by psychologist Albert Ellis. However, there have been no independent, controlled studies to measure its therapeutic value or advantages over classical REBT.
belief bias.