Authority bias

Last updated

Authority bias is the tendency to attribute greater accuracy to the opinion of an authority figure (unrelated to its content) and be more influenced by that opinion. [1] An individual is more influenced by the opinion of this authority figure, believing their views to be more credible, and hence place greater emphasis on the authority figure's viewpoint and are more likely to obey them. This concept is considered one of the social cognitive biases or collective cognitive biases. [2]

Contents

Humans generally have a deep-seated duty to authority and tend to comply when requested by an authority figure. [3] Some scholars explain that individuals are motivated to view authority as deserving of their position and this legitimacy leads people to accept and obey the decisions that it makes. [2] System justification theory articulates this phenomenon, particularly within its position that there is a psychological motivation for believing in the steadiness, stability and justness of the current social system. [4]

Authority bias can be measured concerning respect for authority, where higher respect for authority positively correlates with the increased likelihood of exhibiting authority bias. [5]

Definition

Authority bias, a term popularised by American psychologist Stanley Milgram, is defined as having an unreasonably high confidence in the belief that the information verified by a person with formal authority is correct, and therefore an individual is likely to be more influenced by them.[ citation needed ] Individuals in positions of authority are seen to be treated more favourably, where people believe their views with increased certainty, though the role of authority alone is not always significant enough to directly affect decision-making without this phenomenon being used in conjunction with other heuristics and biases.

The antonym of authority bias is blanket opposition to authority, disregarding their knowledge and believing authority figures to have inherently false claims. This relates to the view of anti-authoritarianism.

Authority bias has many explanations, rooted in the human need to obey authority figures. Namely, authority bias can be explained through evolutionary and social means.

The majority of people accept the police as a legitimate authority, with their position in the social hierarchy giving them the right to apprehend those violating the law, such as this Extinction Rebellion protester. A protester is arrested in Whitehall during the Extinction Rebellion protest, October 2019.jpg
The majority of people accept the police as a legitimate authority, with their position in the social hierarchy giving them the right to apprehend those violating the law, such as this Extinction Rebellion protester.

Legitimacy of authority

Authority bias is greater when the authority figure in question is seen to be legitimate, [6] that is they are accepted to be in an authorized position of relative power, and have the right to demand obedience. Research has highlighted certain important characteristics that can mark one as a legitimate authority figure, strengthening authority bias, with individuals more likely to be influenced by and obey such figures. These markers can affect whether people regard an individual as an authority figure.

Relative position in the social hierarchy

All societies are structured in a hierarchical manner, [7] and often those who possess a higher social status are situated further up the hierarchy. The majority of people accept that in order to maintain a co-operative, functioning society and avoid social chaos, a certain degree of individual choice must be relinquished to authority figures, often representing government institutions, whose role means they have greater status in the social hierarchy. For example, most accept that in the criminal justice system, the police have the right to exercise power over others and apprehend wrongdoers, as their role indicates their position in the social hierarchy [8] and their authority. This is often instilled in children from a young age, with parents instructing them to defer to authority figures such as teachers and policemen and accept the opinions of authority figures as correct. There is also a tendency to attribute greater accuracy to and respect the opinions of experts such as doctors. For example, most people have complete confidence in their doctor treating them, as a trusted authority figure with a high social status. [9] This normative behavior can be a useful shortcut, but blind acceptance of expert advice can raise issues if it becomes an automatic response. This was demonstrated by a study conducted by the psychiatrist Charles Hofling, who found that 21 out of 22 nurses would have given patients a potentially fatal dose of a drug when asked to by a doctor despite this instruction contravening official guidance forbidding the administration of the drug. Similar studies in other areas of society have demonstrated the problems that can arise as a result of unquestioningly accepting the opinions of those regarded as having higher status in the social hierarchy. [10]

Uniform

Outward appearance can signify an individual's social status and relative position in the social hierarchy, and consequently whether they are perceived as a legitimate authority figure. When examining authority bias, outward appearance, manifested in clothing, in particular uniform, seems to have a profound effect on whether an individual is respected and obeyed as an authority figure. Research conducted by Bickman and colleagues found that passers-by were twice as likely to obey a confederate dressed as a security guard rather than a milkman when asked by the individual to complete tasks such as picking up litter or lending the confederate a coin for the parking meter. [11] The effects of uniform on authority bias has been a fairly consistent and reliable finding as other studies, such as those run by Milgram, have reached similar conclusions. In a variation of his original experiment, Milgram found that obedience levels dropped from 65% to 20% when the instructions to enable shocks were given by an ordinary member of the public wearing plain clothes rather than a researcher (signified by a white lab coat). [12] In the real world, attention bias may also play a role here, with people selectively paying more attention to the uniform as an important symbol of power, attributing greater accuracy to the opinions of the wearer and displaying more obedience to that figure in response. However, it is important to recognize that factors like uniform are situational and easily changed. Thus, an individual may masquerade as an authority figure and can elicit the authority bias response from others, even though their authority may not be truly legitimate. Though situational factors may point to an authority figure, the effects of authority bias may be stronger when such situational factors are combined with innate factors, such as gender.

Gender

Evidence also suggests that gender can be important in signifying an authority figure. A subset of authority bias, namely gender-authority bias explains how in particular, women may be more susceptible to authority bias than males. In addition to women being more influenced by authority figures than males, female authority figures may be less influential than their male counterparts. Research has shown that both men and women associated men with high authority roles and women with low authority roles, showing how gender can signify an authority figure and subsequently influence authority bias. [13] As a result, confirmation bias may also play a role here in what people interpret to be a characteristic of an authority figure; as many associate higher-authority roles with males, confirmation bias suggests that people will tend to look for, favor, interpret a recall information that fits with this belief, perpetuating this bias, reinforcing the idea of males being more legitimate as authority figures. Gender-authority bias has been demonstrated across a variety of real-world sectors. In politics, research into leadership using the Implicit Association Test (IAT) found that female political leaders tend to face greater resistance to their authority in comparison to their male counterparts. [14] Authority bias is therefore strengthened when an authority figure is male instead of female in politics. In finance, males are often favored as authority figures, being perceived to have greater control of resources, and able to make better decisions, demonstrated by the fact that they hold higher positions and on average earn 66% more than females in the finance industry. [15] Research also suggests that women are more susceptible to authority bias, as they are more influenced than men by hearing fake news from an authority figure. [16]

Discovery

Prior to psychological research, the most common example of this phenomenon was when people obeyed Hitler during World War II, though such effects have been evident throughout history. This relates to pluralistic ignorance, in which authority figures are obeyed regardless of immorality.

The term “authority bias” was first mentioned in literature in reference to state authority bias, in which it simply indicated a preference for being pro-state or anti-state in the US federal election. [17]

Nevertheless, the first-time authority bias was referenced in literature as a cognitive bias was a result of Milgram's experiment, in which it was used to explain obedience to authority figures. [18] Whilst Milgram did not directly use the term "authority bias" in his 1963 paper, the obedience effect identified from his study became the primary example of authority bias. Milgram's findings succeeded the reason why people during World War II obeyed Hitler; participants voluntarily submitted to the authority figure (the experimenter wearing a white lab coat, signifying professionalism). Authority bias is further strengthened through the use of uniforms to signify authority, initially investigated in Milgram's situational variable (where obedience decreased when the uniform of the experimenter was changed from a lab coat to everyday clothes), but further replicated through Bickman's infamous research into obedience, [19] where security guards are more likely to be obeyed without question and thus contributing directly to authority bias.

Real-world effects

Advertising

Authority bias is used as a marketing strategy in order to increase the legitimacy of claims made about a product. A common example in advertising is where toothpaste companies such as Sensodyne promote the validity of their claims by ensuring the dentists wear lab coats, resulting in the consumer being more trustworthy of the product and consequently more likely to buy the product.

Personalised advertising in relation to political voting attitudes (particularly in the US election) relies on authority bias. [20] Political campaigns are targeted specifically towards female voters (who are more susceptible to the bias), amplified through the use of social media, in which political leaders and other figures of authority are used in ad-campaigns to increase the effectiveness of their claims.

Medicine

The expert halo effect is synonymous with authority bias in medicine, where the expert is seen as infallible. [21] Issues arise in pharmaceutical settings, in which non-experts blindly follow expert's commands, resulting in the distribution of harmful drugs [22] and inappropriate healthcare practices. [23]

A further issue concerning the extent to which an authority figure is perceived to be providing accurate information is apparent in cases such as that of Willie Jackson. Forensic dentistry falsely proved Jackson to be guilty, yet the authority bias strengthened the doctor's standpoint in a court of law as they had expert authority bias. Consequently, the negative effect of authority bias has led to wrongful convictions.

Business

The authority bias is demonstrated in the case of the highest-paid persons' opinion (HIPPO) impact, which describes how employees and other stakeholders in the solution environment tend to go with the opinions and impressions of the highly paid people in an organization. [24]

Explanations

Evolution

Evolution has established a dominance hierarchy in which it is an evolutionary advantage to obey authority figures, as figures of authority have a greater allocation of resources and other means of survival. The logical fallacy of ad verecundiam is evolutionary, highlighting that experts are more reliable due to a lack of opposing information, increasing trustworthiness. [25]

Societal benefits

In any society, a diverse and widely accepted system of authority allows the development of sophisticated structures for the production of resources, trade, expansion and social control. Notions of obedience and loyalty to the legitimate rule of others are generally accorded values in schools, the law, the military and in political systems. The strength of the bias to obey a legitimate authority figure comes from systemic socialization practices designed to instill the perception that such obedience constitutes correct behavior, and that genuine authority figures usually possess higher degrees of knowledge, wisdom and power. Different societies vary the terms of this dimension. [26] [ self-published source ] As a result, authority bias can be rooted in the underlying social norms of society. Consequently, deference to authority can occur mindlessly as a kind of decision-making short cut. [27]

The role of other heuristics and biases

Research support for the strength of authority bias is evident, however, the effect is not significant in some instances. [28] Research is merely correlational, and hence other behavioural effects experienced in conjunction with authority bias strengthen its effects.

Confirmation bias

An individual exhibiting authority bias may also be subject to experiencing confirmation bias, which is the tendency to search for information that confirms your own existing beliefs. Research suggests that greater authority is given to financial advisors who confirm one's existing opinions, implying that authority bias is strengthened when it coincides with confirmation bias. [29]

Bandwagon effect

The bandwagon effect is where people adopt the ideologies of those surrounding them. Society favours the opinions of authority figures, hence it is a majority view which others support. [30]

See also

Related Research Articles

<span class="mw-page-title-main">Milgram experiment</span> Series of social psychology experiments

Beginning on August 7, 1961, a series of social psychology experiments were conducted by Yale University psychologist Stanley Milgram, who intended to measure the willingness of study participants to obey an authority figure who instructed them to perform acts conflicting with their personal conscience. Participants were led to believe that they were assisting an unrelated experiment, in which they had to administer electric shocks to a "learner". These fake electric shocks gradually increased to levels that would have been fatal had they been real.

<span class="mw-page-title-main">Persuasion</span> Umbrella term of influence and mode of communication

Persuasion or persuasion arts is an umbrella term for influence. Persuasion can influence a person's beliefs, attitudes, intentions, motivations, or behaviours.

Social psychology is the scientific study of how thoughts, feelings, and behaviors are influenced by the actual, imagined, or implied presence of others. Social psychologists typically explain human behavior as a result of the relationship between mental states and social situations, studying the social conditions under which thoughts, feelings, and behaviors occur, and how these variables influence social interactions.

<span class="mw-page-title-main">Stanley Milgram</span> American social psychologist

Stanley Milgram was an American social psychologist known for his controversial experiments on obedience conducted in the 1960s during his professorship at Yale.

Confirmation bias is the tendency to search for, interpret, favor, and recall information in a way that confirms or supports one's prior beliefs or values. People display this bias when they select information that supports their views, ignoring contrary information, or when they interpret ambiguous evidence as supporting their existing attitudes. The effect is strongest for desired outcomes, for emotionally charged issues, and for deeply entrenched beliefs.

Obedience, in human behavior, is a form of "social influence in which a person yields to explicit instructions or orders from an authority figure". Obedience is generally distinguished from compliance, which some authors define as behavior influenced by peers while others use it as a more general term for positive responses to another individual's request, and from conformity, which is behavior intended to match that of the majority. Depending on context, obedience can be seen as moral, immoral, or amoral. For example, in psychological research, individuals are usually confronted with immoral demands designed to elicit an internal conflict. If individuals still choose to submit to the demand, they are acting obediently.

Social influence comprises the ways in which individuals adjust their behavior to meet the demands of a social environment. It takes many forms and can be seen in conformity, socialization, peer pressure, obedience, leadership, persuasion, sales, and marketing. Typically social influence results from a specific action, command, or request, but people also alter their attitudes and behaviors in response to what they perceive others might do or think. In 1958, Harvard psychologist Herbert Kelman identified three broad varieties of social influence.

  1. Compliance is when people appear to agree with others but actually keep their dissenting opinions private.
  2. Identification is when people are influenced by someone who is liked and respected, such as a famous celebrity.
  3. Internalization is when people accept a belief or behavior and agree both publicly and privately.

The authoritarian personality is a personality type characterized by a disposition to treat authority figures with unquestioning obedience and respect. Conceptually, the term authoritarian personality originated from the writings of Erich Fromm, and usually is applied to people who exhibit a strict and oppressive personality towards their subordinates. Regardless of whether authoritarianism is more of a personality, attitude, ideology or disposition, scholars find it has significant influence on public opinion and political behavior.

<span class="mw-page-title-main">Risk perception</span>

Risk perception is the subjective judgement that people make about the characteristics and severity of a risk. Risk perceptions often differ from statistical assessments of risk since they are affected by a wide range of affective, cognitive, contextual, and individual factors. Several theories have been proposed to explain why different people make different estimates of the dangerousness of risks. Three major families of theory have been developed: psychology approaches, anthropology/sociology approaches and interdisciplinary approaches.

Moral disengagement is a meaning from Developmental psychology, educational psychology and social psychology for the process of convincing the self that ethical standards do not apply to oneself in a particular context. This is done by separating moral reactions from inhumane conduct and disabling the mechanism of self-condemnation. Thus, moral disengagement involves a process of cognitive re-construing or re-framing of destructive behavior as being morally acceptable without changing the behavior or the moral standards.

<i>Obedience to Authority: An Experimental View</i> 1974 book by Stanley Milgram

Obedience to Authority: An Experimental View is a 1974 book by social psychologist Stanley Milgram concerning a series of experiments on obedience to authority figures he conducted in the early 1960s. This book provides an in-depth look into his methods, theories and conclusions.

Compliance is a response—specifically, a submission—made in reaction to a request. The request may be explicit or implicit. The target may or may not recognize that they are being urged to act in a particular way.

The Semmelweis reflex or "Semmelweis effect" is a metaphor for the reflex-like tendency to reject new evidence or new knowledge because it contradicts established norms, beliefs, or paradigms.

Conformity is the act of matching attitudes, beliefs, and behaviors to group norms, politics or being like-minded. Norms are implicit, specific rules, guidance shared by a group of individuals, that guide their interactions with others. People often choose to conform to society rather than to pursue personal desires – because it is often easier to follow the path others have made already, rather than forging a new one. Thus, conformity is sometimes a product of group communication. This tendency to conform occurs in small groups and/or in society as a whole and may result from subtle unconscious influences, or from direct and overt social pressure. Conformity can occur in the presence of others, or when an individual is alone. For example, people tend to follow social norms when eating or when watching television, even if alone.

<span class="mw-page-title-main">Social experiment</span> Psychological or sociological research

A social experiment is a method of psychological or sociological research that observes people's reactions to certain situations or events. The experiment depends on a particular social approach where the main source of information is the participants' point of view and knowledge. To carry out a social experiment, specialists usually split participants into two groups — active participants and respondents. Throughout the experiment, specialists monitor participants to identify the effects and differences resulting from the experiment. A conclusion is then created based on the results. Intentional communities are generally considered social experiments.

An argument from authority is a form of argument in which the opinion of an authority figure is used as evidence to support an argument.

Social determinism is the theory that social interactions alone determine individual behavior.

Moral blindness, also known as ethical blindness, is defined as a person's temporary inability to see the ethical aspect of a decision they are making. It is often caused by external factors due to which an individual is unable to see the immoral aspect of their behavior in that particular situation.

Belief perseverance is maintaining a belief despite new information that firmly contradicts it.

References

  1. Milgram, Stanley (1963). "Behavioral study of obedience". The Journal of Abnormal and Social Psychology. 67 (4): 371–378. doi:10.1037/h0040525. PMID   14049516. S2CID   18309531.
  2. 1 2 Juárez Ramos, Veronica (2019). Analyzing the Role of Cognitive Biases in the Decision-Making Process. Hershey, PA: IGI Global. p. 113. ISBN   9781522529798.
  3. Milgram, Stanley (1974). Obedience to Authority: An Experimental View .
  4. Browstein, Michael (2016). Implicit Bias and Philosophy, Volume 1: Metaphysics and Epistemology. Oxford: Oxford University Press. p. 240. ISBN   9780198766186.
  5. "Respect for Authority". Cato Institute. Retrieved 2021-02-16.
  6. van der Toorn, Jojanneke; Tyler, Tom R.; Jost, John T. (2011-01-01). "More than fair: Outcome dependence, system justification, and the perceived legitimacy of authority figures". Journal of Experimental Social Psychology. 47 (1): 127–138. doi:10.1016/j.jesp.2010.09.003. ISSN   0022-1031.
  7. Cloutier, Jasmin; Cardenas-Iniguez, Carlos; Gyurovski, Ivo; Barakzai, Anam; Li, Tianyi (2016-01-01), Absher, John R.; Cloutier, Jasmin (eds.), "Chapter 9 - Neuroimaging Investigations of Social Status and Social Hierarchies", Neuroimaging Personality, Social Cognition, and Character, San Diego: Academic Press, pp. 187–203, doi:10.1016/b978-0-12-800935-2.00009-9, ISBN   978-0-12-800935-2 , retrieved 2022-02-09
  8. Jackson, Jonathan; Bradford, Ben; Hough, Mike; Myhill, Andy; Quinton, Paul King; Tyler, Tom (2012). "Why Do People Comply with the Law? Legitimacy and the Influence of Legal Institutions". SSRN Electronic Journal. doi:10.2139/ssrn.1994490. ISSN   1556-5068.
  9. "Confidence and trust in clinicians". The Nuffield Trust. 2018-10-15. Retrieved 2022-02-10.
  10. Tarnow, Eugen (2000-01-01). "Towards the Zero Accident Goal: Assisting the First Officer: Monitor and Challenge Captain Errors". Journal of Aviation/Aerospace Education & Research. 10 (1). doi: 10.15394/jaaer.2000.1269 . ISSN   2329-258X.
  11. Bickman, Leonard (1974). "The Social Power of a Uniform1". Journal of Applied Social Psychology. 4 (1): 47–61. doi:10.1111/j.1559-1816.1974.tb02599.x. ISSN   1559-1816.
  12. "Milgram Experiment | Simply Psychology". www.simplypsychology.org. Retrieved 2022-02-14.
  13. Rudman, Laurie A.; Kilianski, Stephen E. (2000). "Implicit and Explicit Attitudes Toward Female Authority". Personality and Social Psychology Bulletin. 26 (11): 1315–1328. doi:10.1177/0146167200263001. ISSN   0146-1672. S2CID   6962810.
  14. Yang, Ke; Zhu, John Jianjun; Santoro, Michael D. (2017-08-01). "Inter-Firm Managerial Social Ties and Strategic Alliances Formation: A Multiplexity Perspective". Academy of Management Proceedings. 2017 (1): 10151. doi: 10.5465/AMBPP.2017.27 . ISSN   0065-0668.
  15. "Women finance firm directors earn 66% less than men in UK, study finds". Reuters. 2021-03-01. Retrieved 2022-02-09.
  16. Jost, Peter J.; Pünder, Johanna; Schulze-Lohoff, Isabell (2020-04-01). "Fake news - Does perception matter more than the truth?". Journal of Behavioral and Experimental Economics. 85: 101513. doi:10.1016/j.socec.2020.101513. ISSN   2214-8043. S2CID   213652004.
  17. Sprague, John D.; Incorporated, Macmillan Publishing Company (1968). Voting Patterns of the United States Supreme Court: Cases in Federalism, 1889-1959. Ardent Media.
  18. Ellis RM (2015). Middle Way Philosophy: Omnibus Edition. Lulu Press. ISBN   9781326351892.
  19. Bickman, Leonard (1974). "The Social Power of a Uniform1". Journal of Applied Social Psychology. 4 (1): 47–61. doi:10.1111/j.1559-1816.1974.tb02599.x. ISSN   1559-1816.
  20. Kruikemeier, Sanne; Sezgin, Minem; Boerman, Sophie C. (June 2016). "Political Microtargeting: Relationship Between Personalized Advertising on Facebook and Voters' Responses". Cyberpsychology, Behavior, and Social Networking. 19 (6): 367–372. doi:10.1089/cyber.2015.0652. ISSN   2152-2715. PMID   27327063.
  21. Austin, Jared P.; Foster, Byron A. (2019-07-01). "How Pediatric Hospitalists Must Contend With the Expert Halo Effect". Hospital Pediatrics. 9 (7): 560–562. doi: 10.1542/hpeds.2019-0053 . ISSN   2154-1663. PMID   31175143.
  22. Austin, Jared P.; Halvorson, Stephanie A. C. (2019-02-05). "Reducing the Expert Halo Effect on Pharmacy and Therapeutics Committees". JAMA. 321 (5): 453–454. doi:10.1001/jama.2018.20789. ISSN   1538-3598. PMID   30657521. S2CID   58618466.
  23. Seshia, Shashi; Makhinson, Michael; Young, G Bryan (2015). "'Cognitive biases plus': covert subverters of healthcare evidence". Evidence-Based Medicine. 21 (2): 41–5. doi: 10.1136/ebmed-2015-110302 . PMID   26612071. S2CID   353523.
  24. Garrett, John (2018-10-24). Data Analytics for IT Networks: Developing Innovative Use Cases. Cisco Press. ISBN   9780135183441.
  25. Oswald, Steve; Hart, Christopher (2013-05-22). "Trust based on bias: Cognitive constraints on source-related fallacies". OSSA Conference Archive.
  26. Hinnosaar, Marit; Hinnosaar, Toomas (2012). "Authority Bias". marit.hinnosaar.net.[ permanent dead link ]
  27. Huczynski, Andrzej (2004). Influencing within organizations. Routledge.
  28. Scruggs, Meredith (2020-01-01). "Impact of the CSI Effect and Authority Bias on Juror Decisions". Undergraduate Honors Theses.
  29. Zaleskiewicz, Tomasz; Gasiorowska, Agata (2020). "Evaluating experts may serve psychological needs: Self-esteem, bias blind spot, and processing fluency explain confirmation effect in assessing financial advisors' authority". Journal of Experimental Psychology: Applied. 27 (1): 27–45. doi:10.1037/xap0000308. PMID   32597675. S2CID   220258183.
  30. Howard, Jonathan (2019), Howard, Jonathan (ed.), "Bandwagon Effect and Authority Bias", Cognitive Errors and Diagnostic Mistakes: A Case-Based Guide to Critical Thinking in Medicine, Cham: Springer International Publishing, pp. 21–56, doi:10.1007/978-3-319-93224-8_3, ISBN   978-3-319-93224-8, S2CID   81486453 , retrieved 2021-02-17