Cultural cognition

The cultural cognition of risk, sometimes called simply cultural cognition, is the hypothesized tendency to perceive risks and related facts in relation to personal values. Research examining this phenomenon draws on a variety of social science disciplines including psychology, anthropology, political science, sociology, and communications. The stated objectives of this research are both to understand how values shape political conflict over facts (like whether climate change exists, whether gun control increases crime, whether vaccination of school girls for HPV threatens their health) and to promote effective deliberative strategies for resolving such conflicts consistent with sound empirical data.

Theory and evidence

The cultural cognition hypothesis holds that individuals are motivated by a variety of psychological processes to form beliefs about putatively dangerous activities that match their cultural evaluations of them. Persons who subscribe to relatively individualistic values, for example, tend to value commerce and industry and are inclined to disbelieve that such activities pose serious environmental risks. Persons who subscribe to relatively egalitarian and communitarian values, in contrast, readily credit claims of environmental risks, which is consistent with their moral suspicion of commerce and industry as sources of inequality and symbols of excessive self-seeking.[1][2]

Scholars have furnished two types of evidence to support the cultural cognition hypothesis. The first consists of general survey data that suggest that individuals’ values more strongly predict their risk perceptions than do other characteristics such as race, gender, economic status, and political orientations.[3][4]

The second type of evidence consists in experiments that identify discrete psychological processes that connect individuals’ values to their beliefs about risk and related facts.[5] Such experiments suggest, for example, that individuals selectively credit or dismiss information in a manner that reinforces beliefs congenial to their values.[6] They also show that individuals tend to be more persuaded by policy experts perceived to hold values similar to their own rather than by ones perceived to hold values different from them.[7] Such processes, the experiments suggest, often result in divisive forms of cultural conflict over facts, but can also be managed in fashions that reduce such disagreement.[8]

Cultural cognition project at Yale Law School

Funded by governmental and private foundation grants, much of the work on cultural cognition has been performed by an interdisciplinary group of scholars affiliated with the Cultural Cognition Project.[9] There are currently over a dozen project members from a variety of universities. Two members of the project—Dan Kahan and Douglas Kysar—are Yale Law School faculty, although other members (such as Donald Braman of George Washington University Law School and Geoffrey Cohen of Stanford University) were previously affiliated with Yale Law School or Yale University. Students from Yale University also contribute to Project research.

Significant findings

Science comprehension and cultural polarization

A study conducted by Cultural Cognition Project researchers (using a nationally representative U.S. sample) found that ordinary members of the public do not become more concerned about climate change as their science comprehension increases.[10] Instead, the degree of polarization among cultural groups with opposing predispositions increases.

Nanotechnology

The Cultural Cognition Project has conducted a series of studies on public perceptions of nanotechnology risks and benefits. Combining survey and experimental methods, the studies present evidence that individuals culturally predisposed to be skeptical of environmental risks are both more likely to seek out information on nanotechnology and more likely to infer from that information that nanotechnology’s benefits will outweigh its risks. Individuals culturally predisposed to credit environmental risks construe that same information, when exposed to it in the lab, as implying that nanotechnology’s risks will predominate.[6] The studies also present evidence that individuals tend to credit expert information on nanotechnology—regardless of its content—based on whether they share the perceived cultural values of the expert communicator.[11] The studies were issued by the Project on Emerging Nanotechnologies at the Woodrow Wilson International Center for Scholars, one of the research sponsors.

"Scientific consensus"

The same dynamics that motivate individuals of diverse cultural outlooks to form competing perceptions of risks are likely to cause them to form opposing perceptions of "scientific consensus", cultural cognition researchers have concluded.[12] In an experimental study, the researchers found that subjects were substantially more likely to count a scientist (of elite credentials) as an "expert" in his field of study when the scientist was depicted as taking a position consistent with the one associated with the subjects' cultural predispositions than when that scientist took a contrary position. A related survey showed that members of opposed cultural groups hold highly divergent impressions of what most scientific experts believe on various matters, a finding consistent with the ubiquity of culturally biased recognition of who counts as an "expert". Across a range of diverse risks (including climate change, nuclear waste disposal, and private handgun possession), members of no particular cultural group, the study found, were more likely than any other to hold perceptions of scientific consensus that consistently matched those adopted in "expert consensus reports" issued by the U.S. National Academy of Sciences.

Law

Scholars have also applied the cultural cognition of risk to legal issues. One such study examined how individuals reacted to a videotape of a high-speed police chase. In Scott v. Harris,[13] the U.S. Supreme Court (by a vote of 8-1) had held that no reasonable jury could view the tape and fail to find that the driver posed a lethal risk to the public large enough to justify deadly force by the police (namely, ramming the fleeing driver's vehicle, causing it to crash). The majority of study subjects agreed with the Court, but there were significant divisions along cultural lines.[14] Other studies have found that individuals' cultural worldviews influence their perceptions of consent in an acquaintance or date rape scenario,[15] and of the imminence of violence and other facts in self-defense cases involving either battered women or interracial confrontations.[16]

Relationship to other risk perception theories

Cultural cognition is a descendant of two other theories of risk perception. The first is the cultural theory of risk associated with anthropologist Mary Douglas and political scientist Aaron Wildavsky.[17] The cultural cognition hypothesis is derived from Douglas and Wildavsky's claim, advanced most notably in their controversial book Risk and Culture: An Essay on the Selection of Technical and Environmental Dangers (1982), that individuals selectively attend to risks in a manner that expresses and reinforces their preferred way of life.

Cultural cognition researchers, along with other scholars who have investigated Douglas and Wildavsky's theory empirically,[18] use attitudinal scales that reflect Douglas's worldview typology. That typology characterizes worldviews, or preferences about how society should be organized, along two cross-cutting dimensions: "group", which refers to how individualistic or group-oriented a society should be; and "grid", which refers to how hierarchical or egalitarian a society should be.[19]

The second theory is the "psychometric paradigm", to which Paul Slovic, a member of the Cultural Cognition Project, has made significant contributions. The psychometric paradigm links risk perceptions to various cognitive and social mechanisms that generally evade simpler, rational choice models associated with economics.[20][21] Cultural cognition theory posits that these mechanisms mediate between, or connect, individuals' cultural values to their perceptions of risk and other policy-relevant beliefs.

Combining the cultural theory of risk and the psychometric paradigm, cultural cognition, its exponents claim, remedies difficulties with each.[22] The mechanisms featured in the psychometric paradigm (and in social psychology generally) furnish a cogent explanation of why individuals adopt states of mind that fit and promote the aims of groups, including ones featured in Douglas’s culture theory. They do so, moreover, in a manner that avoids "functionalism," a criticized form of analysis that identifies group interests, rather than individual ones, as a cause for human action.[23][24] At the same time cultural theory, by asserting the orienting role of values, explains how the mechanisms featured in the psychometric paradigm can result in differences in risk perception among persons who hold different values. The interrelationship between individual values and perceptions of risk also calls into doubt the depiction of risk perceptions deriving from these mechanisms as products of irrationality or cognitive defect.[25]

Criticisms

Cultural cognition has been subjected to criticisms from a variety of sources. The rational choice economists Fremling & Lott (2003), as well as the psychologist Sjöberg (1998) have suggested that the theory (and others based on the cultural theory of risk generally) explain only a small fraction of the variation in popular risk perceptions. Mary Douglas herself has criticized cultural cognition for a conception of values that is too tightly modeled on American political disputes and that implicitly disparages the "hierarchical" worldview.[26] Finally, some scholars who emphasize elements of the psychometric paradigm suggest that the influence of cultural values on risk perceptions is best understood as simply an additional source of interference with the rational processing of information.[27]

See also

Notes

  1. Kahan (2010a), p. 296.
  2. Kahan et al. (2006), pp. 1083-84.
  3. Kahan & Braman (2006), pp. 155-158.
  4. Kahan et al. (2006), pp. 1086-87.
  5. Kahan et al. (2007), p. ?.
  6. Kahan et al. (2009), p. ?.
  7. Kahan et al. (2010), p. ?.
  8. Kahan (2007), p. 138-139.
  9. .
  10. Kahan et al. (2012), p. ?.
  11. Kahan et al. (2008), pp. 5, 14.
  12. Kahan, Jenkins-Smith & Braman (2010), p. 27.
  13. 550 U.S. 372 (2007).
  14. Kahan, Hoffman & Braman (2009), p. 837.
  15. Kahan (2010b).
  16. Kahan & Braman (2008).
  17. Rayner (1992), p. ?.
  18. For example, Dake (1991), Jenkins-Smith (2001), and Peters & Slovic (1996)
  19. Rayner (1992), pp. 87-91.
  20. Slovic (2000), p. ?.
  21. Kahneman, Slovic & Tversky (1982), p. ?.
  22. Kahan (2008), p. ?.
  23. Boholm (1996), pp. 68, 79-80.
  24. Kahan & Braman (2006), p. 252.
  25. Kahan et al. (2006), pp. 1088-1106.
  26. Douglas (2003).
  27. Sunstein (2006).

References

  • Boholm, Åsa (1996), "Risk perception and social anthropology: critique of cultural theory", Ethnos, 68 (2): 159–178, doi:10.1080/0014184032000097722, S2CID 145127724
  • Dake, Karl (1991), "Orienting Dispositions in the Perception of Risk: An Analysis of Contemporary Worldviews and Cultural Biases", Journal of Cross-Cultural Psychology, 22: 61–82, doi:10.1177/0022022191221006, S2CID 145367186
  • Douglas, Mary (2003), "Being Fair to Hierarchists", University of Pennsylvania Law Review, 151 (4): 1349–1370, doi:10.2307/3312933, JSTOR 3312933
  • Fremling, Gertrud M.; Lott, John R. (2003), "The Surprising Finding That "Cultural Worldviews" Don't Explain People's Views On Gun Control.", University of Pennsylvania Law Review, 151 (4): 1341–1348, doi:10.2307/3312932, JSTOR 3312932
  • Jenkins-Smith, Hank (2001), "Modeling Stigma: An Empirical Analysis of Nuclear Waste Images of Nevada.", in Flynn, James; Slovic, Paul; Kunreuther, Howard (eds.), Risk, Media, and Stigma, London; Sterling, VA: Earthscan., pp. 107–132
  • Kahan, Dan (2007), "The Cognitively Illiberal State", Stanford Law Review, 60: 115–154, SSRN 963929
  • Kahan, Dan (2008), "Cultural Cognition as a Conception of the Cultural Theory of Risk", Cultural Cognition Project Working Paper No. 73
  • Kahan, Dan (2010a), "Fixing the Communications Failure", Nature, 463 (7279): 296–297, doi:10.1038/463296a, PMID 20090734, S2CID 205052651
  • Kahan, Dan (2010b), "Culture, Cognition, and Consent: Who Perceives What, and Why, in Acquaintance Rape Cases", University of Pennsylvania Law Review, 158: 729–812, SSRN 1437742
  • Kahan, Dan; Braman, Donald (2006), "Cultural Cognition of Public Policy", Yale Journal of Law and Public Policy, 24: 147–170, SSRN 746508
  • Kahan, Dan; Braman, Donald (2008), "The Self-defensive Cognition of Self-defense", American Criminal Law Review, 45 (1): 1–65, SSRN 1012967
  • Kahan, Dan; Braman, Donald; Slovic, Paul; Gastil, John; Cohen, Geoffrey (2007), The Second National Risk and Culture Study: Making Sense of–and Making Progress in–the American Culture War of Fact, Harvard Law School Program on Risk Regulation Research Paper, 08–26, hdl:1794/22049, SSRN 1017189
  • Kahan, Dan; Braman, Donald; Cohen, Geoffrey; Slovic, Paul; Gastil, John (2010), "Who Fears the HPV Vaccine, Who Doesn't, and Why: An Experimental Study of the Mechanisms of Cultural Cognition", Law and Human Behavior, 34 (6): 501–516, doi:10.1007/s10979-009-9201-0, PMID 20076997, S2CID 28108691, SSRN 1160654
  • Kahan, Dan; Hoffman, D.A.; Braman, Donald (2009), "Whose Eyes are You Going to Believe? Scott v. Harris and the Perils of Cognitive Illiberalism", Harvard Law Review, 122 (3): 837–906
  • Kahan, Dan; Jenkins-Smith, Hank; Braman, Donald (2010), "Cultural Cognition of Scientific Consensus", Journal of Risk Research, 14 (2): 147–174, doi:10.1080/13669877.2010.511246, hdl:10.1080/13669877.2010.511246, S2CID 216092368, SSRN 1549444
  • Kahan, Dan; Peters, Ellen; Wittlin, Maggie; Slovic, Paul; Ouellette, Lisa Larrimore; Braman, Donald; Mandel, Gregory (2012), "The polarizing impact of science literacy and numeracy on perceived climate change risks", Nature Climate Change, 2 (10): 732–735, CiteSeerX 10.1.1.709.2300, doi:10.1038/nclimate1547, S2CID 53631207
  • Kahan, Dan; Slovic, Paul; Braman, Donald; Cohen, Geoffrey; Gastil, John (2009), "Cultural Cognition of the Risks and Benefits of Nanotechnology", Nature Nanotechnology, 4 (2): 87–90, doi:10.1038/nnano.2008.341, PMID 19197308
  • Kahan, Dan; Slovic, Paul; Braman, Donald; Gastil, John (2006), "Fear of Democracy: A Cultural Critique of Sunstein on Risk", Harvard Law Review, 119: 1071–1109
  • Kahneman, Daniel; Slovic, Paul; Tversky, Amos (1982), "Judgment Under Uncertainty: Heuristics and Biases", Science, Cambridge ; New York: Cambridge University Press, 185 (4157): 1124–31, doi:10.1126/science.185.4157.1124, PMID 17835457, S2CID 143452957
  • Peters, Ellen; Slovic, Paul (1996), "The Role of Affect and Worldviews as Orienting Dispositions in the Perception and Acceptance of Nuclear Power", Journal of Applied Social Psychology, 26 (16): 1427–1453, doi:10.1111/j.1559-1816.1996.tb00079.x, hdl:1794/22509
  • Rayner, Steve (1992), "Cultural Theory and Risk Analysis.", in Krimsky, Sheldon; Golding, Dominic (eds.), Social Theories of Risk, Westport, Conn: Praeger, pp. 83–115
  • Sjöberg, Lennart (1998), "World Views, Political Attitudes, and Risk Perception", Risk: Health, Safety and Environment, 9: 137–152
  • Slovic, Paul (2000), The Perception of Risk, London: Sterling, VA: Earthscan Publications
  • Sunsein, CR (2006), "Misfearing: A Reply" (PDF), Harvard Law Review, 119 (4): 110–1125, archived from the original (PDF) on 2012-03-19, retrieved 2011-08-17
  • Sunsein, CR (2006), "On the Divergence of American Reactions to Terrorism and Climate Change", John M. Olin Law and Economics Working Paper, 295

Further reading

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.