Disbelief or minimization in response to threat warnings
Normalcy bias, or normality bias, is a cognitive bias which leads people to disbelieve or minimize threat warnings.[1] Consequently, individuals underestimate the likelihood of a disaster, when it might affect them, and its potential adverse effects.[2] The normalcy bias causes many people to prepare inadequately for natural disasters, market crashes, and calamities caused by human error. About 80% of people reportedly display normalcy bias during a disaster.[3]
Normalcy bias has also been called analysis paralysis, the ostrich effect,[4] and by first responders, the negative panic.[5] The opposite of normalcy bias is overreaction, or worst-case scenario bias,[6][7] in which small deviations from normality are dealt with as signals of an impending catastrophe.
Phases
Amanda Ripley, author of The Unthinkable: Who Survives When Disaster Strikes – and Why, identifies common response patterns of people in disasters and explains that there are three phases of response: "denial, deliberation, and the decisive moment". With regard to the first phase, described as "denial", Ripley found that people were likely to deny that a disaster was happening. It takes time for the brain to process information and recognize that a disaster is a threat. In the "deliberation" phase, people have to decide what to do. If a person does not have a plan in place, this causes a serious problem because the effects of life-threatening stress on the body (e.g. tunnel vision, audio exclusion, time dilations, out-of-body experiences, or reduced motor skills) limit an individual's ability to perceive information and make plans. Ripley asserts that in the third and final phase, described as the "decisive moment", a person must act quickly and decisively. Failure to do so can result in injury or death. She explains that the faster someone can get through the denial and deliberation phases, the quicker they will reach the decisive moment and begin to take action.[8]
Examples
Journalist David McRaney wrote that "Normalcy bias flows into the brain no matter the scale of the problem. It will appear whether you have days and plenty of warning or are blindsided with only seconds between life and death."[9] It can manifest itself in phenomena such as car crashes. Car crashes occur very frequently, but the average individual experiences them only rarely, if ever. It also manifests itself in connection with events in world history. According to a 2001 study by sociologist Thomas Drabek, when people are asked to leave in anticipation of a disaster, most check with four or more sources of information before deciding what to do. The process of checking in, known as milling, is common in disasters.[10]
It can explain why thousands of people refused to leave New Orleans as Hurricane Katrina approached[11] and why at least 70% of 9/11 survivors spoke with others before evacuating.[10] Officials at the White Star Line made insufficient preparations to evacuate passengers on the Titanic and people refused evacuation orders, possibly because they underestimated the odds of a worst-case scenario and minimized its potential impact.[12] Similarly, experts connected with the Fukushima nuclear power plant were strongly convinced that a multiple reactor meltdown could never occur.[13]
A website for police officers has noted that members of that profession have "all seen videos of officers who were injured or killed while dealing with an ambiguous situation, like the old one of a father with his young daughter on a traffic stop". In the video referred to, "the officer misses multiple threat cues...because the assailant talks lovingly about his daughter and jokes about how packed his minivan is. The officer only seems to react to the positive interactions, while seeming to ignore the negative signals. It's almost as if the officer is thinking, 'Well I've never been brutally assaulted before so it certainly won't happen now.' No one is surprised at the end of the video when the officer is violently attacked, unable to put up an effective defense." This professional failure, notes the website, is a consequence of normalcy bias.[14]
Normalcy bias, David McRaney has written, "is often factored into fatality predictions in everything from ship sinkings to stadium evacuations". Disaster movies, he adds, "get it all wrong. When you and others are warned of danger, you don't evacuate immediately while screaming and flailing your arms." McRaney notes that in the book Big Weather, tornado chaser Mark Svenvold discusses "how contagious normalcy bias can be. He recalled how people often tried to convince him to chill out while fleeing from impending doom. Even when tornado warnings were issued, people assumed it was someone else's problem. Stake-holding peers, he said, would try to shame him into denial so they could remain calm. They didn't want him deflating their attempts at feeling normal".[9]
Hypothesized cause
The normalcy bias may be caused in part by the way the brain processes new data. Research suggests that even when the brain is calm, it takes 8–10 seconds to process new information. Stress slows the process, and when the brain cannot find an acceptable response to a situation, it fixates on a single and sometimes default solution that may or may not be correct. An evolutionary reason for this response could be that paralysis gives an animal a better chance of surviving an attack and predators are less likely to see prey that is not moving.[10]
Effects
About 80% of people reportedly display normalcy bias in disasters.[3] Normalcy bias has been described as "one of the most dangerous biases we have". The lack of preparation for disasters often leads to inadequate shelter, supplies, and evacuation plans. Even when all these things are in place, individuals with a normalcy bias often refuse to leave their homes.[15][16]
Normalcy bias can cause people to drastically underestimate the effects of the disaster. Therefore, people think that they will be safe even though information from the radio, television, or neighbors gives them reasons to believe there is a risk. The normalcy bias causes a cognitive dissonance that people then must work to eliminate. Some manage to eliminate it by refusing to believe new warnings coming in and refusing to evacuate (maintaining the normalcy bias), while others eliminate the dissonance by escaping the danger. The possibility that some people may refuse to evacuate causes significant problems in disaster planning.[17]
Prevention
The negative effects of normalcy bias can be combated through the four stages of disaster response:[18]
preparation, including publicly acknowledging the possibility of disaster and forming contingency plans.
warning, including issuing clear, unambiguous, and frequent warnings and helping the public to understand and believe them.
impact, the stage at which the contingency plans take effect and emergency services, rescue teams, and disaster relief teams work in tandem.
aftermath, reestablishing equilibrium after the fact, by providing both supplies and aid to those in need.
Zebra, a medical slang term referring to a rare condition that may go undiagnosed or misdiagnosed due to medical staff identifying its symptoms as that of as a much more common (and often less severe) condition
^Drabek, Thomas E. (1986). Human system responses to disaster : an inventory of sociological findings. New York: Springer Verlag. p. 72. ISBN978-1-4612-4960-3. OCLC852789578. The initial response to a disaster warning is disbelief.
^Omer, Haim; Alon, Nahman (April 1994). "The continuity principle: A unified approach to disaster and trauma". American Journal of Community Psychology. 22 (2): 275–276. doi:10.1007/BF02506866. PMID7977181. S2CID21140114. ... normalcy bias consists in underestimating the probability of disaster, or the disruption involved in it ...
Omer, Haim; Alon, Nahman (1994). "The continuity principle: A unified approach to disaster and trauma". American Journal of Community Psychology. 22 (2): 273–287. doi:10.1007/BF02506866. PMID7977181. S2CID21140114.