Psychologically Illogical – Part 1
Unhelpful human psychological tendencies, such as avoiding cognitive dissonance.
In this series of essays, and following up on a previous series of essays on “myside” bias, we’ll explore evidence regarding some particularly unhelpful human psychological tendencies that inhibit good decision-making. Indeed, these human psychological tendencies that often lead us to judge poorly are a main reason the principles of free speech should be supported, as the existence of these psychological tendencies should put into at least some doubt even the most fervently held beliefs.
In her book Mistakes Were Made (But Not By Me): Why We Justify Foolish Beliefs, Bad Decisions, and Hurtful Acts, Carrol Tavris writes:
Most people, when directly confronted by evidence that they are wrong, do not change their point of view or course of action but justify it even more tenaciously. Even irrefutable evidence is rarely enough to pierce the mental armor of self-justification … [T]here is a big difference between what a guilty man says to the public to convince them of something he knows is untrue (“I did not have sex with that woman”; “I am not a crook”), and the process of persuading himself that he did a good thing. In the former situation, he is lying and knows he is lying to save his own skin. In the latter, he is lying to himself. That is why self-justification is more powerful and more dangerous than the explicit lie. It allows people to convince themselves that what they did was the best thing they could have done.
Tavris describes a weird tradition among some African tribes, and explores possible explanations for it:
The Dinka and Nuer tribes of the Sudan have a curious tradition. They extract the permanent front teeth of their children—as many as six bottom teeth and two top teeth—which produces a sunken chin, a collapsed lower lip, and speech impediments. This practice apparently began during a period when tetanus (lockjaw, which causes the jaws to clench together) was widespread. Villagers began pulling out their front teeth and those of their children to make it possible to drink liquids through the gap. The lockjaw epidemic is long past, yet the Dinka and Nuer are still pulling out their children's front teeth. How come?
The answer is “cognitive dissonance, the hardwired psychological mechanism that creates self-justification and protects our certainties, self-esteem, and tribal affiliations.” As Tavris explains:
Once we make a decision, we have all kinds of tools at our disposal to bolster it … The more costly a decision, in terms of time, money, effort, or inconvenience, and the more irrevocable its consequences, the greater the dissonance and the greater the need to reduce it by overemphasizing the good things about the choice made. Therefore, when you are about to make a big purchase or an important decision—which car or computer to buy, whether to undergo plastic surgery, or whether to sign up for a costly self-help program—don't ask someone who has just done it. That person will be highly motivated to convince you that it is the right thing to do. Ask people who have spent twelve years and $50,000 on a particular therapy if it helped, and most will say, "Dr. Weltschmerz is wonderful! I would never have found true love [got a new job] [lost weight] if it hadn't been for him." After all that time and money, they aren't likely to say, "Yeah, I saw Dr. Weltschmerz for twelve years, and boy, was it ever a waste." If you want advice on what product to buy, ask someone who is still gathering information and is still open-minded. And if you want to know whether a program will help you, don't rely on testimonials: Get the data from controlled experiments … Self-justification is complicated enough when it follows our conscious choices; at least we know we can expect it. But it also occurs in the aftermath of things we do for unconscious reasons, when we haven't a clue about why we hold some belief or cling to some custom but are too proud to admit it. For example, in the introduction we described the custom of the Dinka and Nuer tribes of the Sudan, who extract several of the permanent front teeth of their children—a painful procedure, done with a fish hook. Anthropologists suggest that this tradition originated during an epidemic of lockjaw; missing front teeth would enable sufferers to get some nourishment. But if that were the reason, why in the world would the villagers continue this custom once the danger had passed? … A practice that makes no sense at all to outsiders makes perfect sense when seen through the lens of dissonance theory. During the epidemic, the villagers would have begun extracting the front teeth of all their children, so that if any later contracted tetanus, the adults would be able to feed them. But this is a painful thing to do to children, especially since only some would become afflicted. To further justify their actions, to themselves and their children, the villagers would need to bolster the decision by adding benefits to the procedure after the fact. For example, they might convince themselves that pulling teeth has aesthetic value—say, that sunken-chin look is really quite attractive—and they might even turn the surgical ordeal into a rite of passage into adulthood. And, indeed, that is just what happened. "The toothless look is beautiful," the villagers say. "People who have all their teeth are ugly: They look like cannibals who would eat a person. A full set of teeth makes a man look like a donkey." The toothless look has other aesthetic advantages: "We like the hissing sound it creates when we speak." And adults reassure frightened children by saying, "This ritual is a sign of maturity." The original medical justification for the practice is long gone. The psychological self-justification remains.
Tavris write on how humans’ natural aggression can actually make people feel more justified in their being wrong:
Throw that doll, hit a punching bag, shout at your spouse; you'll feel better afterward. Actually, decades of experimental research have found exactly the opposite: that when people vent their feelings aggressively they often feel worse, pump up their blood pressure, and make themselves even angrier. Venting is especially likely to backfire if a person commits an aggressive act against another person directly, which is exactly what cognitive dissonance theory would predict. When you do anything that harms someone else—get them in trouble, verbally abuse them, or punch them out—a powerful new factor comes into play: the need to justify what you did. Take a boy who goes along with a group of his fellow seventh graders who are taunting and bullying a weaker kid who did them no harm. The boy likes being part of the gang but his heart really isn't in the bullying. Later, he feels some dissonance about what he did. "How can a decent kid like me," he wonders, "have done such a cruel thing to a nice, innocent little kid like him?" To reduce dissonance, he will try to convince himself that the victim is neither nice nor innocent: "He is such a nerd and crybaby. Besides, he would have done the same to me if he had the chance." Once the boy starts down the path of blaming the victim, he becomes more likely to beat up on the victim with even greater ferocity the next chance he gets. Justifying his first hurtful act sets the stage for more aggression. That's why the catharsis hypothesis is wrong. The first experiment that demonstrated this actually came as a complete surprise to the investigator. Michael Kahn, then a graduate student in clinical psychology at Harvard, designed an ingenious experiment that he was sure would demonstrate the benefits of catharsis. Posing as a medical technician, Kahn took polygraph and blood pressure measurements from college students, one at a time, allegedly as part of a medical experiment. As he was taking these measurements, Kahn feigned annoyance and made some insulting remarks to the students (having to do with their mothers). The students got angry; their blood pressure soared. In the experimental condition, the students were allowed to vent their anger by informing Kahn's supervisor of his insults; thus, they believed they were getting him into big trouble. In the control condition, the students did not get a chance to express their anger. Kahn, a good Freudian, was astonished by the results: Catharsis was a total flop. The people who were allowed to express their anger about Kahn felt far greater animosity toward him than did those who were not given that opportunity. In addition, expressing their anger increased their already heightened blood pressure; the high blood pressure of those who were not allowed to express their anger soon returned to normal. Seeking an explanation for this unexpected pattern, Kahn discovered dissonance theory, which was just getting attention at the time, and realized it could beautifully account for his results. Because the students thought they had gotten him into serious trouble, they had to justify their action by convincing themselves that he deserved it, thus increasing their anger against him—and their blood pressure … [A] vicious circle is created: Aggression begets self-justification, which begets more aggression … Fortunately, dissonance theory also shows us how a person's generous actions can create a spiral of benevolence and compassion, a "virtuous circle." When people do a good deed, particularly when they do it on a whim or by chance, they will come to see the beneficiary of their generosity in a warmer light. Their cognition that they went out of their way to do a favor for this person is dissonant with any negative feelings they might have had about him. In effect, after doing the favor, they ask themselves: "Why would I do something nice for a jerk? Therefore, he's not as big a jerk as I thought he was—as a matter of fact, he is a pretty nice guy who deserves a break."
As usual, Benjamin Franklin led the way in recognizing this positive psychological dynamic:
Although scientific research on the virtuous circle is new, the general idea may have been discovered in the eighteenth century by Benjamin Franklin, a serious student of human nature as well as science and politics. While serving in the Pennsylvania legislature, Franklin was disturbed by the opposition and animosity of a fellow legislator. So he set out to win him over. He didn't do it, he wrote, by "paying any servile respect to him"—that is, by doing the other man a favor—but by inducing his target to do a favor for him— loaning him a rare book from his library: He sent it immediately and I returned it in about a week with another note, expressing strongly my sense of the favor. When we next met in the House, he spoke to me (which he had never done before), and with great civility; and he ever after manifested a readiness to serve me on all occasions, so that we became great friends, and our friendship continued to his death. This is another instance of the truth of an old maxim I had learned, which says, "He that has once done you a kindness will be more ready to do you another than he whom you yourself have obliged."
Tavris also explores how people have a tendency to explain away their investment-backed errors:
Because most people have a reasonably positive self-concept, believing themselves to be competent, moral, and smart, their efforts at reducing dissonance will be designed to preserve their positive self-images … [A]s political scientist Philip Tetlock shows in his book Expert Political Judgment: How Good Is It? How Can We Know?, even professional "experts" who are in the business of economic and political forecasting are usually no more accurate than us untrained folks … Hundreds of studies have shown that predictions based on an expert's "personal experience" or "years of training" are rarely better than chance, in contrast to predictions based on actuarial data. But when experts are wrong, the centerpiece of their professional identity is threatened. Therefore, as dissonance theory would predict, the more self-confident and famous they are, the less likely they will be to admit mistakes … And that is just what Tetlock found. Experts reduce the dissonance caused by their failed forecasts by coming up with explanations of why they would have been right "if only"—if only that improbable calamity had not intervened; if only the timing of events had been different; if only blah blah blah.
Nobel Prize-winning psychologist Daniel Kahneman, in his book Thinking, Fast and Slow, also uses Tetlock’s studies to elaborate on the issue:
Philip Tetlock, a psychologist at the University of Pennsylvania, explored these so-called expert predictions in a landmark twenty-year study, which he published in his 2005 book Expert Political Judgment: How Good Is It? How Can We Know? Tetlock has set the terms for any future discussion of this topic. Tetlock interviewed 284 people who made their living “commenting or offering advice on political and economic trends.” He asked them to assess the probabilities that certain events would occur in the not too distant future, both in areas of the world in which they specialized and in regions about which they had less knowledge. Would Gorbachev be ousted in a coup? Would the United States go to war in the Persian Gulf? Which country would become the next big emerging market? In all, Tetlock gathered more than 80,000 predictions. He also asked the experts how they reached their conclusions, how they reacted when proved wrong, and how they evaluated evidence that did not support their positions. Respondents were asked to rate the probabilities of three alternative outcomes in every case: the persistence of the status quo, more of something such as political freedom or economic growth, or less of that thing. The results were devastating. The experts performed worse than they would have if they had simply assigned equal probabilities to each of the three potential outcomes. In other words, people who spend their time, and earn their living, studying a particular topic produce poorer predictions than dart-throwing monkeys who would have distributed their choices evenly over the options. Even in the region they knew best, experts were not significantly better than nonspecialists. Those who know more forecast very slightly better than those who know less. But those with the most knowledge are often less reliable. The reason is that the person who acquires more knowledge develops an enhanced illusion of her skill and becomes unrealistically overconfident. “We reach the point of diminishing marginal predictive returns for knowledge disconcertingly quickly,” Tetlock writes. “In this age of academic hyperspecialization, there is no reason for supposing that contributors to top journals—distinguished political scientists, area study specialists, economists, and so on—are any better than journalists or attentive readers of The New York Times in ‘reading’ emerging situations.” Tetlock also found that experts resisted admitting that they had been wrong, and when they were compelled to admit error, they had a large collection of excuses: they had been wrong only in their timing, an unforeseeable event had intervened, or they had been wrong but for the right reasons. Experts are just human in the end. They are dazzled by their own brilliance and hate to be wrong. Experts are led astray not by what they believe, but by how they think, says Tetlock. He uses the terminology from Isaiah Berlin’s essay on Tolstoy, “The Hedgehog and the Fox.” Hedgehogs “know one big thing” and have a theory about the world; they account for particular events within a coherent framework, bristle with impatience toward those who don’t see things their way, and are confident in their forecasts. They are also especially reluctant to admit error. For hedgehogs, a failed prediction is almost always “off only on timing” or “very nearly right.” They are opinionated and clear, which is exactly what television producers love to see on programs. Two hedgehogs on different sides of an issue, each attacking the idiotic ideas of the adversary, make for a good show. Foxes, by contrast, are complex thinkers. They don’t believe that one big thing drives the march of history (for example, they are unlikely to accept the view that Ronald Reagan single-handedly ended the cold war by standing tall against the Soviet Union). Instead the foxes recognize that reality emerges from the interactions of many different agents and forces, including blind luck, often producing large and unpredictable outcomes. It was the foxes who scored best in Tetlock’s study, although their performance was still very poor. But they are less likely than hedgehogs to be invited to participate in television debates.
Kahneman also explores the evidence regarding the poor predictive results of personal evaluations by “experts”:
I never met [Paul] Meehl, but he was one of my heroes from the time I read his Clinical vs. Statistical Prediction: A Theoretical Analysis and a Review of the Evidence. In the slim volume that he later called “my disturbing little book,” Meehl reviewed the results of 20 studies that had analyzed whether clinical predictions based on the subjective impressions of trained professionals were more accurate than statistical predictions made by combining a few scores or ratings according to a rule. In a typical study, trained counselors predicted the grades of freshmen at the end of the school year. The counselors interviewed each student for forty-five minutes. They also had access to high school grades, several aptitude tests, and a four-page personal statement. The statistical algorithm used only a fraction of this information: high school grades and one aptitude test. Nevertheless, the formula was more accurate than 11 of the 14 counselors. Meehl reported generally similar results across a variety of other forecast outcomes, including violations of parole, success in pilot training, and criminal recidivism. Not surprisingly, Meehl’s book provoked shock and disbelief among clinical psychologists, and the controversy it started has engendered a stream of research that is still flowing today, more than fifty years after its publication. The number of studies reporting comparisons of clinical and statistical predictions has increased to roughly two hundred, but the score in the contest between algorithms and humans has not changed. About 60% of the studies have shown significantly better accuracy for the algorithms. The other comparisons scored a draw in accuracy, but a tie is tantamount to a win for the statistical rules, which are normally much less expensive to use than expert judgment. No exception has been convincingly documented. As Meehl pointed out with justified pride thirty years after the publication of his book, “There is no controversy in social science which shows such a large body of qualitatively diverse studies coming out so uniformly in the same direction as this one.”
The predictions of algorithms using raw data are often much better than the subjective judgment of “experts.” As Kahneman writes:
The research suggests a surprising conclusion: to maximize predictive accuracy, final decisions should be left to formulas, especially in low-validity environments. In admission decisions for medical schools, for example, the final determination is often made by the faculty members who interview the candidate. The evidence is fragmentary, but there are solid grounds for a conjecture: conducting an interview is likely to diminish the accuracy of a selection procedure, if the interviewers also make the final admission decisions. Because interviewers are overconfident in their intuitions, they will assign too much weight to their personal impressions and too little weight to other sources of information, lowering validity. Similarly, the experts who evaluate the quality of immature wine to predict its future have a source of information that almost certainly makes things worse rather than better. A classic application of this approach is a simple algorithm that has saved the lives of hundreds of thousands of infants. Obstetricians had always known that an infant who is not breathing normally within a few minutes of birth is at high risk of brain damage or death. Until the anesthesiologist Virginia Apgar intervened in 1953, physicians and midwives used their clinical judgment to determine whether a baby was in distress. Different practitioners focused on different cues. Some watched for breathing problems while others monitored how soon the baby cried. Without a standardized procedure, danger signs were often missed, and many newborn infants died. One day over breakfast, a medical resident asked how Dr. Apgar would make a systematic assessment of a newborn. “That’s easy,” she replied. “You would do it like this.” Apgar jotted down five variables (heart rate, respiration, reflex, muscle tone, and color) and three scores (0, 1, or 2, depending on the robustness of each sign). Realizing that she might have made a breakthrough that any delivery room could implement, Apgar began rating infants by this rule one minute after they were born. A baby with a total score of 8 or above was likely to be pink, squirming, crying, grimacing, with a pulse of 100 or more—in good shape. A baby with a score of 4 or below was probably bluish, flaccid, passive, with a slow or weak pulse—in need of immediate intervention. Applying Apgar’s score, the staff in delivery rooms finally had consistent standards for determining which babies were in trouble, and the formula is credited for an important contribution to reducing infant mortality. The Apgar test is still used every day in every delivery room. Atul Gawande’s recent A Checklist Manifesto provides many other examples of the virtues of checklists and simple rules.
Studies by government and other “experts” recommending this or that policy or program routinely rely on what overwhelmingly turn out the be unrealistic best-case scenarios, making for terrible predictions on the public benefits of proposed public projects. As Kahneman writes:
Amos [Tversky] and I coined the term planning fallacy to describe plans and forecasts that are unrealistically close to best-case scenarios could be improved by consulting the statistics of similar cases. Examples of the planning fallacy abound in the experiences of individuals, governments, and businesses. The list of horror stories is endless. A 2005 study examined rail projects undertaken worldwide between 1969 and 1998. In more than 90% of the cases, the number of passengers projected to use the system was overestimated. Even though these passenger shortfalls were widely publicized, forecasts did not improve over those thirty years; on average, planners overestimated how many people would use the new rail projects by 106%, and the average cost overrun was 45%. As more evidence accumulated, the experts did not become more reliant on it. In 2002, a survey of American homeowners who had remodeled their kitchens found that, on average, they had expected the job to cost $18,658; in fact, they ended up paying an average of $38,769. The optimism of planners and decision makers is not the only cause of overruns. Contractors of kitchen renovations and of weapon systems readily admit (though not to their clients) that they routinely make most of their profit on additions to the original plan. The failures of forecasting in these cases reflect the customers’ inability to imagine how much their wishes will escalate over time. They end up paying much more than they would if they had made a realistic plan and stuck to it. Errors in the initial budget are not always innocent. The authors of unrealistic plans are often driven by the desire to get the plan approved—whether by their superiors or by a client—supported by the knowledge that projects are rarely abandoned unfinished merely because of overruns in costs or completion times. In such cases, the greatest responsibility for avoiding the planning fallacy lies with the decision makers who approve the plan. If they do not recognize the need for an outside view, they commit a planning fallacy.
Even when experts are so often wrong, their status as “experts” can lead people to take the advice they give, even when they would otherwise understand such advice to be significantly flawed. Tavris describes the famous “Milgram experiment”:
3,000 people … took part in the famous experiment created by social psychologist Stanley Milgram. In Milgram's original version, two-thirds of the participants administered what they thought were life-threatening levels of electric shock to another person, simply because the experimenter kept saying, "The experiment requires that you continue." This experiment is almost always described as a study of obedience to authority. Indeed it is. But it is more than that: It is also a demonstration of long-term results of self-justification. Imagine that a distinguished-looking man in a white lab coat walks up to you and offers you twenty dollars to participate in a scientific experiment. He says, "I want you to inflict 500 volts of incredibly painful shock to another person to help us understand the role of punishment in learning." Chances are you would refuse; the money isn't worth it to harm another person, even for science. Of course, a few people would do it for twenty bucks and some would not do it for twenty thousand, but most would tell the scientist where he could stick his money. Now suppose the scientist lures you along more gradually. Suppose he offers you twenty dollars to administer a minuscule amount of shock, say 10 volts, to a fellow in the adjoining room, to see if this zap will improve the man's ability to learn. The experimenter even tries the 10 volts on you, and you can barely feel it. So you agree. It's harmless and the study seems pretty interesting. (Besides, you've always wanted to know whether spanking your kids will get them to shape up.) You go along for the moment, and now the experimenter tells you that if the learner gets the wrong answer, you must move to the next toggle switch, which delivers a shock of 20 volts. Again, it's a small and harmless jolt. Because you just gave the learner 10, you see no reason why you shouldn't give him 20. And because you just gave him 20, you say to yourself, 30 isn't much more than 20, so I'll go to 30. He makes another mistake, and the scientist says, "Please administer the next level—40 volts." Where do you draw the line? When do you decide enough is enough? Will you keep going to 450 volts, or even beyond that, to a switch marked XXX DANGER? When people are asked in advance how far they imagine they would go, almost no one says they would go to 450. But when they are actually in the situation, two-thirds of them go all the way to the maximum level they believe is dangerous. They do this by justifying each step as they went along: This small shock doesn't hurt; 20 isn't much worse than 10; if I've given 20, why not 30? As they justified each step, they committed themselves further. By the time people were administering what they believed were strong shocks, most found it difficult to justify a sudden decision to quit. Participants who resisted early in the study, questioning the very validity of the procedure, were less likely to become entrapped by it and more likely to walk out. The Milgram experiment shows us how ordinary people can end up doing immoral and harmful things through a chain reaction of behavior and subsequent self-justification. When we, as observers, look at them in puzzlement or dismay, we fail to realize that we are often looking at the end of a long, slow process down that pyramid.
Finally, Tavris explains our brain’s tendency toward “selective memory,” which also feeds our biases:
The metaphors of memory fit our times and technology. Centuries ago, philosophers compared memory to a soft wax tablet that would preserve anything imprinted on it. With the advent of the printing press, people began to think of memory as a library that stores events and facts for later retrieval. (Those of us of a certain age still think of it that way, muttering about where we "filed" information in our cluttered mental cabinets.) With the inventions of movies and tape recorders, people started thinking of memory as a video camera, clicking on at the moment of birth and automatically recording every moment thereafter. Nowadays we think of memory in computer terms, and although some of us wish for more RAM, we assume that just about everything that happens to us is "saved." Your brain might not choose to screen all those memories, but they are in there, just waiting for you to retrieve them, bring out the popcorn, and watch. These metaphors of memory are popular, reassuring, and wrong. Memories are not buried somewhere in the brain, as if they were bones at an archeological site; nor can we uproot them, as if they were radishes; nor, when they are dug up, are they perfectly preserved. We do not remember everything that happens to us; we select only highlights. (If we didn't forget, our minds could not work efficiently, because they would be cluttered with mental junk—the temperature last Wednesday, a boring conversation on the bus, every phone number we ever dialed.) Moreover, recovering a memory is not at all like retrieving a file or replaying a tape; it is like watching a few unconnected frames of a film and then figuring out what the rest of the scene must have been like. We may reproduce poetry, jokes, and other kinds of information by rote, but when we remember complex information we shape it to fit it into a story line … By far, the most important distortions and confabulations of memory are those that serve to justify and explain our own lives. The mind, sense-making organ that it is, does not interpret our experiences as if they were shattered shards of glass; it assembles them into a mosaic. From the distance of years, we see the mosaic's pattern. It seems tangible, unchangeable; we can't imagine how we could reconfigure those pieces into another design. But it is a result of years of telling our story, shaping it into a life narrative that is complete with heroes and villains, an account of how we came to be the way we are. Because that narrative is the way we understand the world and our place in it, it is bigger than the sum of its parts.
In the next essay in this series, we’ll explore the psychological insights of Nobel Prize-winner Daniel Kahneman.