Guide to the Most Common Cognitive Biases and Heuristics

Our brains have to process an infinite amount of stimuli and make an endless number of decisions. So to save time and mental energy, our brains rely on heuristics, or short-cuts. Think of heuristics like guidelines, or rules of thumb: they’re usually good enough most of the time, but they can result in errors.

Cognitive biases are systematic errors in thinking that interfere with how we reason, process information, and perceive reality. Basically, biases deviate our thinking away from objective reality and cause us to draw incorrect conclusions. 

Biases and heuristics are part of our automatic or intuitive system of thinking, so they occur without our awareness. But because they impact nearly all of our thinking and decision making, familiarity with the most common errors is a great way to become a better critical thinker.

[Before reading about individual biases and heuristics, the following post is strongly recommended: Should you trust your intuition? The elephant and rider inside your mind]

A note on how to use this post: This page is a resource of the most common cognitive biases and heuristics, and is not intended to be read from top to bottom. Feel free to share the graphics to help others learn more about how to be better thinkers.

A very brief background

Cognitive biases and heuristics were first described by Amos Tversky and Daniel Kahneman in the 1970s. While Tversky died in 1996, Kahneman won the 2002 Nobel Prize in Economics for their work, which he later summarized in his best-selling (and must-read) book, Thinking, Fast and Slow

Resources
Positive Psychology: Cognitive biases defined: 7 examples and resources
The Decision Lab: Cognitive biases
Effectivology: Cognitive biases: What they are and how the affect people
Big Think: 200 cognitive biases rule our everyday thinking. A new codex boils them down to 4. (Updated Codex)

Confirmation Bias

Confirmation bias is the tendency to search for, interpret, and remember information that confirms our beliefs. Shown is a Venn diagram with facts and evidence that we ignore, and our beliefs or what we want to be true. The overlap is what we look for and recall. Tip: Confirmation bias can give us a false sense of knowledge. Instead try to prove yourself wrong.

Definition and explanation: Confirmation bias refers to the tendency to search for, interpret, and remember information that confirms our beliefs. In short, we prefer information that tells us we’re right…and we’re more likely to remember the hits and forget the misses.

In a world full of too much information, our brains need to take short-cuts. Unfortunately, some of these short-cuts can lead us astray. In the case of confirmation bias, the short-cut is: Does this piece of information support what I already think is true? If so, we assume there’s no need to question it. 

Of all the biases, confirmation bias is the most powerful and pervasive, constantly filtering reality without our awareness to support our existing beliefs. It’s also self-reinforcing: because confirmation bias makes it seem like our beliefs are supported by evidence, we grow even more confident we’re right, and thus the more we filter and ignore information that would change our mind. 

[Learn more: The person who lies to you the most…is you]

A prime example of confirmation bias plays out in our modern media environment, where we’re able to select news organizations and even the types of stories that validate our worldview. With the help of algorithms that learn our preferences, we can get trapped in filter bubbles, or personal information ecosystems, where we’re served more and more content that reaffirms our existing beliefs and protected from evidence that we’re wrong. (We really don’t like being wrong.) In essence, we assume our news feed is telling us about reality, when the reality is it’s telling us about us.  

Confirmation bias is also one of the biggest reasons we fall for “fake news.” Why bother spending time and energy fact checking that viral video or news story or meme when it already fits with what you believe? It feels true, so it must be!

Another example of confirmation bias is the common (but mistaken) belief that the full moon impacts behavior. Indeed, the full moon is often blamed by nurses and doctors for an increase in hospital admissions and the police for an increase in crime. (It’s fun to note that the root of the words lunacy and lunatic come from the Latin luna for moon.) To be clear, there is no good evidence that the moon has any of these impacts. 

So why then does this belief persist? Imagine a teacher who believes in the full moon effect. He notices his students seem to be a little rambunctious and thinks, “It must be a full moon.” If it is a full moon, he confirms – and probably becomes even more confident in – his belief. If it’s not a full moon, he quickly interprets their behavior differently or blames it on something else. And in the future, he’s much more likely to remember the examples that supported his belief and forget the others.  

[Learn more: “Why do we still believe in ‘lunacy’ during a full moon?”]

How to overcome confirmation bias: Confirmation bias is amongst the most prevalent and influential of all the biases, so it’s important that critical thinkers try their best to reduce its impact. 

Here are a few tips:

  • Learn to recognize when you’re prone to confirmation bias, such as when a belief is tied to strong emotions and/or you’re confident you’re right. So slow down and don’t let your emotions guide your reasoning. And avoid overconfidence! The more certain you are that a belief is true, the less likely you are to question it. 
  • Be open to being wrong! The more tightly we hold our beliefs, the more contradictory evidence is viewed as a threat. But if you can’t change your mind with new evidence  you’ll never be able to learn. Instead, separate your beliefs from your identity: You aren’t wrong, the belief is.
  • Go one step further and search for evidence that would prove you wrong. In today’s information-saturated environment, if you’re looking for evidence that you’re right, you will find it. So instead, search for disconfirming evidence! If the belief is true it will withstand scrutiny.
Back to the Top

Truth bias

The truth bias is the tendency to believe information, regardless of how true it is. Shown is a person on a computer, saying "Honey, guess what? A Nigerian prince wants to send us money!" Tip: Strive for a healthy balance between trust and suspicion

AKA Truth-Default Theory

Definition and explanation: The truth bias describes our tendency to believe information, regardless of how true it is. This bias can be sneaky, as our natural inclination to believe can cause us to fall for things that just aren’t true. 

In general, it makes sense not to question everything. Most of the time people are honest, and it takes a lot more mental energy to be skeptical than it does to believe. Imagine how difficult (and time consuming) everything would be if you had to question every single thing you heard or saw! 

However, we all know that sometimes people aren’t truthful. Sometimes they stretch it a bit. Leave out important details. And sometimes they’re just plain lying. So it’s crucial to know when to be skeptical and search for confirmation.

In the past, our believing brains were only exposed to what we were told by members of our small tribes, many of whom we were related to. Today our brains swim in fake news, propaganda, hoaxes, pseudoscience, conspiracy theories, and so on. That’s a lot of information to question and think critically about… and so, too often, we don’t. As a result, our brains run a very high risk of being polluted with falsehoods.

Unfortunately, many of us like to think that we’re too smart to fall for a falsehood. We assume our beliefs were formed by following and evaluating evidence. The end result is that we can be wrong, but be confident that we’re right. 

We don’t always believe what we hear, of course, and there are many other factors that can impact our acceptance of claims. One of the most important is simply trust. The problem is, we don’t always trust the source that’s the most credible, but the one that we identify with and/or the one that appeals to our emotions. There’s a reason scammers exploit these weaknesses! 

How to overcome the truth bias: As with most mental shortcuts, awareness is key! Recognize that the truth bias is real, that we’re all prone to it, and that it can influence our judgment and decision making. 

With this in mind, one of the best things we can do is to not expose ourselves to potential falsehoods in the first place! Cleanse your media diet and follow reliable sources as much as possible. Remember, we are all susceptible to believing falsehoods, and it’s dangerous to think otherwise.  

As always, it’s essential to properly place our trust. We have a tendency to trust our in-group members, but they might not be the best source. Organizations and experts with a track record of honesty and accuracy are more likely to be trustworthy. 

The goal isn’t to be overly suspicious, but to strike a healthy balance between trust and skepticism, especially in situations where the accuracy of information is critical or there are good reasons to not trust the source.

In short: trust, but verify. While it’s generally safe (and easier) to trust others, be prepared to look for independent confirmation and supporting evidence.

Back to the Top

Dunning-Kruger Effect

The Dunning-Kruger effect is overconfidence due to limited knowledge or competence. Shown is someone who says, Don't worry, I just Googled how to perform heart surgery! And a man with a tinfoil hat saying, birds aren't real. I did my own research! We're unaware of what we don't know. Instead: Be curious and humble.

Definition and explanation: Have you ever noticed that those who know the least are often the most confident? It has a name: the Dunning-Kruger effect.

The Dunning-Kruger effect is a cognitive bias in which people overestimate their knowledge or abilities, resulting in undeserved overconfidence. Essentially, poor performers are unable to recognize their own mistakes and limits, so they assume they’re awesome. The reason is that (paradoxically) the skills and knowledge required to be competent at a task are the same skills needed to evaluate one’s own competence. Basically, ignorant people don’t know enough to recognize how ignorant they are.

But why are the incompetent so overconfident? Contrary to what many people believe, an ignorant mind is not a clean slate. Our brains are constantly trying to make sense of the world, using prior knowledge and experiences. Once we form beliefs and narratives, confirmation bias kicks in, and we unconsciously seek out information that supports our ideas. The result is a mind cluttered with misleading experiences, random facts, and intuitions that feels a lot like knowledge. And this powerful sense of false knowledge is what leads to overconfidence.

You’ve almost certainly witnessed the Dunning-Kruger effect. From dreadful American Idol contestants who can’t fathom why the judges laughed, to ignorant social media commentators LOUDLY proclaiming their opinions, it can be both hilarious and frustrating. 

That said, it’s important to keep in mind that all of us are prone to this bias. Think for a moment about something you’re really good at. It might be fixing cars, breeding Basset hounds, baking bread, or playing Call of Duty…. anything you’re an expert in. Now consider what the average person knows (or doesn’t). It’s probably not much, and some of it is probably wrong. They probably don’t even realize how much there is to know. 

Now consider that you’re that ignorant in basically every other area. I hope you’re humbled by that realization. We are all blind to our ignorance. We are all overconfident idiots. 

There are real consequences to overconfidence. Not only do the incompetent overestimate their skills, they’re unable to recognize true expertise. And if you already know everything, why would you learn? Or change your mind?

[Learn more: Overconfident Idiots: Why Incompetence Breeds Certainty]

How to overcome the Dunning-Kruger effect: At the root of the Dunning-Kruger effect is a lack of self-awareness. Essentially, we are unable to objectively evaluate our own knowledge and competence. Therefore one key solution is metacognition, or being aware of our thought processes, with the goal of assessing and improving our understanding and performance.

Also important is intellectual humility, or the recognition that you might be wrong. Be curious about what you don’t know. Ask for feedback from experts, and be open to incorporating their suggestions. Most issues are more complicated than we think, and understanding their complexity and nuance requires deep knowledge and expertise. If the answer seems simple and obvious to you, and yet somehow experts have “missed” it, consider it might be you that’s wrong.

Back to the Top

Availability Heuristic

Availability heuristic or bias is the tendency to estimate the likelihood or importance of something based on how easily we can think of an example. Shown is a circle that's all of the information, then inside is a tiny circle that represents the information that's most available to you, which tends to be recent, emotional, and vivid. A quote says: I saw a plane crash on the news. Flying is dangerous. I'll never do it again. And another example: I heard about a terrible shark attack. there's no way I'm getting into the water!

Definition and explanation: The availability heuristic is a mental short-cut in which we estimate how likely or important something is based on how easily we can think of examples. However, because we are more likely to remember events that are recent, vivid, or emotional, we overestimate the likelihood of certain events and may make poor decisions.

Consider the following examples:

  • You’re at the beach, thinking about going into the water, and images of shark attacks pop into your head. You sit and read a book instead. 
  • You recently saw a plane crash on the news, and you were already scared of flying, so you decide to drive on your next trip. 
  • You just watched a documentary about someone who won big on the slot machines, so you plan a trip to the casino. Someone has to win…it might as well be you!
  • You’re worried about someone kidnapping your child because you saw news coverage of an attempted abduction. Thankfully, the child wasn’t harmed, but you don’t want to risk it. Today’s world is so much more dangerous than it was when you were young. 

In all of these cases, you assumed something was likely because you could easily think of examples. Yet shark attacks are exceedingly rare, flying is orders of magnitude safer than driving, the chances of winning at the slots are miniscule, and there’s never been a safer time to be a kid. By confusing ease of recall with the truth, your brain misled you. And as a result you make poor decisions.

One of the biggest influences on our perception of risk is news coverage. By definition, the news covers events that are new and noteworthy, and not necessarily things that are common. News reports of murders and horrible crimes (or shark attacks and plane crashes) can result in us thinking these events are more common than they really are. 

How to overcome the availability heuristic: The first step in overcoming any heuristic is awareness. Remember, the goal is to determine how likely something is in order to make better decisions. Short-cuts help us think fast, but they aren’t always reliable. 

So slow down your thinking and don’t assume the first thing that pops into your head is representative of reality. Try to identify the stories your brain is using as evidence, and notice any emotions connected to them. Then if possible, use statistics instead! 

Back to the Top

anchoring effect

Anchoring effect or heuristic is the tendency to rely on an initial piece of information, or an anchor, when making judgements. Example is a sofa which is expensive at $600 but a deal at 40% off $1000. Tip: Awareness is key! Then slow down and give yourself time to make a decision.

Written by Jon Guy

AKA: Anchoring heuristic

Definition and explanation: The anchoring effect refers to our tendency to “anchor” to the first piece of information we learn about something, and form our beliefs about that thing based on the anchor. Newer information isn’t evaluated objectively, but, rather, through the lens of the anchor. The anchoring effect is an extremely common cognitive bias, and one that can interfere with our abilities to make good decisions and objectively understand reality. Therefore, understanding the anchoring effect can save us time, money, and improve the quality of our thinking.

The anchoring effect occurs when we unwittingly cling to the first bit of information we get about something. However, if we’re not careful, anchoring can result in poor decisions that we may  regret. For example, you discover that the new car you’d like to purchase  costs an average of $25,500 (the anchor). So you take a trip to a local dealership, and the salesperson offers to sell you the vehicle for $24,000. “What an amazing deal,” you think, as you drive off the lot in your new car. Later you learn that several other dealerships around town are selling the same vehicle for $23,000! Since you were anchored to the original $25,500, anything less sounded like a good deal, and that anchor kept you from pursuing prices at other local dealerships.

But as I mentioned earlier, anchoring can be tricky. Not only does it affect many of our decisions, it can even affect decisions that are made for us. For instance, if your doctor anchors to the first symptoms you report about an illness, she might misdiagnose you without pursuing other possible explanations. Or, if you’re waiting on a decision from a jury on an insurance settlement case, their decision on how much to award you might be influenced by a strategically placed anchor.

Most people agree that taking care of our health is one of the more important goals we pursue throughout our lives. But let’s say your grandparents and great-grandparents were all very long-lived. You might anchor to their longevity as an expectation of how long you will live, without taking into consideration that they might have lived much more healthy and active lifestyles than you do. Therefore, by anchoring to one piece of information (how long they lived) and ignoring other, more important pieces of information (how they took care of themselves), you could wind up neglecting your own health by eating poorly or exercising infrequently.

The anchoring effect is so powerful that the anchor doesn’t even have to be relevant to the thing we’re making a decision about! For example, researchers have shown that putting an expensive car on a restaurant menu actually resulted in people spending more money while dining there. Other researchers have shown that by simply asking people the last two digits of their social security number and then showing them a list of products, those whose last two digits were higher were willing to pay more for the products than those whose digits were lower.

How to overcome the anchoring effect: The anchoring effect is an extremely pervasive bias, and even contributes to other biases. Moreover, since anchoring happens outside of our conscious awareness, interrupting the process can be rather challenging.  Therefore, it is important to understand the effects of anchoring, so that we might stand a chance of overcoming them.

For example, thinking long and hard about an important decision always sounds like a good idea, right? Intuitively, this makes sense. However, if we’re merely thinking deeply about the anchor, we’re just amplifying its effects, and probably digging ourselves even deeper into our biases.

Fortunately, there are some strategies we can use to combat, if not completely overcome, the anchoring effect, such as by practicing metacognition. We are cognitive misers, which means overcoming our biases requires us to maximize metacognition; an awareness and understanding of our own thought processes. Or as I like to call it, thinking about thinking.

Another strategy we can employ is to try to consider alternative options. If you see t-shirts on sale for 3 for $10, consider that you may only need one, or that you might not need any! Anything you can do to interrupt the decision-making process can help  to slow down your thinking and give you the time you need to make a better decision.

Back to the Top

Representativeness heuristic

Representativeness heuristic is the tendency to estimate the likelihood of something based on how similar it is to existing mental categories. Shown is a man in a hospital bed talking to a Black female doctor. He says, "I'd like to talk to the doctor." She responds, "I'm the doctor." And he says, "You don't look like a doctor!" Understanding cognitive errors can help us avoid harmful stereotypes.

Written by Jon Guy

Definition and explanation: The representativeness heuristic doesn’t exactly flow easily off the tongue. Nonetheless, this heuristic is well worth our attention. Like all heuristics, the representativeness heuristic is a mental shortcut our brains take to preserve its limited resources, in this case to make quick judgments about the likelihood of something based on how similar it is to existing mental categories.

The representativeness heuristic is an error of reasoning that occurs when we make generalizations based on our mental models of reality. To use a classic example, let’s say I told you that Mary is a quiet, shy introvert who’s not very interested in getting to know people and is also very detail oriented. Based on this description, is it more likely that Mary is a librarian or a mechanic? Our gut tells us that Mary is much more likely to be a librarian because her characteristics sound more representative of our mental models of librarians.

But the representativeness heuristic doesn’t just apply to our perception of people. For example, if you saw a frog with bright colors, you might assume that it’s poisonous based on your mental model of what poisonous frogs look like. Or, imagine shopping for a new phone case. If you find one that looks thick and durable, and you assume that it will provide good protection because of your prior mental model about what durability looks like.

The reason we might assume the frog is poisonous is because most of us have a mental construct of what poisonous animals look like, and a colorful frog fits the bill. Likewise, our mental constructs about what constitutes durability might lead us to believe that a thick case means that it’s durable. In either situation, our default assumption is based on our prior mental models, and questioning these models requires deeper thinking. Without further evidence, we cannot know whether the frog is poisonous or the phone case is durable.

Perhaps the best example of the representativeness heuristic is stereotyping. Stereotyping occurs when we have a mental model of a specific class of people, and we make judgments about all people within that class based on our mental construct. Just as we assumed that “colorful amphibian” means “poisonous animal,” we make similar assumptions about classes of people. For instance, we might assume that all Indians enjoy spicy foods, or that all Native Americans are highly spiritual, or that all rich people are snobs, or that all homeless people are drug addicts and alcoholics. However, in each of these incidents, our mental models will be wrong much of the time, and so it is up to us to think critically about how our brains misinterpret reality.

Heuristics are short-cuts our brains take to save energy. Importantly, they aren’t necessarily bad and don’t always lead us astray. Additionally, our brains categorize things (i.e. people, places, events, objects, etc) based on our previous experiences so we know what to expect and how to react when we encounter something new in that category. If we didn’t categorize things, every new thing we encountered would overwhelm us!  So, while the representativeness heuristic can cause us to misconstrue reality, it’s also a valuable tool that helps us navigate our daily lives. Rather than trying to determine if every moving car is potentially dangerous, we draw on our mental models of moving vehicles, and treat them all with caution. Likewise, if a grizzly bear was running our way, we’d know we were in trouble, but if a tumbleweed was speeding towards us, we’d know we have nothing to fear.

Even though the representativeness heuristic can be a useful tool, it’s still important for us to maintain our skepticism and practice our critical thinking skills. While categorizing things assists our brains in reducing its cognitive load, we need to remember that even our mental constructs might be wrong, and understand how easily stereotypes can become prejudices. Moreover, categorizing limits our ability to see similarities and differences between different things, and, when we pay too much attention to specific boundaries, we further limit our ability to see the whole.  

As the foregoing examples show, the representativeness heuristic can affect our judgment in both trivial and nontrivial ways. It’s not that big of a deal to incorrectly assume that the guy wearing a suit is a businessman, when he’s actually a plumber headed to a costume party. However, judging one individual based on a stereotypical example is a major driver of racism, sexism, classism, and even speciesism. Therefore, becoming better educated about our mental shortcomings goes a long way toward making the world a better place.

How to avoid the representativeness heuristic: Biases are part of human nature, and oftentimes serve valuable purposes. As such, completely avoiding the representativeness heuristic is unlikely, and undesirable. However, there are several things we can do to minimize the number of times we fall for it, as well as reduce its effects when we inevitably slip up.

Awareness of our biases is bias kryptonite, so understanding when we’re vulnerable, sharing what you’ve learned about it with others, and developing critical thinking skills are excellent ways to avoid this bias. 

Finally, education can help us combat cognitive biases. Learning formal logic, becoming a better statistical thinker, or studying other errors of human cognition are all great ways to minimize our biases, become better thinkers, and perceive reality on reality’s terms.

Back to the Top

Base Rate Fallacy

Base rate fallacy, or base rate neglect: Makes assumptions about probability based on incomplete information rather than overall frequency. Shown is aliens and a ufo landing at a basketball game, and concluding that 'humans are tall' TIP: To know how common something is, consider all relevant information

Written by Jon Guy

AKA: Base rate neglect and base rate bias

Definition and explanation: The base rate fallacy occurs when we draw conclusions about probability from limited information, rather than from statistical data. Oftentimes, thinking statistically is a difficult undertaking, and it isn’t something we do intuitively. But to make wise decisions, sometimes we need to understand how likely something is, so let’s explore what the base rate fallacy is and discuss some strategies for avoiding it. 

A base rate is the likelihood that an event will occur, or the proportion of individuals in a population that share certain characteristics. Unfortunately, instead of judging probabilities by looking at base rates, we often rely heavily on personal experiences or vivid stories. Readers might be reminded of the availability heuristic, a bias which causes us to estimate the likelihood or importance of something based on how readily similar examples come to mind. By contrast, the base rate fallacy occurs when we weigh the probability of something using only a small sample size, without considering all of the relevant information.

For example, if everyone you spoke with in your neighborhood voted for candidate X, you may assume that candidate X would win the election by a landslide. But your neighborhood—or even region—is only part of a much larger population of voters. By ignoring the base rate of all voters, or the overall popularity of candidate X, you’ve committed the base rate fallacy.

Concerningly, people often commit this error when it comes to matters of health. For example, a headline that reads “Among those who died from infection, 50% were vaccinated” makes it seem like those who are vaccinated are just as likely to die as those who aren’t. Imagine, though, that the population is 100 individuals, 99% of which are vaccinated. Two people die — the individual who wasn’t vaccinated and one of those who was. The headline is technically accurate, but wildly misleading, as the death rate among the unvaccinated is 100%. 

The base rate fallacy shows up in many aspects of our lives, and sometimes carries severe consequences. For example, doctors might order expensive or invasive tests based on a person’s symptoms, even if the base rate of the disease they’re testing for is low, which could lead to unnecessary costs and risks. Or, judges and juries may convict someone based on circumstantial or anecdotal evidence, even if prior probability of guilt is low, leading to wrongful convictions and miscarriages of justice. Or, investors may back a specific company based on recent performance, even if the likelihood of success is low, leading to poor financial decisions and financial losses. 

As the saying goes, if you hear hooves, think horses, not zebras. (Unless you’re in Africa.)

How to overcome the base rate fallacy: The base rate fallacy, like other thinking errors, occurs when we’re blind to information we don’t have, and as a result, we place greater emphasis on the information we do have. Thus, one way to avoid this fallacy is to apply the “rule of total evidence,” which requires us to consider all relevant information when forming an opinion. Essentially, we want to make visible the information we wouldn’t otherwise consider and then evaluate all of the information before forming judgments. As you can imagine, this is far easier said than done.

Applying the rule of total evidence to some of our examples from above; doctors could use diagnostic algorithms that consider a patient’s symptoms alongside the base rate of specific diseases within the population; judges and juries could receive instructions to consider the base rate of guilt when evaluating evidence and testimony; or investors could use diversification strategies to consider both the performance of specific companies along with the historical performance of the market as a whole.

As with all biases, awareness is key. Remember that we can easily be fooled by our personal experiences or emotional stories into thinking something is more common than it actually is. And by remembering to consider base rates, we can improve our thinking skills and our ability to make wiser decisions.

Back to the Top

Barnum Effect (aka Forer Effect)

Barnum effect or Forer effect is the tendency for people to believe that vague personality descriptions uniquely apply to them. Shown is a crystal ball with Barnum statements and someone saying, “how did you know?”

Definition and explanation: If you’ve ever read your horoscope, visited a psychic, or even taken a personality test and thought, “Wow, that was so accurate! How did they know?!?”, you’ve likely fallen for the Barnum effect. The Barnum effect describes the tendency for people to believe that vague personality descriptions apply uniquely to them, even though they apply to nearly everyone. 

The Barnum effect is named after the famous showman P.T. Barnum, but it was first described by psychologist Bertram Forer in 1948 (hence its other name). In an experiment on psychology’s favorite “lab rat” (i.e. undergraduates), Forer gave his students personality tests and told them he would analyze and provide each with individualized feedback. Overall, the class rated their results as very accurate (an average of 4.3 out of 5). 

The kicker, of course, was that all students received the exact same results, which included statements such as:

  • You have a great need for people to like and admire you.
  • You have a tendency to be critical of yourself.
  • Disciplined and self-controlled outside, you tend to be worrisome and insecure inside.
  • At times you have serious doubts as to whether you have made the right decision or done the right thing.

It’s not hard to understand why the students thought these statements applied to them…they apply to nearly everyone. But the students interpreted the generic statements as applying specifically to them. Thus, they fell for the Barnum effect.

To understand why, let’s take a closer at the statements that produce the Barnum effect. Barnum statements are general and applicable to nearly everyone. But their vagueness is their “strength,” as individuals each interpret them with their own meaning. They’re also mostly positive, as we prefer being flattered to hearing negative things about ourselves. And they often include qualifiers, like “at times”, or simultaneously attribute opposing characteristics, so that they’re almost never wrong.

Forer rightly attributed the results of his experiment to his students’ gullibility. But the truth is, we’re all gullible to some degree, which is why the Barnum effect is frequently exploited by those seeking to convince us they have deep insight into our personal psychology. 

How to avoid falling for the Barnum effect: Your best line of defense against the Barnum effect is awareness and skepticism. 

  • Be on guard for situations where vague information might be giving the impression that results are specifically tailored to you, such as fortune tellers, horoscopes, psychics, personality tests (e.g. MBTI), online quizzes, and even Netflix watchlists.
  • Remember to insist on sufficient evidence to accept any claim. Ask yourself: Could this “personalized” feedback apply to others? Am I falling for flattery? And importantly, do I want to believe? Skepticism is the best way to protect yourself against being fooled…but no one can fool us like we can. 

Final note: While many people associate P.T. Barnum with the saying “there’s a sucker born every minute,” there’s actually no good evidence he said those words. For a phrase that describes our tendency to be gullible, isn’t it ironic?

[Learn more: Learn to be a psychic with these 7 tricks]

Back to the Top

SURVIVORSHIP BIAS

Survivorship Bias: Tendency to draw conclusions from "surviving" examples and overlook those that didn't Tip: It's important to include ALL data, as there's much co learn from "failures." Shown is a pyramid with "Failures we don't see". At the very tip there are examples: "A guy on IG lost weight on a diet, so it must work!" "They don't make music like they used to!" "A 100 year old says the secret is smoking a pack of cigarettes a day!"

Written by Jon Guy

Definition and explanation: The survivorship bias is a cognitive error that occurs when we focus on the existing elements of something (i.e., the ones that “survived”), while ignoring those that didn’t endure. This bias arises when we form conclusions based solely on successes without also considering  failures.  

Whether we’re aware of it or not, the survivorship bias pops up in our everyday lives. A news headline champions the success of some celebrity-endorsed diet while failing to mention the numerous people who failed to achieve the same results. Or a television program suggests that the “surviving” Egyptian pyramids couldn’t have been built with such craftsmanship and precision, without mentioning the many failed attempts that have since crumbled.

But we can be fooled by only considering “winners.” For example, we often hear about the inspiring successes of maverick entrepreneurs such as Mark Zuckerberg or Bill Gates, but by overlooking those who failed we could end up with a distorted understanding of the challenges they faced on their journeys. 

Many of us witnessed survivorship bias throughout the Covid-19 pandemic. Some-expressing anti-vaccination sentiments-claimed “no one regrets not taking the vaccine.” Many examples of people who suffered or caused others to suffer as a result of not being vaccinated refute this claim. However, even if it were true, those who died as a result of not being vaccinated cannot object to the claim. As another example, some argue that SARS-CoV-2 isn’t that dangerous because they (and those they know) survived without serious complications, neglecting those who weren’t so fortunate.

Survivorship bias can also influence our personal growth and decision-making. For instance, when we seek advice from successful individuals, we might overlook the experiences of those who faced serious setbacks and failures. If we only consider the opinions of those who’ve succeeded, we might develop unrealistic expectations and face discouragement when we encounter obstacles on our own paths to success. A more balanced perspective—one that includes examples of both successes and failures—equips us to navigate obstacles and pursue success with resilience and flexibility. 

How to overcome the survivorship bias: While we’re all prone to falling for the survivorship bias, there are several strategies we can use to mitigate it. First, be aware that we are all prone to this thinking error. Second, don’t be fooled by only looking at the winners, and remember that failure is often forgotten. Making informed decisions requires us to consider all data, as the challenges, mistakes, and failures of those who didn’t survive have much to teach us. 

We can also keep in mind that randomness and luck often play a role in success and failure, and sometimes it really is just a matter of chance. And we can make a concerted effort to seek out those invisible data that are lurking beyond our awareness.

Back to the Top

Sunk cost fallacy

Sunk cost fallacy is the tendency to stick with a decision we've invested time, money, or energy into, even if the current costs outweigh the benefits. Shown is a man gambling and he says, "I've already lost a lot of money, so I might as well keep playing." And a woman watching a movie who says, "This movie is terrible, but I might as well finish it. Ask: If I had to make this decision all over, starting from where I am now, what would I choose?

Written by Jon Guy

Other names: Escalation of commitment heuristic

Definition and explanation: The sunk cost fallacy describes our tendency to stick with a decision we’ve invested resources into, regardless of the increasing negative consequences of that decision. If you’ve ever wondered why you continue watching a boring movie, or why you proceed through a terribly uninteresting book, it might be because of the sunk cost fallacy. The better we understand the sunk cost fallacy, the more likely we are to catch ourselves committing it and thus the better positioned we are to prevent ourselves from continuing on with a losing course of action.

For many of us, once we’ve invested into something, it can be really difficult to cut our losses, and instead, we end up “throwing good money after bad.” For instance, if you’ve ever tried to bluff your way through a hand of poker, you know just how hard it can be to fold once you’re “pot committed.” However, instead of folding, you up the ante, hoping that an even bigger bluff will prevent an impending loss. But in life, the stakes are much higher than they are in poker.

If the consequences of our actions have a negative impact, the rational thing to do is stop engaging in that action. The problem is, when we’ve spent time, money, or effort pursuing a decision, rather than rationally looking at the negative consequences of continuing to pursue that decision, we tend to look at it from the perspective of “losing” that time, money, or effort. Of course, that time, money, or effort is already gone (they are, after all, sunk costs), and continuing with our commitment is unlikely to bring any of it back. Typically, we commit the sunk cost fallacy because we don’t want to “lose” our investment, even if we wind up more damaged than if we had simply cut our losses. Put another way, it makes no rational sense to continue losing simply because you’ve already lost, especially since you can’t  protect an investment you can’t get back anyway.

The sunk cost fallacy can dramatically affect us as individuals, and can hugely impact the course of our lives. For example, we might stay in a relationship that isn’t good for us, or one we no longer want to be in, because of how much we’ve invested into the relationship. We might remain at a job we can’t stand, simply because of how much work we’ve done for the company. We might continue spending money fixing a run-down car, when it would clearly be cheaper to buy a new one. Or we might remain in the same social circle because of how much history we have with the people in it, even though we no longer share the same interests or values.

Additionally, the sunk cost fallacy has a number of systemic effects. Oftentimes, government projects that don’t meet expectations continue receiving funding simply because so much has already been invested into them. Both Lyndon B. Johnson and George W. Bush continued sending soldiers into losing wars, and spent billions of dollars on what they both knew were hopeless causes. Thus, when it comes to fallacious thinking, sometimes lives are literally at stake.

Another pressing danger of the sunk cost fallacy is that it leads to the snowball effect, wherein we try to protect our investments by investing even more. The more we invest into something, the more we feel like we need to commit to it, which means we’ll likely expend more resources pursuing the endeavor. 

How to overcome the sunk cost fallacy: Sometimes, when we find ourselves in a hole, the best thing we can do is stop digging. However, in order to do so, we must first recognize that we’re in a hole. Therefore, awareness is our first line of defense against the sunk cost fallacy. If you find yourself focused on prior losses rather than current or future costs, you can use this awareness as a signal to don your skeptical spectacles and think critically about your situation.

In addition, we should also carefully consider whether the investment we’re trying to protect is actually recoverable, or if it truly is just a sunken cost. Therefore, it’s good practice to ask yourself, “If there’s no way to get back what I’ve already put into something, am I better off investing further, or is cutting my losses the better option?”

Lastly, we should try to recognize the difference between making a decision that’s based on our emotions, and making a decision that’s based on reason. Emotions are both powerful and important drivers of our decision-making, and, like it or not, we are emotionally tied to the investments we make. Therefore, since decisions that are based on emotions are not always in our best interest, thinking critically about our investments becomes especially important.

Back to the Top

7 thoughts on “Guide to the Most Common Cognitive Biases and Heuristics”

  1. Pingback: Postanite vidoviti uz sedam trikova - FakeNews Tragač

  2. Pingback: Tipps zur Faktenüberprüfung: 6 wichtige Fragen, um herauszufinden, ob deine Überzeugung stimmt

  3. Pingback: The Pros And Cons Of Heuristics: How Mental Shortcuts Can Help Or Harm Your Decision-Making – thesharpener.net

  4. Pingback: Inappropriate Indicators And How They Can Hurt You - AlltopCash.com

  5. Pingback: 6 WICHTIGE FRAGEN ZUR FAKTENÜBERPRÜFUNG DEINER ÜBERZEUGUNG

  6. Pingback: 6 Key Questions to Find Out if Your Belief is True

  7. Pingback: Sechs wichtige Fragen zur Faktenüberprüfung deiner Überzeugung

Leave a Reply