Insights into Psychology

What Is… a Cognitive Bias

Examples of cognitive biases

In this series, I dig a little deeper into the meaning of psychology-related terms.  This week’s term is cognitive bias.

Do you think you’re biased?  Well, your brain certainly is.  A cognitive bias is a type of shortcut your brain takes to make tasks easier and more automatic.  Sometimes that’s helpful, but often it’s not, especially when we’re unaware of it.

Types of Cognitive Biases

Ambiguity effect

We’re naturally inclined to go for the “sure thing” rather than something with an uncertain outcome; this is true even when the result would be better if the uncertain outcome occurred. This is essentially “better the devil you know than the devil you don’t.”


This involves attributing human characteristics to non-human objects, e.g. my belief that my guinea pigs love me.

Belief bias

The strength of a logical argument is evaluated based on the subjective believability of the conclusion rather than on the strength of the arguments that led to the conclusion. If I believe the earth is flat, I’m more likely to believe a logically weak argument about the movement of the moon that’s consistent with the notion of the earth being flat than I am to believe a strong argument that suggests the moon moves in a way that demonstrates the earth isn’t flat.

Ben Franklin effect

If someone has previously done you a favour, they’re more likely to do you another favour in the future than if you had originally done the favour for them. So, if you’re trying to get someone to view you more positively, ask them to do you a small favour.

Bias blind spot

We tend to believe that we are less biased than other people are, regardless of our own actual level of bias. This means that we believe that our own beliefs are more based in facts and objective reality while we view other people’s beliefs as influenced more by prejudice.

Confirmation bias

We tend to both seek out, believe, and focus on information that confirms what we already believe, and ignore information that goes against our beliefs.

Curse of knowledge

This is the assumption that other people have the same level of background knowledge as you do; I tend to get caught up in this, so I ‘ve never been very good dealing with students in my work as a nurse because I always assume they should have more knowledge than they actually do.

Dunning-Kruger effect

The Dunning-Kruger effect is a type of cognitive bias that causes people who know the least about a subject to feel confident that they have greater knowledge or competence than they actually do. This occurs when people lack the basic knowledge and meta-awareness to recognize just how much they don’t know.

Framing effect

Receiving the same information is evaluated differently depending on how it’s framed. If a news outlet is presenting an accurate piece of information in the context of other information that is negative, it will be evaluated differently than if it was framed within the context of positive information.

Gambler’s fallacy

This is the sort of belief that because you’ve had half an hour of losing at a particular slot machine, the machine is due to pay out imminently. You can also see this with a simple coin toss; if you toss a coin and get ten straight heads, the gambler’s fallacy would tell you the next toss is more likely to be tails. However, the coin doesn’t remember previous toss results; the odds stay 50/50 for each toss regardless of previous results.

Hindsight bias

An outcome seems like it should have been obviously predictable after it’s already happened. It’s very hard to un-know what you’ve learned in the interim. It can be a problem if you apply your superpowers of hindsight with the expectation that you can solve problems to come in the future.

Hyperbolic discounting

Hyperbolic discounting is the preference for a smaller short-term gain over a larger but delayed gain. For example, given the choice, people are more likely to choose being given $100 a month for a year over bring given $1500 at the end of the year, even though waiting would mean an extra $300.

Illusory truth effect

The illusory truth effect means that we’re more likely to believe information the more we’re exposed to it, even if the information itself is false. The effect of familiarity can override rationality, and this is often exploited by political campaigns.

Just world fallacy

We’re naturally inclined to think that good things happen to good people who behave properly, and bad things happen to bad people who do the wrong thing. The just world fallacy can feed into victim-blaming, as people don’t want to believe the bad things happen to people who have properly because that would mean they’re at risk.

Mere ownership effect

We tend to like things more simply because we own them, making us reluctant to give things up once we have them. This plays a role in the psychology of advertising and free trial period offers.

Negativity bias

We’re more likely to notice and remember things that are negative compared to things that are positive. This had a survival function back in the caveman days, because it was important to remember where the poisonous plants and the tigers’ den were. Even though that survival function is far less necessary now, it’s hardwired into our brains.

Observer-expectancy effect

Expecting a certain outcome causes the observer to unconsciously do things to influence the outcome. Clinical research trials are often designed to be double-blinded to account for this. This means that the neither the researchers dealing directly with the study participants nor the participants themselves know whether a given participant is receiving the treatment intervention or placebo.

Pluralistic ignorance

Pluralistic ignorance occurs when members of a group go along with something they don’t agree with because they perceive it to be the group norm, even though it’s actually not. For example, if a professor gives a lecture that didn’t make any sense, students may refrain from asking questions because they assume everyone else understands, and they don’t want to be the one person that looks like an idiot. In the meantime, everyone else is thinking the exact same thing, so the expected group norm (understanding the lecture) isn’t the norm at all.

Post-purchase rationalization

Post-purchase rationalization, also known as the choice-supportive bias, means that even if you spent far too much money on something, you’ll likely try to convince yourself afterwards that it was totally worth it. This may occur by enhancing its attributes, minimizing its flaws, or dismissing the validity of other options.

Proportionality bias

If something big and significant happens, we tend to expect there to be a big and significant reason behind it, even if the actual explanation is simple or totally random. This can help to fuel conspiracy theories, where an elaborate conspiracy may feel like it’s more appropriate than a simple explanation.

Reactive devaluation

If someone you don’t like makes a suggestion, you’ll tend to automatically assume it’s a bad suggestion, regardless of the merit of the suggestion itself. This can occur on an individual level or on a group level, such as a policy proposal from a political party you don’t support.

Rhyme as reason effect

This seems utterly bizarre, but Wikipedia gives this example from O.J. Simpson’s trial: “If the gloves don’t fit, then you must acquit.”

Sunk cost fallacy

This is the tendency to think that you should stick with something because you’ve already put a lot of time/effort/money in. For example, let’s say you bought $20,000 worth of stock, and that stock has dropped to $8,000 in value. Rather than taking your $8K and getting out of that disaster while you can, the sunk cost fallacy makes you likely to hang onto that stock as it keeps on dropping.

Unit bias

Even if you’re only hungry enough to need a small plate of food, if given a large plate, you’ll judge that full plate as being the appropriate amount.

There are many more examples of cognitive biases.Β Some of them have a greater degree of voluntariness, such as the ostrich effect (i.e. sticking one’s head in the sand to ignore something bad happening). Others we likely wouldn’t realize if they weren’t pointed out to us. Regardless, the ways that we perceive the world around us are often not an objectively accurate representation.

Source: Wikipedia: List of cognitive biases

The Psychology Corner: Insights into psychology and psychological tests

The Psychology Corner has an overview of terms covered in the What Is… series, along with a collection of scientifically validated psychological tests.

25 thoughts on “What Is… a Cognitive Bias”

  1. Wow, I could relate to so many of these!! Like, check, check, check, check, check. What a fascinating list! Oh, but I’m sure the guinea pigs love you to death!!

  2. Wow! I’ve heard of some of these but certainly didn’t realise there are so many typical patterns influencing thought, judgement and action. Great post x

Leave a Reply