15 Logical Fallacies You Should Know
15 Logical Fallacies You Should Know
15 Logical Fallacies You Should Know
1. Ad Hominem
2. Straw Man
3. Appeal to Ignorance
4. False Dilemma
5. Slippery Slope
6. Circular Argument
7. Hasty Generalization
8. Red Herring
9. Tu Quoque
10. Causal Fallacy
A logical fallacy is an error in reasoning common enough to warrant a fancy
name. Knowing how to spot and identify fallacies is a priceless skill. It can
save you time, money, and personal dignity. Formal fallacies are
breakdowns in how you say something, the ideas are ordered wrong
somehow. Their form is wrong.
Informal fallacies, like the ones below, have to do with what you are saying
(the “content” of an argument). The ideas might be arranged right, but
something you said isn’t quite right. The content is wrong. Here’s a list of
the 15 informal fallacies you are most likely to encounter in discussion and
debate.
1. Ad Hominem Fallacy
When people think of “arguments,” often their first thought is of shouting matches riddled
with personal attacks. Ironically, personal attacks run contrary to rational arguments. In logic
and rhetoric, personal attacks are called ad hominems. Ad hominem is Latin for “against the
man.” Instead of advancing good sound reasoning, ad hominems replace logical
argumentation with attack-language unrelated to the truth of the matter.
An ad hominem is more than just an insult. It’s an insult used as if it were an argument or
evidence in support of a conclusion. Verbally attacking people proves nothing about the truth
or falsity of their claims. Ad hominems are common known in politics as “mudslinging.”
Instead of addressing the candidate’s stance on the issues, or addressing his or her
effectiveness as a statesman or stateswoman, ad hominemsfocus on personality issues,
speech patterns, wardrobe, style, and other things that affect popularity but have no bearing
on their competence. In this way, ad hominemscan be unethical, seeking to manipulate
voters by appealing to irrelevant foibles and name-calling instead of addressing core issues.
In this last election cycle, personal attacks were volleyed freely from all sides of the political
aisle, with both Clinton and Trump facing their fair share of ad hominems.
Ad hominem is an insult used as if it were an argument or evidence in support of
a conclusion.
A thread on Quora lists the following doozies against Hillary Clinton: “Killary Clinton,”
"Crooked Hillary," "Hilla the Hun," "Shillary," "Hitlery," "Klinton," "Hildebeest," "Defender of
Child rapists," "Corporate Whore," "Mr. President," "Heil Hillary," "Wicked Witch of the West
Wing," "Robberty Hillham Clinton," "Mrs. Carpetbagger", and the decidedly unsubtle, "The
Devil.”
The NY Daily News offers an amusing list of insults against Donald Trump: “Short fingered
Vulgarian,” "Angry Creamsicle," "Fascist Carnival Barker," "F*ckface von Clownstick,"
"Decomposing Jack-O-Lantern," "Chairman of the Saddam Hussein Fanclub," "Racist
Clementine," "Sentient Caps Lock Button," "Cheeto Jesus," "Tangerine Tornado," and
perhaps the most creative/literary reference, "Rome Burning in Man Form.”
Ad hominems often signal the point at which a civil disagreement has descended into a
“fight.” Whether it’s siblings, friends, or lovers, most everyone has had a verbal
disagreement crumble into a disjointed shouting match of angry insults and accusations
aimed at discrediting the other person. When these insults crowd out a substantial argument,
they become ad hominems.
Your Turn:
See if you can tell which of these is an ad hominem argument, and which is just an insult.
Example 1: “MacDougal roots for a British football team. Clearly he’s unfit to be a police
chief in Ireland.”
2. Straw Man
It’s much easier to defeat your opponent’s argument when it’s made of straw. The Strawman
fallacy is aptly named after a harmless, lifeless, scarecrow. In the straw man fallacy,
someone attacks a position the opponent doesn’t really hold. Instead of contending with the
actual argument, he or she instead attacks the equivalent of a lifeless bundle of straw, an
easily defeated effigy, which the opponent never intended upon defending anyway.
Straw man fallacies are a cheap and easy way to make one’s position look stronger than it
is. Using this fallacy, opposing views are characterized as “non-starters,” lifeless, truthless,
and wholly unreliable. By comparison, one’s own position will look better for it. You can
imagine how straw man fallacies and ad hominems can occur together, demonizing
opponents and discrediting their views.
In the straw man fallacy, someone attacks a position the opponent doesn’t really hold.
This fallacy can be unethical if it’s done on purpose, deliberately mischaracterizing the
opponent’s position for the sake of deceiving others. But often the straw man fallacy is
accidental, because one doesn’t realize he or she is oversimplifying a nuanced position, or
misrepresenting a narrow, cautious claim as if it were broad and foolhardy.
Your Turn:
See if you can detect how both of the following statements could qualify as a strawman
fallacy.
Example 1: “The Senator thinks we can solve all our ecological problems by driving a
Prius.”
Example 2: “Quite the contrary, the Senator thinks the environment is such a wreck that
no one’s car choice or driving habits would make the slightest difference.”
Ignorance isn’t proof of anything except that one doesn’t know something.
Interestingly, this fallacy is often used to bolster multiple contradictory conclusions at once.
Consider the following two claims: “No one has ever been able to prove definitively that
extra-terrestrials exist, so they must not be real.” “No one has ever been able to prove
definitively that extra-terrestrials do not exist, so they must be real.” If the same argument
strategy can support mutually exclusive claims, then it’s not a good argument strategy.
Ignorance isn’t proof of anything except that one doesn’t know something. If no one has
proven the non-existence of ghosts or flying saucers, that’s hardly proof that those things
either exist or don’t exist. If we don’t know whether they exist, then we don’t know that
they do exist or that they don’t exist. Ignorance doesn’t prove any claim to knowledge.
Your Turn:
Consider the following examples, can you spot the appeal to ignorance?
Example 1: “We have no evidence that the Illuminati ever existed. They must have been
so clever they destroyed all the evidence.”
Example 2: “I know nothing about Tank Johnson except that he has a criminal record as
long as your leg, but I’ll bet he’s really just misunderstood.”
Dilemma-based arguments are only fallacious when, in fact, there are more than the stated
options. It’s not a fallacy however if there really are only two options. For example, “either
Led Zeppelin is the greatest band of all time, or they are not.” That’s a true dilemma, since
there really are only two options there: A or non-A. It would be fallacious however to say,
"there are only two kinds of people in the world, people who love Led Zeppelin, and people
who hate music." Some people are indifferent about that music. Some sort of like it, or sort
of dislike it, but don’t have strong feelings either way.
Dilemma-based arguments are only fallacious when, in fact, there are more than the
stated options.
The false dilemma fallacy is often a manipulative tool designed to polarize the audience,
heroicizing one side and demonizing the other. It’s common in political discourse as a way of
strong-arming the public into supporting controversial legislation or policies.
Your Turn:
See if you can identify a third option these politicians failed to mention.
5. Slippery Slope
You may have used this fallacy on your parents as a teenager: “But, you have to let me go to
the party! If I don’t go to the party, I’ll be a loser with no friends. Next thing you know I’ll end
up alone and jobless living in your basement when I’m 30!” The slippery slope fallacy works
by moving from a seemingly benign premise or starting point and working through a number
of small steps to an improbable extreme.
This fallacy is not just a long series of causes. Some causal chains are perfectly reasonable.
There could be a complicated series of causes which are all related, and we have good
reason for expecting the first cause to generate the last outcome. The slippery slope fallacy,
however, suggests that unlikely or ridiculous outcomes are likely when there’s just not
enough evidence to think so.
The slippery slope fallacy, however, suggests that unlikely or ridiculous outcomes are likely
when there’s just not enough evidence to think so.
Certain ad campaigns from Dodge, Taco Bell, and notably a recent one for Direct TV,
commit this fallacy to great comic effect.
It’s hard enough to prove one thing is happening or has happened; it’s even harder to prove
a whole series of events will happen. That’s a claim about the future, and we haven’t arrived
there yet. We, generally, don’t know the future with that kind of certainty. The slippery slope
fallacy slides right over that difficulty by assuming that chain of future events without really
proving their likelihood.
Your Turn:
Which of these examples is a slippery slope fallacy and which is not?
Example 1: “Your coach’s policy is that no one can be a starter on game day if they miss
practice. So, if you miss basketball practice today, you won’t be a starter in Friday’s
game. Then you won’t be the first freshman to start on the Varsity basketball team at our
school.”
Example 2: “If America doesn’t send weapons to the Syrian rebels, they won’t be able to
defend themselves against their warring dictator. They’ll lose their civil war, and that
dictator will oppress them, and the Soviets will consequently carve out a sphere of
influence that spreads across the entire Middle East.”
Circular arguments are also called Petitio principii meaning “Assuming the initial
[thing]“ (commonly mistranslated as "begging the question"). This fallacy is a kind of
presumptuous argument where it only appears to be an argument. It’s really just restating
one’s assumptions in a way that looks like an argument. You can recognize a circular
argument when the conclusion also appears as one of the premises in the argument.
Your Turn:
Another way to explain circular arguments is that they start where they finish, and finish
where they started. See if you can identify which of these is a circular argument.
Example 1: “Abstract art isn’t even art. Those pictures and sculptures don’t represent
anything, and that’s how you know its not even art.”
Example 2: “We should be tolerant even of people who believe intolerant ideas. Their
ideas matter too, and we can still learn different things from them even if their particular
intolerant idea is wrong.”
7. Hasty Generalization
Hasty generalizations are general statements without sufficient evidence to support them.
They are general claims too hastily made, hence they commit some sort of illicit assumption,
stereotyping, unwarranted conclusion, overstatement, or exaggeration.
Normally we generalize without any problem. We make general statements all the time: “I
like going to the park,” "Democrats disagree with Republicans,” "It’s faster to drive to work
than to walk," or "Everyone mourned the loss of Harambe, the Gorilla.”
Hasty generalization may be the most common logical fallacy because there’s no single
agreed-upon measure for “sufficient” evidence.
Indeed, the above phrase “all the time” is a generalization — we aren’t all the timemaking
these statements. We take breaks to do other things like eat, sleep, and inhale. These
general statements aren’t addressing every case every time. They are speaking generally,
and, generally speaking, they are true. Sometimes you don’t enjoy going to the park.
Sometimes Democrats and Republicans agree. Sometimes driving to work can be slower
than walking if the roads are all shut down for a Harambe procession.
Hasty generalization may be the most common logical fallacy because there’s no single
agreed-upon measure for “sufficient” evidence. Is one example enough to prove the claim
that "Apple computers are the most expensive computer brand?" What about 12 examples?
What about if 37 out of 50 apple computers were more expensive than comparable models
from other brands?
There’s no set rule for what constitutes “enough” evidence. In this case, it might be possible
to find reasonable comparison and prove that claim is true or false. But in other cases,
there’s no clear way to support the claim without resorting to guesswork. The means of
measuring evidence can change according to the kind of claim you are making, whether it’s
in philosophy, or in the sciences, or in a political debate, or in discussing house rules for
using the kitchen. A much safer claim is that "Apple computers are more expensive than
many other computer brands.”
Meanwhile, we do well to avoid treating general statements like they are anything more
than generalizations. Even if it were generally true that women are bad drivers — and I’m not
saying they are — there are still plenty of women who are good drivers. And those “cases”
just aren’t covered with that general statement even if it were true. In my case, my wife is a
better driver than I am. So I do well not to generalize too widely.
A simple way to avoid hasty generalizations is to add qualifiers like “sometimes,” "maybe,"
"often," or "it seems to be the case that . . . ". When we don’t guard against hasty
generalization, we risk stereotyping, sexism, racism, or simple incorrectness. But with the
right qualifiers, we can often make a hasty generalization into a responsible and credible
claim.
Your Turn:
Which of the following is a hasty generalization?
Example 1: “Some people vote without seriously weighing the merits of the candidate.”
Example 2: “People nowadays only vote with their emotions instead of their brains.”
Red herrings can be difficult to identify because it’s not always clear how different
topics relate.
The phrase “red herring” refers to a kippered herring (salted herring-fish) which was reddish
brown in color and quite pungent. According to legend, this aroma was so strong and
delectable to dogs that it served as a good training device for testing how well a hunting dog
could track a scent without getting distracted. Dogs aren’t generally used for hunting fish so
a red herring is a distraction from what he is supposed to be hunting.
Red herrings can be difficult to identify because it’s not always clear how different topics
relate. A “side” topic may be used in a relevant way, or in an irrelevant way. In the big meaty
disagreements of our day, there are usually a lot of layers involved, with different subtopics
weaving into them. We can guard against the red herring fallacy by clarifying how our part of
the conversation is relevant to the core topic.
Your Turn:
Which of the following examples is a red herring fallacy?
Example 1: “My wife wants to talk about cleaning out the garage, so I asked her what
she wants to do with our patio furniture? Now she’s shopping for new patio furniture and
not bothering me about the garage.”
Example 2: “My wife wants to talk about cleaning out the garage, so I asked her what
she wants to do with the patio furniture? It’s just sitting in the garage taking up space.”
9. Tu Quoque Fallacy
The “tu quoque,” Latin for “you too,” is also called the “appeal to hypocrisy” because it
distracts from the argument by pointing out hypocrisy in the opponent. This tactic doesn’t
solve the problem, or prove one’s point, because even hypocrites can tell the truth. Focusing
on the other person’s hypocrisy is a diversionary tactic. In this way, the tu quoque typically
deflects criticism away from one’s self by accusing the other person of the same problem or
something comparable. If Jack says, “Maybe I committed a little adultery, but so did you
Jason!” Jack is trying to diminish his responsibility or defend his actions by distributing blame
to other people. But no one else’s guilt excuses his own guilt. No matter who else is guilty,
Jack is still an adulterer.
The tu quoque fallacy is an attempt to divert blame, but it really only distracts from the initial
problem. To be clear, however, it isn’t a fallacy to simply point out hypocrisy where it occurs.
For example, Jack may say, “yes, I committed adultery. Jill committed adultery. Lots of us
did, but I’m still responsible for my mistakes.” In this example, Jack isn’t defending himself or
excusing his behavior. He’s admitting his part within a larger problem. The hypocrisy claim
becomes a fallacy only when the arguer uses some (apparent) hypocrisy to neutralize
criticism and distract from the issue.
Your Turn:
Which of the following is a tu quoque fallacy?
Example 1: “But, Dad, I know you smoked when you were my age, so how can you tell
me not to do it?”
Example 2: “Son, yes, I smoked when I was your age, it was dumb then. And it’s dumb
now. That’s why I forbid you to smoke, chew, or vape, or use nicotine gum, or whatever
you kids do with tobacco these days.”
Another causal fallacy is the Post Hoc fallacy. Post hoc is short for post hoc ergo propter
hoc ("after this, therefore because of this"). This fallacy happens when you mistake
something for the cause just because it came first. The key words here are “Post” and
“propter” meaning “after" and "because of." Just because this came before that doesn’t
mean this caused that. Post doesn’t prove propter. A lot of superstitions are susceptible to
this fallacy. For example:
“Yesterday, I walked under a ladder with an open umbrella indoors while spilling salt in front
of a black cat. And I forgot to knock on wood with my lucky dice. That must be why I’m
having such a bad day today. It’s bad luck.”
Now, it’s theoretically possible that those things cause bad luck. But since those
superstitions have no known or demonstrated causal power, and “luck” isn’t exactly the most
scientifically reliable category, it’s more reasonable to assume that those events, by
themselves, didn’t cause bad luck. Perhaps that person’s "bad luck" is just his own
interpretation because he was expecting to have bad luck. He might be having a genuinely
bad day, but we cannot assume some non-natural relation between those events caused
today to go bad. That’s a Post Hoc fallacy. Now, if you fell off a ladder onto an angry black
cat and got tangled in an umbrella, that will guarantee you one bad day.
Another kind of causal fallacy is the correlational fallacy also known as cum hoc ergo propter
hoc (Lat., “with this therefore because of this"). This fallacy happens when you mistakenly
interpret two things found together as being causally related. Two things may correlate
without a causal relation, or they may have some third factor causing both of them to occur.
Or perhaps both things just, coincidentally, happened together. Correlation doesn’t prove
causation.
Consider for example, “Every time Joe goes swimming he is wearing his Speedos.
Something about wearing that Speedo must make him want to go swimming.” That
statement is a correlational fallacy. Sure it’s theoretically possible that he spontaneously
sports his euro-style swim trunks, with no thought of where that may lead, and surprisingly
he’s now motivated to dive and swim in cold, wet nature. That’s possible. But it makes more
sense that he put on his trunks because he already planned to go swimming.
Your Turn:
Which kind of causal fallacy is at work in these examples?
Example 2: “Jimmy has a fever, sinus congestion, a cough, and can’t come to school, so
he probably has a test later today.”
Example 3: “Someone really should move this ’Deer Crossing' sign. This is a dangerous
stretch of highway and the deer really should be crossing somewhere else.”
We are susceptible to this errant behavior when we crave that sense of completion or a
sense of accomplishment
“Sunk cost” is an economic term for any past expenses that can no longer be recovered. For
example, after watching the first six episodes of Battlestar Galactica, you decide the show
isn’t for you. Those six episodes are your “sunk cost.” But, because you’ve already invested
roughly six hours of your life into it, you rationalize that you might as well finish it. All
apologies to Edward James Olmos, but this isn’t "good economics" so to speak. It’s more
cost than benefit.
Psychologically, we are susceptible to this errant behavior when we crave that sense of
completion or a sense of accomplishment, or we are too comfortable or too familiar with this
unwieldy project. Sometimes, we become too emotionally committed to an ‘investment,'
burning money, wasting time, and mismanaging resources to do it.
Your Turn:
Consider the following examples. Which of these is a sunk cost fallacy and which is not?
Example 1: “I know this relationship isn’t working anymore and that we’re both miserable.
No marriage. No kids. No steady job. But I’ve been with him for seven years, so I better
stay with him.”
Example 2: “I’m halfway done with college. This is so tough, and It’s not nearly as fun as
I thought it would be, but I don’t know. I guess I’ll finish it and get my degree.”
Like many of the other fallacies in this list, the argumentum ad verecundiam(“argument from
respect”) can be hard to spot. It’s tough to see, sometimes, because its normally a good
responsible move to cite relevant authorities supporting your claim. It can’t hurt. But if all you
have are authorities, and everyone just has to “take their word for it” without any other
evidence to show that those authorities are correct, well then you have a problem.
Often this fallacy refers to irrelevant authorities — like citing a foot doctor when trying to
prove something about Psychiatry; his or her expertise is in an irrelevant field. When citing
authorities to make your case, you need to cite relevant authorities, but you also need to
represent them correctly, and make sure their authority is legitimate.
Suppose someone says, “I buy Fruit of the Loom™ underwear because Michael Jordan says
it’s the best.” But Michael Jordan isn’t a relevant authority when it comes to underwear. This
is a fallacy of irrelevant authority.
Now consider this logical leap: “4 out of five dentists agree that brushing your teeth makes
your life meaningful.” Dentists generally have expert knowledge about dental hygiene, but
they aren’t qualified to draw far-reaching conclusions about its meaningfulness. This is a
fallacy of misused authority. For all we know, their beliefs about the "meaning of life" are just
opinions, not expert advice.
Or take the assumption that “I’m the most handsome man in the world because my Mommy
says so.” Now, while I might be stunningly handsome, my Mom’s opinion doesn’t prove it.
She’s biased. She’s practically required to tell me I’m handsome because it’s her job as a
mother to see the best in me and to encourage me to be the best I can be. She’s also liable
to see me through “rose-colored glasses." And, in this case, she’s not an expert in fashion,
modeling, or anything dealing in refined judgments of human beauty. She’s in no position to
judge whether I’m the most handsome man in the world. Her authority there is illusory.
(Sorry Mom.)
There’s another problem with relying too heavily on authorities. Even the authorities can be
wrong sometimes. The science experts in the 16th century thought the earth was the center
of the solar system (Geocentrism). Turns out they were wrong. The leading scientists, in the
19th century, thought the universe as we know it always existed (Steady State theory). They
too were wrong. For these reasons, it’s a good general rule to treat authorities as helpful
guides with suggestive evidence. But even authorities deserve a fair share of skepticism
since they can make mistakes, overstep their expertise, and otherwise mislead you.
Your Turn:
Consider the following examples. How do these statements mishandle authorities?
Example 1: “Because Martin Sheen played the president on Television, he’d probably
make a great president in real life.”
Example 2: “One day robots will enslave us all. It’s true. My computer science teacher
says so.”
Example 3: “This internet news site said that the candidate punches babies. We know
that’s true because it’s on the internet.”
Your Turn:
Which of these examples is an equivocation fallacy?
Example 1: “His political party wants to spend your precious tax dollars on big
government. But my political party is planning strategic federal investment in critical
programs.”
Example 2: “I don’t understand why you’re saying I broke a promise. I said I’d never
speak again to my ex-girlfriend. And I didn’t. I just sent her some pictures and text
messages.”
“How can you eat that innocent little carrot? He was plucked from his home in the ground at
a young age, and violently skinned, chemically treated, and packaged, and shipped to your
local grocer and now you are going to eat him into oblivion when he did nothing to you. You
really should reconsider what you put into your body.”
Obviously, this characterization of carrot-eating is plying the emotions by personifying a baby
carrot like it’s a conscious animal. So, by the time the conclusion appears, it’s not well-
supported. If you are to be logically persuaded to agree that “you should reconsider what
you put into your body,” then it would have been better evidence to hear about unethical
farming practices or unfair trading practices such as slave labor, toxic runoffs from fields,
and so on.
Truth and falsity aren’t emotional categories, they are factual categories. They deal in what
is and is not, regardless of how one feels about the matter. Another way to say it is that this
fallacy happens when we mistake feelings for facts. Our feelings aren’t disciplined truth-
detectors unless we’ve trained them that way. So, as a general rule, it’s problematic to treat
emotions as if they were (by themselves) infallible proof that something is true or false.
Children may be scared of the dark for fear there are monsters under their bed, but that’s
hardly proof of monsters.
Truth and falsity aren’t emotional categories, they are factual categories.
To be fair, emotions can sometimes be relevant. Often, the emotional aspect is a key insight
into whether something is morally repugnant or praiseworthy, or whether a governmental
policy will be winsome or repulsive. People’s feelings about something can be critically
important data when planning a campaign, advertising a product, or rallying a group together
for a charitable cause. But it becomes a fallacious appeal to pity when the emotions are
used in substitution for facts or as a distraction from the facts of the matter.
It’s not a fallacy for jewelry and car companies to appeal to your emotions to persuade you
into purchasing their product. That’s an action, not a claim, so it’s can’t be true or false. It
would however be a fallacy if they used emotional appeals to prove that you need this car, or
that this diamond bracelet will reclaim your youth, beauty, and social status from the cold
clammy clutches of Father Time. The fact of the matter is, you probably don’t need those
things, and they won’t rescue your fleeting youth.
Your Turn:
Which of these is a fallacious appeal to emotion, and which one is not?
Example 1: “The government needs to hear our cry because we are scared. We are
scared that this candidate will not respect us or protect us. We are scared about our
future. There’s no hope for people like us with these candidates in office.”
Example 2: “These candidates stated that they would close down the education
department and that has many teachers worried about their jobs in 2017.”
For our purposes, we’ll treat all of these fallacies together as the Bandwagon Fallacy.
According to legend, politicians would parade through the streets of their district trying to
draw a crowd and gain attention so people would vote for them. And whoever supported that
candidate was invited to literally jump on board the bandwagon. Hence the nickname
“Bandwagon Fallacy.”
This tactic is common among advertisers. “If you want to be like Mike (Jordan), you’d better
eat your Wheaties.” “Drink Gatorade because that’s what all the professional athletes do to
stay hydrated.” “McDonald’s has served over 99 billion, so you should let them serve you
too.” The form of this argument often looks like this: “Many people do or think X, so you
ought to do or think X too.”
One problem with this kind of reasoning is that the broad acceptance of some claim or action
is not always a good indication that the acceptance is justified. People can be mistaken,
confused, deceived, or even willfully irrational. And when people act together, sometimes
they become even more foolish — i.e., “mob mentality.” People can be quite gullible, and this
fact doesn’t suddenly change when applied to large groups.
Your Turn:
Which of these is a bandwagon fallacy?
Example 1: “Almost everyone at my school will be at the party Friday night. It must be a
popular thing to do.”
Example 2: “Almost everyone at my school will be at the party Friday night. It must be
the right thing to do.”
⁂
We hope this little primer on logical fallacies helps you to navigate future disputes with
friends, family and unhinged online acquaintances without descending into vitriol or childish
name-calling. Or at least if it does descend into vitriol and childish name-calling, you’ll be in a
great position to rhetorically trounce your opponent with sound reasoning and airtight logic.
Did we miss any commonplace logical fallacies? Let us know in the comments section
below!