A Life Preserver for Staying Afloat in a Sea of Misinformation

Melanie Trecek-King

My goals as a science educator are to teach students the essential skills of science literacy and critical thinking. Helping them understand the process of science and how to draw reasonable conclusions from the available evidence can empower them to make better decisions and protect them from being fooled or harmed.

Yet while educators agree that these skills are important, the stubborn persistence of pseudoscientific and irrational beliefs demonstrates that we have plenty of room for improvement. To help address this problem, I developed a general-education science course (Trecek-King 2022) that, instead of teaching science as a collection of facts to memorize, teaches students how to evaluate the evidence for claims to determine how we know something and recognize the characteristics of good science by evaluating bad science, pseudoscience, and science denial.

In my experience, science literacy and critical thinking skills are difficult to master. Therefore, it helps to provide students with a structured toolkit to systematically evaluate claims and allow for ample opportunities to practice. In previous semesters, I’ve had excellent results with James Lett’s “A Field Guide to Critical Thinking” (Lett 1990), in which he summarized the scientific method with the acronym FiLCHeRS (Falsifiability, Logic, Comprehensiveness of evidence, Honesty, Replicability, and Sufficiency of evidence).

While FiLCHeRS has served my students well, I’ve found myself adding rules and updating examples to help my students navigate today’s misinformation landscape. The result is this guide to evaluating claims, summarized by the (hopefully memorable) acronym FLOATER, which stands for Falsifiability, Logic, Objectivity, Alternative explanations, Tentative conclusions, Evidence, and Replicability.

Logo designed
by Wendy Cook

Think of FLOATER as a life-saving device. By using the seven rules in the toolkit we can protect ourselves from drowning in a sea of bad claims.

The foundation of FLOATER is skepticism. While skepticism has taken on a variety of connotations, from cynicism to denialism, scientific skepticism is simply insisting on evidence before accepting a claim and proportioning the strength of our belief to the strength and quality of the evidence.

Before using this guide, clearly identify the claim and define any potentially ambiguous terms. And remember, the person making the claim bears the burden of proof and must provide enough positive evidence to establish the claim’s truth.

Rule 1: Falsifiability

It must be possible to think of evidence that would prove the claim false. It seems counterintuitive, but the first step in determining if a claim is true is to determine if you can prove it wrong.

Falsifiable claims can be proven false with evidence. If a claim is false, the evidence will disprove it. If it’s true, the evidence won’t be able to disprove it. Scientific claims must be falsifiable. Indeed, the process of science involves trying to disprove falsifiable claims. If the claim withstands attempts at disproof, we are more justified in tentatively accepting it.

Unfalsifiable claims cannot be proven false with evidence. They could be true, but because there is no way to use evidence to test the claim, any “evidence” that appears to support the claim is useless. Unfalsifiable claims are essentially immune to evidence. Four types of claims are unfalsifiable.

  • Subjective claims: Claims based on personal preferences, opinions, values, ethics, morals, feelings, and judgments. For example, I may believe that cats make the best pets and that healthcare is a basic human right, but neither of these beliefs is falsifiable no matter how many facts or pieces of evidence I use to justify them.
  • Supernatural claims: Claims that invoke entities such as gods and spirits, vague energies and forces, and magical human abilities such as psychic powers. By definition, the supernatural is above and beyond what is natural and observable and therefore isn’t falsifiable. This doesn’t mean these claims are necessarily false (or true!) but that there is no way to collect evidence to test them. For example, so-called “energy medicine,” such as reiki and acupuncture, is based on the claim that illnesses are caused by out-of-balance energy fields that can be adjusted to restore health. However, these energy fields cannot be detected and do not correspond to any known forms of energy.

There are, however, cases where supernatural claims can be falsifiable. First, if a psychic claims to be able to impact the natural world in some way, such as moving/bending objects or reading minds, we can test the psychic’s abilities under controlled conditions. And second, claims of supernatural events that leave physical evidence can be tested. For example, young-earth creationists claim that the Grand Canyon was formed during Noah’s flood approximately 4,000 years ago. A global flood would leave behind geological evidence, such as massive erosional features and deposits of sediment. Unsurprisingly, the lack of such evidence disproves this claim. However, even if the evidence pointed to a global flood only a few thousand years ago, we still couldn’t falsify the claim that a god was the cause.

  • Vague claims: Claims that are undefined, indefinite, or unclear. Your horoscope for today says, “Today is a good day to dream. Avoid making any important decisions. The energy of the day might bring new people into your life.” Because this horoscope uses ambiguous and vague terms, such as dream, important, and might, it doesn’t make any specific, measurable predictions. Even more, because it’s open to interpretation, you could convince yourself that it matches what happened to you during the day, especially if you spent the day searching for “evidence.” Due to legal restrictions, many alternative medicine claims are purposefully vague. For example, a supplement bottle says it “strengthens the immune system” or a chiropractic advertisement claims it “reduces fatigue.” While these sweeping claims are essentially meaningless because of their ambiguity, consumers often misinterpret them and wrongly conclude that the products are efficacious.
  • Ad hoc excuses: These entail rationalizing and making excuses to explain away observations that might disprove the claim. While the three types of claims described thus far are inherently unfalsifiable, sometimes we protect false beliefs by finding ways to make them unfalsifiable. We do this by making excuses, moving the goalposts, discounting sources or denying evidence, or proclaiming that it’s our “opinion.”

For example, a psychic may dismiss an inaccurate reading by proclaiming her energy levels were low, or an acupuncturist might excuse an ineffective treatment by claiming the needles weren’t placed properly along the patient’s meridians. Conspiracy theorists are masters at immunizing their beliefs against falsification by claiming that supportive evidence was covered up and contradictory evidence was planted.

The rule of falsifiability essentially boils down to this: Evidence matters. And never assume a claim is true because it can’t be proven wrong.

Rule 2: Logic

Arguments for the claim must be logical. Arguments consist of a conclusion, or claim, and one or more premises that provide evidence, or support, for the claim. In effect, the conclusion is a belief, and the premises are the reasons we hold that belief. Many arguments also contain hidden premises, or unstated assumptions that are required for the conclusion to be true, and therefore must be identified when evaluating arguments.

There are two types of arguments, which differ in the level of support they provide for the conclusion.

Deductive arguments provide conclusive support for the conclusion. Deductive arguments are valid if the conclusion must follow from the premises, and they are sound if the argument is valid and the premises are true. For the conclusion to be considered true, the argument must be both valid and sound. For example: “Cats are mammals. Dmitri is a cat. Therefore, Dmitri is a mammal.” The conclusion has to follow from the premises, and the premises are true. Because this argument is both valid and sound, we must accept the conclusion.

In everyday language, the word valid generally means true. However, in argumentation, valid means the conclusion follows from the premises, regardless of whether the premises are true or not. The following example is valid but unsound: “Cats are trees. Dmitri is a cat. Therefore, Dmitri is a tree.” The conclusion is valid because it follows from the premises, but the conclusion is wrong because of an untrue premise; cats aren’t trees.

Inductive arguments provide probable support for the conclusion. Unlike deductive arguments, in which a conclusion is guaranteed if the argument is both valid and sound, inductive arguments provide only varying degrees of support for a conclusion. Inductive arguments whose premises are true and provide reasonable support are considered to be strong, while those that do not provide reasonable support for the conclusion are weak. For example: “Dmitri is a cat. Dmitri is orange. Therefore, all cats are orange.” Even if the premises are true (and they are), a sample size of one does not provide reasonable support to generalize to all cats, making this argument weak.

Logical fallacies are flaws in reasoning that weaken or invalidate an argument. While there are more logical fallacies than can be covered in this guide, some of the more common fallacies include:

  • Ad hominem: Attempts to discredit an argument by attacking the source.
  • Appeal to (false) authority: Claims that something is true based on the position of an assumed authority.
  • Appeal to emotions: Attempts to persuade with emotions, such as anger, fear, happiness, or pity in place of reason or facts.
  • Appeal to the masses: Asserts that a claim is true because many people believe it.
  • Appeal to nature: Argues that something is good or better because it’s natural.
  • Appeal to tradition: Argues that something is good or true because it’s been around for a long time.
  • False choice: Presents only two options when many more likely exist.
  • Hasty generalization: Draws a broad conclusion based on a small sample size.
  • Mistaking correlation for causation: Assumes that because events occurred together, there must be a causal connection.
  • Red herring: Attempts to mislead or distract by referencing irrelevant information.
  • Single cause: Oversimplifies a complex issue to a single cause.
  • Slippery slope: Suggests an action will set off a chain of events leading to an extreme, undesirable outcome.
  • Straw man: Misrepresents someone’s argument to make it easier to dismiss.

Consider the following example: “GMO foods are unhealthy because they aren’t natural.” The conclusion is “GMO foods are unhealthy,” and the stated premise is “They aren’t natural.” This argument has a hidden premise, “Things that aren’t natural are unhealthy,” which commits the appeal to nature fallacy. We can’t assume that something is healthy or unhealthy based on its presumed naturalness. (Arsenic and botulinum are natural, but neither is good for us!) By explicitly stating the hidden premise and recognizing the flaw in reasoning, we see that we should reject this argument.

Rule 3: Objectivity

The evidence for a claim must be evaluated honestly.

Richard Feynman famously said, “The first principle is that you must not fool yourself, and you are the easiest person to fool.”

Most of us think we’re objective; it’s those who disagree with us who are biased, right?

Unfortunately, every single one of us is prone to flawed thinking that can lead us to draw incorrect conclusions. While there are numerous ways we deceive ourselves, three of the most common errors are:

  • Motivated reasoning: Emotionally biased search for justifications that support what we want to be true.
  • Confirmation bias: Tendency to search for, favor, and remember information that confirms our beliefs.
  • Overconfidence effect: Tendency to overestimate our knowledge and/or abilities.

The rule of objectivity is probably the most challenging rule of all, because the human brain’s capacity to reason is matched only by its ability to deceive itself. We don’t set out to fool ourselves, of course. But our beliefs are important to us; they become part of who we are and bind us to others in our social groups. So when we’re faced with evidence that threatens a deeply held belief, especially one that’s central to our identity or worldview, we engage in motivated reasoning and confirmation bias to search for evidence that supports the conclusion we want to believe and discount evidence that doesn’t. If you’re looking for evidence you’re right, you will find it. You’ll be wrong, but you’ll be confident you’re right.

Ultimately the rule of objectivity requires us to be honest with ourselves—which is why it’s so difficult. The problem is, we’re blind to our own biases.

The poster children for violating the rule of objectivity are pseudoscience and science denial, both of which start from a desired conclusion and work backward, cherry-picking evidence to support the belief while ignoring or discounting evidence that doesn’t. There are, however, key differences:

  • Pseudoscience is a collection of beliefs or practices that are portrayed as scientific but aren’t. Pseudoscientific beliefs are motivated by the desire to believe something is true, especially if it conforms to an individual’s existing beliefs, sense of identity, or even wishful thinking. Because of this, the standard of evidence is very low. Examples of pseudoscience include various forms of alternative medicine, cryptozoology, many New Age beliefs, and the paranormal.
  • Science denial is the refusal to accept well-established science. Denial is motivated by the desire not to believe a scientific conclusion, often because it conflicts with existing beliefs, personal identity, or vested interests. As such, the standard of evidence is set impossibly high. Examples include denying human-caused climate change, evolution, the safety and efficacy of vaccines, and GMO safety.

In both these cases, believers are so sure they’re right, and their desire to protect their cherished beliefs is so strong, they are unable to see the errors in their thinking.

To objectively evaluate evidence for a claim, pay attention to your thinking process. Look at all the evidence—even (especially) evidence that contradicts what you want to believe. No denial or rationalization. No cherry-picking or ad hoc excuse-making. If the evidence suggests you should change your mind, then that’s what you must do.

It also helps to separate your identity from the belief, or evidence that the belief is wrong will feel like a personal attack. And don’t play on a team; be the referee. If defending your beliefs is more important to you than understanding reality, you will likely fool yourself.

Rule 4: Alternative Explanations

Other ways of explaining the observation must be considered. It’s human nature to get attached to a single explanation, often because it came from someone we trust or it fits with our existing beliefs. But if the goal is to know the real explanation, we should keep in mind that we might be wrong and consider alternative explanations.

Start by brainstorming other ways to explain your observation. (The more the better!) Ask yourself: What else could be the cause? Could there be more than one cause? Or could it be a coincidence? In short, propose as many (falsifiable) explanations as your creativity allows. Then try to disprove each of the explanations by comprehensively and objectively evaluating the evidence.

Next, determine which of the remaining explanations is the most likely. One helpful tool is Occam’s razor, which states that the explanation that requires the fewest new assumptions has the highest probability of being the right one. Basically, identify and evaluate the assumptions needed for each explanation to be correct, keeping in mind that the explanation requiring the fewest assumptions is most likely to be correct and that extraordinary claims require extraordinary evidence.

For example, one morning you wake up to find a broken glass on the floor. Naturally, you want to know how it got there! Maybe it was a burglar? Could it have been a ghost? Or maybe it was the cat? You look for other signs that someone was in your house, such as a broken window or missing items; without other evidence, the burglar explanation seems unlikely. The ghost explanation requires a massive new assumption for which we currently don’t have proof: the existence of spirits. So while it’s possible that a specter was in your house during the night, a ghost breaking the glass seems even less likely than the burglar explanation, because it requires additional, unproven assumptions for which there is no extraordinary evidence. Finally, you look up to see your cat watching you clean shards of glass off the floor and remember seeing him push objects off tables and counters. You don’t have definitive proof it was the cat, but it was probably the cat.

Rule 5: Tentative Conclusions

In science, any conclusion can change based on new evidence. A popular misconception about science is that it results in proof, but scientific conclusions are always tentative. Each study is a piece of a larger picture that becomes more clear as the pieces are put together. However, because there is always more to learn (more pieces of the puzzle yet to be discovered), science doesn’t provide absolute certainty; instead, uncertainty is reduced as evidence accumulates. There’s always the possibility that we’re wrong, so we have to leave ourselves open to changing our minds with new evidence.

Some scientific conclusions are significantly more robust than others. Explanations that are supported by a vast amount of evidence are called theories. Because the evidence for many theories is so overwhelming, and from many different independent lines of research, they are very unlikely to be overturned—although they may be modified to account for new evidence.

Importantly, this doesn’t mean scientific knowledge is untrustworthy. Quite the opposite: science is predicated on the humility of scientists and their willingness and ability to learn. If scientific ideas were set in stone, knowledge couldn’t progress.

Part of critical thinking is learning to be comfortable with ambiguity and uncertainty. Evidence matters, and the more and better our evidence, the more justified we are in accepting a claim. But knowledge is not black or white. It’s a spectrum with many shades of gray. Because we can never be 100 percent certain, we shouldn’t be overly confident!

Therefore, the goal of evaluating claims and explanations isn’t to prove them true. Disprove those you can, then tentatively accept those left standing proportional to the evidence available and adjust your confidence accordingly. Be open to changing your mind with new evidence and consider that you might never know for sure.

Rule 6: Evidence

The evidence for a claim must be reliable, comprehensive, and sufficient. Evidence gives us reasons to believe (or not believe) a claim. In general, the more and better the evidence, the more justified we are in accepting a claim. This requires that we assess the quality of the evidence based on the following considerations:

The Evidence Must Be Reliable

Not all evidence is created equal. To determine if the evidence is reliable, we must look at two factors:

  • How the evidence was collected. A major reason science is so reliable is that it uses a systematic method of collecting and evaluating evidence.

However, scientific studies vary in the quality of evidence they provide. Anecdotes and testimonials are the least reliable and are never considered sufficient to establish the truth of a claim. Observational studies collect real-world data and can provide correlational evidence, while controlled studies provide causational evidence. At the top of the hierarchy of evidence are meta-analyses and systematic reviews, as they are a combination of other studies and therefore look at the big picture.

  • The source of the information. Sources matter; unreliable sources do not provide reliable evidence. In general, the most reliable sources are peer-reviewed journals, because as the name suggests, the information had to be approved by other experts before being published. Reputable science organizations and government institutions are also very reliable. The next most reliable sources are high-quality journalistic outlets that have a track record of accurate reporting. Be skeptical of websites or YouTube channels that are known to publish low-quality information and be very wary of unsourced material on social media. In addition, experts are more reliable than nonexperts, because they have the qualifications, background knowledge, and experience necessary to understand their field’s body of evidence. Experts can be wrong, of course, but they’re much less likely to be wrong than nonexperts. If the experts have reached consensus, it is the most reliable knowledge.

The Evidence Must Be Comprehensive

Imagine the evidence for a claim is like a puzzle, with each puzzle piece representing a piece of evidence. If we stand back and look at the whole puzzle, or body of evidence, we can see how the pieces of evidence fit together and the larger picture they create.

You could, either accidentally or purposefully, cherry-pick any one piece of the puzzle and miss the bigger picture. For example, everything that’s alive needs liquid water. The typical person can live for only three or four days without water. In fact, water is so essential to life that, when looking for life outside of Earth, we look for evidence of water. But what if I told you that all serial killers have admitted to drinking water? Or that it’s the primary ingredient in many toxic pesticides? Or that drinking too much water can lead to death?

By selectively choosing these facts (or pieces of the puzzle), we can wind up with a distorted, inaccurate view of water’s importance for life. So if we want to better understand the true nature of reality, it behooves us to look at all the evidence—including (especially!) evidence that doesn’t support the claim. And be wary of those who use single studies as evidence; they may want to give their position legitimacy, but in science you don’t get to pick and choose. You have to look at all the relevant evidence. If independent lines of evidence are in agreement, or what scientists call consilience of evidence, the conclusion is considered very strong.

The Evidence Must Be Sufficient

To establish the truth of a claim, the evidence must be sufficient. Claims made without evidence provide no reason to believe and can be dismissed. In general:

  • Extraordinary claims require extraordinary evidence. Essentially, the more implausible or unusual the claim, the more evidence that’s required to accept it.
  • Claims based on authority are never sufficient. Expertise matters, of course, but experts should provide evidence. “Because I said so,” is never enough.
  • Anecdotes are never sufficient. Personal stories can be very powerful. But they can also be unreliable. People can misperceive their experiences, and, unfortunately, they can also lie.

As an example, let’s say you own a company, and Jamie works for you. She is an excellent employee, always on time, and always does great work. One day, Jamie is late for work. If Jamie tells you her car broke down, you most likely will believe her. You have no reason not to—although if you’re really strict you may ask for a receipt from the tow truck driver or mechanic. But what if Jamie tells you she’s late because she was abducted by aliens? I don’t know about you, but my standard of evidence just shot through the roof. That’s an extraordinary claim, and she bears the burden of proof. If she tells you that one of the aliens took her to another dimension and forced her to bear offspring but then reversed time to bring her back without physical changes … Again, just speaking for myself, I’m either going to assume she’s lying or suggest she see a professional.

Rule 7: Replicability

Evidence for a claim should be able to be repeated. Replicability (and its related terms) can refer to a range of definitions, but for the purpose of this guide it means the ability to arrive at a similar conclusion no matter who is doing the research or what methodology they use. The rule of replicability is foundational to the self-correcting nature of science, because it helps to safeguard against coincidence, error, or fraud.

The goal of science is to understand nature, and nature is consistent; therefore, experimental results should be too. But it’s also true that science is a human endeavor, and humans are imperfect. This can lead to fraud or error. For example, in 1998, Andrew Wakefield published a study claiming to have found a link between the MMR (measles, mumps, rubella) vaccine and autism. After scientists all over the world tried unsuccessfully to replicate Wakefield’s findings—with some studies involving millions of children—it was discovered that Wakefield had forged his data as part of a scheme to profit off a new vaccine. The inability to replicate Wakefield’s study highlights the importance of not relying on any single study.

Conversely, we can be significantly more confident in results that are successfully replicated independently with multiple studies. And we can be the most confident in conclusions that are supported by multiple independent lines of evidence, especially those from completely different fields of science. For example, because evidence for the theory of evolution comes from many diverse lines, including anatomical similarities, shared developmental pathways, vestigial structures, imperfect adaptations, DNA and protein similarities, biogeography, fossils, etc., scientists have great confidence in accepting that all living things share a common ancestor.

 

Conclusion

Using FLOATER’s seven rules to evaluate claims can help us make better decisions and protect us from being fooled (or even harmed) by false or misleading claims. Evaluating claims this way will likely take practice. But don’t get discouraged; it’s worth it!

Acknowledgments

This work would not have been possible without Dr. James Lett. His original article “A Field Guide to Critical Thinking” has been an invaluable resource to educators such as myself for decades, and I greatly appreciate his support and suggestions while I was writing this update. I would also like to extend my deepest gratitude to Matthew P. Rowe, Marcus Gillespie, and John Cook for their support, feedback, and guidance.

 

References

Lett, James. 1990. A field guide to critical thinking. Skeptical Inquirer 14(2): 153–160.

Trecek-King, Melanie. 2022. Teach skills, not facts. Skeptical Inquirer 46(1): 39–42.

Melanie Trecek-King

Melanie Trecek-King is an associate professor of biology at Massasoit Community College. Her website is www.ThinkingIsPower.com.