Learning critical thinking by following uncritically

When I read the new Cook et al 2018 paper for the first time, the one thing that stood out was that the example arguments were simplified versions of skeptical arguments, stripped down of any nuance and context, therefor not representative anymore. I already foresaw many posts in my future about these fabrications…

In the meanwhile I found the discussion of Barry Woods on Twitter, tirelessly calling out the many misrepresentations in the paper. The reaction of some of his opponents, that this doesn’t matter because the compiled arguments are fallacious anyway, puzzled me. I couldn’t grasp that they were just okay with:

  • The authors (or Cook and the SkS team) coming up with simplified, unnuanced arguments based on what they think their opponents believe
  • then Cook et al show that these simplified, unnuanced arguments are logically fallacious
  • thus providing proof that their opponents are wrong and therefor should be safely ignored when it comes to those issues.

That is about as close as one can get to a straw man argument. For those who are not familiar with this type of fallacy, according to wikipedia the definition of a straw man argument is (my emphasis):

A straw man is a common form of argument and is an informal fallacy based on giving the impression of refuting an opponent’s argument, while actually refuting an argument that was not presented by that opponent.

The examples Cook et al used were textbook examples of this type of argument, but the defenders of the paper were undeterred by it or maybe did not understand the concept. It seemed to shed of them like water off a duck’s back. I couldn’t really understand that, given that it is pretty clear for everybody to see.

Until I found following tweet:

Of course! How could I been so shortsighted? Until then I was solely focused on whether the arguments were actual arguments from “denialists” and whether the simplified, unnuanced arguments without context were representative for the claims found in the wild. Because of this focus, I failed to understand that the goal of the paper could be different from faithfully representing “denialist” arguments. Re-reading the paper with that in mind gave me an aha moment.

Before I go on, I still believe that Cook et al blatantly (and probably even knowingly) misrepresent the arguments of their opponents and therefor are attacking straw men, but this focus distracts from understanding what the paper is about. That being of my chest, let’s continue with the paper.

The paper starts with the usual problem definition:

  • Misinformation can have significant societal consequences
  • There is an overwhelming scientific consensus
  • There is however little awareness of the scientific consensus among the general public
  • Therefor support for mitigation policies is stalled.

Luckily there is a solution: the inoculation theory. This can neutralize the influence of misinformation “by explaining the techniques used to distort the facts”.

The last piece of the puzzle is the list of 42 common “denialist” claims in the supplementary data. Each item is a statement, followed by the analysis why this statement is a fallacy. These items could be considered as a weaker form of the arguments found in the wild. These items on the list are “analysed” and “found to be false”. Therefor they could be used to “explain the techniques used to distort the facts”. Meaning, when an argument is found in that list, it could show the technique used in that argument, therefor weakening or even eliminating its effect. Just as with an inoculation in modern medicine, it is not necessary to use the real arguments. That would be too messy and would complicate things.

That list will probably be used by those with no(t much) knowledge or background of the issue yet needing a quick-and-dirty way of evaluating claims, for example communicators (this is an actual target group of this list). For them the list will be a resource that they can fall back to. If they feel insecure about a claim, then they can search that list. This is how I thinks it works in practice:

  • A communicator encounters something that might be a “denialist” claim
  • He/she looks into the list of 42 of fallacious arguments and tries to find something that fits that claim
  • When there is something similar in the list ⇒ it is assumed misinformation and the argument could safely be dismissed.

In theory this may seem okay, but the arguments that could be found in the wild are much more nuanced than those provided by the list. The items on that list have no nuances and no context is considered. For example, if someone claims that “there have been other scientific consensuses that have been wrong so we can’t rely solely on the consensus on climate change, then it will land us in the fourth row. According to the analysis, this is because it has the hidden premise that “if other consensuses proved untrustworthy, then the consensus on AGW must be untrustworthy”.

If that is the actual premise that the argument is build on, then yes, this is most definitely a logical fallacy. But if that hidden premise is not the starting point, but something else is (for example, some similar consensuses have been wrong in the past → we have to be careful with such a consensus and not just accept it at face value), then this is a new construct and this does not necessarily results in a logical fallacy. If that communicator just ticks this argument from the list because it looks similar, then that claim will be considered as misinformation, which is not necessarily the case. It is easy to miscategorize arguments if one doesn’t take care of the actual argument or what is exactly in the analysis of that item.

Personally, it think this is a dangerous methodology for the target group it is designed for. It gives the misleading idea that the debate is rather straight forward and could easily be categorized in 42 slick items. It becomes tricky when one has to deal with real arguments that have nuances and are made in a certain context. Making it not as black & white as it is presented.

I don’t think this technique will empower communicators and educators by means of critical thinking. On the contrary, it will lead to uncritically following the vision of the authors/SkS that denialists are wrong because when they looked at the arguments they think the denialists believe in, then they only found logical fallacies.

via Trust, yet verify


February 25, 2018 at 01:20PM

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google+ photo

You are commenting using your Google+ account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s

%d bloggers like this: