The scientific method


Part one: Introduction and logics


In our enlightened society, we tend to look down on those that blindly follow religious doctrines. We accuse them of being ignorant and close-minded, not questioning unreasonable claims by their authorities, and when challenged, resorting to rhetoric or aggression rather than logic to defend their views.

Perhaps phrased a bit extremely, this is the essence of the attitudes of most ‘civilised’ societies that are based on scientific knowledge. Science encourages critical thinking about the world, and gains knowledge by performing experiments that help us figure out how the world works. Religion, they say, only makes up facts that happened in the past to explain things in our everyday lives. That is more or less the way they tend to think.

But…are those who follow science really that different? Take a moment to think about the last time you actually thought critically about what you read in a science magazine.

It does not take much intellectual process to question a fact or opinion. ‘I think this is wrong, because of A’, where A is fact you have learnt that contradicts that claim. This is not what I would call critical thinking, something our modern society values highly, yet tends to misuse so often it is almost frightening.

Critical thinking is more about questioning the essence or nature of a statement or piece of knowledge. Thinking critically is not to question a claim with another claim that contradicts it and saying that one of them is wrong – that is just pointless argument. When thinking critically, you challenge the claim per se (lat. ‘in itself’), by assessing its fundamental reasoning. An example of critical thinking is detecting fallacies – specific cases of flawed logical reasoning – such as circular reasoning and false dilemma, both which are surprisingly common in science. (Fallacies will be discussed more in detail later.)

Having this in mind, ask yourself again: when did you last think critically about a scientific statement?

Would you then agree with me that, in general, we only rarely – if ever – do this? Hopefully, this should shock you, not only because you have realised that we then are not much different from the blindly religious people we denounce as ‘ignorant’ or even ‘brainwashed’, but also since you understand and appreciate the importance of critical thinking and realise what we are lacking.

If we want to justify our trust in science, we need to be able to convince ourselves (and eventually others) of why scientific knowledge is ‘better’ than, say, religious beliefs. To do that, we need to look at the essence and nature of science, and compare it with the essence and nature of other ways of obtaining knowledge, e.g. religion.

This is not an easy task, and it should not surprise you why so few are even able to go through with it, since not many are familiar with how scientists reason (ideally). We all know that they perform experiments and observe results. Some of us further are aware that the scientist interpret the results of the experiments within a theoretical framework, use it to generalise about similar situations, and so add to our collective knowledge. Moreover, some know of the importance of replicability (that the experiments should be repeatable), and probably fewer have heard about falsificationism (that ideas are confirmed not by trying to support them, but by trying to destroy them and not succeeding).

But how does all of this really help us understand the world? Why do we need to perform experiments? Why do we want to generalise? And why should the experiments be replicable? Why does failing to disprove something make it more reliable than if you can show that it is true? How do all these things interrelate and connect to form a comprehensive whole? In short, what are the theoretical reasons for why science is a good method for making sense of the physical world?

That is what this post series will be all about: assessing the scientific method, by bringing to light how it works, and thinking a lot about what makes it good and what problems it faces. In other words, we will critically think about the strengths and weaknesses of the scientific method.

Note that I am not intending to criticise science, but to make us think critically about it, or evaluate it. There is a difference in purpose, which is really important that you do not misunderstand. I am not going to say that science is wrong, only that it is not as right as many people think it is. Indeed, it may very well seem as I am pointing more toward its flaws and limitations, but that is mostly because that is the part the general public is less familiar with, I believe. I will of course also emphasise the really good aspects of science.

Honestly, considering all that science has achieved, most indisputably shown in its practical applications, it is clear that blindly arguing that science is all bad and wrong is just silly. All I wish to achieve is to encourage you to think for yourself about the way you see science, and whether that view is properly justified, or whether you might need to think again.

In order do this, we first need a solid introduction to the essentials of reasoning and logic, and a detailed walk-through of the scientific method. This will include a whole bunch of new words and definitions that you are probably unfamiliar with, but please bear in mind that understanding these fundamental concepts is key to understanding the rest, so I strongly urge you not to skip this part. In coming posts, I will go through each main step in the chain of the idealised scientific reasoning, analysing them in detail in the light of what we will have learned earlier. I will also address other central concepts of science, such as models, operationalisation, paradigms, the ‘data first vs. theory first’ discussion, and also the role of mathematics.


Logics – the structure of an argument

Perhaps in contrary to what most people think about logics, it is all actually about the structure of the arguments, and the purpose of this structure is to preserve truth when reasoning beyond what is already known. In logics, we want to be able to build on what we know in a way that we can be certain that what comes out of it is as true as the things we base it on. I will soon show you why, but first you might wonder why logic would be useful.

What I said above is basically that ‘logical’ is not the same as ‘true’. ‘Logical’ means that it follows a format in which the conclusion (the output of a chain of reasoning) is as true as the arguments you base it on (the input). If you start with true statements and reason logically, you will end up with true statements. If you start with some true and some false statements, logical reasoning will not result in statements you can rely on to be true.

Why, then, should we bother about logic? It does not seem that helpful if we cannot get anything out of it that is truer than what we already have. One could thus see logic as redundant, since it cannot take us any closer to the truth, since it does not ‘improve’ truths.

I would agree that this is a limitation in some cases, but it all depends on how you use logic. What we ideally want to do with logical (and scientific) reasoning is to extrapolate knowledge. To extrapolate is basically to take something you know or expect to be true in one case and apply the same knowledge to a related but different case, so that you can learn more about that second case without needing to observe it directly. (With more formal wording, I would define extrapolate as: extending an application of a method or conclusion to a related or similar situation, assuming it is applicable there as well.) You want to be able to pick out the essence of something, and use that essence to say something more about similar things.

I happen to like bananas, pineapples and mangos very much. These are all fruits from the tropics. They have that feature in common (their geographic origin, and also climate). Therefore, I can conclude that I probably like all or most tropical fruits. I would like to be able to tell whether I will like coconuts, without trying them first. Coconuts are also tropical fruits. I can then extrapolate my earlier conclusion to this other tropical fruit, and say with some confidence that I will probably like coconuts. This is an extrapolation.

This type of dummy-proof phrasing is mostly to give you a taste of what is to come when we dive deeper into logics, but I promise I will not be as obnoxiously explicit most of the time – only when I really want to be crystal clear.

Anyhow, hopefully you now understand what extrapolation means. It is a critical concept in the philosophy of knowledge, and, actually, we do this nearly all the time without thinking about it. It is instinct – an intuitive way of reasoning – because it is incredibly useful. Without applying our knowledge to other areas, we would be stuck with only knowing what we can see with our own eyes and hear with our own ears. (In a later post, we will come to the potential dangers of that as well.) Extrapolation allows us to break free from the chains of our limited self and explore other worlds we only have hunches about.

When you are about to cross the street, you assume the approaching car will stop, because other cars have in the past. When your friend offers you a new sweet he or she thinks you might like, you accept it because you like usually like sweets, and because you usually like what he or she offers you.

When NASA is searching for life on other planets, it is looking for signs based on what we believe allows us to live on our planet (atmosphere, dynamic planet interior, water, carbon, etc.). When biochemists test new drugs on lab rats, they assume the results in the rats will be similar to the consequences of giving it to a human. Geologists are currently trying to work out whether east Africa is about to split off from the rest of the continent, based on what clues they have about such events in the past and what they observe today and interpret as indicators of a continent splitting.

Extrapolations can be made in many other areas as well, e.g. social sciences, and others where it is less obvious, such as mathematics and art. If you are a keen mathematician or artist, take a moment to think about when and for what purpose you use extrapolation in your work.

I hope you now begin to understand the importance of logic as well. If you think about it, none of these examples are more certain than the facts we base them on. I am actually not as fond of fresh coconut as I am of the other tropical fruits; the hasty conclusion that I might like all tropical fruits was evidently not true. I strongly doubt that every car you have met when crossing a street has stopped for you; therefore, you cannot be absolutely certain that the next driver will stop – the extrapolated statement that the next car probably will stop before you becomes uncertain because it is based on every car so far having stopped, which is not quite the case. I can bet that your friend occasionally gives you something you surprisingly happen not to like, and you have surely come across some sweet that just was not your thing.

If you understand the importance of extrapolating knowledge, and thus the importance of logical reasoning, we can start to look at the structure of logic.

The most basic way of structuring an argument – basic in the sense that it is simple and clear, and therefore easily brings out its essence – is making it into a syllogism. Syllogisms consist of two (or more) so-called premises and one conclusion that follows necessarily – i.e. logically – from the premises. Let me give you a classic example:

     P: All bachelors are unmarried men
     P: Peter is a bachelor
     C: Therefore, Peter is an unmarried man

I let the Ps indicate premises and the C indicate the conclusion; it is not necessary, but I want to avoid confusion.

So, that’s a syllogism! It is that simple! Just to make sure you get the hang of it, try to restructure this into a syllogism: Because John was late for class, and all latecomers must be punished, he will sit in detention for an hour after school.

     P:
     P:
     C:

I made it a bit tricky for you, but, in fairness, few arguments you hear will be laid out so plainly that you can make a syllogism by just copying and pasting the words. A key to good reasoning is to almost instinctively think in terms of syllogisms – whenever you hear an argument, rearrange it in your head – and you will start to notice how many times per day you hear incomplete reasoning all around you. That is one of the strengths of this format: it is so easy to spot gaps in arguments.

Take a simple example: ‘Fossils are remains of past life, so evolution must have occurred in the past.’ For a start, you cannot make a syllogism with only one premise – not without it becoming silly or redundant. You need to add at least one more premise, preferably one that fills the jump between fossils and evolution – i.e. shows the connection. If we add that ‘fossils are more or less different from now-living organisms’, it starts to make more sense. Or does it? We have still not quite touched upon the subject of evolution yet. ‘Evolution is organisms changing over time’ could be just what we need. Shall we try it out?

     P: Fossils are remains of past life
     P: Fossils are more or less different from now-living organisms
     P: Evolution is organisms changing over time
     C: So, evolution must have occurred in the past

Although the wording does not quite fit neatly, the line of reasoning now emerges as more or less complete. It needs some more polishing, but it makes more sense now that some large holes have been filled.

Now that you are more familiar with the layout of syllogisms, I can introduce the concepts of validity and soundness.

In logics, a valid argument is one where the conclusion follows necessarily from the premises. For an argument to be valid, its conclusion must be an indisputable and inevitable consequence of the premises. Note that validity and truth are not the same: arguments can be valid without being true. I will show you how very soon.

A sound argument is one that is valid and with true premises. If the premises are true, and the conclusion is a necessary consequence of them, the argument is sound. Soundness can be thought of as roughly synonymous with truth.

To really show the difference, I think it is best if we consider some (pretty fun) examples.

The validity of an argument is all about structure: we want a water-proof chain of thoughts, and focus on that only. The truth or falseness of the premises is dealt with later – it is not of our concern at this stage. We only want to make sure that the reasoning is correct.

A perfectly valid syllogism could therefore be:

     P: All bings are bongs
     P: Ding is a bing
     C: Therefore, ding is also a bong

Here is another example of validity, which I took from a handout for a philosophy class:

     P: All ostriches are teachers
     P: Richard is an ostrich
     C: Therefore, Richard is a teacher

Albeit utterly nonsensical, both are valid. There is no sensible way of questioning the logic in these statements, regardless of how absurd they sound. If you have trouble seeing why, try exchanging the words for symbols or letters. I like to use uppercase letters in italics, but you can choose any of your preference; there is no rule for that.

     P: All As are B
     P: C is an A
     C: Therefore, C is also a B

This is, so to speak, a general formula for a valid syllogism. There are many other general forms of valid statements, but this is the standard. Try exchanging A for ‘bird(s)’, B for ‘dinosaur(s)’ and C for ‘(a) chicken’.

With the above example, we are moving toward a sound syllogism, i.e. one where the premises are regarded as true. Validity is a requirement for soundness, and so is that the premises are true. The latter can be much more difficult to show, so we should not be hasty to call an argument sound only because we believe the premises to be true – a dangerously common pitfall – we must be absolutely certain, and there are few ways to achieve that.

One type of premises that are indisputably true are those that are true by definition. Look back at the first example, the one about bachelors being unmarried men: if a person is not a man or not unmarried, then he/she simply is not a bachelor. Thus, all bachelors are unmarried men, because if they are not, then they cannot be appropriately called bachelors.

What a silly thing to be arguing about, right? Where does that lead us? All it does is divide the world into bachelors and non-bachelors, and allows us to say something about all bachelors – that they are men and unmarried – and something about all non-bachelors – they are either not men, married, or both cases. Sounds daft. Or? We will return to this in some later post, because it opens up to a quite intriguing discussion, but it is not my purpose to drag you into it here and now.

A second way a premise can be regarded as true is if it is the conclusion of a previous sound syllogism. How come? Consider this generalised example:

Syllogism 1
     P: All As are Bs
     P: C is a B
     C: Therefore, C is also a B

Syllogism 2
     P: C is a B
     P: All Bs are D
     C: Therefore, C is also D

If you regard Syllogism 1 as sound, it means you must accept its conclusion to be true. If that conclusion is used as a premise in the next syllogism, that premise must be regarded true, as it is the conclusion of a sound argument. If the remaining premise also is true, then the valid Syllogism 2 is sound, and its conclusion must be regarded as true as well – and can be used as a premise for more syllogisms, if you so wish.

Anyhow, I hope you begin to realise that syllogisms really are quite simple, as long as you have grasped the essence. But, what is the point of this? Why this structure? Why do we want valid arguments? Why do we want sound arguments?

Please take a moment to reflect on those questions yourself before reading on. I want you to get into the habit of thinking for yourself about the meaning and importance of the things you hear.

I hope you reached some interesting conclusions, and, for the better, that they were different from mine. That way, you might gain double insight into the purpose and usefulness of syllogisms.

One purpose of the particular layout of syllogisms is to put it plainly and simply, so that anyone and everyone can understand it and see the logic (or lack thereof) easily. In short, it is all about clarity. In addition, it can be thought of as a way of systematically approaching intellectual problems.

Validity is, in essence, equal to ‘logic’, in the sense that it is a means of preserving the truth. (Remember my first statement in this section?) If you start with true premises and reason validly, you will end up with truth (note that this is then a sound argument). If you start with false or not entirely true premises and reason validly, you are unlikely to reach a true conclusion, though it may be possible in theory:

     P: All potatoes are presidents of the US
     P: Obama is a potato
     C: Therefore, Obama is a president of the US

If you start with truth and reason invalidly, then you have no clue of what you end up with. For all we know, you can start with false premises, reason invalidly, and – by sheer luck! – reach a true conclusion. However, you cannot be certain about its truth, since you cannot show logically how you came to it. Even though it may be true, if you cannot show it logically, you may have a hard time convincing others. (See the next syllogism example below.)

The one thing you can be absolutely certain about is that you can never ever start with true premises, reason validly and end up with a false conclusion. If your conclusion is false, then either the premises cannot all be true, or the argument cannot be valid. This is so because of the definitions of validity and soundness: they are defined so that, when combined as I have shown you, truth is a necessary and inevitable result; the key issue is showing that an argument is valid and that the premises are true – i.e. the task is to show that your argument is sound, because, once you have, it must be accepted as true. That is all you need to do, but it is easier said than done!  

Take these lessons with care! Beware not to misuse what you have now learned. Although showing the validity of a statement is relatively easy, showing its soundness – i.e. the unquestionable truth of its premises – can be nearly impossible in many many many cases. Falsely describing an argument as sound, perhaps just because you agree with the conclusion, is a pitfall we have all fallen into, and probably will many times over again. We should all take caution not to announce something as sound unless we can be absolutely sure!

It is also easy to be fooled into thinking that an invalid argument is valid. Think about this one, taken from another handout on logics:

     P: Vegetarians do not eat pork sausages
     P: Ghandi did not eat pork sausages
     C: Therefore, Ghandi was a vegetarian

The both premises and the conclusion are all true, but the argument is actually invalid! You might see this more clearly if you try it with symbols:

     P: All As are not Bs
     P: C is not a B
     C: Therefore, C is an A

It only happens to be, that, in this case, C is one of the non-Bs that also are As, but since not all non-Bs are As, the conclusion cannot be regarded as a necessary consequence of the premises. Since the argument is not valid, it is not sound either.

Does that make it less true? No, not at all, but it makes it less convincing to critics. We have no guarantee that the conclusive statement is true (let us disregard the fact that the second premise can be questioned and focus on the problem of invalidity now), since we cannot show that it is a necessary consequence of the premises. The conclusion is questionable, even though the premises are not.

Detecting invalidity is not an easy task for a beginner. You need to drill yourself in the art, and the best way to begin is to start breaking everything down into syllogisms, and exchanging the key words for symbols if required. Eventually, it might become second nature to you, and you can ‘feel’ gaps in arguments more and more easily.

However, there are some common types of invalid reasoning called fallacies. These are, so to speak, pre-described logical pitfalls. They are distinct special cases of invalid reasoning, but strikingly common. I will give you some examples here.

Post hoc ergo propter hoc is Latin for ‘after this, therefore on account of this’, and refers to when you assume that that because B follows from A, then A is the cause of B, or, in other words, assume that correlation means causation.

This reasoning can be stated as a syllogism:

     P: Whenever A occurs, B follows
     P: Whenever B occurs, so has A
     C: Therefore, A is the cause of B

Why this is invalid might not be obvious straight away, because it looks structurally correct. However, you must note that occurrence and cause do not have a two-way connection. Since occurrence and cause are not invariably connected, the conclusion does not follow necessarily from the premises; thus, the argument is invalid.

If A causes B, they will naturally occur together or after one another; however, looking at it from the other end, just because A and B occur in close association, there is no guarantee that either causes the other. For example, there may very well be another factor, C, which causes both!

A really simple example, but that one might not think of immediately, is that although night is followed by day, night is not the cause of day. (Both are caused by the Earth’s rotation with respect to sunlight coming from only one direction.)

This is a fallacy made often in science, and also in other areas of knowledge. However, the scientific method is beautifully capable of dealing with that issue, which will be shown in a later post, when we explore the scientific method model in detail. Ideally, if we notice that two phenomena are correlated, it should give us a hint that it might be worth investigating whether the connection could be causal or just coincidence; we must at least be careful not to jump to conclusions!

Equivocation is the use of different meanings of a word in the same argument. Take this example from a philosophy handout:

P: A hamburger is better than nothing
P: Nothing is better than good health
C: A hamburger is better than good health

The ambiguous use of the word ‘nothing’ makes the argument nonsensical. Strictly, however, the syllogism is formally valid, since the conclusion does follow necessarily from the premises. This is a very important limitation of logics, because there is no method to deal with this. Discussing such reasoning will almost always boil down to arguing about the meaning of the words, as so many arguments tend to do.

This problem is probably not more or less common in science than in any other area of knowledge; ambiguity is an inherent problem in language, introduced by our desire for variation in ways of describing things. The aesthetics of language has thrust a wedge through an aspect of logics.

Ad hominem means ‘against the man’, and is the fallacy of attacking or supporting the person instead of the argument. Clearly, the invalidity is a result of completely missing the point. Instead of examining the statement per se, you accept or refute it because you like or dislike the person that made the claim.

It may very well be that this person you like and agree with is right, but only in the same way that invalid reasoning could result in a true conclusion: by sheer luck. There is a possibility that the person is correct, but it is not guaranteed. Thus, it is a type of fallacy.

In science and history, this problem is more familiar under the term bias, which is favouring or disfavouring a view, person or group, usually in an unfair way. Especially keen students of history should be aware of the importance of carefully assessing the degree and direction of bias in a source, and constantly taking care to circumvent the angled view of the writer. In science, the problem is less pronounced thanks to the principles of replicability and falsification, which will be explained in detail at a later point.

It is the same problem when we appeal to what the majority of the people believe in. Indeed, if most believe in something, there ought to be a good reason for that belief, or else it would probably not have been so wide-spread. However, think about times when society considered slavery, racism and sexism acceptable and even natural. (Note that this case has little relation to logics: in past times, these horrendous things were the order of the day because it had been so for long, and the people in power saw no good reason for change; it is a matter of social norms, far removed from strict logical reasoning. In the same way, trusting in the opinion of the vast majority is not logical reasoning, but simply a matter of trust.)

A sort of grey area is when we trust authorities. Especially in modern times, when our collective knowledge is so vast and deep that a single individual simply cannot know or try out everything for him-/herself, we must, in practice, trust authorities in relevant areas that we have less experience in or knowledge of.

I am sure many of us are eager to know what gravity is. Really, what is it? What makes a body attract others just because it is bigger? That just makes no sense in itself. A good friend tells me there is mathematical proof for it, but when I ask her to explain it, she cannot even give me the basics. Evidently, she does not understand the proof well enough to summarise it to someone who has not done much physics. How can she then claim that there is proof for gravity? Her physics teacher probably told her so, and maybe even demonstrated it, but the essentials are too esoteric (i.e. understood only by very few people with highly specialised knowledge or interest) for her to fully grasp the concept. Perhaps she was convinced of it when she saw it demonstrated, but did not assimilate (i.e. take in) the proof, only recognised and accepted it. What happened here is that she relied on an authority in the relevant area (her physics teacher) that the concept she did not herself understand, and perhaps never will, was true. She placed her trust in that those who understand gravity know what they are doing.

In practice, we have to rely on authorities in many many cases; since we simply cannot investigate everything ourselves, we have no choice but to trust others that have spent much of their lives examining the particular field of knowledge. The best we can do is prioritise: investigate in-depth only the parts we find most important to ourselves, and be content with trusting people to do their thing right in other areas.

Is this bad? Not in all cases! It would be arrogant not to acknowledge that there are plenty of people in the world that can do things better than yourself. I know how clumsy I am in mathematics, so when my teacher gets a different result, I naturally accept that I am probably wrong and try again. I can be equally clumsy in a chemistry laboratory (unless we are handling dangerous chemicals, in which case I naturally take uttermost care), so when the ammeter (a device that measures electric current) reads really weird values from an electrolytic cell (a set-up of various chemicals in a way that their reactions produce electricity), I instinctively presume something is amiss with my set-up, not that my odd results disprove the theory of electrolytic cells! There are many more examples where I would prefer to trust an authority because I am sure that he/she is a more capable investigator (in the particular area, or in general).

What makes a good authority? We do not want to rely on just about anyone; we want to be sure that the person we place our trust in is worthy of it. This is a rather interesting question, but its full complexity is not relevant here, so I will give only one short, intuitive answer: the authority should at least be an expert in the relevant field of knowledge.

Argumentum ad ignorantiam means ‘argument from ignorance’, or otherwise phrased as ‘appeal to ignorance’. This is when someone claims something to be true because there is no evidence to disprove it.

     P: There is no evidence that A is false
     C: Therefore, A must be true

Clearly, something is lacking here. We need at least one more premise for the syllogism to be complete. What if we do like this:

     P: A must be either true or false
     P: There is no evidence that A is false
     C: Therefore, A must be (or is probably) true

Now it is valid, but justifying the new premise will be difficult. It may work in some cases, but experience tells us that very few things are black and white: life is more of a mixture of shades of grey. In fact, assuming the first premise to be true without proper reason is another fallacy, which we will consider next.

False dilemma is when you, without justification, assume that only two alternatives exist. It is also referred to as binary thinking, or ‘black and white thinking’. Repeating what I said above, clearly, this is not true in the vast majority of cases.

Sarcastically, humourist Robert Benchley once said: “There are two kinds of people in the world: those who divide the world into two kinds of people, and those who don't”.

As a side comment, I would like to add that science has put considerable effort to both show that many natural phenomena are determined by a variety of factors that can interact with one another, and to explore how these function by attempting to model them, chiefly with the help of mathematics. Although I expect that there are scientist out there that have tendencies to commit these fallacies, such errors tend to be checked and corrected by other scientists, so that, overall, science works in the general direction of increasing the variety of grey hues we are aware of.

To give an example, consider the rate of a general chemical reaction. The rate of a reaction is the speed at which it occurs or is completed, and naturally depends on the reactiveness of the involved chemicals. However, countless experiments have clearly established that the rate also depends to a strong degree on the concentration (i.e. how many particles there are to react, and, really, how tightly packed they are), temperature (i.e. how much the particles move about spontaneously in a random direction) and pressure (i.e. collision frequency: how often two particles meet) of the compounds. If a gas or liquid reacts with a solid, the size of the solid is also important: one large clump has only a small surface where the other substance can react (it cannot reach below the surface until that goes away), while fine powder has many spots where it can make contact with the fluid. In addition, there are certain substances that speed up a reaction without being directly involved in it; these are called catalysts. Unless you have a special case where you have reasons to suspect that one or two of these factors overshadow all others so that they become negligible (ignorable), you cannot apply binary thinking to chemical reaction rates.

False analogy is the fallacy of (falsely) assuming that because two things are similar in one or more respect they must be similar in another respect. Normally, you would not assume that without a good reason; otherwise you will not be taken seriously. I can think of many silly examples, such as: since squids and knights have mantles, they must both be spineless; some a bit less silly: rocks and tables are hard, so both should burn well; and some more realistic: Joseph Stalin and Fidel Castro were both communist leaders that promoted personal culture, so they must both have been paranoid and conducted mass purges in their countries.

In general, people in my experience have been careful with this (so coming up with examples was actually really hard), perhaps because it is such an easy-to-spot fallacy: you should not even need to set up a syllogism to notice such invalidity in an argument.

Still, we can sometimes be drawn towards such thinking, in particular when we want to extrapolate information from one situation to another. (Recall the meaning of ’extrapolate’ from the beginning of this section.) It is therefore important to be conscious of the risk of committing a fallacy when attempting to make extrapolations.

Circular reasoning is the real pitfall in most rational disciplines. Circular reasoning is the act of assuming the truth of the very thing you intend to prove. It is also known as a vicious circle or begging the question.

In syllogism terms, a circular argument is one where one or more of the premises depend on the conclusion to be true. Thus, the conclusion can only be true if it is true, and the reasoning can only be valid if it is valid. The argument does not show that the conclusion is true. The only way of breaking the circle is by assuming the truth of the conclusion in order to justify one of the premises, but that makes the syllogism, which is designed to demonstrate the truth of the conclusion, utterly redundant since you already have assumed the conclusion to be correct.


It is difficult to think of an example of a formal syllogism with circular reasoning, perhaps because the very structure of a syllogism in some way could be incompatible with circular arguments (an interesting thought indeed). Colloquial examples are easier to recognise, but, again, it is tricky to show formally how they are circular.

An example of a not-so-formal circular syllogism is:

     P: John is a good businessman
     P: John has earned a lot of money through his business
     C: Therefore, John must know how to make good business

So, basically, one of the things that make John a good businessman is that he is a good businessman.

It is called circular reasoning because it goes around and around in its own circle; it leads nowhere. It will not convince many that John makes good business if someone keep using John is a good businessman as an argument. When they ask why he is a good businessman, he/she answers: because he makes good money, and when they ask why is that, he/she repeats: because he is a good businessman.

Such a conversation can be incredibly frustrating if the person in question does not realise his or her argument is circular. Therefore, it would be wise to try to break it down into components and try to arrange it into something like a syllogism, to show why the reasoning is circular and thus invalid.

Although they may appear formally valid, circular arguments are actually not. Their conclusions do not follow from the premises: they are essentially one of their premises, or at least part of one.

Circular arguments are common: I know myself to have made quite many. Luckily, they are not hard to notice – informally or formally. So, as long as you stay attentive to them, and when faced with one, know how to show that it is invalid, and are prepared to attempt to reformulate the argument so that it is no longer circular, you should be just fine.

I have one last fallacy I wish to point out, but I cannot remember its name. It is not really a formal fallacy, so its invalidity cannot be clearly shown in a syllogism. It is when you think you have explained something by giving it a name. For example, say you ask me how the sun warms up the Earth, and I reply: “solar radiation”.

This is not a true logical fallacy, since this type of argument is not always wrong, while logical fallacies are definitely invalid. In this case, whether the question has been answered depends on the asker’s understanding of the term ‘solar radiation’: it is all about whether the person asking the question has enough knowledge on the subject for the explanation to be sufficient. However, in many cases the explanation might not be good enough, but the person who confidently says a few fancy words sounds like he knows what he is talking about, so you accept it rather than look like a fool. Not a good idea: that person might have just as little a clue of what he is saying as you do. There is nothing wrong or shameful in asking for a more comprehensive explanation, especially if it is not your strong subject, so don’t be afraid to ask!


Next

Now you have been introduced to the basics of logical reasoning. Hopefully, you understand and accept that logics aims to produce chains or patterns of reasoning that are always correct, so that if you start with true statements and reason validly, the end product will be just as true. However, what I hope more is that you have asked yourself: ‘well, what does this have to do with science?’ This is a very appropriate question, because science has a quite different way of reasoning!

And that is precisely the point I want to finish this section with: science does not concern itself with reasoning that is conclusive, that cannot be debated. It can be said that science only concerns ideas that can be disputed. For example, the theory of gravity acting on every object on Earth could in theory be disputed if, say, you drop a pen and it does not fall to the ground. An unlikely event, perhaps, but it means that there are ways to question the theory of gravity, and that is one of the qualities that makes the theory scientific: it is testable (more on that in some later post).

Is this bad, or is it for the better? It means that science is never certain, that scientific ‘facts’ always have the possibility to be incorrect. Bad for science. However, it means that science becomes useful as a method to try to deal with those un-provable ideas, a way to handle the uncertainty, as best we can. Good for humanity.

Thus, science is fundamentally rather different from logics. Can they be said to be unrelated? Both are branches of the philosophy of knowledge, but they deal with different aspects of the world. But, in many ways, science strives to be based on logical reasoning. So, in a sense, science uses logics to answer questions that pure logics cannot! One difference lies in the start premises. While pure logics prefers fundamental premises that are true by definition, science deals with premises that are mere generalisations.

This will be the topic of the next section: inductive vs. deductive reasoning






Part Two: Deduction vs. induction


The first part introduced the essence of logics, syllogisms and the concepts of validity and soundness. It was intended to present a fundamental way of thinking philosophically about big words such as truth and certainty, and to emphasise that ‘logical’ does not necessarily mean ‘true’. I would strongly recommend that you read Part One before proceeding. I know it is a long text, but it is also essential to be familiar with the concepts that are discussed there. 

In Part Two, I want to take the discussion a step away from general philosophy and closer to concepts more relevant to science: deductive and inductive reasoning. These are still fundamental to many, if not all branches of philosophy, but inductive reasoning in particular is central to the natural sciences.

Deductive and inductive reasoning are both forms of extrapolation (which was discussed in Part One). Strictly speaking, deduction is the act of reasoning from the general to the particular, while induction is the act of reasoning from the particular to the general. Before you get a headache trying to understand what on Earth that even means, let me talk you through a few simple examples.

Deduction is thinking that if all bananas are sweet, the banana you are about to eat will taste sweet. Induction, on the other hand, would be the reverse: if you eat a banana and find it sweet, you can expect other bananas to be sweet as well. In this case, ‘the general’ refers to all bananas, and ‘the particular’ is the banana you have or ate.

So, knowing that the general (all bananas) has a certain feature or attribute (tasting sweet), you may deduce that the particular (the banana in your hand) might possess that attribute as well. Conversely, observing an attribute (sweet taste) in the particular (the banana you ate) may lead to the inductive conclusion that that feature may belong to the general (all bananas).

If you are observant, you may already be wondering: does induction come before deduction then? How would you know that all bananas are sweet, unless that is a conclusion you have reached inductively by tasting many bananas? Well noted – but do not jump to conclusions! Although that may be true in many cases, there are many exceptions as well. (Trick quiz: was that inductive or deductive thinking?)

The main type of deduction that does not rely on induction is based on definitions. A classical example of something that is true by definition is that all bachelors are unmarried men. This is because the word bachelor means unmarried man. Thus, if you meet someone you know is a bachelor (he may have presented himself as such), you can deduce that he (!) is an unmarried man. This is not because you have met many bachelors and they have all turned out to be unmarried men, but because if he is not an unmarried man, then he is not a bachelor (or… maybe he is lying to you).

You may recall this example from Part One as the first example of a syllogism. I repeat it here because it is not only a sound syllogism, but also one that illustrates deductive reasoning:

     P: All bachelors are unmarried men
     P: Peter is a bachelor
     C: Therefore, Peter is an unmarried man

Here, we go from the general (all bachelors) to the particular (Peter). We know something about the general, something that applies to all of them; and from that, we reason that this something is true also for every particular individual or unit in this general group. We reason from the general to the particular in deduction.

Recall that deduction and induction are inferences, and as such not necessarily true. Again, what we are concerned about is how we reason when we infer. We want the reasoning to be valid, so we can be certain that the deductive or inductive conclusion is as true as the initial premise. In the above example, the first premise is true by definition, and therefore unquestionable (the only grey area would be a widower, a man whose wife has died) – unless you want to challenge the definition, but then it becomes a matter of language, not the nature of the world. The second premise, however, may or may not be true. As a result, the conclusion is only as certain as the second premise. If Peter is not a bachelor, then he is either married or not a man.

Note that that example also can be written as:

     P: All unmarried men are bachelors
     P: Peter is a bachelor
     C: Therefore, Peter is an unmarried man

Or:

     P: All bachelors are unmarried men
     P: Peter is an unmarried man
     C: Therefore, Peter is a bachelor

And so on…

Both attributes (being a bachelor and being an unmarried man) are mutually inclusive. But there are deductive syllogisms that are not true both ways, so to speak. Consider an example:

     P: All tall people wear hats
     P: Gareth is a tall person
     C: Therefore, Gareth wears a hat

This cannot be written as:

     P: All tall people wear hats
     P: Gareth wears a hat
     C: Therefore, Gareth is tall

This is because the key word is not to be. They are not the same. Think about it. The first syllogism claims that all tall people wear hats (a ridiculous statement, perhaps, but let us take it as a true fact for the purpose of this explanation), but that does not imply what the second syllogism states: that all people that wear hats necessarily are tall. In contrast, bachelors are unmarried men, and as a consequence unmarried men are also bachelors!

Hopefully, you now understand the basics of deduction, and recognise that it has important set-backs (think about how much we actually can use deduction for…). Please have them in mind, as I will now explain induction, and later go on to compare deductive and inductive reasoning.

I have been struggling to formulate the following example of inductive reasoning as a sensible syllogism, but I think the message can go through as a simple statement anyway: All swans observed so far are white; therefore, all swans are probably white. This is a generalisation based on observations of the particular (the portion all swans that we have observed), which concludes that the general (all swans that exist) possesses the same traits (colour) as the particular. We reason from the particular to the general.

(As a side note, I might add that this particular example, although classical, becomes problematic when you think about swanlings, which have a dark grey plumage. For that inductive claim to be accurate, swanlings must be excluded from the definition of swans, meaning that sub-adults are not considered members of the species, which further suggests that infants are not humans, and that abortion must be ethical… Woah, maybe I should calm down a bit there, hahaha! Actually, that induction can be improved by specifying that adult swans are white!)

Naturally, you wonder how we can be sure that all swans are white, when we haven’t seen them all. How can we know there are not black, blue or pinkish red swans out there?

Of course, we cannot know that. We can only be fairly certain, or guess that it is so. It is simply a logical impossibility to be absolutely certain about a (non-defining) attribute of a group without having observed every single member.

However, unless we have a reason to suspect that there may be deviations from the observed pattern, we can be quite confident that the pattern will hold even for situations that we have not observed. Still, that is no guarantee, only a probability.

Induction reasoning may be regarded as the core of science. Scientists observe features of the world and then, by working out why these features are the way they are, try to predict what will happen in a similar situation, or, if their understanding is deeper, predict what can happen if the situation changes. This is all based on generalising from a limited number of observations about unobserved events, groups, things, whatever. These generalisations are then usually tested by observing more examples of the unobserved things, and if the generalisations appear true, hooray!

The reason for relying on induction is simply that we cannot observe the whole world. We simply do not have the time nor the labour to do it. If we want to say something about what we have not yet seen, we nearly always have to use inductive reasoning, based on what we have seen so far.

As long as we are all aware of the inescapable problem of uncertainty in induction (deduction may very well be uncertain too, but it is at least possible to be entirely certain of a deductive statement, while absolute confidence in an inductive claim is unfeasible), as long as we treat those conclusions with care, this way of reasoning is amazingly useful, because it allows us to at least approach a truth about something unknown. Deductive reasoning is limited to what we already know – mostly what we know because we defined it that way. We will discuss this more later, in the main comparison between these two ways of thinking. 

However, there is one problematic aspect of induction, one that causes trouble for all branches of science that deal with predicting the future, or understanding the past (e.g. predicting natural disasters such as climate change or volcanic eruptions; predicting evolution and genetic changes, including extinction or survival of species; calculating how much longer our resources will last, and the implications of their over-use; and all sorts of things that lie at the heart of what can be seen as one of the main utilities of science).

To illustrate this fundamental problem with the inductive way of reasoning, let us consider another example: 

P: The sun has risen every day since recorded history
P: Tomorrow is a day
C: Therefore, the sun will probably rise tomorrow

Intuitively, this seems correct, but it is actually not even valid. To make this reasoning valid, we must assume that the world will work the same way tomorrow that it did today and has in the observed past. In other words, we need to assume that the universe works in a consistent manner.

That is a very intuitive claim, since it seems silly to think otherwise: we have little or no evidence to suggest that the world will function differently tomorrow, or any time in the foreseeable future. In the same way, it is against common sense to think that the universal laws of nature were different in a time before our own. Scientific evidence suggests that there was a time when there was no sun, and a time when there was a sun but no Earth. Scientists use other evidence to predict that the sun will collapse one day, if it runs out of ‘fuel’. Hence the word ‘probably’ in the conclusion: it safeguards against the chance that the specific example might change.

Still, what we assume must assume in order to call inductive syllogisms of that nature valid, is that the fundamental laws of the universe – the laws about how things exist and behave – have always been the same, and forever will be. We can accept that particular examples may be different, but the underlying thought is that the basic features of the universe are constant. (I am implying that the sun and Earth are not basic features of the universe.)

But, the problem is, can we really justify that assumption? Based only on what we have seen in the past and present, can we justify the claim that the sun is more likely to rise than not? The painful problem is that the assumption that ‘what has happened every day until now is likely to happen tomorrow as well’ can only be justified if we assume that ‘the world always has and always will work the same way’, which is the very thing we want to show with the first assumption. Thus, we have one assumption that only can be justified by another assumption, which in turn is only justifiable by accepting the first assumption. This circular reasoning is a big bad fallacy, as you may recall from Part One.

I quote a handout from philosophy class:

If nature is uniform and regular in its behaviour, then events in the observed past and present are a sure guide to unobserved events in the unobserved past, present and future. But the only grounds for believing that nature is uniform are the observed events in the past and present. We can’t seem to go beyond the events we observe without assuming the very thing we need to prove – that is, that unobserved parts of the world operate in the same way as the parts we’ve observed.

This is the problem with attempting to extrapolate inductively about the past or future: the necessary assumption cannot be logically justified. Although it seems intuitive, it is actually a fallacy.

Another reason to question the validity of the assumption of the universal laws being constant across time is that scientists also seem to hold the belief that everything is constantly changing. Unless they mean ‘everything except the universal laws is constantly changing’, we are looking at a troublesome inconsistency here (inconsistency, in the philosophical sense, is holding beliefs that contradict one another). The universe cannot be the same and, at the same time, change, since remaining the same is the same thing as not changing, by definition. One of those beliefs must be discarded (except if the constancy solely refers to the universal laws, and the change refers to everything else), which is tricky since both ideas are rather intuitive. We see things changing with time every day, every week, every month, etc. But we also see that many things stay the same. 

Still, even though there are logical reasons for doubt, things that science has accomplished, using those very methods, seem (!) to have hit something right somehow. Although the reasoning seems limping, it appears to have reached the goal. Either, the assumption that the world works in a consistent way just happens to be true (note that circular reasoning does not imply falseness: that an argument is circular only means that is cannot be shown to be true or false, but it has no bearing on whether it is actually true or false – it is either true or false or somewhere in between, but it cannot be demonstrated logically, and is therefore not good reasoning, even though it might be correct, just by luck!), or it is almost true, or it simply does not matter whether it is true or not, and the logical analysis of the flaws of the method is wrong somewhere.

In Notes from Underground, Fyodor Dostoyevsky writes that “man has such a predilection for systems and abstract deductions that he is ready to distort the truth intentionally, he is ready to deny the evidence of his senses only to justify his logic”. I interpret it to be questioning the assumption (really) that our logic has anything to the natural world. How can we know that logics really describes how the world works, and not just the way we humans think? This is a really interesting topic, one which I hope we will come back to later. But, now I think it is about time we get on with comparing deduction and induction.

In general, it can be said that deduction is more certain but less informative, while induction is less certain but more informative. (Or, really, one should say that deduction has greater potential of preserving the certainty of its premises, while a great deal of certainty may be lost through inductive generalisation.)

This is really because of the nature of these ways of reasoning. When you argue from the general to the specific, you just apply what you know about a group to a particular member of that group. If I know that most rocks are hard, I can be fairly sure that if I touch a rock, it will feel hard. In contrast, when you argue from the specific to the general, you are dealing with the uncertainty of whether the feature you are reasoning about applies to the whole group or not. In deduction, you know it applies throughout the group, and can therefore be confident that any particular member will also possess that feature; but, in induction, you cannot be sure that the feature is something that applies to the whole group based on only one or a few particular examples. Although many rocks are grey, it happens to be that colour is not consistent among rocks – there are many rocks that are red, white, beige, green, black, etc., and many have multiple colours!

But, I hope you realise that you cannot reason deductively without knowing anything about the whole group! Therefore, deduction is really not very informative at all. Deduction can only tell you things about particular examples from a broader category of which you already have a lot of information, and the information you apply to these particular examples is already known. Deduction is a way of inferring information from broad category to a narrow part of that or a similar category, rather than discovering new information. It is a tool of inference, not investigation. Deduction is nearly useless in the face of an unknown group of things.

Induction is the method of choice when you want to learn about something new. By making (careful) generalisations about the whole group from a limited sample, it provides a platform for beginning to study this group. These generalisations are usually checked by observing additional examples from the category, and, if the generalisations seem to hold, you keep looking for more particular examples; if you find that there are too many exceptions to your generalisation, you should reconsider or discard that idea and start over, drawing new generalisations from the examples you have now studied. (This systematic approach is not part of the essence of induction, but rather the way it is used effectively in science to learn about the unknown.)

While induction makes it possible to explore, deduction cannot do much more than point.

(But, note well that induction does not allow us to explore completely unknown situations. We always need at least one example of the particular to make generalised or inductive statements. When faced with the complete absence of specific examples, loose inference based on educated guesswork is probably the best we could do. Ideally, these guesses should assume the same predicting nature of inductive generalisations: the guesses should encourage explorers to test the ideas, by finding examples of that unknown, and comparing the guesses with observations.)

As already mentioned before, deduction is often preceded by inductive investigations. Since deduction can only preserve certainty, never increase it, it follows that deductive conclusions made from inductive premises is only as certain as those inductive claims. This is a practical limitation of deduction.

However, induction alone does not take us far. It can help us learn about a category of entities in the world, but that is where it stops. Deduction is one way of making use of information gained from induction, by allowing us to apply the general knowledge to specific situations. If we have observed the consequences of earthquakes and learned about it through induction, we can put this general information to use by applying it deductively on earthquakes we experience now, and better know what to do when they occur and how to act after they have passed.

In other words, deduction and induction seem to come hand in hand, combined, in practice, although they are each others’ opposites, theoretically.

The final point I want to make in this comparison is about the role of language in deduction and induction. In one (philosophical) sense, language is based on inductive generalisations: we implicitly organise the world into general categories by putting names on them, based on common characteristics. We notice that some things have much in common with each other, but still are different from everything else, and call them something unique, e.g. dog, table, teacher. Dogs bark, but neither do tables, teachers or anything else, tables have a specific shape and function that make them tables, etc.

We also implicitly use deductive reasoning by expecting similar behaviour from members of the same ‘thing, which may be distinct from other ‘things’. For example, if you punch a dog, it will probably feel pain, and maybe bite you; if you punch a table, the table might take damage, but will not react, and you will surely hurt yourself; if you punch a teacher, you are in big trouble.

In that sense, by using language, you are implicitly also using deductive and inductive reasoning, often combined.

What is more important, however, is that deduction and induction effectively depend on that the world is organisable into discrete categories, i.e. things/words. What I am trying to argue is that language is not only a way of communicating ideas; language is also a way of organising the world into discrete categories, thus making deductive and inductive logics possible. Deduction is the art of reasoning from such categories to their specific members, and induction is the art of reasoning from specific members to broader categories. If the world is not organisable into discrete groups, such reasoning falls apart.

That is the main reason why many modern taxonomists (scientists who classify living organisms) are struggling with distinguishing different species. The more the biologists learn about the diversity of living organisms, the more they realise that the concept of a ‘species’ is more a made-up grouping than a real, natural unit. There is so much continuous variation within and between ‘species’ that the scientist must make rather arbitrary decisions on how to decide the ‘limit’ between closely related ‘species’. This is just one example of humans trying to impose order on something that apparently is not meant to be organisable.

I admit that this discussion might have been rather confusing, but I hope at least the general message has come across. Scientific research is based mainly on inductive reasoning, or generalisations, basically; scientific inference, on the other hand, relies on deduction, the opposite of induction. Deduction is the art of inferring from the general to the particular, while induction is about generalising from the particular to the general. Induction is less certain that deduction, but much more useful when it comes to investigating the world. Both are often used in combination, however, as they are not very meaningful in isolation.

Part Three will be about practical rather than theoretical issues with inductive reasoning: the limitations of sense perception – the link between our minds and the outside world –  and why we should have the fallibility of our senses close in mind when thinking about science, with its particular emphasis on observing the world.

No comments:

Post a Comment