How to Make a Tree

 

        There is a major defect with the truth table method for determining whether an argument is valid.  When there are many different letters in an argument, the number of rows in the truth table can be very large.  Since the number of rows doubles with each added letter (for 2 letters - 4 rows, 3 letters - 8 rows, 4 letters - 16 rows, etc..) the table contains more than a thousand rows for an argument with only 10 letters.  The tree method is a variation on the truth table method which can vastly shorten validity calculation.  True, a tree may take as long as the corresponding table in the worst case, but the chances are very good that it will yield massive savings in effort.

        The main idea behind the tree method is to avoid calculating any parts of the table which do not contribute to the main objective in the validity test, namely to find a row where all the premises are T and the conclusion is F.  If there is such a row, the argument is invalid, because validity requires that it is impossible for (all) the premises to be T and the conclusion F.  Let us call a row with (all) T premises and a F conclusion a counterexample.  The idea behind tables is to examine all possible combinations of truth values to see whether there is a counterexample.  If there is, the argument is invalid, and if not, it is valid.  The tree method is more economical than tables because it calculates only those portions of the table where a counterexample can be.  To help illustrate the point, consider this table for the following invalid argument:    Q>-P | P>Q.       

 

       

Note that the third row is a counterexample to the argument - it shows that the argument is invalid because it is possible for the premise to be T and the conclusion F.  If we had some way to know to calculate only this third row, and to ignore the other three, we would have the answer (invalid) with only 1/4 the effort. 

        So let us see whether we can find a method to locate rows where counterexamples will be.  Suppose that the argument  Q>-P |  P>Q does have a counterexample.  Then there must be a row where Q>-P is T and P>Q is F.  But there is only one way for a conditional like P>Q to be F, namely when the antecedent (P) is T and the conclusion (Q) is F.  This means that if there is a counterexample at all, it must be on the row with P = T and Q = F, i.e. the third row.  No other rows could possibly count as a counterexample, because on these rows the conclusion will be T.  At this point we can easily check that the premise Q>-P is indeed T on this row, verifying that row 3 is indeed a counterexample. 

        Notice how this reasoning focuses our attention on the crucial third row, and helps us see why it is pointless to calculate out the other three rows.  By reasoning backwards from the fact that a counterexample requires an F conclusion to the values the letters must have, it may be possible to narrow down a search for a counterexample to one or two rows.  (The same idea may also be used on the premises: knowing that a premise has to be T, also yields information on what values the letters have to have in a counterexample.)

        You may get the impression that this idea only works for invalid arguments where a counterexample exists.  But it works for valid arguments as well.  Let us use P>-Q |  Q>-P to illustrate.  If this argument were to have a counterexample, then there would have to be a row where P>-Q = T and Q>-P = F.  From the latter fact, it follows that the row would have to have Q = T and -P = F.  There is only one row like this, namely where Q = T and P = T.  But if this row is a counterexample it must make P>-Q = T.  However, it does not, since P = T and Q = T makes -Q = F, with the result that P>-Q = F.  The only possible counterexample to our argument is the row where P = T and Q = T.  But we have just seen that it does not make the premise T, so it does not qualify as a counterexample.  It follows that the argument has no counterexample, and so it must be valid.  Notice that even in the case of this valid argument, we only needed to consider calculating out values for one of the four rows.

        So far we have only explained ideas behind the tree method.  It is time to learn the notation and details.  The basic idea is to develop a coordinated search for a counterexample to our argument by using the information that the premises must be T and the conclusion F.  One notation for doing this involves labeling formulas and their parts with Ts and Fs to indicate their values.  However, this can be both conceptually and visually confusing.  Instead, the tree method uses the convention that every sentence we write down is considered to be T.  If we want to indicate that sentence P is F, we simply write down -P instead, (for if -P is T, P must be F).    Let us illustrate this idea by actually building the tree for the argument P>-Q | Q>-P.  We begin with the information that if the argument is to have a counterexample in some row, then P>-Q must be T and Q>-P must be F:

P>-Q

   -(Q>-P)

 

Note that to indicate Q>-P = F, we have written -(Q>-P) instead.  Now if -(Q>-P) is to be T, i.e. Q>-P = F, then there is only one possibility:  Q = T and -P = F.  Entering this information using our convention, we have the following.

 

The circled 1 is there to indicate that we have worked on the second line - the numbers will indicate the order in which we worked.  We have carried out the tree rule for negative arrows.  In general, the rule has the form:

       

        -(A>B)

        A

        -B

 

This indicates that whenever we have -(A>B) in a tree, we should then write down A and -B below it.  This principle is strongly related to the AR (Arrow) rule.  You probably remember the following useful move:  -(A>B) | A&-B.  Notice that the tree rule for a negative arrow amounts to almost the same thing, for from -(A>B), we obtain A&-B by AR, from which A and -B follow by &Out.

        Notice that --P now appears on the tree.  Using the double negation rule:

 

--A

 A

 

we may simplify this to P:

 

        The next step in building our tree corresponds to checking to see whether P>-Q can be T.  To do so, we must introduce the tree rule for positive conditionals.  In general the rule has the following form:

 

 

The branch in the rule indicates that for A>B to be true, there are two (and only two) possibilities: either A must be F or B must be T.  Review the truth table for > to convince yourself that this is correct.  Another way to understand (and remember) this rule is to think of it as a variation on the AR rule for converting from v to >:  A>B | -AvB.  When we apply this principle to our tree, we have the following:

 

 

        The circled numbers show that we have worked on every complex line of our tree.  It is now time to evaluate the argument.  The tree shows that if there is to be a counterexample to our argument, there are two possibilities.  The first, is recorded by the left branch, which contains the simple sentences Q, P and -P.  This indicates (according to our convention) that Q must be T (Q), P must be T (P) and P must be F   (-P).  But it is impossible for P to be both T and F on the same row, so this possibility can be ruled out.  To indicate that this situation is impossible, we indicate that this branch is "dead" by adding the contradiction sign:

 

Notice that the right hand branch also contains a contradiction, for the counterexample it describes requires that Q be both F and T.  So we mark that branch with ƒ as well:

 

   VALID

 

        Let us review what this tree tells us about validity.  It provides a complete record of all the possible ways in which a counterexample can be constructed for P>-Q | Q>-P.  There were two options (described by the left and right branches), but both of these turned out to be impossible.  The verdict: there cannot be a counterexample to this argument.  To put it another way, the argument is valid.

        Now we will work out the tree for Q>-P | P>Q to see what happens with an invalid argument.  We begin by writing down the premise and the negative of the conclusion:

 

Q>-P

-(P>Q)

 

Applying the rule for negative conditionals we obtain:

 

 

Now we apply the rule for positive conditionals to Q>-P:

 

 

Note that the right-hand branch contains a contradiction, so we mark it closed.

 

 

However, the left hand branch is still open.  Since we have worked on every complex sentence, we have a full account of the possibilities for counterexamples for this argument.  The left hand branch reports that there is a counterexample, namely one with P, -Q, and -Q all true.  That amounts to saying that there is a counterexample with P = T and Q = F.  We list this counterexample by putting the values in a box:

 

 

Here is a review of the tree construction process.

1.  Write down the premises and the negative of the conclusion.

2.  Apply tree rules to every complex sentence.

3.  Mark every branch that contains a contradiction with ƒ.

4.  If all branches are closed (contain ƒ), the argument is valid; if any branch is open (no ƒ on it), then this branch describes a counterexample and the argument is invalid.

 

 

        Let's review the process with a more complex argument:  (P&Q)>R, P | Q>R.  First enter the premises and the negative of the conclusion:

 

          (P&Q)>R

P

        -(Q>R)

 

Now apply the negative conditional rule to the last line:

 

 

Now apply the positive conditional rule to the first premise:

 

 

Notice that the right hand branch is closed because it contains -R and R, so we have marked it with ƒ.  Now we must apply a rule to -(P&Q).  The rule for negative conjunctions is similar to De Morgan's Law:  -(A&B) | -Av-B.  If A&B is F then either A is false or B is false:

 

 

Applying this rule to -(P&Q), we obtain:

 

 

Since all branches close, the argument is valid.

        There are still a few details about the tree method that need explaining.  To illustrate, we will work out the tree for PvQ | P&Q.  We begin entering the premise and negative conclusion as usual, and we have already applied the negative & rule to the last step:

Now we must work on PvQ.  It is easy enough to see what rule we want here.  If AvB is T then either A is T or B is T:

The problem is where to put the result of applying this rule.  We have two open branches in our tree, one ending with -P and the other ending with -Q, and neither one is closed.  So where do we place the new branch generated by PvQ:

on the left (below -P) or on the right (below -Q)?  It would be wrong to do either one, and it would also be wrong to divide the results of applying the rule between the two sides like this:

 

        To see what we should do, it helps to remember what the tree represents.  The two branches ending with -P and -Q represent two alternatives: for there to be a counterexample, either P = F or Q = F.  However, the fact that PvQ is T means that on top of these two alternatives there are two more, either P = T or Q = T.  So the correct thing to do is represent a total of four alternatives.  Beneath the alternative -P, we need to record the alternative P or Q, and beneath -Q we must do so as well.  This illustrates a basic principle about building a tree when there is more than one open branch. 

      The results of applying a rule to a line must be entered on every open branch below that line. 

      Following this principle, our tree now looks like this:

Note that contradictions appeared on two of the branches, so we marked them closed:  Since we have worked on every step, the tree is complete and we can evaluate the argument.  Note that there are two open branches indicating two different counterexamples to our argument.  So the argument is invalid.  It is a good idea to calculate the truth table for this argument to verify that it has these two counterexamples.  One nice thing about trees is that they always give a full account of all the counterexamples to an argument.

 

 

        Our next project will be to show that (PvQ)>R | (P>R)&(Q>R) is valid using trees.  We begin by negating the conclusion, and applying the negative & rule:

 

 

Next we will work on the -(P>R) on the left hand branch.  According to the negative > rule we enter P and -R below -(P>R).  It is not necessary to place P and -R below the right hand branch.  Remember the rule is to put the results of applying a rule on every open branch below the formula you are working on.

 

 

After working on -(Q>R) on the right branch, we obtain.

 

 

Now we will work on the first line:  (PvQ)>R.  In this case, we must place the result of applying the rule on all open branches below this line.  So the result goes on both branches:

All that remains is to work on the -(PvQ) on each branch.  The rule for a negative v is as follows:

 

-(AvB)

-A

-B

 

This rule is related to DeMorgan's Law.  The idea is that when AvB is F, then both A and B are F.  Applying this rule on each branch we complete the tree:

 

 

       There is a final detail about tree construction worth mentioning.  You may have been curious about the order in which I have carried out the steps in the trees.  You might want to experiment with the problems we have already done to see what happens if you carry out the steps in a different order.  In each case I have ordered the steps to keep the tree as simple as possible.  There is a simple recipe to help economize in tree construction.  The branching rules tend to create open branches, and subsequent steps must be copied onto all these open branches.  For this reason, it is a good idea to do the non-branching steps first.  Then when branching steps are finally carried out, there is a better chance that the branches created will close.

        As a last note, we must discuss the <> rules.  The rules may appear complex.  But there is a simple way to remember them.

In case A<>|B is T, the values of A and B match.  So either A and B are both T or A and B are both F.  So the positive <> rule indicates these two possibilities on two branches.  In case A<>B is F, then the values of A and B disagree.  So either A is T and B is F or A is F and B is T.  This is why the negative <> rule shows two alternatives, one containing A and -B and the other containing -A and B.  Here is a tree for the argument M<>G | G<>M to help you practice these rules. 

 

Since there are two open branches, the result of applying the negative <> rule to -(M<>G) goes on both sides:

 

Note that all branches are closed.  It is not necessary for there to be a contradiction for M and a contradiction for G to close a branch.  One contradiction on a branch is enough to close it.