How to Make a Tree
There is a major defect with the truth table method for determining whether an argument 
is valid.  When there are many different letters in an argument, the number of rows 
in the truth table can be very large.  Since the number of rows doubles with each 
added letter (for 2 letters - 4 rows, 3 letters - 8 rows, 4 letters - 16 rows, etc..) 
the table contains more than a thousand rows for an argument with only 10 letters. 
 The tree method is a variation on the truth table method which can vastly shorten 
validity calculation.  True, a tree may take as long as the corresponding table in the 
worst case, but the chances are very good that it will yield massive savings in effort.
The main idea behind the tree method is to avoid calculating any parts of the table 
which do not contribute to the main objective in the validity test, namely to find 
a row where all the premises are T and the conclusion is F.  If there is such a row, 
the argument is invalid, because validity requires that it is impossible for (all) the 
premises to be T and the conclusion F.  Let us call a row with (all) T premises and 
a F conclusion a counterexample
.  The idea behind tables is to examine all possible combinations of truth values 
to see whether there is a counterexample.  If there is, the argument is invalid, 
and if not, it is valid.  The tree method is more economical than tables because 
it calculates only those portions of the table where a counterexample can be.  To help illustrate 
the point, consider this table for the following invalid argument:    Qç~P   PçQ.

Note that the third row is a counterexample to the argument - it shows that the argument 
is invalid because it is possible for the premise to be T and the conclusion F.  
If we had some way to know to calculate only this third row, and to ignore the other 
three, we would have the answer (invalid) with only 1/4 the effort.  
So let us see whether we can find a method to locate rows where counterexamples will 
be.  Suppose that the argument  Qç~P   PçQ does have a counterexample.  Then there 
must be a row where Qç~P is T and PçQ is F.  But there is only one way for a conditional like PçQ to be F, namely when the antecedent (P) is T and the conclusion (Q) is 
F.  This means that if there is a counterexample at all, it must be on the row with 
P = T and Q = F, i.e. the third row.  No other rows could possibly count as a counterexample, because on these rows the conclusion will be T.  At this point we can easily check 
that the premise Qç~P is indeed T on this row, verifying that row 3 is indeed a counterexample. 
 
Notice how this reasoning focuses our attention on the crucial third row, and helps 
us see why it is pointless to calculate out the other three rows.  By reasoning backwards 
from the fact that a counterexample requires an F conclusion to the values the letters must have, it may be possible to narrow down a search for a counterexample to 
one or two rows.  (The same idea may also be used on the premises: knowing that a 
premise has to be T, also yields information on what values the letters have to have 
in a counterexample.)
You may get the impression that this idea only works for invalid arguments where a 
counterexample exists.  But it works for valid arguments as well.  Let us use Pç~Q 
  Qç~P to illustrate.  If this argument were to have a counterexample, then there 
would have to be a row where Pç~Q = T and Qç~P = F.  From the latter fact, it follows that 
the row would have to have Q = T and ~P = F.  There is only one row like this, namely 
where Q = T and P = T.  But if this row is a counterexample it must make Pç~Q = T. 
 However, it does not, since P = T and Q = T makes ~Q = F, with the result that Pç~Q 
= F.  The only possible counterexample to our argument is the row where P = T and 
Q = T.  But we have just seen that it does not make the premise T, so it does not 
qualify as a counterexample.  It follows that the argument has no counterexample, and so it 
must be valid.  Notice that even in the case of this valid argument, we only needed 
to consider calculating out values for one of the four rows.
So far we have only explained ideas behind the tree method.  It is time to learn the 
notation and details.  The basic idea is to develop a coordinated search for a counterexample 
to our argument by using the information that the premises must be T and the conclusion F.  One notation for doing this involves labeling formulas and their parts 
with Ts and Fs to indicate their values.  However, this can be both conceptually 
and visually confusing.  Instead, the tree method uses the convention that every 
sentence we write down is considered to be T.  If we want to indicate that sentence P is 
F, we simply write down ~P instead, (for if ~P is T, P must be F).    Let us illustrate 
this idea by actually building the tree for the argument Pç~Q   Qç~P.  We begin with 
the information that if the argument is to have a counterexample in some row, then 
Pç~Q must be T and Qç~P must be F:
Pç~Q
   ~(Qç~P)
Note that to indicate Qç~P = F, we have written ~(Qç~P) instead.  Now if ~(Qç~P) is 
to be T, i.e. Qç~P = F, then there is only one possibility:  Q = T and ~P = F.  Entering 
this information using our convention, we have the following.

The circled 1 is there to indicate that we have worked on the second line - the numbers 
will indicate the order in which we worked.  We have carried out the tree rule for 
negative arrows.  In general, the rule has the form:
~(Åç )
Å
~ 
This indicates that whenever we have ~(Åç ) in a tree, we should then write down Å and ~  below it.  This principle is strongly related to the AR (Arrow) rule.  You probably 
remember the following useful move:  ~(Åç )   Å&~ .  Notice that the tree rule for a negative arrow amounts to almost the same thing, 
for from ~(Åç ), we obtain Å&~  by AR, from which Å and ~  follow by &Out. 
Notice that ~~P now appears on the tree.  Using the double negation rule:
~~Å
 Å 
we may simplify this to P:

The next step in building our tree corresponds to checking to see whether Pç~Q can 
be T.  To do so, we must introduce the tree rule for positive conditionals.  In general 
the rule has the following form:

The branch in the rule indicates that for Åç  to be true, there are two (and only two) possibilities: either Å must be F or   must be T.  Review the truth table for ç to convince yourself that this is correct. 
 Another way to understand (and remember) this rule is to think of it as a variation 
on the AR rule for converting from   to ç:  Åç    ~Å  .  When we apply this principle to our tree, we have the following: 

The circled numbers show that we have worked on every complex line of our tree.  It 
is now time to evaluate the argument.  The tree shows that if there is to be a counterexample 
to our argument, there are two possibilities.  The first, is recorded by the left branch, which contains the simple sentences Q, P and ~P.  This indicates (according 
to our convention) that Q must be T (Q), P must be T (P) and P must be F (~P).  But 
it is impossible for P to be both T and F on the same row, so this possibility can 
be ruled out.  To indicate that this situation is impossible, we indicate that this 
branch is "dead" by adding the contradiction sign: 

Notice that the right hand branch also contains a contradiction, for the counterexample 
it describes requires that Q be both F and T.  So we mark that branch with   as well:
 VALID
   VALID
Let us review what this tree tells us about validity.  It provides a complete record 
of all the possible ways in which a counterexample can be constructed for Pç~Q   
Qç~P.  There were two options (described by the left and right branches), but both 
of these turned out to be impossible.  The verdict: there cannot be a counterexample to 
this argument.  To put it another way, the argument is valid.
Now we will work out the tree for Qç~P   PçQ to see what happens with an invalid argument. 
 We begin by writing down the premise and the negative
 of the conclusion:
Qç~P
~(PçQ)
Applying the rule for negative conditionals we obtain:

Now we apply the rule for positive conditionals to Qç~P:

Note that the right-hand branch contains a contradiction, so we mark it closed.

However, the left hand branch is still open.  Since we have worked on every complex 
sentence, we have a full account of the possibilities for counterexamples for this 
argument.  The left hand branch reports that there is a counterexample, namely one 
with P, ~Q, and ~Q all true.  That amounts to saying that there is a counterexample with 
P = T and Q = F.  We list this counterexample by putting the values in a box:

Here is a review of the tree construction process.
-   Write down the premises and the negative
 of the conclusion.
 
-   Apply tree rules to every complex sentence.
 
-   Mark every branch that contains a contradiction with  .
 
-   If all branches are closed (contain  ), the argument is valid; if any branch is 
open (no   on it), then this branch describes a counterexample and the argument is 
invalid. 
 
 
 
Let's review the process with a more complex argument:  (P&Q)çR, P   QçR.  First enter 
the premises and the negative of the conclusion:
          (P&Q)çR
P
        ~(QçR)
Now apply the negative conditional rule to the last line:

Now apply the positive conditional rule to the first premise:

Notice that the right hand branch is closed because it contains ~R and R, so we have 
marked it with  .  Now we must apply a rule to ~(P&Q).  The rule for negative conjunctions 
is similar to De Morgan's Law:  ~(Å& )   ~Å ~ .  If Å&  is F then either Å is false or   is false:

Applying this rule to ~(P&Q), we obtain:

Since all branches close, the argument is valid.
There are still a few details about the tree method that need explaining.  To illustrate, 
we will work out the tree for P Q   P&Q.  We begin entering the premise and negative 
conclusion as usual, and we have already applied the negative & rule to the last step:

Now we must work on P Q.  It is easy enough to see what rule we want here.  If Å   is T then either Å is T or   is T:

The problem is where to put the result of applying this rule.  We have two open branches 
in our tree, one ending with ~P and the other ending with ~Q, and neither one is 
closed.  So where do we place the new branch generated by P Q:

on the left (below ~P) or on the right (below ~Q)?  It would be wrong to do either 
one, and it would also be wrong to divide the results of applying the rule between 
the two sides like this:

To see what we should do, it helps to remember what the tree represents.  The two 
branches ending with ~P and ~Q represent two alternatives: for there to be a counterexample, 
either P = F or Q = F.  However, the fact that P Q is T means that on top of
 these two alternatives there are two more, either P = T or Q = T.  So the correct 
thing to do is represent a total of four
 alternatives.  Beneath the alternative ~P, we need to record the alternative P or 
Q, and beneath ~Q we must do so as well.  This illustrates a basic principle about 
building a tree when there is more than one open branch.  
The results of applying a rule to a line must be entered on every open branch
 below that line.  
Following this principle, our tree now looks like this:

Note that contradictions appeared on two of the branches, so we marked them closed: 
 Since we have worked on every step, the tree is complete and we can evaluate the 
argument.  Note that there are two open branches indicating two different counterexamples 
to our argument.  So the argument is invalid.  It is a good idea to calculate the 
truth table for this argument to verify that it has these two counterexamples.  One 
nice thing about trees is that they always give a full account of all the counterexamples 
to an argument.
Our next project will be to show that (P Q)çR   (PçR)&(QçR) is valid using trees. 
 We begin by negating the conclusion, and applying the negative & rule:

Next we will work on the ~(PçR) on the left hand branch.  According to the negative 
ç rule we enter P and ~R below ~(PçR).  It is not necessary to place P and ~R below the right hand branch.  
Remember the rule is to put the results of applying a rule on every open branch below the formula you are working on
.

After working on ~(QçR) on the right branch, we obtain.

Now we will work on the first line:  (P Q)çR.  In this case, we must place the result 
of applying the rule on all open branches below this line.  So the result goes on 
both branches:

All that remains is to work on the ~(P Q) on each branch.  The rule for a negative 
  is as follows:
~(Å  )
~Å
~ 
This rule is related to DeMorgan's Law.  The idea is that when Å   is F, then both Å and   are F.  Applying this rule on each branch we complete the tree:

       There is a final detail about tree construction worth mentioning.  You may 
have been curious about the order in which I have carried out the steps in the trees. 
 You might want to experiment with the problems we have already done to see what 
happens if you carry out the steps in a different order.  In each case I have ordered the 
steps to keep the tree as simple as possible.  There is a simple recipe to help economize 
in tree construction.  The branching rules tend to create open branches, and subsequent steps must be copied onto all these open branches.  For this reason, it is a 
good idea to do the non-branching steps first
.  Then when branching steps are finally carried out, there is a better chance that 
the branches created will close.
As a last note, we must discuss the   rules.  The rules may appear complex.  But there 
is a simple way to remember them.

In case Å   is T, the values of Å and   match.  So either Å and   are both T or Å and   are both F.  So the positive   rule indicates these two possibilities on two branches.  In case Å   is F, then the values of Å and   disagree.  So either Å is T and   is F or Å is F and   is T.  This is why the negative   rule shows two alternatives, one containing Å and ~  and the other containing ~Å and  .  Here is a tree for the argument M G   G M to help you practice these rules.  

Since there are two open branches, the result of applying the negative   rule to ~(M G) 
goes on both sides:

Note that all branches are closed.  It is not necessary for there to be a contradiction 
for M and a contradiction for G to close a branch.  One contradiction on a branch is enough to close it.