Ensenar logica: romper la nuez dura.
Teaching logic is doubly difficult. First of all, it is as rigorous as any other analytical study, e.g., mathematics. Secondly, it is not easy to convince your students that logic is important. Every student knows that they have to take their mathematical lessons seriously. But many of us have to face the following question quite frequently in our logic class: "Why should we learn logic?" I think a logic course should begin with an attempt to answer this question. And very often teachers do face another problem in a logic class: when they start teaching the basic rules, such as Modus Ponens (if p then q; p; therefore q) some of their students would definitely ask "Are those rules to be learned? We know all of that anyway". In the beginning the logic lessons seem to be deceptively naive. Then you start Predicate Calculus. Students will have a great amount of difficulties understanding the restrictions imposed on the rules. Some will fail to understand them and hence soon lose interest in learning logic any more. In this article, I shall basically share my teaching experience with those of my colleagues who teach logic. I will try to address the following distinct but mutually related questions: 1) Why should one take logic seriously?; 2) How should one handle the problems that are faced while teaching senior secondary students symbolic logic?
P. T. Geach (1979) tells us that medieval writers used to make a distinction between logica utens, "the practice of thinking logically about this or that subject-matter", and logica docens, "the construction of logical theory" (p. 5). There are two ways of learning: mechanical and cordial. Mechanical learning is algorithmic; "you do this, then you do that, and then. ... and you solve the problem". No understanding is involved in this process. Cordial learning means "understanding something". In the developing world, most of the education systems encourage the mechanical process. It is easy to handle, as well as effective--temporarily, of course. On the other hand, cordial learning demands more time and involvement; it allows time for the click of understanding to emerge. Any analytical science is based on arguments. Mechanical learning does not try to see the argumentative threads in a scientific discourse; it helps solving problems. What helps us understand something or learn something cordially is logica utens.
It is not difficult to discover that "logica utens needs the aid of logica docens" (Geach, 1979, p. 6). Why? Logic is based on both intuition and technique. It is a map of our intelligence. The basic laws such as Modus Ponens correspond to the basic moves of intelligence. They are "intuitively" right for (almost) everybody.
When one sees a long argument one may not have an intuitive judgment about its (in)correctness. If one is logically trained then one knows how to split it into smaller pieces. If each piece is intuitively right, i.e., if each piece is a licensed logical move, then the argument is valid. Had logic been completely intuitive, there would have been no space for proving a theorem with heuristic devices. In that case, one should be able to respond to an argument intuitively just by looking at it. The logician (who is trained in logica docens) knows how to cut down a long argument into smaller, intelligible pieces. Analysis is nothing but this "logical splitting". It is not completely mechanical, since the minimum logical moves, i.e., the basic rules are always intuitive. One has to make decisions while solving a problem, because there are infinite options.
Epp (2003) and Bako (2002) observe that many students of pure mathematics cannot write proofs properly; it is mainly because they fail to see the logical moves that underlie the steps of a mathematical proof. Following my previous statements, logic pictures the moves of intelligence and creates a flowchart that helps a learner understand how the (n + 1)th line of a proof (or argument) follows from the nth line. In that sense, logic simulates intelligence.
Our logica docens is still not able to analyze all intuitive procedures. Still it does not mechanically trace the steps of all the arguments that seem to be perfectly valid. But logica docens is enriched every day, like all other disciplines. The main point is that logic aims to understand and record even the most minuscule intuitive moves.
Teaching logic at high schools
I have found that a few techniques help high-school students understand symbolic logic in a non-mechanical way. This discussion will be confined to propositional and predicate calculi. I'm sure these techniques can be further extended.
Natural science has an advantage over logic; the former is empirical to a great extent. You have something in flesh and bone to show to your younger students. But the laboratory of logic is our own mind. How can I demonstrate to my students the universal acceptability of the basic logical rules? Suppose I have to teach them Modus Ponens (p [right arrow] q, p; [therefore] q) and Modus Tolens (p [right arrow] q, ~ q; [therefore] ~ p). Most of them seem to understand those. But what they do not understand is what do those rules look like? I normally do two things.
I exemplify the rules and compare them with their stupid counterparts namely (p [right arrow] q, q; [therefore] p) and (p [right arrow] q, ~ p; [therefore] ~ q) respectively. The students do immediately realize that (p [right arrow] q, q; [therefore] p) is not a rule. But many of them think that (p [right arrow] q, ~ p; [therefore] ~ q) is a rule. They argue: "If it rains then the soil gets wet; it is not raining; therefore the soil is not getting wet. Yes, this is fine". I tell them, "Maybe the soil is wet because you are pouring water on the ground. It need not rain". On the other hand, just see: if it is true that "If it rains then soil gets wet", then it can't rain when the soil is not wet". Normally it works. Little by little, students start appreciating the difference between a rule and a non-rule.
By extending the same technique I teach them the rules of Existential Instantiation (EI) and Universal Generalization (UG) through a demonstration of fallacies. Here I'm discussing the Kalish-Montague version of the rules (1964, pp. 118-122). First the UG rule. Suppose I want to prove that ANY triangle x is such that the sum of its angles is 180[degrees]. Here I must make sure that x is a randomly chosen triangle about which I know just one thing: it has the property of being a triangle. If I already know that it is an equilateral one, then I shall end up proving that "ANY equilateral triangle x is such that the sum of its angles is 180[degrees]", which is definitely not the thing I wanted to prove originally. That means, when I want to prove that "for any x, if x is a triangle [Tx] then it has the property of having three angles whose sum is 180[degrees] [Sx]" or Vx (Tx [right arrow] Sx), I have to just that "if x is T then x is S [Tx [right arrow] Sx]", and I must not know anything about x before I prove that Tx [right arrow] Sx. This has been translated into the technical language of logic in the following way: Prove Fx for proving [for all]xFx and see that x never occurs freely before proving Fx. The underlined part is the restriction on the rule. The point here is the following: When I state that [for all]xFx or [there exists]xFx, I mean to say that "everything is F" or "something is F" respectively. In my interpretation there is no space for a variable. But when I state "Fx", I am saying that "x is F". This x, which appears in just "Fx" is a real variable whereas the x, which appears under the scope of a quantifier (as in [for al]xFx or [there exists]xFx), is a pseudo-variable. Note: in the following formula, the first three occurrences of x are within the scope of the quantifier "[for all]" and, hence are bound, while the last occurrence is free: [for all]x (Fx [conjunction] Gx) [disjunction] Fx [the free occurrence is underlined]. It is clear by now that only a free variable is a real one. Thus, I state that "x is an individual such that the predicate F is true of x" only when I say just "Fx".
Let us see what may happen when we violate the restriction on UG. Suppose my premise is [there exists]xGx (which may be interpreted at "something is good"). From that, we may show [for all]xGx ("everything is good"). The violation of the restriction will allow us to draw this unwelcome conclusion:
[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]
The variable of UG, i.e., y occurs free in line 3 which is before "Show Gy".
Roughly the ei rule is the following: [there exists] xGx. [therefore] Gy [where y is a new variable].
The idea is to express something along these lines: "I know that there is at least one good thing [i.e., 3xGx; interpret Ga as "a is good"]. From that point, when I conclude that y is good, I'm just naming a good individual "y".
Hence, "y" must be a new variable because I should not know anything else about y. Suppose no such restriction is imposed on EI. And we know that By and 3xGx. From [there exists]xGx we may conclude that Gy. Now I can say that "y is B and G [By a Gy]", from which we conclude that "there exists some x such that it is both B and G [[there exists]x (Bx a Gx)]". But my premises just say that "some individual y is bad [By]" and "there exists at least one individual which is good [[there exists]xGx]". They do not say that "there is at least one individual x which is at the same time bad and good [[there exists]x (Bx [conjunction] Gx)]". My conclusion cannot claim anything more than what the premises state. This overstatement is due to the violation of the restriction. Even "Gx" cannot be drawn from [there exists]xGx. Consider the following wrong derivation:
[MATHEMATICAL EXPRESSION NOT REPRODUCIBLE IN ASCII]
The purpose of both the restrictions is to prevent x from being the variable of Universal Generalization when x is used as the variable of Existential Instantiation. In Problem II, x in line 4 comes as a result of EI applied on line 1. And the same x is the variable of UG in line 2. This is undesired. Allowing this would illicitly generalize every specific case. It is evident that without any restriction both of the subsequent derivations can take place anywhere in a proof:
UI: [for all]xGx; Gx or Gy, and EG: Gx; [therefore] [there exists]xGx or [there exists]yGy.
If [for all]xGx, i.e., "everything is G" is true then "x is G" or "y is G". Any variable can fill in the blank in "_ is G". If Gx, i.e., "x is G" is true, then it must be true that "something is G" [[there exists]xGx]. It does not matter which variable appears in the symbolic formula as long as the former is bound.
It is often said that the process of translating ordinary sentences into Predicate Logic (pl) is a lot more complicated than into Propositional Logic. Let us discuss a few translation issues that may trouble learners in the beginning phases. We know that "all humans are mortal" is translated as [for all]x (Hx [right arrow] Mx) and "some human is mortal" as [there exists]x (Hx [conjunction] Mx). This is taught with the assumption that everybody understands the translations. But the assumption is not always right. Many learners do not understand why these should be translated this way only.
According to PL, "all H are M" means that "for every x, if it is H then it is M [[for all]x (Hx [right arrow] Mx)]". You choose anything; if it is not a human (if it is a dog for instance) then you do not have to check any further. If it is a human, then it must be mortal. That means this sentence does not allow any non-mortal human. Funny enough! This translation has no issues with a world that has absolutely no humans; for it says that if there is a human ... etc. On the other hand, "some H is M" means "there exists at least one x, such that it is H and M". You check the things that exist. If at least one of them is both H and M, then this sentence is true. That means this sentence allows non-mortal humans. It is not compatible with a world in which there is no H or M, because it asserts that 1) both the set of humans and the set of mortals are nonempty, and 2) their intersection too is non-empty, i.e., they have at least one common member. We shall see what happens if we translate these sentences differently.
Suppose we do the following: "some human is mortal" = [for all]x (Hx [right arrow] Mx). This is wrong because the translation says that there exists some x such that if it is a human then it is mortal. That means, the translation has no problems if there are no humans. It is compatible with both the world that has no humans and the world that has some mortal humans as well. But the sentence "some human is mortal" claims that there must be at least one mortal human. It is not compatible with the world without humans. So this translation is wrong. Now let us consider, "all humans are mortal" = [for all]x (Hx [conjunction] Mx). This is wrong because the translation makes everything a mortal human; there is space for nothing else. What is wrong with "all humans are mortal" = [there exists]x (Hx [right arrow] Mx)? The sentence, according to pl, says that "if something is a human then it is mortal". The translation too seems to say something similar, i.e., there exists some x such that if it is a human then it is mortal. i.e., no non-mortal human is entertained by the translation. Then what is wrong? The problem is [there exists]x (Hx [right arrow] Mx) is happy when there is at least one individual such that if it is a human then it is mortal. It imposes no upper limit on the number of x. If all humans are mortal, then that is also right. But at the same time, it allows non-mortal humans too, since [there exists]x (Hx [right arrow] Mx) is satisfied even with one mortal human in the world in which both the set of humans and the set of mortal beings are nonempty. It does not mind if all other humans are immortal in that very world. So it does not translate "all humans are mortal". The following table describes the worlds associated with propositions and their proper or improper symbolic representations:
All H are M Some H is M [for all]x(Hx [right arrow] Mx) Satisfied if Satisfied if there Satisfied if every every H is M is something which H is M is both H and M Does not allow Allows non-M H Does not allow any any non-M H non-M H Allows the Does not allow Allows the empty empty set of H the empty set of set of H H or M [there exists]x(Hx [for all]x(Hx [there exists]x(Hx [conjunction] Mx) [conjunction] Mx) [right arrow] Mx) Satisfied if there Satisfied if Satisfied if there is something which everything is is at least one is both H and M both H and M mortal human when the set of H is non-empty Allows non-M H Does not allow Allows non-M H any non-M H or non-H M Does not allow the Does not allow Allows the empty empty set of H or M the empty set of set of H H or M
The system-specific logical interpretations of natural-language sentences may not always seem to be intuitively proper. But we must understand that logic is a model of the operations of our intelligence. Like any other model, the model of logic must be compact. The stoic logic represents Propositional Calculus well. I personally prefer Aristotelian Logic to modern Predicate Logic, for I think that Aristotle's approach is more intuitive than that of PL. For example, when somebody says "all humans are mortal", they normally mean that both the set of humans and the set of mortal beings are nonempty and the former is a subset of the latter. If you ask the speaker whether their universe of discourse can have the empty set of humans, they will probably say No. In sum, there is no implication in their sentence. But in the predicate logical interpretation, there is an implication. Thus both "all unicorns are white" and "all humans are mortal" are interpreted in the same way [[for all]x (Ux [right arrow] Wx) and [for all]x (Hx [right arrow] Mx) respectively], and share the same truth-value "True". But no naive speaker would perhaps say that they are equally true. The unicorn-sentence is true because the set of unicorns is empty. Thus there are gaps between our intuitive grasp and logical representations. But, for the sake of compactness, one may accept the modern mathematical model in which you have propositional axioms and rules at the basic level; you add a few more axioms and rules and you get predicate calculus. With some additional rules, this will give you Modal Calculus, and so on. Thus you have a grand system of logic corresponding to the unitary mind that does logical calculations of several types.
Bako, M. (2002). Why we need to teach logic and how can we teach it? International Journal for Mathematics Teaching and Learning, (October, ISSN 1473-0111.). Available at: http://www.cimt.plymouth.ac.uk/journal/bakom.pdf
Epp, S. S. (2003). The role of logic in teaching proof. American Mathematical Monthly (December).
Geach, P. T. (1979). On teaching logic. Philosophy, 54(207), 5-17.
Kalish, D., and Montague, R. (1964). Logic: Techniques of formal reasoning. New York, NY: Oxford University Press.
Indian Institute of Technology (IIT), India
|Printer friendly Cite/link Email Feedback|
|Title Annotation:||ALEPH; articulo en Ingles|
|Date:||Jan 1, 2014|
|Previous Article:||Los bachilleratos bivalentes, las estructuras organizativas y la incorporacion de las competencias genericas en sus curriculos.|
|Next Article:||Validacion de la escala de motivacion de logro escolar (EME-E) en estudiantes de bachillerato en Mexico.|