Proof by contradiction (original) (raw)

From Wikipedia, the free encyclopedia

Proving that the negation is impossible

In logic, proof by contradiction is a form of proof that establishes the truth or the validity of a proposition by showing that assuming the proposition to be false leads to a contradiction. Although it is quite freely used in mathematical proofs, not every school of mathematical thought accepts this kind of nonconstructive proof as universally valid.[1]

More broadly, proof by contradiction is any form of argument that establishes a statement by arriving at a contradiction, even when the initial assumption is not the negation of the statement to be proved. In this general sense, proof by contradiction is also known as indirect proof, proof by assuming the opposite,[2] and reductio ad impossibile.[3]

A mathematical proof employing proof by contradiction usually proceeds as follows:

  1. The proposition to be proved is P.
  2. We assume P to be false, i.e., we assume ¬P.
  3. It is then shown that ¬P implies falsehood. This is typically accomplished by deriving two mutually contradictory assertions, Q and ¬Q, and appealing to the law of noncontradiction.
  4. Since assuming P to be false leads to a contradiction, it is concluded that P is in fact true.

An important special case is the existence proof by contradiction: in order to demonstrate that an object with a given property exists, we derive a contradiction from the assumption that all objects satisfy the negation of the property.

The principle may be formally expressed as the propositional formula ¬¬P ⇒ P, equivalently (¬P ⇒ ⊥) ⇒ P, which reads: "If assuming P to be false implies falsehood, then P is true."

In natural deduction the principle takes the form of the rule of inference

⊢ ¬ ¬ P ⊢ P {\displaystyle {\cfrac {\vdash \lnot \lnot P}{\vdash P}}} {\displaystyle {\cfrac {\vdash \lnot \lnot P}{\vdash P}}}

which reads: "If ¬ ¬ P {\displaystyle \lnot \lnot P} {\displaystyle \lnot \lnot P} is proved, then P {\displaystyle P} {\displaystyle P} may be concluded."

In sequent calculus the principle is expressed by the sequent

Γ , ¬ ¬ P ⊢ P , Δ {\displaystyle \Gamma ,\lnot \lnot P\vdash P,\Delta } {\displaystyle \Gamma ,\lnot \lnot P\vdash P,\Delta }

which reads: "Hypotheses Γ {\displaystyle \Gamma } {\displaystyle \Gamma } and ¬ ¬ P {\displaystyle \lnot \lnot P} {\displaystyle \lnot \lnot P} entail the conclusion P {\displaystyle P} {\displaystyle P} or Δ {\displaystyle \Delta } {\displaystyle \Delta }."

In classical logic the principle may be justified by the examination of the truth table of the proposition ¬¬P ⇒ P, which demonstrates it to be a tautology:

p ¬p ¬¬p ¬¬p ⇒ p
T F T T
F T F T

Another way to justify the principle is to derive it from the law of the excluded middle, as follows. We assume ¬¬P and seek to prove P. By the law of excluded middle P either holds or it does not:

  1. if P holds, then of course P holds.
  2. if ¬P holds, then we derive falsehood by applying the law of noncontradiction to ¬P and ¬¬P, after which the principle of explosion allows us to conclude P.

In either case, we established P. It turns out that, conversely, proof by contradiction can be used to derive the law of excluded middle.

In classical sequent calculus LK proof by contradiction is derivable from the inference rules for negation:

Γ , P ⊢ P , Δ ( I ) Γ , ⊢ ¬ P , P , Δ ( ¬ R ) Γ , ¬ ¬ P ⊢ P , Δ ( ¬ L ) {\displaystyle {\cfrac {{\cfrac {{\cfrac {\ }{\Gamma ,P\vdash P,\Delta }}\;(I)}{\Gamma ,\vdash \lnot P,P,\Delta }}\;({\lnot }R)}{\Gamma ,\lnot \lnot P\vdash P,\Delta }}\;({\lnot }L)} {\displaystyle {\cfrac {{\cfrac {{\cfrac {\ }{\Gamma ,P\vdash P,\Delta }}\;(I)}{\Gamma ,\vdash \lnot P,P,\Delta }}\;({\lnot }R)}{\Gamma ,\lnot \lnot P\vdash P,\Delta }}\;({\lnot }L)}

Relationship with other proof techniques

[edit]

Refutation by contradiction

[edit]

Proof by contradiction is similar to refutation by contradiction,[4][5] also known as proof of negation, which states that ¬P is proved as follows:

  1. The proposition to be proved is ¬P.
  2. Assume P.
  3. Derive falsehood.
  4. Conclude ¬P.

In contrast, proof by contradiction proceeds as follows:

  1. The proposition to be proved is P.
  2. Assume ¬P.
  3. Derive falsehood.
  4. Conclude P.

Formally these are not the same, as refutation by contradiction applies only when the proposition to be proved is negated, whereas proof by contradiction may be applied to any proposition whatsoever.[6] In classical logic, where P {\displaystyle P} {\displaystyle P} and ¬ ¬ P {\displaystyle \neg \neg P} {\displaystyle \neg \neg P} may be freely interchanged, the distinction is largely obscured. Thus in mathematical practice, both principles are referred to as "proof by contradiction".

Law of the excluded middle

[edit]

Proof by contradiction is equivalent to the law of the excluded middle, first formulated by Aristotle, which states that either an assertion or its negation is true, P ∨ ¬P.

Law of non-contradiction

[edit]

The law of noncontradiction was first stated as a metaphysical principle by Aristotle. It posits that a proposition and its negation cannot both be true, or equivalently, that a proposition cannot be both true and false. Formally the law of non-contradiction is written as ¬(P ∧ ¬P) and read as "it is not the case that a proposition is both true and false". The law of non-contradiction neither follows nor is implied by the principle of Proof by contradiction.

The laws of excluded middle and non-contradiction together mean that exactly one of P and ¬P is true.

Proof by contradiction in intuitionistic logic

[edit]

In intuitionistic logic proof by contradiction is not generally valid, although some particular instances can be derived. In contrast, proof of negation and principle of noncontradiction are both intuitionistically valid.

Brouwer–Heyting–Kolmogorov interpretation of proof by contradiction gives the following intuitionistic validity condition: if there is no method for establishing that a proposition is false, then there is a method for establishing that the proposition is true.[_clarify_]

If we take "method" to mean algorithm, then the condition is not acceptable, as it would allow us to solve the Halting problem. To see how, consider the statement H(M) stating "Turing machine M halts or does not halt". Its negation ¬H(M) states that "M neither halts nor does not halt", which is false by the law of noncontradiction (which is intuitionistically valid). If proof by contradiction were intuitionistically valid, we would obtain an algorithm for deciding whether an arbitrary Turing machine M halts, thereby violating the (intuitionistically valid) proof of non-solvability of the Halting problem.

A proposition P which satisfies ¬ ¬ P ⇒ P {\displaystyle \lnot \lnot P\Rightarrow P} {\displaystyle \lnot \lnot P\Rightarrow P} is known as a ¬¬-stable proposition. Thus in intuitionistic logic proof by contradiction is not universally valid, but can only be applied to the ¬¬-stable propositions. An instance of such a proposition is a decidable one, i.e., satisfying P ∨ ¬ P {\displaystyle P\lor \lnot P} {\displaystyle P\lor \lnot P}. Indeed, the above proof that the law of excluded middle implies proof by contradiction can be repurposed to show that a decidable proposition is ¬¬-stable. A typical example of a decidable proposition is a statement that can be checked by direct computation, such as " n {\displaystyle n} {\displaystyle n} is prime" or " a {\displaystyle a} {\displaystyle a} divides b {\displaystyle b} {\displaystyle b}".

Examples of proofs by contradiction

[edit]

An early occurrence of proof by contradiction can be found in Euclid's Elements, Book 1, Proposition 6:[7]

If in a triangle two angles equal one another, then the sides opposite the equal angles also equal one another.

The proof proceeds by assuming that the opposite sides are not equal, and derives a contradiction.

Hilbert's Nullstellensatz

[edit]

An influential proof by contradiction was given by David Hilbert. HisNullstellensatz states:

If f 1 , … , f k {\displaystyle f_{1},\ldots ,f_{k}} {\displaystyle f_{1},\ldots ,f_{k}} are polynomials in n indeterminates with complex coefficients, which have no common complex zeros, then there are polynomials g 1 , … , g k {\displaystyle g_{1},\ldots ,g_{k}} {\displaystyle g_{1},\ldots ,g_{k}} such that f 1 g 1 + … + f k g k = 1. {\displaystyle f_{1}g_{1}+\ldots +f_{k}g_{k}=1.} {\displaystyle f_{1}g_{1}+\ldots +f_{k}g_{k}=1.}

Hilbert proved the statement by assuming that there are no such polynomials g 1 , … , g k {\displaystyle g_{1},\ldots ,g_{k}} {\displaystyle g_{1},\ldots ,g_{k}} and derived a contradiction.[8]

Infinitude of primes

[edit]

Euclid's theorem states that there are infinitely many primes. In Euclid's Elements the theorem is stated in Book IX, Proposition 20:[9]

Prime numbers are more than any assigned multitude of prime numbers.

Depending on how we formally write the above statement, the usual proof takes either the form of a proof by contradiction or a refutation by contradiction. We present here the former, see below how the proof is done as refutation by contradiction.

If we formally express Euclid's theorem as saying that for every natural number n {\displaystyle n} {\displaystyle n} there is a prime bigger than it, then we employ proof by contradiction, as follows.

Given any number n {\displaystyle n} {\displaystyle n}, we seek to prove that there is a prime larger than n {\displaystyle n} {\displaystyle n}. Suppose to the contrary that no such p exists (an application of proof by contradiction). Then all primes are smaller than or equal to n {\displaystyle n} {\displaystyle n}, and we may form the list p 1 , … , p k {\displaystyle p_{1},\ldots ,p_{k}} {\displaystyle p_{1},\ldots ,p_{k}} of them all. Let P = p 1 ⋅ … ⋅ p k {\displaystyle P=p_{1}\cdot \ldots \cdot p_{k}} {\displaystyle P=p_{1}\cdot \ldots \cdot p_{k}} be the product of all primes and Q = P + 1 {\displaystyle Q=P+1} {\displaystyle Q=P+1}. Because Q {\displaystyle Q} {\displaystyle Q} is larger than all prime numbers it is not prime, hence it must be divisible by one of them, say p i {\displaystyle p_{i}} {\displaystyle p_{i}}. Now both P {\displaystyle P} {\displaystyle P} and Q {\displaystyle Q} {\displaystyle Q} are divisible by p i {\displaystyle p_{i}} {\displaystyle p_{i}}, hence so is their difference Q − P = 1 {\displaystyle Q-P=1} {\displaystyle Q-P=1}, but this cannot be because 1 is not divisible by any primes. Hence we have a contradiction and so there is a prime number bigger than n {\displaystyle n} {\displaystyle n}

Examples of refutations by contradiction

[edit]

The following examples are commonly referred to as proofs by contradiction, but formally employ refutation by contradiction (and therefore are intuitionistically valid).[10]

Infinitude of primes

[edit]

Let us take a second look at Euclid's theorem – Book IX, Proposition 20:[9]

Prime numbers are more than any assigned multitude of prime numbers.

We may read the statement as saying that for every finite list of primes, there is another prime not on that list, which is arguably closer to and in the same spirit as Euclid's original formulation. In this case Euclid's proof applies refutation by contradiction at one step, as follows.

Given any finite list of prime numbers p 1 , … , p n {\displaystyle p_{1},\ldots ,p_{n}} {\displaystyle p_{1},\ldots ,p_{n}}, it will be shown that at least one additional prime number not in this list exists. Let P = p 1 ⋅ p 2 ⋯ p n {\displaystyle P=p_{1}\cdot p_{2}\cdots p_{n}} {\displaystyle P=p_{1}\cdot p_{2}\cdots p_{n}} be the product of all the listed primes and p {\displaystyle p} {\displaystyle p} a prime factor of P + 1 {\displaystyle P+1} {\displaystyle P+1}, possibly P + 1 {\displaystyle P+1} {\displaystyle P+1} itself. We claim that p {\displaystyle p} {\displaystyle p} is not in the given list of primes. Suppose to the contrary that it were (an application of refutation by contradiction). Then p {\displaystyle p} {\displaystyle p} would divide both P {\displaystyle P} {\displaystyle P} and P + 1 {\displaystyle P+1} {\displaystyle P+1}, therefore also their difference, which is 1 {\displaystyle 1} {\displaystyle 1}. This gives a contradiction, since no prime number divides 1.

Irrationality of the square root of 2

[edit]

The classic proof that the square root of 2 is irrational is a refutation by contradiction.[11]Indeed, we set out to prove the negation ¬ ∃ a, b ∈ N {\displaystyle \mathbb {N} } {\displaystyle \mathbb {N} } . a/b = √2 by assuming that there exist natural numbers a and b whose ratio is the square root of two, and derive a contradiction.

Proof by infinite descent

[edit]

Proof by infinite descent is a method of proof whereby a smallest object with desired property is shown not to exist as follows:

Such a proof is again a refutation by contradiction. A typical example is the proof of the proposition "there is no smallest positive rational number": assume there is a smallest positive rational number q and derive a contradiction by observing that ⁠q/2⁠ is even smaller than q and still positive.

Russell's paradox, stated set-theoretically as "there is no set whose elements are precisely those sets that do not contain themselves", is a negated statement whose usual proof is a refutation by contradiction.

Proofs by contradiction sometimes end with the word "Contradiction!". Isaac Barrow and Baermann used the notation Q.E.A., for "quod est absurdum" ("which is absurd"), along the lines of Q.E.D., but this notation is rarely used today.[12] A graphical symbol sometimes used for contradictions is a downwards zigzag arrow "lightning" symbol (U+21AF: ↯), for example in Davey and Priestley.[13] Others sometimes used include a pair of opposing arrows (as → ← {\displaystyle \rightarrow \!\leftarrow } {\displaystyle \rightarrow \!\leftarrow }[_citation needed_] or ⇒ ⇐ {\displaystyle \Rightarrow \!\Leftarrow } {\displaystyle \Rightarrow \!\Leftarrow }),[_citation needed_] struck-out arrows ( ↮ {\displaystyle \nleftrightarrow } {\displaystyle \nleftrightarrow }),[_citation needed_] a stylized form of hash (such as U+2A33: ⨳),[_citation needed_] or the "reference mark" (U+203B: ※),[_citation needed_] or × × {\displaystyle \times \!\!\!\!\times } {\displaystyle \times \!\!\!\!\times }.[14][15]

G. H. Hardy described proof by contradiction as "one of a mathematician's finest weapons", saying "It is a far finer gambit than any chess gambit: a chess player may offer the sacrifice of a pawn or even a piece, but a mathematician offers the game."[16]

Automated theorem proving

[edit]

In automated theorem proving the method of resolution is based on proof by contradiction. That is, in order to show that a given statement is entailed by given hypotheses, the automated prover assumes the hypotheses and the negation of the statement, and attempts to derive a contradiction.[17]

  1. ^ Bishop, Errett 1967. Foundations of Constructive Analysis, New York: Academic Press. ISBN 4-87187-714-0
  2. ^ "Proof By Contradiction". www2.edc.org. Retrieved 12 June 2023.
  3. ^ "Reductio ad absurdum | logic". Encyclopedia Britannica. Retrieved 25 October 2019.
  4. ^ "Proof by contradiction". nLab. Retrieved 7 October 2022.
  5. ^ Richard Hammack, Book of Proof, 3rd edition, 2022, ISBN 978-0-9894721-2-8; see "Chapter 9: Disproof".
  6. ^ Bauer, Andrej (29 March 2010). "Proof of negation and proof by contradiction". Mathematics and Computation. Retrieved 26 October 2021.
  7. ^ "Euclid's Elements, Book 6, Proposition 1". Retrieved 2 October 2022.
  8. ^ Hilbert, David (1893). "Ueber die vollen Invariantensysteme". Mathematische Annalen. 42 (3): 313–373. doi:10.1007/BF01444162.
  9. ^ a b "Euclid's Elements, Book 9, Proposition 20". Retrieved 2 October 2022.
  10. ^ Bauer, Andrej (2017). "Five stages of accepting constructive mathematics". Bulletin of the American Mathematical Society. 54 (3): 481–498. doi:10.1090/bull/1556.
  11. ^ Alfeld, Peter (16 August 1996). "Why is the square root of 2 irrational?". Understanding Mathematics, a study guide. Department of Mathematics, University of Utah. Retrieved 6 February 2013.
  12. ^ "Math Forum Discussions".
  13. ^ B. Davey and H.A. Priestley, Introduction to Lattices and Order, Cambridge University Press, 2002; see "Notation Index", p. 286.
  14. ^ Gary Hardegree, Introduction to Modal Logic, Chapter 2, pg. II–2. https://web.archive.org/web/20110607061046/http://people.umass.edu/gmhwww/511/pdf/c02.pdf
  15. ^ The Comprehensive LaTeX Symbol List, pg. 20. http://www.ctan.org/tex-archive/info/symbols/comprehensive/symbols-a4.pdf
  16. ^ G. H. Hardy, A Mathematician's Apology; Cambridge University Press, 1992. ISBN 9780521427067. PDF p.19 Archived 2021-02-16 at the Wayback Machine.
  17. ^ "Linear Resolution", From Logic to Logic Programming, The MIT Press, pp. 93–120, 1994, doi:10.7551/mitpress/3133.003.0007, ISBN 978-0-262-28847-7, retrieved 21 December 2023

[edit]