General propaganda for doing AI and AR over large formal math libraries
Example of today's AI/AR over such libraries
Some future agenda
Minor foundational remarks
Science and Method: Ideas about the interplay between correct deduction and induction/intuition
"And in demonstration itself logic is not all. The true mathematical reasoning is a real induction [...]"
I believe he was right: strong general reasoning engines have to combine deduction and induction (learning patterns from data, making conjectures, etc.)
"There is no actual infinite; the Cantorians have forgotten this, and that is why they have fallen into contradiction."
David Hilbert: "No one will drive us from the paradise which Cantor created for us"
For me, on the larger scale of things, there is only one ultimate hero in what we do (sorry Freek, John, Tom, George, Andrzej)
This hero was born in 1912, the year when Poincare died (24 days before that)
Turing 1936: "On Computable Numbers, with an Application to the Entscheidungsproblem"
Introduced Universal Turing machines (a.k.a. computers)
Halting problem undecidable, therefore FOL undecidable
Unfortunately, for many today's theoretical computer scientists Turing's work ended in 1936:
"It is not possible to have a decision procedure for math, therefore attempting to do math automatically is futile"
1948-1950: Turing wrote the first chess program Turbochamp
Turing 1950: "Computing machinery and intelligence" - beginning of AI, Turing test
On pure deduction: "For at each stage when one is using a logical system, there is a very large number of alternative steps, any of which one is permitted to apply, so far as obedience to the rules of the logical system is concerned. These choices make the difference between a brilliant and a footling reasoner, not the difference between a sound and a fallacious one."
"We may hope that machines will eventually compete with men in all purely intellectual fields."
Turing 1950 (last section on Learning Machines):
"But which are the best ones [fields] to start [learning on] with?"
"... Even this is a difficult decision. Many people think that a very abstract activity, like the playing of chess, would be best."
Me: "Let's use large formal libraries (MML and Flyspeck)" (1998: considered crazy by practically everybody I talked to)
My today's version of Hilbert: "No one will drive us from the semantic AI paradise which large formal libraries created for us"
I don't care much about the foundations, as long as they allow decent formalization and are not too complex to describe
Turing 1936 tells us that we cannot have a decision procedure for math
So anybody trying to automate reasoning has to ultimatelly fail, right?
But how do WE then prove theorems??
Turing: "The fact that a brain can do it seems to suggest that the difficulties [of trying with a machine] may not really be so bad as they now seem." (Penrose of course still disagrees)
Unfortunatelly, the 1936-style CS theoreticians have largely ignored that
John Shawe-Taylor and Nello Cristiani -- Kernel Methods for Pattern Analysis (2004):
Learning high-level and low-level guidance of ATPs, methods for problem characterization, etc.
Implementation methods dealing with large knowledge bases and signatures
ITP-to-ATP translation methods, proof reconstruction methods
"Hammers" (L. Paulson): strong real-time (often cloud-based) AI/ATP for ITP users
Slower "more-AI" experimental systems developing various feedback loops between deduction and induction
Large-theory AI/ATP benchmarks/competitions (CASC LTB) since 2006/2008
Not very much in general, more in some special domains
Most recent eval on Mizar/MML, October 2013: 40% of about 50k Mizar toplevel theorems fully automated
Was 14-16% in 2003 in the first large-scale eval on 33k-big MML
Similar numbers for Flyspeck in 2012 (14k-20k theorems, very recent methods go between 40-50%). No complete eval for Isabelle/AFP.
These numbers look good, but we mostly prove the easier theorems
Average Mizar proof length of ATP-provable theorem is now 10 lines
ML evaluation (premise recall)
ATP evaluation (problems solved)
Combining with SInE improves the best method from 726 to 797 (10%)
MaLARea
The strategies are like giraffes, the problems are their food
The better the giraffe specializes for eating problems unsolvable by others, the more it gets fed and further evolved
In 20 years, 80% of MML and Flyspeck toplevel theorems will be provable automatically (same hardware, same library versions as today - about 40%)
The same in 30 years - I'll give you 2:1, In 10 years: 60%
In 25 years, 50% of the toplevel statements in LaTeX-written Msc-level math curriculum textbooks will be parsed automatically and with correct formal semantics
No betting: all this could be today done in 5 years with reasonable resources - I believe this technology is getting easy ("This problem is data-driven-AI-easy")
Hurry up: I will only accept bets up to 10k EUR total (negotiable)
Disclaimer: I do not understand type theory.
Mizar is to a considerable extent an implementation of little theories (Bill Farmer, IMPS, 1992)
Artur's example: the abstract theory of topological groups
There is one large theory (ZFC/TG) in which all the little theories ultimately live
This plays the role of the explicit consistency layer, we can (or could) use it for doing model theory
And I do not know anything better than ZFC for doing model theory
Guillame's motivating example: if G is isomorphic to H then G is solvable iff H is solvable
This is trivial in model theory: two isomorphic models of L have the same true sentences of L (i.e., they are "elementarily equivalent").
My "correct version" of Guillame's example: if some ultrapowers of G and H are isomorphic then G is solvable iff H is solvable
Why? Keisler-Shelah, 1960s: M and N are elementary equivalent if and only if they have isomorphic ultrapowers.
Of course, "M and N isomorphic" implies that they have some isomorphic ultrapowers, but not vice versa
So this is the right (both sufficient and necessary) condition for two models to exchange sentences.
Isomorphism is only its special case (sufficient, but unnecessary).
The theorem/proof transporting mechanisms in classical systems like HOL, Isabelle, Mizar can be strengthened to this case
Why not also strengthen the Univalence Axiom in this way?
Let A be the graph consisting of a single infinite beaded chain, or more concretely, the integers under adjacency. That is, A=(Z,~), where n~m just in case they differ by exactly one.
And let B consist of two (or more) disconnected copies of A.
It is easy to see that the ultrapower of either A or B by any ultrafilter on a countable index set consists of continuum many such beaded chains.
Thus, the structures A and B have isomorphic ultrapowers, and so they are elementarily equivalent by Keisler-Shelah.
Another: two pseudofinite fields with the same absolute numbers (i.e., the relative algebraic closure of the prime field) are elementarily equivalent.
I hope I have now introduced another schism in type theory:
Certainly, a pure-blood constructivist/type-theorist cannot accept a model-theoretic argument based on ZFC, large cardinals, and non-principal ultrafilters?
Thanks for your attention!
Invited talks: Eric Weisstein (Wolfram), Yves Bertot (INRIA), Freek Wiedijk (RU Nijmegen), Herbert Van de Sompel (Los Alamos Nat. Lab.), Antonio Leal Duarte (U. Coimbra), Jaime Carvalho e Silva (U. Coimbra)
QED+20, July 18 2014 (between ITP and IJCAR): Twenty Years of the QED Manifesto
Invited talks: Michael Beeson (San Jose State U.), Georges Gonthier (Microsoft Research), Adam Grabowski (U. of Bialystok), John Harrison (Intel), Gerwin Klein (NICTA), Magnus Myreen and Ramana Kumar (U. of Cambridge), Claudio Sacerdoti Coen (U. of Bologna)
Please come, in large cardinalities!