Finally completed the #Lean4 formalization of my Maclaurin-type inequality paper at https://github.com/teorth/symmetric_project . Specifically, if
is established for
Was surprisingly satisfied to get the blue "No goals" message in the Lean infoview.
In the end it took perhaps a hundred hours of effort, spread out over a month. It was about 20x slower to formalize this paper than it was to write it in LaTeX. Initially this was due to my own inexperience with Lean, though near the end I was hitting the limitations of existing Lean tools, as well as the speed of the Lean compiler (though this was in part due some inefficiencies in my coding). However, Lean is a living language still in development, and several helpful people provided me during this project with upgraded Lean tactics (such as an improved `positivity` tactic that could handle finite sums and products, a `rify` tactic to easily convert natural number inequalities to real number ones, and a `rw_ineq` tactic to implement the rewriting of inequalities I had noted was previously missing). These definitely helped expedite the formalization process, particularly in the last week or so. 1/3
Happily, apart from the one bug previously noted, no other errors (other than very small typos) were detected in the paper. I found some mild simplifications to some steps in the proof as a consequence of formalization, but nothing particularly notable. (But this was a rather elementary paper, and I was not really expecting to glean deep mathematical insights from the formalization process.)
The Sapir-Whorf hypothesis https://en.wikipedia.org/wiki/Linguistic_relativity is controversial for natural languages, but for the specific case of proof formalization languages I think the effect is real. The Lean syntax and library of lemmas and tools certainly pushed me to plan ahead and abstract away many small lemmas to a far greater extent than I would ordinarily do; long "stream of consciousness" type arguments, which I tend to favor personally, are possible in Lean, but often not the most efficient way to proceed. I have also started noticing myself mentally converting results I read in regular mathematical literature to a sort of "Lean pseudocode", though this is probably just a version of the "Tetris effect" https://en.wikipedia.org/wiki/Tetris_effect and will likely dissipate now that my project has concluded.
A combination of automated tools, AI tools, and interactions with human Lean experts proved invaluable. GPT greatly sped up the initial learning of syntax, tools such as `exact?`, `apply?`, and Moogle https://moogle-morphlabs.vercel.app/ helped locate key lemmas, and Copilot had an often uncanny ability to autocomplete one or more lines of code at a time. But there was certainly a lot of "folklore" tips and tricks that I had to go to the Lean Zulip to get answers to. 2/3
I found that documenting all the places where I encountered some problem, and recording the solution, was helpful, not just as a text to refer to whenever encountering a similar issue in the future, but also just to organize my thoughts and consolidate my understanding of various subtle aspects of Lean. (This document can be found at https://docs.google.com/spreadsheets/d/1Gsn5al4hlpNc_xKoXdU6XGmMyLiX4q-LFesFVsMlANo ).
While formalization is generally much slower than writing in Mathematical English, I found one aspect in which the formalized writing process was far superior. If, halfway during the writing of a long argument, I found it necessary to tweak some parameter or hypothesis at the very beginning of the argument, the Lean compiler would quickly identify all the other parts of the argument that might need changing, and it was a fairly quick matter to modify the proof to get back to a working state. In contrast, when performing these dozens of small edits in Mathematical English, it is all too common to introduce one or more small errors which tended to be quite tedious to detect and eradicate. In many cases, if sufficiently "smart" tactics were employed, many fewer edits would be required in Lean as compared to Mathematical English; Lean can often propagate changes in hypotheses automatically to the natural changes in conclusions through its advanced ability to fill in "holes".
Would I attempt to formalize more mathematical papers any time soon? Probably not as a solo project; it was significantly more time consuming than expected. But I particularly enjoyed the (unplanned) collaborative aspects, and I could see myself contributing to a larger Lean collaboration in the future. (3/3)
@tao Thanks for this. Your spreadsheet notes are very helpful.
@tao What is the ratio of the sizes of the zipped lean files in your project to the zipped .tex file(s) for the human writeup? I don't know if there's a good delineation between what you should and shouldn't include in the former, I leave the jugdement to you. That ratio will estimate the de Bruijn factor (https://www.cs.ru.nl/~freek/factor/factor.pdf) for your proofs.
@highergeometer I'm not sure that raw file sizes are that good of a measure, but my LaTeX source for the paper is 37KB uncompressed while the essential files for the project (removing newton.lean and maclaurin.lean) add up to a bit over 110KB, plus some additional tactic files that were contributed. It is not evenly distributed; the last three pages of my argument required about 45KB of Lean to formalize. The size expansion of 3x or so is significantly less than the time expansion of 20x I experienced; a reasonable goal perhaps is to have additional development tools to bring the latter factor down to more closely resemble the former.
@tao Yes, it's certainly not a measure of the formalisation effort, but it measures how close the language/libraries is to the mathematical English vernacular. One might imagine that if mathlib had already had a number of basic results you needed to prove yourself, and the tactic automation a year or two more developed, that the proof could have been even shorter. The inhomogeneity of the paper proof:formal proof ratio certainly reflects this.
The recent blog post at Xena Project about the difference between the Ramsey-theoretic formal proof of Mehta and the core of the Liquid Tensor Project also shows this changing ratio (https://xenaproject.wordpress.com/2023/11/04/formalising-modern-research-mathematics-in-real-time/, the section "5 years, 50 pages, 5 months")
@tao Dear Professor Tao, Thank you! Sorry for writing here, but I would like to clarify the steps
"We can thus take expectations and conclude that..."
and "...and hence by (27) we have...", just above Equation (44) in the logarithmically averaged Chowla paper.
Should we sum over
in the representation
where
are the errors?
Is
sufficient?
@consensus This part of the argument does not require the fact that
@tao Thanks! Shouldn't we factor the logarithmic weight
@consensus No. The connection to the original two-point correlation was already worked out in Section 2, where it was calculated that (2.4) implies (2.17). To complete the proof, it suffices to show that (2.17) leads to a contradiction; no further discussion of the original two-point correlation is needed.