*In some better world, Ukrainian women aren't getting blasted from the sky for ten years, but are just having kids while doing lots of award-winning math about extradimensionality #autoformalization en.wikipedia.org/wiki/Maryna_...

Maryna Viazovska - Wikipedia
Maryna Viazovska - Wikipedia

Diary of #autoformalization

The methodology and approaches are not too far from how my brain works or what would be the output if I would record every of my thoughts on such a problem.

But it also becomes clear that claude code is a beginner who does lots of trial-and-error and copy-paste coding.

fly51fly (@fly51fly)

J. Urban의 논문은 단기간(2주) 동안 13만 줄 규모의 형식적 위상수학(formal topology)을 자동 형식화(autoformalization)로 생성한 작업을 보고합니다. 비용과 복잡도를 낮춘 간단한 방법을 제안해 누구나 자동 형식화에 접근할 수 있게 하는 접근법과 실험 결과를 제시하며 정리된 데이터셋과 파이프라인을 공개합니다 (arXiv:2601.03298).

https://x.com/fly51fly/status/2013735633425146080

#autoformalization #formalization #theoremproving #automatedreasoning

fly51fly (@fly51fly) on X

[LG] 130k Lines of Formal Topology in Two Weeks: Simple and Cheap Autoformalization for Everyone? J Urban [AI4REASON] (2026) https://t.co/pGBD5M2ThS

X (formerly Twitter)
Readings shared January 14, 2026

The readings shared in Bluesky on 14 January 2026 are: A lambda-superposition tactic for Isabelle/HOL. ~ Massin Guerdi. #ITP #IsabelleHOL Adding sorts to an Isabelle formalization of superposition. ~

Vestigium
130k lines of formal topology in two weeks: Simple and cheap autoformalization for everyone? ~ Josef Urban. https://arxiv.org/abs/2601.03298v1 #ITP #Mizar #LLMs #Math #Autoformalization
130k Lines of Formal Topology in Two Weeks: Simple and Cheap Autoformalization for Everyone?

This is a brief description of a project that has already autoformalized a large portion of the general topology from the Munkres textbook (which has in total 241 pages in 7 chapters and 39 sections). The project has been running since November 21, 2025 and has as of January 4, 2026, produced 160k lines of formalized topology. Most of it (about 130k lines) have been done in two weeks,from December 22 to January 4, for an LLM subscription cost of about \$100. This includes a 3k-line proof of Urysohn's lemma, a 2k-line proof of Urysohn's Metrization theorem, over 10k-line proof of the Tietze extension theorem, and many more (in total over 1.5k lemmas/theorems). The approach is quite simple and cheap: build a long-running feedback loop between an LLM and a reasonably fast proof checker equipped with a core foundational library. The LLM is now instantiated as ChatGPT (mostly 5.2) or Claude Sonnet (4.5) run through the respective Codex or Claude Code command line interfaces. The proof checker is Chad Brown's higher-order set theory system Megalodon, and the core library is Brown's formalization of basic set theory and surreal numbers (including reals, etc). The rest is some prompt engineering and technical choices which we describe here. Based on the fast progress, low cost, virtually unknown ITP/library, and the simple setup available to everyone, we believe that (auto)formalization may become quite easy and ubiquitous in 2026, regardless of which proof assistant is used.

arXiv.org
Readings shared December 5, 2025

The readings shared in Bluesky on 5 December 2025 are: Formalization of Erdős problems. ~ Boris Alexeev. #ITP #LeanProver #Math #Autoformalization A Rocq formalization of monomial and graded orders.

Vestigium
Reseña de «Formalization of Erdős problems»

El artículo «Formalization of Erdős problems» destaca el auge en la resolución de problemas de Paul Erdős gracias a la web erdosproblems.com y el proyecto Formal Conjectures. De más de 1100 problemas

Vestigium
Formalization of Erdős problems

[This is a guest post by Boris Alexeev. Now over to Boris.] I’m here to tell you about various exciting developments centering on Erdős problems, especially involving the formalization of old and n…

Xena
Readings shared November 18, 2025

The readings shared in Bluesky on 18 November 2025 are: Teaching real analysis as a game. ~ Alex Kontorovich. #ITP #LeanProver #Math Are automated proof assistants ready for semigroup research? Orien

Vestigium
Evaluating autoformalization robustness via semantically similar paraphrasing. ~ Hayden Moore, Asfahan Shah. https://arxiv.org/abs/2511.12784 #AI #Math #Autoformalization #ITP #LeanProver
Evaluating Autoformalization Robustness via Semantically Similar Paraphrasing

Large Language Models (LLMs) have recently emerged as powerful tools for autoformalization. Despite their impressive performance, these models can still struggle to produce grounded and verifiable formalizations. Recent work in text-to-SQL, has revealed that LLMs can be sensitive to paraphrased natural language (NL) inputs, even when high degrees of semantic fidelity are preserved (Safarzadeh, Oroojlooyjadid, and Roth 2025). In this paper, we investigate this claim in the autoformalization domain. Specifically, we evaluate the robustness of LLMs generating formal proofs with semantically similar paraphrased NL statements by measuring semantic and compilation validity. Using the formal benchmarks MiniF2F (Zheng, Han, and Polu 2021) and Lean 4 version of ProofNet (Xin et al. 2024), and two modern LLMs, we generate paraphrased natural language statements and cross-evaluate these statements across both models. The results of this paper reveal performance variability across paraphrased inputs, demonstrating that minor shifts in NL statements can significantly impact model outputs.

arXiv.org