發現的幻覺:AI 生成的「開放式」數學難題證明

發現的幻覺:AI 生成的「開放式」數學難題證明

Hacker News·

本文探討了 AI,特別是像 GPT 5.2 Pro 這樣的語言模型,在解決複雜的開放式數學難題方面的影響。文章質疑 AI 是否真正發現了新的數學真理,還是僅僅變得更擅長引用現有知識。

The Illusion of Discovery: AI-Generated Proofs of 'Open' Math Problems

This article evaluates the practical impact of AI on open math problems and outlines the current community consensus on its technical strengths and limitations.

Image

Do you use AI as a source for:

Generating novel scientific ideas for open research problems?

Reviewing scientific drafts?

Literature review?

Then this article is for you.

Thanks for reading! Subscribe for free to receive new posts and support my work.

A year ago, the idea of an LLM executing multi-step complex proofs was considered a distant goal. Today, they can can function as sophisticated research assistants. But as AI begins 'solving' open math problems like those posed by Paul Erdős, we have to ask: is it discovering new truths, or just reading the map better than we do?

The Breakthrough

Ever since GPT 5.2 Pro was released last month, AI-generated proofs of open math problems are taking off. According to an active contributor who volunteers in verifying AI-written proof of Erdős problems, GPT 5.2 Pro currently is mathematically the strongest model and can actually outperform “independent” checks by Gemini/Claude. Recently, a post on X went viral declaring that they have solved a second Erdős problem (#281) using only GPT 5.2 Pro.

Image

The Reality Check

So, what was the aftermath? In less than 24 hours, a prior proof from literature was located and it turned out that GPT had only discovered new proof of a non-open problem. Is that impressive? Absolutely. Infact, that inspired me to find and submit a solution for an open Erdős problem. Does that mean AGI is already knocking on our doors? Before we answer that question, let’s dive into their wiki page of the AI-contributions.

Image

As shown in Figure 1, we could see that only 2 problems have actually been solved completely by AI without any existence of prior literature review out of which 1 solution was achieved by applying an existing theorem.

Image

When trying to assess the success rate of these tools, reporting bias should be kept in mind. The share of the ‘Failure/No Progress’ category in Figure 1 should be much higher than expected due to underreporting of failures to solve such problems. Quoting from its wiki page: “We have found empirically that problems that could be solved fully by AI without the benefit of prior literature or human interaction tend to be simpler and have shorter proofs than problems that required a combination of AI, human input, and past literature to solve.”

Image

In other cases, a subsequent literature review revealed the problem already had a partial solution or alternate full solutions as shown in Figure 3. In only 1 case was the AI able to generate a full solution with an existing partial solution from past literature. It is also difficult to say whether the AI solution was arrived at independently or was it genuinely influenced by prior literature. At best, we can say that AI is currently better at "re-synthesizing" existing mathematical knowledge than venturing into "uncharted territory."

According to Terence Tao, “AI-written proofs have improved to the point of being human-readable and technically correct, but still feel slightly ‘off’, for instance sometimes stressing easy statements and being overly brief on core aspects of the argument”. AI often treats a foundational axiom and a complex lemma with the same weight. The point 5 on the disclaimer section in the wiki stresses on how a problem solved by an AI tool must be received:

“If an Erdős problem was first posed N years ago, and was recently solved by an AI tool, it may be misleading to make announcements such as “this problem resisted all human efforts at proof for N years” in order to imply that the problem is particularly difficult. Such an announcement may be technically true, especially if a comprehensive literature review turns up no significant prior literature on the problem, but it is also possible that this particular question received very little attention from human mathematicians during this time period, and the absence of any past progress may therefore be more a reflection of the obscurity of the problem than of its difficulty. (If on the other hand there is substantial literature expending significant effort on establishing partial results, without generating a clear path to a full solution, this is more convincing evidence of difficulty than mere absence of literature.)”

The Researcher’s Toolkit

Here are some AI-pitfalls that they have observed:

Self-fulfills the problem’s status as open based on available literature

Gets confused with closely related problems

Lack of remarks and observations unlike human-generated proofs

AI-math slop/Hallucination

Weak performance for problems in other subdomain (eg. geometric)

Usefulness of the solution

So before you take the output of an AI tool seriously for research, they recommend to consider the following checklist for rigor before announcing your solution. This guide also applies to the non-experts in the field and in extension other fields of research as AI-tools have reduced the barriers for entry and participation:

Do I understand why this question was asked, and why the hypotheses and conclusions were formulated the way they were?

Have I performed my own literature review on the problem? How does this problem relate to other results and questions in the field?

Do I understand what the key ideas of the solution are, and how the hypotheses are utilized to reach the conclusion?

How does the proof method compare to that in past literature? What was the key new insight that allowed this solution to succeed when previous methods failed?

Can I generate a formalized version of the proof?

Did I submit the AI-generated proof to a separate AI tool and ask it to critically evaluate the proof for correctness?

AI deep research tools have been successful in locating solutions to some previously “open” Erdős problems. It can find a needle in a haystack of existing literature that humans have forgotten. However, it isn’t yet an architect capable of building entirely new mathematical frameworks. Based on that reality, Mr. Tao suggests that a more realistic and productive use of AI tools proposed in research is to crowd source literature review.

Conclusion

AI is excellent at re-synthesizing obscured knowledge, which is a massive win for productivity, even if it isn’t AGI. We are entering an era of "Citizen Mathematics."

Sources:

Wiki: AI contributions to Erdős problems

erdosproblems.com: TerrenceTao

erdosproblems.com: natso26

Thanks for reading! Subscribe for free to receive new posts and support my work.

Image

Ready for more?

Hacker News

相關文章

  1. AI模型開始破解高等數學難題

    Techcrunch · 3 個月前

  2. 業餘數學家藉助AI解決長期懸而未決的厄多什數學難題

    3 個月前

  3. AI 是否正在解決未解的 Erdős 問題?

    3 個月前

  4. GPT-5.2 在理論物理學上取得新成果

    2 個月前

  5. 邁向自主數學研究

    2 個月前