zlacker

AI just proved Erdos Problem #124

submitted by nl+(OP) on 2025-11-30 05:19:33 | 242 points 98 comments
[view article] [source] [go to bottom]

https://twitter.com/vladtenev/status/1994922827208663383


NOTE: showing posts with links only show all posts
2. menaer+Nb[view] [source] 2025-11-30 08:02:18
>>nl+(OP)
This seems to be 2nd in row proof from the same author by using the AI models. First time it was the ChatGPT which wrote the formal Lean proof for Erdos Problem #340.

https://arxiv.org/html/2510.19804v1#Thmtheorem3

> In over a dozen papers, beginning in 1976 and spanning two decades, Paul Erdős repeatedly posed one of his “favourite” conjectures: every finite Sidon set can be extended to a finite perfect difference set. We establish that {1, 2, 4, 8, 13} is a counterexample to this conjecture.

6. adt+9g[view] [source] 2025-11-30 09:01:48
>>nl+(OP)
Related, independent, and verified:

GPT-5 solved Erdős problem #848 (combinatorial number theory):

https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8...

https://lifearchitect.ai/asi/

7. espere+5h[view] [source] 2025-11-30 09:13:01
>>nl+(OP)
More interesting discussion than on Twitter here:

https://www.erdosproblems.com/forum/thread/124#post-1892

◧◩
11. demirb+Tn[view] [source] [discussion] 2025-11-30 10:32:13
>>adt+9g
I read how GPT-5 contributed to proof. It is not fully solved by GPT-5 instead assisted. For more look here https://www.math.columbia.edu/~msawhney/Problem_848.pdf
15. consum+0s[view] [source] 2025-11-30 11:28:05
>>nl+(OP)
For reference, this is from: https://harmonic.fun/

related article:

> Is Math the Path to Chatbots That Don’t Make Stuff Up?

https://www.nytimes.com/2024/09/23/technology/ai-chatbots-ch...

◧◩◪
16. clucki+Pt[view] [source] [discussion] 2025-11-30 11:51:00
>>singul+kq
Lean verified a proof of a solution to a problem, but was it the same problem as Erdős problem #124?

https://www.erdosproblems.com/forum/thread/124#post-1899

◧◩
24. aaomid+op1[view] [source] [discussion] 2025-11-30 19:13:20
>>demirb+Zm
I think that person is the owner of the website discussing it too.

He was cited https://the-decoder.com/leading-openai-researcher-announced-...

Where a while back OpenAI made a misleading claim about solving some of these problems.

◧◩
26. dang+jh2[view] [source] [discussion] 2025-12-01 02:24:39
>>ares62+Ib
"Please don't post shallow dismissals, especially of other people's work. A good critical comment teaches us something."

"Don't be curmudgeonly. Thoughtful criticism is fine, but please don't be rigidly or generically negative."

https://news.ycombinator.com/newsguidelines.html

◧◩
27. dang+nh2[view] [source] [discussion] 2025-12-01 02:24:48
>>Comple+Kn
(We detached this subthread from >>46094763 .)
◧◩
29. dang+Zi2[view] [source] [discussion] 2025-12-01 02:37:13
>>espere+5h
Ok, we've changed the link to that from https://twitter.com/vladtenev/status/1994922827208663383, but I've included the latter in the toptext. Thanks!
◧◩◪◨
31. AdieuT+cm2[view] [source] [discussion] 2025-12-01 03:13:14
>>Comple+dj1
>> We've had automated theorem proving since the 60s.

> By that logic, we've had LLMs since the 60s!

From a bit earlier[0], actually:

  Progressing to the 1950s and 60s

  We saw the development of the first language models.
Were those "large"? I'm sure at the time they were thought to be so.

0 - https://ai-researchstudies.com/history-of-large-language-mod...

33. magica+Ao2[view] [source] 2025-12-01 03:40:02
>>nl+(OP)
The overhyped tweet from the robinhood guy raising money for his AI startup is nicely brought into better perspective by Thomas Bloom (including that #124 is not from the cited paper, "Complete sequences of sets of integer powers "/BEGL96):

> This is a nice solution, and impressive to be found by AI, although the proof is (in hindsight) very simple, and the surprising thing is that Erdos missed it. But there is definitely precedent for Erdos missing easy solutions!

> Also this is not the problem as posed in that paper

> That paper asks a harder version of this problem. The problem which has been solved was asked by Erdos in a couple of later papers.

> One also needs to be careful about saying things like 'open for 30 years'. This does not mean it has resisted 30 years of efforts to solve it! Many Erdos problems (including this one) have just been forgotten about it, and nobody has seriously tried to solve it.[1]

And, indeed, Boris Alexeev (who ran the problem) agrees:

> My summary is that Aristotle solved "a" version of this problem (indeed, with an olympiad-style proof), but not "the" version.

> I agree that the [BEGL96] problem is still open (for now!), and your plan to keep this problem open by changing the statement is reasonable. Alternatively, one could add another problem and link them. I have no preference.[2]

Not to rain on the parade out of spite, it's just that this is neat, but not like, unusually neat compared to the last few months.

[1] https://twitter.com/thomasfbloom/status/1995083348201586965

[2] https://www.erdosproblems.com/forum/thread/124#post-1899

35. newman+9q2[view] [source] 2025-12-01 03:59:50
>>nl+(OP)
"Recently, yet another category of low-hanging fruit has been identified as within reach of automated tools: problems which, due to a technical flaw in their description, are unexpectedly easy to resolve. Specifically, problem #124 https://www.erdosproblems.com/124 was a problem that was stated in three separate papers of Erdos, but in two of them he omitted a key hypothesis which made the problem a consequence of a known result (Brown's criterion). However, this fact was not noticed until Boris Alexeev applied the Aristotle tool to this problem, which autonomously located (and formalized in Lean) a solution to this weaker version of the problem within hours."

https://mathstodon.xyz/@tao/115639984077620023

◧◩
38. NooneA+Hu2[view] [source] [discussion] 2025-12-01 04:56:24
>>magica+Ao2
reading the original paper and the lean statement that got proven, it's kinda fascinating what exactly is considered interesting and hard in this problem

roughly, what lean theorem (and statement on the website) asks is this: take some numbers t_i, for each of them form all the powers t_i^j, then combine all into multiset T. Barring some necessary conditions, prove that you can take subset of T to sum to any number you want

what Erdosh problem in the paper asks is to add one more step - arbitrarily cut off beginnings of t_i^j power sequences before merging. Erdosh-and-co conjectured that only finite amount of subset sums stop being possible

"subsets sum to any number" is an easy condition to check (that's why "olympiad level" gets mentioned in the discussion) - and it's the "arbitrarily cut off" that's the part that og question is all about, while "only finite amount disappear" is hard to grasp formulaically

so... overhyped yes, not actually erdos problem proven yes, usual math olympiad level problems are solvable by current level of Ai as was shown by this year's IMO - also yes (just don't get caught by https://en.wikipedia.org/wiki/AI_effect on the backlash since olympiads are haaard! really!)

◧◩
78. jvande+od4[view] [source] [discussion] 2025-12-01 17:32:26
>>harshi+Tk3
The most amazing thing about AI is how it's amazing and disappointing at the same time.

Everything is amazing and nobody is happy: https://www.youtube.com/watch?v=nUBtKNzoKZ4

"How quickly the world owes him something, he knew existed only 10 seconds ago"

79. kelsey+2o4[view] [source] 2025-12-01 18:21:07
>>nl+(OP)
An Erdos problem is great, but all major LLMs still score 0 out of 7 on all Millenium Prize Problems[1]. I, and I'm sure other like me, won't consider it "true" AI until AI saturate the MillenniumPrizeProblemBench.

1. https://www.claymath.org/millennium-problems/

1.

◧◩◪
87. vatsac+AQ5[view] [source] [discussion] 2025-12-02 02:06:25
>>emil-l+mG2
In the sense that in 15 years my bet is that an AI system can solve my thesis problem for under 100$

https://ems.press/journals/jems/articles/14298293

[go to top]