Superhuman mathematical problem solving before 2030, assuming no AGI yet?
23
Ṁ10k
2029
53%
chance

Imagine that any math problem you can write down on a piece of paper that a team of Fields medalists can solve, AI can as well. Until recently, I would've predicted that that was an AGI-complete problem. Of course people used to think grandmaster-level chess would require AGI. Until 2022 I was sure that commonsense reasoning and being able to explain jokes would require AGI.

If subhuman general intelligence can be a superhuman mathematical intelligence, that will be another big update for me.

FAQ

1. What if AGI happens first?

This is a conditional prediction market. If AGI, as defined in my other market, happens first, this resolves N/A.

2. Does the AI need to max out the FrontierMath benchmark for this to resolve YES?

Yes, and every math benchmark, plus gold-medal performance on the International Math Olympiad. Even acing the Putnam.

3. What if it's essentially true but there are rare exceptions?

The spirit of the question is that we'd only consider an AI failure to be an exception if it failed for a reason other than being insufficiently brilliant at math. Like tricksy wording, or any trick question. The posing of the question has to be non-adversarial.

4. What about a book-length question?

Tentative answer so far: The problem has to be posed on a single human-readable sheet of paper or equivalent. But a question can cite any peer-reviewed math paper as background. (Dumping an impenetrable tome on the arXiv doesn't count.) If you have an example where this feels limiting, let me know. My suspicion is that all interesting math problems can be posed on a single page and in any case it won't harm the spirit of this question to limit ourselves to such.

4. What about research taste?

That's a big part of being a mathematician and isn't required for this market. The AI just has to be superhuman at answering questions, not asking them.

5. What about cost and speed?

The AI has to dominate the best humans on all metrics. We'll find an authoritative source for the market value of mathematicians' time if it comes down to that.

6. What about availability to the public?

Not required. If there's any doubt about the veracity of claims that this has been achieved, we'll discuss and delay resolution as needed.

Related markets

Get
Ṁ1,000
and
S3.00
Sort by:

What about solving currently unsolved conjectures? Like, would some sort of AI proving the Collatz Conjecture count?

bought Ṁ10 NO

The goalposts keep changing for what counts as an AGI. It’s quite annoying.

@Sebastianus Do they though? Or is it that people have always thought of AGI as roughly "the intelligence of humans" or "able to think, learn and solve problems across an arbitrary range of domains," and similar, while the narrow capabilities that were thought to be bottlenecks to AGI turned out to be easier than AGI?

bought Ṁ50 NO

@DavidHiggs Exactly.We want to find the simplest sufficient condition for AGI. We're not moving goalposts, we just keep learning that candidates like chess, explaining jokes, perhaps all of math, aren't sufficient after all. This is a hard meta problem.

I'm not sure if I'll bet in this myself yet. So far I haven't, because I'm feeling pretty freaking clueless. If I do, I'll say so and figure out a way to deal with my conflict of interest in resolving it.

@dreev Actually, as soon as I've said that, I'm thinking 50% is too high. So I'm going to dive in after all. As usual, I'll be extremely mindful of my conflict of interest and I commit to making the resolution fair, outsourcing the final decision if needed. I'll also be entirely transparent about my thinking. Ask me anything!

© Manifold Markets, Inc.Terms + Mana-only TermsPrivacyRules