Transcript
A (0:00)
Hello, I'm Andrew Main and this is the OpenAI podcast. Today our guests are researchers Sebastian Bubeck and Ernest Rio and we're going to talk about math, how it went from almost laughable to Olympiad level, and why you need math to reach AGI.
B (0:14)
The progress of the last few years has been nothing short of miraculous.
C (0:18)
We will be able to have LLMs, be able to solve problems that require more than 50 pages of thinking.
B (0:24)
Mathematics was just the perfect benchmark to see the model making progress during the last four years.
A (0:34)
Sebastian Ernest, I'd love to know more about you. So how would you explain your roles?
B (0:38)
Yeah, sure. So I have been working in mathematics for almost 20 years now. I used to work in optimization and theory of machine learning. I was a professor at Princeton for a few years before moving to Microsoft, and now I'm a researcher at OpenAI and in the last few years have been really trying to understand how AI can help mathematics and to really evaluate the progress that we're making in terms of solving difficult math problems with AI.
A (1:08)
Ernest, how about you?
C (1:09)
Yeah, so I've recently joined OpenAI as a researcher, but before that I was an applied mathematician working on optimization and machine learning theory. And in my previous job I worked as a professor of mathematics at the UCLA Math department.
A (1:26)
So I think a lot of people have this perception that these models aren't good at math, literally called language models. And how has that changed what's gone on?
B (1:35)
Yeah, I think, you know, the progress of the last few years has been nothing short of miraculous. It's important to remember that two years ago we didn't even have reasoning models, let alone models that could prove, you know, difficult mathematical theorems. Today, two years later, the models, they are able to help field medalists in their day to day work. Really the jump is just simply astounding. Maybe if I can build a little bit more on that. Something which is important to understand is that everybody has been surprised by this progress, including us. To tell you a story, a year and a half ago I was at a workshop at the conference with other fellow mathematicians and there was a debate that I participated in on whether scaling LLMs will help us resolve major open problems. So this was a debate a year and a half ago and the room was very divided. In fact, they did a poll at the beginning and I think it was like 80% said no, impossible that this would happen. So then the debate unfolded and by the end of the debate it was more like 50 50. So pretty good progress during that hour. This obviously was just so wrong. In hindsight, like just mere eight months later, the model we're starting to be able to do research level mathematics.
