ENSPIRING.ai: Why Vlad Tenev and Tudor Achim of Harmonic Think AI Is About to Change Mathand Why It Matters
The discussion in the video revolves around the burgeoning intersection of mathematics and artificial intelligence through the endeavors of Vlad and Tudor, who have founded Harmonic, an AI research lab. With a goal to create mathematical superintelligence, they believe that teaching AI math directly provides a more profound understanding of the world. The conversation explores this philosophy and contrasts it with existing approaches, questioning the efficacy of allowing math to emerge passively through AI training.
The video further investigates the role AI might play in conquering mathematical challenges such as the Riemann Hypothesis or winning millennium prizes. It introduces the concept of recursive self-improvement enabled by objective verification and self-play, a method envisioned to accelerate AI advancement without human intervention. The guests discuss how the right data and methodologies drive AI progress, emphasizing their privilege of collaborating on novel mathematical datasets.
Main takeaways from the video:
Please remember to turn on the CC button to view the subtitles.
Key Vocabularies and Common Phrases:
1. millennium prize [mɪˈlɛnɪəm praɪz] - (n.) - A prestigious award for solving some of the most challenging problems in mathematics. - Synonyms: (award, accolade, honor)
Last I checked, there's a 43% chance that the next millennium prize will be solved by AI...
2. superintelligence [ˌsuːpərˈɪntɛlɪdʒəns] - (n.) - An intelligence surpassing that of the brightest and most gifted human minds. - Synonyms: (genius, extreme intelligence, extraordinary cleverness)
They hope to create mathematical superintelligence with the thesis that understanding math allows you to better understand and reason about much of the broader world.
3. cognitive [ˈkɑːɡnɪtɪv] - (adj.) - Related to mental action or processes of acquiring knowledge and understanding through thought and experience. - Synonyms: (mental, intellectual, cerebral)
And the other really important thing is that it's a way to build these shared cognitive theories that are very objective and clear and transparent.
4. deductive [dɪˈdʌktɪv] - (adj.) - Characterized by the inference of particular instances from a general law or principle. - Synonyms: (logical, inferential, reasoned)
We can agree on what the ground rules are of whatever we're modeling, and then from there we can derive conclusions with humans through deductive rules.
5. quantitative [ˈkwɒntɪtətɪv] - (adj.) - Involving the measurement of quantity or amount. - Synonyms: (numerical, measurable, countable)
What you see is that when people become very good at math, they tend to be good at other quantitative areas in science and engineering.
6. autoregressive [ˌɔːtəʊrɪˈɡrɛsɪv] - (adj.) - Relating to a statistical model in which future values are assumed to depend linearly on past values. - Synonyms: (self-dependent, recursive, iterative)
So autoregressive language models have gotten pretty good.
7. synthesizing [ˈsɪnθəˌsaɪzɪŋ] - (v.) - Combining different ideas or things to create something new. - Synonyms: (blending, integrating, fusing)
I mean, that's exactly kind of how the system where engineering works, and that's what computers are great at, and AI models are great at synthesizing large amounts of information, finding patterns, recursively, self improving, I think
8. formalization [ˌfɔːrmələˈzeɪʃən] - (n.) - The process of creating explicit procedures and rules to make structures or systems rigorous and systematic. - Synonyms: (standardization, codification, systematization)
And make systems create the vast majority of their data and have the data actually increase in complexity as these models continue to iterate.
9. universal characteristic [juːnɪˈvɜːrsəl ˌkærəktəˈrɪstɪk] - (n.) - The idea of a universal language or system that can encapsulate and deduce all knowledge. - Synonyms: (general feature, common attribute, universal trait)
He created this thing called an idea called the universal characteristic, which is essentially the notion of having a deductive language.
10. recursive self-improvement [rɪˈkɜːrsɪv sɛlf-ɪmˈpruːvmənt] - (n.) - The process in which an AI system enhances its intelligence through iterative self-optimization. - Synonyms: (self-enhancement, iterative development, self-optimization)
One of the things that you said I want to hit on, which is this idea of recursive self improvement.
Why Vlad Tenev and Tudor Achim of Harmonic Think AI Is About to Change Mathand Why It Matters
Last I checked, there's a 43% chance that the next millennium prize will be solved by AI or, like a human, with significant AI assist. I think that's an underestimate. I mean, we could be lucky, and Larry Guth might be on the path to the Riemann hypothesis, which would be amazing. But I think that if the next one is solved by a human, it would probably have to be in the very near future. For sure, the next one will probably be significantly solved by AI.
We're excited to welcome Vlad and Tudor to the show. We've had the pleasure of knowing Vlad for many years at Sequoia, but what many of you may not know about Vlad is that in addition to being the founder and CEO of Robinhood, he's also an enormously talented mathematician. Vlad and Tudor have teamed up to create Harmonic, an AI research lab with the goal of pushing the frontiers of human knowledge. Specifically, they hope to create mathematical superintelligence with the thesis that understanding math allows you to better understand and reason about much of the broader world.
We're excited to talk to Tudor and Vlad about Harmonic, about the ingredients that go into creating mathematical superintelligence, including synthetic data, reinforcement, learning, and self-play, and when AI will win the IMO or a millennium prize, or even solve the Riemann hypothesis. All right, Vlad and Tudor, welcome to the show.
Oh, thanks for having us. All right, so you guys, you have this core belief that math is reasoning. And you have what might be a contrarian belief that the best way to train a model to perform well in mathematics is to directly teach it math, versus allowing math to emerge as a property of scale, which is what a lot of the other foundation model companies are doing. Can you talk a bit about that core belief? Why do you need to teach the model directly? Math. And also, maybe what does it mean that math is reasoning?
When we started the company, we had a really big focus on math, and maybe we can get to that later. But if you look around at all fields of science and engineering, well, almost all fields, math is really at their foundation. And math has essentially become the way that people understand the universe. It's the way you model phenomena from black holes to atoms; it's the way you design things in engineering. And there's a couple of reasons for that. First of all, it just happens to be the case that the universe is explainable by math. So you can write down a fairly compact set of symbols that explain things.
But the other really important thing is that it's a way to build these shared cognitive theories that are very objective and clear and transparent. And if you and I are discussing something that's rigorous, we can write down a set of deductive rules. We can agree on what the ground rules are of whatever we're modeling, and then from there, we can derive conclusions with humans. What you see is that when people become very good at math, they tend to be good at other quantitative areas in science and engineering.
And so our bet is that if you make a system that's really good at math, you're probably going to see the same phenomenon where it's true it might not immediately write the world's best history essays, but when you ask it to do something scientific or something in engineering, it's just going to be really, really good at that. That's why we started with math. And where is the boundary between helping me with my math homework and writing a history essay? There is some boundary that it's hard for math to cross.
What do you think the outer edges of what's possible with a model with sort of math at its core? Where are those outer edges? I'll give you the non-AI perspective. I studied math, and I was really good at math from when I was a little kid. And I remember there were always the 7th graders in math class that would raise their hands whenever the teacher would come up with something. It was always an abstract thing, you know, the side, angle, side of triangles. There'd be the annoying kid that was like, when are we ever going to use this? And, you know, the teacher would kind of mumble a little bit, and they'd be like, well, you know, math, just like, you're probably not going to use it soon, but it'll make you really good at other things.
And, you know, the other kids were always skeptical of that, and I bought into it. And so I just kept taking more and more advanced math. You know, I went to Stanford and I majored in it. Then I went to grad school to do a math Ph.D. And my belief was that, okay, if I just focus on math, then I'm going to get really good at problem-solving, and business problems and other problems will be easy compared to solving these really tough abstract algebra problem sets that I was banging my head against the wall for 10 hours every week trying to do. And I think it basically ended up being correct, right? It's like I didn't really pay attention to anything else.
I took maybe like one computer science class, an intro to computer programming at Stanford. And five years later, when I became an entrepreneur, I found it really easy to pick up code. I found it really easy to pick up contracts. And of course, I'm no lawyer, but you could understand that stuff. The logical underpinnings are relatively simple compared to abstract algebra and analysis. So I think for humans at least, I consider myself an example of mathematics transferring to other very monetizable things.
And I think for AI, my intuition seems to suggest that it should be the same. Yeah. And you already see a little bit of evidence of this at this point. It's an open secret in the industry that code training on a lot of code data leads to much better performance on reasoning benchmarks. So you can imagine what that'll look like when you have incredible math datasets that encompass a lot more general types of reasoning. Yeah, yeah. That resonates the idea that math teaches a human how to think critically, how to think logically, and that skill can be ported to a bunch of other domains.
It stands to reason that that would work in AI. Also, Vlad, you casually mentioned that you studied a little bit of math. And just for anybody who's not quite familiar with your background in math, I believe you studied briefly under Terry Tao, who is perhaps the world's greatest living mathematician. Yeah. And then one of the things you mentioned to us was that you still catch up with them every now and then. You have lunch when you're in LA, that sort of thing.
So I'm curious, when you have lunch with Terry Tao, what do you guys talk about? Do you give me stock tips? No, no, I'm not allowed to do that. Yeah. One of the unfortunate things of being a public company CEO in the financial space is I have lots of stock tips, but I can't share any of them. I have to keep them internal, and I can't even use them. I basically can't trade anymore, which is unfortunate because I love trading.
To backtrack, how I got to UCLA, Terry Tao is a professor at UCLA. And I think what's really amazing about him is the breadth of the work. So most mathematicians get very deep into a pretty narrow domain, and Terry can get very deep across dozens of domains. He's made contributions to number theory, combinatorics, harmonic analysis, applied math. He's one of the leading Lean contributors at this point. I'm sure he's formalizing his papers in Lean and actually hopping on the community Zulip and engaging with students. And then he has a very popular blog.
I think the way that he's been able to do this is he's just smarter than 99.999% of people, probably even more than that. So from a very early age, it was very clear that he was on another plane. I studied. I did my math honors thesis in undergrad under this professor Larry Guth, who's also really amazing. I mean, he actually had a recent result that came out that was a groundbreaking result in, I want to say, number theory or something about the Riemann hypothesis. But, yeah, this result in non-AI math really was quite something.
And he kind of suggested I look at UCLA. I was really interested in his field, and I ended up going there and being fortunate to study under Professor Tao. But I should be clear, I am a dropout, and it's amazing that I can claim that after grad school, but I will claim dropout status. So I only did one year of UCLA, so it was an intro to graduate-level analysis. Terry Tao taught my first year, which was pretty amazing.
And one thing I remember was I was doing some reading with Professor Tao, and he gave me this book and he signed it. And I think he signed it because he wanted to make sure I would give it back to him when I was done reading it. And little did he know that by signing it, he guaranteed he would never get that book back. And I bring it up every time I see him. I'm like, hey, you're not getting that book back. It's on my shelf next to all my other autographed first editions.
What does the math community think of AI mathematics? Are people split, or do people think it's the path to the promised land and the way that we're going to solve Riemann and everything else? I think it's split. I think it seems to be split. And there's sort of like, the younger mathematicians, I think, are very pro-AI and pro-verification and tools like Lean. And I think the older folks are a little bit more skeptical, so not surprising. I think you see that in pretty much every field.
I think that my guess would be that this will evolve. My mental model is something like chess, where at first there will be perhaps a lengthy period of humans plus AI assist, and that will lead to a lot of really good results over time. I think the AI will get better and better. And you look at chess right now, and it's sort of like if there's a human assisting the AI, the AI would be annoyed by it. They would just want to just delete all of the input because it would only make the results worse.
So I'm not sure if we're going to get to that point. I think humans will, at some point. I mean, they'll need to guide the algorithms, but I think the kind of definition of what a mathematician will do will fundamentally change. I was talking to one of my friends who's a mathematician at MIT, and I asked him when we were first starting this, what do you think? And this is a young professor, like, very excited about the field.
I was like, are you worried that you're kind of in a field that is going to fundamentally change? He's like, the field of math has always changed. Back in the 1800s, mathematicians used to be kind of like in the royal court, and they would be glorified calculators; they would solve quadratic equations by hand. And of course, they were worried that when computers and calculators came out, the job would no longer exist. But mathematicians get to define what math is. So I'm sure at some point it'll be prompting and kind of guiding these AI's to solve problems. And I think, yeah, I think that's going to be very huge.
Even if an AI solves the Riemann hypothesis, a human will always be in the loop, because the humans kind of pose the Riemann hypothesis to begin with. Yeah, just to hop on that, I think in the future, you're going to have a lot of compute resources dedicated to math, and the question will be a very human one, which is, by which procedure do humans decide where to direct all that reasoning firepower? And I think that's going to be the job of mathematicians. They have to choose, what do we work on? How do we interpret the results? How do we interpret failures to find answers? That kind of thing.
Do you think an AI math system can solve the Riemann? Or where is the ceiling, do you think? I think that it should be able to solve it or prove that it's undecidable. That would also be an interesting result. Yeah, I think if we think about what a great mathematician, like a Terry Tao, for instance, is capable of doing, they're able to synthesize lots of papers, lots of frontier results, and learn from them in a way that the other top human mathematicians can and kind of find connections between these things and sort of use them to create new and more complex theories.
I mean, that's exactly kind of how the system where engineering works, and that's what computers are great at, and AI models are great at synthesizing large amounts of information, finding patterns, recursively, self improving, I think. Now on Metaculous, last I checked, there's a 43% chance that the next millennium prize will be solved by AI or human, with significant AI assist. And I think that's an underestimate. I mean, we could be lucky, and Larry Guth might be on the path to the Riemann hypothesis, which would be amazing. But I think that if the next one is solved by a human, it would probably have to be in the very near future. And for sure, the next one will probably be significantly solved by AI.
One of the things that you said I want to hit on, which is this idea of recursive self improvement, because it seems like in the world of AI, there are, if you were to draw a spectrum of human-only to AI-only, and then human in the loop is sort of the spectrum in the middle from lots of human, a little bit of AI, and lots of AI, a little bit of human. One of the things that is interesting about Harmonic, at least the way I understand it, is because of Lean, you can encapsulate math and code. Because of formal verification, you can objectively determine whether things are right or wrong, which means that you have an objective reward function that you can use with self-play to have very fast cycle times with reinforcement learning, which means that the progress of your model has a chance to be extremely fast, because there are no humans in the loop with that.
recursive self-improvement objective function is clearly defined. You can do self-play to just make the model better and better and better and better and better, which is not something that we see in a lot of domains of AI. Most domains of AI, it's a lot messier to kind of get the cycle time on improvement. Can you just talk through the system a bit? A little bit of what I described, like what feeds into your model? What governs the rate at which it can get better? Because it seems like something that we'll be able to get better at a pretty quick rate.
Yeah, I'm happy to cover that. One point before going to that is just that I think the most exciting part about this is that there are other areas where recursive self-improvement can work, for example, again, in those board games like Alphago. But I think what a lot of people don't realize is that in these, let's say, perfectly observed zero-sum games, you improve recursively just by playing against yourself, but you hit an optimal strategy. So at that point, it doesn't matter what system you have; it won't do better.
The most exciting part about math is that there is no upper bound. So you're just going to keep putting compute in, and it's going to keep getting better, and there's no end to it. And so, when we talk about, do we think AI can solve a Riemann hypothesis or get a millennium prize, those are very human milestones. And I think the real question is, will it ever stop? I mean, because it clearly will get there. And I think we're going to end up solving problems that are much, much harder than we thought.
One hypothesis, which we haven't even conceived of yet, because it's almost like it's beyond us to write down such a hard problem. But coming up, have you guys ever seen that Minecraft video of, like, the AI beating Minecraft in 20 seconds? No. That sounds like a good analogy. It's like, you know what Minecraft is, how a human would play it, and then the AI beating it in 20 seconds is just incomprehensible. You can't even kind of grok what's going on in the video feed.
Yeah, but I think if we just talk about how Harmonic works, you can just think of it as there's a collection of agents that are essentially trying to solve problems. And it's true, because we use Lean, we're able to check whether our answers are correct and thereby derive a variety of training signals that we use to improve the system. But just to be clear, the use of Lean just lets you verify things.
Lean doesn't itself tell you whether you're getting closer to the answer or whether you're getting smarter or not. It's just telling you whether it's correct or nothing. So there's a lot of open scientific challenges to making it get better quickly.
And can you just say a word about what Lean is just in case people aren't familiar? Yeah, totally. Lean is just another programming language, a really great one, created by Leo D'Amora, the best programming language. We might all be writing Lean, or the AIs might just be writing Lean in the future. But the idea is that the mathematical statements are encoded in the type system of the language. So just very simply, you have functions in Lean, and the input types correspond to the assumptions of the mathematical theorem, and the output type is the conclusion.
And the point of Lean is that if you write a program that implements that function and it compiles, that means you can derive the output type from the input type, which in turn implies that you can conclude the conclusion from the assumptions. So that's really the fundamental, that's how you use Lean for math.
And I think one thing that's super interesting about Lean is if you look at Leo de Mora, the creator, who's at Amazon AWS. Now, he's not a mathematician, and he wrote this as a software verification tool. So he has the belief that in the future, software will be verified. And the existing tools, things like Coq and Isabelle, which are kind of multi-decade-old software verification tools, are just not good, and they're frankly unusable.
The experience for a developer is poor, and so he wanted to create a better software verification tool in the hopes of if he builds something better, more people will use it, and we'll have better software, which is a super noble goal in its own right. But then what he didn't realize was software verification. All it is is just proving that software has certain properties. And this thing became very popular in the math community. And you had thousands of mathematicians and math students building up an organic library called Mathlib, which you can think of as kind of like the largest math textbook open source.
It's on GitHub, and it's just growing at a pretty fast clip. And the usage of Lean for math, I think, to some degree has surpassed anyone's expectations. It might be more than the usage for verified software at this point, and that might change as time goes forward.
And with AI, one of the questions we always ask is, why now? Because reinforcement learning has existed for a long time, math has existed for even longer, and it seems like math has really hit an inflection point. You guys have chosen to start Harmonic at this point in time, but why now? Oh, I mean, there's two really good reasons why now that we're excited about.
So, the first one is just that the AI systems have gotten better in an interesting way. So I was actually talking with a close friend about RL for theorem proving back in 2015, in 2016. And one issue back then was that there wasn't even a great notion of an AI system that could predict something in an infinite action space. So in Go, you can place a piece somewhere. It's either a black or white piece, but in math, you can really do anything.
You can just generate any next step. And so we didn't have great systems to do that. So autoregressive language models have gotten pretty good. So that's one thing that makes it possible. I'm talking on the timescale of a decade here, but that's really important. And the other thing that's kind of crazy is that Lean has gotten really good.
So if you had told a mathematician 20 years ago that a large fraction of the field would be excited about formal methods in math, they might have thought you were crazy, because back then, formal methods were really isolated to formal logic or certain types of graph theory, like if you guys have heard the four-color theorem, that was one big success for formal math. But what's changed is that Lean is so flexible and so exciting for people that they've contributed this thing called Mathlib. So now there's a lot of body of knowledge that you can build on to prove things.
And so the combination of AI starting to even be a possible fit for this problem. Plus, Lean working really well, and Lean four was only released officially in September 2023. So those two things happening together really made it the right time to attack this.
Can you say a word about data and specifically synthetic data and what it is that fuels the model that you guys are building? Yeah. So synthetic data is the fuel for the model. There's an amazing resource called Mathlib. It's that open source repository. So that's a lot of human-written Lean, and it's written in a way that's very general and compact. So they're really proving advanced theorems. Right. It's not necessarily the best fit for problem-solving. And so as a result, almost the only data you can use for this is synthetic data you generate yourself, because that original data is not very applicable.
So it's a data-poor regime compared to most areas of AI. That process that I described, where the agents themselves are trying to solve problems and thereby generate training signals, that's the primary way in which you can get data. And the other issue is that you have to progress through levels of intelligence. So you're not necessarily proving the Riemann hypothesis upfront, you're proving really simple things, but then you amplify yourself recursively throughout the process.
Turns out there's not as much math data on the Internet as cat videos. Unfortunately not. Unfortunately not. Well, yeah, it's interesting, though, because there's the data wall that the foundation model, the general-purpose foundation model companies, are running into. And at this point, they've exhausted what's available on the Internet. And if you can generate most of the data that's required to train, that's kind of another advantage of having math at the core versus hoping for math as an emergent property of scale.
Yeah. And I think the data wall kind of manifests itself in two ways. One is just like you said, we're out of Internet data. Yeah. The other is the actual Internet data quality that's out there. You can think of that as providing kind of a ceiling to how smart these models can get, because if you train on the cat videos and all the nice Wikipedia content and the Internet content. It's an open problem, how to get something that's significantly smarter than that.
And so you do need to get into some kind of self-reinforcing, self-play regime, in our opinion, to get to a point where you can surpass the ability of human mathematicians and researchers at multiple tasks. And so I think in many ways, the path, it's inevitable that it takes kind of the AlphaGo to AlphaZero approach, and we learn how to make these models create the vast majority of their data and have the data actually increase in complexity as these models continue to iterate.
I think the great thing about math is there's a simple path to doing this. You can basically measure the complexity of a math problem and how difficult it is by how many lines of Lean it takes to solve. So you can actually look at the complexity of a system, and a lot of problems are solved by breaking it down into smaller chunks and actually solving those chunks.
And if you kind of think about how that works, the smaller chunks are then more manageable because there are sort of fewer lines to solve than the big one. So if you get really good at that, and then you get good at solving the chunks, then you can kind of train your model to do better. And as you kind of keep turning the gears on that, the model gets better at solving incrementally harder and harder and more complex problems.
I think that works very well in math, because it mirrors how we solve math on pen and paper. And we've been able to start with simple axioms and build up just giant complex structures. Maybe the Riemann hypothesis would be hundreds of pages, if not more, to solve. Fermat's Last Theorem was, I think, 200 pages, very, very complex. So, yeah, I do think eventually you'll get to a level where you'll be able to solve these things.
And the math is, to some extent, like the original synthetic data. Yeah. What determines the rate at which your model can get better, the rate at which you can get better? Well, I think the highest level one is energy. So the more energy you can put into it, the more attempts can happen in parallel, which means you generate data faster.
There's no rate-limiting step, so there's a bunch of rate-limiting steps, but there's no fundamental constraint on how fast it gets better. So it's really just about how much compute you put in. I think it's also, I mean, there's still a lot of unsolved problems in this field. Right.
Like, we benefit a lot from core theorems that have been proved in the past. And, you know, there's. If you think about, like, competition, math context, there's theorems that every student would just learn and use, like the AM-GM inequality, things of that nature. And so, to some degree, like, Mathlib is incomplete.
There's very little content about geometry, for instance, and it's very theoretical and abstract. And so a limiting step is like, what's in Mathlib. And, of course, at some point, the models have to solve the problem of, like, creating new theories and new structures, kind of like expanding to other domains and getting really, really good at formalizing things that haven't been formalized by humans. I think that'll be a big unlock, and that'll certainly happen within the next several years.
You'll be able to say, hey, here's just like this situation. It could be as simple as a baseball team, and they're, like, throwing balls back and forth to each other. Systems would be able to auto formalize that and turn that into Lean code on the fly. And I don't think we're quite there yet to the point where that's reliable. But when it does get reliable, I think that'll be a really big unlock.
If everything goes right, what do you think Harmonic becomes? Well, our mission statement is to explore the frontiers of human knowledge. So it's very important to us that the things we produce are correct and useful to humans. So I think in the best case, we're able to build a tool that a lot of mathematicians use to close all the millennium prize problems and to go far beyond that. I think that'll be a great service to humanity.
There's also other areas of commercial application for the software. The dream for software engineering is to be able to just check that code is correct. To do that, you need to have a very good model of how code works. You need to be able to understand how the libraries work, what they promise to think. You can imagine a future where safety-critical software is proven correct, and the general software is proven correct, and the way software engineering is done can change as well.
There's just a lot of applications. If you can make a system that's very good at math reasoning and very good at checking its reasoning. Yeah, really? We think there's a lot of applications. Yeah. And I think that, I mean, math and software are two fairly obvious ones. I think software engineering as a discipline is changing really quickly.
I'm sure you guys are seeing all the reports of people doing crazy things with cursor and Claude 3.5 I think in the future, software engineering will be less about reviewing and collaborating on code as an artifact and will be more about collaborating on specs. What do we want the code to do? Can we be more rigorous about that?
And I think that's where verification will become a bigger thing, because as the cost of software goes to zero, the cost of verified software will also go to zero. And suddenly, this thing that was very impractical and expensive because you need specialist humans to do it, will just accelerate dramatically. With AI, I think you look out 5 to 10 years from now, the vast majority, I mean, if we progress along the capability curve as we have been, the vast majority of software written will be verified and provably correct. I think that's a really exciting future.
I also think on the more theoretical side, it's not just math, but physics is essentially math. Theoretical physics is one of the main ways the frontier of math gets implemented. I think it would be amazing to me personally to accelerate some of the fundamental physics research at the frontier and really just develop an understanding of why the universe is the way it is, why the laws of physics exist, and also help figure out experiments to test those. So I would be very proud if we contributed to that effort.
And do you think you'll mostly be contributing to math and math adjacent areas like physics and software engineering, or do you think anything that involves reasoning is in spec for Harmonic? Yeah, I mean, I think we try to be focused on the things that we're still a small company over the long term. I think if you believe math is reasoning, and we do, then, yeah, if we get really, really good at math and computer science is a very natural analogous, then, yeah.
I mean, anything is in scope for those models. Even the history essay, I think we'll see. History essays? Yeah. I really enjoyed writing history essays, even though my parents were like, you know, humanities, language arts, just ignore all that stuff. But I think my math skills led me to write great history essays, too. So hopefully, Aristotle will be no different one day.
Aristotle wrote some great historical commentary. You are truly a polymath. Yeah, I mean, poetics is. If you've read poetics, it's. Should we wrap it up with some rapid fire? Let's do it.
Lightning round. Okay. In what year will you win the IMO? What do you think, Tudor? Soon. All right. 2025 soon. Maybe 2024. All right. We'll sign you up for 2024. All right. How about the millennium prize? Ooh, that's a tough one. I would guess 2029. 2029, yeah. Okay. I heard 2028. Is that what they're. Yeah, I guess it's a fully AI, unassisted millennium prize, or AI human hybrid. Well, how about. What do you think for hybrid?
Hybrid? I could see 2028. Are we talking an easy millennium prize or hard. Yeah. Is it, like, not very easy, easy millennium prize, Navier Stokes might be 2026. Riemann hypothesis. I'll give you 2029. All right. All right. There we go. Good. Given we can't even do arithmetic today with LLMs, that's pretty amazing.
When do you think we'll have human or superhuman-level reasoning, more broadly defined? I think to some degree, if you define it as something that can reason and solve math problems in excess of any human, like something that Terry Tao would, you know, would give Terry Tao a run for his money. I think we're a couple of years away, but I think the models within the next year will get to probably like 99.99 percentile.
Would Terry agree with that? I think so, yeah. I don't know. You'd have to ask him, but I think he would agree with that. One of our favorite questions is, who in the world of AI do you admire most? And we'll modify it slightly for you guys. Who in the world of AI or mathematics do you admire most? I like von Neumann.
We were just talking before about von Neumann's biography. I think what I find really interesting about him was he started as a mathematician and he was discouraged. I think his father, who was like a Hungarian businessman, was trying to discourage him from doing math because it wasn't very monetizable.
And so he got his friend, who was a great mathematician, to try to talk him out of it. But the friend came back and he's like, I can't do it. This guy's too good. It would be just disservice to society if he didn't use his talents for math. And then he pioneered computer science, and the von Neumann machine was the blueprint for all modern computers.
He contributed to the Manhattan project, which is a little controversial, but very, very practical and impactful, and created probably the canonical text in game theory as well. So, yeah, I think it's pretty amazing. Also, a fellow eastern European. While some people debate whether Hungary is in eastern Europe. Yeah, it's an interesting question.
I think I definitely admire almost all scientists and mathematicians, but I think that if you've heard of the mathematician Leibniz, what was shocking to learn during the course of working on this company is that Leibniz was competing with Newton to create calculus, and Newton's formulation went out, but Leibniz was basically there.
But one thing people don't know is that Leibniz also had a lot of other work and one piece of work that is just incredibly prescient. Keep in mind, this is the late 1600s. He created this thing called an idea called the universal characteristic, which is essentially the notion of having a deductive language, an automated procedure to deduce things using that language, and an encyclopedia of work in that language that you can build on to derive things.
And so the amazing thing to me is that this thinker hundreds of years ago essentially predicted what would be happening in 2024. And it seems that the only thing that was required was having AI get a little better and having computers that can do something like Lean. I think it's just incredible to have a human being predict that with no concept whatsoever of what's going to come later, but to understand that that's such a fundamental thing that we're going to end up working on that hundreds of years later.
Awesome. Thank you guys. Thank you. Thanks for having us. Thanks for having us.
Artificial Intelligence, Education, Innovation, Mathematics, Research, Ai Development, Sequoia Capital
Comments ()