ENSPIRING.ai: Eric Schmidt Unveils The Future Of AI Challenges And Insights
Eric Schmidt, the former CEO of Google, discusses the vast implications of AI advancements in a compelling interview at Stanford. He anticipates that AI, particularly large language models and context window extensions, could transform human interaction with technology to unprecedented levels. Schmidt envisions scenarios where anyone can use language to generate digital commands and emphasizes the potential for individual programmers, which could significantly reduce costs and increase accessibility.
Schmidt impresses the necessity of maintaining America's lead in AI technology amidst competition from China. He underscores the importance of strategic investments, energy considerations, and collaboration with countries like Canada. He also addresses the complex challenges faced by big tech companies and emphasizes his views on work ethic and innovation, highlighting transformative examples like unmanned drones in warfare.
Main takeaways from the video:
Please remember to turn on the CC button to view the subtitles.
1. Context [ˈkɑːn.tɛkst] - (n.) The circumstances or setting surrounding a particular event or situation.
In a context window, you can basically use that as short term memory.
2. Arbitrary [ˈɑːrbɪˌtrɛri] - (adj.) Based on random choice or personal whim, rather than any reason or system.
You understand how powerful that is if you can go from Arbitrary language to Arbitrary digital command.
3. Suppression [səˈprɛʃən] - (n.) The act of stopping or reducing something, typically by force.
If TikTok is banned, here's what I propose.
4. Oscillation [ˌɑː.səˈleɪ.ʃən] - (n.) Movement back and forth at a regular speed.
Every six months, I oscillate.
5. Hypothesis [haɪˈpɒθ.ə.sɪs] - (n.) A proposed explanation for a phenomenon used as a starting point for further investigation.
They run it to generate chemistry hypotheses.
6. Recency [ˈriː.sən.si] - (n.) The characteristic of being recent or the state of having occurred not long ago.
The current models take a year to train.
7. Misinformation [ˌmɪsˌɪn.fəˈmeɪ.ʃən] - (n.) False or misleading information spread regardless of intent to deceive.
How can we stop AI from influencing public opinion Misinformation...
8. Adversarial [ˌæd.vɚˈsɛr.i.əl] - (adj.) Involving or characterized by conflict or opposition.
So in the example that I gave...
9. Infrastructure [ˈɪn.frəˌstrʌk.tʃər] - (n.) The basic physical and organizational structures and facilities needed for the operation of a society or enterprise.
They need 10 billion, 20 billion, 50 billion...
10. Asymmetry [ˌeɪˈsɪm.ə.tri] - (n.) Lack of equality or equivalence between parts or aspects of something.
So there's this asymmetric warfare happening...
Eric Schmidt Unveils The Future Of AI Challenges And Insights
Eric Schmidt, the former CEO of Google, just did an interview at Stanford where he talked about a lot of controversial stuff. Initially, the interview was uploaded on Stanford's YouTube channel, but a couple of days later, the interview was taken down. But today, I was somehow able to access the interview video after spending multiple hours. So let's watch it together and dissect some important parts of the interview. In the next year, you're going to see very large context windows agents and text to action.
When they are delivered at scale, it's going to have an impact on the world at a scale that no one understands yet, much bigger than the horrific impact we've had by social media, in my view. So here's why. In a context window, you can basically use that as short-term memory. And I was shocked that context windows get this long. The technical reasons have to do with the fact that it's hard to serve, hard to calculate, and so forth. The interesting thing about short-term memory is when you feed, you ask it a question, read 20 books, you give it. The text of the books is the query, and you say, tell me what they say. It forgets the middle, which is exactly how human brains work too, right?
That's where we are with respect to agents. There are people who are now building essentially LLM agents. And the way they do it is they read something like chemistry. They discover the principles of chemistry, and then they test it. And then they add that back into their understanding. Right? That's extremely powerful. And then the third thing, as I mentioned, is text action. So I'll give you an example. The government is in the process of trying to ban TikTok. We'll see if that actually happens.
If TikTok is banned, here's what I propose. Each and every one of you do say to your LLM the following; Make me a copy of TikTok. Steal all the users, steal all the music, put my preferences in it, produce this program in the next 30 seconds, release it, and in 1 hour, if it's not viral, do something different along the same lines. That's the command. Boom, boom, boom, boom. Right? You understand how powerful that is if you can go from Arbitrary language to Arbitrary digital command, which is essentially what Python in this scenario is.
Imagine that each and every human on the planet has their own programmer that actually does what they want, as opposed to the programmers that work for me who don't do what I ask. Right. The programmers here know what I'm talking about. So imagine a non-arrogant programmer that actually does what you want, and you don't have to pay all that money to. And there's an infinite supply of these programs, and this is all within the next year or two, very soon.
So we've already discussed on this channel a number of different versions of this. Whether you're talking about Aitor Devon Pythagora or just using agents to collaborate with each other in code, there are just so many great options for coding assistance right now. However, AI coders that can actually build full-stack complex applications. We're not quite there yet, but hopefully soon. And also what he's describing of just saying, download all the music and the secrets and recreate, that's not really possible right now. Obviously, all of that stuff is behind security walls, and you can't just download all that stuff. So if he's saying, hey, reproduce the functionality, you can certainly do that.
Those three things. And I'm quite convinced it's the union of those three things. It will happen in the next wave. So you asked about what else is going to happen. Every six months, I oscillate. So we're on a. It's an even odd Oscillation. So at the moment, the gap between the frontier models, which they're now only three, very few who they are and everybody else appears to me to be getting larger. Six months ago, I was convinced that the gap was getting smaller.
So I invested lots of money in the little companies. Now I'm not so sure. And I'm talking to the big companies, and the big companies are telling me that they need 10 billion, 20 billion, 50 billion, 100 billion. Stargate is 100 billion. Right. They're very, very hard. I talked. Sam Altman is a close friend. He believes that it's going to take about 300 billion, maybe more. I pointed out to him that I'd done the calculation on the amount of energy required. And I then, in the spirit of full disclosure, went to the White House on Friday and told them that we need to become best friends with Canada, because Canada has really nice people, helped invent AI, and lots of hydropower, because we as a country do not have enough power to do this.
The alternative is to have the Arabs fund it. And I like the Arabs, personally. I spent lots of time there, right? But they're not going to adhere to our national security rules, whereas Canada and the US are part of a treaty where we all are. So these $300 billion data centers, electricity starts becoming the scarce resource now. First of all, we definitely don't have enough energy resources to achieve AGI. It's just not possible right now. And Eric is also assuming that we're going to need more and more data and larger models to reach AGI.
And I think that's also not actually true. Sam Altman has said similar things. He has said that we need to be able to do more with less or even the same amount of data. Because we've already used all the data that humanity has ever created. There's really no more left. So we're going to need to either figure out how to create synthetic data that is valuable, not just derivative, and we're also going to have to do more with the data that we do have.
You were at Google for a long time, and they invented the transformer architecture. It's all Peter's fault, thanks to brilliant people over there like Peter and Jeff Dean and everyone. But now it doesn't seem like they've kind of lost the initiative to open AI. And even the last leaderboard I saw, Anthropic's Claude was at the top of the list. I asked Sundar this. He didn't really give me a very sharp answer. Maybe you have a sharper or more objective explanation for what's going on there. I'm no longer a Google employee.
In this beautiful disclosure, Google decided that work-life balance and going home early and working from home was more important than winning. Okay, so that is the line that got him in trouble. It was everywhere, all over Twitter, all over the news when he said Google prioritized work-life balance, going home early, not working as hard as the competitor. To winning, they chose work-life balance over winning. And that's actually a pretty common perception of Google and the startups.
The reason startups work is because the people work like hell. And I'm sorry to be so blunt, but the fact of the matter is if you all leave the university and go found a company, you're not gonna let people work from home and only come in one day a week if you wanna compete against the other startups. In the early days of Google, Microsoft was like that. Exactly. But now it seems to be. There's a long history of, in my industry, our industry, I guess, of companies winning in a genuinely creative way and really dominating a space and not making the next transition. It's very well documented.
And I think that the truth is founders are special. The founders need to be in charge. The founders are difficult to work with, push people hard as much as we can. Dislike Elon's personal behavior. Look at what he gets out of people. I had dinner with him and he was flying. I was in Montana. He was flying that night at 10:00 p.m. to have a meeting at midnight with x AI. Think about it.
I was in Taiwan, different country, different culture, and they said that. And this is TSMC, who I'm very impressed with. And they have a rule that the starting PhDs coming out of the, they're good physicists, work in the factory on the basement floor. Now, can you imagine getting American physicists to do that with PhDs? Highly unlikely. Different work ethic. And the problem here, the reason I'm being so harsh about work, is that these are systems which have network effects. So time matters a lot.
And in most businesses, time doesn't matter that much. You have lots of time. Coke and Pepsi will still be around, and the fight between Coke and Pepsi will continue to go along. And it's all glacial. When I dealt with telcos, the typical telco deal would take 18 months to sign. There's no reason to take 18 months to do anything. Get it done. We're in a period of maximum growth, maximum gain. So here he was asked about competition with China's AI and AGI, and that's his answer. We're ahead, we need to stay ahead, and we need money.
AI is going to play a role, or competition with China as well. So I was the chairman of an AI commission that sort of looked at this very carefully, and you can read it, it's about 752 pages. And I'll just summarize it by saying we're ahead, we need to stay ahead, and we need lots of money to do so. Our customers were the Senate and the House, and out of that came the chips act and a lot of other stuff like that.
A rough scenario is that if you assume the frontier models drive forward and a few of the open-source models, it's likely that a very small number of companies can play this game. Countries. Excuse me, what are those countries or who are they? Countries with a lot of money and a lot of talent, strong educational systems, and a willingness to win. The US is one of them, China is another one. How many others are there? Are there any others? I don't know, maybe. But certainly in your lifetimes, the battle between the US and China for knowledge supremacy is going to be the big fight.
So the US government banned, essentially the Nvidia chips, although they weren't allowed to say that was what they were doing, but they actually did that into China. They have about a ten-year chip advantage. We have a roughly ten-year chip advantage in terms of sub duv. That is sub 5 years. That long? Roughly ten years. Wow. And so you're going to have. So an example would be, today we're a couple of years ahead of China.
My guess is we'll get a few more years ahead of China. And the Chinese are whopping mad about this. It's like, hugely upset about it. Well, let's talk about a real war that's going on. I know that something you've been very involved in is the Ukraine war, and in particular, I don't know how much you can talk about white Stork and your goal of having $500 drones destroy $5 million tanks. How's that changing warfare?
I worked for the secretary of defense for seven years and tried to change the way we run our military. I'm not a particularly big fan of the military, but it's very expensive, and I wanted to see if I could be helpful. And I think, in my view, I largely failed. They gave me a medal, so they must give medals to failure or whatever. But my self-criticism was, nothing has really changed, and the system in America is not going to lead to real innovation.
So watching the Russians use tanks to destroy apartment buildings with little old ladies and kids just drove me crazy. So I decided to work on a company with your friend Sebastian Thrun, as a former faculty member here, and a whole bunch of Stanford people. And the idea basically is to do two things, use AI in complicated, powerful ways for these essentially robotic war. And the second one is to lower the cost of the robots.
Now, you sit there and you go, why would a good liberal like me do that? And the answer is that the whole theory of armies is tanks, artilleries, and mortar, and we can eliminate all of them. So here, what he's talking about is that Ukraine has been able to create really cheap and simple drones by spending just a couple of hundred dollars. Ukraine is creating 3d printed drones. They carry a bomb, drop it on a million-dollar tank, and they've been able to do that over and over again. So there's this asymmetric warfare happening between drones and more traditional artillery.
So there was an article that you and Henry Kissinger and Dan Huttenlacher wrote last year about the nature of knowledge and how it's evolving. I had a discussion the other night about this as well. So for most of history, humans sort of had a mystical understanding of the universe. And then there's the scientific revolution and the enlightenment. And in your article, you argue that now these models are becoming so complicated and difficult to understand that we don't really know what's going on in them. I'll take a quote from Richard Feynman.
He says, what I cannot create, I do not understand the saddest quote the other day, but now people are creating things they do not that they can create, but they don't really understand what's inside of them. Is the nature of knowledge changing? In a way? Are we going to have to start just taking the word for these models without them being able to explain it to us? The analogy I would offer is to teenagers. If you have a teenager, you know they're human, but you can't quite figure out what they're thinking. But somehow we've managed in society to adapt to the presence of teenagers.
Right. And they eventually grow out of it. And I'm just serious. So it's probably the case that we're going to have knowledge systems that we cannot fully characterize, but we understand their boundaries. Right. We understand the limits of what they can, and that's probably the best outcome we can get.
Do you think we'll understand the limits? We'll get pretty good at it. He's referencing the way that large language models work, which is really essentially a black box. You put in a prompt, you get a response, but we don't know why. Certain nodes within the algorithm light up, and we don't know exactly how the answers come to be. It is really a black box. There's a lot of work being done right now trying to kind of unveil what is going on behind the curtain, but we just don't know.
The consensus of my group that meets every week is that eventually the way you'll do this so-called Adversarial AI is that there will actually be companies that you will hire and pay money to to break your AI system, like a red team. So it'll be the red instead of human red teams, which is what they do today. You'll have whole companies and a whole industry of AI systems whose jobs are to break the existing AI systems and find their vulnerabilities, especially the knowledge that they have that we can't figure out.
That makes sense to me. It's also a great project for you here at Stanford, because if you have a graduate student who has to figure out how to attack one of these large models and understand what it does, that is a great skill to build the next generation. So it makes sense to me that the two will travel together.
All right, let's take some questions from the student. There's one right there in the back. Just say your name. Earlier you mentioned, and this is related to the comment. Right now I'm getting an AI that actually does what you want. You just mentioned Adversarial AI, and I'm wondering if you could elaborate on that more. This seems to be, besides, obviously, compute will increase and get more performant models, but getting them to do what you want issue seems partially unanswered, in my view.
Well, you have to assume that the current hallucination problems become less as the technology gets better and so forth. I'm not suggesting it goes away. And then you also have to assume that there are tests for efficacy. So there has to be a way of knowing that the thing succeeded. So in the example that I gave of the TikTok competitor, and by the way, I was not arguing that you should illegally steal everybody's music.
What you would do, if you're a Silicon Valley entrepreneur, which hopefully all of you will be, is if it took off, then you'd hire a whole bunch of lawyers to go clean the mess up. But if nobody uses your product, it doesn't matter that you stole all the content. And do not quote me. You're on camera. Yeah, that's right. But you see my point. In other words, Silicon Valley will run these tests and clean up the mess, and that's typically how those things are done.
So my own view is that you'll see more and more performative systems with even better tests and eventually Adversarial tests, and that'll keep it within a box. The technical term is called chain of thought reasoning. And people believe that in the next few years, you'll be able to generate a thousand steps of chain of thought reasoning, right? Do this. Do this. It's like building recipes, right? That the recipes, you can run the recipe, and you can actually test that it produced the correct outcome.
Now, that was maybe not my exact understanding of chain of thought reasoning. My understanding of chain of thought reasoning, which I think is accurate, is when you break a problem down into its basic steps and you solve each step, allowing for progression into the next step. Not only it allows you to kind of replay the steps, it's more of how do you break problems down and then think through them step by step?
The amounts of money being thrown around are mind-boggling. And I've chosen, I essentially invest in everything because I can't figure out who's going to win. And the amounts of money that are following me are so large, I think some of it is because the early money's been made, and the big money people who don't know what they're doing have to have an AI component, and everything is now an AI investment, so they can't tell the difference.
I define AI as learning systems, systems that actually learn. So I think that's one of them. The second is that there are very sophisticated new algorithms that are sort of post transformers. My friend, my collaborator for a long time, has invented a new non-transformer architecture. There's a group that I'm funding in Paris that claims to have done the same thing. So there's enormous invention there, a lot of things at Stanford, and the final thing is that there is a belief in the market that the invention of intelligence has infinite return.
So let's say you put $50 billion of capital into a company, you have to make an awful lot of money from intelligence to pay that back. So it's probably the case that we'll go through some huge investment bubble and then it'll sort itself out. That's always been true in the past, and it's likely to be true here. And what you said earlier. Yeah, so there's been something like a trillion dollars already invested into artificial intelligence and only 30 billion of revenue.
I think those are accurate numbers. And really, there just hasn't been a return on investment yet. But again, as he just mentioned, that's been the theme of previous waves of technology. Huge upfront investment, and then it pays off in the end. Well, I don't know what he's talking about here. Because didn't he run Google? And Google has always been about being closed source and always tried to protect the algorithm at all costs.
So I don't know what he's referring to there. You think that the leaders are pulling away from right now, and this is a really. The question is roughly the following. There's a company called Mistral in France. They've done a really good job, and I'm obviously an investor. They have produced their second version. Their third model is likely to be closed because it's so expensive. They need revenue and they can't give their model away.
So this open source versus closed source debate in our industry is huge. And my entire career was based on people being willing to share software in open source. Everything about me is open source. Much of Google's underpinnings were open source. Everything I've done, technically, what didn't he run? Google. And Google was all about staying closed-source, and everything about Google was kept secret at all times.
So I don't know what he's referring to there. Everything I've done technically, and yet it may be that the capital costs, which are so immense, fundamentally changes how software is built. You and I were talking. My own view of software programmers is that software programmers' productivity will at least double. There are three or four software companies that are trying to do that. I've invested in all of them in the spirit, and they're all trying to make software programmers more productive.
The most interesting one that I just met with is called augmented. And I always think of an individual programmer, and they said, that's not our target. Our target is these 100-person software programming teams on millions of lines of code where nobody knows what's going on. Well, that's a really good AI thing. Will they make money? I hope so. A lot of questions here. Hi.
So at the very beginning. Yes, ma'am. At the very beginning, you mentioned that there's the combination of the context window expansion, the agents, and the text to action is going to have unimaginable impacts. First of all, why is the combination important? And second of all, I know that you're not like a crystal ball and you can't necessarily tell the future, but why do you think it's beyond anything that we could imagine?
I think largely because the context window allows you to solve the problem of Recency. The current models take a year to train, roughly six. There's 18 months, six months of preparation, six months of training, six months of fine-tuning. So they're always out of date. Context window, you can feed what happened. Like, you can ask it questions about the Hamas Israel war in a context that's very powerful. It becomes current. Like Google. Yeah.
So that's essentially how search GPT works. For example, the new search product from OpenAI can scour the web, scrape the web, and then take all of that information and put it into the context text window. That is the Recency he's talking about in the case of agents. I'll give you an example.
I set up a foundation which is funding a nonprofit, which starts. There's a. I don't know if there's chemists in the room that I don't really understand chemistry. There's a tool called Chemcrow, C R O W, which was an LLM-based system that learned chemistry. And what they do is they run it to generate chemistry hypotheses about proteins, and they have a lab which runs the tests overnight, and then it learns.
That's a huge acceleration accelerant in chemistry, material science, and so forth. So that's an agent model. And I think the text to action can be understood by just having a lot of cheap programmers. Right. And I don't think we understand what happens. And this is, again, your area of expertise, what happens when everyone has their own programmer. And I'm not talking about turning on and off the lights. You know, I imagine another example, for some reason you don't like Google.
So you say, build me a Google competitor. Yeah. You personally build me a Google competitor. Search the web, build a UI, make a good copy, add generative AI in an interesting way, do it in 30 seconds and see if it works. Right. So a lot of people believe that the incumbents, including Google, are vulnerable to this kind of an attack.
Now we'll see. How can we stop AI from influencing public opinion Misinformation, especially during the upcoming election. What are the short and long-term solutions? Most of the Misinformation in this upcoming election, globally will be on social media. And the social media companies are not organized well enough to police it. If you look at TikTok, for example, there are lots of accusations that TikTok is favoring one kind of Misinformation over another. And there are many people who claim, without proof that I'm aware of, that the Chinese are forcing them to do it. I think we just, we have a mess here and the country's going to have to learn critical thinking.
That may be an impossible challenge for the US, but the fact that somebody told you something does not mean that it's true. I think that the greatest threat to democracy is Misinformation because we're going to get really good at it. When I ran managed YouTube, the biggest problems we had on YouTube were that people would upload false videos, and people would die as a result. And we had a no-death policy. Shocking. Yeah.
And also, it's not even about potentially making deep fakes or kind of Misinformation. Just muddying the waters is enough to make the entire topic kind of untouchable. I'm really curious about the text to action and its impact on, for example, computer science education. I'm wondering what you have thoughts on, like how CN's education should transform to kind of meet the age. Well, I'm assuming that computer scientists as a group in undergraduate school will always have a programmer buddy with them. So when you learn your first for loop and so forth and so on, you'll have a tool that will be your natural partner, and then that's how the teaching will go on, that the professor he or she will talk about the concepts, but you'll engage with it that way.
And that's my guess. Yes, ma'am. Behind you. So here I have a slightly different view. I think in the long run, there probably isn't going to be the need for programmers. Eventually, the LLMs will become so sophisticated, they're writing their own kind of code. Maybe it gets to a point where we can't even read that code anymore.
So there is this world in which it is not necessary to have programmers, researchers, or computer scientists. I'm not sure that's the way it's going to be, but there is a timeline in which that happens. The most interesting country is India, because the top AI people come from India to the US, and we should let India keep some of its top talent, not all of them, but some of them. And they don't have the kind of training facilities and programs that we so richly have here. To me, India is the big swing state in that regard.
China's lost. It's not going to come back. They're not going to change the regime as much as people wish them to do. Japan and Korea are clearly in our camp. Taiwan is a fantastic country whose software is terrible.
So that's not going to work. Amazing hardware. And in the rest of the world, there are not a lot of other good choices that are big. Europe is screwed up because of Brussels. It's not a new fact. I spent ten years fighting them and I worked really hard to get them to fix the EU act.
And they still have all the restrictions that make it very difficult to do our kind of research in Europe. My French friends have spent all their time battling Brussels, and Macron, who's a personal friend, is fighting hard for this. And so France, I think, has a chance. I don't see Germany coming, and the rest is not big enough. Given the capabilities that you envision these models having, should we say, still spend time learning to code? Yeah. So here she asked, should we still learn to code? Because. Because ultimately it's the old thing of, why do you study English?
If you can speak English? You get better at it, right? You really do need to understand how these systems work. And I feel very strongly. Yes, sir. So these were the most important parts of the interview.
AI, Technology, Innovation, Eric Schmidt, AI Ethics, Competition
Comments ()