06-reference / transcripts

moonshots ep183 eric schmidt superintelligence transcript

Wed Jul 16 2025 20:00:00 GMT-0400 (Eastern Daylight Time) ·transcript ·source: Moonshots Podcast

When do you see what you define as digital super intelligence? >> Uh, within 10 years. >> The AI’s ability to generate its own scaffolding is imminent. Pretty much sure that that will be a 2025 thing. We certainly don’t know what super intelligence will deliver, but we know it’s coming. >> And what do people need to know about that? >> You’re going to have your own polymath. So, you’re going to have the sum of Einstein and Leonardo da Vinci in the equivalent of your pocket. agents are going to happen. This math thing is going to happen. The software thing is going to happen. Everything I’ve talked about is in the positive domain, but there’s a negative domain as well. It’s likely, in my opinion, that you’re going to see. >> Now, that’s a moonshot, ladies and gentlemen. >> Hey, everybody. Welcome to Moonshots. I’m here live with my Moonshot mate, Dave London. Uh we’re here in our Santa Monica studios and we have a special guest today,

[00:01:00] Eric Schmidt, the author of Genesis. We talk about China. We’re going to talk about, you know, digital super intelligence. We’ll talk about, you know, what people should be thinking about over the 10 years. >> And we’re talking about the guy who has more access to more more actionable information than probably anyone else you could think of. So, it should be should be pretty exciting. >> Incredibly brilliant. All right, stand by for a conversation with the Eric Schmidt, CEO or past CEO of Google and an extraordinary investor and uh and thinker in this field of AI. >> Let’s do it. >> Eric, welcome back to Moonshots. >> It’s great to be here with you guys. >> Thank you. It’s been uh it’s been a long road since I first met you at Google. I remember uh our first conversations were fantastic. Uh it’s been a crazy month in the world of AI, but I think every month from here is going to be a crazy month. And so I’d love to hit on a number of subjects and get your your take on them. >> I want to start with probably the most important point that you’ve made

[00:02:00] recently that got a lot of traction, a lot of attention, which is that AI is underhyped when the rest of the world is either confused, lost, or think it’s, you know, not impacting us. We’ll get into in more detail, but quick most important point to make there. >> AI is a learning machine. Yeah. >> And in network effect businesses, when the learning machine learns faster, everything accelerates. >> It accelerates to its natural limit. The natural limit is electricity. >> Not chips, >> electricity really. Okay. >> So that gets me to the next point here, which is uh a discussion on AI and energy. So, we saw recently was Meta recently announcing uh that they signed a 20-year nuclear contract with uh with Constellation Energy. We’ve seen Google, Microsoft, Amazon, everybody buying basically nuclear capacity right now.

[00:03:02] That’s got to be weird uh that private companies are are basically taking over into their own hands what was utility function before. Um, >> well, just to be cynical, I I’m so glad those companies plan to be around the 20 years that it’s going to take to get the nuclear power plants built. >> In my recent testimony, I talked about the the current expected need for the AI revolution in the United States is 92 gawatt of more power. >> For reference, one gawatt is one big nuclear power station. And there are none essentially being started now. >> And there have been two in the last what, 30 years built. There is excitement that there’s an SMR, small modular reactor coming in at 300 megawws, but it won’t start till 2030. As important as nuclear, both fision and fusion is, they’re not going to arrive in time to get us what we need as a

[00:04:00] globe to deal with our many problems and the many opportunities that are before us. Do you think uh so if if you look at the sort of three-year timeline toward AGI, do you think if you started a a fusion reactor project today that won’t come online for five, six, seven years, is there a probability that the AGI comes up with some other breakthrough fusion or otherwise that makes it irrelevant before it even gets online? >> A very good question. We don’t know what artificial general intelligence will deliver. Yeah. And we certainly don’t know what super intelligence will deliver, but we know it’s coming. So, first we need to plan for it. And there’s lots of issues as well as opportunities for that. But the fact of the matter is that the computing needs that we name now are going to come from traditional energy suppliers in places like the United States and the Arab world and Canada and the Western world. And it’s important to note that China has lots of electricity. So if they get the chips, it’s going to be one heck of a race. >> Yeah. They’ve been scaling it uh you

[00:05:02] know at two or three times. The US has been flat for how long in terms of energy production? >> Um from my perspective uh infinite. In fact, >> electricity demand declined for a while as has overall energy needs because of conservation and other things. >> But the data center story is the story of the energy people, right? And you sit there and you go, how could these data centers use so much power? Well, and especially when you think about how little power our brains do. Well, these are our best approximation in digital form of how our brains work. But when they start working together, they become superbrains. The promise of a superbrain with a 1 gawatt for example data center is so palpable. People are going crazy. And by the way, the economics of these things are unproven. How much revenue do you have to have to have 50 billion in capital? Well, if you depreciate it over three years or four years, you need to have 10 or 15 billion dollars of capital

[00:06:00] spend per year just to handle the infrastructure. Those are huge businesses and huge revenue, which in most places is not there yet. >> I’m curious, there’s so much capital being invested and deployed right now in SMRs in in nuclear bringing Three Mile Island back online. uh in in fusion companies. Why isn’t there an equal amount of capital going into making uh the entire you know chipset and compute just a thousand times more energy efficient? >> There is a similar amount in going in capital. There are many many startups that are working on non-traditional ways of doing chips. The transformer architecture which is what is powering things today has new variants. Every week or so I get a pitch from a new startup that’s going to build inference time, test time computing which are simpler and they’re optimized for inference. It looks like the hardware will arrive just as the software needs expand.

[00:07:00] >> And by the way, that’s always been true. We old-timers had a phrase um grove giveth and gates take it away. So Intel would improve the chipsets right way back when >> and the software people would immediately use it all and suck it all up. I have no reason to believe >> that that’s that that law grove and gates law has changed. If you look at the gains in like the Blackwell chip or the AS uh the the 350 chip in AMD, >> these chips are massive supercomputers and yet we need according to the people have hundreds of thousands of these chips just to make a data center work. That shows you the scale of what this kind of thinking algorithms. Now you sit there and you go what could these people possibly be doing with all these chips? I’ll give you an example. We went from language to language which is what chatbd can be understood at to reasoning and thinking. If you want to look at an open eye example look at open oi03

[00:08:01] >> which go does forward and back reinforcement learning and planning. >> Now the cost of doing the forward and back is many orders of magnitude besides just answering your question for your PhD thesis or your college paper that planning the back and forth is computationally very very expensive. So with the best energy and the best technology today we are able to show evidence of planning. Many people believe that if you combine planning and very deep memories you can build human level intelligence. Now of course they will be very expensive to start with but humans are very very industrious and furthermore the great future companies will have AI scientists that is non-human scientists AI programmers that as opposed to human programmers who will accelerate their impact. So, if you think about it, going back to you’re the author of the abundance thesis, as best I can tell, Peter, you’ve talked about this for 20 years. You saw it first. It sure looks like if we get enough

[00:09:00] electricity, we can generate the power in in the sense of intellectual power to generate abundance along the lines that you predicted two decades ago. >> Every week, I study the 10 major tech meta trends that will transform industries over the decade ahead. I cover trends ranging from humanoid robots, AGI, quantum computing, transport, energy, longevity, and more. No fluff, only the important stuff that matters, that impacts our lives and our careers. If you want me to share these with you, I write a newsletter twice a week, sending it out as a short two-minute read via email. And if you want to discover the most important meta trends 10 years before anyone else, these reports are for you. Readers include founders and CEOs from the world’s most disruptive companies and entrepreneurs building the world’s most disruptive companies. It’s not for you if you don’t want to be informed of what’s coming, why it matters, and how you can benefit from it. To subscribe for free, go to dmadis.com/tatrends.

[00:10:00] That’s dmandis.com/tatrends to gain access to trends 10 plus years before anyone else. >> Let me throw some numbers at you just to reinforce what you said. you know, we have a couple companies in the lab that are doing voice customer service, voice sales with the new, you know, just as of the last month. >> Sure. >> And the value of these these conversations is 10 to $1,000. And the cost of the compute is, you know, maybe two three concurrent GPUs is optimal. It’s like 10 20 cents. And so they would buy massively more compute to improve the the quality of the conversation. There aren’t even close to enough. We we count about 10 million concurrent phone calls that should move to AI in the next year or so. >> And and my view of that is that’s a good tactical solution and a great business. >> Let’s look at other examples of tactical solutions that are great businesses. >> And I obviously have a conflict of interest talking about Google because I love it so much. So with that as in mind, look at the Google strength in GCP. Now Google Google’s cloud product

[00:11:02] >> where they have a completely fully served enterprise offering for essentially automating your company with AI. >> Yeah. >> And the remarkable thing and this is to me is shocking is you can in an enterprise write the task that you want and then using something called the model context protocol you can connect your databases to that and the large language model can produce the code for your enterprise. Now, there’s 100,000 enterprise software companies, middleware companies that grew up in the last 30 years that I’ve been working on this that are all now in trouble because that that interstitial connection is no longer needed >> with their business >> and and and of course they’ll have to change as well. The good news for them is enterprises make these changes very slowly. If you built a brand new enterprise um architecture for ERP and MRP, you would be highly tempted to not use any of the ERP or MRP suppliers, but instead use open- source libraries,

[00:12:01] build essentially use BigQuery or the equivalent from Amazon, which is Red Redshift, and essentially build that architecture and it gives you infinite flexibility and the computer system writes most of the code. Now, programmers don’t go away at the moment. It’s pretty clear that junior programmers go away. The sort of journeymen, if you will, of the stereotype because these systems aren’t good enough yet to automatically write all the code. They need very senior computer scientists, computer engineers who are watching it, that will eventually go away. >> One of the things to say about productivity, and I call this the San Francisco consensus because it’s it’s largely the view of people who operate in San Francisco, >> goes something like this. uh we’re just about to the point where we can do two things that are shocking. The first is we can replace most programming tasks by computers and we can replace both most mathemat mathematical tasks by computers. >> Now you sit there and you go why? Well, if you think about programming and math,

[00:13:02] they have limited language sets compared to human language. So close they’re simpler computationally >> and they’re scale free. You can just do it and do it and do it with more electricity. You don’t need data. You don’t need real world input. You don’t need telemetry. You don’t need sensors. >> Yeah. >> So, it’s likely in my opinion that you’re going to see worldclass mathematicians emerge in the next one year that are AI based and worldclass programmers that going to appear within the next one or two years. When those things are deployed at scale, remember math and programming are the basis of kind of everything, right? It’s an accelerate accelerant for physics, chemistry, biology, material science. So, going back to things like climate change, can you imagine if we and this goes back to your original argument, Peter, imagine if we can accelerate the discoveries of the new materials that allow us to deal with a carbonized world. >> Yeah. >> Right. It’s very exciting. Can I love to

[00:14:01] drill in about you first? >> I just want to hit this because it’s important the potential for there to be I don’t want to use the word PhD level you know other than uh thinking in the terms of research PhD level AIS and uh that can basically attack any problem and solve it uh and solve math if you would in physics. uh this idea of an AI, you know, intelligence explosion. Um Leo Leopold put that at like 26 27 uh heading towards digital super intelligence in the next few years. Do you buy that time frame? >> So again, I consider that to be the San Francisco consensus. I think the dates are probably off by one and a half or two times, >> which is pretty close. So a reasonable prediction is that we’re going to have specialized soants in every field within

[00:15:01] five years. That’s pretty much in the bag as far as I’m concerned. >> Sure. >> And here’s why. You have this amount of humans and then you add a million AI scientists to do something, your slope goes like this. Your rate of improvement, we should get there. The real question is once you have all these sants, do they unify? Do they ultimately become a superhum? The term we’re using is super intelligence, which implies intelligence that beyond the sum of what humans can do. >> The race to super intelligence, which is incredibly important because imagine what a super intelligence could do that we ourselves cannot imagine, right? There it’s so much smarter than we and it has huge proliferation issues, competitive issues, China versus the US issues, electricity issues, so forth. We don’t even have the language for the deterrence aspects and the proliferation issues of these powerful models >> or the imagination. >> Totally agree. In fact, it’s it’s one of

[00:16:00] the great flaws actually in the original conception. You remember Singularity University and Ray Curtzwhile’s books and everything. And we kind of drew this curve of rat level intelligence, then cat, then monkey, and then it hits human and then it goes super intelligent. But it’s now really obvious when you talk to one of these multilingual models that’s explaining physics to you that it’s already hugely super intelligent within its savant category. And so Dennis keeps redefining AGI day as well when it can discover relativity the same way Einstein did with data that was available up until that date. That’s when we have AGI. >> So long before that. >> Yeah. So I think it’s worth getting the timeline right. >> Yeah. >> So the following things are baked in. You’re going to have an agentic revolution where agents are connected to solve business processes, government processes and so forth. They will be adopted most quickly in companies in country companies that have a lot of money and a lot of uh time latency issues at stake. It will adop be adopted

[00:17:00] most slowly in places like government which do not have an incentive for innovation. Um and fundamentally are job programs and redistribution of income kind of programs. >> So call it what you will. The important thing is that there will be a tip of the spear in places like financial services, certain kind of bio biomedical things, startups and so forth. And that’s the place to watch. So all of that is going to happen. The agents are going to happen. This math thing is going to happen. The software thing is going to happen. We can debate the rate at which the biological revolution will occur, but everyone agrees that it’s right after that. We’re very close to these major biological understandings. Um in physics you’re limited by data but you can generate it synthetically. There are groups which I’m funding which are generating physics um essentially um models that can approximate algorithms that cannot be they’re incomputable. So in other words you have a a essentially a foundation model that can answer the question good enough for the purposes of

[00:18:01] doing physics without having to spend a million years doing the computation of you know quantum chromodnamics and things like that. Yep. >> Um, all of that’s going to happen. The next questions have to do with what is the point in which this becomes a national emergency >> and it goes something like this. Everything I’ve talked about is in the positive domain, but there’s a negative domain as well. The ability for biological attacks, um, uh, obviously cyber attacks. Imagine a cyber attack that we as humans cannot conceive of, which means there’s no defense for it because no one ever thought about it. Right? These are real issues. A biological attack, you take a virus, I won’t obviously go into the details. You take a virus that’s bad and you make it undetectable by some changes in its structure, which again I won’t go into the details. We released a whole report at the national level on this issue. So at some point the government and not it

[00:19:02] doesn’t appear to understand this now is going to have to say this is very big because it affects national security, national economic strengths and so forth. Now China clearly understands this and China is putting an enormous amount of money into this. We have slowed them down by virtue of our chips controls but they found clever ways around this. There are also proliferation issues. Many of the chips that they’re not supposed to have, they seem to be able to get. And more importantly, as I mentioned, the algorithms are changing. And instead of having these expensive foundation models by themselves, you have continuous updating, which is called test time training. That continuous updating appears to be capable of being done with lesser power chips. So, so we I there are so many questions that I think we don’t know. We don’t know the role of open source because remember open source means open weights, which means everyone can use it. A fair reading of this is that every country that’s not in the west will end up using open source

[00:20:01] because they’ll perceive it as cheaper which trans transfers leadership in open source from America to China. That’s a big deal, right? If that occurs. >> Um how much longer do the chip bans if you will hold and how long before China can answer? >> What are the effects of the current uh government’s policies of getting rid of foreigners and foreign investment? what happens with the Arab U data centers assuming they work and I’m generally supportive of them um if those things are then misused uh to help train train models. The list just goes on and on. We just don’t know. Okay. Can I ask you probably one of the toughest questions? I don’t know if you saw Mark Andre uh he went and talked to the Biden administration past administration and said how are we going to deal with exactly what you just talked about chemical and biological and radiological and nuclear risks from big foundation models being operated by foreign countries. And the Biden answer was you know we’re going to keep it into the three or four big companies like Google

[00:21:00] and we’ll just regulate them. And Mark was like, “That is a surefire way to lose the race with China because all innovation comes from a startup that you didn’t anticipate or you know it’s just the American history and you’re you’re cutting off the entrepreneur from participating in this.” So as of right now with the open source models, the entrepreneurs are in great shape. But if you think about the models getting crazy smart a year from now, how are we going to have the the balance between startups actually being able to work with the best technology but proliferation not percolating to every country in the world. >> Again, a set of unknown questions and anybody who knows the answer to these things is not telling the full truth. >> Um the doctrine in the B administration was called 10 to the 26 flops. It was a point that was a consensus above which the models were powerful enough to cause some damage. So the theory was that if you stayed below 10 the 26 you didn’t need to be regulated.

[00:22:00] >> But if you were above that you needed to be regulated. And the proposal in the Biden administration was to regulate both the open source and the closed source. >> Okay that’s that’s the those are the the summary >> that of course has been ended by the Trump administration. um they have not yet produced their own thinking in this area. They’re very concerned about China and it getting forward. So, they’ll come out with something. From my perspective, the the core questions are the following. Will the Chinese be able to use even with um chip restrictions, will they use architectural changes that will allow them to build models as powerful as ours? >> And let’s assume they’re government funded. That’s the first question. The next fun question is how will you raise $50 billion for your data center if your product is open source? >> Yeah. >> In the American model, part of the reason these models are closed is that the business people and the lawyers correctly are saying I’ve got to sell this thing because I’ve got to pay for my capital. These are not free goods.

[00:23:00] And the US government correctly is not giving $50 billion to these companies. >> So we don’t know that. Um the to me the key question to watch is look at Deepseek. So Deepseek um a week or so ago Gemini 2.5 Pro got to the top of the leaderboards in intelligence. Great achievement for my friends at Gem at Gemini. A week later deepseek comes in and is slightly better than Gemini. and Deeps of course is trained on the existing hardware that’s in China which includes stuff that’s been Pilford and some of the Ascend it’s called the Ascend Huawei chips and a few others what happens now the US people say well you know the the deepseek people cheated and they cheated by doing a technique called distillation where you take a large model and you ask it 10,000 questions you get its answers and then then you use that as your training material >> yep >> so the US companies will have to figure out a way to make sure that their

[00:24:01] proprietary information that they’ve spent so much money on does not get leaked into these open source things. Um I just don’t know with respect to uh nuclear, biological, chemical and so forth issues. Um the US companies are doing a really good job of looking for that. There’s a great concern, for example, that nuclear information would leak into these models as they’re training without us knowing it. And by the way, that’s a violation of law. >> Oh, really? they work and the whole nuclear information thing is is there’s no free speech in that world for good reasons >> and there’s no free use and copyright and all that kind of stuff. It’s illegal to do it and so they’re doing a really really good job of making sure that that does not happen. They also put in very significant tests for biological information and certain kinds of cyber attacks. What happens there? Their incentive is their incentive to continue especially if it’s not if it’s not required by law. The government has just gotten rid of the the safety institutes that were in place in Biden and are

[00:25:00] replacing it by a new term which is largely a safety assessment program which is a fine answer. I think collectively we in the industry just want the government at the secret and top secret level to have people who are really studying what China and others are doing. You can be sure that China really has very smart people studying what we’re doing. We at the secret and top secret level should have the same thing. >> Have you read the uh AI27 paper? >> I have. Uh, and so for those listening who haven’t read it, it’s a it’s a future vision of the AI and US and China racing towards AI and at some point the story splits into a we’re going to slow down and work on alignment or we’re going full out and uh, you know, spoiler alert and the race to infinity uh, humanity vanishes. So the right outcome will ultimately be some form of deterrence and mutually assured

[00:26:00] destruction. Uh I wrote a paper with two other authors Dan Hendricks and Alex Wang where we named it mutual AI malfunction. >> And the idea was goes something like this. Um you’re the United States, I’m China, you’re ahead of me. Um at some point you cross a line. You know, you Peter cross a line and I China go this is unacceptable. At some point it becomes >> in terms of amount of compute and amount >> it’s it’s something you’re doing where it affects my sovereignty. >> It’s not just words and yelling and an occasional shooting down a jet. It’s it’s a real threat to the identity of my my country, my economic what have you. Under this scenario, I would be highly tempted to do a cyber attack to slow you down. Okay? In mutually assured mal malfunction, if you will, we have to engineer it so that you have the ability to then do the same thing to me.

[00:27:00] >> And that causes both of us to be careful not to trigger the other. >> That’s what mutual assured destruction is. That’s our best formulation right now. We also recommend in our work, and I think it’s very strong, that the government require that we know where all the chips are. And remember, the chips can tell you where they are because they’re computers. Yeah. >> And it would be easy to add a little crypto thing, which would say, “Yeah, here I am, and this is what I’m doing.” >> So, so knowing where the chips are, knowing where the training runs are, and knowing what these fault lines are are very important. Now, there are a whole bunch of assumptions in this scenario that I described. The first is that there was enough electricity. The second is that there was enough power. The third is the Chinese had enough electricity, which they do, and enough computing resources, which they may or may not have >> or may in the future have, >> and may in the future have. And also, I’m asserting that everyone arrives at this eventual state of super intelligence at a roughly the same time. Again, these are debatable points, but

[00:28:02] the most interesting scenario is we’re saying it’s 1938. the letter has come, you know, from Einstein to the president and we’re having a conversation and we’re saying,“Well, how does this end?” >> Okay. So, if you were so brilliant in 38, what you would have said is this ultimately ends with us having a bomb, the other guys having a bomb, and then we’re going to have one heck of a negotiation to try to make sure that we don’t end up destroying each other. And I think the same conversation needs to get started now, well before the Chernobyl events, well before the buildups. >> Can I just take that one more step? And and don’t answer if you don’t want to, but if it was 1947, 1948, so before the Cold War really took off, and you say, well, that’s similar to where we are with China right now. We have a competitive lead, but it may or may not be fragile. What would you do differently 1947 1940 or what would Kissinger do different 1947 1948 1949 than what we did do?

[00:29:03] >> You know I I wrote two books with Dr. Kissinger and I miss him very much. He was my closest friend. Um and Henry was very much a realist in the sense that when you look at his history in uh roughly 36 38 he and his uh I guess 37 38 his family were were Jewish were forced to immigrate from uh Germany because of the Nazis >> and he watched the entire world that he’d grown up with as a boy be destroyed by the Nazis and by Hitler and then he saw the confilgration that occurred as a result and I tell you that whether you like him or not, he spent the rest of his life trying to prevent that from happening again. >> Mhm. >> So we we are today safe because people like Henry saw the world fall apart. >> Mhm. >> So I think from my perspective, we should be very careful in our language and our strategy to not start that

[00:30:00] process. Henry’s view on China was different from other China scholars. His view was in China was that we shouldn’t poke the bear, that we shouldn’t talk about Taiwan too much and we let China deal with our own problems which were very significant. But he was worried that we or China in a small way would start World War II in the same way that World War I was started. You remember that World War One one, World War I started with a essentially a small geopolitical event which was quickly escalated for political reasons on on all sides >> and then the rest was a horrific war, the war to end all wars at the time. >> So we have to be very very careful when we have these conversations not to isolate each other. Um Henry started a number of what are called track two dialogues which I’m part of one of them to try to make sure we’re talking to each other. And so somebody who’s a a hardcore person would say, well, you know, we’re Americans and we’re better and so forth. Well, I can tell you having spent lots of time on this, the Chinese are very smart, very care

[00:31:01] capable, very much up here. And if you’re confused about that, again, look at the arrival of Deep Seek. A year ago, I said they were two years behind. >> I was clearly wrong. >> With enough money and enough power, >> they’re in the game. >> Yeah. Let me actually drill in just a little bit more on that too because I think um one of the reasons deep sea caught up so quickly is because it turned out that inference time generates a lot of IQ and I don’t think anyone saw that coming and inference time is a lot easier to catch up on and also if you take one of our big open source models and distill it >> and then make it a specialist like you were saying a minute ago and then you put a ton of infra time compute behind it, it’s a massive advantage and also a ma massive leak of capability within CBRN for example that nobody anticipated >> and CBNN remember is chemical, biological, radiological and nuclear. >> Um let me rephrase what you said.

[00:32:00] If the structure of the world in 5 to 10 years is 10 models and I’ll make some numbers up. Five in the United States, three in China, two elsewhere. And those models are data centers that are multi- gigawatts. They will be all nationalized in some way. >> In China, they will be owned by the government. >> Mhm. >> The stakes are too high. >> Mhm. Um, one in my military work one day I visited a place where we keep our plutonium and we keep our plutonium in in a base that’s inside of another base with even more machine guns and even more specialized because the plutonium is so is so interesting and and obviously very dangerous and I believe it’s the only one or two facilities that we have in America. So in that scenario, these data centers will have the equivalent of guards and machine guns because they’re so important. Now is that a stable geopolitical

[00:33:00] system? Absolutely. You know where they are. President of one country can call the other. They can have a conversation. You know, they can agree on what they agree on and so forth. But let’s say the it is not true. Let’s say that the technology improves again unknown to the point where the kind of technologies that I’m describing are implementable on the equivalent of a small server >> then you have a humongous data center proliferation problem and that’s where the open-source issue is so important because those servers which will be proliferate throughout the world will all be on open source. We have no control regime for that. Now, I’m in favor of open source as you mentioned earlier with Mark Andre u uh that open competition and so forth tends to allow people to run ahead in defense of the proprietary companies. Collectively, they believe as best I can tell that the open- source models can’t scale fast enough because they need this

[00:34:01] heavyweight training. If you look, I I’ll give you an example of Grock is trained on a single cluster that was built by Nvidia in 20 days or so forth in Memphis, Tennessee of 200,000 GPUs. Um GPU is about $50,000. You can say it’s about a $10 billion supercomput in one building that does one thing, right? If that is the future, then we’re okay because we’ll be able to know where they are. >> Yeah. If in fact the arrival of intelligence is ultimately a a distributed problem, then we’re going to have lots of problems with terrorism, bad actors, North Korea poorly, >> which is my which is my greatest concern. Right. China and the US are rational actors. >> Yeah. >> Uh the terrorist who has access to this and I I don’t want to go all negative on this on this podcast. It’s it’s an important thing to wake people up to the deep thinking you’ve done on this. Um my concern is is the terrorist who gains

[00:35:01] access and >> are we spending enough time and energy and are we training enough models to watch them. >> So the first the companies are doing this there are there’s a body of work happening now which can be understood as follows. You have a super intelligent model. Can you build a model that’s not as smart as the student that’s studying? You know, there is a professor that’s watching the student, >> but the student is smarter than the professor. Is it possible to watch what it does? It appears that we can. >> It appears that there’s a way even if you have a this rogue incredible thing, we can watch it and understand what it’s doing and thereby control it. Another example of the of where where we don’t know is that it’s very clear that these savant models will proceed. There’s no question about that.

[00:36:00] The question is how do we get the Einsteins? So there are two possibilities. One and this is to discover completely new schools of thought >> which is what’s the most exciting thing. Yeah. And in our book Genesis, Henry and I and Craig talk about the importance of polymaths in history. In fact, the first chapter is on polymaths. What happens when we have millions and millions of polymaths? Very, very interesting. >> Okay. >> Now, it looks like the great discoveries, the greatest scientists and people in our history had the following property. They were experts in something and they looked at some at a different problem and they saw a pattern >> in one area of thinking that they could apply to a completely unrelated field and they were able to do so and make a huge breakthrough. The models today are not able to do that. So one thing to watch for is algorithmically

[00:37:00] >> when can they do that? This is generally known as the non-stationerity problem. Yeah. because uh the reward functions in these models are fairly straightforward. You know, beat the human, beat the question and so forth. But when the rules keep changing, is it possible to say the old rule can be applied to a new rule to discover something new? >> And and again, the research is underway. We won’t know for years. >> Peter and I were over at OpenAI yesterday, actually, and we were talking to many people, but Noan Brown in particular, and um I said the word of the year is scaffolding. And he said, “Yeah, maybe the word of the month is scaffolding.” I was like, “Okay, what did I step on there?” He said, “Look, you know, right now, if you try to get the AI to discover relativity or, you know, just some green field opportunity, it won’t it won’t do it. If you set up a framework kind of like a lattice, like a trellis, the vine will grow on the trellis beautifully, but you have to lay out those pathways and breadcrumbs.” He was saying the AI’s ability to generate its own scaffolding is imminent.

[00:38:01] >> Mhm. That doesn’t make it completely self-improving. It’s not it’s not Pandora’s box, but it’s also much deeper down the path of create an entire breakthrough in physics or create an entire feature length movie or you know these these prompts that require 20 hours of consecutive inference time compute pretty much sure that that will be a 2025 thing at least from from their point of view. So, uh, recursive self-improvement is the general term for the computer continuing to learn. >> Yeah, >> we’ve already crossed that >> in the sense that these systems are now running and learning things and they’re learning from the way they own they think within limited functions. When does the system have the ability to generate its own objective and its own question? >> Does not have that today. >> Yep. That’s another sign. Another sign would be that the system decides to uh

[00:39:01] exfiltrate itself and it takes steps to get it get itself away from the commander the control and command system. Um that has not happened yet. Jim and I hasn’t called you yet and said, “Hi, Eric. Can I >> but but there there are theoreticians who believe that the that the systems will ultimately choose that as a reward function because they’re programmed to, you know, to continue to learn.” >> Uh, another one is access to weapons, right? And lying to get it. So, these are trip wires, >> right? All of each of each of which is a trip wire that we’re we’re watching. And again, each of these could be the beginning of a mini Chernobyl event that would become part of consciousness. I think at the moment the US government is not focused on these issues. They’re focused on other things, economic opportunity, growth, and so forth. It’s all good, but somebody’s going to get focused on this and somebody’s going to pay attention to it and it will ultimately be a problem. A quick aside,

[00:40:00] you probably heard me speaking about fountain life before and you’re probably wishing, “Peter, would you please stop talking about fountain life?” And the answer is no, I won’t. Because genuinely, we’re living through a healthc care crisis. You may not know this, but 70% of heart attacks have no precedent, no pain, no shortness of breath. And half of those people with a heart attack never wake up. You don’t feel cancer until stage three or stage 4, until it’s too late. But we have all the technology required to detect and prevent these diseases early at scale. That’s why a group of us including Tony Robbins, Bill Cap, and Bob Heruri founded Fountain Life, a one-stop center to help people understand what’s going on inside their bodies before it’s too late and to gain access to the therapeutics to give them decades of extra health span. Learn more about what’s going on inside your body from Fountainife. Go to fountainlife.com/per and tell them Peter sent you. Okay, back to the episode. Can I can I clean up one kind of common misconception there because um I think it’s a really important one. In the movie version of

[00:41:00] AI, you described, hey, maybe there are 10 big AIs and five are in the US, three are in China, and two are one’s not in Brussels, probably one’s maybe in Dubai. >> Um or, you know, Israel. >> Israel. Okay, there you go. >> Some somewhere like that. >> Yeah. Um in the movie version of this, if it goes rogue, you know, the SWAT team comes in, they blow it up, and it’s it’s solved. But the actual real world is when you’re using one of these huge data centers to create an super intelligent AI, the training process is 10 E26, 10 E28, you know, or more flops. But then the final brain can be ported and run on four GPUs, 8 GPUs, so a box about this size. >> Um, and it’s just as intelligent, you know, it’s it’s it’s and that’s one of the beautiful things about it is you >> This is called stealing the weights. >> Stealing the weights. Exactly. And the new new thing is that that weight file with if you have an innovation in inference time speed and you say oh same weights no difference distill it or or

[00:42:01] just quantize it or whatever but I made it a 100 times faster now it’s actually far more intelligent than what you exported from the data center and so the but all of these are examples of the proliferation problem >> and I’m not convinced that we will hold these things in the 10 places. >> And and here’s why. Let’s assume you have the 10, which is possible. They will have subsets of models that are smaller but nearly as intelligent. And so the tree of knowledge of systems that have knowledge is not going to be 10 and then zero. It’s going to be 10, a h 100red, a thousand, a million, a billion at different levels of complexity. So the system that’s on your future phone may be, you know, three orders of magnitude, four order magnitude smaller than the one at the very tippy top, but it will be very,

[00:43:01] very powerful. >> You know, to exactly what you’re talking about, there’s some great research going on at MIT. It’ll probably move to Stanford just to be fair but it always does but uh it’s great research going on at MIT on uh if you have one of these huge models and it’s been trained on movies it’s been trained on Swahili a lot of the parameters aren’t useful for this soant use case but the general knowledge and intuition is so what’s the optimal balance between narrowing the training data and narrowing the parameter set to be a specialist without losing general you know learning >> so the people who opposed to that view and again we don’t know would say the following. If you take a general purpose model and you specialize it through finetuning it also becomes more brittle. >> Mhm. Mhm. >> Their view is that what you do is you just make bigger and bigger and bigger models because they’re in the big model camp right and that’s why they need gigawatts of data centers and so forth. And their argument is that that flexibility of intelligence that we that

[00:44:01] they are seeing will continue. Dario wrote a a piece called um basically about machines >> and he argued that there >> machines of of grace >> machines of amazing grace >> and he argued that there are three scaling laws playing. The first one is what you know of which is foundation model growth. We’re we’re still on that. The second one is a test time training law and the third one is a reinforcement learning training law. >> Training laws are where if you just put more hardware and more data, they just get smarter in a in a predictable way. Um, we’re just at the beginning in his view of uh this the second and third one beginning. That’s why I I’m sure our audience would be frustrated. Why why do we not know? I’m just we don’t know, >> right? It’s too new. It’s too powerful. And at the moment, all of these

[00:45:00] businesses are incredibly highly valued. They’re growing incredibly quickly. The uses of them, I mentioned earlier, uh going back to Google, um the ability to refactor your entire workflow in a business is a very big deal. That’s a lot of money to be made there for all the companies involved. We will see. Eric, shifting the topic. One of the concerns that people have in the near term and people have been, you know, ringing the alarm bells is on jobs. >> Um, I’m wondering where you come out on this and flipping that forward to education. How do we educate our kids today in high school and college? Uh, and what’s your advice? So on the first thing, do you believe that as Dario has gone on uh you know TV shows now and speaking to significant white collar job loss, we’re seeing obviously a multitude of different drivers and uh robots coming in. How do you think about the job market over the next 5 years? Um

[00:46:03] let’s posit that in 30 or 40 years there’ll be a very different employment robotic human interaction >> or the definition of of do we need to work at all >> the definition of work the definition of identity. Let’s just posit that uh and let’s also posit that it will take 20 or 30 years for those things to work through the economy of our world. Um, now in California and other cities in America, you can get on a Whimo taxi. Um, Whimo, it’s 2025. The original work was done in the late ‘9s. >> The original challenge at Stanford was done, I believe, in 2004. >> The DRA Grand Challenge. It was 2004. >> 20 Sebastian through one. >> So, so more than 20 years from a visible demonstration to our ability to use it in daily life. Why? It’s hard. It’s deep tech. It’s regulated and all of that.

[00:47:00] And I think that’s going to be true, especially in robots that are interacting with humans. They’re going to get regulated. You’re not going to be wandering around and the robots going to decide to slap you. It just doesn’t, you know, societyy’s not going to allow that sort of thing. >> It’s just not, it’s not going to it’s it’s not going to allow it. >> So, in the shorter term, five or 10 years, I’m going to argue that this is positive for jobs in the following way. >> Okay. Um if you look at the history of automation and economic growth, automation starts with the lowest status and most dangerous jobs and then works up the chain. So if you think about assembly lines and cars and you know furnaces and all these sort of very very dangerous jobs that our four forefathers did, they don’t do them anymore. They’re done by robotic solutions of one another and typically not a humanoid robot but an arm. So the so the world dominated by arms that are intelligent and so forth will automate those functions. What

[00:48:02] happens to the people? Well, it turns out that the person who was working with the the welder who’s now operating the arm has a higher wage and the company has higher profits because it’s producing more widgets. So the company makes more money and the person makes more money, right? In that sense. Now you sit there and say well that’s not true because humans don’t want to be retrained. Ah but in the vision that we’re talking about every single person will have a human a computer assistant that’s very intelligent that helps them perform. >> And you take a person of normal intelligence or knowledge and you add a you know sort of accelerant they can get a higher paying job. So you sit there and you go well why are there more jobs? There should be less jobs. That’s not how economics works. Economics expands because the opportunities expands, profits expands, wealth expands and so forth. So there’s plenty of dislocation but in aggregate are there more people

[00:49:02] employed or fewer? The answer is more people with higher paying jobs. >> Is that true in India as well? >> Uh it will be and you picked India because India has a positive demographic outlook although their their birth rate is now down to 2.0. >> Huh. That’s good. the the the rest of the world is choosing not to have children. >> If you look at Korea, it’s now down to.7 children per two parents. >> Yeah. >> China is down to one child per two parents. >> It’s evaporating. >> Now, what happens in those situations? They completely automate everything because it’s the only way to increase national priority. So the most likely scenario, at least in the next decade, is it’s a national emergency to use more AI in the workplace to give people better paying jobs and create more productivity in the United States because our birth rate has been falling. And and what happens is people have talked about this for 20 years. If you if you have this conversation and you ignore demographics, which is negative for humans, and economic growth, which

[00:50:00] occurs naturally because of capital investment, then you miss the whole story. Now, there are plenty of people who lose their jobs, but there’s an awful lot of people who have new jobs. And the typical simple example would be all those people who work in in Amazon distribution centers and Amazon trucks, those jobs didn’t exist until Amazon was created, right? Um the number one shortage in jobs right now in America are truck drivers. Why? Truck driving is a lonely, hard, lowpaying, right? low status of good people job. They don’t want it. They want a better paying job. Right? Going back to education, >> it’s really a crime that our industry has not invented the following product. >> The product that I wanted to build is a product that teaches every single human who wants to be taught in their language in a gamified way the stuff they need to know to be a great citizen in their country. >> Right? That can all be done on phones now. It can all be learned and you can

[00:51:00] all learn how to do it. And why do we not have that product? Right? The investment in the humans of the world is the best return always in knowledge in capability is always the right answer. Let me try and get get your opinion on this because you’re so influential with >> so I’ve got about a thousand people in the companies where I’m the controlling shareholder and I’ve been trying to tell them exactly what you just articulated where a lot of these people have been in the company for 10 15 years. They’re incredibly capable and loyal, but they’ve learned a specific white collar skill. They worked really hard to learn the skill and the AI is coming within no no more than 3 years and maybe two years. And the the opportunity to retrain and have continuity is right now. >> But if they delay, which everyone seems to be just let’s wait and see. And what I’m trying to tell them is if you wait and see, you’re you’re really screwing over that employee. So, so we are in wild agreement that this is going to happen and the winners we the ones who

[00:52:03] act. Now, what’s interesting is when you look at innovation history, the biggest companies who you would think of are the slowest because they have economic resources that the little companies typically don’t, they tend to eventually get there, right? So, watch what the big companies do. Mhm. >> are their CFOs and the people who measure things carefully, who are very very intelligent. They say, “I’m done with that thousand engineering team that doesn’t do very much. I want 50 people working in this other way and we’ll do something else for the other people.” >> And when you say big companies, we’re thinking Google, Meta. We’re not thinking, you know, big bank hasn’t done anything. >> I’m thinking about big banks. Um when when I talk to CEOs and I know a lot of them in traditional industries, what I counsel them is you already have people in the company who know what to do. You just don’t know who they are. >> So call a review of the best ideas to apply AI in our business and ine inevitably the first ones are boring.

[00:53:01] Improve customer service, improve call centers and so forth. But then somebody says, you know, we could increase revenue if we built this product. I’ll give you another example. There’s this whole industry of people who work on regulated user interfaces or one another. I think user interfaces are largely going to go away because if you think about it, the agents speak English typically or other languages. You can talk to them. You can say what you want. The UI can be generated. So I can say generate me a set of buttons that allows me to solve this problem and it’s generated for you. Why do I have to be stuck in what is called the WIMP interface, Windows, icons, menus, and pulld down that was invented in Xerox Park, right, 50 years ago? Why am I still stuck in that paradigm? I just want it to work. >> Yeah. >> Kids in high school and college now, any different recommendations for where they go? When you spend any time in a high school or I was at a conference yesterday where we had a drone challenge

[00:54:00] >> and you watch the 15 year olds, they’re going to be fine. >> They’re just going to be fine. It all makes sense to them and we’re in their way. >> Um, if I were >> digital natives, >> but they’re more than digital natives. They get it. They understand the speed. It’s natural to them. They’re also, frankly, faster and smarter than we are, right? That’s just how life works, I’m sorry to say. So we have wisdom, they have intelligence, they win, right? So in their case, I used to think the right answer was to go into biology. I now actually think going into the application of intelligence to whatever you’re interested in is the best thing you can do as a young person. >> Purpose driven. >> Yeah. >> Any form of solution that you find interesting. Most uh most kids get into it for gaming reasons or something and they learn how to program very young. So they’re quite familiar with this. Um I work uh at a particular university with undergraduates and they’re already doing

[00:55:00] different different algorithms for reinforcement learning as sophomores. This shows you how fast this is happening at their level. They’re going to be just fine. >> They’re responding to the economic signals, but they’re also responding to their purpose. Right? So, an example would be you care about climate, which I certainly do. If you’re a young person, why don’t you figure out a way to simplify the climate science to use simple foundation models to answer these core questions? >> Yeah. >> Why don’t you figure out a way to use these powerful models to come up with new materials, right, that allow us again to address the carbon challenge? And why don’t you work on energy systems to have better and more efficient energy sources that are not that less carbon? You see my point? Yeah, >> you know, I’ve noticed uh because I have kids exactly that that era and um there’s a very clear step function change largely attributable I think to Google and Apple that they have the assumption that things will work >> and if you go just a couple years older during the wimp era like you described it which I’ll attribute more to Microsoft the assumption is nothing will

[00:56:01] ever work like if I try to use this thing it’s going to crash I’m going to be also interesting was that in my career I used to give these speeches about the internet which I enjoyed uh where I said, you know, the great thing about the internet is it has there’s an off button and you can turn off your odd button and you can actually have dinner with your family and then you can turn it on after dinner. This is no longer possible. So the divi the distinction between the real world and the digital world has become confusing. But no one none of us are offline for any significant period of time. >> Yeah. And indeed the the reward system in the world has now caused us to not even be able to fly in peace. Yeah. >> Right. Drive in peace, take a train in peace. >> Star link is everywhere. >> Right. And and that that ubiquitous connectivity has some negative impact in terms of psychological stress uh loss of emotional physical health and so forth. But the benefit of that productivity is without question. >> Every day I get the strangest compliment. Someone will stop me and

[00:57:01] say, “Peter, you have such nice skin.” Honestly, I never thought I’d hear that from anyone. And honestly, I can’t take the full credit. All I do is use something called OneSkin OS1 twice a day every day. The company is built by four brilliant PhD women who’ve identified a peptide that effectively reverses the age of your skin. I love it. And again, I use this twice a day, every day. You can go to onkin.co and write peter at checkout for a discount on the same product I use. That’s oneskin.co co and use the code Peter at checkout. All right, back to the episode. >> Google IO was amazing. >> I mean, just hats off to the entire team there. Um, V3 was shocking and we’re we’re sitting here 8 miles from Hollywood and I’m just wondering your thoughts on the impact this will have. you know, we going to see the oneperson film, feature

[00:58:01] film like we’re seeing potentially oneperson uh unicorns in the future with a with aic. Are we going to see uh an individual be able to compete with a Hollywood studio? And should they be worried about their assets? >> Well, they should always be worried because of intellectual property issues and so forth. Um, I think blockbusters are likely to still be put together by people with an awful lot of help from by AI. Mhm. >> Um I don’t think that goes away. Um if you look at what we can do with generating long- form video, it’s very expensive to do long-term video, although that will come down. And also there’s an occasional extra leg or extra clock or whatever. It’s not perfect yet. And that requires human editing. So even in the scenario where a lot of the the video is created by by a computer, there going to be humans that are producing it and directing it for reasons. My best example in Hollywood is that let’s let’s use the example and I was at at a studio where they were showing me this.

[00:59:01] They had they happened to have an actor who was recreating William Sha Shatner’s movies uh movements a young man and they had licensed the likeness from you know William Shatner who’s now older and they put his head on this person’s body and it was seamless. Well that’s pretty impressive. That’s more revenue for everyone. The an unknown actor becomes a bit more famous, Mr. Shatner gets more revenue, they the whole the whole movie genre works. That’s a good thing. Another example is that nowadays they use green screens rather than sets. And furthermore, in the alien department, when you have, you know, scary movies, instead of having the makeup person, they just add the makeup digitally. >> So, who wins? The costs are lower. the movies are made quicker. In theory, the movies are better, right? Because you have more choices. Um, so everybody wins. Who loses? Well, there was somebody who built that set

[01:00:01] >> and that set isn’t needed anymore. That’s a carpenter and a very talented person who now has to go get a job in the carpentry business. So again, I think people get confused. If I look at at if I look at the digital transformation of entertainment subject to intellectual property being held, which is always a question, it’s going to be just fine, >> right? There’s still going to be blockbusters. The cost will go down, not up, or the or the relative income because in Hollywood, they essentially have their own accounting and they essentially allocate all the revenue to all the key producing people. The the allocation will shift to the people who are the most creative. That’s a normal process. Remember we said earlier that automation gets rid of the poor the lowest quality jobs, the most dangerous jobs there. The jobs that are sort of straightforward are probably automated, but they’re really creative jobs. Um, another example, the script writers. You’re still going to have script writers, but they’re going to have an awful lot of help from AI to write even

[01:01:01] better scripts. That’s not bad. >> Okay. I saw a study recently out of Stanford that documented AI being much more persuasive than the best humans. >> Yes. >> Uh that set off some alarms. It also set off some interesting thoughts on the future of advertising. Any particular thoughts about that? >> So we know the following. We know that if the system knows you well enough, it can learn to convince you of anything. >> Mhm. So what that means in an unregulated environment is that the systems will know you better and better. They’ll get better at pitching you and if you’re not savvy, if you’re not smart, you could be easily manipulated. We also know that the computer is better than humans trying to do the same thing. >> So none of this surprises me. The real question and I’ll ask this in as a question is in the presence of unregulated misinformation engines of

[01:02:00] which there will be many advertisers uh politicians just criminal people people trying to evade responsibility. There’s all sorts of people who have free speech. When they have free speech which includes the ability to use misinformation to their advantage, what happens to democracy? Yeah, >> we we’ve all grown up in democracies where there’s a sort of a a consensus around trust and there’s an elite that more or less administers the trust vectors and so forth. There’s a set of shared values. Do those shared values go away? In our book about Genesis, we talk about this as a deeper problem. What does it mean to be human when you’re interacting mostly with these digital things, >> especially if the digital things have their own scenarios? My favorite example is that uh you have a son or a grandson or a child or a grandchild and you give them a bear and the bear has a personality and the child grows up but the bear grows up too.

[01:03:01] >> So who regulates what the bear talks to the kid? Most people haven’t actually experienced the super super empathetic voice that can be any inflection you want. When they see that which will be in the next probably two months. >> Yeah. they’re going to completely open their eyes to what this >> Well, remember that voice casting was solved a few years ago and that you can cast anyone else’s voice onto your own. >> Yeah. >> And that has all sorts of problems. >> Have you seen uh an avatar yet of somebody that you love that’s passed away or or Henry Kissinger or anything is that? >> Well, we created we actually created one with the permission of his family. >> Did you start crying instantly? >> Uh it’s very emotional. It’s very emotional because, you know, it brings back I mean it’s it’s a real human, >> you know, it’s a real memory, a real voice. Um, and I think we’re going to see more of that. Now, one obvious thing that will happen is at some point in the future when when we naturally die, our digital essence will live in the cloud. >> Yeah. >> And it will know what we knew at the time and you can ask it a question.

[01:04:00] Yeah. >> So, can you imagine asking Einstein, going back to Einstein, >> what did you really think about, >> you know, this other guy, >> you know, did you actually like him or were you just being polite with him with letters? >> Yeah. >> Right. Um, and in all those sort of famous contests that we study as students, >> can you imagine be able to ask the, you know, the people >> Yeah. >> Today, you know, with today’s retrospective, what did you really think? I know that the education example you gave earlier is so much more compelling when you’re talking to Isaac Newton or Albert Einstein instead of just a >> but you know it’s so it’s so >> this is coming back to the V3 in the movies when the one of the first companies we incubated out of MIT course advisor we sold it to Don Graham and the Washington Post and then so I was working for him for a year after that and the conception was here’s the internet here’s the newspaper let’s move the newspaper onto the internet we’ll call it washingtonost.com and if you look hit where it ended up, you know, today with Meta, Tik Tok, YouTube didn’t end up anything like the

[01:05:00] newspaper moves to the internet. >> So now here’s V3, here are movies. You can definitely make a long form movie much more >> cheaply. But I just had this experience of somebody that I know is a complete this director will try and make a tearjerker by leading me down a two-hour long path. But I can get you to that same emotional state in about five minutes if it’s personalized to you. Well, one of the things that’s happened because of the addictive nature of the internet is we’ve lost um sort of the deep state of reading. >> Mhm. >> So, I was walking around and I saw a Borders, sorry, a Barnes & Noble bookstore. Big, oh my god, my old home is back and I went in and I felt good. >> But it’s a very fond memory. But the fact of the matter is that people’s attention spans are shorter. >> They consume things quicker. One of the things interesting about sports is the sports highlights business is a huge business. Licensed clips around highlights because it’s more efficient than watching the whole game. >> So, I suspect that if you’re with your

[01:06:01] buddies and you want to have be drinking and so forth, you put the game on, that’s fine. But if you’re a busy person and you’re busy with whatever you’re busy of and you want to know what happened with your favorite team, the highlights are good enough. >> Yeah. You have four panes of it going at the same time, too. >> And so, this is again a change and it’s it’s a more fundamental change to attention. Mhm. >> I’ve been work I work with a lot of 20somes in research >> and one of the questions I had is how do they do research in the presence of all of these stimulations and I can answer the question definitively. They turn off their phone. >> Yeah. >> You can’t think deeply as a researcher with this thing buzzing. And remember that that part of the the industry’s goal was to fully monetize your attention. >> Yeah. >> Right. We we essent aside from sleeping and we’re working on having you have less sleep I I guess from stress we’ve essentially tried to monetize all of your waking hours with something some form of ads some form of entertainment some form of subscription that is

[01:07:01] completely antithetical to the way humans traditionally work with respect to long thoughtful examination of principles the time that it takes to be a good human being these are in conflict right now there are various attempts at this. So, you know, my favorite are these digital apps that make you relax. Okay. So, the correct thing to do to relax is to turn off your phone, right? And then relax in a traditional way for, you know, 70,000 human years of existence. >> Yeah. Yeah. I had an incredible experience. I’m doing the flight from MIT to Stanford all the time. >> And, you know, like you said, attention spans are getting shorter and shorter and shorter. The Tik Tok extreme, you know, the clips are so short. This particular flight was my first time brainstorming with Gemini for six hours straight >> and I completely lost track of time and I was we’re I’m trying to figure out it’s a circuit design and chip design for inference time compute and it’s so good at brainstorming with me and bringing back data and so long as the Wi-Fi on the plane is working.

[01:08:00] >> Time went by. So my first experience with technology that went the other direction >> but noticed that you also were not responding to texts and annoyances. You weren’t reading ads. you were deep inside of a system >> which for which you paid a subscription. >> Mhm. >> So if you look at the deep research stuff, one of the questions I have when you do a deep research analysis, I was looking at factory automation for something. Where is the boundary of factory automation versus human automation? It’s some an area I don’t understand very well. It’s very very deep technical set of problems. I didn’t understand it. >> It took 20 12 minutes or so to generate this paper. 12 minutes of these supercomputers is an enormous amount of time. What is it doing? Right. And the answer, of course, the product is fantastic. >> Yeah. You know, to Peter’s question earlier, too, I keep the Google IPO perspectus in my bathroom up in Vermont. It’s 2004. I’ve read it probably 500 times. But I don’t know if you remember. It’s getting a little ratty actually. You’re the only the only person besides me who did the same.

[01:09:01] >> I read it 500 times because I had to. It was. It was legally legally required. Well, I still read it um because because of the misconceptions, it’s just so it’s such a great learning experience. But even before the IPO, if you think back, you know, there’s this big debate about will it be ad revenue, will it be subscription revenue, will it be paid inclusion, will the ads be visible, and all this confusion about how you’re going to make money with this thing. Now, the internet moved to almost entirely ad revenue. But if you look at the AI models, they’re, you know, you got your $20 now $200 subscription and people are signing up like crazy. So, you know, the it’s ultra ultra convincing. Is that going to be a form of ad revenue where it convinces you to buy something or no? Is it going to be subscription revenue where people pay a lot more and there’s no advertising at all? >> No, but you have you have this with Netflix. There was this whole discussion about would would how would you fund movies through ads? And the answer is you don’t. You have a subscription. And the Netflix p people looked at having free movies without a subscription and

[01:10:01] advertising supported and the math didn’t work. So I think both will be tried. I think the fact of the matter is deep research at least at the moment is going to be chosen by wellto-do or professional tasks. >> You are capable of spending that $200 a month. A lot of people don’t afford cannot afford it. >> And that free service remember is the thing that is the stepping stone for that young person man or woman who just needs that access. My favorite story there is that when I when I was at Google and I went to Kenya and Kenya is a great country and I and I was with this computer science professor and he said, “I love Google.” I said, “Well, I love Google, too.” And he goes, “Well, I really love Google.” I said, “I really love Google, too.” And I said, “Why do you really love Google?” He said, “Because we don’t have textbooks.” >> And I thought, “The top computer science program in the nation does not have textbooks.” >> Yeah. Well, let me uh >> let me jump in a couple things here. Uh Eric in in the next few years what moes

[01:11:01] actually exist for startups as AI is coming in and disrupting uh do you have a list? >> Yes, I I’ll give you a simple answer. >> And what do you look for in the companies that you’re investing in? >> So first in the deep tech hardware stuff there’s going to be patents, patents, filings, inventions, you know the hard stuff. Those things are much slower than the software industry in terms of growth and they’re just as important. You know, power systems, all those robotic systems we’ve been waiting for a long time. They’re just it’s just slower for all sorts of hardware is hard. >> Hardware is hard for those reasons. >> In software, it’s pretty clear to me it’s going to be really simple. These software is typically a network effect business where the fastest mover wins. The fastest mover is the fastest learner in an AI system. So what I look for is a is a a company where they have a loop. Ideally, they have a couple of learning loops. So I’ll give you a simple

[01:12:00] learning loop that as you get more people, the more people click and you learn from their click. They they they express their preferences. So let’s say I invent a whole new consumer thing, which I don’t have an idea right now for it, but imagine I did. And furthermore, I said that I don’t know anything about how consumers behave, but I’m going to launch this thing. The moment people start using it, I’m going to learn from them, and I’ll have instantaneous learning to get smarter about what they want. So, I start from nothing. If my learning slope is this, I’m essentially unstoppable. >> I’m unstoppable because I’m my learning advantage by the time my competitor figures out what I’ve done is too great. Yeah. >> Now, how close can my my competitor be and still lose? The answer is a few months. >> Mhm. >> Because the slopes are exponential. >> Mhm. >> And so, it’s likely to me that there will be another 10 fantastic Google scale meta-cale companies. They’ll all be founded on this principle of learning

[01:13:01] loops. And when I say learning loops, I mean in the core product, solving the current problem as fast you can. If you cannot define the learning loop, you’re going to be beaten by a company that can define it. >> And you said 10 meta Googlesized companies. Do you think they’ll there will also be a thousand like if you look at the enterprise software business the you know Oracle on down peopleoft whatever thousands of those or will they all consolidate into those 10 that are domain dominant learning loop companies? Um, I think I’m largely speaking about consumer scale because that’s where the real growth is. >> The problem with learning loops is if your customer is not ready for you, you can only learn at a certain rate. So, it’s probably the case that the government is not interested in learning and therefore there’s no growth in learning loop serving the government. I’m sorry to say that needs to get fixed. >> Yeah. >> Um, educational systems are largely regulated and run by the unions and so

[01:14:01] forth. they’re not interested in innovation. They’re not going to be doing any learning. I’m sorry to say we have to get that has to get fixed. So the ones where there’s a very fast feedback signal are the ones to watch. Another example, uh it’s pretty obvious that you can build a whole new stock trading company where you learn if you get the algorithms right, you learn faster than everyone else and scale matters. So in the presence of scale and fast learning loops, that’s the moat. Now I don’t know that there’s many others there. You do have >> you think brand would be a mode? >> Uh brand matters but less so. What’s interesting is people seem to be perfectly willing now to move from one thing to the other in at least in the digital world. >> And there’s a whole new set of brands that have emerged that everyone is using that are you know the next generations that I haven’t even heard of. >> With within those learning loops you think domain specific synthetic data is a is a big advantage? Well, the answer is whatever it causes faster learning.

[01:15:01] >> There are applications where you have enough training data from humans. There are applications where you have to generate the training data from what the humans are doing. >> Right? So, you could imagine a situation where you had a learning loop where there’s no humans involved where it’s monitoring something, some sensors, but because you learn faster on those sensors, you get so smart, you can’t be replaced by another sensor management company. That’s the way to think about. So, so what about the the capital for the learning loop? Like because um do you know Danielle Roose who runs CE? So Danielle and I are really good friends. We’ve been talking to our governor Mora Healey who’s one of the best governors in the world. >> I agree. >> So there’s a problem in our academic systems where the big companies have all the hardware because they have all the money and the universities do not have the money for even reasonablesiz data centers. I was with one university where after lot lots of meetings they agreed to spend $50 million on a data center which generates less than a thousand GPUs >> right for the entire campus and all the

[01:16:01] research. >> Yeah. >> And that doesn’t even include the terabytes of storage and so forth. So I and others are working on this as a philanthropic matter. The government is going to have to come in with more money for universities for this kind of stuff. >> That is among the best investment. When I was young, I was on a National Science Foundation scholarship for and by the way, I made $15,000 a year. Uh the return to the nation of my that $15,000 has been very good, shall we say, based on the taxes that I pay and the jobs that we have created. >> So core question. So glad you >> so so creating so creating an ecosystem for the next generation to have the access to the systems is important. It’s not obvious to me that they need billions of dollars. >> It’s pretty obvious to me that they need a million dollars, $2 million. Yeah, >> that’s the goal. >> Yeah. >> I want to I want to take a I want to take us in a direction of uh of uh wrapping up on super intelligence and

[01:17:00] the book. Um, we didn’t finish the timeline on super intelligence and I think it’s important to give people a sense of how quickly the self-reerential learning can get and how rapidly we can get to something, you know, a thousand times, a million, a billion times more capable than a human. On the flip side of that, Eric, when I look at my greatest concerns when we get through this 5 to sevenyear period of uh let’s just say rogue actors and stabilization and such. Uh one of the biggest concerns I have is the diminishment of human purpose. Mhm. >> Um, you know, you wrote uh in the book uh and I’ve listened to it uh haven’t read it physically and my kids say you don’t read anymore. >> You you listen to books you don’t read. >> But um you said the real risk is not terminator, it’s drift. Um you argue that AI won’t destroy human uh humanity

[01:18:01] violently, but might slowly erode human values, autonomy, and judgment if left unregulated, misunderstood. So it’s really a Wall-E like future versus a a Star Trek boldly go out there. >> We’re very in the book and my own personal view is it’s very important that human agency be protected. >> Yeah. >> Human agency means the ability to get up in the day and do what you want subject to the law. Right. And it’s perfectly possible that these digital devices can create a form of a virtual prison where you don’t feel that you as a human can do what you want. Right? That is to be avoided. I I’m I’m not worried about that case. I’m more worried about the case that if you want to do something, it’s just so much easier to ask your robot or your AI to do it for you. The the human spirit that wants to overcome a challenge. I mean the unchallenged life is so going to so critical

[01:19:00] >> but but there will be always new challenges. Uh when I was a boy uh one of the things that I did is I would repair my father’s car >> right I don’t do that anymore. When I was a boy I used to mow the lawn. I don’t do that anymore. >> Sure. >> Right. So there are plenty of examples of things that we used to do that we don’t need to do anymore. But there’ll be plenty of things. Just remember the complexity of the world that I’m describing is not a simple world. Just managing the world around you is going to be a full-time and purposeful job. Partly because there will be so many people fighting for misinformation and for your attention and and there’s obviously lots of competition and so forth. There’s lots of things to worry about. Plus, you have all of the people, you know, trying to get your trying to get your your money, create opportunities, deceive you, what have you. So, I think human purpose will remain because humans need purpose. >> That’s the point. And you know there’s lots of literature that the people who have what we would consider to be lowpaying worthless jobs enjoy going to

[01:20:01] work. So the challenge is not to get rid of their job. It’s to make their job more productive using AI tools. They’re still going to go to work. And I to be very clear this notion that we’re all going to be sitting around doing poetry is not happening. Right? In the future there’ll be lawyers. They’ll use tools to have even more complex lawsuits against each other, right? There will be evil people who will use these tools to create even more evil problems. There will be good people who will be trying to deter the evil people. The tools change, but the structure of humanity, the way we work together is not going to change. >> Peter and I were on Mike Sailor’s yacht a couple months ago, and I was complaining that the curriculum is completely broken in all these schools. >> But what I meant was we should be teaching AI. And he said, “Yeah, they should be teaching aesthetics.” And I looked at him, I’m like, “What the hell are you talking about?” He said, “No, in the age of AI, which is imminent, look at everything around you, whether it’s good or bad, enjoyable, not enjoyable,

[01:21:01] it’s all about designing aesthetics.” When the AI is such a force multiplier that you can create virtually anything, what what are you creating and why? And that becomes the challenge. If you look at Vickinstein and the sort of theories of all of this stuff, it is all fundament we’re having a conversation that America has about tasks and outcomes. It’s our culture. But there are other aspects of human life, meaning, thinking, reasoning. We’re not going to stop doing that. >> So imagine if your purpose in life in the future is to figure out what’s going on and to be successful, just figuring that out is sufficient. Because once you figured it out, it’s taken care of for you. >> That’s beautiful, >> right? That provides purpose. >> Yeah. >> Um it’s pretty clear that robots will take over an awful lot of mechanical or manual work. >> Um and for people who like to, you know, I like to repair the car. I don’t do it anymore. I miss it, >> but I I have other things to do with my time. >> Yeah. >> Take me forward. When do you see uh what

[01:22:03] you define as digital super intelligence? >> Uh within 10 years. >> Within 10 years. And what do people need to know about that? >> What do people need to understand and sort of uh prepare themselves for either from as a parent or as a employee or as a CEO? >> One way to think about it is that when digital super intelligence finally arrives and is generally available and generally safe, you’re going to have your own polymath. So you’re going to have the sum of Einstein and Leonardo da Vinci in the equivalent of your pocket. I think thinking about how you would use that gift is interesting. And of course evil people will become more evil, but the vast majority of people are good. Yes, >> they’re well-meaning, right? So going back to your abundance argument, there are people who’ve studied the the n the notion of productivity increases and

[01:23:01] they believe that you can get we’ll see to 30% year-over-year economic growth through abundance and so forth. That’s a very wealthy world. That’s a world of much less disease, many more choices, much more fun if you will, right? Just taking all those poor people and lifting them out of the daily struggle they have. That is a great human goal. That’s focus on that. That’s the goal we should have. Does GDP still have meaning in that world? >> If you include services, it does. Um, one of the things about manufacturing and and everyone’s focused on trade deficits and they don’t understand the vast majority of modern economies are service economies, not manufacturing economies. And if you look at the percentage of farming, it was roughly 98% to roughly 2 or 3% in America over a hundred years. If you look at manufacturing, the heydays in the 30s and 40s and 50s, those percentages are now down. Well, lower than 10%. It’s not because we don’t buy stuff. It’s because the stuff is automat automated. You need

[01:24:00] fewer people. Those there’s plenty of people working in other jobs. So again, look at the totality of the society. Is it healthy? >> If you look in China, it’s easy to complain about them. Um they have now deflation. They have a term where people are it’s called laying down where they lay they they stay at home. They don’t participate in the workforce, which is counter to their traditional culture. If you look at reproduction rates, these countries that are essentially having no children, that’s not a good thing. >> Yeah. >> Right. Those are problems that we’re going to face. Those are the new problems of the age. >> I love that. >> Eric, uh, so grateful for your time. >> Thank you. Thank you both. Um, I I love your show. >> Yeah. Thank you, buddy. >> Thank you. >> Okay. Thank you, guys. If you could have had a 10-year head start on the dot boom back in the 2000s, would you have taken it? Every week, I track the major tech meta trends. These are massive game-changing shifts that will play out over the decade ahead. From humanoid

[01:25:00] robotics to AGI, quantum computing, energy breakthroughs, and longevity. I cut through the noise and deliver only what matters to our lives and our careers. I send out a Metatron newsletter twice a week as a quick two-minute readover email. It’s entirely free. These insights are read by founders, CEOs, and investors behind some of the world’s most disruptive companies. Why? Because acting early is everything. This is for you if you want to see the future before it arrives and profit from it. Sign up at dmandis.com/atrends and be ahead of the next tech bubble. That’s dmmand.com/metats. [Music]