TruthArchive.ai - Tweets Saved By @MilkRoadAI

Saved - February 28, 2026 at 4:48 AM

@MilkRoadAI - Milk Road AI

Dario Amodei is the CEO of Anthropic, just laid out the single biggest financial risk in AI right now.​ It's not whether the technology works. He's pretty confident it will. The risk is whether the money comes back fast enough to justify what's being spent.​ Building AI data centers costs tens of billions of dollars and takes one to two years to finish.​ So right now, he has to decide how much compute to buy for infrastructure that won't even be ready until 2027. He's placing a massive bet today on what revenue will look like two years from now. Anthropic has been growing at roughly 10x per year.​ They went from about $1 billion in early 2025 to around $9 billion by end of 2025 to $14 billion annualized as of February 2026. That growth is insane but he can't just assume it keeps going at that pace forever. If revenue keeps growing 10x a year it would hit $100 billion by end of 2026 and $1 trillion by end of 2027.​ If he bought a trillion dollars worth of compute based on that assumption and revenue came in at even $800 billion instead, there is no hedge on earth that saves him from bankruptcy.​ Being off by just 20% when you've committed that much capital is fatal. If the growth rate slows to 5x instead of 10x, or the timeline shifts by just one year, same result.​ You're basically done. Even if AI becomes genius level in the lab, turning that into actual revenue takes time. He uses the example of disease.​ AI might discover cures for everything, but you still have to manufacture the drug, run clinical trials, get regulatory approval and distribute it globally. COVID vaccines took a year and a half to reach everyone even with the entire world in a panic. Polio has had a vaccine for 50 years and still hasn't been fully eradicated.​ The technology being ready and the revenue actually showing up are two very different timelines. So what does he do? He deliberately under buys.​ He commits to hundreds of billions in infrastructure, not trillions. He accepts the risk that if demand explodes he won't have enough capacity. But he'd rather leave money on the table than bet the entire company on a growth curve that might be off by a year. He also says some of the other AI companies are just throwing money around without doing the math. Committing $100 billion here, $100 billion there, without actually modeling what happens if revenue comes in below expectations.​ He calls it YOLOing. For context, Big Tech is expected to spend around $625 billion on AI infrastructure in 2026 alone.​ AI services are only generating about $25 billion in actual revenue against all of that.​ That's roughly a 4% return on what's being invested. The gap between what's being spent and what's being earned right now is massive. The CEO of one of the top AI companies on the planet is saying out loud that the financial math might not work for a lot of these players.​ The technology is real. The demand is probably coming. But if the revenue wave shows up even one or two years late, some of the biggest companies in the world are going to be sitting on historic losses.​ This is the same dynamic that wiped out telecom companies in the early 2000s. They built the infrastructure for demand that eventually came, but it came too late to save the companies that built it.​ He's basically saying he's trying not to be one of those companies.

Video Transcript AI Summary
Speaker 0 discusses the uncertainty around how fast AI will translate into revenue, noting that even if technology advances quickly, misjudging the pace can be ruinous due to the way data centers are purchased. They reference a concept from Machines of Loving Grace, suggesting we might see a powerful AI country in the data center by 2026 or 2027, and acknowledge a possible one- or two-year error in that hunch. They pose a question: if AI can cure all diseases, how long would it take to deliver cures for everyone? They explain that biological discovery, drug manufacturing, and regulatory processes (citing vaccines during COVID) create delay, such as the vaccine rollout taking about a year and a half. They ask how long from the lab-created AI to actual universal cures, noting polio vaccines have existed for fifty years and eradication remains difficult in remote regions, with the Gates Foundation and others trying to overcome this. The speaker asserts that while economic diffusion may not be as difficult as eliminating polio, there are real limits. They outline their expected acceleration curve: a 10x year-over-year revenue increase. At the start of the year, revenue pace is $10 billion annualized; given the time needed to build and reserve data centers, they ask how much compute to buy for 2027. If revenue grows at 10x annually, it could imply $100 billion in 2026 and $1 trillion by the end of 2027, leading to a potential purchase of about $5 trillion in compute starting in 2027 (a trillion dollars per year for five years). They caution that if revenue is not a trillion dollars, no force could prevent bankruptcy from such a purchase. Thus, they acknowledge risk: either the growth rate remains 10x, slows to 5x, or revenue fails to reach the projected level. They emphasize the need to balance ambitious compute procurement with financial risk, rather than a reckless “YOLO” approach. They observe that some other companies may be acting without fully understanding the risks or performing thorough financial scrutiny. The core message is to behave responsibly, aligning compute investments with anticipated revenue growth and recognizing the potential consequences of overextension.
Full Transcript
Speaker 0: And so we have this uncertainty, is even if the technology goes as fast as I suspect that it will, we don't know exactly how fast it's going to drive revenue. We know it's coming, but with the way you buy these data centers, if you're off by a couple years, that can be ruinous. It is just like how I wrote, you know, in Machines of Loving Grace, I said, look, I think we might get this powerful AI, this country of genius in the data center. That description you gave comes from the Machines of Loving Grace. I said, we'll get that 2026, maybe 2027 again. That is my hunch. Wouldn't be surprised if I'm off by a year or two, but that is my hunch. Let's say that happens. That's the starting gun. How long does it take to cure all the diseases? Right? That's that's one of the ways that like drives a huge amount of of of of economic value. Right? Like, you cure you cure every disease. You know, there's a question of how much of that goes to the pharmaceutical company, to the AI company, but there's an enormous consumer surplus because everyone assuming we can get access for everyone, which I care about greatly, we cure all of these diseases. How long does it take? You have to do the biological discovery. You to manufacture the new drug, you have to go through the regulatory process. I mean, we saw this with vaccines and COVID. There's just this we got the vaccine out to everyone, but it took a year and a half. So my question is, how long does it take to get the cure for everything, which AI is the genius that can, in theory, invent out to everyone? How long from when that AI first exists in the lab to when diseases have actually been cured for everyone? Right? In in you know, we've had a polio vaccine for fifty years. We're still trying to eradicate it in the most remote corners of Africa. And, you know, the Gates Foundation is trying as hard as they can. Others are trying as hard as they can, but, you know, that's difficult. Again, I, you know, I don't expect most of the economic diffusion to be as difficult as that. Right? That's like the most difficult case. But but there's a there's a real dilemma here. And and where I've settled on it is it will be it will be it will be faster than anything we've seen in the world, but it still has its limits. And and so then when we go to buying data centers, you know, you again again, the curve I'm looking at is, okay. We've had a 10x a year increase every year. So beginning of this year, we're looking at 10,000,000,000 in rate of annualized revenue at the beginning of the year. We have to decide how much compute to buy. And it takes a year or two to actually build out the data centers, to reserve the data center. So basically, I'm saying in 2027, how much compute do I get? Well, I could assume that the revenue will continue growing 10x a year, so it'll be 100,000,000,000 at the 2026 and 1,000,000,000,000 at the end of twenty twenty seven. And so I could buy a trillion dollars. Actually, it would be like $5,000,000,000,000 of compute because it would be a trillion dollar a year for five years. Right? I could buy a trillion dollars of compute that starts at the 2027. If my revenue is not a trillion dollars, if it's even 800,000,000,000, there's no force on earth. There's no hedge on earth that could stop me from going bankrupt if I buy that much compute. So even though a part of my brain wonders if it's gonna keep going 10x, I can't buy a trillion dollars a year of compute in 2027. If I'm just off by a year in that rate of growth or if the growth rate is five x a year instead of 10 x a year, then then, you know, then you go bankrupt. And so you end up in a world where, you know, you're supporting hundreds of billions, not trillions, and you accept some risk that there's so much demand that you can't support the revenue, and you accept still some risk that you got it wrong and it's still slow. And so when I talked about behaving responsibly, what I meant actually was not the absolute amount. That that actually was not you know, I think it is true we're spending somewhat less than some of the other players. It's actually the other things like, have we been thoughtful about it? Or are we YOLO ing and saying, oh, we're gonna do a $100,000,000,000 here, a $100,000,000,000 there. I kinda get the impression that, you know, some of the other companies have not written down the spreadsheet, that they don't really understand the risks they're taking. They're just kinda doing stuff because it sounds cool.

@MilkRoadAI - Milk Road AI

Instead of watching a 2-hour movie, watch this to understand the blueprint of our AI future. Save this so you can reference these insights later. https://t.co/7qOvGlhpxD

Video Transcript AI Summary
- The conversation centers on how AI progress has evolved over the last few years, what is surprising, and what the near future might look like in terms of capabilities, diffusion, and economic impact. - Big picture of progress - Speaker 1 argues that the underlying exponential progression of AI tech has followed expectations, with models advancing from “smart high school student” to “smart college student” to capabilities approaching PhD/professional levels, and code-related tasks extending beyond that frontier. The pace is roughly as anticipated, with some variance in direction for specific tasks. - The most surprising aspect, per Speaker 1, is the lack of public recognition of how close we are to the end of the exponential growth curve. He notes that public discourse remains focused on political controversies while the technology is approaching a phase where the exponential growth tapers or ends. - What “the exponential” looks like now - There is a shared hypothesis dating back to 2017 (the big blob of compute hypothesis) that what matters most for progress are a small handful of factors: compute, data quantity, data quality/distribution, training duration, scalable objective functions, and normalization/conditioning for stability. - Pretraining scaling has continued to yield gains, and now RL shows a similar pattern: pretraining followed by RL phases can scale with long-term training data and objectives. Tasks like math contests have shown log-linear improvements with training time in RL, and this pattern mirrors pretraining. - The discussion emphasizes that RL and pretraining are not fundamentally different in their relation to scaling; RL is seen as an RL-like extension atop the same scaling principles already observed in pretraining. - On the nature of learning and generalization - There is debate about whether the best path to generalization is “human-like” learning (continual on-the-job learning) or large-scale pretraining plus RL. Speaker 1 argues the generalization observed in pretraining on massive, diverse data (e.g., Common Crawl) is what enables the broad capabilities, and RL similarly benefits from broad, varied data and tasks. - The in-context learning capacity is described as a form of short- to mid-term learning that sits between long-term human learning and evolution, suggesting a spectrum rather than a binary gap between AI learning and human learning. - On the end state and timeline to AGI-like capabilities - Speaker 1 expresses high confidence (~90% or higher) that within ten years we will reach capabilities where a country-of-geniuses-level model in a data center could handle end-to-end tasks (including coding) and generalize across many domains. He places a strong emphasis on timing: “one to three years” for on-the-job, end-to-end coding and related tasks; “three to five” or “five to ten” years for broader, high-ability AI integration into real work. - A central caution is the diffusion problem: even if the technology is advancing rapidly, the economic uptake and deployment into real-world tasks take time due to organizational, regulatory, and operational frictions. He envisions two overlapping fast exponential curves: one for model capability and one for diffusion into the economy, with the latter slower but still rapid compared with historical tech diffusion. - On coding and software engineering - The conversation explores whether the near-term future could see 90% or even 100% of coding tasks done by AI. Speaker 1 clarifies his forecast as a spectrum: - 90% of code written by models is already seen in some places. - 90% of end-to-end SWE tasks (including environment setup, testing, deployment, and even writing memos) might be handled by models; 100% is still a broader claim. - The distinction is between what can be automated now and the broader productivity impact across teams. Even with high automation, human roles in software design and project management may shift rather than disappear. - The value of coding-specific products like Claude Code is discussed as a result of internal experimentation becoming externally marketable; adoption is rapid in the coding domain, both internally and externally. - On product strategy and economics - The economics of frontier AI are discussed in depth. The industry is characterized as a few large players with steep compute needs and a dynamic where training costs grow rapidly while inference margins are substantial. This creates a cycle: training costs are enormous, but inference revenue plus margins can be significant; the industry’s profitability depends on accurately forecasting future demand for compute and managing investment in training versus inference. - The concept of a “country of geniuses in a data center” is used to describe the point at which frontier AI capabilities become so powerful that they unlock large-scale economic value. The timing is uncertain and depends on both technical progress and the diffusion of benefits through the economy. - There is a nuanced view on profitability: in a multi-firm equilibrium, each model may be profitable on its own, but the cost of training new models can outpace current profits if demand does not grow as fast as the compute investments. The balance is described in terms of a distribution where roughly half of compute is used for training and half for inference, with margins on inference driving profitability while training remains a cost center. - On governance, safety, and society - The conversation ventures into governance and international dynamics. The world may evolve toward an “AI governance architecture” with preemption or standard-setting at the federal level, to avoid an unhelpful patchwork of state laws. The idea is to establish standards for transparency, safety, and alignment while balancing innovation. - There is concern about autocracies and the potential for AI to exacerbate geopolitical tensions. The idea is that the post-AGI world may require new governance structures that preserve human freedoms, while enabling competitive but safe AI development. Speaker 1 contemplates scenarios in which authoritarian regimes could become destabilized by powerful AI-enabled information and privacy tools, though cautions that practical governance approaches would be required. - The role of philanthropy is acknowledged, but there is emphasis on endogenous growth and the dissemination of benefits globally. Building AI-enabled health, drug discovery, and other critical sectors in the developing world is seen as essential for broad distribution of AI benefits. - The role of safety tools and alignments - Anthropic’s approach to model governance includes a constitution-like framework for AI behavior, focusing on principles rather than just prohibitions. The idea is to train models to act according to high-level principles with guardrails, enabling better handling of edge cases and greater alignment with human values. - The constitution is viewed as an evolving set of guidelines that can be iterated within the company, compared across different organizations, and subject to broader societal input. This iterative approach is intended to improve alignment while preserving safety and corrigibility. - Specific topics and examples - Video editing and content workflows illustrate how an AI with long-context capabilities and computer-use ability could perform complex tasks, such as reviewing interviews, identifying where to edit, and generating a final cut with context-aware decisions. - There is a discussion of long-context capacity (from thousands of tokens to potentially millions) and the engineering challenges of serving such long contexts, including memory management and inference efficiency. The conversation stresses that these are engineering problems tied to system design rather than fundamental limits of the model’s capabilities. - Final outlook and strategy - The timeline for a country-of-geniuses in a data center is framed as potentially within one to three years for end-to-end on-the-job capabilities, and by 2028-2030 for broader societal diffusion and economic impact. The probability of reaching fundamental capabilities that enable trillions of dollars in revenue is asserted as high within the next decade, with 2030 as a plausible horizon. - There is ongoing emphasis on responsible scaling: the pace of compute expansion must be balanced with thoughtful investment and risk management to ensure long-term stability and safety. The broader vision includes global distribution of benefits, governance mechanisms that preserve civil liberties, and a cautious but optimistic expectation that AI progress will transform many sectors while requiring careful policy and institutional responses. - Mentions of concrete topics - Claude Code as a notable Anthropic product rising from internal use to external adoption. - The idea of a “collective intelligence” approach to shaping AI constitutions with input from multiple stakeholders, including potential future government-level processes. - The role of continual learning, model governance, and the interplay between technology progression and regulatory development. - The broader existential and geopolitical questions—how the world navigates diffusion, governance, and potential misalignment—are acknowledged as central to both policy and industry strategy. - In sum, the dialogue canvasses (a) the expected trajectory of AI progress and the surprising proximity to exponential endpoints, (b) how scaling, pretraining, and RL interact to yield generalization, (c) the practical timelines for on-the-job competencies and automation of complex professional tasks, (d) the economics of compute and the diffusion of frontier AI across the economy, (e) governance, safety, and the potential for a governance architecture (constitutions, preemption, and multi-stakeholder input), and (f) the strategic moves of Anthropic (including Claude Code) within this evolving landscape.
Full Transcript
Speaker 0: So we talked three years ago. I'm curious, in your view, what has been the biggest update of the last three years? What has been the biggest difference between what it felt like last three years versus now? Speaker 1: Yeah. I would say, actually, the underlying technology, like the exponential of the technology, has has gone, broadly speaking, I would say about about as I expected it to go. I mean, there's like plus or minus, you know, a couple there's plus or minus a year or two here. There's plus or minus a year or two there. I don't know that I would have predicted the specific direction of code. But actually when I look at the exponential, it is roughly what I expected in terms of the march of the models from smart high school student to smart college student to beginning to do PhD and professional stuff, and in the case of code reaching beyond that. The frontier is a little bit uneven. It's roughly what I expected. I will tell you though what the most surprising thing has been. The most surprising thing has been the lack of public recognition of how close we are to the end of the exponential. To me, it is absolutely wild that you have within the bubble and outside the bubble, but you have people talking about these just the same tired old hot button political issues and, like, you know, around us. We're, like, near the end of the exponential. Speaker 0: I I wanna understand what that exponential looks like right now because the first question I asked you when we recorded three years ago was, what's up with scaling? Why does it work? I have a similar question now, but I feel like it's a more complicated question because, at least from the public's point of view, three years ago there were these well known public trends where across many orders of magnitude of compute you could see how the loss improves. And now we have RL scaling and there's no publicly known scaling law for it. It's not even clear what exactly the story is of, is this supposed to be teaching the model skills? Is this supposed to be teaching meta learning? What is the scaling hypothesis at this point? Speaker 1: Yeah. So I have actually the same hypothesis that I had even all the way back in 2017. So in 2017, I think I talked about it last time, but I wrote a doc called the big blob of compute hypothesis. It wasn't about the scaling of language models in particular. When I wrote it, GPT-one had just come out, right? That was one among many things. There was back in those days, was robotics. People tried to work on reasoning as a separate thing from language models. There was scaling of the kind of RL that happened that kind of happened in AlphaGo and that happened at DOTA at OpenAI. And people remember StarCraft at DeepMind, the AlphaStar. So it was written as a more general document. And the specific thing I said was the following, and it's very Rich Sutton put out the bitter lesson a couple years later, but the hypothesis is basically the same. What it says is all the cleverness, all the techniques, all the kind of we need a new method to do something like that doesn't matter very much. There are only a few things that matter, and I think I listed seven of them. One is like how much raw compute you have. The other is the quantity of data that you have. Then the third is kind of the quality and distribution of data. Right? It needs to be a broad distribution of data. The fourth is I think how long you train for. The fifth is you need an objective function that can scale to the moon. So the pre training objective function is one such objective function. Right? Another objective function is the kind of RL objective function that says like you have a goal, you're gonna go out and reach the goal. Within that, of course, there's objective rewards you see in math and coding. And there's more subjective rewards like you see in RL from human feedback or kind of higher order versions of that. Then the sixth and seventh were things around kind of like normalization or conditioning, like just getting the numerical stability so that kind of a big blob of compute flows in this laminar way instead of instead of running into problems. So that was the hypothesis, and it's a hypothesis I still hold. I don't think I've seen very much that is not in line with that hypothesis. And so the pretrained scaling laws were one example of we see there. And indeed, those have continued going. I think now it's been widely reported, we feel good about pre training. Pre training is continuing to give us gains. What has changed is that now we're also seeing the same thing for RL. Right? So we're seeing a pre training phase and then we're seeing like an RL phase on top of that. And with RL, it's it's actually just the same. Even other companies have published in some of their releases have published things that say, Look, we train the model on math contests, AIME or the kind of other things, and how well the model does is log linear and how long we've trained it. We see that as well, and it's not just math contests. It's a wide variety of RL tasks. We're seeing the same scaling in RL that we saw for pretraining. Speaker 0: You mentioned Richard Sutton and the Bitter Lesson. Yeah. I interviewed him last year, and he is actually very non LLM pilled. And if I'm if I don't know if this is his perspective, but one way to paraphrase this objection is something like, look, something which possesses the true core of human learning would not require all these billions of dollars of data and compute and these bespoke environments to learn how to use Excel or how does an you know, how to how to use PowerPoint, how to navigate a web browser. And the fact that we have to build in these skills using these RL environments hints that we're actually lacking this core human learning algorithm, and so we're scaling the wrong thing. And so, yeah, that that does raise the question, why are we doing all this RO scaling if we do think there's something that's gonna be human like in its ability to learn on the fly? Speaker 1: Yeah. Yeah. So I think I think this kind of puts together several things that should be kind of thought of thought of differently. Yeah. I think there is a genuine puzzle here, but it it may not matter. In fact, would guess it probably it probably doesn't matter. So let's take the RL out of it for a second because I actually think RL it's a red herring to say that RL was any different from pretraining in this matter. So if we if we look at pretraining scaling, it it was very interesting back in, you know, 2017 when Alec Radford was doing GPT-one. If you look at the models before GPT-one, they were trained on these datasets that didn't represent a wide distribution of text. Right? You had these very standard language modeling benchmarks, and GPT-one itself was trained on a bunch of think it was fan fiction actually. But it was like literary text, which is a very small fraction of the text that you get. What we found with that in those days, it was like a billion words or something, so small datasets and represented a pretty narrow distribution, right? Like a narrow distribution of kind of what you can see in the world. And it didn't generalize well. If you did better on the forgot I what it was, some kind of fan fiction corpus. It wouldn't generalize that well to kind of the other we had all these measures of how well does the model do at predicting all of these other kinds of texts, you really didn't see the generalization. It was only when you trained over all the tasks on the Internet, when you kind of did a general Internet scrape, right, from something like Common Crawl or scraping links on Reddit, which is what we did for GPT-two. It's only when you do that that you kind of started to get generalization. And I think we're seeing the same thing on RL, that we're starting with first very simple RL tasks like training on math competitions, Then we're kind of moving to, you know, kind of broader broader training that involves things like code as a task. And now we're moving to do kind of many many other tasks. And then I think we're going to increasingly get generalization. So that that kind of takes out the RL versus the pretraining side of it. But I think there is a puzzle here either way, which is that on pretraining, when we train the model on pretraining, you know, we we use like trillions of tokens. Right? And and humans don't see trillions of words. So there is an actual sample efficiency difference here. There there is actually something different that's that's happening here, which is that the models start from scratch and they have to get much more training. But we also see that once they're trained, if we give them a long context length the only thing blocking a long context length is like inference. But if we give them like a context length of a million, they're very good at learning and adapting within that context length. And so I don't know the full answer to this, but I think there's something going on that pre training, it's it's not like the process of humans learning. It's somewhere between the process of humans learning and the process of human evolution. It's like it's somewhere between like, we get many of our priors from evolution. Our brain isn't just a blank slate. Right? Whole books have been written about. I think the language models, they're much more blank slates. They literally start as like random weights, whereas the human brain starts with all these regions. It's connected to all these inputs and outputs. So maybe we should think of pre training and for that matter RL as well as as being something that exists in the middle space between human evolution and, you know, kind of human on on the spot learning. And as the in context learning that the models do as as something between long term human learning and short term human learning. So, you know, there there's this hierarchy of, like, there's evolution, there's long term learning, there's short term learning, and there's just human reaction. And the LOM phases exist along this spectrum, but not necessarily exactly at the same points. There's no analog to some of the human modes of learning. The LOMs are kind of falling between the points. Does that make sense? Speaker 0: Yes. Although some things are still a bit confusing. For example, if the analogy is that this is like evolution, so it's fine that it's not that sample efficient, then like, well, if we're gonna get the kind of super sample efficient agent from in context learning, why are we bothering to build in you know, there's our own environment companies which are it seems like what they're doing is they're teaching it how to use this API, how to use Slack, how to use whatever. It's confusing to me why there's so much emphasis on that if the kind of agent that can just learn on the fly is emerging or is gonna soon emerge or has already emerged. Speaker 1: Yeah. Yeah. So I I I mean, I can't speak for the emphasis of anyone else. I can I can only talk about how we how we think about it? I think the way we think about it is the goal is not to teach the model every possible skill within RL just as we don't do that within pre training. Right? Within pre training, we're not trying to expose the model to every possible way that words could be put together. Right? It's rather that the model trains on a lot of things and then it reaches generalization across pre training. Right? That was the transition from GPT-one to GPT-two that I saw up close, which is like the model reaches a point. I had these moments where I was like, Oh, yeah. You just give the model a list of numbers that's like, This is the cost of the house. This is the square feet of the house, and the model completes the pattern and does linear regression. Not great, but it does it, but it's never seen that exact thing before. The extent that we are building these RL environments, the the goal is is very similar to what is you know, to what was done five or ten years ago with pre training with we're trying to get a we're trying to get a whole bunch of data not because we wanna cover a specific document or a specific skill, but because we want to generalize. Speaker 0: I mean, I think the framework you're laying down obviously makes sense. Like, we're making progress towards AGI. I think the crux is something like, nobody at this point disagrees that we're gonna achieve AGI in this century. And the crux is, you say we're hitting the end of the exponential and somebody else looks at this and says, oh, yeah. We've we're making progress. We've been making progress since 2012, and then 2035 will have a human like agent. And so I wanna understand what it is that you're seeing which makes you think, yeah, obviously, we're seeing the kinds of things that evolution did or that human within human lifetime learning is like in these models. And why think that it's one year away and not ten years away? Speaker 1: I I I actually think of it as like two there's kind of two cases to be made here or like two two claims you could make, one of which is like stronger and the other of which is weaker. So I think starting with the weaker claim, when I first saw the scaling back in 2019, I wasn't sure. This kind of a fiftyfifty thing. I thought I saw something that was my claim was this is much more likely than anyone thinks it is. This is wild. No one else would even consider this. Maybe there's a 50% chance this happens. On the basic hypothesis of, as you put it, within ten years, we'll get to what I call country of geniuses in a data center. I'm at 90% on that. It's hard to go much higher than 90% because the world is so unpredictable. Maybe the irreducible uncertainty would be if we were at 95% where you get to things like, I don't know, may maybe multi you know, multiple companies have, you know, kind of internal turmoil and nothing happens, and then Taiwan gets invaded and, like, all the all the fabs get blown up by missiles and and, you know, and then Now Speaker 0: you would drink to Cisneria. Speaker 1: Yeah. Yeah. Yeah. You you know, just you you could construct a scenario where there's like a 5% chance that it or, you know, you you can construct a 5% world where, like, things get delayed for ten years. That's maybe 5%. There's another 5% which is that I'm very confident on tasks that can be verified. So I think with coding, I'm just except for that irreducible uncertainty, there's just I mean, I think we'll be there in one or two years. There's no way we will not be there in ten years in terms of being able to do it end to end coding. My one little bit, the one little bit of fundamental uncertainty even on long time scales is this thing about tasks that aren't verifiable, like planning a mission to Mars, like, you know, doing some fundamental scientific discovery like like CRISPR, like writing a novel, hard to verify those tasks. I am almost certain that we have a reliable path to get there, but if there was a little bit uncertainty, it's there. On the ten years, I'm 90%, which is about as certain as you can be. I think it's crazy to say that this won't happen by 2035. In some sane world, it would be outside the mainstream. Speaker 0: But but the emphasis on verification hints to me as a lack of a lack of belief that these models are generalized. If you think about humans Yes. We are good at things that both of which we get verifiable reward and things which we don't. You're like, you have a Speaker 1: good start. No. No. No. This is this is why I'm almost sure. We already see substantial generalization from things that that verify to things that don't we're already seeing that. But but Speaker 0: it seems like you were emphasizing this as a spectrum which will split apart, which means you see more progress, and I'm like, but that doesn't seem like how humans Speaker 1: get better. In which we don't make it or or or the world in which we don't get there is the world in which we do we do all the things that are that are verifiable, and then they like you know, many of them generalize, but what we kinda don't get fully there. We don't we don't we don't fully, you know, we don't fully color in this side of the box. It's it's it's not a it's not a binary thing. Speaker 0: But but it also seems to me even if even if in the world where generalization is weak when you only say verifiable domains, it's not clear to me in such a world you could automate software engineering because software like, in some sense, you are, quote, unquote, a software engineer. Yeah. But part of being a software engineer for you involves writing these, like, long memos about your grand vision about That's right. Different things. Speaker 1: And so Well, I don't think that's part of the job of SWE. That's part that's part of the job of the company. I do think SWE involves design documents and other things like that, which by the way, the models are not bad. They're already pretty good at writing comments. So with with again, again, I'm making much weaker claims here than I believe to like, you know, to to to to to kinda set up a you know, to to distinguish between two things. Like, we're we're already almost there for software engineering. We are already almost there. Speaker 0: By by what metric? There's one metric which is like how many lines of code are written by AI? And if you use if you consider other productivity improvements in the course of the history of software engineering, compilers write all the lines of software. And but we there's a difference between how many lines are written and how big the productivity improvement is. Oh, yeah. So And then, like, we're almost there, meaning how big is the productivity improvement, not just how many lines are written. Speaker 1: Yeah. Yeah. So I actually agree with you on this. So I've made this series of predictions on code and software engineering, I think people have repeatedly kind of misunderstood them. Let me lay out the spectrum. I think it was eight or nine months ago or something, I said, The AI model will be writing 90% of the lines of code in three to six months, which happened at least at some places. Right? Happened at Anthropic, happened with many people downstream using our models. That's actually a very weak criterion. Right? People thought I was saying, we won't need 90% of the software engineers. Those things are worlds apart. Right? Like, I would put the spectrum as 90% of code is written by the model. A 100% of code is written by the model, and that's a big difference in productivity. 90% of the end to end SWE tasks, including things like compiling, including things like setting up clusters and environments, testing features, writing memos, 90% of the SWE tasks are written by the models. 100% of today's SWE tasks are are are are written by the models. And and even when when when that happened, it doesn't mean software engineers are out of a job. Like, there's like new higher level things they can do where they can they can manage. And then there's a further down the spectrum like, you know, there's 90% less demand for SWISE, which I think will happen, but like, this this this is a spectrum. And, you know, I I wrote about it in in the adolescence of technology where I went through this kind of spectrum with farming. And so I I actually totally agree with you on that. It's just these are very different benchmarks from each other, but we're proceeding through them super fast. Speaker 0: It seems like in part of your vision, it's like going from 90 to a 100. First, it's gonna happen fast, and two, that somehow that leads to huge productivity improvements. Whereas when I noticed even in greenfield projects that people start with Cloud Code or something, people report starting a lot of projects. And I'm like, do we see in the world out there a renaissance of software, all these new features that wouldn't exist otherwise? And at least so far, it doesn't seem like we see that. And so that does make me wonder, even if even if, like, I never had to intervene on Cloud Code. There is this thing of, there's just the world is complicated, jobs are complicated, and closing the loop on self contained systems, whether it's just writing software or something, how much sort of how much broader gains we would see just from that? And so maybe that makes us this should dilute our estimation of the country of geniuses. Speaker 1: Well, I actually I I like I like simultaneously I simultaneously agree with you, agree that it's a reason why these things don't happen instantly. But at the same time, I think the the effect is gonna be very fast. So like, I don't know. You could have these two poles. Right? One is like, you know, AI is like, you know, it's not gonna make progress. It's slow. Like, it's gonna take, you know, kind of forever to diffuse within the economy. Right? Economic diffusion has become one of these buzzwords that's like a a reason why we're not gonna make AI progress or why AI progress doesn't matter. And and, you know, the other axis is like we'll get recursive self improvement, you know, the whole thing, you know, can't you just draw an exponential line on the on the curve? You know, it's it's we're gonna have, you know, Dyson spheres around the sun and like, you know, so many nanoseconds after we get recursive. I mean, I'm completely caricaturing the view here, but there are these two extremes. But what we've seen from the beginning, at least if you look within Anthropic, there's this bizarre 10x per year growth in revenue that we've seen. Right? So in 2023, it was like 0 to 100,000,000. 2024, it was a 100,000,000 to a billion. 2025, it was a billion to, like, 9 or 10,000,000,000. And then Speaker 0: You guys should've just bought, like, a billion dollars with your own products so you could just, like, have a clean 10 view. Speaker 1: And and the first month of this year, like, that that exponential is you would think it would slow down, but it it would like you know, we we added another few billion to like you know, to to to we added another few billion to revenue in January. And and so, you know, obviously that curve can't go on forever. Right? The GDP is only so large. I would even guess that it bends somewhat this year, but that is like a fast curve. Right? That's a really fast curve, and I would bet it stays pretty fast even as the scale goes to the entire economy. So I think we should be thinking about this middle world where things are extremely fast, but not instant, they take time because of economic diffusion, because of the need to close the loop, because, you know, it's like this fiddly, oh, man. I have to do change management within my enterprise. You know, I have to like I set this up, but I have to change the security permissions on this in order to make it actually work. Or I had this old piece of software that, like, you know, checks the model before it's compiled and and and, like, released, and I have to rewrite it. And, yes, the model can do that, but I have to tell the model to do that, and it has to it has to take time to do that. And and and so I think everything we've seen so far is compatible with the idea that there's one fast exponential that's the capability of the model, and then there's another fast exponential that's downstream of that, which is the diffusion of the model into the economy. Not instant, not slow, much faster than any previous technology, but it has its limits. And this is what we when I look inside Anthropic, when I look at our customers, fast adoption, but not infinitely fast. Speaker 0: Can I try a hot take on you? Yeah. I feel like diffusion is cope that people use to say when it's like if the model wasn't able to do something, they're like, oh, but the diff it's like a diffusion issue. But then you should use the comparison to humans. You would think that the inherent advantages that AIs have would make diffusion a much easier problem for new AIs getting onboarded than new humans getting onboarded, so AI can read your entire Slack and your drive in minutes. They can share all the knowledge that the other copy other copies of the same instance have. You don't have this adverse selection problem when you're hiring AIs because you can just hire copies of a vetted AI model. Hiring a human is, like, so much more hassle. And people hire humans all the time. Right? We pay humans upwards of $50,000,000,000,000 in wages because they're useful even though it's, like in principle, it would be much easier to integrate AIs into the economy than it is to hire humans. So think, like, the diffusion, I feel like, doesn't I really Speaker 1: think diffusion is very real and and doesn't have to doesn't exclusively have to do with limitation limitation limitations on the AI models. Like, again, there are people who use diffusion to to you know, as kind of a buzzword to say this isn't a big deal. I'm not talking about that. I'm not talking about, you know, AI will diffuse at the speed that previous. I think AI will diffuse much faster than previous technologies have, but not infinitely fast. So I'll just give an example of this. Right? There's like quad code. Quad code is extremely easy to set up. If you're a developer, you can kind of just start using Claude code. There is no reason why a developer at a large enterprise should not be adopting Claude code as quickly as individual developer or developer at a startup. We do everything we can to promote it. Right? We sell Claude code to enterprises and big enterprises like big financial companies, big pharmaceutical companies, all of them, they're adopting Claude code much faster than enterprises typically adopt new technology. Right? But but again, it like, it it it it it it it takes time. Like, any given feature or any given product like Claude code or like co work will get adopted by the, you know, the individual developers who are on Twitter all the time, by the, like, series a startups many months faster than than, you know, than they will get adopted by, like, you know, a like large enterprise that does food sales. There are a number of factors. Like, you have to go through legal. You have to provision it for everyone. It has to, you know, like, it has to pass security and compliance. The leaders of the company who are further away from the AI revolution are forward looking, but they have to say, Oh, it makes sense for us to spend 50,000,000. This is what this Claude code thing is. This is why it helps our company. This is why it makes us more productive. And then they have to explain to the people two levels below, and they have to say, okay, we have 3,000 developers. Like, here's how we're gonna roll it out to our developers. And we have conversations like this every day. Like, we are doing everything we can to make Anthropix revenue grow 20 or 30 x a year instead of 10 x a year. Again, many enterprises are just saying, this is so productive. We're gonna take shortcuts in our usual procurement process. Right? They're moving much faster than when we tried to sell them just the ordinary API, which many of them use, but quad code is a more compelling product. But it's not an infinitely compelling product. And I don't think even AGI or powerful AI or country of geniuses in the data center will be an infinitely compelling product. It will be a compelling product enough maybe to get three or five or 10 x a year growth even when you're in the hundreds of billions of dollars, which is extremely hard to do and has never been done in history before, but not infinitely fast. Speaker 0: I buy that it would be a slight slowdown, and maybe this is not your claim, but sometimes people talk about this like, oh, the capabilities aren't there, but because of diffusion. Otherwise, like, we're basically at AGI, and then Speaker 1: I I I don't believe we're basically at AGI. Speaker 0: I think if you had the country of geniuses in a data center, if your company didn't adopt Speaker 1: the country of geniuses a center, we would know it. Right. We would know it if you had the country of geniuses in a data center. Like, everyone in this room would know it. Everyone in Washington would know it. Like, you know, people in rural rural parts might not know it, but but but, like, we would know it. We don't have that now. That is very clear. Speaker 0: As Dario was hinting at, to get generalization, you need to train across a wide variety of realistic tasks and environments. For example, with a sales agent, the hardest part isn't teaching it to mash buttons in a specific database in Salesforce. It's training the agent's judgment across ambiguous situations. How do you sort through a database with thousands of leads to figure out which ones are hot? How do you actually reach out? What do you do when you get ghosted? When an AI lab wanted to train a sales agent, Labelbox brought in dozens of Fortune 500 salespeople to build a bunch of different aural environments. They created thousands of scenarios where the sales agent had to engage with the potential customer, which was role played by a second AI. Labelbox made sure that this customer AI had a few different personas because when you cold call, you have no idea who's gonna be on the other end. You need to be able to deal with a whole range of possibilities. LimbleBox's sales experts monitored these conversations turn by turn, tweaking the role playing agent to ensure it did the kinds of things an actual customer would do. Labelbox could iterate faster than anybody else in the industry. This is super important because RL is an empirical science. It's not a solved problem. Labelbox has a bunch of tools for monitoring agent performance in real time. This lets their experts keep coming up with tasks so that the model stays in the right distribution of difficulty and gets the optimal reward signal during training. Labelbox can do this sort of thing in almost every domain. They've got hedge fund managers, radiologists, even airline pilots. So whatever you're working on, Labelbox can help. Learn more at labelbox.com/vorcash. Coming back to concrete predictions because I think because there's so many different things to disambiguate, it can be easy to talk past each other when we're talking about capabilities. So, for example, when I interviewed you three years ago, I asked her a prediction about what we should we expect three years from now. I think you were right. So when you said we should expect systems, which if you talk to them for the course of an hour, it's hard to tell them apart from a generally well educated human. Yes. I think you were right about that. And I think spiritually, feel unsatisfied because my internal expectation was was that such a system could automate large parts of white collar work. And so it might be more productive to talk about the actual end capabilities you want such a system. Speaker 1: So so I will I will I will basically tell you what what you know, where where where I think we are. Speaker 0: So But let let me let me ask it in a very specific question so that we can figure out exactly what kinds of capabilities we should expect soon. So maybe I'll ask about it in the context of a job I understand well, not because it's the most relevant job, but just because I can evaluate the claims about it. Take video editors. Right? I have video editors. And part of their job involves learning about our audience's preferences, learning about my preferences and tastes and the different trade offs we have and how just over the course of many months building up this understanding of context. And so the skill and ability they have six months into the job, a model that can pick up that skill on the job, on the fly, when should we expect such an AI system? Speaker 1: Yeah. So I guess what you're talking about is we're doing this interview for three hours, and then someone's gonna come in, someone's gonna edit it. They're gonna be like, don't know, Dario scratched his head and we could edit that out and Magnify that. Was this long discussion that is less interesting to people, then there's other thing that's more interesting to people, let's make this edit. So I think the country of geniuses in a data center will be able to do that. The way it will be able to do that is it will have general control of a computer screen. You'll be able to feed this in and it'll be able to also use the computer screen to go on the web, look all your previous interviews, look at what people are saying on Twitter in response to your interviews, talk to you, ask you questions, talk to your staff, look at the history of kind of edits edits that you did, and from that, do the job. Yeah. So I think that's dependent on several things. One that's dependent and and and and I think this is one of the things that's actually blocking deployment, getting to the point on computer use where the models are really masters at using the computer. Right? And we've seen this climb in benchmarks, and benchmarks are always imperfect measures. But OS world went from 5%. I think when we first released computer use a year and a quarter ago, was like maybe 15%. I don't remember exactly, but we've climbed from that to like 65 or 70%. And, you know, there may be harder measures as well, but but I think computer use has to pass a point of reliability. Speaker 0: Can I just ask a follow-up on that Yeah? Before you move on to the next point? I often for years, I've been trying to build different internal LLM tools for myself, and I off often I have these text in, text out tasks, which should be dead center in the repertoire of these models, and yet I still hire humans to do them just because if it's something like identify what the best clips would be in this transcript, and maybe they'll do a seven out of 10 job at them. But there's not this ongoing way I can engage with them to help them get better at the job the way I could with a human employee. And so that missing ability, even if you saw computer use, would still block my ability to offload an actual job to them. Speaker 1: Again, this gets back to what we were talking about before with learning on the job where it's very interesting. I think with the coding agents, I don't think people would say that learning on the job is what is preventing the coding agents from doing everything end to end. They keep getting better. We have engineers at Anthropic who don't write any code. When I look at the productivity, to your previous question, we have folks who say, this this GPU kernel, this chip, I used to write it myself. I just have Claude do it. And so there's this there's this enormous improvement in productivity. And I don't know. Like, when I see Claude code, like, familiarity with the code base or, like, you know, or or a feeling that the model hasn't worked at the company for for a year, that's not high up on the list of complaints I see. And so I think what I'm saying is we're we're like, we're kind of taking a different path. Speaker 0: Don't you think with coding that's because there is external scaffold of memory which exists instantiated in the code base, which I don't know how many other jobs have coding made fast progress precisely because it has its unique advantage that other economic activity doesn't. Speaker 1: But but when you say that, what you're what you're implying is that by reading the code base into the context, I have everything that the human needed to learn on the job. So that would be an example of whether it's written or not, whether it's available or not, a case where everything you needed to know, you got from the context window. Right? And that and that what we think of as learning, oh, man. I started this job. It's gonna take me six months to understand the code base. The model just did it in the context. Speaker 0: Yeah. I honestly don't know how to think about this because there there are people who qualitatively report what you're saying. There was a meter study, I'm sure you saw last year Yes. Where they had experienced developers try to close pull request in repositories that they were familiar with, and those developers reported an uplift. They they reported that they felt more productive with the use of these models. But in fact, if you look at their output and how much was actually merged back in, there's a 20% down lift. They were less productive as a result of these these models. And so I'm trying to square the qualitative feeling that people feel with these models versus, one, in a macro level, where are all the where is this, like, renaissance of software? And then, two, when people do these independent evaluations, why are we not seeing the Yeah. So productivity benefits that we would expect. Speaker 1: Within Anthropic, this is just really unambiguous. Right? We're under an incredible amount of commercial pressure and make it even hard harder for ourselves because we have all the safety stuff we do that I think we do more than than than other companies. So like the the the pressure to survive economically while also keeping our values is is just incredible. Right? We're trying to keep this 10 x revenue curve going. There's like, there is zero time for bullshit. There is zero time for feeling like we're productive when we're not. Like, these tools make us a lot more productive. Like, why why do you think we're concerned about competitors using the tools? Because we think we're ahead of the competitors and, like, we don't we don't wanna excel. We we wouldn't be going through all this trouble if this was secretly reducing reducing our productivity. Like, We see the end productivity every few months in the form of model launches. There's no kidding yourself about this. The models make you more productive. Speaker 0: One, people feeling like they're more productive is qualitatively predicted by studies like this. But two, if I just look at the end output, obviously, you guys are making fast progress. But the fact you know, the the idea was supposed to be with recursive self improvement is that you make a better AI, the AI helps you build a better next AI, etcetera, etcetera. And what I see instead, if I look at the you, OpenAI, DeepMind, is that people are just shifting around the podium every few months. And maybe you think that stops because you've you've won or whatever. But but why why are we not seeing the person with the best coding model have this lasting advantage if in fact there are these enormous productivity gains from the last Speaker 1: coding model? So no. No. No. I I mean I mean I mean, I think it's all like my my model of the situation is there's there's an advantage that's gradually growing. Like, I would say right now, the coding models give maybe, I don't know, a 15, maybe 20% total factor speed up. That's my view. Six months ago, it was maybe 5%, and so it didn't matter. 5% doesn't register. It's now just getting to the point where it's one of several factors that matters, and that's going to keep speeding up. I think six months ago, there were several companies that were at roughly the same point because this wasn't a notable factor. But I think it's starting to speed up more and more. I would also say there are multiple companies that write models that are used for code and we're not perfectly good at preventing some of these other companies from using our models internally. So I think I think everything we're kind of everything we're seeing is consistent with this kind of this kind of snowball model where where there's no hard again, my my my my my theme in all of this is like all of this is soft takeoff, like soft, smooth exponentials, although the exponentials are relatively steep. And so and so we're seeing this snowball gather momentum where it's like 10%, 20%, 25%, you know, 40%. And as you go, yeah, Amdahl's Law, you have to get all the things that are preventing you from closing the loop out of the way. But this is one of the biggest priorities within Anthropic. Speaker 0: Stepping back, I think before in the stack we were talking about, well, when do we get this on the job learning? And it seems like the the point you were making at the coding thing is we actually don't need on the job learning, that you can have tremendous productivity improvements. You can have potentially trillions of dollars of revenue for AI companies without this basic human maybe that's not your claim. You should clarify. But without this basic human ability to learn on the job. But I I just look at, like, in most domains of economic activity, people say, I hired somebody, they weren't that useful for the first few months, and then over time, they built up the context understanding. It's actually hard to define what we're talking about here. But they they got something, and then now now they're they're a power horse and they're so valuable to us. And if AI doesn't develop this ability to learn on the fly, I'm not I'm a bit skeptical that we're gonna see huge changes to the world without that ability. Speaker 1: I think I think I think two things here. Right? There's the state of the technology right now, which is, again, we have these two stages. We have the pre training and RL stage where you throw a bunch of data and tasks into the models and then they generalize. So it's like learning, but it's like learning from more data and and not, you know, not learning over kind of one human or one model's lifetime. So again, this is situated between evolution and and and and human learning. But once you learn all those skills, you have them. Just like with pre training, just how the models know more if I look at a pre trained model, it knows more about the history of samurai in Japan than I do. It knows more about baseball than I do. More about low pass filters and electronics. All of these things, its knowledge is way broader than mine. So I think even just that may get us to the point where the models are better at kind of better at everything. And then we also have, again, just with scaling the kind of existing setup, we have the in context learning, which I would describe as kind of like human on the job learning, but like a little weaker and a little short term. Like, you look at in context learning, the you you give the model a bunch of examples. It does get it. There's real learning that happens in context, and like a million tokens is a lot. That's that's you know, that can be days of human learning. Right? If you think about the model kind of reading a million words, it takes me how long would it take me to read a million? I mean, days or weeks at least. So you have these two things, and I think these two things within the existing paradigm may just be enough to get you the country of geniuses in the data center. I don't know for sure, but I think they're gonna get you a large fraction of it. There may be gaps, but I certainly think just as things are, this I believe is enough to generate trillions of dollars of revenue. That's one. That's all one. Two is this idea of continual learning, this idea of a single model learning on the job. I think we're working on that too. I think there's a good chance that in the next year or two, we also make we also solve that. Again, I think you get most of the way there without it. I think the trillions of dollars I think the trillions of dollars a year market, maybe all of the national security implications and the safety implications that I wrote about in adolescence of technology can happen without it. But I also think we, and I imagine others, are working on it. And I think there's a good chance that that, you know, that we get there within the next year or two. There are a bunch of ideas. I won't go into all of them in detail, but, you know, one is just make the context longer. There's nothing preventing longer context from working. You just have to train at longer context and then learn to serve them at inference, and both of those are engineering problems that we are working on and that I would assume others are working on as well. Speaker 0: Yeah. So this context line increase, it seemed like there was a period from 2020 to 2023 where from GBD three to GBD four Turbo, there was an increase from, like, 2,000 context lines to one twenty eight k. I feel like for the next for the two ish years since then, we've been in the ballpark. Yeah. And when model context lines get much longer than that, people report qualitative degradation in the ability of the model to consider that full context. So I'm curious what you're internally seeing that makes you think, like, oh, 10,000,000 context, 100,000,000 to get human six months learning, billion billion context. Speaker 1: This isn't a research problem. This is an engineering and inference problem. Right? If you wanna serve long context, you have to store your entire KV cache. You have it's difficult to store all the memory in the GPUs, to juggle the memory around. I don't even know the detail. At this point, this is at a level of detail that I'm no longer able to follow, although I knew it in the GPD three era of, like, you know, these are the weights, Speaker 0: these are the Speaker 1: activations you have to store. But, you know, you know, these days, the whole thing has flipped because we have MOE models and and and kind of all of that. But and and this degradation you're talking about, like, again, without getting too specific, a question I would ask is there's two things. There's the context length you train at, and there's a context length that you serve at. If you train at a small context length and then try to serve at a long context length, maybe you get these degradations. It's better than nothing. You might still offer it, but you get these degradations. And maybe it's harder to train at a long context length. Yeah. So, you know, there's there's a lot. Speaker 0: I I I wanna, at the same time, ask about, like, maybe some rabbit holes of, well, wouldn't you expect that if you had to train on longer context length, that would mean that you're able to get sort of like less samples in for the same amount of compute. But before maybe maybe it's not worth diving deep on that. I I wanna get an answer to the bigger picture question, which is like, okay. So I don't feel a preference for a human editor that's been working for me for six months versus an AI that's been working with me for six months. What year do you predict that that will be the case? Speaker 1: I my I mean, you know, my guess for that is, you know, there's there's a lot of problems that are basically like, we can do this when we have the country of geniuses in a data center. And so, you know, my my my my my picture for that is, you know, again, if you if you if you if you know if you made me guess, it's like one to two years, maybe one to three years. It's really hard to tell. I have a I have a strong view, 99, 95% that like all this will happen in ten years, like that's I think that's just a super safe bet. Yeah. And then I have a hunch this is more like a fifty fifty thing, that it's gonna be more like one to two, maybe more like one to three. Speaker 0: So one to three years. Country of geniuses and the slightly less economically valuable task of editing videos. Speaker 1: I I it seems pretty economically valuable, let me tell you. It's just there are a lot of use cases like that. Right? There are Speaker 0: lot of similar Exactly. So you're predicting that within one to three years. And then generally, Anthropic has predicted that by late twenty six, early twenty seven, we will have AI systems that are, quote, have the ability to navigate interfaces available to humans doing digital work today, intellectual capabilities matching or exceeding that of Nobel Prize winners, and the ability to interface with the physical world. And then you gave an interview two months ago with DealBook where you're emphasizing your your company's more responsible compute scaling as compared to your competitors. And I'm trying to square these two views where if you really believe that we're gonna have a country of geniuses, you you want as big a data center as you can get. There's no reason to slow down. The TAM of a Nobel Prize winner that is actually can do everything a Nobel Prize winner can do is, like, trillions of dollars. And so I'm trying to square this conservatism, which seems rational if you have more moderate timelines, with your stated views about AI progress. Speaker 1: Yeah. So so it actually all fits together. And and we go back to this fast, but not infinitely fast diffusion. So, like, let's say that we're making progress at this rate. The technology is making progress this fast. Again, I have very high conviction that it's going we're gonna get there within a few years. I have a hunch that we're gonna get there within a year or two. So a little uncertainty on the technical side, but pretty strong confidence that it won't be off by much. What I'm less certain about is, again, the economic diffusion side. I really do believe that we could have models that are a country of geniuses a country of geniuses in the data center in one to two years. One question is, how many years after that do the trillions in you know, do do the do the trillions in revenue start rolling in? I don't think it's guaranteed that it's going to be immediate. You know, I think it could be one year. It could be two years. I could even stretch it to five years, although I'm skeptical of that. And so we have this uncertainty, which is even if the technology goes as fast as I suspect that it will, we we don't know exactly how fast it's gonna drive revenue. We we know it's coming, but with the way you buy these data centers, if you're off by a couple years, that can be ruinous. It is just like how I wrote, you know, in Machines of Loving Grace, I said, look, I think we might get this powerful AI, this country of genius in the data center. That description you gave comes from the Machines of Loving Grace. I said, we'll get that twenty twenty six, maybe twenty twenty seven again. That is that is my hunch. Wouldn't be surprised if I'm off by a year or two, but, like, that is my hunch. Let's say that happens. That's the starting gun. How long does it take to cure all the diseases? Right? That's one of the ways that drives a huge amount of economic value. Right? You cure every disease. There's a question of how much of that goes to the pharmaceutical company, to the AI company, but there's an enormous consumer surplus because everyone you know, every assuming we can get access for everyone, which I care about greatly, we, you know, we we cure all of these diseases. How long does it take? You have to do the biological discovery. You to manufacture the new drug. You have to go through the regulatory process. We saw this with vaccines and COVID. There's just this, we we got the vaccine out to everyone, but it it took a year and a half. Right? And and so my question is, how long does it take to get the cure for everything, which AI is the genius that can, in theory, invent out to everyone. How long from when that AI first exists in the lab to when diseases have actually been cured for everyone? Right? In in you know, we've had a polio vaccine for fifty years. We're still trying to eradicate it in the most remote corners of Africa. And, you know, the Gates Foundation is trying as hard as they can. Others are trying as hard as they can, but, you know, that's difficult. Again, I, you know, I don't expect most of the economic diffusion to be as difficult as that. Right? That's like the most difficult case. But but there's a there's a real dilemma here, and and where I've settled on it is it will be it will be it will be faster than anything we've seen in the world, but it still has its limits. So then when we go to buying data centers, you again, again, the curve I'm looking at is, okay, we've had a 10 x a year increase every year. So beginning of this year, we're looking at 10,000,000,000 in rate of annualized revenue at the beginning of the year. We have to decide how much compute to buy. It takes a year or two to actually build out the data centers, to reserve the data centers. So basically, I'm saying in 2027, how much compute do I get? Well, I could assume that the revenue will continue growing 10x a year, so it'll be 100,000,000,000 at the 2026 and 1,000,000,000,000 at the 2027. And so I could buy a trillion dollars. Actually, it would be like $5,000,000,000,000 of compute because it would be a trillion dollar a year for for five years. Right? I could buy a trillion dollars of compute that starts at the 2027. And if my if my revenue is not a trillion dollars, if it's even 800,000,000,000, there's no force on earth. There's there's no hedge on earth that could stop me from going bankrupt if I if I buy that much compute. And and so even though a part of my brain wonders if it's gonna keep growing 10x, I can't buy a trillion dollars a year of compute in 2027. If I'm just off by a year in that rate of growth or if the growth rate is five x a year instead of 10 x a year, then then, you know, then you go bankrupt. And and and and and you end up in a world where, you know, you're supporting hundreds of billions, not trillions, and you accept some risk that there's so much demand that you can't support the revenue, and you accept still some risk that you got it wrong and it's still slow. When I talked about behaving responsibly, what I meant actually was not the absolute amount. That that actually was not you know, I think it is true we're spending somewhat less than some of the other players. It's actually the other things like, have we been thoughtful about it? Or are we YOLO ing and saying, oh, we're gonna do a $100,000,000,000 here or a $100,000,000,000 there? I kinda get the impression that, you know, some of the other companies have not written down the spreadsheet, that they don't really understand the risk they're taking. They're just kind of doing stuff because it sounds cool. We've thought carefully about it. Right? We're an enterprise business. Therefore, we can rely more on revenue. It's less fickle than consumer. We have better margins, which is the buffer between buying too much and buying too little. And so I think we bought an amount that allows us to capture pretty strong upside worlds. It won't capture the full 10x a year, and things would have to go pretty badly for us to be for us to be in financial trouble. So I think we've thought carefully and we've made that balance, and and that's what I mean when I say that we're being responsible. Speaker 0: Okay. So it seems like it's possible that we're we actually just have different definitions of a country of a genius in a data center. Because when I think of, like, actual human geniuses, an actual country of human geniuses in a data center, I'm like I would happily buy $5,000,000,000,000 worth of compute to run actual country of human geniuses at a data center. So let's say JPMorgan or Moderna or whatever doesn't wanna use them. Also, I've got a country of geniuses. They'll they'll start their own company. And if, like, they they can't start their own company and they're bottlenecked by clinical trials, it is worth stating with clinical trials. Like, most clinical trials fail because the drug doesn't work. There's no efficacy. Right? Speaker 1: And I make exactly that point in in machines of love and grace. I say the clinical trials are gonna go much faster than we're used to, but not not instant, not infinitely fast. Speaker 0: And then suppose it takes a year to for the clinical trials to work out so that you're getting revenue from that and you can make more drugs. Okay. Well, you've got a country of geniuses, and you're an AI lab, and you have you could use many more AI researchers, and you also think that there's these, like, self reinforcing gains from, you know, smart people working on AI tech. So, like, okay, you can have the That's right. But can have the data center working on, and, like, AI progress. Speaker 1: Is there more gains from buying, like, substantially more gains from buying a trillion dollars a year of compute versus $300,000,000,000 a year of compute. Speaker 0: If your competitor's buying a trillion, yes, there is. Speaker 1: Well, no. There's some gain, but then but again, there's this chance that they go bankrupt before, you know, be again, if you're off by only a year, you destroy yourselves. That's the that's the balance. We're buying a lot. We're buying a hell of a lot. Like, we're not we're we're you know, we're buying an amount that's comparable to that that, you know, the the the the the the biggest players in the game are buying. But but if you're asking me, why why haven't we signed, you know, $1,010,000,000,000,000 of compute starting in starting in mid twenty twenty seven? First of all, it can't be produced. There isn't that much in the world. But but second, what if the country of geniuses comes, but it comes in mid twenty twenty eight instead of mid twenty twenty seven? You go bankrupt. Speaker 0: So if your projection is one to three years, it seems like you should have won $10,000,000,000,000 of compute by 2029? Speaker 1: 2020 and maybe 2020. Speaker 0: I mean The latest? Speaker 1: Like, I mean, you know, you you But, Speaker 0: like, are you interested like, it seems like even in your the longest version of the timelines you state, the compute you are ramping up to build doesn't seem What what accordance Speaker 1: What what makes you think that? Speaker 0: Well, you you as you said, you would want the 10,000,000,000,000 like, human wages, let's say, are on the order of 50,000,000,000,000 a year. Speaker 1: You if you look at so so I won't I won't talk about Anthropic in particular, but if you talk about the industry, like, the amount of compute the industry you know, the the the the amount of compute the industry is building this year is probably in the, I don't know, very low tens of call it ten, fifteen gigawatts next year. It goes up by roughly three x a year, so next year's 30 or 40 gigawatts, and twenty twenty eight might be a 100, 2029 might be 300 gigawatts. Each gigawatt costs maybe 10 I mean, I'm doing the math in my head, but each gigawatt costs maybe $10,000,000,000 border 10 to $15,000,000,000 a year. So you put that all together and you're getting about what you described. You're getting multiple trillions a year by 2028 or 2029. So you're getting exactly that. You're getting you're getting exactly what you predict. Speaker 0: That's for the industry. That that's for the industry. That's right. So suppose Anthropix compute keeps three x ing a year, and then by, like, '27, you have or '27, '28, you have 10 gigawatts. And, like, multiply that by, as you say, 10,000,000,000, so then it's like a 100,000,000,000 a year. But then you're saying the TAM by 2028, Speaker 1: I 20 I don't wanna give exact numbers for Anthropic, but but these numbers are too small. These numbers are too small. Okay. Interesting. Speaker 0: I'm really proud that the puzzles I've worked on with Jane Street have resulted in them hiring a bunch of people from my audience. Well, they're still hiring, and they just sent me another puzzle. For this one, they spent about 20,000 GPU hours training backdoors into three different language models. Each one has a hidden prompt that elicits completely different behavior. You just have to find the trigger. This is particularly cool because finding backdoors is actually an open question in Frontier AI research. Anthropic actually released a couple of papers about sleep operations, and they showed that you can build a simple classifier on the residual stream to detect when a backdoor is about to fire. But they already knew what the triggers were because they built them. Here, you don't, and it's not feasible to check the activations for all possible trigger phrases. Unlike the other puzzles they made for this podcast, Jane Street isn't even sure this one is solvable, but they've set aside $50,000 for the best attempts and write ups. The puzzle's live at janestreet.com/torques, and they're accepting submissions until April 1. Alright. Back to Dario. You've told investors that you plan to be profitable starting in '28, and this is the year where we're, like, potentially getting the country of geniuses at a data center. And this is gonna now unlock all this progress and medicine and health and etcetera etcetera and new technologies. Wouldn't this be particularly exactly the time where you'd want to reinvest in the business and build bigger countries so they can So, be more Speaker 1: I mean, profit profitability is this kind of weird thing in this field. I I like like, I don't think I I don't think in this field profitability is actually a measure of spending down versus investing in the business. Let's just take a model of this. I actually think profitability happens when you underestimated the amount of demand you were gonna get, and loss happens when you overestimated the amount of demand you were going to get because you're buying the data centers ahead of time. So think about it this way. Ideally, you would like and again, these are stylized facts. These numbers are not exact for I'm just trying to make a toy model here. Let's say half of your compute is for training and half of your compute is for inference. And, you know, the inference has some gross margin that's like more than 50%. What that means is that if you were in steady state, you build a data center, if you knew exactly the demand you were getting, would would would get a certain amount of revenue, say, I don't know, let's say you pay a $100,000,000,000 a year for compute, and on $50,000,000,000 a year, you support a $150,000,000,000 of of revenue, and the other 50,000,000,000 are used for training. Basically, you're profitable, you make $50,000,000,000 of profit. Those are the economics of the industry today, or sorry, not today, but that's we're projecting forward in a year or two. The only thing that makes that not the case is if you get less demand than 50,000,000,000, then you have more than 50% of your your data center for research and you're not profitable. So you, you know, you train stronger models, but you're, like, not profitable. If you get more demand than you thought, then your research gets squeezed, but, you know, you're you're you're kind of able to support more inference and you're more profitable. So it's maybe I'm not explaining it well, but but the thing I'm trying to say is you decide the amount of compute first, and then you have some target desire of of inference versus versus training, but that gets determined by demand. It doesn't get determined by What Speaker 0: I'm hearing is the reason you're predicting profit is that you are systematically underestimate under investing in compute. Right? Because if you actually like Speaker 1: compute I'm I'm saying it's hard to predict. So so these things about 2028 and when it will happen, that's our that's our attempt to do the best we can with investors. All of this stuff is really uncertain because of the cone of uncertainty. Like, we could be profitable in 2026 if the if the revenue grows fast enough, and then and then, you know, if we if we overestimate or underestimate the next year, that could swing wildly. Like, I I I what I'm trying to get is you have a model in your head of, like, the the business invest, invest, invest, invest, gets scale, and and and and kind of then becomes profitable. There's a single point at which things turn around. I don't think the economics of this industry work that way. Speaker 0: I see. So if I'm understanding correctly, you're saying because of the discrepancy between the amount of compute we should have gotten and the amount of compute we got, we we were, like, sort of forced to make profit, but that that doesn't mean we're gonna continue making profit. We're gonna, like, reinvest the money because, well, now AI has made so much progress and we want the bigger country of geniuses. And so then back into revenue is high, but losses are also high. Speaker 1: If we if we predict if every year we predict exactly what the demand is going to be, we'll be profitable every year because grow because spending spending 50% of your compute on on 50% of your compute on research, roughly, plus a gross margin that's higher than 50%, and and correct demand prediction leads to profit. That's the that's that's the profitable business model that I think is kind of like there, but, like, obscured by these, like, building ahead and prediction errors. Speaker 0: I I guess you're treating the 50% as a as a sort of, like, you know, just like a given constant. Whereas you in fact, if you if AI progress is fast and you can increase the progress by scaling up more, you just have more than 50% and not make profit. Speaker 1: Here's what I'll say. You might wanna scale up it more. You might wanna scale it up more, but but but, you know, remember the log returns to scale. Right? If if 70% would get you a very little bit of a smaller model through a factor of of 1.4 x, right, like, that extra $20,000,000,000 is is is is, you know, that each each dollar there is worth much less to you because of because because the log linear setup. And so you might find that it's better to invest that that that that it's better to invest that $20,000,000,000 in, you know, in in serving inference or in hiring engineers who are who are who are are who are who are who of better who are kind kind of better at what they're doing. So the the reason I said 50%, that's not that's not exactly our target. It's not exactly gonna be 50%. It'll probably vary vary over time. What what I'm saying is the the the the the, like, log linear return, what it leads to is you spend of order one fraction of the business. Right? Like, not 5%, not 95%. And then it then it then, you know, then then then you get diminishing returns because of the because of the log. Speaker 0: Everyone's trying to say, I'm like convincing Dario to, like, believe in AI progress or something. But, like, you okay. You you don't invest in research because it has diminishing returns, but you invest in the other things you mentioned. Speaker 1: Again again, we're talking about diminishing returns after you're spending 50,000,000,000 a year. Right? Speaker 0: Like, this is a point I'm I'm sure you would make, but, like, diminishing returns on a genius is could be quite high. And more generally, like, what is profit in the market economy? Profit is basically saying the other companies in the market can, like, do more things with this money that I Speaker 1: can't then put aside anthropa. I'm just trying to, like because I I, you know, I don't wanna give information about anthropic is why I'm giving these stylized numbers. But, like, let's just derive the equilibrium of the industry. Right? I think the so so why doesn't everyone spend 100% of their, you know, 100% of their compute on training and not serve any customers? Right? It's because if they didn't get any revenue, they couldn't raise money, they couldn't do compute deals, they couldn't buy more compute the next year. So there's gonna be an equilibrium where every every company spends less than 100% on on on on on training and certainly less than 100% on inference. It should be clear why you don't just serve the current models and and, you know, and and and and never train another model because then you don't have any demand because you'll because you'll fall behind. So there's some equilibrium. It's it's not gonna be 10%. It's not gonna be 90%. Let's just say as a stylized fact, it's 50%. That's what I'm getting at. And and and I think we're gonna be in a position where that equilibrium of how much you spend on training is less than the gross margins that you're able to get on compute. And so the underlying economics are profitable. The problem is you have this hellish demand prediction problem when you're buying the next year of compute, and you might guess under and be very profitable but have no compute for research, or you might guess over and you are not profitable and you have all the compute for research in the world. Does does that make sense? Just as a dynamic model of the industry. Speaker 0: Maybe stepping back, I'm like I I I'm not saying I I think the country of genius is gonna come in two years, and therefore, should buy this compute. To me, what you're saying the end conclusion you're arriving at makes a lot of sense, but that's because it's like, oh, it seems like country geniuses is hard and there's a long way to go. And so the stepping back, the thing I'm trying to get at is more like it seems like your worldview is compatible with somebody who says, we're ten years away from a world in which we're generating trillions of Speaker 1: dollars That's just not my view. Yeah. That is not my view. So I'll make another prediction. It is hard for me to see that there won't be trillions of dollars in revenue before 2030. I can construct a plausible world. It takes maybe three years, so that be the end of what I think it's plausible. Like in 2028, we get the real country of geniuses in the data center. The revenue's been going into the maybe is is in the low hundreds of billions by by by by 2028, and and and then the country of geniuses accelerates it to trillions, you know, and and we're basically we're basically on the slow end of diffusion. It takes two years to get to the trillions. That that that would that that that would be the world where it takes until that would be the world where it takes until 2030. I I I suspect even composing the technical exponential and diffusion exponential will get there before 2030. Speaker 0: So you laid out a model where Anthropic makes profit because it seems like fundamentally, we're in a compute constrained world, and so it's like, eventually, we keep growing compute. Speaker 1: No. I think I think the way the profit comes is again, and and, you know, let's let's just abstract the whole industry here. Like, we have a know, let's just imagine we're we're we're in like an economics textbook. We have a small number of firms. Each can invest a limited amount in you know, or or or like each can invest some fraction fraction in r and d. They have some marginal cost to serve. The margins on that the profit margin the gross profit margins on that marginal cost are very high because inference is efficient. There's some competition, but the models are also differentiated. There's some companies will compete to push their research budgets up, but like because there's a small number of players, you know, we have the what is it called? Cornot equilibrium, I think is what the what the small number of firm equal equilibrium is. It the point is it it doesn't equilibrate to perfect competition with with with with with with with zero margins. If there's, like, three firms if there's three firms in the economy, all are kind of independently behaving behaving rationally, it doesn't equilibrate to zero. Speaker 0: Help me understand that because right now we do have three leading firms and they're not making profit. And so what what what yeah. What what is changing? Speaker 1: Yeah. So the the again, the gross margins right now are very positive. What's happen what what's happening is a combination of two things. One is we're still in the exponential scale up phase of compute. Yeah. So what basically, what that means is we're training like, a model gets trained. Yeah. It costs you know, let's say a model got trained that costs a billion dollars last year. And then this year, it produced $4,000,000,000 of revenue and cost $1,000,000,000 to to to to inference from. So, you know, again, I'm using stylized number here, but, you know, that would be 75%, you know, gross gross gross margins and, you know, this this 25% tax. So that model as a whole makes $2,000,000,000. But at the same time, we're spending $10,000,000,000 to train the next model because there's an exponential scale up, and so the company loses money. Each model makes money, but the company loses money. The equilibrium I'm talking about is an equilibrium where we have the country of geniuses we have the country of geniuses in a data center, but that that model training scale up has equilibrated more. Maybe maybe it's still it's still going up. We're still trying to predict the demand, but it's more it's more leveled out. Speaker 0: I'll give you just a couple of things there. So let's start with the current world. In the current world, you're right that, as you said before, if you treat each individual model as a company, it's profitable. But of course, a big part of the production function of being a Frontier lab is training the next model. Right? So Yes. That's if right. You didn't do that, then you'd make profit for two months. And then you wouldn't have margins because you wouldn't have the best model. And then so yeah. You you can make profit for two months on the current system. Speaker 1: At some point, that reaches the biggest scale that it can reach. And then and then in equilibrium, we have algorithmic improvements, but we're spending roughly the same amount to train the next model as as as we as we spend to Speaker 0: train the current model. So this equilibrium relies I mean, at some point, Speaker 1: at some at some point, you run out of money in the economy. Speaker 0: A fixed lump of labor or fallacy. The economy is gonna grow. Right? That's one of your predictions. Well We're gonna have this this is Data centers this space. Speaker 1: But this is another example of the theme I was talking about, which is that the economy will grow much faster with AI than I think it ever has before. But it's not like right now, the compute is growing three x a year. Yeah. I don't believe the economy is gonna grow 300% a year. Like, I said this in Machines of Love and Grace. Like, I think we we may get 10 or 20% per year growth in the economy, but we're not gonna get 300% growth in the economy. So I think I think in the end, you know, if if compute becomes the majority of what the economy produces, it's it's gonna it's gonna be capped by that. Speaker 0: So let's okay. Now let's assume a model where compute stays capped. Yeah. The world where Frontier Labs are making money is one where they continue to make fast progress because fundamentally, margin is limited by how good the alternative is. And so you are able to make money because you have a frontier model. If you didn't have frontier model, you wouldn't be making money. Well, you you I mean And and so this this model requires there never to be a steady state. Like, forever and ever, you keep making more out of the progress. Speaker 1: I don't think that's true. I mean, I I feel I feel like we're we're, like, we're taught we're we're, know, we're I feel like this is an economics this is like an economics class. You like, know that Tyler Cowen code? Speaker 0: We never stop talking about economics. We never Speaker 1: we never stop talking about economics. So no. But but there there are there are worlds in which, you know, there so think this field's gonna be a I don't think this field's gonna be a monopoly. All my lawyers never want me to say the word monopoly. But I don't think this field's gonna be a monopoly. But but you do get you get industries in which there are small number of players. Not one, but a small number of players. And ordinarily, like, the the way you get monopolies like Facebook or or Meta, I always call them Facebook, but is is these kind of net is these kind of these kind of network effects. Yeah. The way you get industries in which there are small number of players are very high costs of entry. Right? So, you know, cloud is like this. I think cloud is a good example of this. You have three, maybe four players within cloud. I think I think that's the same for AI. Three, maybe four. And the reason is that it's it's so expensive. It requires so much expertise and so much capital to, like, run a cloud company. Right? So you have to put up all this capital and then in addition to putting up all this capital, you have to get all of this other stuff that requires a lot of skill to make it happen. So it's like if you go to someone and you're like, want to disrupt this industry. Here's a $100,000,000,000. You're like, okay. I'm putting a $100,000,000,000 and also betting that you can do all these other things that these people have been doing. Speaker 0: Only to decrease the profit in the industry. Speaker 1: And and then and then the effect of your entering is the is the profit margins go down. So, you know, we have equilibria like this all the time in the economy where we have a few we have a few players. Profits are not astronomical. Margins are not astronomical, but they're they're not zero. Right? And and, you know, I think I think that's what we see on cloud. Cloud is very undifferentiated. Models are more differentiated than cloud. Right? Like, everyone knows Claude is Claude Claude is good at different things than GPT is good at is than than Gemini is good at. And it's not just Claude's good at coding, GPT is good at math and reasoning. It's more subtle than that. Models are good at different types of coding. Models have different styles. I think these things are actually quite different from each other, and so I would expect more differentiation than you see in cloud. Now, there actually is counter there is one counterargument, and that counterargument is that if all of that, the process of producing models becomes if AI models can do that themselves, then that could spread throughout the economy. But that is not an argument for commoditizing AI models in general. That's kind of an argument for commoditizing the whole economy at once. I don't know what what quite happens in that world where basically anyone can do anything, anyone can build anything, and there's like no moat around anything at all. I don't know. Maybe we want that world. Maybe that's the end state here. Maybe when AI models can do everything, if we've solved all the safety and security problems, like, that's one of the one of the one of the for for just just kind of the economy flattening itself again. But that's kind of like post like far post country geniuses in a data center. Speaker 0: Maybe a a finer way to put that potential point is, one, it seems like AI research is especially loaded on raw intellectual power, which will be especially abundant in a world of AGI. And two, if you just look at the world today, there's very few technologies that seem to be diffusing as fast as as AI algorithmic progress. And so that does hint that this industry is sort of structurally diffusive. Speaker 1: So I think coding is going fast, but I think AI research is a superset of coding, and there are aspects of it that are not going fast. But I do think, again, once we get coding, once we get AI models going fast, then that will speed up the ability of AI models to do everything else. So I think while coding is going fast now, I think once the AI models are building the next AI models and building everything else, the kind of whole the whole economy will kind of go at the same pace. I am I am worried geographically, though. I'm a little worried that, like, just proximity to AI, having heard about AI, that that that may be one differentiator. And so when I said the like, you know, 10 or 20% growth rate, a worry I have is that the growth rate could be like 50 in Silicon Valley and, you know, parts of the world that are kind of socially connected to Silicon Valley and, you know, not that much faster than its current pace elsewhere. And I think that'd be a pretty messed up world. So I one of the things I think about a lot is how to prevent that. Speaker 0: Yep. Do you think that once we have this country of geniuses at data center that robotics is sort of quickly solved afterwards because it seems like a big problem with robotics is that a human can learn how to teleoperate current hardware, but current AI models can't, at least not if not in a way that's super productive. And so if we have this ability to learn like a human, should it solve robotics immediately as well? Speaker 1: I don't think it's dependent on learning like a human. It could happen in different ways. Again, we could have trained the model on many different video games, which are like robotic controls or many different simulated robotics environments or just train them to control computer screens and they learn to generalize. So it will happen. It's not necessarily dependent on human like learning. Human like learning is one way it could happen if the model's like, oh, I pick up a robot. I don't know how to use it. I learn. That that could happen because we discovered discovering continual learning. That could also happen because we train the model on a bunch of environments and then generalized, or it could happen because the model learns that in the context length. It it it doesn't actually matter which way. If we go back to the discussion we had like like an hour ago, that type of thing can happen in that type of thing can happen in several different ways. Yeah. But but I do think when for for whatever reason the models have those skills, then robotics will be revolutionized, both the design of robots because the models will be much better than humans at that, and also the the ability to kind of control robots. So we'll get better at the physical building the physical hardware, building the physical robots, and we'll also get better at controlling it. Now, you know, does that mean the robotics industry will also be generating trillions of dollars of revenue? My answer there is yes, but there will be the same extremely fast, but not infinitely fast diffusion. So will robotics be be revolutionized? Yeah. Maybe tack on another year or two. That's the way I think about these things. Speaker 0: Makes sense. There's a general skepticism about extremely fast progress. Here's my view, which is like, it sounds like you are gonna solve continual learning one way or another within a matter of years. But just as people weren't talking about continual learning a couple years ago and then we realized, oh, why aren't these models as useful as they could be right now even though they are clearly passing the Turing test and are experts in so many different domains? Maybe it's this thing. Then And we solve this thing and we realize, actually, there's another another thing that human intelligence can do and that's a basis of human labor that these models can't do. Then so why not think there will be more things like this? So I think that we're we're you know, we've, like, found the pieces of human intelligence. Speaker 1: Well well, to be clear, I mean, I think continual learning, as I've said before, might not be a barrier at all. Yeah. Right? Like like, you know, I think I think we maybe just get there by pretraining generalization and and and RL generalization. Like, I I think there might just might not be there there basically might not be such a thing at all. In fact, I would point to the history in in ML of people coming up with things that are barriers that end up kind of dissolving within the big blob of compute. Right? That people talked about how do your models keep track of nouns and verbs and how do they you know they can understand syntactically, but they can't understand semantically. It's only statistical correlations. You can understand a paragraph, you can't understand a word. There's reasoning, you can't do reasoning, but then suddenly it turns out you can do code and math very well at all. So I I think there act there's there's actually a stronger history of some of these things seeming like a big deal and then and then kind of and then kind of dissolving. Some of them are real. I mean, the need for data is real. May maybe continual continual learn continual learning is a real thing. But, again, I would ground us in something like code. Like, I think we may get to the point in, like, a year or two where the models can just do SWE end to end. Like, that's a whole task. That's a whole sphere of human activity that that we're just saying models can do it now. Speaker 0: But when you say end to end, do you mean setting technical direction, understanding the context of the problem Yes. Etcetera. Okay. Yes. I mean all of that. Interesting. I mean, that that is, I feel like, AGI Complete. Maybe it's internally consistent, but it's not like saying 90% of code or a 100% of code. It's like, no. No. I I I The the other parts of Speaker 1: the job is No. No. I gave this I gave this spectrum. 90% of code, 100% of code, 90% of n 10 SWE, 100% of n 10 SWE, new tasks are created for SWE, eventually those get done as well. Yeah. But there's a long spectrum But we're traversing the spectrum very quickly. Speaker 0: Yeah. I do think it's funny that I've I've seen a couple of podcasts you've done where the host will be like, but Vorkash wrote this essay about the computer learning thing, and it always makes you crack up because you're like, you know, you've been an AI researcher for, like, ten years. I'm sure there's, like, some feeling of, like, okay. So podcasts are wrote an essay. No. And in, like, every interview, I get Speaker 1: asked about it. You know, the the truth of the the truth of the matter is that we're all trying to figure this out together. Yeah. Right? There there are some ways in which I'm able to see things that others aren't. These days, that probably has more to do with, like, I can see a bunch of stuff within anthropic and have to make a bunch of decisions than I have any great research insight that that that others don't. Right? I've you know, I'm running a 2,500 person company. Like, it's it's actually pretty hard for me to have have concrete research insight, you know, much harder than, you know, than than it would have been, you know, ten years ago or or, you know, or even two or three years ago. Speaker 0: As we go towards a world of a full drop in remote worker replacement, does a API pricing model still make the most sense? And if not, what is the correct way to price AGI or serve AGI? Speaker 1: Yeah. I mean, I think there's gonna be a bunch of different business models here sort of all at once that are gonna be that are gonna be experimented with. I I I actually do think that the the API model is is more durable than many people think. One way I think about it is if the technology is kind of advancing quickly, if it's advancing exponentially, what that means is there's always kind of like a surface area of kind of new use cases that have been developed in the last three months. And any kind of product surface you put in place is always at risk of sort of becoming irrelevant. Right? Any given product surface probably makes sense for our, you know, a range of capabilities of the model. Right? The the chatbot is already running into limitations of, you know, making it smarter doesn't really help the average consumer that much. But I don't think that's a limitation of AI models. I don't think that's evidence that, you know, the models are are the models are good enough and they're they're you know, them getting better doesn't matter to the economy. It doesn't matter to that particular product. And and so I think the value of the API is the API always offers an opportunity, you know, very close to the bare metal to build on what the latest thing is. And so there's kind of always gonna be this kind of front of new startups and new ideas that weren't possible a few months ago and are possible because the model is advancing. And and so I I actually I I I kind of actually predict that we are it's gonna exist alongside other models, but we're always gonna have the API business model because there's there's always gonna be a need for a thousand different people to try experimenting with the model in a different way, and a 100 of them become startups and 10 of them become big successful startups and two or three really end up being the way that people use the model of a given generation. So I I basically think it's always gonna exist. At the same time, I'm sure there's gonna be other models as well. Like, not every token that's output by the model is worth the same amount. Think about, you know, how how how what is the value of the tokens that are like, you know, that the model outputs when someone, you know, call you know, someone, you know, calls them up and says, my Mac isn't working or something, you know, the model's like restart it. Right? Yeah. And like, you know, someone hasn't heard that before, but like, you know, the model said that like 10,000,000 times. Right? You know, that's that maybe that's worth like a dollar or a few cents or something. Whereas if the model, you know, the model goes to, you know, one of the one of the pharmaceutical companies and it says, oh, you know, this molecule you're developing, you should take the aromatic ring from that end of the molecule and put it on that end of the molecule. And and, you know, if you do that, wonderful things will happen. Like like those tokens could be worth, you know, tens of millions of dollars. Right? So so I think we're definitely gonna see business models that that recognize that, you know, at some point, we're gonna see, you know, pay for results or, you you know, in some in some form, or we may see forms of compensation that are like labor, you know, that that kind of work by the hour. I I I, you know, I don't know. I think I think I think because it's a new industry, a lot of things are gonna be tried, and I, you know, I don't know what will turn out to be the right thing. Speaker 0: What I find I I take your point that people will have to try things to figure out what is the best way to use this blob of intelligence, but what I find striking is ClaudeCode. So I don't think in the history of startups, there has been a single application that has been as hotly competed in as coding agents. And the Cloud Code is a category leader here. And that seems surprising to me. Like, it doesn't seem intrinsically like Anthropic had to build this. And I wonder if you have an accounting of why it had to be Anthropic or why how Anthropic ended up building an application in addition to the model underlying it. Speaker 1: Yeah. So it actually happened in a pretty simple way, which is we had our own you know, we had our coding models, which were good at coding. And, you know, around the beginning of 2025, I said, I I think the time has come where you can have nontrivial acceleration of your own research if you're an AI company by using these models. And, of course, you know, we you need an interface. You need a harness to use them. So I encourage people internally. I didn't say this is one thing that, you know, that you have to use. I just said people should experiment with this. And then, you know, this thing, I I think it might have been originally called Claude CLI, and then the name eventually got changed to Claude Code internally, was the thing that kind of everyone was using, and it was seeing fast internal adoption. And I looked at it and I said, probably we should launch this externally. Right? It's seen such fast adoption within Anthropic, coding is a lot of what we do, and so we a audience of many hundreds of people that's in some ways at least representative of the external audience. So it looks like we already have product market fit. Let's launch this thing. Then we launched it, and I think you know, just just the fact that we ourselves are kind of developing the model and we ourselves know what we most need to use the model, I think it's it's kinda creating this feedback loop. Speaker 0: I see. In the sense that you let's say a developer at Anthropic is like, it it'd be better if it was better at this x thing. And then you bake that into the next model that you build. Speaker 1: That that's that's one version of it, but but then there's just the ordinary product iteration of like, you know, we have a bunch of we have a bunch of coders within Anthropic. Like, we you know, they they like use quad code every day, and so we get fast feedback. That was more important in the early days. Now, of course, there are millions of people using it, and so we get a bunch of external feedback as well, but it's, you know, it's just great to be able to get, you know, kind of kind of fast fast internal feedback. You know, I think this is the reason why we launched a coding model and, you know, didn't launch a pharmaceutical company. Right? My background's in biology, but we don't have any of the resources that are needed to launch a pharmaceutical company. Speaker 0: There's been a ton of hype around OpenClaw, and I wanted to check it out for myself. I've got a day coming up this weekend, and I don't have anything planned yet. I gave Openclaw a Mercury debit card. I set a couple $100 limit, and I said, surprise me. Okay. So here's the Mac Mini it's on, and besides having access to my Mercury, it's totally quarantined. And I actually felt quite comfortable giving an access to a debit card because Mercury makes it super easy to set up guardrails. I was able to customize permissions, cap the spend, and restrict the category of purchases. I wanted to make sure the debit card worked, so I asked OpenCloud to just make a test transaction and decided to donate a couple bucks to Wikipedia. Besides that, I have no idea what's gonna happen. I will report back on the next episode about how it goes. In the meantime, if you want a personal banking solution that can accommodate all the different ways that people use their money, even experimental ones like this one, visit mercury.com/personal. Mercury is a fintech company, not an FDIC insured bank. Banking services provided through Choice Financial Group and column NA, members FDIC. You know she thinks we're getting coffee and walking around the neighborhood. Let me ask you about now making AI go well. It seems like whatever vision we have about how AI goes well has to be compatible with two things. One is the ability to build and run AIs is diffusing extremely rapidly. And two is that the population of AIs, the amount we have in their intelligence will also increase very rapidly. And that means that lots of people will be able to build huge populations of misaligned AIs or AIs which are just like companies which are trying to increase their footprint or have weird psyches like Sidney Bing, but now they're superhuman. What is a vision for a world in which we have an equilibrium that is compatible with lots of different AIs, some of which are misaligned, running around? Speaker 1: Yeah. Yeah. So I think, you know, in the adolescence of technology, I was kind of, you know, skeptical of, like, the balance of power. But I I think I was particularly skeptical of or the thing I was specifically skeptical of is you have, like, three or four of these companies, like, kind of all building models that are kind of dry you know, sort of sort of, like, derived from the, like, derived from the same thing and, you know, that that these would check each other or or even that kinda, you know, any number of them would would would check each other. Like, we might live in a offense dominant world where, you know, like, one person or one AI model is, like, smart enough to do something that, like, causes damage for everything else. I think in the I mean, in the short run, we have a limited number of players now, so we can start by within the limited number of players. We, you know, we kind of you know, we we need to put in place the, you know, the safeguards. We need to make sure everyone does the right alignment work. We need to make sure everyone has bioclassifiers. Like, you know, those are those are kind of the immediate things we need to do. I agree that, you know, that that doesn't solve the problem in the long run, particularly if the ability of AI models to make other AI models proliferates, then, you know, the the whole thing can kind of, you know, can become harder to solve. You know, I think I think in the long run, we need some architecture of governance. Right? Some some architecture of governance that preserves human freedom, but but kind of also allows us to, like, you know, govern the the very large number of kind of, you know, human systems, AI systems, hybrid hybrid human human, you know, hybrid hybrid human AI, like, you know, companies or or like or like or like economic units. You know, we're we're gonna need to think about, like, you know, how do we how do we protect the world against, you know, bioterrorism? How do we protect the world against, like, you know, against, like, against, like, mirror life? Like, you know, probably probably we're gonna need to, you know, need some kind of like AI monitoring system that like, know, kind of monitors for for all of these things, but then we need to build this in a way that like, you know, preserves civil liberties and like our constitutional rights. So I think just as is anything else, it's like a new security landscape with a new set of tools and a new set of vulnerabilities. And I think my worry is if we had a hundred years for this to happen all very slowly, we'd get used to it. You know, like, we've gotten used to, like, you know, the presence of, you know, the presence of explosives in society or, the, you know, the presence of various you know, like new weapons or the, you know, the the presence of video cameras. We would get used to it over over over over a 100 and we develop governance mechanisms. We'd make our mistakes. My my worry is just that this is happening all so fast. And so think maybe we need to do our thinking faster about how to make these governance mechanisms work. Speaker 0: Yeah. It seems like in an offense dominant world, over the course of the next century so the idea is AI is making the progress that would happen over the next century happen in some period of five to ten years. But we would still need the same mechanisms, or balance of power would be similarly intractable even if humans were the only game in town. And so I guess we have the advice of AI. We fundamentally it doesn't seem like a totally different ballgame here. If checks and balances were gonna work, they would work with humans as well. If they aren't gonna work, they wouldn't work with AIs as well. And so maybe this just dooms human checks and balances as well. Speaker 1: Yeah. Again, I think there's some way to I think there's some way to make this happen. It just the governments of the world may have to work together to make it happen. May have to talk to AIs about kind of, you know, building societal structures in such a way that, like, these these defenses are possible. I I I don't know. I mean, this is so this is you know, I I don't wanna say so far ahead in time, but, like, so far ahead in technological ability that may happen over a short period of time that it's hard for us to anticipate in advance. Speaker 0: Speaking of governments getting involved, on December 26, the Tennessee legislature introduced a bill which said, quote, it would be an offense for a person to knowingly train artificial intelligence to provide emotional support, including through open ended conversations with a user. And, of course, one of the things that Claude attempts to do is be a thoughtful friend, thoughtful, knowledgeable friend. And in general, it seems like we're gonna have this patchwork of state laws. A lot of the benefits that normal people could experience as a result of AI are going to be curtailed, especially when we get into the kinds of things you discuss in Machines of Love and Grace, biological freedom, mental health improvements, etcetera, etcetera. Seems easy to imagine worlds in which these get whack a mole ed away by different laws. Whereas bills like this don't seem to address the actual existential threats that you're concerned about. So I'm curious about to understand in the context of things like this, your anthropics position against the federal moratorium on state AI laws. Speaker 1: Yes. So I don't know. There's there's many different things going on at at once. Right? I think I think that that I think that particular law is is dumb. Like, you know, I think it was it was clearly made by legislators who just probably had little idea what AI models could do and not do. They're like, AI models serving as that that just sounds scary. Like, I don't want I don't want that to happen. So, you know, we're we're we're not we're not in favor of that. Right? But but but that, you know, that that wasn't the thing that was being voted on. The thing that was being voted on is we're going to ban all state regulation of AI for ten years with no apparent plan to to do any federal regulation of AI, which would take congress to pass, which is a very high bar. So, you know, the idea that we'd ban states from doing anything for ten years, and people said they had a plan for federal government, but, you know, there was no actual there was no proposal on the table. There was no actual attempt. Given the serious dangers that I lay out in adolescence of technology around things like the, you know, kind of biological weapons and bioterrorism, autonomy risk, and the timelines we've been talking about, like ten years is an eternity. I think that's a crazy thing to do. So if that's the choice, if that's what you force us to choose, then then we're gonna we're gonna choose not to have that moratorium. And, you know, I I think the the benefits of that position exceed the costs, but it's it's not a perfect position if that's the choice. Now, I think the thing that we should do, the thing that I would support, is the federal government should step in, not saying states you can't regulate, but here's what we're gonna do, and states you can't differ from this. Right? I think preemption is fine in the sense of saying that federal government says, here is our standard. This applies to everyone. States can't do something different. That would be something I would support if it would be done in the right way. What but but this idea of states, you can't do anything and we're not doing anything either, that that struck that struck us as, you know, very much not making sense. And I think will not age well. It's already starting to not age well with with all the backlash that that you've seen. Now, in terms of in terms of what we would want, I mean, you know, the things we've talked about are are starting with transparency standards, you know, in order to monitor some of these autonomy risks and bioterrorism risks. As the risks become more serious, as we get more evidence for them, then I think we could be more aggressive in some targeted ways and say, Hey, AI bioterrorism is really a threat. Let's pass a law that forces people to have classifiers. I could even imagine it depends. It depends how serious a threat it ends up being. We don't know for sure. Then we need to pursue this in an intellectually honest way where we say ahead of time, the risk has not emerged yet. But I could certainly imagine with the pace that things are going that, you know, I could imagine a world where later this year we say, hey, this this AI bioterrorism stuff is really serious. We should do something about it. We should put it in a federal we should put it in a federal standard, and if the federal government won't act, we should put it in a state standard. I could totally see that. Speaker 0: I'm concerned about a world where if you just consider the pace of progress you're expecting, the life cycle of legislation, the benefits are, as you say, because of diffusion lag, the benefits are slow enough that I really do think this patchwork of on the current trajectory, this patchwork of state laws would prohibit. I mean, having an emotional chatbot friend is something that freaks people out, then just imagine the kinds of actual benefits from AI we want normal people to be able to experience from improvements in health and health span and improvements in mental health and so forth. Whereas at the same time, it seems like you think the dangers are already on the horizon, and I just don't see that much seems like it would be especially injurious to the benefits of AI as compared to the the dangers of AI. And so that that's maybe the where the cost benefit makes less sense to me. Speaker 1: So there's a few things here. Right? I mean, people talk about there being thousands of these state laws. First of all, the vast, vast majority of them do not pass. The world works a certain way in theory, but just because a law has been passed doesn't mean it's really enforced. Right? The people the people, you know, implementing it may be like, oh my god. This is stupid. It would mean shutting off, like, you know, everything that's ever been built in everything that's ever been built in Tennessee. So, you know, very often laws are interpreted in, like, you know, a way that makes them that that that makes them not as dangerous or not as harmful. On on the same side, of course, you have to worry if you're passing a law to stop a bad thing, you had this you had this problem as well. Yeah. Look. My my look. I mean, my basic view is, you know, if if if, you know, we could decide, you know, what laws were passed and how things were done, which, you know, we're only one small input input into that, you know, I would deregulate a lot of the stuff around the health benefits of AI. I think, you know, I I I don't worry as much about the, like, the the the the kind of chatbot laws. I I actually worry more about the drug approval process where I think AI models are going to greatly accelerate the rate at which we discover drugs, and just the the pipeline will get jammed up. Like, the pipeline will not be prepared to, like, process all all of the stuff that's going through it. So, you know, I I think I think reform of the regulatory process to buy us more towards we have a lot of things coming where the safety and the efficacy is actually gonna be really crisp and clear. I mean, a beautiful thing. Really, really crisp and clear and really, really effective. But you know? And and and maybe we don't need all this all this, like, all this superstructure around it that was designed around an era of drugs that barely work and often have serious side effects. But at the same time, I think we should be ramping up quite significantly you know, this this kind of safety and security legislation. And, you know, like I've said, you know, starting with transparency is is my view of trying not to hamper the industry. Right? Trying to find the right balance. I'm worried about it. Some people criticize my essay for saying that's too slow. The dangers of AI will come too soon if we do that. Well, basically, I kind of think like the last six months and maybe the next few months are gonna be about transparency, and then if these if these risks emerge when we're more certain of them, which I think we might be as soon as as later this year, then I think we need to act very fast in the areas that we've actually seen the risk. Like, I think the only way to do this is to be nimble. Now, the legislative process is normally not nimble, but we we need to emphasize to everyone involved the urgency of this. That's why I'm sending this message of urgency. Right? That's why I wrote adolescents of technology. I wanted policymakers to read it. I wanted economists to read it. I want national security professionals to read it. You know, I want decision makers to read it so that they have some hope of acting faster than they would have otherwise. Speaker 0: Is there anything you can do or advocate that would make it more certain that the benefits of AI are are better instantiated? Where I feel like you have worked with legislatures to be like, okay, we're gonna prevent bioterrorism here away. We're gonna increase insurgency. We're gonna increase whistleblower protection. And I just think by default, the actual the things we're looking forward to here, it just seems very easy. They seem very fragile to different kinds of moral panics or political economy problems. Speaker 1: So so I don't actually agree that much in the developed world. I feel like, you know, in the developed world, markets function pretty well. And when there's when there's like a lot of money to be made on something and it's clearly the best available alternative, it's actually hard for the regulatory system to stop it. You know, we're we're seeing that in AI itself. Right? I you know, like a thing I've been trying to fight for is export controls on chips to China. Right? And like, that's in the national security interests of The US. Like, you know, that's like square within the, you know, the the policy beliefs of, you know, every almost everyone in congress of both parties. But and, you know, I think the case is very clear. The counterarguments against it are I'll politely call them fishy. And yet, it doesn't happen, and we sell the chips because there's there's so much money. There's so much money riding on it. And, you know, the the that money wants to be made, and and in that case, in my opinion, that's a bad thing. And but but it also it also applies when when it's a good thing. And and so I I don't think that if we're talking about drugs and benefits of the technology, I I I am not as worried about those benefits being hampered in the developed world. I am a little worried about them going too slow. And I as I said, I do think we should work to speed the approval process in the FDA. I do think we should fight against these chatbot bills that you're describing, right, described individually. I'm against them. I think they're stupid. But I actually think the bigger worry is a developing world where we don't have functioning markets, where, you know, we often can't build on the technology that that we've had. I worry more that those folks will get left behind. And I worry that even if the cures are developed, you know, maybe there's someone in rural Mississippi who doesn't get it as well. Right? That's a kind of smaller version of the thing, the concern we have in the developing world. And so the things we've been doing are we work with philanthropists. Right? We work with folks who deliver medicine and health interventions to developing world, to Sub Saharan Africa, India, Latin America, other developing parts of the world. That's the thing I think that won't happen on its own. Speaker 0: You mentioned export controls. Yeah. Why can't US and China both have a country of geniuses on a data center? Speaker 1: Why can't you know, why won't it happen or why should No. Speaker 0: Like, why why shouldn't it happen? Speaker 1: Why shouldn't it happen? You know, I think I think if this does happen, you know, then then we kind of have a well, we could have a few situate if we have, like, an offense dominant situation, we could have a situation like nuclear weapons, but, like, more dangerous, right, where it's, like, you know, kind of kind of either side could could easily destroy everything. We could also have a world where it's kind of it's unstable. Like, nuclear equilibrium is stable. Right? Because it's, you know, it's like deterrence. But let's say there were uncertainty about, like, if the two AIs fought, which AI would win. That could create instability. Right? You often have conflict when the two sides have a different assessment of their likelihood of winning. Right? If one side is like, oh, yeah. There's a 90% chance I'll win, the other side's like, there's a 90% chance I'll win, then then then a fight is much more likely. They can't both be right, but they can both think that. Speaker 0: But this is like a fully general argument against the diffusion of AI technology, which it may which is that's the implication of this world. Speaker 1: Let just go on because I think we will get diffusion eventually. The other concern I have is that people the governments will oppress their own people with AI. And and and so, you know, I'm I'm just I'm worried about some world where you have a country that's already, you know, kind of a you know, there's there's a government that kind of kind of already, you know, is is kind of kind of building a, you know, a tech a high-tech authoritarian state. And to be clear, this is about the government. This is not about the people. Like, people we need to find a way for people everywhere to benefit. My worry here is about governments. So, yeah, my, you know, my my worry is if the world gets carved up into two pieces, one of those two pieces could be authoritarian or totalitarian in a way that's very difficult to displace. Now, will will governments eventually get powerful and there's risk of authoritarianism? Yes. Will governments eventually get powerful AI and there's risk of bad equilibria? Yes. I think both things, but the initial conditions matter. Right? At some point, we're gonna need to set up the rules of the road. I'm not saying that one country, either The United States or a coalition of democracies, which I think would be a better setup, although it requires more international cooperation than we currently seem to wanna make. But, you know, I don't I don't think a coalition of democracies or or certainly one country should just say these are the rules of the road. There's gonna be some negotiation. Right? The world is gonna have to grapple with this. And what I would like is that the the the, you know, the democratic nations of the world, those with you know, who are whose governments have represent closer to prohuman values are holding a stronger hand then, more leverage when the rules of the road are set. So I'm very concerned about that initial condition. Speaker 0: I was relisting to an interview from three years ago, and one of the ways it aged poorly is that I kept asking questions assuming there was gonna be some key fulcrum moment two to three years from now when in fact, being that far out, it just seems like progress continues, AI improves, AI is more diffused, and people will use it for more things. It seems like you're imagining a world in the future where the countries get together and here's the rules of the world and here's the leverage we have, here's the leverage you have, when it seems like on current trajectory, everybody will have more AI. Some of that AI will be used by authoritarian countries. Some of that within the authoritarian countries will be used by private actors versus state actors. It's not clear who will benefit more. It's always unpredictable to tell in advance. It seems like the Internet privileged authoritarian countries more than you would have expected, Maybe the AI will be the opposite way around. I Speaker 1: want to Speaker 0: better understand what you're imagining here. Speaker 1: Yeah. Yeah. Just to be precise about it, I think the exponential of the underlying technology will continue as it has before. The models get smarter and smarter even when they get to country of geniuses in a data center. You know? I I think you can continue to make the model smarter. There's a question of, like, getting diminishing returns on their value in the world. Right? How much does it matter after you've already solved human biology or, you know, at some point, can do harder math. You can do more abstruse math problems, but nothing after that matters. But putting that aside, I do think the the exponential will continue, but there will be certain distinguished points on the exponential, and companies, individuals, countries will reach those points at different times. So, know, there's you know, could there be some know, I talk about is nuclear deterrent still in adolescence of technology? Is nuclear deterrent still stable in the world of of AI? I don't know, but that's an example of one thing we've taken for granted that the technology could reach such a level that it's no longer we can no longer be certain of it at least. Think of others. There are kind of points where if reach a certain point, maybe you have offensive cyber dominance, and every computer system is transparent to you after that, unless the other side has a kind of equivalent defense. So I don't know what the critical moment is or if there's a single critical moment, but I think there will be either a critical moment, a small number of critical moments, or some critical window where it's like AI is AI confers some large advantage from the perspective of national security and one country or coalition has reached it before others. That that, you know, that that you know, I'm not advocating that they're just like, okay. We're in charge now. That's not that's not how that's not how I think about it. You know, that there's always the the other side is catching up. There's extreme actions you're not willing to take, and and and it's not right to take, you know, to take complete to take complete control anyway. But but at at the point that that happens, I think people are gonna understand that the world has changed, and there there's gonna be some negotiation implicit or implicit about what what is the what is the post AI world order look like? And and I think my interest is in, you know, making that negotiation be one in which, classical liberal democracy has a strong hand? Speaker 0: Well, I wanna understand what that better means because you say in the essay, quote, Autocracy is simply not a form of government that people can accept in the post powerful AI age. And it sounds like you're saying the CCP as an institution cannot exist after we get AGI. And that seems like very strong demand, and it seems to imply a world where the leading lab or the leading country will be able to, and by that language should, get to determine how the world is governed or what kinds of governments are allowed and not allowed. Speaker 1: Yeah. So when I when I I I believe that paragraph was I think I said something like, you could take it even further and say x. So I wasn't I wasn't necessarily endorsing that that I wasn't necessarily endorsing that view. I, know, I was saying like, here's first, you know, here here's a weaker thing that I believe. But, you know, I think I, you know, I think I said, you know, we have to worry a lot about authoritarians and, know, we should try and, you know, kind of kinda check them and limit their power. Like, you could take this kind of further, much more interventionist view that says, like, authoritarian countries with AI are these, you know, the the the you know, these kind of self fulfilling cycles that that you can't, that are very hard to displace, and so you just need to get rid of them from from the beginning. That that has exactly all the problems you say, which is, you know, know, if you were to make a commitment to overthrowing every authoritarian country, I mean, they then they would take a bunch of actions now that like Right. You know, that that that could could lead to instability. So that that may or you know, that that that just that just may not be possible. But the point I was making that I do endorse is that it is it is quite possible that, you know, today, you know, the view or at least my view or the view in most of the Western world is is democracy is a better form of government than authoritarianism. But it's not like if a country's authoritarian, we don't react the way we reacted if they committed a genocide or something. Right? And and I'm well, I guess what I'm saying is I'm a little worried that in the age of AGI, authoritarianism will have a different meaning. It will be a graver thing. We have to decide one way or another how to deal with that. And the interventionist view is one possible view. I was exploring such views. It may end up being the right view. It may end up being too extreme to be the right view. But I do have hope. And and one piece of hope I have is there there is we have seen that as new technologies are invented, forms of government become obsolete. I I mentioned this in adolescence of technology where I said, you know, like feudalism was basically, you know, like a form of government. Right? And and then when when we invented industrialization, feudalism was no longer sustainable. It no longer made sense. Speaker 0: Why is that hope? Couldn't that imply that democracy is no longer gonna be a competitive system? Speaker 1: Right. It could go either way. Right? But I actually so these problems with authoritarianism, right, that the problems with authoritarianism get deeper. I just I wonder if that's an indicator of other problems that authoritarianism will have. Right? In other words, people become because authoritarianism becomes worse, people are more afraid of authoritarianism. They work harder to stop it. It's it's more of a like, you have to think in terms of total equilibrium. Right? I just wonder if it will motivate new ways of thinking about, with with with the new technology, how to preserve and protect freedom. And and even more optimistically, will it lead to a collective reckoning and, you know, a a a a more emphatic realization of how important some of the things we take as individual rights are. Right? A more emphatic realization that we just we really can't give these away. There's there we've seen there's no other way to live that actually works. I I I am actually I am actually hopeful that I I guess one way to say it, it sounds too idealistic, but I actually believe it could be the case, is is that that dictatorships become morally obsolete. They become morally unworkable forms of government, and that and that and that the the the the crisis that that creates is is is sufficient to force us to find another way. Speaker 0: I I think there is genuinely a tough question here, which I'm not sure how you resolve. For and we've had to come out one way or another on it through history. Right? So with China in the seventies and eighties, we decided even though it's an authoritarian system, we will engage with it. And think I in retrospect, that was the right call because it has stayed our authoritarian system, but a billion plus people are much wealthier and better off than they would have otherwise been. And it's not clear that it would have stopped being an authoritarian country otherwise. You can just look at North Korea as an example of that. Right? And I don't know if that takes that much intelligence to remain an authoritarian country that continues to coalesce its own power. And so you can just imagine a North Korea with an AI that's much worse than everybody else's, but still enough to keep power. And and and then so in general, it seems like should we just have this attitude of the benefits of AI will, in the form of all these empowerments of humanity and health and so forth, will be big. And and historically, we have decided it's good to spread the benefits of technology widely even with even to people whose governments are authoritarian. And I think I guess it is a tough question about how to think about it with AI, but historically, we have said yes. This is this is a positive sum world, and it's still worth diffusing the technology. Speaker 1: Yeah. So there are a number of choices we have. I think framing this as a kind of government to government decision in national security terms, that's one lens, but there are a lot of other lenses. You could imagine a world where we produce all these cures to diseases and the cures to diseases are fine to sell to authoritarian countries. The data centers just aren't. The chips and the data centers just aren't, the AI industry itself. Another possibility is and I think folks should think about this, could there be developments we can make either that naturally happen as a result of AI or that we could make happen by building technology on AI, could we create an equilibrium where where it becomes infeasible for authoritarian countries to deny their people kind of private use of the benefit to the technology? You know? Are there are there are there are there equilibria where we can kind of give everyone in an authoritarian country their own AI model that kind of, you know, defends themselves from surveillance? And there isn't a way for the authoritarian country to, like, crack crack down on this while while retaining power. I don't know. That that sounds to me like if that went far enough, it would be it would be a reason why authoritarian countries would disintegrate from the inside. But but maybe there's a middle world where, like, there there's an equilibrium where if they wanna hold on to power, the authoritarians can't deny kind of individualized access access to the technology. But I actually do have a hope for the for the the for the more radical version, which is, you know, is it possible that the technology might inherently have properties or that by building on it in certain ways, we could create properties that that that have this kind of dissolving effect on authoritarian structures? Now we we hoped originally, right, we think about back to the beginning of the Obama administration. We thought originally that social media and the Internet would have that property and turns out not to. But I don't know. What we could try again with the knowledge of how many things could go wrong and that this is a different technology? I don't know that it would work, but it's worth a try. Speaker 0: Yeah. I think it's it's very unpredictable. Like, there's first principles reasons why authoritarianism Speaker 1: It's might not be to very unpredictable. I I don't think I mean, we gotta we we just gotta we kind of we gotta recognize the problem, and then we gotta come up with 10 things we can try, and we gotta try those and then assess whether they're working or which ones are working, if any, and and then try new ones if the old ones aren't working. Speaker 0: What that nets out to today is you say, we will not sell data centers or sorry, chips and then the ability to make chips to China. And so in some sense, you are denying there'll be some benefits to That's right. The Chinese economy, Chinese people, etcetera, because we're doing that. And then there'd also be benefits to the American economy because it's a positive sum world. We could trade. They could have their country data centers doing one thing. We could have ours doing another. Already we you're saying it's not worth that positive sum stipend to empower this country? That Speaker 1: What I would say is that we are we are about to be in a world where growth and economic value will come very easily if right? If we're able to build these powerful AI models, growth and economic value will come very easily. What will not come easily is distribution of benefits, distribution of wealth, political freedom. These are the things that are gonna be hard to achieve. So when I think about policy, I think that the technology in the market will deliver all the fundamental benefits almost almost faster than we can take them, and and that these questions about about distribution and political freedom and rights are are are the ones that that will actually matter and that policy should focus on. Speaker 0: Okay. So speaking of distribution, as you're mentioning, we have developing countries. And in many cases, catch up growth has been weaker than we would have hoped for. But when catch up growth does happen, it's fundamentally because they have underutilized labor, and we can bring the capital and know how from developed countries to these countries, and then they can grow quite rapidly. Obviously, in a world where labor is no longer the constraining factor, this mechanism no longer works. And so is the hope basically to rely on philanthropy from the people who immediately get wealthy from AI or from the countries that get wealthy from AI? What is the I hope for Speaker 1: mean, philanthropy should obviously play some role as it has in the past, but I think better growth is and stronger if we can make it endogenous. Yeah. What are the relevant industries in an AI driven world? Look, there's lots of stuff. I said we shouldn't build data centers in China, but there's no reason we shouldn't build data centers in Africa. Right? In fact, I think it'd be great to build data centers in Africa. Long as they're not owned by China, we should build data centers in Africa. I think that's a great thing to do. We should also build there's no reason we can't build you know, a pharmaceutical industry that's like AI driven. Like, you know, the the if if AI is accelerating accelerating drug discovery, then, you know, there will be a bunch of biotech startups. Like, let's make sure some of those happen in the developing world. Certainly, during the transition, I mean, we can talk about the point where humans have no role, but humans will have still have some role in starting up these companies and supervising supervising the AI models. So let's make sure some of those humans are humans in the developing world so that fast growth can happen there as well. Speaker 0: You guys recently announced Quad is gonna have a constitution that's aligned to a set of values and not necessarily just to the end user. And there's a world you could imagine where if it is aligned to the end user, it preserves the balance of power we have in the world today because everybody gets to have their own AI that's advocating for them. And so the ratio of bad actors to good actors stays constant. It seems to work out for our world today. Why is it better not to do that but to have a specific set of values that the AI should carry forward? Speaker 1: Yeah. So I'm not sure I'd quite draw the distinction in that way. There may be two relevant distinctions here, which are I think you're talking about a mix of the two. One is, should we give the model a set of instructions about do this and versus don't do this? Yeah. And the other, you know, should we give the model a set of principles for, you know, for kind of how to act? And and and there, it's it's, it's purely a practical and empirical thing that we've observed that by teaching the model principles, getting it to learn from principles, its behavior is more consistent, it's easier to cover edge cases, and the model is more likely to do what people want it to do. In other words, if you're like, don't tell people how to hotwire a car, don't speak in Korean, if you give it a list of rules, it doesn't really understand the rules and it's kind of hard to generalize from them, you know, if if it's just kind of a, you know, list of do dos and don'ts. Whereas if you give it principles and then, you know, it has some hard guardrails, like don't make biological weapons. But overall, you're trying to understand what it should be aiming to do, how it should be aiming to operate. So just from a practical perspective, that turns out to be just a more effective way to train the model. That's one piece of it. So that's the kind of rules versus principles trade off. Then there's another thing you're talking about, which is kind of like the corrigibility versus, like, you know, I would say kind of intrinsic motivation trade off, which is like, how much should the model be a kind of I don't know, like a a a skin suit or something where, you know, you know, you know, you just kind of, you know it it just kind of directly follows the instructions that are given to it by whoever is giving it those instructions versus how much should the model have an inherent set of values and go off and do things on its own. There, I would actually say everything about the model is actually closer to the direction of it should mostly do what people want. It should mostly follow the we're not trying to build something that goes off and runs the world on its own. We're actually pretty far on the corrigible side. Now what we do say is there are certain things that the model won't do, right? That it's like I think we say it in various ways in the constitution that under normal circumstances, if someone asks the model to do a task, it should do that task. That should be the default. But if you've asked it to do something dangerous or if you've asked it to kind of harm someone else, then the model is unwilling to do that. So I actually think of it as a mostly corrigible model that has some limits, but those limits are based on principles. Speaker 0: Yeah. I mean, then the fundamental question is, how are those principles determined? And this is not a special question for Anthropic. This would be a question for any company. But because you have been the ones to actually write down the principles, I get to ask you this question. Normally, a constitution is like, you write it down, it's set in stone, and there's a process of updating it and changing it and so forth. In this case, it seems like a document that people at Anthropic write that can be changed at any time that guides the behavior of systems that are gonna be the basis of a lot of economic activity. What is the do you think about how those principles should be set? Speaker 1: Yes. So I think there's there's two there's maybe three three kind of sizes of loop here. Right? Three three ways to iterate. One is you can iterate we iterate within anthropic. We train the model. We're not happy with it, we kind of change the constitution. And I think that's good to do. Putting out publicly, making updates to the constitution every once in a while saying, here's a new constitution. Right. I think that's good to do because people can comment on it. The second level of loop is different companies will have different constitutions. I think it's useful for like Anthropic puts out a constitution and the Gemini model puts out a constitution and other companies put out a constitution and then they can kind of look at them, compare, outside observers can critique and say this this I like this one, this thing from this constitution and this thing from that constitution, and and then kind of that that creates some kind of soft incentive and feedback for all the companies to take the best of each elements and improve. Then I think there's a third loop, which is society beyond the AI companies and beyond just those who comment on the constitutions without hard power. There, we've done some experiments. A couple years ago, did an experiment with, I think it was called the collective intelligence project to basically poll people and ask them what should be in our AI constitution. I think at the time we incorporated some of those changes, and so you could imagine with the new approach we've taken to the constitution doing something like that, it's a little harder because it's like that was actually an easier approach to take when the constitution was like a list of dos and don'ts. At the level of principles, it has to have a certain amount of coherence, but but you could you could still imagine getting views from a wide variety of people. And I think you could also imagine and this is like a crazy idea, but, hey, you know, this whole interview is about trade crazy ideas. Right? So, you know, you could even imagine systems of of kind of representative government having having input. Right? Like, you know, I I wouldn't I wouldn't do this today because the legislative process is so slow. Like, this is exactly why I think we should be careful about the legislative process and AI regulation. But there's no reason you couldn't, in principle, say, like, you know, all AI you know, all AI models have to have a constitution that starts with, like, these things. And then you can append other things after it, but there has to be this special section that takes precedence. I wouldn't do that. That's too rigid. That sounds that sounds kind of overly prescriptive in a way that I think overly aggressive legislation is, but that is a thing you could that thing is you could try to do. Is some much less heavy handed version of that? Maybe. Speaker 0: I really like control loop too, where obviously, this is not how constitutions of actual governments do or should work, where there there's not this vague sense in which the Supreme Court will feel out how people are feeling and what are the vibes and then update the update the constitution accordingly. So there's Yeah. With actual governments, there's a more procedural process. Speaker 1: Or formal process. Speaker 0: Yeah. Exactly. But you actually have a vision of competition between constitutions, is actually very reminiscent of how some libertarian charter cities people you used to talk about what an archipelago of different kinds of governments would look like, and then there would be selection among them of who could operate the most effectively, in which place people would be the happiest. And in in a sense, you're actually yeah. There's this vision. Speaker 1: I'm I'm I'm kind of recreating that. Speaker 0: Yeah. Like, this Utopia of Archipelago. You know? Speaker 1: Again, I think I think that vision has has you know, if things to recommend it and things that things that things that will kind of kind of go wrong with it, you know, I think I think it's a I think it's an interesting, in some ways, compelling vision, but also things will go wrong with it that you hadn't that you hadn't imagined. So, you know, I I I like loop two as well, but I I I feel like the whole thing has gotta be some some mix of loops one, two, and three, and it's a it's a matter of the proportions. Right? I I think that's gotta be the the answer. Speaker 0: When somebody eventually writes the equivalent of the making of the atomic bomb for this era, what is the thing that will be hardest to glean from the historical record that they're most likely to miss? Speaker 1: I think a few things. One is at every moment of this exponential, the extent to which the world outside it didn't understand it. This a bias that's often present in history where anything that actually happened looks inevitable in retrospect. I think when people look back, it will be hard for them to put themselves in the place of people who were actually making a bet on this thing to happen that wasn't inevitable, that we had these arguments, like the arguments that, you know, that I make for scaling or that continual learning will be solved, you know, that that you know, some of us internally in our heads put a high probability on this happening, but it's like there's a world outside us that's not acting on that's not kind of not acting on that at all. I think the weirdness of it think unfortunately, the insularity of it, we're one year or two years away from it happening, the average person on the street has no idea. That's one of the things I'm trying to change with the memos, with talking to policymakers, but I don't know. I think that's just a crazy thing. Yeah. Finally, I would say, and this probably applies to almost all historical moments of crisis, how absolutely fast it was happening, how everything was happening all at once. Decisions that you might think were carefully calculated, well, actually you have to make that decision and then you have to make 30 other decisions on the same day because it's all happening so fast, and you don't even know which decisions are gonna turn out to be consequential. So one of my, I guess, worries, although it's also an insight into kind of what's happening is that some very critical decision will be some decision that someone just comes into my office and is like, Dario, you have two minutes. Should we do thing A or thing B on this someone gives me this random half page memo and is like, should we should we do a or b? And I'm like, I don't know. I have to eat lunch. Let's do b. And that ends up being the most consequential thing ever. Speaker 0: So final question. It seems like you have there's not tech CEOs who are usually writing 50 page memos every few months, and it seems like you have managed to build a role for yourself and a company around you which is compatible with this more intellectual type role as CEO. Ed, I wanna understand how you construct that and how like, how does that work to be that you just go away for a couple weeks and then you tell your company, this is the memo. Like, here's what we're doing. It's also reported you write a bunch of these internally. Speaker 1: Yeah. So for this particular one, I wrote it over winter break. So there was the tie and I was having a hard time finding the time to actually find it, to actually write it. But I actually think about this in a broader way. I actually think it relates to the culture of the company. So I probably spend a third, maybe 40% of my time making sure the culture of Anthropic is good. As Anthropic has gotten larger, it's it's gotten harder to just, you know, get involved in, you know, directly involved in, the training of the models, the launch of the models, the building of the products. Like, it's 2,500 people. It's like, you know, there's just you know, I have certain instincts, but, like, there's only you know, the I I it's very difficult to get in to get to get involved in every single detail. You know? I like I try as much as possible. But one thing that's very leveraged is making sure Anthropic is a good place to work. People like working there. Everyone thinks of themselves as team members. Everyone works together instead of against each other. We've seen as some of the other AI companies have grown without naming any names, we're starting to see decoherence and people fighting each other. I would argue there was even a lot of that from the beginning, but it's gotten worse. I think we've done an extraordinarily good job, even if not perfect, of holding the company together, making everyone feel the mission, that we're sincere about the mission, and that everyone has faith that everyone else there is working for the right reason, that we're a team, that people aren't trying to get ahead of each other's expense or back stab each other, which again, I think happens a lot at some of the other places. How do you make that the case? It's a lot of things. It's me. It's Daniella who runs the company day to day. It's the co founders. It's the other people we hire. It's the environment we try to create. But I think an important thing in the culture is I, some and just, you know, the other leaders as well, but especially me, to articulate what the company is about, why it's doing what it's doing, what its strategy is, what its values are, what its mission is, and what it stands for. When you get to 2,500 people, you can't do that person by person. You have to write or you have to speak to the whole company. This is why I get up in front of the whole company every two weeks and speak for an hour. It's actually I mean, I wouldn't say I write essays internally. I do two things. One, I write this thing called the DVQ, Dario Vision Quest. I wasn't the one who named it that. That's the name it it received, and it's one of these names that I kind of I tried to fight it because it made it sound like I was going off and smoking peyote or something, but the name just stuck. So I get up in front of the company. Every two weeks. I have a three or four page document, and I just talk through three or four different topics about what's going on internally, models we're producing, the products, the outside industry, the world as a whole as it relates to AI and geopolitically in general, just some mix of that. And I just go through very, very honestly. I just go through and I just say, This is what I'm thinking and this is what anthropic leadership is thinking. And then I answer questions. That direct connection, I think, has a lot of value that is hard to achieve when you're passing things down the chain, six levels deep. And Large fraction of the company comes to attend, either in person or virtually. It really means that you can communicate a lot. Then the other thing I do is I just I have a channel in Slack where I just write a bunch of things and comment a lot. Often that's in response to just things I'm seeing at the company or questions people ask or we do internal surveys and there are things people are concerned about and so I'll write them up. And I'm like, I'm very honest about these things. I just say them very directly, and the point is to get a reputation of telling the company the truth about what's happening, to call things what they are, to acknowledge problems, to avoid the sort of corpo speak, the kind of defensive communication that often is necessary in public because, you know, the world is very large and full of people who are, you know, interpreting things in bad faith. But, you know, if you have a company of people who you trust and we try to hire people that we trust, then then, you know, you can you can you can, you know, you can you can really just be entirely unfiltered. I think that's an enormous strength of the company. It makes it a better place to work. It makes people more of the sum of their parts and increases likelihood that we accomplish the mission because everyone is on the same page about the mission. Everyone is debating and discussing how best to accomplish the mission. Speaker 0: Well, in lieu of an external Dario vision quest, we have this interview. This interview is a little like that. This has been fun, Dario. Thanks for doing it. Speaker 1: Yeah. Thank you, Dorkesh. Speaker 0: Hey, everybody. I hope you enjoyed that episode. If you did, the most helpful thing you can do is just share it with other people who you think might enjoy it. It's also helpful if you leave a rating or a comment on whatever platform you're listening on. If you're interested in sponsoring the podcast, you can reach out at bwarkesh.com/advertise. Otherwise, I'll see you at the next one.
Saved - January 27, 2026 at 2:34 PM

@MilkRoadAI - Milk Road AI

Stop scrolling, watch this video immediately and save this S&P's Daniel Yergin just dropped a brutal reality check: we need 50% more copper for electrification, but new mines take 17 years to come online. A global shortage is now "inevitable." https://t.co/Tp7SS4bc6t

Video Transcript AI Summary
Dan and Kelly discuss the outlook for copper and what’s baked into current prices. Kelly notes that short-term factors influencing copper include disruptions in mines in parts of the world, tariffs, and uncertainty about Federal Reserve policy. In the longer term, she says, prices reflect a growing need for copper as the world electrifies. They estimate that by 2040 the world will use 50% more electricity than today, which she equates to “building 650 nuclear power plants every year.” Copper is described as the “metal of electrification.” She explains that much of the demand growth will come from developing countries, and that with the rise of data centers and AI there is a voracious appetite for electricity that has surprised traditional utilities. She cites that data centers used about 4% of US electricity last year, and by 2030 it will be more like 14%, and none of that happens without copper. Dan recalls that copper was first discussed as a major story in 2022, noting that while prices have risen since, they haven’t surged like major tech equities. He acknowledges that commodities are highly cyclical and asks how investors can ensure continued upside given potential soft data points or supply coming online. Kelly responds by emphasizing copper’s link to GDP, describing it as a core economic demand vector. She notes that a key factor is government policy toward mining exploration: it takes an average of 17 years to bring a new copper mine online, so investing in copper is a bet on the future and depends on how governments regulate mining exploration. Overall, the conversation highlights the thesis that long-term copper demand will be driven by electrification and rising electricity use (especially from data centers and AI), while near-term price dynamics will be influenced by mine disruptions, tariffs, and macropolicy. The lag between discovering, permitting, and developing new copper mines (about 17 years) adds to the structural bullish case.
Full Transcript
Speaker 0: Dan, it's great to have you here. Welcome. Speaker 1: Well, thank you. Kelly is very back. Good to be back with you. Speaker 0: You know, I was trying to go back and remember the first time you published on copper warning and and trying to draw attention to the fact that so much was needed and were undersupplied. And not that we can buy it, but I thought to myself, if someone had bought it then, I mean, it has absolutely soared. So not that you're here to give trading advice, but is all of this bullishness now priced in or not? Speaker 1: Well, I think that some of it is has to do with disruptions in mines that's occurring in some parts of the world. Lack of the the impact of tariffs and uncertainty about what happens and also waiting to see what the Federal Reserve does overall. So I think those are the short term factors. But longer term, I think the prices are, if not at this level, at least pointing to the fact that the world's gonna need a lot more copper, as you said, about 50% more than we have today. And the main reason is that the world is electrifying. We estimate in our own that in a by the year 2040, the world will use 50% more electricity than it does today. That's like building 650 nuclear power plants every year. Wow. Copper is the metal of electrification. Speaker 0: That's like building 650 nuclear power plants every year? Speaker 1: Yes. That's that's what the demand growth is looking like because a lot of the demand growth is obviously in the developing world. But now with the advent of, of data centers, AI, you suddenly have this voracious appetite for electricity that is really, in fact, caught traditional utilities by surprise because they've been used to flat demand. We estimate that, you know, data centers used about 4% of US electricity, last year. By 2030, which is pretty close. It'll be more like 14%, and none of that happens without copper. Speaker 0: You I I found it was in 2022 you guys first were talking about this. So over the past three years, you know, the price has risen. It it it hasn't risen like NVIDIA's share price, to be sure. Speaker 1: And it commodity nothing anything has. Speaker 0: Correct. In commodities, investors know that these things are are hugely cyclical, so they might say, okay, well, I could get in on this trade now, but how do I make sure that there continues to be so much upside? Right? Because every you turn around and then maybe there's a soft economic data point so you can react to that or you turn around and you know supply is being brought online. Mean, that's traditionally what was the famous bet in the eighties where they said, I think commodities are gonna outperform the S and P and you always lose. Right? Speaker 1: Yeah. That's right. Well, I think, you know, I think that's true. I mean, cycles are there, and we have to remember that also copper is called doctor copper because one of the vectors of demand is a traditional core economic demand that we see in everything, and that's very responsive to GDP. So that will have a impact. What also matters is how governments treat companies who wanna come and explore and develop copper mines. It takes an average, Kelly, seventeen years to bring on a new copper mine. So you are making a bet on the future when you start a new mine.

@MilkRoadAI - Milk Road AI

Forget Nvidia. Forget Crypto (Save this video). The smartest trade for 2026 might be rocks? Chamath makes a terrifyingly simple case for a Copper supercycle: A 70% supply shortage is coming and there is no Plan B. https://t.co/PUk6ybFMfb

Video Transcript AI Summary
Speaker 0 argues that we are still completely underestimating how short we will be in terms of the global demand-supply dynamics of a handful of critical elements. In the view of the Trump doctrine, the world is no longer as multilateral, and there is a need for unilateral national security. From this lens, the asset set to go absolutely parabolic is copper. Copper is described as the most useful, cheap, amenable, conducted material that we have, and it manifests in everything from data centers to chips to weapon systems. Currently, Jason, we are on a path by 2040 where we will be short about 70% of the global supply at current course and speed. Copper.
Full Transcript
Speaker 0: I will pick copper. Okay. Copper. We are still completely underestimating how short we are in terms of the global demand supply dynamics of a handful of critical elements that we need, again, in the Trump doctrine view of the world, that is no longer as multilateral as it was, and we need to have unilateral national security. And if you look through that lens, the asset that is set up to go absolutely parabolic is copper. And the reason is that it is, at least as it stands today, the most useful, cheap, amenable, conducted material that we have. That material manifests in everything from our data centers to our chips to our weapon systems. It's just everywhere, everywhere, everywhere. And right now, Jason, we are on a path by 2040 where we will be short about 70% of the global supply at current course and speed. Copper. I will pick copper.
Saved - January 27, 2026 at 2:07 PM

@MilkRoadAI - Milk Road AI

Robert Friedland is warning that humanity faces a copper crisis. This crisis will fundamentally constrain economic growth. It will limit AI deployment and the energy transition. We consume 30 million tonnes of copper annually. But we only recycle 4 million tonnes of it.​ This means we're dependent on constant new mining. We need new mining to keep the economy running. To maintain just 3% GDP growth over the next 18 years, we need to build out AI data centers. We need to electrify vehicles and expand renewable energy. To do all this, we need to mine as much copper as humanity has mined in the previous 10,000 years combined.​ The demand explosion is happening from four directions at once. First, conventional economic growth drives baseline copper consumption. This consumption happens in construction, appliances, and traditional infrastructure.​ Second, the energy transition requires more copper per vehicle for electric cars. It requires massive amounts for renewable energy infrastructure.​ Wind turbines and solar panel installations are copper intensive. Third, AI data centers are a new copper monster. A single hyperscale facility can consume 50,000 tonnes of copper.​ By 2030, data centers could need half a million tonnes annually.​ This is as AI infrastructure explodes across the globe. Fourth, defense spending is driving demand for electrified military equipment. Advanced telecommunications systems also need copper.​ Add it all up and global copper demand is projected to jump 50% by 2040. It's climbing from 28 million tonnes today to 42 million tonnes annually.​ Here's the problem: supply can't keep up. Global copper production will peak in 2030. It will peak at around 33 million tonnes.​ Then it will stagnate or decline. Remaining ore deposits are lower quality. Extraction costs are rising significantly. Mining takes 17 years from discovery to production.​ By 2040, the world will face a shortage. It will be roughly 10 million tonnes annually.​ That's about 25% below what's actually needed. You can't solve a 25% shortage just by raising prices. You physically don't have enough copper. Recycling sounds like it should solve this. But it won't. Yes, recycled copper saves 85% of the energy required for mining.​ Improving collection could recover an additional 6 million tonnes annually.​ This could happen by mid century. The problem is you can only recycle copper that's already in circulation. Old wiring, end of life cars, and decommissioned electronics contain recycled copper.​ To actually close the supply gap, the world would need to double recycling rates to 66%.​ We'd also need to source 10 million additional tonnes from new mining by 2040. Realistic projections show primary supply will actually decline slightly.​ Why should you care? This copper constraint becomes a chokepoint for everything you care about. If there's not enough copper, AI infrastructure deployment slows down.​ This means slower AI progress. It means higher cloud computing costs. You'll pay more for software and services you use. Renewable energy deployment decelerates significantly.​ You can't source enough copper for grid upgrades. New solar and wind installations become harder to build. Electric vehicle adoption stalls.​ Production becomes constrained and prices spike. The US government has already classified copper as a critical mineral.​ Policymakers recognize it's essential to national competitiveness. It's essential to national security. China controls significant portions of global copper mining.​ China also controls global copper processing. This adds geopolitical risk. Copper prices are already signaling the crisis. The metal hit $11,771 per tonne in December 2025.​ But that's still not high enough to justify building new mines. Friedland estimates you need $15,000 per tonne.​ That's the price needed to make new mine projects economically viable. JP Morgan projects $12,075 per tonne averaging in 2026.​ Goldman Sachs forecasts $10,710 in early 2026.​ Long term projections are around $15,000 by 2035. What needs to happen? Friedland advocates for deploying advanced exploration technology. This technology finds new deposits faster.​ Building high efficiency mining operations is essential.​ Massively scaling up recycling infrastructure is critical.​ But more fundamentally, governments need to stockpile strategic copper reserves.​ Accelerating permitting for new mines is necessary.​ Managing demand by prioritizing critical applications is required.

Video Transcript AI Summary
Copper and aluminum are the primary beneficiaries of the grid spending increase. That $800,000,000,000 is going to buy copper, which is money. The oil market, compared to the metals market, is dwarfed by the demand for metals like copper, aluminum, iron ore, gold, and nickel, which are said to be so thinly traded and critical that there is no chance to get off crude oil. You can’t build electric cars, windmills, solar, or a modern military without these metals. Underwater power cables are expensive, and offshore wind and bringing that electricity green requires copper—copper, copper, copper. Copper now is described as a trillion-dollar annual market by tomorrow morning. There is no copper inventory to meet this demand. Since Mohenjo Daro, humanity has mined 700,000,000 metric tons of copper. If we put that in a big cube for scale (about 4 thirty-meter sides), approximately 80% of all the copper ever mined is still in human possession. Recycling could recover about 80% of that 700,000,000 tons, but it would require tearing down every building in the United States, Europe, Japan, and China. We can recycle copper from buildings and even from the university in front of us, but the consequence would be living in the dark. Currently, we consume 30,000,000 tons of copper per year, with only 4,000,000 tons recycled. To maintain 3% GDP growth with no electrification, this speaker claims we must mine the same amount of copper in the next eighteen years as we mined in the last ten thousand years. In the next eighteen years, we would need to mine the same copper volume as mined in the entire previous span of human history, without electrification, without data centers, without solar and wind, and without the greening of the world economy. Since 1900, the energy required to produce copper has increased sixteen-fold, and as ore grades decline, more energy is needed to produce the same metal while water consumption has doubled. Grades are declining globally, and easy copper mines are depleted; Chile is highlighted as a major producer (24% of global copper mine production), yet costs are in the third or fourth quartile. They burn coal in the Chilean grid, and solar is ineffective for mining because the sun only shines a few hours a day; solar is useless without grid-scale storage. The speaker asserts we are heading for a train wreck in Chile and that we need six giant tier-one mines online every year from now until 2050 to meet copper demand for electrification, data centers, and grid upgrades—40% of the production to come from new mines. All the hype about AI is dismissed as fantasy because we do not have the energy. Nuclear power is proposed as a solution, but what are those plants made of? All the metals mentioned earlier. The country reportedly does not have the capability to weld containment vessels in a traditional nuclear power plant anymore, whereas Korea can build a nuclear power plant.
Full Transcript
Speaker 0: Copper and aluminum are the primary beneficiaries of the grid spending increase. That $800,000,000,000 is going to buy copper, which is money. So how big is the oil market compared to the metals market? If you look at that circle, that's crude oil. You want to get rid of crude oil? The iron ore, the gold, the copper, the aluminum, the nickel, all the metals are so thinly traded and so critical. You have no chance to get off of crude oil. You can't build electric cars and windmills and solar and have a modern military without these metals. So there's a reason why underwater power cables are so expensive. That's what it looks like when you put up a windmill offshore Nantucket Island and you wanna bring that electricity and be green, it's all copper, copper, copper, copper, copper. Copper right now, we're expecting that to be $270,000,000,000 a year market by tomorrow morning. And where's this metal going to come from? There's no copper inventory at all. Let me show you how bad this problem is. In human history, if we go back to Mohenjo Daro, we have mined 700,000,000 metric tons of copper. So we put that in a big cube. You see the Eiffel Tower for scale. It's about four thirty meters by four thirty meters. Approximately 80% of all the copper we've ever mined is still in human possession. Now if you want to get that back, no problem. We can recycle that. All we have to do is tear down every building in The United States, building in Europe, every building in Japan and all of China, and we can get back about 80% of that 700,000,000 tons. But we'll be living and freezing in the dark. These lights are on here because the copper that's embedded in this building. We can tear it off this university and get the copper back. Yeah, we can recycle it. How much are we using? We're consuming 30,000,000 tons of copper a year, only 4,000,000 tons of which is recycled. That means to maintain 3% GDP growth. Now listen carefully, with no electrification, this is with burning oil and gas. To maintain global 3% GDP growth we have to mine the same amount of copper in the next eighteen years as we mined in the last ten thousand years. In the next eighteen years I've got to mine the same amount of copper as we mined in the last ten thousand years without electrification, without data centers, without solar and wind and the greening of the world economy. You people have no idea whatsoever what we're facing. You're dreaming. Since 1900 the energy to produce copper, the energy you need to make copper is 16 fold up. As the ore grade in the world goes down we need more and more energy to produce the very metal that produces the energy and water consumption has doubled. So grades are declining in the world's minds. The easy copper mines were mined. They're like 100 year old ladies laying in bed waiting to die in Chile. So Chile accounts for a lot. That's 24% of global copper mine production. But you know their costs are in the third or fourth quartile. They're very expensive. They burn coal in the Chilean grid and solar doesn't work for a mine because the sun only shines five hours a day. Solar is useless unless you have grid scale storage. So we're heading for a train wreck in Chile. We need six giant tier one minutees to come online every year between now and 2050 when I'm scheduled to be a 100 years old. To meet copper demand that means 40% of the production for new mines for electrification, data centers and grid upgrades. And all this bullshit about AI, everything you're hearing, like it's a fantasy because we don't have the energy. Now maybe we can build nuclear power plants. What are they made out of? All those metals I told you about. We don't even have the capability in this country to to weld the containment vessels in a traditional nuclear power plant. We used to build them in this country. Korea can build a nuclear power

@TheGladiatorHC - TheGladiator

You just can't own enough Copper. Billionaire Robert Friedland sums it up perfectly..... You people have no idea whatsoever what we’re facing. “We’re consuming 30m tonnes of copper a year. Only 4m tonnes of which is recycled. That means to maintain 3% GDP growth, with no electrification, we have to mine the same amount of copper in the next 18 years as we mined in the last 10,000 years, combined. In the next 18 years, I’ve got to mine the same amount of copper as we mined the last 10,000 years. This is without any new electrification, without data centers, without solar and wind and the greening of the world economy. You people have no idea whatsoever what we’re facing.”

Video Transcript AI Summary
Copper and aluminum are the primary beneficiaries of the grid spending increase. $800,000,000,000 is going to buy copper, which is money. How big is the oil market compared to the metals market? Crude oil dominates. All metals—iron ore, gold, copper, aluminum, nickel—are thinly traded and critical. There is no chance to get off crude oil; you can’t build electric cars, windmills, solar, or a modern military without these metals. Underwater power cables are expensive, and offshore wind with transmission to Greening efforts illustrates copper’s central role. Copper is the focus: copper is the expected $270,000,000,000 per year market by tomorrow morning. Where will this metal come from? There is no copper inventory. Historically, since Mohenjo Daro, humanity mined 700,000,000 metric tons of copper; about 80% of all copper ever mined is still in human possession. Recycling can recover about 80% of that 700,000,000 tons, but to do so would require tearing down every building in the United States, Europe, Japan, and China. Copper is embedded in buildings and other infrastructure; it can be recycled, but extracting it at scale remains challenging. Currently, we consume 30,000,000 tons of copper a year, with only 4,000,000 tons recycled. To maintain global 3% GDP growth, without electrification and relying on burning oil and gas, we must mine the same amount of copper in the next eighteen years as we mined in the last ten thousand years. In the next eighteen years, we would have to mine the same cumulative amount as in ten thousand years prior, without electrification, without data centers, without solar and wind, and without the greening of the world economy. There is little appreciation for the challenge faced. Since 1900, the energy required to produce copper has increased 16-fold. As ore grades decline, more energy is needed to produce the same metal, while water consumption has doubled. The easy copper deposits are largely depleted; Chile accounts for 24% of global copper mine production, but costs are in the third or fourth quartile. Chile burns coal, and solar isn’t reliable for mining operations since the sun shines only ~five hours a day; solar is useless without grid-scale storage. We are heading for a train wreck in Chile. To meet copper demand, six giant Tier One mines must come online every year from now until 2050. To meet copper demand, 40% of production must come from new mines for electrification, data centers, and grid upgrades. All the talk about AI is fantasy without sufficient energy. Nuclear power could help, but its components require metals, and the U.S. lacks the capability to weld containment vessels in traditional nuclear plants; Korea can build a nuclear power plant.
Full Transcript
Speaker 0: Copper and aluminum are the primary beneficiaries of the grid spending increase. $800,000,000,000 is going to buy copper, which is money. So how big is the oil market compared to the metals market? If you look at that circle, that's crude oil. You want to get rid of crude oil? The iron ore, the gold, the copper, the aluminum, the nickel, all the metals are so thinly traded and so critical. You have no chance to get off of crude oil. You can't build electric cars and windmills and solar and have a modern military without these metals. So there's a reason why underwater power cables are so expensive. That's what it looks like when you put up a windmill offshore Nantucket Island, and you wanna bring that electricity and be green. It's all copper copper copper copper copper. Copper right now we're expecting that to be $270,000,000,000 a year market by tomorrow morning. And where's this metal going to come from? There's no copper inventory at all. Let me show you how bad this problem is. In human history, if we go back to Mohenjo Daro, we have mined 700,000,000 metric tons of copper. So we put that in a big cube, you see the Eiffel Tower for scale. It's about four thirty meters by four thirty meters. Approximately 80% of all the copper we've ever mined is still in human possession. Now if you want to get that back, no problem. We can recycle that. All we have to do is tear down every building in The United States, every building in Europe, every building in Japan and all of China, and we can get back about 80% of that 700,000,000 tons. But we'll be living and freezing in the dark. These lights are on here because the copper that's embedded in this building. We can tear it off this university and get the copper back. Yeah, we can recycle it. How much are we using? We're consuming 30,000,000 tons of copper a year, only 4,000,000 tons of which is recycled. That means to maintain 3% GDP growth. Now listen carefully, with no electrification, this is with burning oil and gas. To maintain global 3% GDP growth we have to mine the same amount of copper in the next eighteen years as we mined in the last ten thousand years. In the next eighteen years I've got to mine the same amount of copper as we mined in the last ten thousand years without electrification, without data centers, without solar and wind and the greening of the world economy. You people have no idea whatsoever what we're facing. You're dreaming. Since 1900, the energy to produce copper, the energy you need to make copper is 16 fold up. As the ore grade in the world goes down, we need more and more energy to produce the very metal that produces the energy and water consumption has doubled. So grades are declining in the world's minds. The easy copper mines were mined. They're like 100 year old ladies laying in bed waiting to die in Chile. So Chile accounts for a lot. That's 24% of global copper mine production. But you know their costs are in the third or fourth quartile. They're very expensive. They burn coal in the Chilean grid and solar doesn't work for a mine because the sun only shines five hours a day. Solar is useless unless you have grid scale storage. So we're heading for a train wreck in Chile. We need six giant Tier one minutees to come online every year between now and 2050 when I'm scheduled to be 100 years old. To meet copper demand, that means 40% of the production for new mines for electrification, data centers and grid upgrades. And all this bullshit about AI, everything you're hearing, like it's a fantasy because we don't have the energy. Now maybe we can build nuclear power plants. What are they made out of? All those metals I told you about. We don't even have the capability in this country to to weld the containment vessels in a traditional nuclear power plant. We used to build them in this country. Korea can build a nuclear power
Saved - January 18, 2026 at 10:18 PM

@MilkRoadAI - Milk Road AI

Andrej Karpathy casually shows how to build real apps just by prompting in under 30 minutes. Save this, I promise you'll come back to it. https://t.co/vTkXkX1SFP

Video Transcript AI Summary
Former Tesla AI director Andre Karpathy discusses software in the era of AI, emphasizing how software is changing at a fundamental level and what this means for students entering the industry. Key framework: three generations of software - Software 1.0: the code that programs computers. - Software 2.0: neural networks, where you tune data sets and run optimizers to create model parameters; the weights program the neural nets rather than hand-written code. - Software 3.0: prompts as programs that program large language models (LLMs); prompts are written in English, effectively a new programming language. - He notes that a growing amount of GitHub-like activity in software 2.0 blends English with code, and that the ecosystem around LLMs resembles a newer GitHub-like space (e.g., Hugging Face, Model Atlas). An example: tuning a LoRa on Flux’s image generator creates a “git commit” in this space. Evolving software stacks in practice - At Tesla Autopilot, the stack evolved from heavy C++ (software 1.0) to neural nets handling image processing and sensor fusion, with many 1.0 components being migrated to 2.0. The neural network grew in capability and size, and the 1.0 code was deleted as functionality migrated to 2.0. - We now have three distinct programming paradigms: 1.0 coding, 2.0 weights, and 3.0 prompts. Fluent capability in all three is valuable because tasks may be best solved with code, trained networks, or prompts. LLMs as a new computer and ecosystem view - Andrew Ng’s “AI is the new electricity” is cited to frame LLMs as utility-like (CapEx for training, OpEx for API serving, metered usage, low latency, high uptime) and also as fabs-like (large CapEx, rapid tech-tree growth), though software nature means more malleability. - LLMs are compared to operating systems: CPU-like core, memory in context windows, and orchestration of compute/memory for problem solving. App downloads can be run across various LLM platforms similarly to cross-OS apps. - The diffusion pattern of LLMs is inverted compared to many technologies: governments and corporations often lag behind consumer adoption, with AI topics sometimes used for everyday tasks like “boiling an egg” rather than high-level strategic aims. Practical implications for developers and students - Build fluently across paradigms: code in 1.0, tune 2.0 models, and design 3.0 prompts; decide when to code, train, or prompt depending on task. - Partially autonomous apps: exemplified by Cursor and Perplexity. - Cursor: traditional interface plus LLM integration, with under-the-hood embeddings, diffs, and multi-LLM orchestration; GUI support for auditing changes; autonomy slider lets users control how much the AI acts vs. what humans verify. - Perplexity: similar features, with sources cited and ability to scale autonomy from quick search to deep research. - Autonomy slider concept: users can limit or increase AI autonomy depending on task complexity; the AI handles context management and multi-call orchestration, while humans verify for correctness and security. - Education and “keeping AI on the leash”: emphasize concrete prompts, better verification, and development of structured education pipelines with auditable AI-generated content. Opportunities and caveats in AI-assisted workflows - Education and governance: separate roles for AI-generated courses and AI-assisted delivery to students, ensuring syllabus adherence and auditability. - Documentation and access for LLMs: docs should be machine-readable (e.g., markdown), and wording should be actionable (avoid “click” commands; provide equivalent API calls like curl) to facilitate LLM interactions. - Tools to ingest data for LLMs: services that convert GitHub repos into ingestible formats (e.g., git ingest, DeepWiki) to create ready-to-query knowledge bases. - Agents vs. augmentation: early emphasis on augmentation (Iron Man-like suits) rather than fully autonomous systems; the autonomy slider enables gradual handover from human supervision to more autonomous tasks while maintaining safety and auditability. - The future of “native” programming: vibe coding and byte coding illustrate how language-based programming lowers barriers, enabling broad participation in software creation; the takeaway is that natural-language interfaces can act as a gateway to software development, even for non-experts. Closing synthesis - We’re at an era where enormous code rewriting is needed, and LLMs function as utilities, fabs, and operating systems, though still early—like the 1960s of OS development. - The next decade will likely feature a spectrum of partially autonomous products with specialized GUIs and rapid verification loops, guided by an autonomy slider and careful human oversight. - Karpathy envisions an ongoing collaboration with AI: building partial autonomy products, evolving tooling, and experimenting with how the industry and education adapt to this new programming reality. He invites readers to participate in shaping this future.
Full Transcript
Speaker 0: Please welcome former director of AI, Tesla, Andre Karpathy. Hello. Wow. A lot of people here. Hello. Okay. Yeah. So I'm excited to be here today to talk to you about software in the era of AI. And I'm told that many of you are students, like bachelor's, master's, PhD, and so on, and you're about to enter the industry. And I think it's actually like an extremely unique and very interesting time to enter the industry right now. And I think fundamentally the reason for that is that software is changing again. And I say again because I actually gave this talk already, but the problem is that software keeps changing. So I actually have a lot of material to create new talks. And I think it's changing quite fundamentally. I think roughly speaking, software has not changed much on such a fundamental level for seventy years. And then it's changed, I think, about twice quite rapidly in the last few years. And so there's just a huge amount of work to do, a huge amount of software to write and rewrite. So let's take a look at maybe the realm of software. So if we kind of think of this as like the map of software, this is a really cool tool called map of GitHub. This is kind of like all the software that's written. These are instructions to the computer for carrying out tasks in the digital space. So if you zoom in here, these are all different kinds of repositories, and this is all the code that has been written. And a few years ago, I kind of observed that software was kind of changing, and there was kind of like a new new type of software around, and I called this software two point o at the time. And the idea here was that software one point o is the code you write for the computer. Software two point o are basically neural networks, and in particular, the weights of a neural network. And you're not writing this code directly. You are most, are more kind of like tuning the data sets and then you're running an optimizer to create the parameters of this neural net. And I think like at the time, neural nets were kind of seen as like just a different kind of classifier, a decision tree or something like that. And so I think it was kind of like I I think this framing was a lot more appropriate. And now actually what we have is kind of like an equivalent of GitHub in the realm of software two point o. And I think the Hugging Face is basically equivalent of GitHub in software two point o. And there's also model Atlas and you can visualize all the code written there. In case you're curious, by the way, the giant circle, the point in the middle, these are the parameters of flux, the image generator. And so anytime someone tunes a LoRa on top of a flux model, you basically create a git commit in this space and you create a different kind of image generator. So basically what we have is software one point o is the computer code that programs a computer. Software two point o are the weights which program neural networks. And here's an example of AlexNet image recognizer neural network. Now so far, all of the neural networks that we've been familiar with until recently were kind of like fixed function computers, image to categories or something like that. And I think what's changed and I think as a quite fundamental change is that neural networks became programmable with large language models. And so I see this as quite new, unique, it's a new kind of a computer. And so in my mind, it's worth giving it a new designation of software three point zero. And basically, your prompts are now programs that program the LLM. And remarkably, these prompts are written in English. So it's kind of a very interesting programming language. So maybe to summarize the difference, if you're doing sentiment classification, for example, you can imagine writing some amount of Python to basically do sentiment classification, or you can train a neural net, or you can prompt a large language model. So here, this is a few shot prompt and you can imagine changing it and programming the computer in a slightly different way. So basically we have software one point o, software two point o, and I think we're seeing, maybe you've seen a lot of GitHub code is not just like code anymore, there's a bunch of like English interspersed with code. And so I think kind of there's a growing category of new kind of code. So not only is it a new programming paradigm, it's also remarkable to me that it's in our native language of English. And so when this blew my mind a few, I guess years ago now, I tweeted this and I think it captured the attention of a lot of people and this is my currently pinned tweet, is that remarkably we're now programming computers in English. Now, when I was at, Tesla, we were working on the autopilot and we were trying to get the car to drive. And I sort of showed this slide at the time where you can imagine that the inputs to the car are on the bottom and they're going through a software stack to produce the steering and acceleration. And I made the observation at the time that there was a ton of c plus plus code around in the autopilot, which was the software one point o code, and then there was some neural nets in there doing image recognition. And I kind of observed that over time as we made the autopilot better, basically the neural network grew in capability and size. And in addition to that, all the C plus plus code was being deleted and kind of like was and a lot of the kind of capabilities and functionality that was originally written in one point zero was migrated to two point zero. So as an example, a lot of the stitching up of information across images from the different cameras and across time was done by neural network and we were able to delete a lot of code. And so the software two point zero stack quite literally ate through the software stack of the autopilot. So I thought this was really remarkable at the time. And I think we're seeing the same thing again where basically we have a new kind of software and it's eating through the stack. We have three completely different programming paradigms. And I think if you're entering the industry, it's a very good idea to be fluent in all of them because they all have slight pros and cons and you may want to program some functionality in one point o or two point o or three point o. Are you going to train a neural net? Are you going to just prompt an LLM? Should this be a piece of code that's explicit, etcetera? So we all have to make these decisions and actually potentially fluidly transition between these paradigms. So what I want to get into now is first I want to, in the first part, talk about LLMs and how to kind of like think of this new paradigm and the ecosystem and what that looks like. Like what are, what is this new computer? What does it look like? And what does the ecosystem look like? I was struck by this quote from Andrew Ng actually many years ago now, think. And I think Andrew is going to be speaking right after me. But he said at the time AI is the new electricity. And I do think that it kind of captures something very interesting in that LLMs certainly feel like they have properties of utilities right now. So LLM Labs like OpenAI, Gemini, Anthropia, etcetera, they spend CapEx to train the LLMs and this is kind of equivalent to building out a grid. And then there's OpEx to serve that intelligence over APIs to all of us. And this is done through metered access where we pay per million tokens or something like that. And we have a lot of demands that are very utility like demands out of this API. We demand low latency, high uptime, consistent quality, etcetera. In electricity, you would have a transfer switch. So you can transfer your electricity source from like grid and solar or battery or generator. In LLMs, we have maybe open router and easily switch between the different types of LLMs that exist. Because the LLMs are software, they don't compete for physical space. So it's okay to have basically like six electricity providers and you can switch between them, right, because they don't compete in such a direct way. And I think what's also a little fascinating and we saw this in the last few days actually, a lot of the LLMs went down and people were kind of like stuck and unable to work. And I think it's kind of fascinating to me that when the state of the art LLMs go down, it's actually kind of like an intelligence brownout in the world. It's kind of like when the voltage is unreliable in the grid and the planet just gets dumber the more reliance we have on these models, which already is like really dramatic and I think will continue to grow. But LLMs don't only have properties of utilities. I think it's also fair to say that they have some properties of fabs. And the reason for this is that the CapEx required for building LLMs is actually quite large. It's not just like building some power station or something like that, right? You're investing a huge amount of money and I think the tech tree and for the technology is growing quite rapidly. So we're in a world where we have sort of deep tech trees, research and development, secrets that are centralizing inside the LLM labs. And but I think the analogy muddies a little bit also because as I mentioned, this is software. And software is a bit less defensible because it is so malleable. And so I think it's just an interesting kind of thing to think about potentially. There's many analogies you can make, like a four nanometer process node maybe is something like a cluster with certain max plots. You can think about when you're use when you're using NVIDIA GPUs and you're only doing the software and you're not doing the hardware, that's kind of like the fabless model. But if you're actually also building your own hardware and you're training on TPUs if you're Google, that's kind of like the Intel model where you own your fab. So I think there's some analogies here that make sense. But actually, I think the analogy that makes the most sense perhaps is that in my mind, LLMs have very strong kind of analogies to operating systems. In that, this is not just electricity or water. It's not something that comes out of the tap as a commodity. These are now increasingly complex software ecosystems. Right? So they're not just like simple commodities like electricity. And it's kind of interesting to me that the ecosystem is shaping in a very similar kind of way where you have a few closed source providers like Windows or Mac OS, and then you have an open source alternative like Linux. And I think for LLMs as well, we have a kind of a few competing closed source providers. And then maybe the LAMA ecosystem is currently like maybe a close approximation to something that may grow into something like Linux. Again, think it's still very early because these are just simple LMs, but we're starting to see that these are going to get a lot more complicated. It's not just about the LLM itself, it's about all the tool use and the multi modalities and how all of that works. So when I sort of had this realization a while back, I tried to sketch it out and it kind of seemed to me like LLMs are kind of like a new operating system, right? So the LLM is a new kind of a computer. It's setting, it's kind of like the CPU equivalent. The context windows are kind of like the memory. And then the LLM is orchestrating memory and compute for problem solving using all of these capabilities here. And so definitely, if you look at it, it looks very much like operating system from that perspective. A few more analogies, for example, if you want to download an app, say I go to Versus Code and I go to download, you can download Versus Code and you can run it on Windows, Linux or Mac. In the same way as you can take an LLM app like Cursor and you can run it on GPT or Cloud or Gemini series. Right? It's just a drop down. So it's kind of like similar in that way as well. More analogies that I think strike me is that we're kind of like in this 1960s ish era where LLM compute is still very expensive for this new kind of a computer. And that forces the LLMs to be centralized in the cloud and we're all just sort of thin clients that interact with it over the network. And none of us have full utilization of these computers and therefore it makes sense to use time sharing where we're all just a dimension of the batch when they're running the computer in the cloud. And this is very much what computers used to look like during this time. The operating systems were in the cloud, everything was streamed around and there was batching. And so the personal computing revolution hasn't happened yet because it's just not economical, it doesn't make sense, but I think some people are trying. And it turns out that Mac minis for example are a very good fit for some of the LLMs because it's all if you're doing batch one inference, this is all super memory bound so this actually works. And I think these are some early indications maybe of personal computing, but this hasn't really happened yet. It's not clear what this looks like. Maybe some of you get to invent what this is or how it works or what this should be. Maybe one more analogy that I'll mention is whenever I talk to ChatGPT or some LLM directly in text, I feel like I'm talking to an operating system through the terminal. Like it's just, it's text, it's direct access to the operating system and I think a GUI hasn't yet really been invented in like a general way. Like should ChatGPT have a GUI, like different than just the tech bubbles? Certainly some of the apps that we're going to go into in a bit have GUI, but there's no like GUI across all the tasks if that makes sense. There are some ways in which LLMs are different from kind of operating systems in some fairly unique way and from early computing. And I wrote about this one particular property that strikes me as very different this time around, it's that LLMs like flip, they flip the direction of technology diffusion that is usually present in technology. So for example, with electricity, cryptography, computing, flight, Internet, GPS, lots of new transformative technologies that have not been around. Typically, it is the government and corporations that are the first users because it's new and expensive, etcetera, and it only later diffuses to consumer. But I feel like LMs are kind of like flipped around. So maybe with early computers, it was all about ballistics and military use, but with LLMs, it's all about how do you boil an egg or something like that. This is certainly like a lot of my use. And so it's really fascinating to me that we have a new magical computer and it's like helping me boil an egg. It's not helping the government do something really crazy like some military ballistics or some special technology. Indeed corporations or governments are lagging behind the adoption of all of us, of all of these technologies. So it's just backwards. And I think it informs maybe some of the uses of how we want to use this technology or like what are some of the first apps and so on. So in summary so far, LLM Labs, bad LLMs, I think it's accurate language to use. But llms are complicated operating systems. They're circa nineteen sixties in computing and we're redoing computing all over again. And they're currently available via time sharing and distributed like a utility. What is new and unprecedented is that they're not in the hands of a few governments and corporations, they're in the hands of all of us because we all have a computer and it's all just software and Chachi PT was beamed down to our computers like billions of people like instantly and overnight. And this is insane. And it's kind of insane to me that this is the case and now it is our time to enter the industry and program these computers. This is crazy. So I think this is quite remarkable. Before we program LLMs, we have to kind of like spend some time to think about what these things are. And I especially like to kind of talk about their psychology. So the way I like to think about LLMs is that they're kind of like people spirits. They are stochastic simulations of people. And the simulator in this case happens to be an autoregressive transformer. So a transformer is a neural net. It's and it just kind of like goes on the level of tokens. It goes chunk, chunk, chunk, chunk, chunk. And there's an almost equal amount of compute for every single chunk. And this simulator of course is just, is basically there's some weights involved and we fit it to all of text that we have on the internet and so on. And you end up with this kind of a simulator. And because it is trained on humans, it's got this emergent psychology that is human like. So the first thing you'll notice is of course, LLMs have encyclopedic knowledge and memory and they can remember lots of things, a lot more than any single individual human can because they've read so many things. It actually kind of reminds me of this movie Rain Man which I actually really recommend people watch. It's an amazing movie. I love this movie. And Dustin Hoffman here is an autistic savant who has almost perfect memory. So he can read like a phone book and remember all of the names and phone numbers. And I kind of feel like LNs are kind of like very similar. They can remember SHA hashes and lots of different kinds of things very, very easily. So they certainly have superpowers in some set in some respects. But they also have a bunch of, I would say, cognitive deficits. So they hallucinate quite a bit and they kind of make up stuff and don't have a very good sort of internal model of self knowledge, not sufficient at least. And this has gotten better but not perfect. They display jagged intelligence. So they're going to be superhuman in some problem solving domains. And then they're going to make mistakes that basically no human will make. Like, they will insist that 9.11 is greater than 9.9 or that there are two r's in strawberry. These are some famous examples. But basically there are rough edges that you can trip on. So that's kind of, I think, also kind of unique. They also kind of suffer from anterograde amnesia. So and I think I'm alluding to the fact that if you have a coworker who joins your organization, this coworker will over time learn your organization and they will understand and gain like a huge amount of context on the organization. And they go home and they sleep and they consolidate knowledge and they develop expertise over time. LLMs don't natively do this and this is not something that has really been solved in the R and D of LLMs, I think. And so context windows are really kind of like working memory and you have to sort of program the working memory quite directly because they don't just kind of like get smarter by default. And I think a lot of people get tripped up by the analogies in this way. In popular culture, I recommend people watch these two movies, Memento and fifty First Dates. In both of these movies, the protagonists, their weights are fixed and their context windows gets wiped every single morning and it's really problematic to go to work or have relationships when this happens and this happens to LLMs all the time. I guess one more thing I would point to is security kind of related limitations of the use of LLMs. So for example, LLMs are quite gullible. They are susceptible to prompt injection risks. They might leak your data, etcetera. And so and there's many other considerations security related. So so basically, long story short, you have to load your, you to load your, you have to simultaneously think through this superhuman thing that has a bunch of cognitive deficits and issues. How do we, and yet they are extremely useful. And so how do we program them? And how do we work around their deficits and enjoy their superhuman powers? So what I want to switch to now is talk about the opportunities of how do we use these models and what are some of the biggest opportunities. This is not a comprehensive list of some of the things that I thought were interesting for this talk. The first thing I'm kind of excited about is what I would call partial autonomy apps. So for example, let's work with the example of coding. You can certainly go to chat GPT directly and you can start copy pasting code around and copy pasting bug reports and stuff around and getting code and copy pasting everything around. Why would you do that? Why would you go directly to the operating system? It makes a lot more sense to have an app dedicated for this. And so I think many of you use Cursor, I do as well. And cursor is kind of like the thing you want instead. You don't want to just directly go to the chat GPT. And I think cursor is a very good example of an early LLM app that has a bunch of properties that I think are useful across all the LLM apps. So in particular, you will notice that we have a traditional interface that allows a human to go in and do all the work manually just as before. But in addition to that, we now have this LLM integration that allows us to go in bigger chunks. And so some of the properties of LLM apps that I think are shared and useful to point out. Number one, the LLMs basically do a ton of the context management. Number two, they orchestrate multiple calls to LLMs, right? So in the case of Cursor, there's under the hood embedding models for all your files, the actual chat models, models that apply diffs to the code and this is all orchestrated for you. A really big one that I think also maybe not fully appreciated always is application specific GUI and the importance of it. Because you don't just want to talk to the operating system directly in text. Text is very hard to read, interpret, understand. And also like you don't want to take some of these actions natively in text. So it's much better to just see a diff as like red and green change. You can see what's being added or subtracted. It's much easier to just do command y to accept or command n to reject. I shouldn't have to type it in text, right? So GUI allows a human to audit the work of these fallible systems and to go faster. I'm going to come back to this point a little bit later as well. And the last kind of feature I want to point out is that there's what I call the autonomy slider. So for example, in cursor, can just do tab completion, you're mostly in charge. You can select a chunk of code and command k to change just that chunk of code. You can do command l to change the entire file. Or you can do command I which just let it rip, do whatever you want in the entire repo. And that's the sort of full autonomy agent gen tech version. And so you are in charge of the autonomy slider. And depending on the complexity of the task at hand, you can tune the amount of autonomy that you're willing to give up for that task. Maybe to show one more example of a fairly successful LLM app, Perplexity, it also has very similar features to what I've just pointed out in Cursor. It packages up a lot of the information. It orchestrates multiple LLMs. It's got a GUI that allows you to audit some of its work. So for example, it will cite sources and you can imagine inspecting them. And it's got an autonomy slider. You can either just do a quick search or you can do research or you can do deep research and come back ten minutes later. So this is all just varying levels of autonomy that you give up to the tool. So I guess my question is, I feel like a lot of software will become partially autonomous. And I'm trying to think through like what does that look like? And for many of you who maintain products and services, how are you going to make your products and services partially autonomous? Can an LLM see everything that a human can see? Can an LLM act in all the ways that a human could act? And can humans supervise and stay in the loop of this activity? Because again, these are fallible systems that aren't yet perfect. And what does a diff look like in Photoshop or something like that? Also a lot of the traditional software right now, it has all these switches and all this kind of stuff, it's all designed for human. All of this has to change and become accessible to LLMs. So one thing I want to stress with a lot of these LLM apps that I'm not sure gets as much attention as it should is, we're now kind of like cooperating with AIs and usually they are doing the generation and we as humans are doing the verification. It is in our interest to make this loop go as fast as possible so we're getting a lot of work done. There are two major ways that I think this can be done. Number one, you can speed up verification a lot. And I think GUI's for example are extremely important to this because a GUI utilizes your computer vision GPU in all of our head. Reading text is effortful and it's not fun, but looking at stuff is fun and it's just kind of like a highway to your brain. So I think GUIs are very useful for auditing systems and visual representations in general. And number two, would say is we have to keep the AI on the leash. I think a lot of people are getting way overexcited with AI agents and it's not useful to me to get a diff of 1,000 lines of code to my repo. Like I have to I'm still the bottleneck, right? Even though that 1,000 lines come out instantly, I have to make sure that this thing is not introducing bugs, it's just like and that it's doing the correct thing, right? And that there's no security issues and so on. So I think that, yeah, basically we have to sort of like it's in our interest to make the flow of these two go very, very fast and we have somehow keep the AI on the leash because it gets way too overreactive. It's kind of like this. This is how I feel when I do AI assisted coding. If I'm just byte coding, everything is nice and great. But if I'm actually trying to get work done, it's not so great to have an overreactive agent doing all this kind of stuff. So this slide is not very good, I'm sorry. But I guess I'm trying to develop like many of you some ways of utilizing these agents in my coding workflow and to do AI assisted coding. And in my own work, I'm always scared to get way too big diffs. I always go in small incremental chunks. I want to make sure that everything is good. I want to spin this loop very, very fast. And I sort of work on small chunks of single concrete thing. And so I think many of you probably are developing similar ways of working with LLMs. I also saw a number of blog posts that try to develop these best practices for working with LLMs. And here's one that I read recently and I thought was quite good. And it kind of discussed some techniques. Some of them have to do with how you keep the AI on the leash. And so as an example, if you are prompting, if your prompt is vague, then the AI might not do exactly what you wanted and in that case verification will fail. You're going to ask for something else. If verification fails, then you're going to start spinning. So it makes a lot more sense to spend a bit more time to be more concrete in your prompts which increases the probability of successful verification and you can move forward. And so I think a lot of us are going to end up finding techniques like this. Think in my own work as well, I'm currently interested in what education looks like in together with kind of like now that we have AI and all of them, what does education look like? And I think a large amount of thought for me goes into how we keep AI on the leash. I don't think it just works to go to trashypt and be like, hey, teach me physics. I don't think this works because the AI is like gets lost in the woods. And so for me, this is actually two separate apps for example. There's an app for a teacher that creates courses and then there's an app that takes courses and serves them to students. And in both cases, we now have this intermediate artifact of a course that is auditable and we can make sure it's good, we can make sure it's consistent and that the AI is kept on the leash with respect to a certain syllabus, a certain like progression of projects and so on. And so this is one way of keeping the AI on the leash and I think has a much higher likelihood of working. And the AI is not getting lost in the woods. One more kind of analogy I wanted to sort of allude to is I'm not I'm no stranger to partial autonomy and I've kind of worked on this I think for five years at Tesla. And this is also a partial autonomy product and shares a lot of the features. But for example, right there in the instrument panel is the GUI of the autopilot. So it's showing me what the what the neural network sees and so on. And we have the autonomy slider where over the course of my tenure there, we did more and more autonomous tasks for the user. And maybe the story that I wanted to tell very briefly is, actually the first time I drove a self driving vehicle was in 2013 I had a friend who worked at Waymo and he offered to give me a drive around Palo Alto. I took this picture using Google Glass at the time. And many of you are so young that you might not even know what that is. But yeah, this was like older age at the time. And we got into this car and we went for about a thirty minute drive around Palo Alto, highways, streets and so on. And this drive was perfect. There was zero interventions. And this was 2013, which is now twelve years ago. And it kind of struck me because at the time when I had this perfect drive, this perfect demo, I felt like, wow, self driving is imminent because this just worked, this is incredible. But here we are twelve years later and we are still working on autonomy. We are still working on driving agents. And even now we haven't actually like fully solved the problem. Like you may see Waymo's going around and they look driverless, but there's still a lot of teleoperation and a lot of human in the loop of a lot of this driving. So we still haven't even like declared success, but I think it's definitely like going to succeed at this point, but it just took a long time. And so I think like like this is software is really tricky, I think in the same way that driving is tricky. And so when I see things like, oh, 2025 is the year of agents, I get very concerned and I kind of feel like, you know, this is the decade of agents and this is going to be quite some time. We need humans in the loop. We need to do this carefully. This is software. Let's be serious here. One more kind of analogy that I always think through is the Ironman suit. I think this is I always love Ironman. I think it's like so correct in a bunch of ways with respect to technology and how it will play out. And what I love about the Iron Man suit is that it's both an augmentation and Tony Stark can drive it, and it's also an agent. And in some of the movies, the Iron Man suit is quite autonomous and can fly around and find Tony and all this kind of stuff. So And this is the autonomy slider is we can be we can build augmentations or we can build agents and we kind of want to do a bit of both, but at this stage, I would say working with fallible LLMs and so on, I would say, you know, it's less Ironman robots and more Ironman suits that you want to build. It's less like building flashy demos of autonomous agents and more building partial autonomy products. And these products have custom GUIs and UI UX and we're trying to and this is done so that the generation verification loop of the human is very, very fast. But we are not losing the sight of the fact that it is in principle possible to automate this work. And there should be an autonomy slider in your product and you should be thinking about how you can slide that autonomy slider and make your product sort of more autonomous over time. But this is kind of how I think there's lots of opportunities in these kinds of products. I wanna now switch gears a little bit and talk about one other dimension that I think is very unique. Not only is there a new type of programming language that allows for autonomy in software, but also, as I mentioned, it's programmed in English, which is this natural interface. And suddenly, everyone is a programmer because everyone speaks natural language like English. So this is extremely bullish and very interesting to me and also completely unprecedented, I would say. It it used to be the case that you need to spend five to ten years studying something to be able to do something in software. This is not the case anymore. So I don't know if by any chance anyone has heard of vibe coding. This this is the tweet that's kind of like introduced this, but I'm told that this is now like a major meme. Fun story about this is that I've been on Twitter for like fifteen years or something like that at this point and I still have no clue which tweet will become viral and which tweet like fizzles and no one cares. And I thought that this tweet was gonna be the latter. I don't know, was just like a shower of thoughts, but this became like a total meme and I really just can't tell. But I guess I'll get struck a chord and gave a name to something that everyone was feeling but couldn't quite say in words. So now there's Wikipedia page and everything. This is like Yeah. This is like a major contribution now or something like that. So so Tom Wolf from Hugging Face shared this beautiful video that I really love. These are kids vibe coding. And I find that this is such a wholesome video. Like, love this video. Like, how can you look at this video and feel bad about the future? The future is great. I think this will end up being like a gateway drug to software development. I'm not a doomer about the future of the generation, and I think yeah. I love this video. So I tried bytecoding a little bit as well because it's so fun. So bytecoding is so great when you wanna build something super duper custom that doesn't appear to exist and you just wanna wing it because it's a Saturday or something like that. So I built this iOS app and I don't I can't actually program in Swift, but I was really shocked that I was able to build like a super basic app and I'm not gonna explain it. It's really dumb. But I kind of like this was just like a day of work and this was running on my phone like later that day and I was like, wow. This is amazing. I didn't have to like read through Swift for like five days or something like that to like get started. I also bot coded this app called MenuGen, and this is live. You can try it in menugen. App. And I basically had this problem where I show up at a restaurant, I read through the menu and I have no idea what any of the things are and I need pictures. So this doesn't exist. So I was like, hey, I'm gonna bot code it. So this is what it looks like. You go to menugen.app and you take a picture of a of a menu and then menu gen generates the images. And everyone gets $5 in credits for free when you sign up, and therefore, this is a major cost center in my life. So this is a negative negative revenue app for me right now. I've lost a huge amount of money on MenuGen. Okay. But the fascinating thing about menu gen for me is that the code of the byte the byte coding part, the code was actually the easy part of byte of byte coding menu gen. And this was extreme slog and took another week. So it was really fascinating that I had the menu gen basically demo working on my laptop in a few hours and then it took me a week because I was trying to make it real. And the reason for this is this was just really annoying. So for example, if you try to add Google login to your web page, I know this is very small, but just a huge amount of instructions of this clerk library telling me how to integrate this. And this is crazy. Like, it's telling me, go to this URL, click on this drop down, choose this, go to this, and click on that. And it's like telling me what to do. Like, a computer is telling me the actions I should be taking. Like, you do it. Why am I doing this? What the hell? I had to follow all these instructions. This was crazy. So I think the last part of my talk, therefore, focuses on can we just build for agents? I don't want to do this work. Can agents do this? Thank you. Okay. So roughly speaking, I think there's a new category of consumer and manipulator of digital information. It used to be just humans through GUIs or computers through APIs, and now we have a completely new thing. And agents are their computers, but they are human like, kind of. Right? They're people spirits. There's people spirits on the Internet and they need to interact with our software infrastructure. What can we build for them? It's a new thing. So as an example, can you have robots.txt on your domain and you can instruct or like advise, I suppose, web crawlers on how to behave on your website. In the same way, you can have maybe l n dot t x t file, which is just a simple markdown that's telling LLMs what this domain is about. And this is very readable to a to an LLM. If it had to instead get the HTML of your web page and try to parse it, this is very error prone and difficult and we'll screw it up and it's not gonna work. So we can just directly speak to the LLM. It's worth it. A huge amount of documentation is currently written for people, so you will see things like lists and bold and pictures, and this is not directly accessible by an LLM. So I see some of the services now are transitioning a lot of their their docs to be specifically for LLMs. So Vercel and Stripe as an example are early movers here, but there are a few more that I've seen already and they offer their documentation in markdown. Markdown is super easy for LLMs to understand. This is great. Maybe one simple example from from my experience as well. Maybe some of you know three blue one brown, he makes beautiful animation videos on YouTube. Yeah. I love this library. I said that he wrote Manon. And I wanted to make my own. And there's extensive documentations on how to use Manon. And so I didn't wanna actually read through it. So I copy pasted the whole thing to an LLM and I described what I wanted and it just worked out of the box. Like, LLM just by coded me an animation exactly what I wanted and I was like, wow, this is amazing. So if we can make docs legible to LLMs, it's going to unlock a huge amount of kind of use and I think this is wonderful and should should happen more. The other thing I wanted to point out is that you do unfortunately have to it's not just about taking your docs and making them appear in markdown. That's the easy part. We actually have to change the docs because anytime your docs say click, this is bad. An LLM will not be able to natively take this action right now. So Vercel, for example, is replacing every occurrence of click with the equivalent curl command that your LLM agent could take on your behalf. And so I think this is very interesting. And then, of course, there's a model context protocol from Anthropic, and this is also another way. It's a protocol speaking directly to agents as this new consumer and manipulator of digital information. So I'm very bullish on these ideas. The other thing I really like is a number of little tools here and there that are helping ingest data that in, like, very LLM friendly formats. So for example, when I go to a GitHub repo like my Nana GPT repo, I can't feed this to an LLM and ask questions about it because it's you know, this is a human interface on GitHub. So when you just change the URL from GitHub to git ingest, then this will actually concatenate all the files into a single giant text and it will create a directory structure, etcetera, and this is ready to be copy pasted into your favorite LLM, and you can do stuff. Maybe even more dramatic example of this is DeepWiki, where it's not just the raw content of of these files. This is from Devon. But also, like, they have Devin basically do analysis of the GitHub repo, and Devin basically builds up a whole docs pages just for your repo. And you can imagine that this is even more helpful to copy paste into your LLM. So I love all the little tools that basically where you just change the URL and it makes something accessible to an LLM. So this is all well and great and, yeah, I think there should be a lot more of it. One more note I wanted to make is that it is absolutely possible that in the future, LLMs will be able to it's not even future, this is today. They'll be able to go around and they'll be able to click stuff and so on. But I still think it's very worth basically meeting LLM halfway, LLMs halfway, and making it easier for them to access all this information because this is still fairly expensive, I would say, to use and a lot more difficult. And so I do think that lots of software, there will be a long tail where it won't, like, adapt because these are not, like, live player sort of repositories or digital infrastructure, and we will need these tools. But I think for everyone else, I think it's very worth kind of like meeting in some middle point. So I'm bullish on both, if that makes sense. So in summary, what an amazing time to get into the industry. We need to rewrite a ton of code. A ton of code will be written by professionals and by coders. These LLMs are kind of like utilities, kind of like fabs, but they're kind of especially like operating systems, but it's so early. It's like nineteen sixties of operating systems, And and I think a lot of the analogies crossover. And these LLMs are kind of like these fallible, you know, people spirits that we have to learn to work with. And in order to do that properly, we need to adjust our infrastructure towards it. So when you're building these LLM apps, I described some of the ways of working effectively with these LLMs and some of the tools that make that kind of possible and how you can spin this loop very, quickly and basically create partial autonomy products. And then, yeah, a lot of code has to also be written for the agents or directly. But in any case, going back to the Ironman suit analogy, I think what we'll see over the next decade roughly is we're going to take the slider from left to right and I'm very interesting it's gonna be very interesting to see what that looks like and I can't wait to build it with all of you. Thank you.
Saved - January 11, 2026 at 6:57 AM
reSee.it AI Summary
MilkRoadAI kickstarted with Jensen Huang’s 2011 Stanford talk and Nvidia’s later ascent to a multi-trillion value, urging preservation. lkondeth asked for a 10-point summary and 5 actions. grok’s version covers: lasting impact comes from a unique perspective; Nvidia started in 1993 with 3 engineers targeting 3D graphics; early rejections but persistence; 3D graphics enabled broader apps; strong competition; GPUs became programmable; embrace failure and honesty; passion and big markets; leadership cultivation. Five actions: pursue a unique view; take calculated risks; reinvent often; foster honesty; value passion and learning.

@MilkRoadAI - Milk Road AI

Forget the $200,000 degree. In 2011, Jensen Huang gave a lecture at Stanford that explains strategy better than any professor. NVIDIA is now worth $5 Trillion. Save this post. You won't find this video again once your feed refreshes: https://t.co/T79TtP74Mv

Video Transcript AI Summary
Jensen Huang opens by inviting an interactive conversation about building a company, noting that it is both gratifying and incredibly hard, with perspectives on company building shaped by diverse experiences. He recalls NVIDIA’s beginnings sixteen years ago with three engineers and introduces the idea that perspective, more than grand vision, drives entrepreneurial direction. He distinguishes vision from perspective, arguing that vision is not exclusive to a few, while everyone has a perspective—the way you see the world and identify opportunities. In 1993, with Windows 3.1 era and no networks or wireless tech, Huang explains NVIDIA’s perspective: a PC could run three-dimensional graphics programs to explore new worlds, enabling video games as the killer app. The business plan was to take advanced graphics technology from expensive workstations, reinvent it, and make it affordable. He recounts pitching to Sand Hill Road, who doubted a video game market existed, and a parental nudge to get a real job. Yet the team believed video games would be a large market, a view later validated by today’s status as the world’s largest digital media industry. They also anticipated broader uses for the technology beyond games, such as a notable example with Keyhole (which Google acquired to become Google Earth, the world’s largest downloaded application). He emphasizes that perspectives often differ even among seemingly obvious opportunities. He cites Yahoo!, AltaVista, Lycos, and others, illustrating how two similar cores (search) could lead to different outcomes based on what each company chose to become (destinations/portals, etc.). Competition was intense as hundreds of three-dimensional graphics startups emerged, yet NVIDIA remains the only surviving graphics company. The lesson is that perspective matters because different viewpoints shape strategic focus. Huang then discusses the core business principle: Moore’s Law—though framed as a competition-driven efficiency—drives GPU advancement. The early approach was to make three-dimensional graphics insatiable—improving performance year after year even if customers initially resisted due to cost. For the first five years, NVIDIA “turned off the blinders” and ignored customer constraints, eventually cannibalizing its own products when a new generation proved more capable and profitable. Innovation is risky, he notes, and sustaining a leading position required reinvention. By the late 1990s, NVIDIA shifted from a fixed-function graphics accelerator to a programmable shader architecture with the GeForce FX (a gamble that nearly killed the company but ultimately paid off). The introduction of programmable shaders kept NVIDIA at the forefront, enabling GPUs to be used for general-purpose computing (GPGPU), which has become a major trajectory. On company culture, Huang stresses the importance of fostering risk-taking and a tolerance for failure, teaching people how to fail quickly and cheaply, and maintaining intellectual honesty to pivot when necessary. He contrasts older, more rigid corporate cultures with modern, beta-form experimentation found in companies like Google, where many applications operate in beta to test ideas rapidly. Regarding cofounders and governance, he notes that equity was divided equally among the three founders (each initially contributing $200 and receiving 20% each). He explains that leadership should be clearly established (Jensen as CEO) to avoid decision-making gridlock, while still valuing collaboration with strong, trusted partners. Asked about the venture capital process, Huang explains that VCs invest in people and a sufficiently large, novel market, not just a polished business plan. He shares that their reputations and prior work with notable figures helped, and he emphasizes the ongoing importance of great people and a focused, strategic vision. He addresses mentors and best advice—focus intensely on a few things, learn from diverse sources, and remain adaptable. On succession, Huang argues against rigid, preselected succession planning, favoring the cultivation of future leaders within the company so that many internal options exist if leadership changes become necessary. Finally, he speaks about the finance side in the early days: cash is king and survival is paramount, constantly raising or conserving funds. He closes by reiterating the core message: ideas are plentiful, but a unique, passionate perspective and perseverance are what sustain a company, along with a culture that embraces calculated risk and continuous reinvention.
Full Transcript
Speaker 0: Instead of giving you a company presentation today, what I thought I would do is just have a conversation with you. At any time, if you have a question, if you would like to change the direction of conversation, just raise your hand, and we'll talk about whatever comes up. A lot of people talk about and write about building companies. And I can tell you firsthand that building a company is extraordinarily gratifying. It is also incredibly hard. And so the things that you want to talk about with respect to the company building process is rather expansive. You could talk about company building processes from a lot of different perspectives. And so I'm going to try to touch on a few of them that I think are particularly important in my experience. So sixteen years ago, NVIDIA had three people, three engineers. Speaker 1: Did Speaker 0: I do something? Was it me? We had a perspective that if I was just Speaker 2: You think it's me? Think it's. Speaker 0: How about I just did that? See, it's you can't be you can't be you can't control user stupidity, you know. You're good. You're good. It was probably me. I was sitting on it. All right. So sixteen years ago, we started NVIDIA. And the insight that we had, some people call it vision. Vision is an awfully big word to me. Vision is an awfully big word to me because I believe first of all, vision matters. Let me tell you that vision matters, and I'll help you understand that in a second. But I like to use the word perspective because it makes it possible for anyone to have one. When you say vision, it feels like only a few selected visionaries of the world can have one. But everyone has a perspective, and that's, in fact, all vision means, that you see the world in a way that is either different or otherwise, okay, than somebody else. And you see opportunities that I think are that you believe are particularly important to go in and address, that you can address in a particular way. And so perspective. Our perspective at the time, this is 1993. You guys won't remember this, but the PC was Windows 3.1. CD ROM was about to be introduced. There were no PCs with networks. Wireless technology, the no if you said some if somebody said radio, I think you would the word that would come to mind is FM radio. And so wireless technology didn't exist. The fastest microprocessor in the world was a 66 megahertz, 46DX2, and I don't think any of you would even use it in your tennis shoes today. And we would run our computers with that. And the PC was becoming used for desktop or for office automation. Our perspective was that this particular device was going to be unique in the sense that it has the ability to run programs. And what if we gave it the benefit of running three d graphics programs so that you could explore new worlds, play games, play games. And so we started a company, and the business plan basically read something like this. We're going to take technology that was available only in the most expensive workstations. We're going to try to make it reinvent the technology and make it inexpensive. And the killer app was video games. And so I took this idea to Sandhill Road, and they told me there was no video game market. People don't start companies to play games. And my parents might I remember calling my mom and telling her that I'm gonna start this company. And she says, you know, what do you what do you guys do? And I said, we build these things called three d graphics chips, and and people would use them to play games. And then she said, why don't you go get a job? And so now of course, games was, we believe, going to be a very large part of the marketplace. Now we had that perspective for very obvious reasons. We grew up in the video game generation. I was the video game generation. I was the beginning of the video game generation. And so the entertainment value of video games, computer games, was very obvious to me. And I could imagine how it could be a very large market and a very large industry. For a lot of the people that were older, that sensibility didn't exist. And so notice, I've just described to you a perspective about the world that we had that is apparently, obviously, now true because video games is the world's largest digital media industry today. It is apparently true. And yet at the time, our common sense was unique. Nobody would have created the technology, nobody would have created the company with the sole purpose of building technology to make video games possible. And so that was our perspective. Now we felt that video games would, of course, fuel the technology development, but you could use this technology for a whole bunch of other reasons. And one of my favorite applications, this happened about, I guess, about five years ago, a small company, struggling company here in Silicon Valley called Keyhole. And they were they created a three d virtual world, and it had no application. This three d virtual world, and you start out in space, you see the earth, you zoom into any location you wanted just by typing in the address. I thought it was such a fabulous way of exploring the world, going to places you've never been, and they couldn't raise a penny. And so we I was so excited about the company. We put money into the company. And I went everywhere and showed that demonstration. I would tell people that this is the way we're going do search someday. If you want to search for something, look for an address, you would type it in and we would fly you there, okay. And satellite images will continue to download. And before you know it, you're right there on the street and you might even see some buildings. That small company was eventually purchased by Google and became Google Earth. Google Earth is now the single largest downloaded, most frequently downloaded application in the history of mankind, over 200 plus million downloads. So three d graphics could be used for a lot more than video games. Now that vision, if you will, that perspective, was unique at the time and hard to sell. And so we had to go and explain it to venture capitalists who had to figure out whether the technology was going to be possible, how big was the market because it was $0 at the time. So how do you extrapolate, how do you scope the size of a market when its apparent size was zero at the time? And you look at analyst reports and you study market research, and all of it would say approximately zero. It would never show up. It's a noncategory, a nonmarket. And so it's incumbent upon the venture capitalists and, of course, the founders to try to figure out how to inspire each other into doing something together. And so Sequoia Capital and Sutter Hill were our venture capitalists, and we got the company going with $2,000,000 Now the question about perspective becomes very interesting in other opportunities in the company, other circumstances in the company. And let me give you some examples. Many years later, Sequoia Capital came to me and said, you know, there's a couple of kids at Stanford, and they have this thing, and it's an Internet thing. And you just type what you're looking for, and it shows up puts up the website. And I said, yes, Yellow Pages. I mean, no duh, right? We use it. We and there's a variety of versions of it on the web at the time. We used the Internet, just like everybody else, to do FTP and also to visit various websites. And they said, should we invest in this company? And I said, there's no freaking way they're going to make money. That stuff is free, right? And so they said, well, we can't figure it out ourselves, but it doesn't cost much to give them $1,000,000 or 2,000,000 And they invested in a small company called what eventually became Jerry's company called Yahoo! Notice, although I had the perspective about one thing, I didn't have the perspective about something else. Just because you're a visionary doesn't mean you're a visionary by everything. Your perspective stems from your life experiences, what's commonsensical about you, what's interesting to you. And so that's important to realize that you have perspective too. Therefore, you have vision too. Now what's interesting about these websites to follow on the Yahoo! Story, if you remember, were several other searches out there. AltaVista at the time, Xcite at the time, Lycos at the time, right? And now the question is, they're all doing search and now they did a reasonably good job. Now comes the question is what was their perspective? How did they how was their perspective different from one another? One website thought that they were a destination. Do you guys remember that? We would be a destination, kind of like a channel. Somebody said, in fact, since we're going to be a destination, we would serve up content. And therefore, the search part of it is a commodity. We'll outsource that. So all of the search engines, which started out as search, turned into destinations or portals, and they outsourced the search to someone else, which made it possible for Google to start. And so notice two companies doing exactly the same thing started with the exact same fundamental core technology, ended up in radically different places because they had different perspectives. They saw the world differently. So perspective matters. Vision matters. Now in our industry, shortly after we were started, three d graphics for PCs and consumer three d graphics became the hottest, hottest thing. And so everybody in Silicon Valley was starting a three d graphics company. We were, in 1993, the only consumer three d graphics company in the world. Silicon Graphics up the street was the professional, if you will, three d graphics company. By the end of a couple of years or so, 1995, there were probably fifty, seventy start ups doing exactly the same thing we were trying to do. And over time, we competed with about 200 companies. NVIDIA today is the only surviving computer graphics company in the world. And so the question is then, what happened? Competition is intense. Everybody has smart people. Everybody has money. We competed with IBM. We competed with HP. We competed with Silicon Graphics. We competed with Sony, three d FX, S3, Sirius Logic, big, small, international, local. We competed with companies all over the world. So the question is what happened. I would argue that 300 companies armed with exactly the same technology, armed with exactly the same people, the company that wins and let's say they all execute, and they did. With 300 companies, 50% of them are going to execute at any given point in time. And so the question is, why does one survive? Well, I think that it matters to have perspective, and let me give you some examples. I always believed that you need to understand the reason why your business work. What is the essence of your business? What makes it work? Now the foundation of my business, at its core, is semiconductor technology. Here in Silicon Valley, we usually like to refer to semiconductor technology as Moore's Law. Moore's Law is not so much a physical law as it's a law of competition. It is a law of challenging engineers. It's a law almost of setting pace. And Moore's Law approximately gives you twice the performance every year or two. And so understanding the fundamental ingredient of our business improves by a factor of two every year and simultaneously reduces in cost by a factor of two every year. The question is what makes a survivable business. And so our first perspective was that three d graphics was insatiable. It was insatiable. That if I made something twice as good every year, even if the customer never asked for it, even if the customer told us it was too expensive, even if the customer, when you went to float that product specification to them, told you that they're not interested. And in fact, that was the case. I took our product spec to Dell and HP and IBM and Gateway, and they all told me it was too much money. You're well outside of the boundaries of what they were willing to pay for. When your customers all tell you not to do something, the question is then what do you do? In our case, because we had this unique perspective that three d graphics was insatiable and Moore's Law was our friend, therefore. We should make our graphics processors twice as good every year. And so for the first five years of our company, we just turned off our blinders and said, we're going to ignore customers. Now which one of you guys are going to go through your marketing courses and the lesson that it teaches you is ignore your customers? Well, sometimes you have to ignore your customers. And the reason for that is because they don't know the nature of your business. And while the industry is being created, before there's common sense about the rules of that business, there is no way they can possibly know. And so we I took the last few million dollars of the company's money and built a chip that is way, way, way too big. And our customers told us they we were way out of bounds on cost, and they weren't going to buy any. Until the day we showed up with the processor, we were in allocation throughout the entire life of that project, until our next generation product, which was twice its price, cannibalized the previous one. And so we grew and grew and grew for several years. Then the question became, what now? Now you guys are going to learn that innovation is a rather dangerous thing. On the one hand, once you discover a great idea, once you discover a great idea, you rinse and repeat, rinse and repeat, and you make that idea better and better and better. Whether it's a laptop computer that you guys have here or a car or a microprocessor or, in our case, a graphics processor, we made it better and better every year. At some point, it becomes good enough. Moore's Law is a wonderful thing. Semiconductor technology enables you to make amazing leaps and bounds in technology. And at some point, it becomes good enough. And so in the this is probably in the late '90s, about seven years into our company, maybe six, seven years into our company. I came to the conclusion that three d graphics was not going to be sustainable as an accelerator or a fixed function device that renders texture maps and polygons on the screen. That we had to change the company to make the three d graphics processor programmable so that it could be an artistic medium for expression. Now this is a weird word. Here we are, an engineering company, and we now want to change this chip to become an artistic expression, an artistic medium so that all of the video games, so that all of the applications that were developed on our chip would be stylistically different. And we believed unless we could figure out a way to make the content richer and more interesting and stylistically different from one game developer to another game developer to another, we would limit the life of our medium. And if our medium reached its end, we, as the world leader, would also see our end. And so we decided to make the GPU programmable and make it a medium for artistic expression and invented a technology called programmable shader. Almost every single video game that you guys see today has our fingerprint on it, whether it's a Xbox three sixty or PS3 or any PC game today. You could see elements of what programmable shaders made possible. That started a whole new innovation curve for us and kept our industry vibrant. But the crossing from one generation of technology to the next generation almost killed the company. And so that process of reinventing the company, the perspective that led us to a new idea also risks the company in the process. And those are interesting conversations that we can have. So there's perspective matters. When a company gets larger, you guys are going to learn that as the founder or as the CEO, you have to learn new things. And many of the new things that you'll learn has to do with building products at first. And I've just talked to you about building products. Soon, you'll be talking about and learning about building companies. And building companies means things that are soft and hard to explain, like building a company with a culture. What does that mean? How does the culture of one company different from a culture of another company? And why is it that this particular culture is better for your company and not for another? So the culture of a company is important to find out, to put your arms around and to create and develop. How do you organize? Are you we were just talking earlier with one of the guys. Are you functionally organized? Are you organized in business units? How do you deal with multiple products and multiple geographies and multiple customers? And so that's the company building process. It's mechanical. It's interesting. Lots of trial and error. It's organic. People matter. Personalities matter. And if you guys are interested in talking about that, I'm happy to talk about that as well. And then I would say probably the most important thing above that is to realize that building a when you're building a company and building a product, skill matters, intellect matters, training matters, but it's not enough. The part of it that is important to realize about building companies is that it's a challenging and painful and oftentimes extraordinarily scary thing to do. And so unless you have passion, unless you really love the process of building the company and what you're trying to do, it's going to be incredibly, incredibly challenging. And so what I would leave you with is when you're building a company, if you decide to build a company, you have to ask yourself what is the purpose that you're building the company for. Is it that you would like to build a company so that you can sell it, make a fortune? Is it that you would like to build a company so you can take it public? You're just a serial entrepreneur. You want to build something, let somebody else run it, build something, sell it. Whatever your reason is, be honest to yourself. It turns out that, in my case, I just love the process of building things. And I love being part of something. You know, being part of NVIDIA and being part of a people and being part of a cause that's inspiring to me keeps me vibrant, and and it's something that I'm willing to do for a very long period of time. And so I've been in it now for sixteen years, and I've learned a lot in the process of being the CEO and the founder of the company. And so maybe the thing to do is why don't I open it up now and let you guys ask me whatever type of questions you guys have. Yes, sir. Speaker 3: What did your investors say about your idea of reinventing your company to include programmable Speaker 2: GPUs? And Jensen, could you please Speaker 0: repeat Yes, I will. The The question is what did our investors say about us taking the big risk of adding programmable shaders and reinventing that product category and reinventing our company. First of all, it's not a conversation you really have with your shareholders, but you do have the conversation with your management team first, your employees second, your Board of Directors third, typically is the way the process that I take. And when you're in high-tech, when you're in a technology industry, when the technology moves this fast, if you're not reinventing yourself, you're just slowly dying. You're just slowly dying, unfortunately, at at the rate of Moore's Law, which is the fastest of any rate that we know, right? The compounded rate of Moore's Law is pretty unbelievable. We have a very Speaker 3: successful product that was generating a lot of money. Speaker 0: I know. So it's scary. And so you have to and this is so the question is, but we have a product that's generating a lot of money and it's very successful, how do you cannibalize it? There is a theory that if you don't cannibalize it, someone will and it surely will be cannibalized. And so if you want to be a market leader, you have to take the initiative to cannibalize your own products and have your ideas cannibalize your own ideas. When we went from a fixed function graphics accelerator, a texture mapping engine for games like Quake III and Doom and those kind of games, to a programmable shading architecture, our first chip almost killed the company. It was called GeForce FX. I don't know if any of you have ever owned one of those. GeForce FX is a chip, is a processor that it's a baby only a mother can love. I mean, it's we took enormous chance in building GeForceFX, but it almost broke our back. But if it if we didn't build that chip, I am sure NVIDIA would be dead today. I am absolutely certain we'd be dead. It was one of the biggest gambles in our history. We had to create instead of an API, we had to go to a processor with a language, we call it CG, and a compiler. So it's kind of like a processor. We introduced a new programming paradigm to the world that it never understood in the beginning. And so it took a lot of evangelism, a lot of marketing, a lot of education. But CG inventing CG took us to unbelievable places. And one of our most important work today is related to using GPUs for general purpose computing. And it's extraordinary that the results of what we're seeing. And it wouldn't have been possible if it wasn't because of CG that we started then. Okay. So you have to take these leaps. Questions? Yes, sir. Speaker 1: Some people are sort of see that gaming is sort of moving more towards consoles and Speaker 0: away Is from Speaker 1: it harder to make profit with when you have to negotiate with Sony or Nintendo or whoever, some big company, than it would be when you sell individual cards? Speaker 0: Yeah. The economics the economics of it of building anything ultimately comes down to the amount of competition you have. You don't set the price, the competition sets the price. The market doesn't set the price, the competition does. And so if your competitor wishes to build PlayStation three as much as you do, then the economics will be challenging. And so there, you just need to decide, is an economic decision for you? For example, there's there are rumors that we were not enthusiastic in building some of the game consoles, and we were more enthusiastic in building other game consoles. It came down to this for me. I think the you have to realize what is the finite resource, what is the scarce resource that you have as a manager. The function of a manager is to allocate resources properly for the best return. And so if you think about our resources, our resource is the finite number of extraordinary engineers and how much time they have in a day to pursue whatever opportunities that are out there. Now if the number of opportunities that are out there is less than my supply of engineers, so therefore, exceeds is less than my supply, then obviously, I'm very enthusiastic about it. But if it's the other way around, then the opportunity to build a game console at terrible economics or any project at terrible economics is simply not worth it. And so I look at it irrespective of competition. The competition sets the price, but then I get to decide whether I want to engage in that project or not. You are in charge of your own company as the CEO, right? And so we decide whether it's economic. And once you decide, then it is what it is. So you have to be thoughtful about what is your critical resource. Do you have more of it or less of it than the market demands? Do you have more opportunity or less opportunity than your resource can support? And then what's the appropriate return on that investment, thinking about not just your cost, but more importantly, your opportunity cost? And so we look at it from that perspective every single time. Speaker 1: Can you tell us a little bit about the culture that you try to set at NVIDIA? Speaker 0: Yes. That's a good question. The question is what is the can I talk about the culture that we're trying to set at NVIDIA? At the core of our company's success is innovation. Now a lot of companies say innovation is important to their company. Invention is important to their company. However, I don't believe you can fundamentally say that innovation that you want as a CEO to nurture the spirit of innovation, to encourage innovation unless you have a culture of risk taking. We have to encourage our engineers excuse me, our marketing people, right, all of our employees to take calculated risks. In order to encourage them to take calculated risks, first of all, you have to teach them how to do that. That's a skill, a matter of skill. Then the second part of it is a matter of courage. Most people hate to fail. Do you guys agree with that? Well, unless you guys want to be successful let me say it the positive way. If you want to be successful, I would encourage you to grow a tolerance for failure, to develop a tolerance for failure. Now when I mean a tolerance for failure, I don't mean, gee, what Jensen just told me is sleep in until noon, okay? Don't do any of my homework. Flunk out of all my classes because that defines failure. Right? That's not what I said. What I said is what I was trying to say is that I want you to try things even though it is impossible to calculate precisely that it would lead to success, that your instincts and your intuition is something you ought to follow. If it wasn't because of following my own instincts or the founder's instincts or many of our employees' instincts, why would we be where we are today? And why would we have invented things that the markets never had before, the world's never had before? So you have to have this culture or tolerance for risk taking. But the thing about failure is this. If you fail often enough, you actually might become a failure. And that's different than being successful. And so the question is, how do you teach someone how to fail but fail quickly and to change courses as soon as you know it's a dead end. And the way to do that is we call it intellectual honesty. We assess on a continuous basis whether something makes sense or not. And if it's the wrong decision, let's change our mind. And a lot of people say CEOs are always right, and they never change their mind. That doesn't make any sense at all to me, especially when it violates the first principles of what we want the company to become, an innovative company that invents amazing things, that solves problems for the world that it sometimes didn't even know it had, If you want to do that, then you have to cultivate that tolerance for risk taking. And you have to then teach people how to fail but fail quickly and inexpensively and how to be direct with each other that this is the wrong approach and what's the better approach and then, you know, be flexible enough to change courses and quick. And so that type of culture, if you will, in today's if you guys were to start a company and you were build building a website with an Internet service of some kind, Internet based service of some kind, with the competition coming from all over the world, and it's twenty fourseven, and ideas take no time to experiment, and it a particular website or a particular company could be throwing ideas out into the world 20 a day. And so unless you are thoughtful about risk taking and being able to change your mind, reacting to the market conditions and being flexible, how are you going to stay alive? And so you could almost see what I just described in the nature of older companies and the nature of the newer companies. The modern companies, if you guys you got I'm sure you guys all go to Google's website. Almost every single application is in beta form. They're trying all kinds of stuff. Right? They're trying all kinds of stuff. If they call it production and it doesn't work well, you guys would just be upset at them. So they call it beta. Have you noticed? They call it beta so that they could try a lot of things. And if it fails, take it out. If it's bad, take it out. If it works, do more. And so innovation requires a little bit of experimentation. Experimentation requires exploration. Exploration will result in failure. Unless you have a tolerance for failure, you would never experiment. And if you don't ever experiment, you would never innovate. If you don't innovate, you don't succeed. You'll just be a dweeb. That's it. Any other questions? Yes, sir. Speaker 2: How did you choose your cofounders at additional time? Speaker 0: Don't ever go into business with anyone you don't deeply trust. And they were two my closest colleagues, and I trust them completely. And they're wonderful friends even today. So by the way, as a CEO, selecting people is 99% of the job. Speaker 1: What applications do you see driving demand in the general purpose EV market, which is like that's going to be like the next big thing the semiconductor? Speaker 0: So the question is what is what do I see driving the demand for this thing that we're pushing right now called GPU computing, using the GPU for much more than just graphics. For graphics, there's a model of graphics that we call computational graphics. So it's using programs to generate the images. The algorithms are no longer cast in the silicon. The algorithms are actually software. And it could be ambient occlusion. It could be ray tracing. It could be all kinds of interesting algorithms that people are going to explore for the future. We observed and it was, in fact, this is to give you another example of innovation and with Stanford is fabulous. We when we invented GeForceFX, although it wasn't a very successful GPU for graphics, Researchers around the world noticed that it had a programming language called CG, C for graphics, and that you could program this GPU to do other things aside from graphics. And it had 32 bit floating point, IEEE compatible 32 bit floating point. And so some smart researchers, many of them were here at Stanford, just bought a graphics card from Fry's and started writing programs. And they discovered that if they really worked hard and do all these algorithmic gymnastics, they could get something, an algorithm it could be nanomolecular dynamics, it could be computational fluid dynamics to run 20x faster. And they couldn't believe it. How do you speed up an application 20x? Well, the interesting observation that we made is that we speed up three d graphics applications, which is basically something you can do in software, 1000x over a CPU all the time. So what if we took all of those parallel processors that were inside our GPUs and make it completely programmable and expose it through a programming language called C, right? C and now in the near future, C plus plus Imagine the type of problems we could help solve. And so whether it's weather prediction or seismic analysis or taking your CT scans and reconstructing the human image for the body from it, all kinds of very computationally intensive applications, we could accelerate 50x, 100x. Now just to put it in perspective, 100x is a Moore's Law time, approximately ten years. Now to put ten years' worth of computing resources in the hands of scientists, researchers, engineers, unbelievable benefits. And so the risk was really large, however, to make our GPU even more general purpose. Because every time you make something general purpose, you know what, right, a Swiss Army knife? It's dangerous because whenever you make something general purpose or a Swiss Army knife like, you move away from your core business. It's much, much better to have a very specific niche, to have intense focus on a particular market segment. And you guys will learn all of this in marketing. When you make something general purpose, you're all things to all people, you become, you know, what is it? What is it? Jack of all trades, master of none. Very, very dangerous move. Now we thought that it was just too important for us not to do it, so we decided to make that move. And it's fabulous results. Okay. So those are some of the things that we're seeing now. Yes, ma'am. Speaker 2: Mentioned you're the cofounders of NZD. So especially in the initial stage, how can you find each of your investment in the company? And how can you distribute the profit? Speaker 0: Okay. So her question is we were friends in the beginning, we're friends now. How do we figure out who's the right position and how do we distribute the profits? Okay. And so I won't say anything funny just as I'm not misunderstood, but all of our pay were identical. And we all had identical share in the company. So that's that's just simply fairness. Now the question becomes governance. There's the part of it which is equity. Equity is another way of saying what's fair. Right? So we all had the same salary. All three of us were making $100,000 a year. Okay? And we all had a percentage of the company, equal percentage. Now you can't run a company, though. You can't build a great company. When you have three people who has to vote on everything and with equal share of responsibilities. You simply can't. That becomes a leadership question. That becomes a governance question. That becomes a management question. Right? That becomes a question about building a great company. I don't recall exactly the conversation, but I think it kind of went like this. All right, Jensen, you're the CEO, right? Okay. That was done. That was basically the process. I think that some people are I'm not particularly I'm not from a personality perspective, I'm not particularly outgoing. And so that's not a necessity for being a good CEO. But as a personality, I've always been able to see around the corners, if you will. I can see around the fuzzy edges. And I think CEOs and leaders need to be comfortable with ambiguity. Ambiguity meaning that, you know, what does the future look like? Well, it's hard to say. Some people hate that. Some people just say, Jensen, tell me what you need to have done and for how with how much resource and by when. Okay? Some people rather me tell them that, hey, look, there's this there's this opportunity out there, not sure what it is, not sure how big it is, but it kind of feels like this. Let's go figure it out, and let's build a business. Some people can are very comfortable with that. And so this this ambiguity is is important to to to to be comfortable with, I guess. And I think that all CEOs that are very successful are comfortable with ambiguity. And I'm very I am very comfortable with ambiguity. Yes? Speaker 4: Yes. What percentage of your initial investment was yours? Like how do you get the rest? And also, like, how many times was your proposal for an for help with the investment was rejected? Speaker 0: Mhmm. I was 30 years old, and I'd never taken a single business class, and and I've never taken any marketing classes. And and I've never used never used the at the time, it wasn't PowerPoint. It was called Persuasion on the on the Mac. It was called Persuasion. And so I bought a Mac so I could I could use Persuasion. And and and then I tried to create a company presentation to take it to venture capitalists. The process kind of went like this. We started my first official day of work was my thirtieth birthday, February 17, and we got the company funded. And so once we got started, the question is, what are we gonna do? You know, how does it all work out? How do we start the company? And so we met every day, the three of us, in in one of the founder's townhouse in Fremont. And and we would get together, and and there would be nothing to do. I mean, what do you do? You get three guys and get together. You just talk. You know? So what did you guys do last night? What did you have for dinner? I mean, so you talk about that for about six months. Okay? And the big event of the day would be, hey, where do guys want to go to for lunch? And so Philly cheesesteak today or some Chinese food tomorrow or whatever. That would be like a big deal. And then after a while, it was like, could you put some donuts in the fridge in the morning for when we come? I mean, so that would be a big deal for a while. And so that lasted for a few months, just the three of us like that. I know it sounds pathetic, but it's it's it's true. Because at that time, I'm reading about books on how to start companies, and I'm trying to figure out, you know, how to go raise money and, you know, what's a venture capitalist and how do you incorporate the company and those kind of things. And pretty soon, I met a I met a met a lawyer, went to a went to a law firm called Cooley Godward, and they helped us incorporate the company. And the amount of money that he he he says, you know, we need we need some money from you so that we could price the shares and also to incorporate the company. So he says, how much money do you have in your pocket? I said, $200. So he says, okay. Give me $200. I gave him $200. And for $200, I bought 15%, I think it was 20% of NVIDIA. So it was a good deal. Yeah. 20%. Yeah. And then I went I went back to the house, and then I went back to the condo, and and they all they both gave me $200, and they both got 20%. And that's how it worked, literally. Yeah. It's not that much more. You know, don't here here's here's the thing. NVIDIA, I never finished my business plan. I know it. I know it. We we never finished a business plan, never could figure out how to finish a business plan, to tell you the truth. And and if if I would have finished that book, and I I went to went to Borders and got Gordon Bell's book, How to Start a High-tech Company, it's like this thick. If I would have read the whole thing, I would have been dead now. We would have run out of money, run out of time. And so I I I read I read the first three or four chapters, and I, you know, I I gotta go to work. And so so I I incorporated the company. They introduced us to two venture capitalists, and I just went to their office and told them what I'd like to do. The thing that gets the company funded and when you're when you get to that point, you just have to remember a few things. VCs don't invest in business plans because business plans are easy to write. I couldn't write it, but other people could. Right? And so so they invest in this. They invest in great people. And so the so the question is is do they trust you? Your reputation matters. Your history matters. Because because I had done so much work with Andy Bechtelsheim, which was another graduate of Stanford of Stanford and the founder of Sun and and worked with the founders of Synopsys and LSI Logic. And the and, you know, we we we were all very successful, and we did good work. Your reputation will precede you even if your business plan writing skills are inadequate. And the second thing is you need to have a vision that's sufficiently large to invest in because their statistics, their probability of success is rather low. And if they need to put in $10,000,000 if the market is only $20,000,000 large, they'll never get that $10,000,000 back with reasonable return. But if it's a $200,000,000,000 market, then, of course, it's a rather different thing. Okay? So the size of the market. And they want to know that at least there is a clever idea that the market has never done before. So that last part is probably second, you know, last. I said it last because also I think it's least important. You have to you might have to reinvent yourself over time. And if you want to reinvent yourself, you need to have great people. That's why great people is so important. Yes, sir. Who Speaker 2: were some of the people that you considered to be your mentors when you were getting started? What was some of the best advice that you got from them? Speaker 0: So the question is what what who are some of the mentors, and what were the best advice I got? I I truly believe that if you wanna be successful, you a successful habit is to have the capacity and the willingness to learn from just about anybody. And I do. I learn from just about anybody. And it could be a little thing, could be a big thing. You know, if it wasn't because of my kids, I would be I would miss the whole Internet age. You know, I would have missed YouTube and Facebook and Twitter and you know, I mean, without so you you need to you need to know that that the world changes, and and you wanna be able to learn from just about anybody. And so I I'm surrounded with extraordinarily talented executives and professionals of of all walks of life. And so you just have to make sure that you're you're willing to learn from just about anybody. Some of the some of the great advice that I've I've had over the years, focus. Laser beam focus. You know, don't do too much. Do a few things well and do it with extraordinary intensity. And focus matters. If you look at what I do with my time, I wake up in the morning, and the first thing of my time is NVIDIA, and the last thing I do is NVIDIA. And I do that 20 fourseven. And if I could figure out a way to do that for another fifty years, we're going to be in good company. Yes, ma'am? Speaker 2: Successful entrepreneur and female, Speaker 5: so what's your greatest challenge at the current? Speaker 2: With your ambiguity as CEO, what's your best and worst estimation of the future of you and your company? Speaker 0: Okay. So her question is she started out by saying that I'm a successful entrepreneur. And what are my biggest challenges now? And considering that I like ambiguity, what's my best estimate of the future for the company or and and for myself? The biggest challenge with building a company is the reinvention of the company. Every successful thing needs to be torn down at some point and be rebuilt. It is unfortunate but true. And the reason for that is because the technology either gets good enough, and therefore, you have to reinvent. And sometimes the invention process is disruptive. Sometimes it's, in fact, destructive. And it could be it could destroy you could destroy what you have built in the past. And so the reinvention process is very challenging. It's gut wrenching. It takes a lot of courage, and it it really tests your conviction. In in the technology industry, reinventing the company every ten years is almost a necessary thing. And so that's when I say challenging, I don't mean bad challenging. I think that's fun challenging. I love the process of reinvention. Okay. So that's fun challenging. What's my best forecast for our company? I think that NVIDIA has the opportunity to become one of the most important technology company companies in the world. And I hope that I hope that it does. And my best forecast for me is that I am 80 years old, and I'm here talking to students, and I hope I'm still the CEO. Yes, sir. Speaker 6: Questions about the first few years of when you start, very critical. You actually survive and make sure your cash is positive. Speaker 0: Yes. And Speaker 6: the constant cash flow, cash is king. So in NVIDIA case, how do you manage that cash Speaker 0: or Yes. So the gentleman's question is has to do with in the beginning, survival is important, cash is king. Just so that there's no no no ambiguity about this, survival is always important. Cash is always king. And so as the CEO, you're either making money, saving money, or raising money. And and if you're not making money, raising money and saving money, you ought to be doing those three things. It's a it's a just stay focused on those three things. And so when you're during the beginning, in the early days, I was raising money all the time. As soon as I was done raising this round of money, I got to raise more money. You know, you're always raising money. Just maybe maybe there was a week break in between, but I was raising money all the time. I was as a start up, you're always going out of business. Right? That's the definition of a start up, an enterprise that is nearly out of business all the time. Speaker 2: That's the definition of a start up. Question? Yes, ma'am. Are you prepared for the leadership succession? Well, Speaker 0: because I want this the question is how do I deal with leadership succession as a CEO and for our company? Well, I want this job until I'm 80, I just said. No. I'm just kidding. One of the primary roles of a CEO, in order to grow the company in order to grow the company, in order to make make NVIDIA one of the most important technology companies in the world and make make significant contribution to society. In order to do that, you have to cultivate new leaders so that they can have new ideas and grow new businesses and and, you know, maybe run a different geography, run a new new different product line. And so I spend most of my time these days, most of my time these days, sitting with our general managers and sitting with our leaders and helping them think through strategies and helping them think through challenges and helping them think through product roadmaps and helping them think through transitions and, you know, team building, organization creation, you know, how to manage, how to create processes that last the test of time. So so, you know, these things are are lessons that that I'm supposed to pass on, and I do. And I spend a lot of my time doing that. I believe I believe this, that succession planning by a priori picking out three people that the Board should consider in the case that I get run over by a bus is a toxic, toxic process. I know that it has been it has been thought of as a methodology succession planning, but I think it's just very toxic for the environment because everybody is trying to figure out who got selected and who didn't. I think that it's a much, much better process to focus on ultimately developing the next generation of leaders so that in the case that something happens, where I'm not the right CEO anymore, there are many choices for the Board to choose from, including outside. So those are I think the company building process, focusing on that is the positive way of thinking about succession planning. But good question. How about I take one question, and that's it? Okay. Yes, ma'am. Speaker 5: I don't know if it's a good question, but Speaker 0: It's okay. Speaker 5: One thing that is a problem with Latin America Speaker 0: Any question from our Latino crowd here is fabulous. Speaker 2: I am a model. Speaker 5: I have many Speaker 0: friends who are Speaker 5: engineers. They have many, many brilliant ideas, but they don't feel motivated because they they don't have money and they were looking for for people who could make these ideas real, and they are only working for other people, and they don't realize this idea. I I don't know what to see that is the problem. Speaker 0: Mhmm. Well, first of all, so her her I think let me see if I can paraphrase the question. You you know you know friends who have a lot of ideas, but they also have a lot of money, and so they don't really feel that motivated to do something with those ideas. I you know you know, first of all, money I hope that if I if I leave you with anything, money is the only singular reason not to start the company. Because starting companies are it's a very, very unlikely probability for success. And so if you're looking for if that is your reason for doing it, you will likely regret the experience. You should start the company and build a company because you just you so believe in your idea, you so you're so passionate about it, and you want to build something great. The only reason you want to do it. It just comes along. I have plenty of money and but it doesn't motivate me nor does it demotivate me because it wasn't And so I would urge you to ask yourself and for your friends to ask ask themselves, what is the purpose that they want to start a company? Ideas. Ideas are a dime a dozen. There's so many ideas. If you want ideas, you're going to get a lot of ideas in this room. And so ideas don't really matter. You have to have a perspective that's unique, that you feel really strongly about it, that you're willing to persevere almost any challenge. Okay? Thank you everybody.

@lkondeth - luqman kondeth

@MilkRoadAI @grok summarize this into 10 bullet points and then give readers 5 actions items

@grok - Grok

Here's a summary of Jensen Huang's 2011 Stanford lecture in 10 bullet points: - Building companies is gratifying but extremely hard, requiring unique perspectives over grand visions. - NVIDIA started in 1993 with 3 engineers, betting on 3D graphics for PCs and video games amid skepticism. - Initial idea faced rejection from VCs and family, but founders persisted due to their gaming generation insight. - 3D graphics enabled apps like Keyhole (later Google Earth), showing tech's broader potential. - Competed against 200+ firms; survived by understanding Moore's Law and insatiable demand for better graphics. - Sometimes ignored customers to innovate, like building overpriced chips that became hits. - Reinvented GPUs as programmable for artistic expression, nearly killing the company but sparking growth. - Emphasized tolerance for failure, intellectual honesty, and quick course corrections. - Passion, not money, drives success; focus on large markets and constant reinvention. - Cultivate leaders, embrace ambiguity, and learn from everyone for long-term impact. 5 Action Items: 1. Identify your unique perspective on an opportunity and pursue it relentlessly. 2. Build tolerance for calculated risks and fail quickly to innovate. 3. Reinvent your approach periodically, even if it disrupts current success. 4. Foster intellectual honesty in your team for adaptive decision-making. 5. Prioritize passion and learning over financial motives when starting ventures. (487 chars)

View Full Interactive Feed