Get Started FreePricing

Dario Amodei CEO of Anthropic: Claude, New models, AI safety and Economic impact

2024-06-26 01:11:02

The CEO of the largest single investor in the world, Norges Bank Investment Management, interviews leaders of some of the largest companies in the world. You will get to know the leader, their strategy, leadership principles, and much more.<br /><hr><p style='color:grey; font-size:0.75em;'> Hosted on Acast. See <a style='color:grey;' target='_blank' rel='noopener noreferrer' href='https://acast.com/privacy'>acast.com/privacy</a> for more information.</p>

2
Speaker 2
[00:00:00.82 - 00:00:19.56]

Hi, everybody. Welcome to In Good Company. Today, really exciting. We have Dario Amodei, the CEO and co-founder of Entropic, Visiting. Now, Dario, he's a superstar in the AI world and, together with his team, has developed the Claude language model, one of the best out there, and they are backed by Amazon and Google.

[00:00:20.28 - 00:00:32.96]

Now, you are a leading figure, Dario, on AI safety and ethics, and you even interrupted your holiday to come here and to talk to us. So big thanks for coming. Thank you for having me.

1
Speaker 1
[00:00:42.00 - 00:00:59.36]

Now, what are the latest breakthrough in AI? Yes. So a few things I could talk about. One is, I think, the scaling trends of AI are continuing. So I think we're going to see, over the next year, much bigger and more powerful models that are able to do greater tasks.

[00:00:59.48 - 00:01:29.98]

In fact, by the time this podcast airs, a new model will be out from Entropic that will probably be the most intelligent and powerful model in the world. But one area I'm particularly excited about that we're developing in parallel with that is interpretability of models, the ability to see inside our AI models and see why they make the decisions they make. That area has been mainly a research area for the last few years and it's just at the beginning of starting to have practical applications. So that's one area I'm very excited about. Why is that so important?

[00:01:30.78 - 00:02:01.00]

So, if you look at what AI models do today, often you won't understand why an AI model does what it does. I was just talking to someone at lunch. Let's say you want to consider your industry. Let's say, you want an AI model to be trained on some data, to be able to predict what happened with a particular set of financial data. One problem you have with training a model to work on.

[00:02:01.00 - 00:02:19.18]

that is that if you trained it on data in the past, the model might have memorized it because it was trained on, it basically knows what happens. It knows the future. in that case. Interpretability might allow you to tell the difference. Is the model deducing the answer to the question or is it memorizing the answer to the question?

[00:02:19.90 - 00:02:48.70]

Similarly, if a model acts in a way that, say, shows prejudice against a particular group or appears to do so, can we look at the reasoning of the model? Is it really being driven by prejudice? There are also a number of legal requirements, right? In the EU, there's a right explanation. And so interpretability, being able to see inside the model, could help us to understand why the models do and say the things that they do and say, and even to intervene in them and change what they do and say.

2
Speaker 2
[00:02:48.84 - 00:02:54.66]

So, a while back, you stated that we still don't know how the advanced AI models work. Does this mean that this will solve this problem?

1
Speaker 1
[00:02:55.34 - 00:03:02.60]

You know, I wouldn't say solve. I would say we're at the beginning. Maybe we now like understand 3% of how they work.

[00:03:04.48 - 00:03:50.86]

We're at the level where we can look inside the model and we can find features inside it that correspond to very complex concepts, like one feature might represent the concept of hedging or hesitating, the particular genre of music, a particular type of metaphorical situation that a character could be in, or the idea of, you know, again, prejudice for or against various groups. So we have all of these features, but we think we've only found a small fraction of what there is. And what we still don't understand is we don't understand how all of these things interact to give us the behaviors we see from models every day. So, you know, it's a little like the brain, right? We can do brain scans.

[00:03:51.00 - 00:04:01.00]

We can say a little about the human brain, but, you know, we don't have a spec sheet for it. We can't go and say, well, this is why that person did exactly what they did.

2
Speaker 2
[00:04:01.28 - 00:04:03.20]

Will we ever understand fully how they work?

1
Speaker 1
[00:04:03.20 - 00:04:11.42]

I don't know about, you know, fully down to the last detail, but I think progress is happening fast and I'm optimistic about getting.

2
Speaker 2
[00:04:11.42 - 00:04:15.28]

But is progress happening faster than complexity of the new models?

1
Speaker 1
[00:04:15.58 - 00:04:42.96]

That is a great question. And that is the thing we're contending with. So we are putting a lot of resources behind interpretability of language models to try and keep pace with the rate at which the complexity of the model is increasing. I think this is one of the biggest challenges in the field. The field is moving so fast, including by our own efforts, that we want to make sure that our understanding keeps pace with our abilities, our capabilities to produce powerful models.

2
Speaker 2
[00:04:43.16 - 00:04:46.78]

What's so good about your model? So this is a CLAWD model, right?

1
Speaker 1
[00:04:46.80 - 00:05:06.58]

This is the CLAWD models. Yeah. So, to give some context, we recently released a set of CLAWD 3 models. They're called Opus, Sonnet, and Haiku. They're different trade-offs between power and intelligence, and speed and low cost while still being intelligent.

[00:05:07.74 - 00:05:30.10]

At the time that Opus was released, it was actually the best all-around model in the world. But I think one thing that particularly made it good is that we put a lot of engineering into its character. And we recently put out a post about how do we design CLAWD's character? People have generally found the CLAWD models are warmer, more human. They enjoy interacting with them more.

[00:05:30.42 - 00:05:43.80]

Some of the other models sound more robotic, more uninspired. We're continuing to innovate quickly. And, as I said, by the time this podcast comes out, we'll probably have at least part of a new generation of models out.

2
Speaker 2
[00:05:44.14 - 00:05:44.78]

Tell me about the new one.

1
Speaker 1
[00:05:45.96 - 00:06:08.46]

So I can't say too much about it, but if I had to say a bit, I would say that we're pushing the frontier. Right now, there's a trade-off between speed and low cost of models and qualities. You can imagine that as a trade-off curve, a frontier. There's going to be a new generation of models that pushes that frontier outward.

[00:06:10.02 - 00:06:35.42]

And so you're going to see, by the time this podcast is out, we'll have a name for it, for at least some of those models. And we'll see that things that you needed the most powerful model to be able to do, you'll be able to do with some of the mid-tier or low-tier models that are faster, cheaper, and even more capable than the past generation, the previous model.

2
Speaker 2
[00:06:35.76 - 00:06:41.82]

So, Daria, what's going to be the wow factor here? So, when I get this model, what is it going to do to me?

1
Speaker 1
[00:06:41.82 - 00:07:09.30]

Yeah. You're going to see models that are better at, say, things like code, math, better at reasoning. One of my favorite is biology and medicine. That's one of the sets of applications I'm most excited about for the new models. So the models we have today, they're kind of like early undergrads in their knowledge of many things, or like interns.

[00:07:09.72 - 00:07:36.64]

And I think we're starting to push that boundary towards advanced undergrads or even graduate level knowledge. And so when we think of use of models for drug development or, in your own industry, use of the models for thinking about investing or even trading, I think the models are just going to get a good deal more sophisticated at those tasks. And we're hoping that every few months we can release a new model that pushes those boundaries further and further.

2
Speaker 2
[00:07:37.40 - 00:07:50.66]

Now, one of the things which has accelerated lately is just how we kind of weave AI into everything we do. And with recently the announcement from Apple and OpenAI,

[00:07:52.34 - 00:07:53.86]

how do you look at this?

1
Speaker 1
[00:07:54.72 - 00:08:25.02]

Yeah. Yeah. So Anthropic thinks of itself more as providing services to enterprises than it does on the consumer side. And so we're thinking a lot about how to integrate AI in work settings. So if you think about today's models, today's chatbots, it's a bit like if I use them in an enterprise setting, it's like if I took some random person on the street who is very smart, but who knew nothing about your company, and I brought them in and I asked them for advice.

[00:08:25.46 - 00:08:55.04]

What I'd really like is someone that acts more like an AI model, that acts more like someone that's been trained with knowledge of your company for many years. And so we're working on connecting our AI models to knowledge databases, having them site work, having them be able to use internal enterprise tools and really integrate with the enterprise as sort of a virtual assistant to an employee. So that's one way I think about really driving the integration.

2
Speaker 2
[00:08:56.46 - 00:09:00.42]

So, if you look at the long-term goal of Anthropic, what is the long-term goal?

1
Speaker 1
[00:09:00.96 - 00:09:02.96]

Yeah. You know, if I think about our long-term goal...

2
Speaker 2
[00:09:02.96 - 00:09:04.18]

Because you're only like three years old, right?

1
Speaker 1
[00:09:04.26 - 00:09:18.02]

Yeah. We're only three and a half years old. Yeah. We're by far the newest player in the space that's been able to build models on the frontier. You know, we're a public benefit corporation, and I think our long-term goal is to make sure all of this goes well.

[00:09:19.12 - 00:09:58.56]

And that's being done, obviously, through the vehicle of a company. But if you think about our long-term strategy, what we're really trying to do is create what we call a race to the top. So, you know, race to the bottom is this well-known thing where everyone, you know, fights to cut corners because the market competition is so intense. We think that there's a way to have the reverse effect, which is that if you're able to produce higher standards, innovate in ways that make the technology more ethical, then others will follow suit. They'll either be inspired by it, or they'll be kind of, you know, bullied into it by their own employees or public sentiment, or ultimately, the law will go in that direction.

[00:09:58.70 - 00:10:25.30]

And so we're hoping to kind of provide an example of how to do AI right and pull the rest of the industry along with us. That's a lot of the work behind our interpretability work, behind our safety work, behind how we think about responsible scaling. We have something called a responsible scaling policy. So I think our overall goal is kind of to try and help the whole industry be better. So you kind of pitch yourself as a good guys?

[00:10:25.88 - 00:10:41.18]

I mean, you know, I wouldn't say anything that grandiose, right? It's more like I want to, you know, I think more in terms of incentives and structures, more than I think of good guys and bad guys. I want to help change the incentives so that everyone can be the good.

2
Speaker 2
[00:10:41.18 - 00:10:55.08]

guys. Yeah. Do you think we will care who, which model we interact with, or are we going to have, just like one agent who picks the model, which is the best for, you know, for that purpose? That was kind of what Bill Gates said when he was on the podcast. Yeah.

[00:10:55.08 - 00:10:55.44]

You.

1
Speaker 1
[00:10:55.44 - 00:11:12.36]

know, I think, I think it really depends on the setting. A few points on this. One is I think, we are increasingly going in the direction where models are good at different things. So, for example, I was just talking about Claude's character, right? Claude is more warm and friendly to interact with.

[00:11:12.62 - 00:11:30.58]

For a lot of applications and use cases, that's very desirable. For other applications and use cases, a model which focuses on different things might be helpful. Some people are going in the direction of agents. Some people are going in the direction of models that are good as code. Claude, for example, another thing it's good at is creative writing.

[00:11:31.08 - 00:11:53.90]

And so I think we're going to have an ecosystem where people use different models for different purposes. Now, in practice, does that mean you're kind of, there's something that's choosing models for you? I think in some consumer context, that will be the case. I think in other contexts, someone will say, oh yeah, no, you know, the, the job I'm doing or the kind of person that I am, I want to use this.

2
Speaker 2
[00:11:53.90 - 00:12:03.22]

particular model all the time. Well, what makes a warm model? I mean, how can you make a model friendly? Is it more humoristic or more polite? Or just like put in some red?

1
Speaker 1
[00:12:03.22 - 00:12:29.22]

hearts in between? And we actually try and avoid too many emojis because it gets annoying. But you know, if you, I don't know, if you go on Twitter and see some of the comments when people, when people interact with Claude, it, you know, it, it, it, it just kind of, I don't know how to describe it. It just kind of sounds more like a human, right? I think a lot of these bots have certain ticks, right?

[00:12:29.22 - 00:12:44.46]

Like, you know, models, will, you know, there are certain phrases, I apologize, but as an AI language model, I can't do X, Y, and Z. Right. That's kind of like a common phrase. And we've helped the model to, to vary their thinking more, to sound more like a human, that, that kind of thing.

2
Speaker 2
[00:12:46.78 - 00:12:54.50]

Now, when you launch new models, you got a pretty, pretty good predictions on how accurate it will be, right? It's a function of number of parameters and so on.

[00:12:56.76 - 00:13:05.36]

Now to get to AGI, what, how far out are we there? This is the general intelligence. So, more, more intelligence.

1
Speaker 1
[00:13:05.36 - 00:13:36.16]

So I've said this, I've said this a few times, but you know, back in 10 years ago, when all of this was kind of science fiction, I used to talk about AGI a lot. I now have a different perspective where I don't think of it as one point in time. I just think we're on this smooth exponential, the models are getting better and better over time. There's no one point where it's like, oh, the models weren't generally intelligent and now, and now they are. I just think, you know, like, like a human child, learning and developing, they're getting better and better, smarter and smarter, more and more knowledgeable.

[00:13:36.32 - 00:14:20.28]

And I don't think there'll be any single point of note, but I think there's a phenomenon happening where, over time, these models are getting, you know, better, better and better than even the best humans. Um, I do think that if we continue to increase the scale, the amount of funding for the models, if it goes to, say, 10 billion, so now, so now a model would cost 100 million. right now, a hundred million. There are models in training today that are more like a billion. Um, I think, if we go to 10 or a hundred billion, and I think that will happen in 2025, 2026, maybe 2027, um, and the algorithmic improvements continue a pace and the chip improvements continue a pace.

[00:14:20.36 - 00:14:30.54]

Then I think there, there is in my mind, a good chance that by that time we'll be able to get models that are better than most humans at most things. So

[00:14:32.28 - 00:14:46.96]

10 billion, you think a model will be next year? Um, I think that the training of, of order, $10 billion models. Yeah. It could start sometime in 2025.. I'm not, not many people can participate in that race.

[00:14:47.16 - 00:15:13.62]

No, no, no, no. And you know, I, you know, of course, I think there's going to be a vibrant downstream ecosystem and there's going to be an ecosystem for small models, but you don't have money. You don't have that much money. Uh, I mean, we have of order, of order, that, um, we've raised, uh, I believe it's a little over eight, 8 billion to date. Um, so, generally, generally of order, that, and, uh, you know, of course, we're, we're, you know, we're, we're, we're always, uh, we're always interested.

2
Speaker 2
[00:15:13.62 - 00:15:29.80]

in getting to the next level of scale. Yeah. Um, now this is of course also a function of, of the chips. Um, and we just learned that NVIDIA is halving the time now between launches, right? So in the past, every other year now is more like every year.

[00:15:29.86 - 00:15:30.18]

So what,

1
Speaker 1
[00:15:30.26 - 00:15:44.70]

what are the implications of this? Yeah. Um, you know, I think that is a, you know, I can't speak for NVIDIA, but I think that is a natural consequence of the recognition that, uh, chips are going to be super important. Right. And also facing competition.

[00:15:45.24 - 00:16:07.10]

Um, Google is building their own, you know, chips, as we know, Amazon is building their own chips. Uh, you know, Anthropic is collaborating with, with both to, uh, to, uh, you know, to work with those chips. And, you know, without getting specific, what I can say is that the, you know, the chip industry is getting very competitive and there are some very strong offerings from.

2
Speaker 2
[00:16:07.10 - 00:16:10.66]

a large number of players. How far behind, how far behind, with Google and Amazon being.

1
Speaker 1
[00:16:10.66 - 00:16:14.36]

in the chip development? I, you know, that's not something I could say, and it's not, it's.

2
Speaker 2
[00:16:14.36 - 00:16:18.52]

not one dimensional, but just some kind of indication. Yeah. I mean, you know what, what,

1
Speaker 1
[00:16:18.58 - 00:16:27.46]

again, you know, I would just, I would just repeat that. I think there are now strong offerings from, from multiple players that have been useful to us and will be useful.

2
Speaker 2
[00:16:27.46 - 00:16:31.36]

to us in different ways. Okay. So it's not only about NVIDIA anymore, as I was saying.

1
Speaker 1
[00:16:31.38 - 00:16:43.60]

I don't think it's only about NVIDIA anymore. Um, but you know, of course, you look at their stock value, you look at their stock price, which, you know, certainly, certainly aware of, and, you know, it's, it's, it's an indicator, I think both about them and the industry.

2
Speaker 2
[00:16:44.06 - 00:17:00.00]

Yeah. You mentioned that you were more on the enterprise side and not necessarily on the consumer side, but, um, just lately there has been more talk about, uh, having chips in, in phones and we talk about, uh, AI, PC and so on. What, um, how do you look at this?

1
Speaker 1
[00:17:00.44 - 00:17:48.64]

Yeah, no, I think that's going to be an important development. And again, if we go back to the curve, I talked about, right, the trade-off curve between powerful, smart, but, you know, relatively expensive and slow models and models that are super cheap, super fast, but very smart, for how fast, fast and cheap they are. As that curves shifts outward, we are going to have models that are very fast and cheap, that are smarter than the best models of today, even though the best models then will be smarter than that. Um, and, and I think, uh, we'll be able to put those models on phones and on mobile chips, and, uh, you know, they'll, they'll pass some threshold where the things that you need to call to a cloud or the server for today, you can do there. And so I'm, I'm very excited about the implications of that.

[00:17:49.10 - 00:17:57.04]

Um, you know, I'm, of course, even more excited about pushing the frontier of where things will go, but as this curve shifts outward, an implication is that both things will happen.

2
Speaker 2
[00:17:57.32 - 00:18:07.64]

We have from Mistral, the French, uh, competitor, that they have developed some really efficient, uh, kind of low costs or lower cost models. What, what are you, how do you view that?

1
Speaker 1
[00:18:07.64 - 00:18:33.58]

Yeah. I mean, I, you know, I can't, I can't comment on, uh, you know, what's going on in other companies, but I think we are seeing this, this kind of general moving of the curve. And so it is definitely true. We're seeing efficient, low cost models, but I think of it less as, like things are leveling out, you know, costs are going down and more as the curve is shifting, right. We can do more with less, but we can also do even more with more resources.

[00:18:33.86 - 00:18:35.94]

Yeah. So I think both, both trends coexist.

2
Speaker 2
[00:18:41.66 - 00:18:45.46]

Dario, changing tack a bit here, uh, your background, you kicked off in physics.

1
Speaker 1
[00:18:46.04 - 00:18:51.84]

Yes. I was an undergrad in physics and then, uh, uh, did grad school in neuroscience.

2
Speaker 2
[00:18:52.22 - 00:18:53.98]

Yeah. So how come you ended up in AI?

1
Speaker 1
[00:18:54.52 - 00:19:20.54]

Yeah. So, you know, when I finished my physics degree, I, you know, I wanted to do something that, you know, would, would have an impact on, you know, humanity in the world. Um, and I felt that, you know, an important component of that would be understanding intelligence, you know, that, that that's one of the things that's obviously shaped our world. And that was back in the mid two thousands. And in those days, I wasn't particularly, to be honest, that excited about the AI of the day.

[00:19:21.10 - 00:19:51.36]

Uh, and so I felt like the best way to study intelligence in those days was to study the human brain. So I went into neuroscience for grad school, computational neuroscience that use some of my physics background and, and, you know, study kind of collective properties of neurons. But by the end of that, by, you know, by the end of grad school, after that, I did a short postdoc, um, AI was really starting to work. We really saw, you know, the deep learning revolution. I saw the work of, you know, Ilya Sutskever back, back then.

[00:19:51.96 - 00:20:05.66]

Uh, and so I decided, based on that, to go into the AI field. Uh, and I worked, you know, different places. I was at Baidu for a bit. I was at Google for a year. I worked at open AI for five years and, and, and, you know, you were

2
Speaker 2
[00:20:05.66 - 00:20:10.56]

instrumental in developing Chachapiti. two and three, right? Uh, yes, yes. I led the.

1
Speaker 1
[00:20:10.56 - 00:20:39.98]

development of both of those. Why did you leave? Um, you know, we had reached around the end of 2020.. Um, we had kind of reached a point, the set of us who worked in this, these, on these projects, in these areas, where we kind of had our own vision for how to do things. Uh, so, uh, you know, again, you know, we had this picture that I think I've already kind of implicitly laid out of one, you know, real belief in this scaling hypothesis, and two, in the importance of safety and interpretability.

[00:20:40.16 - 00:20:56.08]

So it was a safety side, which made you leave? I mean, I think, you know, we just, we just had our own vision of things. There were a set of us who were co-founders, who really felt like we were on the same page, really felt like we trusted each other, really felt like, you know, we just wanted to do something.

2
Speaker 2
[00:20:56.08 - 00:21:03.18]

together. Right. But you were a bit more AI doomsday before than you are now. Um, I, you.

1
Speaker 1
[00:21:03.18 - 00:21:32.10]

know, I wouldn't say that my view has always been that, uh, there, there are important risks and there are benefits, and that is, the technology goes on. It's exponential. The risks become greater and the benefits become greater. Um, and so we are, you know, including that anthropic, very interested in these questions of catastrophic risk, right? We have this thing called responsible scaling policy, and that's basically about measuring models at each step for catastrophic risk.

[00:21:32.30 - 00:21:59.44]

What, what is catastrophic risk? So this would be, um, I would put it in two categories. Um, one is misuse of the models, which could include things in the realm of biology or cyber or kind of, um, you know, election operations at scale, um, things that are really disruptive to society. Um, so that misuse should be one bucket. And then the other bucket would be autonomous, unintended behavior of the model.

[00:21:59.58 - 00:22:09.10]

So, you know, today it might be just, you know, the model doing something unexpected, but increasingly, as models act in the world, we have to worry about them behaving in ways.

2
Speaker 2
[00:22:09.10 - 00:22:17.48]

that you wouldn't expect. And what was it that you saw exactly with the chachapiti? Well, three, I guess, then, which made you particularly concerned about this?

1
Speaker 1
[00:22:17.66 - 00:23:10.52]

Yeah, it wasn't about any particular model. Um, you know, if we go all the way back to 2016,, you know, before I even worked at open AI, when I was at Google, I wrote a paper called with some colleagues, some of whom are now anthropic co-founders, concrete problems in AI safety, um, and concrete problems in AI safety laid out this concern that, you know, we have these powerful AI models, neural nets, but they're fundamentally statistical systems. And so that's going to create all these problems about predictability and uncertainty. And if you combine that with the scaling hypothesis, and I really came to believe in the scaling hypothesis as I worked on GPT two and GPT three, those two things together told me, okay, we're going to have something powerful and it's not going to be trivial to control it. And so we put those two things together.

[00:23:10.52 - 00:23:14.62]

and, and that makes me think, oh, this is an important problem that we have to solve.

2
Speaker 2
[00:23:15.16 - 00:23:18.96]

Hmm. How do you solve, um, the two catastrophic risk problems?

1
Speaker 1
[00:23:19.26 - 00:23:43.96]

Yes. Yes. Um, so, uh, one of the biggest tools for this is our RSP, or responsible scaling policy. And so the way that works is every time we have a new model that you, you know, that, that represents a significant leap, a certain amount of compute above an old model, we measure it for both the misuse risks and the autonomous self-replication risks.

2
Speaker 2
[00:23:44.10 - 00:23:44.66]

And how do you do that?

1
Speaker 1
[00:23:45.24 - 00:24:16.78]

Um, so we have a set of evaluations that we run. Um, we've in fact worked with, for the misuse risks folks in the national security community. So, for example, we've worked with this company called Griffin Biosciences that, um, contracts with the U S government that does biosecurity work and they're, they're the experts on responding to biological risk. And so they say, what is the stuff that's not on the internet that if the model knew it would be concerning, and they run their tests and, you know, give them access to the new model, they run their tests. And every time.

[00:24:16.78 - 00:24:23.60]

so far, they've said, well, it's better at the task than it was before, but you know, it's not yet at the level where it's a serious concern.

2
Speaker 2
[00:24:23.88 - 00:24:32.58]

So a misuse test would be, for instance, if I put in, Hey, can you come up with a virus which is just going to wipe out the earth? The people is that that's, that's an example, right?

1
Speaker 1
[00:24:32.58 - 00:24:57.10]

Conceptually. Yes. Although it's less about answering one question, it's more about, can the model go through a whole workflow? Like, could some, could some bad actor over the period of weeks, use this model to, to deuce as they were doing something nefarious in the real world, could the model give them hints on how to help them? Could the model help them through the task over a long period of time?

2
Speaker 2
[00:24:57.10 - 00:25:03.52]

Okay. So what you're saying is that, uh, the AI models so far cannot, uh, do this.

1
Speaker 1
[00:25:03.84 - 00:25:12.42]

They, they, they know individual, isolated things, which are concerning. Right. Um, and they get better at it every time we release a new model, but they haven't reached this.

2
Speaker 2
[00:25:12.42 - 00:25:17.98]

point yet. Okay. And, and my guess is, what about the other one, the autonomous? Yeah. So how far are we away from that?

1
Speaker 1
[00:25:18.18 - 00:25:50.46]

We, we test the models there for things like ability to train their own models, ability to provision cloud compute accounts and, you know, take actions on those accounts, ability to, like, you know, simultaneously, you know, sign up for accounts and engage in financial transactions. Just some of the measures of things that would kind of unbind the model and enable them to take actions. How far are we away from that? Um, I think it's kind of the same story as with misuse. They're getting better and better at individual pieces of the task.

[00:25:50.82 - 00:26:17.58]

There's a clear trend towards ability to do that, but we're not there yet. Um, I, I again point to the 2025, 2026, maybe 2027 window, just as I think a lot of the, the extreme positive economic applications of AI are, are going to arrive sometime around. then. Um, I think some of the negative concerns may, may start to arise then as well, but you know, I, I'm not a crystal ball. I'm sorry, 25, 26.

[00:26:18.52 - 00:26:19.32]

Around that, you.

2
Speaker 2
[00:26:19.32 - 00:26:22.70]

know, I mean, so what do you, so? what do you, what do you do then? Do you build in like?

1
Speaker 1
[00:26:22.70 - 00:26:44.78]

a kill switch or what do you? Yeah, well, I mean, there's a number of things. Um, I think, on the autonomous behavior, a lot of our work on interpretability, a lot of our work on, um, you know, we haven't discussed constitutional AI, but that's another way. we provide kind of values and principles for the AI system. on the autonomous risk.

[00:26:44.92 - 00:27:09.88]

What we really want to do is understand what's going on inside the model and make sure that we design it and can iterate on it so that it doesn't do these dangerous things we don't want it to do. Um, on misuse risk, again, it's, it's there, it's more about putting safeguards into the model so that people can't ask it to do dangerous things and we can monitor when people try to use it to do dangerous things. So how can, so, so, Jen, so generally.

2
Speaker 2
[00:27:09.88 - 00:27:13.42]

speaking, I mean, there's been a lot of talk about this, but how can one regulate AI? Can?

1
Speaker 1
[00:27:13.42 - 00:27:30.58]

companies self-regulate? Yeah. So, um, you know, one way I think about it is, uh, the RSP, the responsible scaling policy that I was describing, is maybe the beginning of a process, right? That represents voluntary self-regulation. And, you know, I mentioned this concept of race to the top.

[00:27:30.58 - 00:28:13.54]

last September, we put in place our RSP. since then, other companies like Google, open AI, have put in place, similar frameworks, they've given them different names, but they operate in roughly the same way. And now we've heard, uh, you know, Amazon, Microsoft, even meta reportedly public reporting are, are at least considering similar frameworks. And so I would like it if that process continues, right? Where we have some time for companies to experiment with different ways of voluntarily self-regulating, some kind of consensus emerges from some mixture of public pressure, experimentation with what, what is unnecessary versus what is really needed.

[00:28:14.30 - 00:28:49.26]

Um, and then I, I would imagine, the real way for things to go is once there's some consensus, once there's industry best practices, probably the role for legislation is to, is to look in and say, Hey, there's this thing that 80% of the companies are already doing. That's a consensus for how to make it safe. The job of legislation is just to enforce, force those 20% who aren't doing it forced that companies are telling the truth about what they're doing. I don't think regulation is good at coming up with a bunch of new concepts. So how do you view the EU AI act?

[00:28:49.64 - 00:29:07.96]

Yeah. So the, you know, the EU AI act, I should say, first of all, and the California safety bill as well. Yeah. Yeah. You know, I, I should say, I should say, the EU AI act, you know, kind of, it's still being, it's still being, um, you know, the, the kind of details of it are still be, even though the act was passed, the, you know, many details are still being worked out.

[00:29:08.26 - 00:29:40.78]

So, you know, I think a lot of this, a lot of this depends on the details. Um, you know, the, the, the, the, the, the California bill, you know, I would say that, um, it is very, you know, it has some structures in it that are very much like kind of the RSP. Uh, and, and so, you know, I think something that resembles that structure at, at some point could be a good thing. Um, do you think, if I, if I have a concern, though, I think it's that we're very early in the process, right? I described this process.

[00:29:40.78 - 00:29:57.98]

That's like, you know, first, first one company has an RSP, then many have RSPs, then these, these kind of industry consensus comes into place. My only question would be, are we, are we too early in that process? Too early in regulation? to, yeah. Or, or, you know, that, that?

[00:29:57.98 - 00:30:00.80]

maybe regulation should be the last step of a series.

2
Speaker 2
[00:30:00.80 - 00:30:06.60]

of steps. Yeah. And so what's the danger of regulating too early? Um, I don't know. One,

1
Speaker 1
[00:30:06.82 - 00:30:37.30]

one, um, one thing I could say is that I'll look at our own experience with RSPs. Um, so if I look at what we've done with RSP, you know, we wrote an RSP in September. Um, and you know, since then we've deployed one model, we're soon going to deploy another. You, you see so many things that not that it was too strict or not strict enough, but you just didn't anticipate them in the RSP, right? Like, you know, there are various kinds of like AB tests.

[00:30:37.30 - 00:30:55.28]

you can run on your models that are even informative about safety. And our RSP didn't speak one way or another about like when those are okay and when there's not. And so we're updating our RSP to say, Hey, how should we handle this issue? We've never, we've never even thought of. And so I think in the early days, that flexibility is easy.

[00:30:55.56 - 00:31:23.76]

If you don't have that flexibility, if your RSP was written by a third party, um, and you didn't have the ability to change it in the process for changing, it was very complicated. I think it, it could create, it could create a version of the RSP that doesn't protect against the risks, but also is very onerous. And then people could say, Oh man, all this regulation stuff, all this catastrophic, it's all nonsense. It's all a pain. So I I'm not, I'm not against it.

[00:31:23.80 - 00:31:25.62]

You just have to do it delicately and in the.

2
Speaker 2
[00:31:25.62 - 00:31:39.94]

right order. But we build AI into, you know, into the race between the superpowers, right? We're building into the weapons, the cars, the medical research, into everything. How can you, how can you regulate when it's part of the power balance in the world? Yeah.

[00:31:40.06 - 00:31:40.28]

Yeah.

1
Speaker 1
[00:31:40.34 - 00:31:59.32]

So I think there's different, there's different questions, right? One question is, uh, you know, how do you regulate the use domestically? And you know, there, I think there's, there's a history of it, right? Um, you know, I think an analogy I would make is like, you know, I don't know the way cars and airplanes are regulated, right? I think that's been a reasonable story.

[00:31:59.40 - 00:32:16.28]

I don't know that much about Europe, but like in the U S, I think that's been a reasonable story. Everyone understands there's huge economic value. Everyone understands that these things are dangerous and they can kill people. And you know, everyone understands, yes, you have to do this kind of basic safety testing. Um, and you know, that's, that's evolved over years.

[00:32:16.28 - 00:32:36.40]

I think that's generally, um, you know, gone reasonably well. It hasn't been perfect. Um, so, so I think, for domestic regulation, you know, that's, that's what we should aim to. Things are moving fast, but you know, we should try to go through all the steps to get there. Uh, from, from an international point of view, I mean, you know, I think that's a completely different question.

[00:32:36.40 - 00:33:05.60]

that's less about regulation and more about. there's an international race to the, to the bottom. And how do you, how do you handle, how do you handle that race to the bottom? I mean, I think it's an inherently difficult question because on one hand we don't want to, you know, you know, we, we don't want to just recklessly build as fast as we can, particularly on the weapons side. Uh, you know, on, on the other side, I think, looking, you know, as, as a citizen of the U S, here, I, here I am in Norway, another democracy.

[00:33:06.24 - 00:33:13.86]

Um, I I'm very worried about if autocratic regimes were to lead, lead in this technology. I think that's very, very dangerous.

2
Speaker 2
[00:33:14.28 - 00:33:19.70]

Um, and so how far behind are they now? or are they behind? Uh, I mean, it's hard to.

1
Speaker 1
[00:33:19.70 - 00:33:43.24]

say. I would say that with some of the restrictions that have been put in place, uh, on, for example, you know, shipment of chips and equipment to Russia and China. Um, I think if the U.S. Government plays its, plays, its cards right, then, uh, you know, those countries could be kept behind, I don't know, maybe two or three years. Um, that doesn't give us much.

2
Speaker 2
[00:33:43.24 - 00:33:49.20]

margin. Um, talking about democracies, will AI impact the U? S election? Yes. Uh, you know,

1
Speaker 1
[00:33:49.22 - 00:34:17.22]

I, I, I, I am concerned about that. You know, when Thropic actually just put out a post about what we're doing to, you know, to counter election interference. Um, how could it interfere? Um, so I, you know, if we look back at, say, the 2016 election, something that happened in that election was that there were large numbers of people who were being, you know, who were being, uh, paid to provide content. I don't know how effective that was in the end.

[00:34:17.28 - 00:34:42.74]

It's very hard to measure. Um, but a lot of the things that, you know, you had, you had, you know, farms of, of people being paid to do, could, could now be done by AI. I think it's less that, like, you know, you could make content that people necessarily believe. It's, it's more that you could kind of flood the information ecosystem with a bunch of very low quality content. that would make it hard for people to believe things.

2
Speaker 2
[00:34:42.74 - 00:34:47.32]

that really are true. Did that happen, for instance, in India, in the European election?

1
Speaker 1
[00:34:48.28 - 00:35:11.10]

I mean, is it, is it really happening this year? We don't, we don't have particular evidence of the use of our models. We've banned their use for electioneering and we monitor use of the models. Um, you know, occasionally we shut things down, but I don't think we've ever seen a super large scale operation. I can only speak for use of our models, but I don't, I don't think we've ever seen a super large scale operation there.

2
Speaker 2
[00:35:16.46 - 00:35:27.76]

Changing, um, topic slightly. You mentioned that you thought we were going to see some extreme positive effects of AI in 25, 26.. What are these extremely positive things?

1
Speaker 1
[00:35:27.98 - 00:35:36.30]

Yeah. So again, if we go back to the analogy of like, today's models are like undergraduates, um, if we get to the point where the models are...

2
Speaker 2
[00:35:36.30 - 00:35:38.18]

I suspect you were a better undergrad than me, though.

1
Speaker 1
[00:35:39.36 - 00:36:10.24]

I can kind of feel it. I couldn't speak to it, but, um, uh, but if, uh, you know, let's say those models get to the point where, you know, they're kind of, you know, graduate level or strong professional level, think of biology and drug discovery. Think of, um, a model that is as strong as, you know, a Nobel prize winning scientist. or, you know, the head of the, you know, the head of, head of drug discovery at a major pharmaceutical company. Um, uh, I look at all the things that have been invented.

[00:36:10.24 - 00:36:51.70]

You know, if I look back at biology, you know, CRISPR, the ability to like, edit genes. If I look at, um, you know, CAR T therapies, which have cured certain kinds of cancers, there's probably dozens of discoveries like that lying around. And if we had a million copies of an AI system that are as knowledgeable and as creative about the field as all of those scientists that invented those things, then I think the rate of, of those discoveries could really proliferate. And, you know, some of our really, really longstanding diseases, uh, you know, could be, could be addressed or even cured. Now, I don't think all of that will come to fruition in 2025, 2026..

[00:36:52.20 - 00:37:07.14]

At most, I think that the caliber of AI that's, that's capable of starting the process of addressing all those things could be ready. Then it's another question of like applying it all, putting it through the regulatory system.

2
Speaker 2
[00:37:07.14 - 00:37:09.24]

But what can you do to productivity in society?

1
Speaker 1
[00:37:10.12 - 00:37:30.56]

Yeah. Um, uh, you know, I think of, again, virtual assistants, like, uh, you know, uh, like a chief of staff for everyone, right? I have a chief of staff, but not everyone has a chief of staff. Uh, you know, could, could everyone have a chief of staff who helps them, uh, you know, just deal with every, deal with everything that lands on their desk?

2
Speaker 2
[00:37:30.72 - 00:37:35.34]

So if everybody had that kind of thing, what would you do? Could you put a number on productivity gain?

1
Speaker 1
[00:37:35.96 - 00:38:02.94]

Uh, you know, I'm not an economist. I couldn't tell you, you know, X, I couldn't tell you X percent. Um, but if we look at kind of the exponential, right, if we look at like, you know, revenues for AI companies, like that, seems like they've been growing roughly 10 X a year. And so you could imagine getting, you can imagine getting to the hundreds of billions in, you know, two to three years, and getting even to the trillions, which per year, which no company has reached.

2
Speaker 2
[00:38:02.94 - 00:38:08.16]

But I'm saying that's revenue for the company. Revenue for the company. Right. And then what about productivity? What about productivity?

1
Speaker 1
[00:38:08.68 - 00:38:21.06]

in society? Right. So that depends on how much is this replacing something that was already being done versus doing new things. I think with things like biology, we're probably going to be doing new things. So I don't know.

[00:38:21.06 - 00:38:31.20]

if, let's say, you extend people's, you know, let's say, you know, you extend people's productive ability to work by 10 years, right? That could be, you know, one sixth of the whole economy.

2
Speaker 2
[00:38:31.84 - 00:38:34.30]

Do you think that's a realistic target?

1
Speaker 1
[00:38:34.70 - 00:38:46.24]

I mean, again, like I know some biology, I know something about how the AI models are going to happen. I wouldn't be able to tell you exactly what would happen, but like, I can tell a story where it's possible.

2
Speaker 2
[00:38:46.54 - 00:38:54.76]

So 15%, and when will we, so when could we have added the equivalent of 10 years to our life? I mean, how, what was the timeframe?

1
Speaker 1
[00:38:55.22 - 00:39:23.30]

Again, like, you know, this involves so many unknowns, right? If I, if I try and give an exact number, it's just going to sound like hype. But like a thing I could, I think I could imagine is like, I don't know, like two to three years from now, we have like AI systems that are like capable of making that kind of discovery. Like five years from now, like, you know, those, those discoveries are actually being made. And five years after that, it's all gone through the regulatory apparatus and, and really has.

[00:39:23.38 - 00:39:39.36]

So, you know, we're talking about more, we're talking about, you know, a little over a decade, but really I'm just pulling things out of my hat here. Like, I don't know that much about drug discovery. I don't know that much about biology. And, frankly, although I invented AI scaling, I don't know that much about that either. I can't predict it.

2
Speaker 2
[00:39:40.02 - 00:39:44.08]

I think you know more about these things than, than most of us.

1
Speaker 1
[00:39:45.16 - 00:39:47.82]

And yet it is also hard to predict.

2
Speaker 2
[00:39:48.00 - 00:39:49.82]

Absolutely. Have you thought about what it could do to inflation?

1
Speaker 1
[00:39:51.04 - 00:40:19.12]

Yeah. So again, I'm not, I'm not an economist. If we look at, if we look at inflation, I mean, again, using my limited economic reasoning, I think if we had very large, real productivity gains, that would tend to be deflationary rather than inflationary, right? Like you would be able to do, you'd be able to do more with less, the dollar would go further. So, directionally, at least, that suggests disinflation, but...

2
Speaker 2
[00:40:19.12 - 00:40:21.18]

Totally. But what kind of magnitude?

1
Speaker 1
[00:40:21.38 - 00:40:27.48]

What kind of magnitude? I mean, that you are more of the expert on than I am. Maybe I should ask you to predict that.

2
Speaker 2
[00:40:27.96 - 00:40:33.78]

How do you work with the hyperscalers? Like, you know, some of your shareholders, like Google and Amazon?

1
Speaker 1
[00:40:34.08 - 00:40:36.98]

Yeah. Yeah. So, you know, I think...

2
Speaker 2
[00:40:36.98 - 00:40:40.28]

I'm sorry, just to get it straight. These are called hyperscalers because why?

1
Speaker 1
[00:40:40.96 - 00:40:54.02]

I actually don't know the reason for the name, but, you know, they're, they're, they're hypercap companies in terms of their valuation, but, but also they make very large, you know, very large AI data centers. I assume you refer to the second one, but...

2
Speaker 2
[00:40:54.02 - 00:40:55.06]

Absolutely. How do you work with them?

1
Speaker 1
[00:40:56.84 - 00:41:22.34]

So, you know, I would say that the relationship with these companies makes sense in the sense that we have complementary inputs, right? That they provide the chips in the cloud, and then we provide the model. And then that model is something that, again, can be sold to customers on the cloud. So there's kind of a layered cake where we provide some layers and they provide the other layers. So this partnership, these partnerships, make sense on, on multiple grounds.

[00:41:22.50 - 00:41:49.14]

You know, at the same time, we've always been very careful, right? We have our own kind of value as values as a company, our own way of doing things. And so we try to stay as independent as possible. And one of the things we've done is, of course, we have relationships with multiple of these cloud providers, right? We work with both Google and Amazon, and, and that has allowed us some flexibility in, in, in our ability to make sure that there isn't too much.

[00:41:51.40 - 00:41:57.06]

exclusivity and that we're kind of, you know, free to deploy our models on multiple surfaces.

2
Speaker 2
[00:41:57.54 - 00:42:02.08]

The fact that these companies are becoming so incredibly powerful, what kind of systemic.

1
Speaker 1
[00:42:02.08 - 00:42:26.48]

risk does that pose? Yeah. I mean, you know, I would say that, and this is maybe broader than AI, it maybe relates to just kind of the era that we're, that we're living in. There are certain eras in history where, you know, there's a powerful, there's a powerful technology or there's an economic force that kind of tends to concentrate resources.

[00:42:28.46 - 00:42:54.92]

You know, I think, you know, probably the same thing happened in the 19th century. And so I think it, I think it actually is important to make sure that the benefits are shared by all. So one thing that's often on my mind is there's been, for example, very little penetration of AI and language models in some parts of the developing world, right? In, like Sub-Saharan Africa. And so how do we, how do we bring these models to those areas?

[00:42:55.32 - 00:43:13.38]

How do we even help with challenges in those areas like health or education? So I definitely agree. We're, we're, you know, living in an era of more concentrated wealth, and that's an area of concern and area that, you know, we should, we should, we should, you know, we should do what we can to find countervailing forces.

2
Speaker 2
[00:43:13.42 - 00:43:18.68]

But is it, what's the risk in that these companies are now becoming more powerful?

1
Speaker 1
[00:43:18.68 - 00:43:58.62]

than the countries and governments? Yeah. I mean, you know, this is, this is kind of, you know, what I said on, in terms of, in terms of regulation, like, you know, I think that AI is a very powerful technology and, you know, our governments, our democratic governments, do need to step in and set some basic rules of the road, right? It needs to be done in the right order. It can't be stifling, but I think it, I think it does, it does need to be done because, you know, because, because, as you said, like we're getting to a point where the amount of concentration of power can be, you know, can be greater than that of national economy, you know, national governments.

[00:43:58.84 - 00:44:13.98]

And, you know, we don't want that to happen at the end of the day, you know, you know, all the people of the country and all entities, including companies, that, that work in it, you know, they ultimately have to be accountable to democratic processes, right? There's, there's no other, there's no other way.

2
Speaker 2
[00:44:14.32 - 00:44:18.62]

Will AI increase or decrease the difference between rich and poor countries?

1
Speaker 1
[00:44:19.76 - 00:44:23.24]

That, I think, depends on what we choose to do with it.

2
Speaker 2
[00:44:23.36 - 00:44:28.66]

The way you look, at, the way you look at the path forward just now.

1
Speaker 1
[00:44:29.32 - 00:44:37.74]

Yeah. So, you know, I would say that we are looking for ways for it, not to make, you know. Sure. But is that happening?

[00:44:39.30 - 00:45:32.44]

I would, I mean, it's too early to say with like how, you know, with how the technology is being deployed. I would definitely say I do see something related to it. That's a little worrying to me and that we're trying to counter, which is that if you look at the natural applications of the technology, the, you know, the, the things that customers are, are, you know, the, the most eager customers that come to us. often, I think, because we're a Silicon Valley company, often the most eager customers are other kind of technologically forward Silicon Valley companies that kind of also use the technology. And so I think there's this danger of what you might call like a, a kind of closed loop, where it's like an AI company, you know, supply is a like AI legal company, which supplies an AI productivity company, which supplies a, you know, some other company in Silicon Valley.

[00:45:32.64 - 00:45:33.68]

And, you know, is it all.

2
Speaker 2
[00:45:33.68 - 00:45:38.72]

a closed ecosystem where it's all being used by the most highly educated people?

1
Speaker 1
[00:45:39.26 - 00:45:59.38]

Exactly. And, and so how do we break out of that loop? And so we thought about a number of ways to break out of that loop. One of the reasons I talk about biology and health is that I think biology and health can be used to help us to break out of that loop. You know, innovations in health, assuming we distribute them well, can apply to everyone.

[00:46:00.20 - 00:46:42.88]

I think things like education can help here. Another area that I'm very excited about is use of AI for provision of everyday government services. You know, I don't know what the names of these services are in Norway, you know, in the US, every time you interact with, like the DMV, the IRS, various social services, people almost always have a bad experience and it drives cynicism about the role of government. And I would, I would love it if we can modernize government services that everyone use so that they can actually deliver what people across the world need. I have to say that I think.

2
Speaker 2
[00:46:42.88 - 00:46:48.18]

in this country, we are, we are fortunate in that we are not so many people, and we are, uh, we've got, you know,

1
Speaker 1
[00:46:48.24 - 00:46:57.38]

it's heavily digitalized. You are probably much better than we are at this. I'm reacting to my experience in the United States, which, which, you know, I think could be better.

2
Speaker 2
[00:46:58.74 - 00:47:06.68]

Yeah. So, net net, what do you think? Well, in 10 years time, will the gap between rich and poor be bigger or smaller?

1
Speaker 1
[00:47:07.70 - 00:47:13.00]

I, I, I just have to say, like, if we handle this the right way, I hear what you say, but I mean, what is the right way?

2
Speaker 2
[00:47:13.08 - 00:47:17.62]

We can narrow the gap. What do you think? What do you think will happen? I don't know what I.

1
Speaker 1
[00:47:17.62 - 00:47:26.62]

think will happen. I know that if we are not extremely thoughtful about this, if we're not extremely deliberate about it, then yes, it will increase the gap. Okay.

2
Speaker 2
[00:47:26.62 - 00:47:37.02]

Who will make the most money on, on AI? Will it be the chip manufacturers, or will it be, uh, you guys or the scalers, or, uh, all the consumers or companies?

1
Speaker 1
[00:47:37.32 - 00:48:02.78]

I, I, my boring answer is that I think it's going to be distributed among all of them. And that the pie is going to be so large, um, that in some ways it may not even matter. Like, certainly right now, the, you know, the, the chip companies, are making the most money. I think that's because training of models comes before, deployment of models comes before revenue. So I think the way I think about it is the valuation of the chip companies is a leading indicator.

[00:48:03.62 - 00:48:12.78]

Um, the, um, valuation of the AI companies is maybe a present indicator and the valuation of lots of things downstream is a lagging indicator,

2
Speaker 2
[00:48:12.78 - 00:48:27.80]

but, but the wave is going to reach everyone. So when you look at the, the market cap of, uh, you, that's an indicator. I mean, what? do you multiply that by to find, uh, the potential impact of AI?

1
Speaker 1
[00:48:28.48 - 00:48:33.60]

Yeah. I mean, what. I think that, and you know, obviously I can't stock advice on podcast.

2
Speaker 2
[00:48:33.60 - 00:48:39.34]

about chips. So that's $3 trillion, right? Yeah. But, but $3 trillion. So, so why is that?

[00:48:39.36 - 00:48:47.86]

Which is nearly twice the size of this fund, which is the largest sovereign wealth fund in the world. Yes. Um, you know,

1
Speaker 1
[00:48:47.92 - 00:49:03.14]

if I think about that again, like speaking very abstractly and conceptually, um, what's that? driven by? Probably that's driven by, like anticipated demand. Like people are building very large AI clusters. Those clusters involve lots of revenue for Nvidia.

[00:49:04.24 - 00:49:32.14]

Presumably, companies like us are paying for those clusters because they think that the models they build with them will generate lots of revenue, but that revenue is not present yet. And so what we're seeing so far is just, man, people want to buy a lot of chips. And of course it's possible. It's consistent with the whole picture that all of this will be a bust. The models don't turn out to be, you know, that powerful, like companies like Anthropic and the other companies in the space, don't do as well as we expected, because the models don't keep getting better.

[00:49:32.40 - 00:49:42.56]

That always could happen. That's not my bet. That's not what I think is going to happen. What I think is going to happen is that these models are going to produce a great deal of revenue. And then there's going to be even more demand for chips.

[00:49:43.10 - 00:49:58.66]

Nvidia's value will go up. The AI company's value will go up. All these downstream company, you know, that's the up, that's the bullish scenario that I'm, that I'm, I'm betting on by, by leading this company. Um, but I'm, I'm not sure it could go the other way. I don't think anyone knows.

2
Speaker 2
[00:49:58.66 - 00:50:06.78]

Where is the biggest constraint just now? I mean, is it in, is it in chips, uh, talent, uh, algorithms, electricity?

1
Speaker 1
[00:50:07.36 - 00:50:21.44]

I would say, um, you know, a big bottleneck we're dealing with is data. Um, but, as I've said elsewhere, uh, we, we and other companies are working very hard on synthetic data. Um, and I think that bottleneck is going to be lifted.

2
Speaker 2
[00:50:21.64 - 00:50:25.00]

So data, just to get it straight, that's just information you feed into your models to.

1
Speaker 1
[00:50:25.00 - 00:50:55.88]

Yeah. Based information, that's, that's fed into the models, but we're getting increasingly good at synthesizing the data. Tell me what is synthetic data? Uh, so, synthetic data, um, the example I like to give is, uh, seven years ago, uh, you know, deep mind, uh, as part of Google, uh, produce the AlphaGo model, which was able to beat the world champion in go. And there was no, that version, or there was a version of it called AlphaGo zero that was not trained on any humans playing go.

[00:50:56.18 - 00:51:33.38]

All it did was the model played go against itself for, for, for a long time, basically for forever. Um, uh, and, and so, basically, with just the little tiny rules of go and the models playing against each other, pushing against each other, using that rule, they were able to get better and better to the level where they were better than any human. And so you can think of those models as having been trained on synthetic data that are created by other models with the help of these, this kind of logical structure of the rules of go. And so I think there are things analogous to that, that can be done for language models.

2
Speaker 2
[00:51:35.76 - 00:51:39.86]

How do you think AI will affect geopolitics? Yeah, I, you know,

1
Speaker 1
[00:51:39.88 - 00:52:08.56]

I think that's a big one. Um, uh, my, my view is that, again, if we get to the level of AI systems that are better than the best professionals at a wide range of tasks, um, you know, then tasks like military and intelligence are going to be among those tasks. Uh, and, uh, you know, we shouldn't be naive. Everyone is going to try to deploy those. Um, I think we should try to create cooperation and restraints where we can.

[00:52:09.36 - 00:52:40.14]

Um, but that, uh, you know, in many cases that won't be possible. And when it isn't possible, you know, I'm, I'm on the side of democracies in the free world. Um, I, I want to make sure that the future is democratic, that as much as possible of the world is democratic and that democracies have, um, a lead and an advantage on the world stage. Um, the idea of powerful AI plus autocracies terrifies me and I don't want it to happen.

2
Speaker 2
[00:52:40.36 - 00:52:47.62]

Should each country have its own language model? Yeah. Um, should Norway build a language model, uh, you know,

1
Speaker 1
[00:52:48.26 - 00:53:28.68]

5 million people, it, it, it, it kind of really depends on what you're aiming to do. Right. Um, uh, you know, it, it may make sense, from a national security perspective, for every country to have language models. I think an idea that might work, uh, that, you know, like another direction we could go in, is imagining some kind of democratic coalition or cooperation in which democratic countries, you know, work together to provide for their mutual security, to protect each other, to protect the integrity of their democratic processes. Maybe it makes sense for them all to pull their resources and make a very small number of very large language models.

[00:53:29.12 - 00:53:34.84]

But then there, you know, there may also be value in decentralization. I don't have a strong opinion on which of those is better.

2
Speaker 2
[00:53:35.54 - 00:53:43.54]

Is it a national security issue that US controls AI? Um, you know, uh, should we, should Europe be worried about this?

1
Speaker 1
[00:53:44.30 - 00:54:18.70]

Yeah. I mean, again, you know, I, I, I would go to, you know, each, each country has to kind of worry about its own security, even separately from its allies. Um, you know, I think that's, that's more of a question for, for kind of, for kind of, for kind of individual governments. I mean, you know, I would think of it, probably this is a provocative analogy, but a little like nuclear weapons, right? Some countries, even though their allies feel the need to have their own nuclear weapons, for example, France, um, other countries say, no, we trust that we're being protected by the US and the UK and France.

[00:54:19.04 - 00:54:34.46]

Um, I think it may be somewhat similar with these more powerful models. And I think it's less important how many of them exist within the democratic world, as that the democratic world is in a strong position relative to relative to autocracies.

2
Speaker 2
[00:54:35.67 - 00:54:42.14]

You talk about, um, corporation and partners and so on. Do you guys in, in AI actually like each other? Do, do we,

1
Speaker 1
[00:54:42.26 - 00:55:08.46]

do we in AI actually like each other? I mean, we've done a number of collaborations. Um, so, you know, uh, I think, uh, very early on, when I was at open AI, you know, I drove the original RL from human feedback paper was considered safety work. And this ended up being a collaboration between deep mind and open AI. And we've worked together, um, you know, in organizations like the frontier model forum, uh, to collaborate with each other.

[00:55:08.46 - 00:55:25.02]

that said, I mean, you know, I'll be honest. I don't think every company in this space takes issues of safety and responsibility equally seriously from, from, from all the other companies, but, you know, instead of pointing fingers and saying,

2
Speaker 2
[00:55:25.38 - 00:55:38.04]

is that the kind of things that make you not being so keen on other companies? Is it their view on safety and security? I mean, it's one of the few industries where you, where you even consider, you know, having a cage fight between, you know.

1
Speaker 1
[00:55:38.40 - 00:55:41.78]

Yeah. So I'm, I'm, I'm not a fan. I'm not a fan of the cage fights.

2
Speaker 2
[00:55:41.86 - 00:55:48.48]

I'm not a fan of the, I mean, you do well, but you know, I, I know. Even though I suspect it won't be your strength.

1
Speaker 1
[00:55:48.54 - 00:56:14.92]

I know I fighting in cage fights is not, not my, not my forte, but the thing I was going to say is, look, instead of, instead of pointing fingers, instead of having feuds and saying, this guy's the bad guy, this guy's the good guy, let's think systemically. Right. Um, going back to like the race to the top idea, right. The idea that it's like, let's set standards. instead of pointing fingers, that people doing something bad, let's do something good.

[00:56:15.16 - 00:56:32.74]

And, and then, a lot of the time, people just follow along. We invent an interpretability idea. You know, just a few weeks ago we, we put out, you know, I was talking about it a few minutes ago, this innovation in interpretability, being able to see inside the model. Yeah. A few weeks later, we got similar things from open AI.

[00:56:33.00 - 00:56:55.28]

We've seen internally other companies increase their prioritization on it. So, a lot of the time, you can just do something good and you can inspire others to do something good. Now, if you've done a lot of that, if you've set these standards, if they're industry standards, and then there's someone who's not complying with them, there's something that's really wrong, then you can talk about pointing fingers. Yeah.

2
Speaker 2
[00:57:00.32 - 00:57:04.46]

Let's spend a few minutes talking about culture. How many people are you in the firm?

1
Speaker 1
[00:57:04.78 - 00:57:11.78]

We are, we are about 600. as of, as of a couple of weeks ago. I've been on vacation, so it may be even higher now.

2
Speaker 2
[00:57:13.20 - 00:57:14.22]

What's the culture like?

1
Speaker 1
[00:57:14.58 - 00:57:42.56]

Yeah. I would describe a few elements of the culture. One, one element of the culture is, is what I describe as do the stupid, simple thing that works. A number of folks at Anthropic are ex-physicists, because, you know, I myself had that background and a couple of my co-founders had that background, including one person who was actually a professor of physics before he, before he co-founded Anthropic. and, you know, physicists look for simple explanations of things.

[00:57:43.16 - 00:58:06.14]

So one of the elements of our culture is, you know, don't do something overcomplicated, right? A lot of academic ML research tends to overcomplicate things. We go for the simplest thing possible that works. We have the same view in engineering. And again, we have the same view, even on things like safety and ethics, on interpretability, on, you know, our constitutionally eye methods.

[00:58:06.30 - 00:58:19.56]

They're all incredibly simple ideas that we just try and push as far as we can. Even this race to the top thing, right? You can say it in a sentence or two, right? It's not complicated. Like you, don't need a hundred page paper to talk about it.

[00:58:19.82 - 00:58:24.84]

It's a simple strategy. Do good things and try and encourage others to follow.

2
Speaker 2
[00:58:25.24 - 00:58:29.84]

When you hire 600 people in three years, how can you be confident that they are good?

1
Speaker 1
[00:58:30.70 - 00:58:54.14]

Yeah. So, you know, I think, I think candidly. one challenge of the AI industry, right, is how fast everything moves. So, you know, in a normal startup, things, you know, might grow 1.5 X or 2 X a year. We recognize that in this field, things move so fast that faster growth is required in order to meet the needs of the market.

[00:58:56.08 - 00:59:13.52]

And that ends up entailing faster growth than usual. I was actually worried about this at the beginning of the call. I said, oh my God, we have this dilemma. How do we deal with it? I have generally been positively surprised at how well we've been able to handle it so far, right?

[00:59:13.54 - 00:59:28.54]

How good we've been able to scale hiring processes. How much. I feel everyone is both technically talented, knowledgeable, and just generally kind and compassionate people, which I think are equally important as hiring technically talented people.

2
Speaker 2
[00:59:28.56 - 00:59:34.22]

So what do you look for? So, here I'm sitting, you are interviewing me now for a position. What do you look for?

1
Speaker 1
[00:59:34.64 - 00:59:56.50]

Yeah. I mean, again, you know, we look for willingness to do the simple thing that works. You know, we look for talent. It generally, you know, we don't necessarily look at years of experience in the AI field. Like a number of folks we hire are physicists or other natural scientists who, you know, have, have maybe only been doing, doing AI for a month or so.

[00:59:56.52 - 01:00:18.26]

Right. Have only, have only been doing a project on their own. And so we look for, we look for ability to learn. We look for, you know, we look for curiosity, ability to quickly get to the heart of the matter. And then, in terms of values, you know, we, we just look for thinking in terms of the public benefit, right?

[01:00:18.28 - 01:00:54.16]

Like it's, it's less that we have particular opinions on what the right policies for anthropic is or what the right things to do in the world. It's, it's more, we want to carry a spirit as we, as we scale the company. And it gets increasingly hard as the company gets bigger and bigger. because, you know, how do you, how do you find, how do you find all these people? But we, we want people who, who carry some amount of public spirit, who understand, on one hand, that anthropic needs to be a commercial entity, to be, to be close enough to the center of this, to have an impact.

[01:00:54.34 - 01:01:03.34]

But when you hire to understand that, that, that in the long run, we're aiming to, you know, we're aiming for this, this public benefit, the societal impact.

2
Speaker 2
[01:01:03.60 - 01:01:05.60]

When you hire, do you feel you have unlimited amount of money?

1
Speaker 1
[01:01:07.20 - 01:01:38.40]

You know, I think compute is almost all of our expenses. I won't give an exact number, but you know, I think, you know, can be publicly backed out that it's more than 80%. And so, so salaries doesn't matter really in terms of, in terms of paying people, we think more about what is fair, right? We want to do something that's, you know, that's fair, that meets the market, that treats people. Well, it's, it's less of a consideration of like, you know, how much, how much money are we spending?

[01:01:38.40 - 01:01:49.62]

Because compute is, is the biggest expenditure. It's more, how can we, how can we create a place where everyone feels they're treated fairly and people who do equal work get equal pay.

2
Speaker 2
[01:01:50.48 - 01:02:01.20]

Now you work with all these brilliant minds and kind of geniuses and, and perhaps even some, some prima donnas, what's the best way to manage them or lead them?

1
Speaker 1
[01:02:01.76 - 01:02:19.38]

Yeah. You know, I think, I guess they can't be managed. One of the most important principles is, is just the thing you said, which is letting creativity happen. You know, if things are too top down, then it's, it's hard for people to be fully creative.

[01:02:21.70 - 01:02:59.88]

It, you know, if you look at a lot of the big innovations in the ML field over the last 10 years, like the invention of the transformer, you know, no one at Google kind of ordered, you know, Oh, you know, here's the project, here's what we're trying to produce. It was, it was just kind of, you know, it was, it was a decentralized effort. At the same time, you have to make a product and everyone has to work together to make a single thing. And I think that creative tension between we need new ideas, but we need everyone to kind of contribute to one thing. I think that creative tension is where, is where the magic is finding, finding the right combination so that you can get the best of both worlds.

2
Speaker 2
[01:02:59.88 - 01:03:02.14]

You run this company together with your sister, right?

1
Speaker 1
[01:03:02.54 - 01:03:11.34]

Yes. Yes. How is that? We both worked at OpenAI and then we both founded Anthropic together. It's really great.

[01:03:11.56 - 01:03:54.30]

So, you know, the real division of labor is, you know, she, she does most of the things you would describe as running the company day to day, managing people, figuring out the structure of the company, you know, making sure we have, you know, a CFO, a chief product officer, you know, making sure comp is set up in a reasonable way, making sure the culture is good. I, I think more in terms of kind of ideas and strategy. every couple of weeks, I'll give a talk to the company, basically a vision talk, where I say, here's some things we're thinking about strategically. These aren't decisions. This is kind of a picture of what leadership is thinking about.

[01:03:54.40 - 01:04:02.34]

What do we think is going to be big in the next year? Where do we think things are going, both on the commercial side, the research side, the public benefit side?

2
Speaker 2
[01:04:02.40 - 01:04:03.42]

Is she younger or older than you?

1
Speaker 1
[01:04:03.86 - 01:04:05.26]

She is four years younger than me.

2
Speaker 2
[01:04:05.38 - 01:04:06.18]

Is she clever than you?

1
Speaker 1
[01:04:06.94 - 01:04:10.00]

We are both extremely skilled in different ways.

2
Speaker 2
[01:04:11.10 - 01:04:12.30]

What did your parents do?

1
Speaker 1
[01:04:12.88 - 01:04:24.94]

So my, my father is deceased. He was, he was previously a craftsman. My mother's, my mother's retired. She was, she was a project, a project manager for public libraries.

2
Speaker 2
[01:04:24.94 - 01:04:26.40]

How were you raised?

1
Speaker 1
[01:04:26.94 - 01:04:59.20]

How was I raised? You know, it, it, it, there really was, I think, a big focus on social responsibility and helping the world. Like that was, that was, I think, a big thing for my, for my parents. You know, they really thought about how do you, how do you, you know, how do you make things better? How do you, how do people who have been born in a fortunate position, you know, reflect their responsibilities and, you know, deliver their responsibilities to those who are, who are less fortunate.

[01:04:59.20 - 01:05:02.94]

And, you know, you can kind of see that in the public benefit orientation of the company.

2
Speaker 2
[01:05:03.36 - 01:05:05.78]

So, like the 14 year old Dario, what was he up to?

1
Speaker 1
[01:05:06.72 - 01:05:21.98]

I mean, I was really into, you know, math and science. Like, you know, I did like math competitions and all of that, but, you know, I was, I was just also thinking about, like, you know, what, you know, or how could I, how could I apply those skills to, you know, invent something that would, that would help people.

2
Speaker 2
[01:05:21.98 - 01:05:22.80]

Did you have any friends?

1
Speaker 1
[01:05:23.10 - 01:05:38.78]

Did I have any friends? You know, less than, less than I would have, less than I would have liked, you know, I was, I was, I was a little bit, I was a little bit introverted, but you know, that there were, there were people who, who, you know, who I knew back then, who I still, who I still know now.

2
Speaker 2
[01:05:38.84 - 01:05:40.78]

So is Entropic like Revenge of the Nerd?

1
Speaker 1
[01:05:41.54 - 01:05:44.40]

You know, I wouldn't really put it, I wouldn't really put it in those terms.

2
Speaker 2
[01:05:44.40 - 01:05:45.80]

And I think that's a good thing. I love that kind of stuff.

1
Speaker 1
[01:05:45.80 - 01:06:24.06]

I wouldn't really put it in those terms, if only because, you know, I'm kind of reluctant to like set different groups against off each other. You know, like people are, different kinds of people are good at different things. You know, like we have a whole sales team, like they're good at a whole different set of things than I am. Like, you know, of course I'm the CEO, so I have to learn how to do some sales as well, but they're just very different skills. And one of the things you think about, you realize in a company is that different kinds of people with very different kind of skills, just, you know, like you, you recognize the value of a very wide range of skills, including ones that, you know, that, that you're, that, that you have no ability in yourself.

[01:06:24.06 - 01:06:24.42]

Right.

2
Speaker 2
[01:06:24.46 - 01:06:26.12]

So what, what drives you now?

1
Speaker 1
[01:06:26.66 - 01:06:48.60]

You know, I think we're in a very special time, in kind of like the AI world. Like, you know, this is, you know, these, these things I've said about, you know, how crazy things could be in 2025 or 2026.. You know, I think it's important to get that right. Right. And you know, running Anthropic, I'm, you know, that's only one small piece of that, right.

[01:06:48.70 - 01:07:48.20]

There are, there are other companies, you know, some of them are bigger or better known than we are. And so, on one hand, you know, we have only one small part to play, but you know, I think, given the importance of what's happening for, you know, for the economy, for, for, for, for, for humanity, I think we have an important opportunity to, you know, make sure that these things go well, there's a lot of variance in how things could go. And I think we have the ability to affect that, you know, of course, day to day, we have to, we have to grow the business, we have to hire people, we have to sell products. And, you know, I think that's important and it's, you know, it's, it's important to do that well, so that the company is relevant, but, but I think, in the long run, the thing that drives me, or that, at least I hope, drives me is, is, you know, is, is, is the desire to capture some of that variance and push things in a good direction. How do you relax?

[01:07:48.48 - 01:08:11.66]

How do I relax? So, you know, I'm in Norway now. This is not relaxing, but I, I came here from, from my vacation, from my vacation in, from my vacation in Italy. So, you know, every, every year I take, I take a few weeks, I take a few weeks off to kind of relax and think about the deeper concepts. I go swimming every day.

[01:08:12.66 - 01:08:23.72]

Actually, me and my sister still play video games. We used to do this since high school. And, you know, now, now, you know, I'm, I'm over 40 and she's, she's like, you know,

2
Speaker 2
[01:08:23.98 - 01:08:25.04]

What kind of games do you play?

1
Speaker 1
[01:08:25.38 - 01:08:52.02]

Well, we recently got the new, the new Final Fantasy game. So we played Final Fantasy in high school. It was, like, you know, a game made in the nineties and they recently made a remake of it. And, and so, so, you know, we recently started playing, like, you know, the new version with all the like fancy graphics from, like, you know, 20 years of, 20 years of progress in, well, actually, GPUs. And, you know, we were ourselves noticing, it was like, wow, you know, this is like, we used to do this when we were in high school.

[01:08:52.06 - 01:08:53.38]

Now we're like running this company.

2
Speaker 2
[01:08:54.00 - 01:08:57.08]

Well, well, I'm glad to hear that. Some people never grow up.

1
Speaker 1
[01:08:57.46 - 01:09:02.30]

I, I, I, I don't think we've grown up in certain ways. I don't think we've grown up. Hopefully we have in others.

2
Speaker 2
[01:09:02.98 - 01:09:09.10]

Talking of which we always finish off these podcasts with, with, with a question, what kind of advice do you have to young people?

1
Speaker 1
[01:09:10.46 - 01:09:37.02]

Yeah. You know, I would, I would say, you know, get to, you know, gain familiarity with these new AI technologies. You know, I'm not, I'm not going to offer some kind of bromide about, you know, I know exactly which jobs are going to, you know, be, be big and which, which aren't. I think we don't know that. And, and also, you know, we don't, we don't know that AI won't, won't touch every area.

[01:09:38.10 - 01:10:29.80]

But I think it's safe to say that there will all, you know, there, there's going to be a role for humans, in kind of using these technologies and working alongside them at the very least, at the very least, understanding them in the public debate, that's going to come from them. I guess the other thing I would say, and this is already important advice, but I think it's going to get really more important is just the, the faculties about skepticism about information. as AI generates more and more information and content, being discerning about that, information is going to become more and more important and more and more necessary. I hope that we'll have AI systems that help us sift through everything that, you know, that help us understand the world. So that we're kind of less vulnerable to these kinds of, you know, to, to, to, to, to, to these kinds of attacks.

[01:10:30.30 - 01:10:40.12]

But, but at the end of the day, it has to come from you. You have to have some basic desire, some basic curiosity, some basic discernment. And so I think developing those is important.

2
Speaker 2
[01:10:40.90 - 01:10:51.10]

Well, that's a really great, great advice. Well, big thanks. This has been a true blast and I wish you all the best. Get back to Italy and get some more rest and do some more deep conceptual.

1
Speaker 1
[01:10:51.10 - 01:10:54.42]

thinking. Yes. Thank you so much for having me on the podcast. Thank you.

v1.0.0.251205-1-20251205132707_os