Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing
Podcast Image

Lex Fridman Podcast

#452 – Dario Amodei: Anthropic CEO on Claude, AGI & the Future of AI & Humanity

11 Nov 2024

Transcription

Full Episode

0.149 - 27.065 Lex Fridman

The following is a conversation with Dario Amadei, CEO of Anthropic, the company that created Claude, that is currently and often at the top of most LLM benchmark leaderboards. On top of that, Dario and the Anthropic team have been outspoken advocates for taking the topic of AI safety very seriously, and they have continued to publish a lot of fascinating AI research on this and other topics.

0

28.089 - 49.73 Lex Fridman

I'm also joined afterwards by two other brilliant people from Anthropic. First, Amanda Askell, who is a researcher working on alignment and fine-tuning of Claude, including the design of Claude's character and personality. A few folks told me she has probably talked with Claude more than any human at Anthropic.

0

50.631 - 62.608 Lex Fridman

So she was definitely a fascinating person to talk to about prompt engineering and practical advice on how to get the best out of Claude. After that, Chris Ola stopped by for a chat.

0

63.248 - 86.918 Lex Fridman

He's one of the pioneers of the field of mechanistic interpretability, which is an exciting set of efforts that aims to reverse engineer neural networks to figure out what's going on inside, inferring behaviors from neural activation patterns inside the network. This is a very promising approach for keeping future super-intelligent AI systems safe.

0

87.619 - 102.433 Lex Fridman

For example, by detecting from the activations when the model is trying to deceive the human it is talking to. And now a quick few second mention of each sponsor. Check them out in the description. It's the best way to support this podcast.

102.794 - 127.871 Lex Fridman

We got Encore for machine learning, Notion for machine learning powered note taking and team collaboration, Shopify for selling stuff online, BetterHelp for your mind. and element for your health. Choose Wisely, my friends. Also, if you want to work with our amazing team, or just want to get in touch with me for whatever reason, go to lexfriedman.com slash contact. And now onto the full ad reads.

128.031 - 145.562 Lex Fridman

I try to make these interesting, but if you skip them, please still check out our sponsors. I enjoy their stuff. Maybe you will too. This episode is brought to you by Encore, a platform that provides data-focused AI tooling for data annotation, curation, and management, and for model evaluation.

145.902 - 159.394 Lex Fridman

We talk a little bit about public benchmarks in this podcast, I think mostly focused on software engineering, SWEBench. There's a lot of exciting developments about how do you have a benchmark that you can't cheat on.

160.074 - 178.342 Lex Fridman

But if it's not public, then you can use it the right way, which is to evaluate how well is the annotation, the data curation, the training, the pre-training, the post-training, all of that, how's that working? Anyway, a lot of the fascinating conversation with the anthropic folks was focused on the language side.

Comments

There are no comments yet.

Please log in to write the first comment.