Menu
Sign In Pricing Add Podcast

Lex Fridman Podcast

#452 – Dario Amodei: Anthropic CEO on Claude, AGI & the Future of AI & Humanity

4537.009 - 4547.365 Lex Fridman

It's both. And so deception, and that's where mechanistic interpretability comes into play. And hopefully the techniques used for that are not made accessible to the model.

0
💬 0

Comments

There are no comments yet.

Log in to comment.