Menu
Sign In Pricing Add Podcast

Lex Fridman Podcast

#434 – Aravind Srinivas: Perplexity CEO on Future of AI, Search & the Internet

4724.514 - 4745.18 Aravind Srinivas

You take Common Crawl and instead of 1 billion, go all the way to 175 billion. But that was done through analysis called the scaling loss, which is for a bigger model, you need to keep scaling the amount of tokens and you train on 300 billion tokens. Now it feels small. These models are being trained on like tens of trillions of tokens. and trillions of parameters.

0
💬 0

Comments

There are no comments yet.

Log in to comment.