Menu
Sign In Pricing Add Podcast

Lex Fridman Podcast

#459 – DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate, and AI Megaclusters

8488.257 - 8498.829 Dylan Patel

Related to our previous discussion, this multi-head latent attention can save about 80% to 90% in memory from the attention mechanism, which helps especially along context.

0
💬 0

Comments

There are no comments yet.

Log in to comment.