DiscoverThinking Machines: AI & PhilosophyPre-training LLMs: One Model To Rule Them All? with Talfan Evans, DeepMind
Pre-training LLMs: One Model To Rule Them All? with Talfan Evans, DeepMind

Pre-training LLMs: One Model To Rule Them All? with Talfan Evans, DeepMind

Update: 2024-05-18
Share

Description

Talfan Evans is a research engineer at DeepMind, where he focuses on data curation and foundational research for pre-training LLMs and multimodal models like Gemini. I ask Talfan: 

  • Will one model rule them all?
  • What does "high quality data" actually mean in the context of LLM training?
  • Is language model pre-training becoming commoditized?
  • Are companies like Google and OpenAI keeping their AI secrets to themselves?
  • Does the startup or open source community stand a chance next to the giants?

Also check out Talfan's latest paper at DeepMind, Bad Students Make Good Teachers.

Comments 
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Pre-training LLMs: One Model To Rule Them All? with Talfan Evans, DeepMind

Pre-training LLMs: One Model To Rule Them All? with Talfan Evans, DeepMind

Daniel Reid Cahn