DiscoverData Science at HomeFrom Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)
From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)

From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)

Update: 2025-11-11
Share

Description

LLMs generate text painfully slow, one low-info token at a time. Researchers just figured out how to compress 4 tokens into smart vectors & cut costs by 44%—with full code & proofs! Meanwhile OpenAI drops product ads, not papers.

We explore CALM & why open science matters. 🔥📊


 


Sponsors


This episode is brought to you by Statistical Horizons 

At Statistical Horizons, you can stay ahead with expert-led livestream seminars that make data analytics and AI methods practical and accessible.

Join thousands of researchers and professionals who’ve advanced their careers with Statistical Horizons.

Get $200 off any seminar with code DATA25 at https://statisticalhorizons.com


 

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)

From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)

Francesco Gadaleta