DiscoverSample SpaceTime for some (extreme) distillation with Thomas van Dongen - founder of the Minish Lab
Time for some (extreme) distillation with Thomas van Dongen - founder of the Minish Lab

Time for some (extreme) distillation with Thomas van Dongen - founder of the Minish Lab

Update: 2025-01-15
Share

Description

Word embeddings might feel like they are a little bit out of fashion. After all, we have attention mechanisms and transformer models now, right? Well, it turns out that if you apply distillation the right way you can actually get highly performant word embeddings out. It's a technique featured by the model2vec project from the Minish lab and in this episode we talk to the founder to learn more about the technique.

We have a Discord these days, feel free to discuss the podcast with us there! https://discord.probabl.ai

This podcast is part of the open efforts over at probabl.

To learn more you can check out website or reach out to us on social media.

Website: https://probabl.ai/

Bluesky: https://bsky.app/profile/probabl.bsky.social

LinkedIn: https://www.linkedin.com/company/probabl

Twitter: https://x.com/probabl_ai

#probabl

Comments 
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Time for some (extreme) distillation with Thomas van Dongen - founder of the Minish Lab

Time for some (extreme) distillation with Thomas van Dongen - founder of the Minish Lab