DiscoverVanishing GradientsEpisode 21: Deploying LLMs in Production: Lessons Learned
Episode 21: Deploying LLMs in Production: Lessons Learned

Episode 21: Deploying LLMs in Production: Lessons Learned

Update: 2023-11-14
Share

Description

Hugo speaks with Hamel Husain, a machine learning engineer who loves building machine learning infrastructure and tools πŸ‘·. Hamel leads and contributes to many popular open-source machine learning projects. He also has extensive experience (20+ years) as a machine learning engineer across various industries, including large tech companies like Airbnb and GitHub. At GitHub, he led CodeSearchNet, a large language model for semantic search that was a precursor to CoPilot. Hamel is the founder of Parlance-Labs, a research and consultancy focused on LLMs.



They talk about generative AI, large language models, the business value they can generate, and how to get started.



They delve into




  • Where Hamel is seeing the most business interest in LLMs (spoiler: the answer isn’t only tech);

  • Common misconceptions about LLMs;

  • The skills you need to work with LLMs and GenAI models;

  • Tools and techniques, such as fine-tuning, RAGs, LoRA, hardware, and more!

  • Vendor APIs vs OSS models.



LINKS



CommentsΒ 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Episode 21: Deploying LLMs in Production: Lessons Learned

Episode 21: Deploying LLMs in Production: Lessons Learned

Hugo Bowne-Anderson