ML Infrastructure Without The Ops: Simplifying The ML Developer Experience With Runhouse
Update: 2024-11-111
Description
Summary
Machine learning workflows have long been complex and difficult to operationalize. They are often characterized by a period of research, resulting in an artifact that gets passed to another engineer or team to prepare for running in production. The MLOps category of tools have tried to build a new set of utilities to reduce that friction, but have instead introduced a new barrier at the team and organizational level. Donny Greenberg took the lessons that he learned on the PyTorch team at Meta and created Runhouse. In this episode he explains how, by reducing the number of opinions in the framework, he has also reduced the complexity of moving from development to production for ML systems.
Announcements
Parting Question
Machine learning workflows have long been complex and difficult to operationalize. They are often characterized by a period of research, resulting in an artifact that gets passed to another engineer or team to prepare for running in production. The MLOps category of tools have tried to build a new set of utilities to reduce that friction, but have instead introduced a new barrier at the team and organizational level. Donny Greenberg took the lessons that he learned on the PyTorch team at Meta and created Runhouse. In this episode he explains how, by reducing the number of opinions in the framework, he has also reduced the complexity of moving from development to production for ML systems.
Announcements
- Hello and welcome to the AI Engineering Podcast, your guide to the fast-moving world of building scalable and maintainable AI systems
- Your host is Tobias Macey and today I'm interviewing Donny Greenberg about Runhouse and the current state of ML infrastructure
- Introduction
- How did you get involved in machine learning?
- What are the core elements of infrastructure for ML and AI?
- How has that changed over the past ~5 years?
- For the past few years the MLOps and data engineering stacks were built and managed separately. How does the current generation of tools and product requirements influence the present and future approach to those domains?
- There are numerous projects that aim to bridge the complexity gap in running Python and ML code from your laptop up to distributed compute on clouds (e.g. Ray, Metaflow, Dask, Modin, etc.). How do you view the decision process for teams trying to understand which tool(s) to use for managing their ML/AI developer experience?
- Can you describe what Runhouse is and the story behind it?
- What are the core problems that you are working to solve?
- What are the main personas that you are focusing on? (e.g. data scientists, DevOps, data engineers, etc.)
- How does Runhouse factor into collaboration across skill sets and teams?
- Can you describe how Runhouse is implemented?
- How has the focus on developer experience informed the way that you think about the features and interfaces that you include in Runhouse?
- How do you think about the role of Runhouse in the integration with the AI/ML and data ecosystem?
- What does the workflow look like for someone building with Runhouse?
- What is involved in managing the coordination of compute and data locality to reduce networking costs and latencies?
- What are the most interesting, innovative, or unexpected ways that you have seen Runhouse used?
- What are the most interesting, unexpected, or challenging lessons that you have learned while working on Runhouse?
- When is Runhouse the wrong choice?
- What do you have planned for the future of Runhouse?
- What is your vision for the future of infrastructure and developer experience in ML/AI?
Parting Question
- From your perspective, what are the biggest gaps in tooling, technology, or training for AI systems today?
- Thank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast covers the latest on modern data management. Podcast.__init__ covers the Python language, its community, and the innovative ways it is being used.
- Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.
- If you've learned something or tried out a project from the show then tell us about it! Email hosts@aiengineeringpodcast.com with your story.
- To help other people find the show please leave a review on iTunes and tell your friends and co-workers.
- Runhouse
- PyTorch
- Kubernetes
- Bin Packing
- Linear Regression
- Gradient Boosted Decision Tree
- Deep Learning
- Transformer Architecture)
- Slurm
- Sagemaker
- Vertex AI
- Metaflow
- MLFlow
- Dask
- Ray
- Spark
- Databricks
- Snowflake
- ArgoCD
- PyTorch Distributed
- Horovod
- Llama.cpp
- Prefect
- Airflow
- OOM == Out of Memory
- Weights and Biases
- KNative
- BERT language model
Comments
Top Podcasts
The Best New Comedy Podcast Right Now – June 2024The Best News Podcast Right Now – June 2024The Best New Business Podcast Right Now – June 2024The Best New Sports Podcast Right Now – June 2024The Best New True Crime Podcast Right Now – June 2024The Best New Joe Rogan Experience Podcast Right Now – June 20The Best New Dan Bongino Show Podcast Right Now – June 20The Best New Mark Levin Podcast – June 2024
In Channel