Where and how can we use foundation models in AutoML? Richard Song, researcher at Google DeepMind, has some answers. Starting off from his position paper on leveraging foundation models for optimization, we chat about what makes foundation models valuable for AutoML, how the next steps could look like, but also why the community is not currently embracing the topic as much as it could. Paper Link: https://arxiv.org/abs/2405.03547 Richard's website: https://xingyousong.github.io/
Oscar Beijbom is talking about what it's like to run an AutoML startup: Nyckel. Beyond that, we chat about the differences between academia and industry, what truly matters in application and more. Check out Nyckel at: https://www.nyckel.com/
Colin White, head of research at Abacus AI, takes us on a tour of Neural Architecture Search: its origins, important paradigms and the future of NAS in the age of LLMs. If you're looking for a broad overview of NAS, this is the podcast for you!
There are so many great foundation models in many different domains - but how do you choose one for your specific problem? And how can you best finetune it? Sebastian Pineda has an answer: Quicktune can help select the best model and tune it for specific use cases. Listen to find out when this will be a Huggingface feature and if hyperparameter optimization is even important in finetuning models (spoiler: very much so)!
Designing algorithms by hand is hard, so Chris Lu and Matthew Jackson talk about how to meta-learn them for reinforcement learning. Many of the concepts in this episode are interesting to meta-learning approaches as a whole, though: "how expressive can we be and still perform well?", "how can we get the necessary data to generalize?" and "how do we make the resulting algorithm easy to apply in practice?" are problems that come up for any learning-based approach to AutoML and some of the...
AutoML can be a tool for good, but there are pitfalls along the way. Rahul Sharma and David Selby tell us about how AutoML systems can be used to give us false impressions about explainability metrics of ML systems - maliciously, but also on accident. While this episode isn't talking about a new exciting AutoML method, it can tell us a lot about what can go wrong in applying AutoML and what we should think about when we build tools for ML novices to use.
In today's episode, we're introducing the very special Theresa Eimer to the show. Theresa will be taking over the hosting of many of the future episodes. Theresa has already recorded multiple episodes and we are stoked to air those shortly. We also spend a few moments explaining my relative absence in the last few months (since the war in the middle east erupted) and what I'm up to now. Theresa, we are all so excited to be doing this together! To learn more about Theresa, Follow her on T...
Today we're talking with Nick Erickson from AutoGluon. We discuss AutoGluon's fascinating origin story, its unique point of view, the science and engineering behind some of its unique contributions, Amazon's Machine Learning University, AutoGluon's multi-layer stack ensembler in all its detail, their feature preprocessing pipeline, their feature type inference, their adaptive approach to early stopping, controlling for inference speeds, the different multi-modal architectures, the ML culture...
Today we're talking with Joseph Giovanelli about his work on integrating logic and argumentation into AutoML systems. Joseph is a PhD student at the University of Bologna. He was more recently in Hannover working on ethics and fairness with Marius’ team. The paper he published presents his framework, HAMLET, which stands for Human-centric AutoML via Logic and Argumentation. It allows a user to iteratively specify constraints in a formal manner and, once defined, those constraints become logi...
Today we're talking with Caitlin Owen, a post-doc at the University of Otago about her work on error decomposition. She recently published a paper titled "Towards Explainable AutoML Using Error Decomposition" about how a more granular view of the components of error can lead the construction of better AutoML systems. Read her paper here: https://link.springer.com/chapter/10.1007/978-3-031-22695-3_13 Follow her on Twitter here: @CaitAshfordOwen Connect with her on LinkedIn here: https://www...
Today we're talking with Gaurav Rao, the EVP & GM of Machine Learning and AI at AtScale, a company centered around the semantic layer. For some time now, I've been feeling that there is a deep connection between a formal articulation of business context and the realization of the dream of AutoML, so I searched for people in the space who can help shine light on this direction. Gaurav is one of the few who can speak about this. As you'll hear, he's extremely pedagogic and he's walking us...
Today Ankush Garg is speaking with Rishi Bommasani, PhD student at Stanford and one of the originator of the term Foundation Models. They’re talking about the origins of the term Foundation Model, which he and his group advanced, in the paper "On the Opportunities and Risks of Foundation Models". They’ll talk about self-supervision, issues of scale, the motivation behind the terminology, the origins of the Research for Foundation Models Institute at Stanford, outcome homogenization, emergenc...
Today we're talking with Raymond Peck, a senior engineer and director in the AutoML space. He spent time at H2O, dotData, Alteryx and many other places. This is a fascinating conversation about the business, engineering, and science of machine learning automation in production. Learning about his experience is crucial for understanding the biography of the space. We discuss the early motivations behind AutoML, the initial value propositions that propelled the first movers in the market, the...
Today we’re talking to Noah Hollmann and Samuel Muller about their paper on TabPFN - which is an incredible spin on AutoML based on Bayesian inference and transformers. [Quick note on audio quality]: Some of the tracks have not recorded perfectly but I felt that the content there was too important not to release. Sorry for any ear-strain! In the episode, we spend some time discussing posterior predictive probabilities before discussing how exactly they’ve pre-fitted their network, how they g...
Today we’re talking to Sean Sexton, the Director of Modeling and Analytics Consulting at KPMG, about the role of models in financial institutions and how the risks associated with them is managed. This turned out to be an incredibly deep and interesting topic, and we really only scratched the surface of it. Sean has a unique ability to summarize developments in an entire space. If you're interested to learn more about modeling in financial institutions and about the history of how we got he...
Today we’re talking to Matt Levine. Matt is a PhD student in computing and mathematical sciences at Caltech, and he focuses on improving the prediction and inference of physical systems by blending together both mechanistic modeling and machine learning. This episode is one of my favorites: we go pretty deep into dynamical systems, and into Matt's new framework for solving them by blending traditional, mechanistic, approaches with machine learning. This is a fascinating use of machine ...
Today we’re chatting with Junhong Shen, a PhD student at Carnegie Mellon. Junhong and her team are working on the generalizability of NAS algorithms across a diverse set of tasks. Today we'll be talking about DASH, a NAS algorithm that takes diversity of tasks at its center. In order to implement DASH, Junhong and her team implemented three clever ideas that she'll share with us. Efficient Architecture Search for Diverse Tasks - https://arxiv.org/pdf/2204.07554.pdf Tackling Diverse Tasks ...
Today we're speaking with Marius Lindauer and it is certainly one of my favorite episodes! As you’ll hear, Marius is full of ideas for where AutoML systems can and should go. These ideas are crystallized in a blog-post, published here: https://www.automl.org/rethinking-automl-advancing-from-a-machine-centered-to-human-centered-paradigm/ If you’re searching for research directions, this conversation left me with dozens of ideas. Marius and his team are doing phenomenal work to make AutoML syst...
Today Ankush Garg is talking to Mehdi Bahrami about his recent project: BERT-Sort. BERT-Sort is an example of how large language models can add useful context to tabular datasets, and to AutoML systems. Mehdi is a Member of Research Staff at Fujitsu and, as he describes, he began using AutoML systems for his research, yet he came across some crucial limitations of existing solutions. The modifications he made highlight a promising future for the relationship between language models and Auto...
In today's episode, we’re talking to Lars Kothoff about the fascinating origin story of AutoML (as he sees it), and how it emerged from the SAT community. While talking to many of you, it became clear that this origin story is one that a lot of people have some vague sense about, but not a very concrete knowledge of so hopefully this episode can help to flesh out the narrative with greater clarity. We'll be discussing his survey paper from 2016, "Algorithm selection for combinatorial search ...