Discover
MLOps.community
![MLOps.community MLOps.community](https://is1-ssl.mzstatic.com/image/thumb/Podcasts115/v4/4b/6a/cc/4b6acc53-d792-b8bf-530d-d65d6b55a366/mza_16675100765413598218.jpg/400x400bb.jpg)
403 Episodes
Reverse
Alex Milowski is a researcher, developer, entrepreneur, mathematician, and computer scientist.Evolving Workflow Orchestration // MLOps Podcast #291 with Alex Milowski, Entrepreneur and Computer Scientist.// AbstractThere seems to be a shift from workflow languages to code - mostly annotation pythons - happening and getting us. It is a symptom of how complex workflow orchestration has gotten. Is it a dominant trend or will we cycle back to “DAG specifications”? At Stitchfix, we had our own DSL that “compiled” into airflow DAGs and at MicroByre, we used a external workflow langauge. Both had a batch task executor on K8s but at MicroByre, we had human and robot in the loop workflows.// BioDr. Milowski is a serial entrepreneur and computer scientist with experience in a variety of data and machine learning technologies. He holds a PhD in Informatics (Computer Science) from the University of Edinburgh, where he researched large-scale computation over scientific data. Over the years, he's spent many years working on various aspects of workflow orchestration in industry, standardization, and in research.// MLOps Swag/Merchhttps://shop.mlops.community/// Related LinksWebsite: https://www.milowski.com/ --------------- ✌️Connect With Us ✌️ -------------Join our slack community: https://go.mlops.community/slackFollow us on Twitter: @mlopscommunitySign up for the next meetup: https://go.mlops.community/registerCatch all episodes, blogs, newsletters, and more: https://mlops.community/Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/Connect with Alex on LinkedIn: https://www.linkedin.com/in/alexmilowski/
Willem Pienaar is the Co-Founder and CTO ofCleric. He previously worked at Tecton as a Principal Engineer. Willem Pienaar attended the Georgia Institute of Technology.Insights from Cleric: Building an Autonomous AI SRE // MLOps Podcast #289 with Willem Pienaar, CTO & Co-Founder of Cleric.// AbstractIn this MLOps Community Podcast episode, Willem Pienaar, CTO of Cleric, breaks down how they built an autonomous AI SRE that helps engineering teams diagnose production issues. We explore how Cleric builds knowledge graphs for system understanding, and uses existing tools/systems during investigations. We also get into some gnarly challenges around memory, tool integration, and evaluation frameworks, and some lessons learned from deploying to engineering teams.// BioWillem Pienaar, CTO of Cleric, is a builder with a focus on LLM agents, MLOps, and open source tooling. He is the creator of Feast, an open source feature store, and contributed to the creation of both the feature store and MLOps categories.Before starting Cleric, Willem led the open-source engineering team at Tecton and established the ML platform team at Gojek, where he built high-scale ML systems for the Southeast Asian Decacorn.// MLOps Swag/Merchhttps://shop.mlops.community/// Related LinksWebsite: willem.co --------------- ✌️Connect With Us ✌️ -------------Join our slack community:https://go.mlops.community/slackFollow us on Twitter:@mlopscommunitySign up for the next meetup:https://go.mlops.community/registerCatch all episodes, blogs, newsletters, and more:https://mlops.community/Connect with Demetrios on LinkedIn:https://www.linkedin.com/in/dpbrinkm/Connect with Willem on LinkedIn:https://www.linkedin.com/in/willempienaar/
Vinu Sankar Sadasivan is a CS PhD ... Currently, I am working as a full-time Student Researcher at Google DeepMind on jailbreaking multimodal AI models.
Robustness, Detectability, and Data Privacy in AI // MLOps Podcast #289 with Vinu Sankar Sadasivan, Student Researcher at Google DeepMind.
// Abstract
Recent rapid advancements in Artificial Intelligence (AI) have made it widely applicable across various domains, from autonomous systems to multimodal content generation. However, these models remain susceptible to significant security and safety vulnerabilities. Such weaknesses can enable attackers to jailbreak systems, allowing them to perform harmful tasks or leak sensitive information. As AI becomes increasingly integrated into critical applications like autonomous robotics and healthcare, the importance of ensuring AI safety is growing. Understanding the vulnerabilities in today’s AI systems is crucial to addressing these concerns.
// Bio
Vinu Sankar Sadasivan is a final-year Computer Science PhD candidate at The University of Maryland, College Park, advised by Prof. Soheil Feizi. His research focuses on Security and Privacy in AI, with a particular emphasis on AI robustness, detectability, and user privacy. Currently, Vinu is a full-time Student Researcher at Google DeepMind, working on jailbreaking multimodal AI models. Previously, Vinu was a Research Scientist intern at Meta FAIR in Paris, where he worked on AI watermarking.
Vinu is a recipient of the 2023 Kulkarni Fellowship and has earned several distinctions, including the prestigious Director’s Silver Medal. He completed a Bachelor’s degree in Computer Science & Engineering at IIT Gandhinagar in 2020. Prior to their PhD, Vinu gained research experience as a Junior Research Fellow in the Data Science Lab at IIT Gandhinagar and through internships at Caltech, Microsoft Research India, and IISc.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://vinusankars.github.io/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Richard on LinkedIn: https://www.linkedin.com/in/vinusankars/
Richard Cloete is a computer scientist and a Laukien-Oumuamua Postdoctoral Research Fellow at the Center for Astrophysics, Harvard University. He is a member of the Galileo Project working under the supervision of Professor Avi, having recently held a postdoctoral position at the University of Cambridge, UK.
AI & Aliens: New Eyes on Ancient Questions // MLOps Podcast #288 with Richard Cloete,
Laukien-Oumuamua Postdoctoral Research Fellow at Harvard University.
// Abstract
Demetrios speaks with Dr. Richard Cloete, a Harvard computer scientist and founder of SEAQR Robotics, about his AI-driven work in tracking Unidentified Aerial Phenomena (UAPs) through the Galileo Project. Dr. Cloete explains their advanced sensor setup and the challenges of training AI in this niche field, leading to the creation of AeroSynth, a synthetic data tool.
He also discusses his collaboration with the Minor Planet Center on using AI to classify interstellar objects and upcoming telescope data. Additionally, he introduces Seeker Robotics, applying similar AI techniques to oceanic research with unmanned vehicles for marine monitoring. The conversation explores AI’s role in advancing our understanding of space and the ocean.
// Bio
Richard is a computer scientist and Laukien-Oumuamua Postdoctoral Research Fellow at the Center for Astrophysics, Harvard University. As a member of the Galileo Project under Professor Avi Loeb's supervision, he develops AI models for detecting and tracking aerial objects, specializing in Unidentified Anomalous Phenomena (UAP).
Beyond UAP research, he collaborates with astronomers at the Minor Planet Center to create AI models for identifying potential interstellar objects using the upcoming Vera C. Rubin Observatory.
Richard is also the CEO and co-founder of SEAQR Robotics, a startup developing advanced unmanned surface vehicles to accelerate the discovery of novel life and phenomena in Earth's oceans and atmosphere.
Before joining Harvard, he completed a postdoctoral fellowship at the University of Cambridge, UK, where his research explored the intersection of emerging technologies and law.Grew up in Cape Town, South Africa, where I used to build Tesla Coils, plasma globes, radio stethoscopes, microwave guns, AM radios, and bombs...
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: www.seaqr.net
https://itc.cfa.harvard.edu/people/richard-cloete
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Richard on LinkedIn: https://www.linkedin.com/in/richard-cloete/
A software engineer based in Delft, Alex Strick van Linschoten recently built Ekko, an open-source framework for adding real-time infrastructure and in-transit message processing to web applications. With years of experience in Ruby, JavaScript, Go, PostgreSQL, AWS, and Docker, I bring a versatile skill set to the table. I hold a PhD in History, have authored books on Afghanistan, and currently work as an ML Engineer at ZenML.
Real LLM Success Stories: How They Actually Work // MLOps Podcast #287 with Alex Strick van Linschoten, ML Engineer at ZenML.
// Abstract
Alex Strick van Linschoten, a machine learning engineer at ZenML, joins the MLOps Community podcast to discuss his comprehensive database of real-world LLM use cases. Drawing inspiration from Evidently AI, Alex created the database to organize fragmented information on LLM usage, covering everything from common chatbot implementations to innovative applications across sectors. They discuss the technical challenges and successes in deploying LLMs, emphasizing the importance of foundational MLOps practices. The episode concludes with a call for community contributions to further enrich the database and collective knowledge of LLM applications.
// Bio
Alex is a Software Engineer based in the Netherlands, working as a Machine Learning Engineer at ZenML. He previously was awarded a PhD in History (specialism: War Studies) from King's College London and has authored several critically acclaimed books based on his research work in Afghanistan.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://mlops.systemshttps://www.zenml.io/llmops-databasehttps://www.zenml.io/llmops-databasehttps://www.zenml.io/blog/llmops-in-production-457-case-studies-of-what-actually-workshttps://www.zenml.io/blog/llmops-lessons-learned-navigating-the-wild-west-of-production-llmshttps://www.zenml.io/blog/demystifying-llmops-a-practical-database-of-real-world-generative-ai-implementationshttps://huggingface.co/datasets/zenml/llmops-database
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Alex on LinkedIn: https://www.linkedin.com/in/strickvl
In his 13 years of software engineering, Ilya Reznik has specialized in commercializing machine learning solutions and building robust ML platforms. He's held technical lead and staff engineering roles at premier firms like Adobe, Twitter, and Meta. Currently, Ilya channels his expertise into his travel startup, Jaunt, while consulting and advising emerging startups.
Navigating Machine Learning Careers: Insights from Meta to Consulting // MLOps Podcast #286 with Ilya Reznik, ML Engineering Thought Leader at Instructed Machines, LLC.
// Abstract
Ilya Reznik's insights into machine learning and career development within the field. With over 13 years of experience at leading tech companies such as Meta, Adobe, and Twitter, Ilya emphasizes the limitations of traditional model fine-tuning methods. He advocates for alternatives like prompt engineering and knowledge retrieval, highlighting their potential to enhance AI performance without the drawbacks associated with fine-tuning.
Ilya's recent discussions at the NeurIPS conference reflect a shift towards practical applications of Transformer models and innovative strategies like curriculum learning. Additionally, he shares valuable perspectives on navigating career progression in tech, offering guidance for aspiring ML engineers aiming for senior roles. His narrative serves as a blend of technical expertise and practical career advice, making it a significant resource for professionals in the AI domain.
// Bio
Ilya has navigated a diverse career path since 2011, transitioning from physicist to software engineer, data scientist, ML engineer, and now content creator. He is passionate about helping ML engineers advance their careers and making AI more impactful and beneficial for society.
Previously, Ilya was a technical lead at Meta, where he contributed to 12% of the company’s revenue and managed approximately 30 production ML models. He also worked at Twitter, overseeing offline model evaluation, and at Adobe, where his team was responsible for all intelligent services within Adobe Analytics.
Based in Salt Lake City, Ilya enjoys the outdoors, tinkering with Arduino electronics, and, most importantly, spending time with his family.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: mlepath.com
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Ilya on LinkedIn: https://www.linkedin.com/in/ibreznik/
Tomaž Levak is the Co-founder and CEO of Trace Labs – OriginTrail core developers. OriginTrail is a web3 infrastructure project combining a decentralized knowledge graph (DKG) and blockchain technologies to create a neutral, inclusive ecosystem.
Collective Memory for AI on Decentralized Knowledge Graph // MLOps Podcast #285 with Tomaz Levak, Founder of Trace Labs, Core Developers of OriginTrail.
// Abstract
The talk focuses on how OriginTrail Decentralized Knowledge Graph serves as a collective memory for AI and enables neuro-symbolic AI. We cover the basics of OriginTrail’s symbolic AI fundamentals (i.e. knowledge graphs) and go over details how decentralization improves data integrity, provenance, and user control. We’ll cover the DKG role in AI agentic frameworks and how it helps with verifying and accessing diverse data sources, while maintaining compatibility with existing standards.
We’ll explore practical use cases from the enterprise sector as well as latest integrations into frameworks like ElizaOS. We conclude by outlining the future potential of decentralized AI, AI becoming the interface to “eat” SaaS and the general convergence of AI, Internet and Crypto.
// Bio
Tomaz Levak, founder of OriginTrail, is active at the intersection of Cryptocurrency, the Internet, and Artificial Intelligence (AI). At the core of OriginTrail is a pursuit of Verifiable Internet for AI, an inclusive framework addressing critical challenges of the world in an AI era. To achieve the goal of Verifiable Internet for AI, OriginTrail's trusted knowledge foundation ensures the provenance and verifiability of information while incentivizing the creation of high-quality knowledge. These advancements are pivotal to unlock the full potential of AI as they minimize the technology’s shortfalls such as hallucinations, bias, issues of data ownership, and model collapse.
Tomaz's contributions to OriginTrail span over a decade and across multiple fields. He is involved in strategic technical innovations for OriginTrail Decentralized Knowledge Graph (DKG) and NeuroWeb blockchain and was among the authors of all three foundational White Paper documents that defined how OriginTrail technology addresses global challenges. Tomaz contributed to the design of OriginTrail token economies and is driving adoption with global brands such as British Standards Institution, Swiss Federal Railways and World Federation of Haemophilia, among others.
Committed to the ongoing expansion of the OriginTrail ecosystem, Tomaz is a regular speaker at key industry events. In his appearances, he highlights the significant value that the OriginTrail DKG brings to diverse sectors, including supply chains, life sciences, healthcare, and scientific research. In a rapidly evolving digital landscape, Tomaz and the OriginTrail ecosystem as a whole are playing an important role in ensuring a more inclusive, transparent and decentralized AI.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://origintrail.io
Song recommendation: https://open.spotify.com/track/5GGHmGNZYnVSdRERLUSB4w?si=ae744c3ad528424b
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Tomaz on LinkedIn: https://www.linkedin.com/in/tomazlevak/
Krishna Sridhar is an experienced engineering leader passionate about building wonderful products powered by machine learning.
Efficient Deployment of Models at the Edge // MLOps Podcast #284 with Krishna Sridhar, Vice President of Qualcomm.
Big shout out to Qualcomm for sponsoring this episode!
// Abstract
Qualcomm® AI Hub helps to optimize, validate, and deploy machine learning models on-device for vision, audio, and speech use cases.
With Qualcomm® AI Hub, you can:
Convert trained models from frameworks like PyTorch and ONNX for optimized on-device performance on Qualcomm® devices.
Profile models on-device to obtain detailed metrics including runtime, load time, and compute unit utilization.
Verify numerical correctness by performing on-device inference.
Easily deploy models using Qualcomm® AI Engine Direct, TensorFlow Lite, or ONNX Runtime.
The Qualcomm® AI Hub Models repository contains a collection of example models that use Qualcomm® AI Hub to optimize, validate, and deploy models on Qualcomm® devices.
Qualcomm® AI Hub automatically handles model translation from source framework to device runtime, applying hardware-aware optimizations, and performs physical performance/numerical validation. The system automatically provisions devices in the cloud for on-device profiling and inference. The following image shows the steps taken to analyze a model using Qualcomm® AI Hub.
// Bio
Krishna Sridhar leads engineering for Qualcomm™ AI Hub, a system used by more than 10,000 AI developers spanning 1,000 companies to run more than 100,000 models on Qualcomm platforms.
Prior to joining Qualcomm, he was Co-founder and CEO of Tetra AI which made its easy to efficiently deploy ML models on mobile/edge hardware.
Prior to Tetra AI, Krishna helped design Apple's CoreML which was a software system mission critical to running several experiences at Apple including Camera, Photos, Siri, FaceTime, Watch, and many more across all major Apple device operating systems and all hardware and IP blocks.
He has a Ph.D. in computer science from the University of Wisconsin-Madison, and a bachelor’s degree in computer science from Birla Institute of Technology and Science, Pilani, India.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://www.linkedin.com/in/srikris/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Krishna on LinkedIn: https://www.linkedin.com/in/srikris/
Machine Learning, AI Agents, and Autonomy // MLOps Podcast #283 with Zach Wallace, Staff Software Engineer at Nearpod Inc.
// Abstract
Demetrios chats with Zach Wallace, engineering manager at Nearpod, about integrating AI agents in e-commerce and edtech. They discuss using agents for personalized user targeting, adapting AI models with real-time data, and ensuring efficiency through clear task definitions. Zach shares how Nearpod streamlined data integration with tools like Redshift and DBT, enabling real-time updates. The conversation covers challenges like maintaining AI in production, handling high-quality data, and meeting regulatory standards. Zach also highlights the cost-efficiency framework for deploying and decommissioning agents and the transformative potential of LLMs in education.
// Bio
Software Engineer with 10 years of experience. Started my career as an Application Engineer, but I have transformed into a Platform Engineer. As a Platform Engineer, I have handled the problems described below
- Localization across 6-7 different languages
- Building a custom local environment tool for our engineers
- Building a Data Platform
- Building standards and interfaces for Agentic AI within ed-tech.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
https://medium.com/renaissance-learning-r-d/data-platform-transform-a-data-monolith-9d5290a552ef
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Zach on LinkedIn: https://www.linkedin.com/in/zachary-wallace/
Since three years, Egor is bringing the power of AI to bear at Wise, across domains as varied as trading algorithms for Treasury, fraud detection, experiment analysis and causal inference, and recently the numerous applications unlocked by large language models. Open-source projects initiated and guided by Egor include wise-pizza, causaltune, and neural-lifetimes, with more on the way.
Machine Learning, AI Agents, and Autonomy // MLOps Podcast #282 with Egor Kraev, Head of AI at Wise Plc.
// Abstract
Demetrios chats with Egor Kraev, principal AI scientist at Wise, about integrating large language models (LLMs) to enhance ML pipelines and humanize data interactions. Egor discusses his open-source MotleyCrew framework, career journey, and insights into AI's role in fintech, highlighting its potential to streamline operations and transform organizations.
// Bio
Egor first learned mathematics in the Russian tradition, then continued his studies at ETH Zurich and the University of Maryland. Egor has been doing data science since last century, including economic and human development data analysis for nonprofits in the US, the UK, and Ghana, and 10 years as a quant, solutions architect, and occasional trader at UBS then Deutsche Bank. Following last decade's explosion in AI techniques, Egor became Head of AI at Mosaic Smart Data Ltd, and for the last four years is bringing the power of AI to bear at Wise, in a variety of domains, from fraud detection to trading algorithms and causal inference for A/B testing and marketing. Egor has multiple side projects such as RL for molecular optimization, GenAI for generating and solving high school math problems, and others.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
https://github.com/transferwise/wise-pizza
https://github.com/py-why/causaltune
https://www.linkedin.com/posts/egorkraev_a-talk-on-experimentation-best-practices-activity-7092158531247755265-q0kt?utm_source=share&utm_medium=member_desktop
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Egor on LinkedIn: https://www.linkedin.com/in/egorkraev/
Re-Platforming Your Tech Stack // MLOps Podcast #281 with Michelle Marie Conway, Lead Data Scientist at Lloyds Banking Group and Andrew Baker, Data Science Delivery Lead at Lloyds Banking Group.
// Abstract
Lloyds Banking Group is on a mission to embrace the power of cloud and unlock the opportunities that it provides. Andrew, Michelle, and their MLOps team have been on a journey over the last 12 months to take their portfolio of circa 10 Machine Learning models in production and migrate them from an on-prem solution to a cloud-based environment. During the podcast, Michelle and Andrew share their reflections as well as some dos (and don’ts!) of managing the migration of an established portfolio.
// Bio
Michelle Marie Conway
Michelle is a Lead Data Scientist in the high-performance data science team at Lloyds Banking Group. With deep expertise in managing production-level Python code and machine learning models, she has worked alongside fellow senior manager Andrew to drive the bank's transition to the Google Cloud Platform. Together, they have played a pivotal role in modernising the ML portfolio in collaboration with a remarkable ML Ops team.
Originally from Ireland and now based in London, Michelle blends her technical expertise with a love for the arts.
Andrew Baker
Andrew graduated from the University of Birmingham with a first-class honours degree in Mathematics and Music with a Year in Computer Science and joined Lloyds Banking Group on their Retail graduate scheme in 2015.
Since 2021 Andrew has worked in the world of data, firstly in shaping the Retail data strategy and most recently as a Data Science Delivery Lead, growing and managing a team of Data Scientists and Machine Learning Engineers. He has built a high-performing team responsible for building and maintaining ML models in production for the Consumer Lending division of the bank.
Andrew is motivated by the role that data science and ML can play in transforming the business and its processes, and is focused on balancing the power of ML with the need for simplicity and explainability that enables business users to engage with the opportunities that exist in this space and the demands of a highly regulated environment.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://www.michelleconway.co.uk/
https://www.linkedin.com/pulse/artificial-intelligence-just-when-data-science-answer-andrew-baker-hfdge/
https://www.linkedin.com/pulse/artificial-intelligence-conundrum-generative-ai-andrew-baker-qla7e/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Michelle on LinkedIn: https://www.linkedin.com/in/michelle--conway/
Connect with Andrew on LinkedIn: https://www.linkedin.com/in/andrew-baker-90952289
Jineet Doshi is an award-winning Scientist, Machine Learning Engineer, and Leader at Intuit with over 7 years of experience. He has a proven track record of leading successful AI projects and building machine-learning models from design to production across various domains which have impacted 100 million customers and significantly improved business metrics, leading to millions of dollars of impact.
Holistic Evaluation of Generative AI Systems // MLOps Podcast #280 with Jineet Doshi, Staff AI Scientist or AI Lead at Intuit.
// Abstract
Evaluating LLMs is essential in establishing trust before deploying them to production. Even post deployment, evaluation is essential to ensure LLM outputs meet expectations, making it a foundational part of LLMOps. However, evaluating LLMs remains an open problem. Unlike traditional machine learning models, LLMs can perform a wide variety of tasks such as writing poems, Q&A, summarization etc. This leads to the question how do you evaluate a system with such broad intelligence capabilities? This talk covers the various approaches for evaluating LLMs such as classic NLP techniques, red teaming and newer ones like using LLMs as a judge, along with the pros and cons of each. The talk includes evaluation of complex GenAI systems like RAG and Agents. It also covers evaluating LLMs for safety and security and the need to have a holistic approach for evaluating these very capable models.
// Bio
Jineet Doshi is an award winning AI Lead and Engineer with over 7 years of experience. He has a proven track record of leading successful AI projects and building machine learning models from design to production across various domains, which have impacted millions of customers and have significantly improved business metrics, leading to millions of dollars of impact. He is currently an AI Lead at Intuit where he is one of the architects and developers of their Generative AI platform, which is serving Generative AI experiences for more than 100 million customers around the world.
Jineet is also a guest lecturer at Stanford University as part of their building LLM Applications class. He is on the Advisory Board of University of San Francisco’s AI Program. He holds multiple patents in the field, is on the steering committee of MLOps World Conference and has also co chaired workshops at top AI conferences like KDD. He holds a Masters degree from Carnegie Mellon university.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://www.intuit.com/
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Jineet on LinkedIn: https://www.linkedin.com/in/jineetdoshi/
Robert Caulk is responsible for directing software development, enabling research, coordinating company projects, quality control, proposing external collaborations, and securing funding. He believes firmly in open-source, having spent 12 years accruing over 1000 academic citations building open-source software in domains such as machine learning, image analysis, and coupled physical processes. He received his Ph.D. from Université Grenoble Alpes, France, in computational mechanics.
Unleashing Unconstrained News Knowledge Graphs to Combat Misinformation // MLOps Podcast #279 with Robert Caulk, Founder of Emergent Methods.
// Abstract
Indexing hundreds of thousands of news articles per day into a knowledge graph (KG) was previously impossible due to the strict requirement that high-level reasoning, general world knowledge, and full-text context *must* be present for proper KG construction.
The latest tools now enable such general world knowledge and reasoning to be applied cost effectively to high-volumes of news articles. Beyond the low cost of processing these news articles, these tools are also opening up a new, controversial, approach to KG building - unconstrained KGs.
We discuss the construction and exploration of the largest news-knowledge-graph on the planet - hosted on an endpoint at AskNews.app. During talk we aim to highlight some of the sacrifices and benefits that go hand-in-hand with using the infamous unconstrained KG approach.
We conclude the talk by explaining how knowledge graphs like these help to mitigate misinformation. We provide some examples of how our clients are using this graph, such as generating sports forecasts, generating better social media posts, generating regional security alerts, and combating human trafficking.
// Bio
Robert is the founder of Emergent Methods, where he directs research and software development for large-scale applications. He is currently overseeing the structuring of hundreds of thousands of news articles per day in order to build the best news retrieval API in the world: https://asknews.app.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://emergentmethods.ai
News Retrieval API: https://asknews.app
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Rob on LinkedIn: https://www.linkedin.com/in/rcaulk/
Timestamps:
[00:00] Rob's preferred coffee
[00:05] Takeaways
[00:55] Please like, share, leave a review, and subscribe to our MLOps channels!
[01:00] Join our Local Organizer Carousel!
[02:15] Knowledge Graphs and ontology
[07:43] Ontology vs Noun Approach
[12:46] Ephemeral tools for efficiency
[17:26] Oracle to PostgreSQL migration
[22:20] MEM Graph life cycle
[29:14] Knowledge Graph Investigation Insights
[33:37] Fine-tuning and distillation of LLMs
[39:28] DAG workflow and quality control
[46:23] Crawling nodes with Phi 3 Llama
[50:05] AI pricing risks and strategies
[56:14] Data labeling and poisoning
[58:34] API costs vs News latency
[1:02:10] Product focus and value
[1:04:52] Ensuring reliable information
[1:11:01] Podcast transcripts as News
[1:13:08] Ontology trade-offs explained
[1:15:00] Wrap up
Guanhua Wang is a Senior Researcher in DeepSpeed Team at Microsoft. Before Microsoft, Guanhua earned his Computer Science PhD from UC Berkeley.
Domino: Communication-Free LLM Training Engine // MLOps Podcast #278 with Guanhua "Alex" Wang, Senior Researcher at Microsoft.
// Abstract
Given the popularity of generative AI, Large Language Models (LLMs) often consume hundreds or thousands of GPUs to parallelize and accelerate the training process. Communication overhead becomes more pronounced when training LLMs at scale. To eliminate communication overhead in distributed LLM training, we propose Domino, which provides a generic scheme to hide communication behind computation. By breaking the data dependency of a single batch training into smaller independent pieces, Domino pipelines these independent pieces of training and provides a generic strategy of fine-grained communication and computation overlapping. Extensive results show that compared with Megatron-LM, Domino achieves up to 1.3x speedup for LLM training on Nvidia DGX-H100 GPUs.
// Bio
Guanhua Wang is a Senior Researcher in the DeepSpeed team at Microsoft. His research focuses on large-scale LLM training and serving. Previously, he led the ZeRO++ project at Microsoft which helped reduce over half of model training time inside Microsoft and Linkedin. He also led and was a major contributor to Microsoft Phi-3 model training. He holds a CS PhD from UC Berkeley advised by Prof Ion Stoica.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://guanhuawang.github.io/
DeepSpeed hiring: https://www.microsoft.com/en-us/research/project/deepspeed/opportunities/
Large Model Training and Inference with DeepSpeed // Samyam Rajbhandari // LLMs in Prod Conference: https://youtu.be/cntxC3g22oU
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Guanhua on LinkedIn: https://www.linkedin.com/in/guanhua-wang/
Timestamps:
[00:00] Guanhua's preferred coffee
[00:17] Takeaways
[01:36] Please like, share, leave a review, and subscribe to our MLOps channels!
[01:47] Phi model explanation
[06:29] Small Language Models optimization challenges
[07:29] DeepSpeed overview and benefits
[10:58] Crazy unimplemented crazy AI ideas
[17:15] Post training vs QAT
[19:44] Quantization over distillation
[24:15] Using Lauras
[27:04] LLM scaling sweet spot
[28:28] Quantization techniques
[32:38] Domino overview
[38:02] Training performance benchmark
[42:44] Data dependency-breaking strategies
[49:14] Wrap up
Thanks to the High Signal Podcast by Delphina:
https://go.mlops.community/HighSignalPodcast
Aditya Naganath is an experienced investor currently working with Kleiner Perkins. He has a passion for connecting with people over coffee and discussing various topics related to tech, products, ideas, and markets.
AI's Next Frontier // MLOps Podcast #277 with Aditya Naganath, Principal at Kleiner Perkins.
// Abstract
LLMs have ushered in an unmistakable supercycle in the world of technology. The low-hanging use cases have largely been picked off. The next frontier will be AI coworkers who sit alongside knowledge workers, doing work side by side. At the infrastructure level, one of the most important primitives invented by man - the data center, is being fundamentally rethought in this new wave.
// Bio
Aditya Naganath joined Kleiner Perkins’ investment team in 2022 with a focus on artificial intelligence, enterprise software applications, infrastructure and security. Prior to joining Kleiner Perkins, Aditya was a product manager at Google focusing on growth initiatives for the next billion users team. He previously was a technical lead at Palantir Technologies and formerly held software engineering roles at Twitter and Nextdoor, where he was a Kleiner Perkins fellow. Aditya earned a patent during his time at Twitter for a technical analytics product he co-created.
Originally from Mumbai India, Aditya graduated magna cum laude from Columbia University with a bachelor’s degree in Computer Science, and an MBA from Stanford University. Outside of work, you can find him playing guitar with a hard rock band, competing in chess or on the squash courts, and fostering puppies. He is also an avid poker player.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Faith's Hymn by Beautiful Chorus: https://open.spotify.com/track/1bDv6grQB5ohVFI8UDGvKK?si=4b00752eaa96413b Substack: https://adityanaganath.substack.com/?utm_source=substack&utm_medium=web&utm_campaign=substack_profileWith thanks to the High Signal Podcast by Delphina: https://go.mlops.community/HighSignalPodcastBuilding the Future of AI in Software Development // Varun Mohan // MLOps Podcast #195 - https://youtu.be/1DJKq8StuToDo Re MI for Training Metrics: Start at the Beginning // Todd Underwood // AIQCON - https://youtu.be/DxyOlRdCofo
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Aditya on LinkedIn: https://www.linkedin.com/in/aditya-naganath/
Dr Vincent Moens is an Applied Machine Learning Research Scientist at Meta and an author of TorchRL and TensorDict in Pytorch.
PyTorch for Control Systems and Decision Making // MLOps Podcast #276 with Vincent Moens, Research Engineer at Meta.
// Abstract
PyTorch is widely adopted across the machine learning community for its flexibility and ease of use in applications such as computer vision and natural language processing. However, supporting reinforcement learning, decision-making, and control communities is equally crucial, as these fields drive innovation in areas like robotics, autonomous systems, and game-playing. This podcast explores the intersection of PyTorch and these fields, covering practical tips and tricks for working with PyTorch, an in-depth look at TorchRL, and discussions on debugging techniques, optimization strategies, and testing frameworks. By examining these topics, listeners will understand how to effectively use PyTorch for control systems and decision-making applications.
// Bio
Vincent Moens is a research engineer on the PyTorch core team at Meta, based in London. As the maintainer of TorchRL (https://github.com/pytorch/rl) and TensorDict (https://github.com/pytorch/tensordict), Vincent plays a key role in supporting the decision-making community within the PyTorch ecosystem.
Alongside his technical role in the PyTorch community, Vincent also actively contributes to AI-related research projects.
Before joining Meta, Vincent worked as an ML researcher at Huawei and AIG.
Vincent holds a Medical Degree and a PhD in Computational Neuroscience.
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Musical recommendation: https://open.spotify.com/artist/1Uff91EOsvd99rtAupatMP?si=jVkoFiq8Tmq0fqK_OIEglg
Website: github.com/vmoens
TorchRL: https://github.com/pytorch/rl
TensorDict: https://github.com/pytorch/tensordict
LinkedIn post: https://www.linkedin.com/posts/vincent-moens-9bb91972_join-the-tensordict-discord-server-activity-7189297643322253312-Wo9J?utm_source=share&utm_medium=member_desktop
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Vincent on LinkedIn: https://www.linkedin.com/in/mvi/
Matt Van Itallie is the founder and CEO of Sema. Prior to this, they were the Vice President of Customer Support and Customer Operations at Social Solutions.
AI-Driven Code: Navigating Due Diligence & Transparency in MLOps // MLOps Podcast #275 with Matt van Itallie, Founder and CEO of Sema.
// Abstract
Matt Van Itallie, founder and CEO of Sema, discusses how comprehensive codebase evaluations play a crucial role in MLOps and technical due diligence. He highlights the impact of Generative AI on code transparency and explains the Generative AI Bill of Materials (GBOM), which helps identify and manage risks in AI-generated code. This talk offers practical insights for technical and non-technical audiences, showing how proper diligence can enhance value and mitigate risks in machine learning operations.
// Bio
Matt Van Itallie is the Founder and CEO of Sema. He and his team have developed Comprehensive Codebase Scans, the most thorough and easily understandable assessment of a codebase and engineering organization. These scans are crucial for private equity and venture capital firms looking to make informed investment decisions. Sema has evaluated code within organizations that have a collective value of over $1 trillion. In 2023, Sema served 7 of the 9 largest global investors, along with market-leading strategic investors, private equity, and venture capital firms, providing them with critical insights.
In addition, Sema is at the forefront of Generative AI Code Transparency, which measures how much code created by GenAI is in a codebase. They are the inventors behind the Generative AI Bill of Materials (GBOM), an essential resource for investors to understand and mitigate risks associated with AI-generated code.
Before founding Sema, Matt was a Private Equity operating executive and a management consultant at McKinsey. He graduated from Harvard Law School and has had some interesting adventures, like hiking a third of the Appalachian Trail and biking from Boston to Seattle.
Full bio: https://alistar.fm/bio/matt-van-itallie
// MLOps Swag/Merch
https://shop.mlops.community/
// Related Links
Website: https://en.m.wikipedia.org/wiki/Michael_Gschwind
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Matt on LinkedIn: https://www.linkedin.com/in/mvi/
Dr. Michael Gschwind is a Director / Principal Engineer for PyTorch at Meta Platforms. At Meta, he led the rollout of GPU Inference for production services.
// MLOps Podcast #274 with Michael Gschwind, Software Engineer, Software Executive at Meta Platforms.
// Abstract
Explore the role in boosting model performance, on-device AI processing, and collaborations with tech giants like ARM and Apple. Michael shares his journey from gaming console accelerators to AI, emphasizing the power of community and innovation in driving advancements.
// Bio
Dr. Michael Gschwind is a Director / Principal Engineer for PyTorch at Meta Platforms. At Meta, he led the rollout of GPU Inference for production services. He led the development of MultiRay and Textray, the first deployment of LLMs at a scale exceeding a trillion queries per day shortly after its rollout. He created the strategy and led the implementation of PyTorch donation optimization with Better Transformers and Accelerated Transformers, bringing Flash Attention, PT2 compilation, and ExecuTorch into the mainstream for LLMs and GenAI models. Most recently, he led the enablement of large language models on-device AI with mobile and edge devices.
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://en.m.wikipedia.org/wiki/Michael_Gschwind
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Michael on LinkedIn: https://www.linkedin.com/in/michael-gschwind-3704222/?utm_source=share&utm_campaign=share_via&utm_content=profile&utm_medium=ios_app
Timestamps:
[00:00] Michael's preferred coffee
[00:21] Takeaways
[01:59] Please like, share, leave a review, and subscribe to our MLOps channels!
[02:10] Gaming to AI Accelerators
[11:34] Torch Chat goals
[18:53] Pytorch benchmarking and competitiveness
[21:28] Optimizing MLOps models
[24:52] GPU optimization tips
[29:36] Cloud vs On-device AI
[38:22] Abstraction across devices
[42:29] PyTorch developer experience
[45:33] AI and MLOps-related antipatterns
[48:33] When to optimize
[53:26] Efficient edge AI models
[56:57] Wrap up
//Abstract
In this segment, the Panel will dive into the evolving landscape of AI, where large language models (LLMs) power the next wave of intelligent agents. In this engaging panel, leading investors Meera (Redpoint), George (Sequoia), and Sandeep (Prosus Ventures) discuss the promise and pitfalls of AI in production. From transformative industry applications to the challenges of scalability, costs, and shifting business models, this session unpacks the metrics and insights shaping GenAI's future. Whether you're excited about AI's potential or wary of its complexities, this is a must-watch for anyone exploring the cutting edge of tech investment.
//Bio
Host: Paul van der Boor
Senior Director Data Science @ Prosus Group
Sandeep Bakshi
Head of Investments, Europe @ Prosus
Meera Clark
Principal @ Redpoint Ventures
George Robson
Partner @ Sequoia Capital
A Prosus | MLOps Community Production
Luke Marsden, is a passionate technology leader. Experienced in consultant, CEO, CTO, tech lead, product, sales, and engineering roles. Proven ability to conceive and execute a product vision from strategy to implementation, while iterating on product-market fit.
We Can All Be AI Engineers and We Can Do It with Open Source Models // MLOps Podcast #273 with Luke Marsden, CEO of HelixML.
// Abstract
In this podcast episode, Luke Marsden explores practical approaches to building Generative AI applications using open-source models and modern tools. Through real-world examples, Luke breaks down the key components of GenAI development, from model selection to knowledge and API integrations, while highlighting the data privacy advantages of open-source solutions.
// Bio
Hacker & entrepreneur. Founder at helix.ml. Career spanning DevOps, MLOps, and now LLMOps. Working on bringing business value to local, open-source LLMs.
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://helix.ml
About open source AI: https://blog.helix.ml/p/the-open-source-ai-revolution
Ratatat Cream on Chrome: https://open.spotify.com/track/3s25iX3minD5jORW4KpANZ?si=719b715154f64a5f
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Luke on LinkedIn: https://www.linkedin.com/in/luke-marsden-71b3789/
Top Podcasts
The Best New Comedy Podcast Right Now – June 2024The Best News Podcast Right Now – June 2024The Best New Business Podcast Right Now – June 2024The Best New Sports Podcast Right Now – June 2024The Best New True Crime Podcast Right Now – June 2024The Best New Joe Rogan Experience Podcast Right Now – June 20The Best New Dan Bongino Show Podcast Right Now – June 20The Best New Mark Levin Podcast – June 2024
"in Kaggle you normally see a 1-1 ratio of positive to negative examples" huh? has he ever done a Kaggle competition? this statement is totally off