DiscovernCast: The Cloud Optimization Podcast from nOps#8 - Operating ML and GenAI at scale: the latest on Kubernetes, Karpenter and Bedrock
#8 - Operating ML and GenAI at scale: the latest on Kubernetes, Karpenter and Bedrock

#8 - Operating ML and GenAI at scale: the latest on Kubernetes, Karpenter and Bedrock

Update: 2023-10-11
Share

Description

In Episode 8, we’re joined by AWS Partner Solutions Architects Andrew Park and Mike McDonald to discuss the complexities and cost of running today’s ML and AI workloads on the cloud.


From anecdotes of the bad old days before container orchestration, our panelists take you to the present challenge of how to simplify efficient infrastructure operation — with the aim of freeing up Data Scientists and Engineers to focus on building and innovating.


Our panelists discuss the merits, pitfalls, and potential of various cost-optimizing tools and approaches (Ray, Karpenter, Spot, timeslicing) — key to addressing the demand for the expensive computing power generated by ML and AI models at scale.


Watch the full episode for:



  • The lowdown on AWS Bedrock and where it fits into the current stack of the latest AWS ML and AI offerings — how it works, use cases, the access it grants to new generative AI models



  • How Karpenter can make your life easy and save you SO much money (especially if you set-it-and-forget-it with nKS)



  • And hot takes on the controversial question: is ECS dead?!

Comments 
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

#8 - Operating ML and GenAI at scale: the latest on Kubernetes, Karpenter and Bedrock

#8 - Operating ML and GenAI at scale: the latest on Kubernetes, Karpenter and Bedrock

nOps