Confluent Developer ft. Tim Berglund, Adi Polak & Viktor Gamov

Confluent Developer ft. Tim Berglund, Adi Polak & Viktor Gamov

<p>Hi, we’re Tim Berglund, Adi Polak, and Viktor Gamov and we’re excited to bring you the Confluent Developer podcast (formerly “Streaming Audio.”) Our hand-crafted weekly episodes feature in-depth interviews with our community of software developers (actual human beings - not AI) talking about some of the most interesting challenges they’ve faced in their careers. We aim to explore the conditions that gave rise to each person’s technical hurdles, as well as how their experiences transformed their understanding and approach to building systems. </p><p><br /></p><p>Whether you’re a seasoned open source data streaming engineer, or just someone who’s interested in learning more about Apache Kafka®,  Apache Flink® and real-time data, we hope you’ll appreciate the stories, the discussion, and our effort to bring you a high-quality show worth your time.</p>

We're back! Welcome to the Confluent Developer Podcast.

Weekly episodes launching Sept. 22! | Hi, I'm Tim Berglund. It's been about four years since I've been podcasting at Confluent, and "Streaming Audio" has been on hiatus for a little more than two, but I've got great news: we are back! We're back with a new name, a new format, and new hosts. Welcome to the Confluent Developer Podcast, where we talk to software developers of all stripes about some of the most interesting problems they've solved in their career. I'll be joined by my ...

09-02
01:20

Apache Kafka 3.5 - Kafka Core, Connect, Streams, & Client Updates

Apache Kafka® 3.5 is here with the capability of previewing migrations between ZooKeeper clusters to KRaft mode. Follow along as Danica Fine highlights key release updates. Kafka Core: KIP-833 provides an updated timeline for KRaft.KIP-866 now is preview and allows migration from an existing ZooKeeper cluster to KRaft mode.KIP-900 introduces a way to bootstrap the KRaft controllers with SCRAM credentials.KIP-903 prevents a data loss scenario by preventing replicas with stale broker epochs fr...

06-15
11:25

A Special Announcement from Streaming Audio

After recording 64 episodes and featuring 58 amazing guests, the Streaming Audio podcast series has amassed over 130,000 plays on YouTube in the last year. We're extremely proud of these achievements and feel that it's time to take a well-deserved break. Streaming Audio will be taking a vacation! We want to express our gratitude to you, our valued listeners, for spending 10,000 hours with us on this incredible journey. Rest assured, we will be back with more episodes! In the meantime, feel f...

04-13
01:18

How to use Data Contracts for Long-Term Schema Management

Have you ever struggled with managing data long term, especially as the schema changes over time? In order to manage and leverage data across an organization, it’s essential to have well-defined guidelines and standards in place around data quality, enforcement, and data transfer. To get started, Abraham Leal (Customer Success Technical Architect, Confluent) suggests that organizations associate their Apache Kafka® data with a data contract (schema). A data contract is an agreement between a ...

03-21
57:28

How to use Python with Apache Kafka

Can you use Apache Kafka® and Python together? What’s the current state of Python support? And what are the best options to get started? In this episode, Dave Klein joins Kris to talk about all things Kafka and Python: the libraries, the tools, and the pros & cons. He also talks about the new course he just launched to support Python programmers entering the event-streaming world. Dave has been an active member of the Kafka community for many years and noticed that there were a lot of Ka...

03-14
31:57

Next-Gen Data Modeling, Integrity, and Governance with YODA

In this episode, Kris interviews Doron Porat, Director of Infrastructure at Yotpo, and Liran Yogev, Director of Engineering at ZipRecruiter (formerly at Yotpo), about their experiences and strategies in dealing with data modeling at scale. Yotpo has a vast and active data lake, comprising thousands of datasets that are processed by different engines, primarily Apache Spark™. They wanted to provide users with self-service tools for generating and utilizing data with maximum flexibility, but e...

03-07
55:55

Migrate Your Kafka Cluster with Minimal Downtime

Migrating Apache Kafka® clusters can be challenging, especially when moving large amounts of data while minimizing downtime. Michael Dunn (Solutions Architect, Confluent) has worked in the data space for many years, designing and managing systems to support high-volume applications. He has helped many organizations strategize, design, and implement successful Kafka cluster migrations between different environments. In this episode, Michael shares some tips about Kafka cluster migration with K...

03-01
01:01:30

Real-Time Data Transformation and Analytics with dbt Labs

dbt is known as being part of the Modern Data Stack for ELT processes. Being in the MDS, dbt Labs believes in having the best of breed for every part of the stack. Oftentimes folks are using an EL tool like Fivetran to pull data from the database into the warehouse, then using dbt to manage the transformations in the warehouse. Analysts can then build dashboards on top of that data, or execute tests. It’s possible for an analyst to adapt this process for use with a microservice application u...

02-22
43:41

What is the Future of Streaming Data?

What’s the next big thing in the future of streaming data? In this episode, Greg DeMichillie (VP of Product and Solutions Marketing, Confluent) talks to Kris about the future of stream processing in environments where the value of data lies in their ability to intercept and interpret data. Greg explains that organizations typically focus on the infrastructure containers themselves, and not on the thousands of data connections that form within. When they finally realize that they don't have a...

02-15
41:29

What can Apache Kafka Developers learn from Online Gaming?

What can online gaming teach us about making large-scale event management more collaborative in real-time? Ben Gamble (Developer Relations Manager, Aiven) has come to the world of real-time event streaming from an usual source: the video games industry. And if you stop to think about it, modern online games are complex, distributed real-time data systems with decades of innovative techniques to teach us. In this episode, Ben talks with Kris about integrating gaming concepts with Apache...

02-08
55:32

Apache Kafka 3.4 - New Features & Improvements

Apache Kafka® 3.4 is released! In this special episode, Danica Fine (Senior Developer Advocate, Confluent), shares highlights of the Apache Kafka 3.4 release. This release introduces new KIPs in Kafka Core, Kafka Streams, and Kafka Connect. In Kafka Core: KIP-792 expands the metadata each group member passes to the group leader in its JoinGroup subscription to include the highest stable generation that consumer was a part of. KIP-830 includes a new configuration setting that allows you ...

02-07
05:13

How to use OpenTelemetry to Trace and Monitor Apache Kafka Systems

How can you use OpenTelemetry to gain insight into your Apache Kafka® event systems? Roman Kolesnev, Staff Customer Innovation Engineer at Confluent, is a member of the Customer Solutions & Innovation Division Labs team working to build business-critical OpenTelemetry applications so companies can see what’s happening inside their data pipelines. In this episode, Roman joins Kris to discuss tracing and monitoring in distributed systems using OpenTelemetry. He talks about how monitoring ea...

02-01
50:01

What is Data Democratization and Why is it Important?

Data democratization allows everyone in an organization to have access to the data they need, and the necessary tools needed to use this data effectively. In short, data democratization enables better business decisions. In this episode, Rama Ryali, a Senior IT and Data Executive, chats with Kris Jenkins about the importance of data democratization in modern systems. Rama explains that tech has unprecedented control over data and ignores basic business needs. Tech’s influence has larg...

01-26
47:27

Git for Data: Managing Data like Code with lakeFS

Is it possible to manage and test data like code? lakeFS is an open-source data version control tool that transforms object storage into Git-like repositories, offering teams a way to use the same workflows for code and data. In this episode, Kris sits down with guest Adi Polak, VP of DevX at Treeverse, to discuss how lakeFS can be used to facilitate better management and testing of data. At its core, lakeFS provides teams with better data management. A theoretical data engineer on a large t...

01-19
30:42

Using Kafka-Leader-Election to Improve Scalability and Performance

How does leader election work in Apache Kafka®? For the past 2 ½ years, Adithya Chandra, Staff Software Engineer at Confluent, has been working on Kafka scalability and performance, specifically partition leader election. In this episode, he gives Kris Jenkins a deep dive into the power of leader election in Kafka replication, why we need it, how it works, what can go wrong, and how it's being improved. Adithya explains that you can configure a certain number of replicas to be distributed ac...

01-12
51:06

Real-Time Machine Learning and Smarter AI with Data Streaming

Are bad customer experiences really just data integration problems? Can real-time data streaming and machine learning be democratized in order to deliver a better customer experience? Airy, an open-source data-streaming platform, uses Apache Kafka® to help business teams deliver better results to their customers. In this episode, Airy CEO and co-founder Steffen Hoellinger explains how his company is expanding the reach of stream-processing tools and ideas beyond the world of programmers. Air...

01-05
38:56

The Present and Future of Stream Processing

The past year saw new trends emerge in the world of data streaming technologies, as well as some unexpected and novel use cases for Apache Kafka®. New reflections on the future of stream processing and when companies should adopt microservice architecture inspired several talks at this year’s industry conferences. In this episode, Kris is joined by his colleagues Danica Fine, Senior Developer Advocate, and Robin Moffatt, Principal Developer Advocate, for an end-of-year roundtable on this year...

12-28
31:19

Top 6 Worst Apache Kafka JIRA Bugs

Entomophiliac, Anna McDonald (Principal Customer Success Technical Architect, Confluent) has seen her fair share of Apache Kafka® bugs. For her annual holiday roundup of the most noteworthy Kafka bugs, Anna tells Kris Jenkins about some of the scariest, most surprising, and most enlightening corner cases that make you ask, “Ah, so that’s how it really works?” She shares a lot of interesting details about how batching works, the replication protocol, how Kafka’s networking stack dances with L...

12-21
01:10:58

Learn How Stream-Processing Works The Simplest Way Possible

Could you explain Apache Kafka® in ways that a small child could understand? When Mitch Seymour, author of Mastering Kafka Streams and ksqlDB, wanted a way to communicate the basics of Kafka and event-based stream processing, he decided to author a children’s book on the subject, but it turned into something with a far broader appeal. Mitch conceived the idea while writing a traditional manuscript for engineers and technicians interested in building stream processing applications. He wished ...

12-20
31:29

Building and Designing Events and Event Streams with Apache Kafka

What are the key factors to consider when developing event-driven architecture? When properly designed, events can connect existing systems with a common language and allow data exchange in near real time. They also help reduce complexity by providing a single source of truth that eliminates the need to synchronize data between different services or applications. They enable dynamic behavior, allowing each service or application to respond quickly to changes in its environment. Using events, ...

12-15
53:06
Remember those childhood dreams where you’d build the most fantastic machine out of LEGOs, only to find out that real-life engineering is much harder? Double Cloud is here to make those dreams come true—but with data, not LEGOs. Imagine constructing an entire data pipeline in just one day. That’s right—one day. Check out https://double.cloud/services/managed-kafka/. Double Cloud’s platform is like a giant box of digital LEGOs, offering managed services like ClickHouse and Kafka, all while making sure everything fits together perfectly. Their no-code ELT tool is the magic brick that snaps everything into place, letting you sync data in real time without breaking a sweat. And maintenance? It’s like having a team of LEGO master builders on call, handling sharding, backups, and scaling, so you can focus on the fun part—innovating. Worried about security? Double Cloud has that covered too, with encryption and compliance so tight, it’s like locking your LEGOs in a vault. Curious about their

Chad Rourke

Remember those childhood dreams where you’d build the most fantastic machine out of LEGOs, only to find out that real-life engineering is much harder? Double Cloud is here to make those dreams come true—but with data, not LEGOs. Imagine constructing an entire data pipeline in just one day. That’s right—one day. Check out https://double.cloud/services/managed-kafka/. Double Cloud’s platform is like a giant box of digital LEGOs, offering managed services like ClickHouse and Kafka, all while making sure everything fits together perfectly. Their no-code ELT tool is the magic brick that snaps everything into place, letting you sync data in real time without breaking a sweat. And maintenance? It’s like having a team of LEGO master builders on call, handling sharding, backups, and scaling, so you can focus on the fun part—innovating. Worried about security? Double Cloud has that covered too, with encryption and compliance so tight, it’s like locking your LEGOs in a vault. Curious about their

08-22 Reply
Remember those screeching modem sounds? The excruciating wait for a single image to load? The 90s internet was like a one-lane highway for information. Today, it's a ten-lane superhighway with self-driving cars! We can download movies in minutes, stream music on-demand, and connect with anyone in the world instantly. It's a data explosion, and companies are scrambling to keep up. That's where services like https://double.cloud/ come in. Imagine Apache Kafka® as a super efficient traffic controller, sorting and delivering all this data at breakneck speed. It helps businesses handle the information overload and make real-time decisions – because in today's world, speed is king.

Chad Rourke

Remember those screeching modem sounds? The excruciating wait for a single image to load? The 90s internet was like a one-lane highway for information. Today, it's a ten-lane superhighway with self-driving cars! We can download movies in minutes, stream music on-demand, and connect with anyone in the world instantly. It's a data explosion, and companies are scrambling to keep up. That's where services like https://double.cloud/ come in. Imagine Apache Kafka® as a super efficient traffic controller, sorting and delivering all this data at breakneck speed. It helps businesses handle the information overload and make real-time decisions – because in today's world, speed is king.

05-07 Reply
Apache Kafka 3.2, the latest version of the popular distributed streaming platform, comes packed with new features and improvements that enhance its performance, scalability, and security. Some of the key updates include support for OAuth 2.0, improved error handling, and better handling of slow consumers.  But to truly harness the power of Apache Kafka, it's important to have the right platform to support your analytics applications. This is where Double Cloud comes in. DoubleCloud is a platform that specializes in building analytics applications with second-by-second processing using proven open-source technologies such as ClickHouse and Kafka.  With DoubleCloud, businesses can create powerful analytics applications that process real-time data with precision and speed. The platform's expertise in Apache Kafka and ClickHouse makes it a valuable partner for companies looking to optimize their data processing capabilities.  If you're interested in learning more about Double Cloud and ho

Chad Rourke

Apache Kafka 3.2, the latest version of the popular distributed streaming platform, comes packed with new features and improvements that enhance its performance, scalability, and security. Some of the key updates include support for OAuth 2.0, improved error handling, and better handling of slow consumers. But to truly harness the power of Apache Kafka, it's important to have the right platform to support your analytics applications. This is where Double Cloud comes in. DoubleCloud is a platform that specializes in building analytics applications with second-by-second processing using proven open-source technologies such as ClickHouse and Kafka. With DoubleCloud, businesses can create powerful analytics applications that process real-time data with precision and speed. The platform's expertise in Apache Kafka and ClickHouse makes it a valuable partner for companies looking to optimize their data processing capabilities. If you're interested in learning more about Double Cloud and ho

04-27 Reply

Recommend Channels