Discover
The Data Playbook Podcast
The Data Playbook Podcast
Author: Dataminded
Subscribed: 1Played: 24Subscribe
Share
Š Dataminded
Description
đď¸ The Data Playbook is a podcast where we aim to build a playbook for data leaders. We do that through a series of interviews with other data leaders, data practitioners and data experts. In each episode, we break down real-world data challenges: from building modern architectures and embracing Data Mesh to navigating cloud sovereignty, we help you make smarter decisions one play at a time.
19Â Episodes
Reverse
Kris Peeters sits down with Amaury Anciaux, founder of River Solutions, to tackle a painful reality for data leaders: critical decisions still depend on fragile Excel models.They explore why Excel wonât disappear, how River turns spreadsheets into visual, explainable and reliable decision models, and what happens when you bring data quality checks, testing and documentation into the analyst workflow.Topics include:Why 99% of models in organisations are still built in ExcelSilent errors, risk, and the real cost of debugging formulasVisual flow-based modelling and model maps inside ExcelBuilt-in checks for missing data, duplicates and broken lookupsHow AI copilots helped build River, and why AI wonât replace transparent modelsThe evolving role of analysts and managers in data-driven decisionsđ§ Listen to more episodes of The Data Playbook for real-world stories on data platforms, GenAI, data products and cloud independence from Europeâs leading data practitioners and leaders.đ More at https://www.dataminded.com/resources Chapters: 00:00 â Intro & episode setup00:45 â Amauryâs background & consulting career02:00 â The hidden reality of Excel decision models04:00 â Why âjust get it out of Excelâ doesnât scale05:10 â What River Solutions does in Excel06:40 â Visual model maps for explainable models08:40 â Removing formulas & adding data quality checks10:50 â Why Excel errors are so risky for big decisions13:15 â Who River is for: analysts, Excel gurus & managers16:05 â Why Amaury started River now & building with Copilot19:00 â Will AI copilots replace River and Excel modelling?22:51 â How River works as an Excel add-in (UX & interactivity)26:25 â How River changes the analyst role (less debugging, more thinking)28:10 â Roadmap: community, cloud, AI & connecting to data warehouses31:14 â Biggest lesson learned: software is easy, change is hard
What happens when a bank decides that AI and IP are so strategic they must be built in-house - then actually follows through for more than a decade?In this episode of The Data Playbook, Dr. Barak Chizi, Chief Data & Analytics Officer at KBC Group, joins Kris Peeters to reveal how KBC built one of Europeâs most mature AI organisations and what it took to bring Kate, their AI assistant, to life, and keep her evolving for 5 years.Youâll hear how KBC:Grew from early machine learning to 2,000+ AI use cases in productionDeveloped an AI-driven anti-money laundering platform and commercialised it for other banksScaled Kate, now celebrating 5 years and upgraded with GPT.Uses the U-model to govern AI safely from idea to productionKeeps ROI at the centre of every AI projectStays vendor-independent while still leveraging hyperscaler LLMsBuilds diverse, high-calibre AI teams with a rigorous recruitment approachExplores soft logic and modelling customer intent as the next frontier of financial AIIf you want to understand how to turn AI from experiments into a true competitive advantage, this conversation is your playbook.đ More at www.dataminded.com and subscribe to our channel.Show notes:The Foundation of Soft Logicđ https://link.springer.com/book/10.1007/978-3-031-58233-2Dan Ariely â Predictably Irrationalđ https://www.amazon.com/Predictably-Irrational-Revised-Expanded-Decisions/dp/0061353248/âąď¸ Chapters00:00 â Intro to The Data Playbook & todayâs guest01:15 â Barakâs backstory: 25 years in AI & high-dimensional data03:02 â What a CDAO does at KBC & enabling 24/7 AI-assisted service04:55 â Towards continuous, machine-supported customer journeys06:37 â The U-Model: KBCâs framework for data & AI projects08:35 â Flagship AI products, finite project lifecycle & retraining10:07 â Prioritising AI use cases across 5 countries12:31 â ROI mindset, conservative risk culture & data as an asset14:21 â Why KBC keeps AI in-house & limits external consultants18:17 â Beyond data warehouses: from reporting to prediction22:21 â AI-driven AML platform & the creation of SKY25:30 â Patents, AI IP and KBCâs competitive positioning27:25 â Generative AI at KBC since 2018 & early transformer experiments29:11 â Pragmatic tech choices: LLMs vs ML vs simple automation31:42 â Avoiding GenAI hype and focusing on customer value33:03 â Why KBC built Kate: 24/7 banking & impatient customers35:28 â From FAQ bot to execution engine: Kateâs end-to-end capabilities37:07 â Customer reactions, branches vs digital & Kateâs 2026 roadmap39:24 â Multi-LLM strategy, vendor independence & design partnerships40:44 â Inside Kateâs architecture: NLU, open source & KBC-built layers42:37 â Proactive AI: timing, context and personalised offers44:51 â Soft logic, consciousness & modelling customer intent49:19 â Building a diverse, 24-nationality AI team at KBC51:37 â Recruitment process, tests & how candidates are evaluated55:21 â What KBC looks for in modern data scientists57:15 â Lessons after 10 years at KBC & book recommendation
EU clouds without the hype. Niels Claeys (Partner & Lead Data Engineer at Dataminded, and our technical hiring lead) breaks down data sovereignty vs. Cloud Act, GDPR realities, and a portable, Kubernetes-first stack with Iceberg, Trino, and Airflow. We compare Scaleway, OVH, Exoscale, UpCloud, look at cost drivers, encryption/KMS, egress policies, and how to avoid vendor lock-in plus when best-of-breed beats all-in-one and why âkeep it simpleâ still wins.What youâll learn:When EU clouds make more sense than hyperscalers (and when they donât)Designing a portable platform: Terraform/Tofu for infra, Argo CD for appsTable formats 101: why Apache Iceberg over plain Parquet/CSVQuery layer choices: Trino for open SQL across object storage & DBsOrchestration in practice: Airflow patterns, dependencies, SLAsSecurity & governance: OPA for fine-grained policies, IAM, catalogsCost & ops: egress, managed services gaps, version lag, troubleshootingTeam skills: what to hire for, and the âhard questionsâ Niels asks in interviewsđ More at â www.dataminded.comâ â and subscribe!Chapters00:00 Intro & why EU clouds now04:40 Compliance & legal: GDPR, Cloud Act, sovereignty11:55 Platform blueprint: Kubernetes + Iceberg + Trino + Airflow20:30 Catalogs, OPA, IAM & access control27:10 EU providers deep dive: Scaleway, OVH, Exoscale, UpCloud36:20 Cost, encryption/KMS, egress & performance43:10 Best-of-breed vs all-in-one (and glue work)51:00 Getting started: IaC, Argo CD, day-2 ops56:40 Hiring: interview signals & practical takeawaysKeywordsEU cloud, European cloud providers, data sovereignty, GDPR, Cloud Act, Kubernetes data platform, Apache Iceberg, Trino, Airflow, vendor lock-in, OPA, Argo CD, Terraform, Exoscale, Scaleway, OVH, UpCloud
Belfius Insuranceâs Head of Data & AI, Hannes Heylen shares how his team scaled GenAI - from a fraud detection flywheel to âNestor,â a claims copilot that speeds summaries, completeness and coverage checks. We unpack AI agents in the claims flow, build-vs-buy decisions, and why content/data governance drives LLM quality. Plus: a pragmatic delivery mantra - make it work, then right, then cheap - for CIOs, CDOs and Heads of Data.What youâll learnHow to pick first AI cases that prove âŹROI (fraud models)Designing a claims copilot: summarization, completeness & coverage checksWhere AI agents fit (GenAI + ML + humans) across the claims flowBuild vs. buy in 2025: foundation models, vendor flexibility, cost controlContent/data governance as the make-or-break for LLM appsâFirst make it work, then right, then cheapâ: an AI operating model for CIO/CDOGuest: Hannes Heylen, Head of Data & AI, Belfius Insuranceđ More at â www.dataminded.comâ Chapters:00:00 Why AI now in financial services06:30 GenAIâs impact on text-heavy insurance processes18:40 AI agents across claims31:00 Governance > model tweaks38:00 Fraud detection: the ⏠case41:30 Claims copilot (âNestorâ) & lab-to-prod55:00 Lessons for CIOs/CDOsTopics: ROI-first use cases ⢠Claims automation ⢠AI agents (GenAI + ML + human-in-the-loop) ⢠Governance ⢠Vendor flexibility & costs
In this episode of The Data Playbook, we go inside imec, one of the worldâs leading semiconductor research institutes, to explore how they scale data governance, self-service, and innovation in one of the most data-intensive environments on Earth.Our guest, Wim Vancuyck, Manager of ICT for Data & Research Enablement, leads imecâs data strategy - bridging IT, researchers, and business to accelerate R&D through digital solutions. Wimâs mission: make imec a data-driven research organisation that turns raw measurements into insights and intellectual property faster and more securely.Wim explains how imec built a research data platform that empowers thousands of scientists through:Purpose-based access control, linking people, platforms, and data assetsFour self-service workbenches for Power BI, Data Engineering, Data Science & AI, and Application DevelopmentA clear platform vision built on efficiency, scalability, and reliabilityA governance model that supports both compliance and creativityA pragmatic stance on shadow IT: embrace, standardise, and professionalize itA bottom-up adoption strategy driven by early adopters and community engagementHe also discusses his evolution from technical architect to data leader, and what it takes to manage change in a 5,000-person R&D organisation, balancing technical depth with people leadership.đď¸ Guest: Wim Vancuyck - Manager ICT, Data & Research Enablement, imecđ More at: www.dataminded.com#DataPlaybook #imec #SemiconductorR&D #DataGovernance #PurposeBasedAccessControl #DataMesh #PlatformEngineering #SelfServiceAnalytics #CIO #CDO #DataLeadership #ResearchDataPlatform #DataStrategy
How do you turn governance from a bottleneck into a business accelerator - in a bank?In this episode of The Data Playbook, host Kris Peeters talks with Jan Mark Pleijsant, Senior Data Strategy & Governance Advisor at ABN AMRO Bank, about their move from 300+ dispersed data owners to 15 clear, business-aligned data domains and what it took to make federated governance work in a highly regulated environment.Youâll learn:Why âgovernance policeâ fails and how a federated model balances autonomy with oversight.How ABN AMRO defined domains by the nature of data (customer, loans, payments) rather than shifting business units driving stability across reorganisations.What âgovernance by designâ looks like in practice (policies, standards, data lineage, quality, clear ownership).The evolution from golden (source-aligned) datasets to consumer-ready data products (âpre-packaged saladsâ) that speed time-to-insight.Why executive sponsorship, domain maturity, and delivery discipline are the real success factors.How to measure domain maturity, prioritize critical reports (e.g., regulatory & exec dashboards), and avoid endless debates over definitions by making differences explicit.Who should listen: CIOs, CDOs, Heads of Data, and Data Leaders building scalable, compliant data platforms in complex organizations - especially in financial services.Chapters:00:00 Intro & context (POA Summit)03:12 Why banks need strong data governance07:45 Federation vs. centralization (and pitfalls)13:10 From 300 owners to 15 domains19:40 Designing domains that donât break with reorgs26:05 Data products vs. datasetsâwhat changed33:20 Governance by design: policy â product40:05 Measuring maturity & building momentum47:30 Risks, success factors, and the next 24 monthsđ More at https://www.dataminded.com/#DataGovernance #FederatedGovernance #DataDomains #DataProducts #DataLeadership #BankingData #ABNAMRO #DataStrategy
What does it take to scale data products across an organization? In this episode of The Data Playbook, we sit down with Simon Harrer, CEO and Co-Founder of Entropy Data, recorded live at the POA Summit 2025 in Stuttgart.Simon unpacks his journey from developer to founder, the creation of Data Mesh Manager, and why data contracts are becoming the backbone of modern data governance. We dive deep into:The evolution from consulting to product-based data companiesHow data products and contracts drive interoperabilityWhy AI and MCPs will redefine how data is shared and governedThe future of Data Mesh and the rise of data marketplacesA conversation packed with real-world lessons for Data Leaders, CIOs, and CDOs driving digital transformation.đ More at www.dataminded.comđ Resources from the episode:Entropy DataData Mesh ManagerData Contract CLIData Product MCPEntropy Spin-off Story
In this episode of the Data Playbook podcast, we dive into how publiq, the organization behind Belgiumâs largest cultural event database, is building RADAR, an AI-powered framework that enriches and structures event data at scale.Host Kris Peeters is joined by Sven Houtmeyers (CTO) and Elia Van Wolputte (Data Scientist) from Publiq, who share how their team uses LLMs, semantic parsing, and linked data to improve search, recommendations, and user experience, all while respecting publiq values like privacy, transparency, and digital inclusion.Topics covered:Why unstructured data is a challenge for cultural event discoveryHow RADAR leverages LLMs for smarter enrichment and entity resolutionFrom batch jobs to live services: Scaling across GCP and AWSDesigning ethical AI in the public sector: avoiding filter bubbles and over-personalizationThis is a behind-the-scenes look at how public organizations can use modern AI tools, not to manipulate users, but to empower them.đ More at â www.dataminded.com
đď¸ In this episode, host Kris Peeters talks with Jelle De Vleminck, consultant at Dataminded, about what it really takes to build a data platform that people actually want to use.Together, they explore:What separates a platform from just âa bunch of toolsâHow to reduce cognitive load for developersThe 5 biggest mistakes in platform designWhy adoption matters more than featuresHow data contracts, products, and cloud IDEs improve usabilityWhy enabling your users beats controlling themIf youâre building internal tooling or scaling data across teams, this episode is packed with practical insight.đ More at www.dataminded.com
In this episode of The Data Playbook, we explore what it really takes to turn AI into meaningful business impact.Host Kris Peeters talks with Joris Renkens, founder of AI product studio Guatavita, about how organizations can build AI solutions that truly work in practice.They discuss:Why innovation starts with the right problem, not the latest techHow to validate adoption early, before writing complex modelsWhat makes a high-performing AI product teamHow to manage technical debt while moving fastWhy flexible strategy matters more than fixed roadmapsđ Listen & subscribe on Spotifyđ More at www.dataminded.com
In this episode of The Data Playbook, we explore what it really takes to build high-performance data teams.Host Kris Peeters is joined by Rushil Daya, Senior Data Engineer at Dataminded, who shares practical lessons from years of leading successful data teams across industries.They discuss:The link between data success and business valueHow mentoring beats documentation when upskilling teamsWhy testing and CI/CD matter more than flashy toolsWhat makes stakeholder communication essentialWhy Agile is nothing without real feedback loopsđ Watch on YouTube: https://youtu.be/JEPPVakHfhAđ More at www.dataminded.com
In this episode of the Data Playbook podcast, we explore what it really takes to build sustainable, data-centric organizations, moving beyond tooling and dashboards toward lasting value.Host Kris Peeters is joined by Jonny Daenen (Knowledge Lead at Dataminded), who shares insights from years of helping organizations evolve their data strategy across sectors. Together, they discuss why data platforms, domain-owned data products, and people-first operating models are the foundations of modern data success.đ Topics covered:Why dashboards arenât enough for true data maturityFrom central chaos to federated teams and self-serviceThe role of governance in enabling deliveryHow AI and LLMs reshape data tooling, ownership & valueWhat âdata-centricâ really means , and why most fail to get theređ Hosted by Kris PeetersđĽ With Jonny Daenen, Datamindedđ Visit our website for more.
In this episode of The Data Playbook, we take a technical look at SQLMesh, a data transformation framework designed to improve the workflow and reliability of SQL-based data pipelines. Hosted by Kris Peeters, the episode features Michiel De Muynck, Senior Data Engineer at Dataminded, who provides a deep dive into SQLMeshâs internal mechanics, including its use of semantic analysis and isolated runtime environments.Michiel outlines how SQLMesh differentiates itself from tools like dbt by incorporating a semantic parser for SQL, enabling structural validation and more precise error diagnostics during pipeline development. He also explains the implementation of virtual data environments, which allow data engineers to stage, test, and version transformations without impacting production datasets, supporting safer iteration and deployment processes.đ§ Listen to more episodes on Spotify: Data Playbook Podcastđ Visit our website for more: Website Link
In this special episode of "The Data Playbook" podcast, recorded live at the Data Mesh Live Event in Antwerp, Kris Peeters speaks with Data Mesh pioneers Jacek Majchrzak and Andrew Jones. They explore how Data Mesh addresses critical challenges in data management, including data bottlenecks, governance, and decentralization. With years of experience in the field, both Jacek and Andrew share practical lessons from their journeys and offer actionable insights into implementing Data Mesh effectively.The conversation covers:Solving data bottlenecks through decentralized architecturesImproving governance with federated modelsAligning data strategy with business goals for impactful resultsUnderstanding the importance of incremental implementationMoving beyond "data silos" towards a more flexible, scalable approachJacek and Andrew provide real-world examples of how Data Mesh can transform your data infrastructure, sharing lessons on what works, what doesnât, and how to manage a successful Data Mesh implementation. If you're looking to overcome common data management challenges like governance and scalability, this episode is packed with practical advice.Books Referenced by Our Speakers:đ Data Mesh in Action by Jacek Majchrzak - https://a.co/d/4i5HUcYđ Driving Data Quality with Data Contracts by Andrew Jones - https://amzn.eu/d/aMQRFH1Stay tuned for more episodes on Data Mesh and other important topics in data architecture by following "The Data Playbook" on Spotify.đ§ Watch the full episode on Youtubeâ â đ Learn more on our website
Join us in this episode of The Data Playbook as we explore the sense and nonsense of data modeling with Jonas De Keuster, VP of Product at VaultSpeed. Jonas takes us through his journey in the world of data automation, discussing the role of data integration, data vaulting, and how modern data products are built using structured models. From dimensional modeling to the complexities of integrating data across multiple systems, Jonas shares practical insights into how organizations can scale their data operations.Topics covered include:Data Modeling Techniques: Data Vault vs. Dimensional ModelingData Automation and IntegrationBuilding Data Products with Scalable ModelsHow to Manage Data Changes and Evolving Business NeedsReal-world Challenges in Data PlatformsWhether you're leading a data team or just beginning your journey, this episode is a must-listen for anyone interested in the future of data architecture. Tune in for expert advice on building integrated data solutions that deliver real business value.To learn more, visit our website, or Watch more episodes on YouTube.
What do you do when GDPR forces your cloud project to stopâand years later, you need to go back? In this episode, Niels Melotte, Data Engineer at Dataminded, unpacks the journey of a government agency that migrated from the cloud to on-prem and then back to the cloud again.And hereâs the kicker: the Big Bang migration only took 14 hours. No downtime. No data loss. No angry users.đ In this episode, we discuss:Schrems II and why it sent European governments off the cloudAWS Nitro Enclaves & external key management for GDPR complianceWhy the on-prem platform failed to meet uptime guaranteesWhat âpurpose-based access controlâ means and why it mattersThe value of standardizing with dbt and StarburstHow data product thinking shaped the migration strategyLessons learned about trust, stakeholder communication, and platform maturityThis isnât a fluffy case study. Itâs a practical guide full of engineering tradeoffs, real-world headaches, and long-term lessons. A must-listen for data leaders, engineers, architects, and anyone dealing with sensitive data and complex infrastructure decisions.đ§ Want more episodes?Watch or Listen to all episodes of The Data Playbook on Spotify: đ https://open.spotify.com/show/78z3kdyBSKiURz1VnTVP9l?si=781abec722264306Show notes, episodes & resources:đ https://www.dataminded.com/resources/podcast#CloudMigration #PublicSector #GDPR #DataGovernance #AWS #DataPlatform #dbt #Starburst #BigBangMigration #TheDataPlaybook #Dataminded
In this episode of The Data Playbook, we dive deep into a critical, often-overlooked question: What does it mean to build sustainable data products? And no, weâre not just talking ESG dashboards or carbon reporting.đď¸ Host Kris Peeters is joined by Geert Verstraeten, a seasoned data scientist, founder of Python Predictions, and now a Co-Lead at The Data Forestâa consultancy that puts purpose and sustainability at the core of every data project.Over a candid and rich conversation, Geert shares:How he transitioned from startup founder to corporate leader and back to startup lifeâwith a sustainability mission.Why many data projects fail not because of tech, but because of missed alignment and poor adoption.What sustainable data products really are: tools that not only minimize environmental impact, but also stand the test of timeâwell-documented, actually used, and aligned with real business needs.Why selecting the right clients and projects is the first step toward impact, not just profit.How the Data Forest scores potential engagements using a unique framework: Head, Heart, and Hands.đĄ Along the way, youâll hear thought-provoking takes on:The role of documentation in sustainability.Why good data work isnât about building everything real-time or at scaleâespecially when youâre not Google.The paradox of GenAI and compute-heavy models in a world striving for tech responsibility.Whether you're a data engineer, architect, scientist, or team lead, this episode challenges you to rethink what a "good" data project looks like.đ If youâve ever built something technically brilliant that no one used, this episode is for you.Hit play to hear:What drives Geert and his team at The Data ForestHow to make better decisions on project scoping and client fitWhy data professionals need to talk to stakeholders much earlierâand more often
What Not to Build with AI: Avoiding the New Technical Debt in Data-Driven OrganizationsIn this episode of The Data Playbook, we explore a crucial and often overlooked question in the age of generative AI: not what to buildâbut what not to build.Host Kris Peeters (CEO of Dataminded) is joined by seasoned data leaders Pascal Brokmeier (Head of Engineering at Every Cure) and Tim SchrĂśder (AI & Data Transformation Lead in Biopharma), to talk about the dark side of unlimited AI capabilities: technical debt, fragmented systems, and innovation chaos.Topics we dive into:Why generative AI lowers the barrier to buildingâbut increases long-term complexityThe risks of treating LLMs as âmagical oraclesâ without governanceHow RAG systems became the default architectureâand why thatâs dangerousThe zoo vs. factory dilemma: how to balance AI experimentation with structureMaster data vs. knowledge graphs vs. embeddings â when and why each breaks downWhat Klarna got right (and wrong) by replacing SaaS tools with AI-generated internal appsThe growing importance of AI literacy, data maps, and platform thinkingReal-world examples of AI agents autonomously debugging systemsâand when thatâs terrifyingWe ask tough questions like:Are enterprises just building themselves into a new kind of mess, faster than ever before?Is the AI hype driving us toward âbuild now, regret laterâ?Should you really let every department build their own AI stack?Whether you're a data engineer, data architect, AI product lead, or a data strategist, this episode is a must-listen. Weâre cutting through the hype to figure out where the real value isâand where the future tech debt is quietly piling up.đ§ Key quote:"If you can't tell me why you're building it, maybe you shouldn't be building it at all."đĄ Tune in to learn how to stay smart, intentional, and strategic when it comes to building with AI.#TheDataPlaybook #DataEngineering #AIinBusiness #TechnicalDebt #RAG #LLMs #DataStrategy #EnterpriseAI #DataGovernance #DataLeadership #KnowledgeGraphs #GenerativeAI #AIinHealthcare #AIProduct #Dataminded
What does it really take to build a modern data architecture from the ground up? In our very first episode of The Data Playbook, host Kris Peeters, founder and CEO of Dataminded, sits down with Thorsten Foltz, seasoned data architect and engineer, to unpack what works (and what doesnât) when designing scalable, future-proof data platforms.With a focus on real-world tradeoffs, this episode explores:Cloud vs on-prem vs hybrid: how to choose the right infrastructureThe rise of Data Mesh and when it actually makes senseWhy fake news isnât just a media problemâitâs a data problem inside companies tooVendor lock-in, cloud sovereignty, and the growing relevance of European alternativesThe balance between open source and managed services: cost, control, and complexityWhy team culture and communication often make or break your data strategyWhat engineers can really expect from LLMs in the data stack (spoiler: they're not replacing data modeling any time soon)Whether you're a data engineer, architect, analyst, or tech leader, this conversation goes far beyond buzzwords. Youâll hear practical lessons, hard-earned insights, and a few uncomfortable truths about how companies actually manage data todayâand how they should rethink it for tomorrow.




