DiscoverGnarly Data Waves by Dremio
Gnarly Data Waves by Dremio
Author: Dremio (The Open Data Lakehouse Platform)
Subscribed: 4Played: 27Subscribe
Share
© Dremio (The Open Data Lakehouse Platform)
Description
Gnarly Data Waves is a weekly show about the world of Data Analytics and Data Architecture. Learn about the technologies giving the company access to cutting-edge insights. If you work datasets, data warehouses, data lakes or data lakehouses, this show it for you!
Join us for our live recordings to participate in the Q&A:
dremio.com/events
Subscribe to the Dremio youtube channel on:
youtube.com/dremio
Take the Dremio Platform for a free test-drive:
https://www.dremio.com/test-drive/
Join us for our live recordings to participate in the Q&A:
dremio.com/events
Subscribe to the Dremio youtube channel on:
youtube.com/dremio
Take the Dremio Platform for a free test-drive:
https://www.dremio.com/test-drive/
62 Episodes
Reverse
This session will provide a comprehensive overview of Iceberg's journey, its current role within the data ecosystem, and the promising future it holds with the integration of Polaris (incubating). We will discuss how these technologies redefine table formats and catalog management, empowering organizations to efficiently manage and analyze large-scale data. Attendees will gain valuable insights into the evolving landscape, ensuring they remain at the forefront of innovation and continue to shape thought leadership in the data ecosystem.
Try Out Dremio on your Laptop: https://drmevn.fyi/youtubelakehouse102924
Legacy data platforms often fall short of the performance, processing and scaling requirements for robust AI/ML initiatives. This is especially true in complex multi-cloud (public, private, edge, airgapped) environments.
The combined power of MinIO and Dremio creates a data lakehouse platform that overcomes these challenges, delivering scalability, performance and efficiency to ensure successful AI initiatives.
Watch Brenna Buuck, Sr. Technical Evangelist at MinIO and Alex Merced, Sr. Technical Evangelist at Dremio provided insights on:
- AI Workflows: How a data lakehouse simplifies critical AI tasks like model training, refinement, feature selection and real-time inference for faster decisions
- Scalability and Performance: How a data lakehouse architecture scales seamlessly to meet the fast-growing demands of AI applications
- Data Management Efficiency: How a data lakehouse streamlines data management for IT teams, allowing them to focus on innovation
Dremio unveiled new features in our latest release that enhance the creation, performance, and management of Apache Iceberg data lakehouses.
You will learn how Dremio delivers market-leading SQL query and write performance, improved federated query security and management as well as streamlined data ingestion by delivering:
- Live Reflections on Iceberg tables that will accelerate performance, ensure up-to-date data and reduce management overhead.
- Result Set Caching that can accelerate query performance up to 28X
- Merge-on-Read that can enhance write and ingestion speed
- Auto Ingest Pipes that eliminate complex pipeline setup and maintenance
- User Impersonation for federated queries that allows for granular permissions, better access control, and user workload tracking
Organizations want to empower teams with data at their fingertips and in every part of their business. They want their teams to move quickly with data never as a bottleneck, but an accelerant to decision making —all without the curiosity tax too common in consumption-based cloud platforms. Dremio enables data teams to unify all of their disparate data, from Snowflake to Iceberg and other sources, by combining an intelligent semantic layer with a powerful SQL platform that eliminates silos, optimizes costs through intelligent query acceleration, and enables self-service analytics.
You will how Dremio enables Snowflake users to:
- Unify all of your data from Snowflake and all sources
- Optimize analytics costs and performance
- Use easy self-service analytics for faster time-to-insight
- Ensure Apache Iceberg native compatibility for future-proof data access
Learn how to master semantic layers with Dremio. We will provide a high-level overview of their purpose in modern analytics, showing how they act as a bridge between complex data sources and business users.
You’ll learn how semantic layers simplify data access, ensure consistency, and empower users to derive meaningful insights from data, regardless of their technical expertise.
- The definition and core purpose of a semantic layer in data analytics: How it acts as a bridge between complex data and business users, simplifying data access and interpretation.
- Key benefits and use cases of semantic layers: How they enable self-service analytics, ensure data consistency, and accelerate time-to-insight.
- How Dremio's semantic layer technology can transform your data strategy: Dremio makes it easier to manage and leverage your data for faster, data-driven decision-making.
Watch Vishnu Vardhan, Director of Product Management StorageGRID at NetApp and Alex Merced,Senior Technical Evangelist at Dremio, as they explore the future of data lakes and discover how NetApp and Dremio can revolutionize your analytics by delivering the next generation of lakehouse with Apache Iceberg.
Transitioning to a modern data lakehouse environment allows organizations to increase business insight, reduce management complexity, and lower overall TCO of their analytics environments. The growing adoption of Apache Iceberg is a key enabler for building the next generation lakehouse. Its robust feature set, coupled with an open ecosystem for analytics use cases, including ACID transactions, time travel, and schema evolution, continues to drive rapid adoption.
Vishnu and Alex will delve into market trends surrounding Iceberg, as well as key drivers for lakehouse adoption and modernization.
You will learn about:
- Iceberg adoption trends
- NetApp StorageGRID and its benefits
- The Dremio and NetApp data lakehouse solution
- Key Iceberg data lakehouse modernization use cases
- Customer examples
Watch and learn about Apache Iceberg. A 10 part web series designed to help you master Apache Iceberg.
https://hello.dremio.com/webcast-an-apache-iceberg-lakehouse-crash-course-reg.html?utm_medium=social-free&utm_source=youtube&utm_content=webcast-gdw-se-the-architecture-of-apache-iceberg-apache-hudi-and-delta-lake-intro&utm_campaign=webcast-gdw-se-the-architecture-of-apache-iceberg-apache-hudi-and-delta-lake-intro
The "An Apache Iceberg Lakehouse Crash Course" an in-depth webinar series designed to provide a comprehensive understanding of Apache Iceberg and its pivotal role in modern data lakehouse architectures.
Over the course of ten sessions, you'll explore a wide range of topics:
- foundational concepts like data lakehouses
- table formats to advanced features such as partitioning, optimization, and streaming with Apache Iceberg
- Each session will offer detailed insights into the architecture and capabilities of Apache Iceberg, alongside practical demonstrations of data ingestion using tools like Apache Spark and Dremio.
As the demand for data analytics grows, and with a decentralized approach at its core, Major Swedish manufacturer Scania needed to balance domain autonomy and alignment, while implementing a self-serve data & governance platform, coupled with a unified way of accessing data.
Discover how Scania addressed these challenges by adopting a data mesh strategy, and how using Dremio and Witboost has facilitated their journey. Learn about the cultural shifts, changes, and partnerships that are driving tangible business impacts. Additionally, gain insights and trends from Dremio’s Field CDO and the co-founder and CTO Witboost.
Ready to Get-Started: https://www.dremio.com/get-started/?utm_medium=website&utm_source=youtube&utm_content=gdw-od&utm_campaign=gdw-ep51
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-waves/?utm_medium=website&utm_source=youtube&utm_content=gdw-od&utm_campaign=gdw-ep51
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm_medium=website&utm_source=youtube&utm_content=gdw-od&utm_campaign=gdw-ep51
Events: https://www.dremio.com/events/?utm_medium=website&utm_source=youtube&utm_content=gdw-od&utm_campaign=gdw-ep51
Join us for a captivating recap of Subsurface 2024—the leading conference at the intersection of data engineering, open source technology, and modern data architecture.
This webinar will distill:
- highlights of the conference,
- curated clips of inspiring keynotes,
- insightful discussions on real-world data lakehouse implementations by industry leaders such as Nomura, NetApp, and Blue Cross.
- and deep dives into the transformative potential of open source projects like Apache Iceberg, Apache XTable, and Ibis.
Whether you missed the conference or want to revisit its most impactful moments, this webinar offers a unique opportunity to stay ahead of the curve in the rapidly evolving data landscape. Don't miss this chance to gain valuable insights from the experts and innovators who are shaping the future of data.
- Article on Dremio Auto-Ingest: https://www.dremio.com/blog/introducing-auto-ingest-pipes-event-driven-ingestion-made-easy/
- Article on Dremio and Hybrid Data Lakehouses (Vast, Netapp, Minio): https://www.dremio.com/blog/3-reasons-to-create-hybrid-apache-iceberg-data-lakehouses/
---------------------------------------------------------------
Get Hands-on with the Data Lakehouse
----------------------------------------------------------------
- Apache Iceberg Lakehouse on your Laptop: https://bit.ly/am-dremio-lakehouse-laptop
- SQLServer to Iceberg to Dashboard: https://bit.ly/am-sqlserver-dashboard
- MongoDB to Iceberg to Dashboard: https://bit.ly/am-mongodb-dashboard
- Postgres to Iceberg to Dashboard: https://bit.ly/am-postgres-to-dashboard
- MySQL to Iceberg to Dashboard: https://bit.ly/am-dremio-mysql-dashboard
- Elasticsearch to Iceberg to Dashboard: https://bit.ly/am-dremio-elastic
- Apache Druid to Iceberg to Dashboard: https://bit.ly/am-druid-dremio
- JSON/CSV/Parquet to Iceberg to Dashboard: https://bit.ly/am-json-csv-parquet-dremio
- From Kafka to Iceberg to Dremio: https://bit.ly/am-kafka-connect-dremio
- Lowering Snowflake Costs with Dremio: https://bit.ly/am-dremio-snowflake-spend
Watch Alex Merced, Senior Technical Evangelist at Dremio on "Optimize Analytics Workloads with Dremio + Snowflake". This session will delve into the key cost drivers of Snowflake and demonstrate how integrating Apache Iceberg and Dremio with a Data Lakehouse architecture can significantly reduce your data warehousing expenses. Discover strategies to optimize your data operations and achieve cost efficiency with cutting-edge technologies.
Ready to Get-Started: https://www.dremio.com/get-started/?u...
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa...
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm...
Events: https://www.dremio.com/events/?utm_me...
Dremio is making it easier than ever to build and manage an Apache Iceberg data lakehouse. Mark Shainman will share the new Dremio capabilities that help you achieve the fastest, most scalable, and easiest-to-manage lakehouse for analytics and AI.
In this video you’ll learn how:
- Dremio can help you accelerate Apache Iceberg adoption with seamless ingest
- Enhanced Reflections query acceleration can optimize performance and streamline management
- New capabilities continue to improve reliability, stability and scalability
- Dremio is delivering new capabilities to increase observability for ease of administration and management
We will embark on a journey that begins with a brief history of data analytics, tracing its development through the advent of the data lakehouse concept. This exploration sets the stage for a deeper understanding of the unique position Dremio occupies within this ecosystem, highlighting its innovative approach to bridging the gap between vast data lakes and the analysts striving to extract actionable insights.
The core of this presentation features a live demonstration, showcasing the end-to-end process of data connection and evaluation within the Dremio platform. Attendees will witness firsthand how Dremio facilitates a seamless flow of data from storage in data lakes to its transformation into a format ready for analysis, ultimately culminating in the delivery of valuable insights to analysts. This demonstration not only illustrates Dremio’s capabilities but also emphasizes its role in enabling a win-win scenario for both data engineers and analysts, by simplifying access to data and enhancing the efficiency of the analytics process.
In this video, we’ll cover:
- A short overview of the power of Dremio
- What is a semantic layer and why you need it
- Why Dremio is faster than anything else
Watch to gain a deeper understanding of the Dremio Data Lakehouse and discover how it can revolutionize your approach to data analytics, from enhancing data accessibility to streamlining the journey from raw data to actionable insights.
Ready to revolutionize your data management approach and learn how to maximize your environment with Dremio?
Watch Alex Merced in this workshop where he’ll guide you step-by-step through building a lakehouse on your laptop with Dremio, Nessie and Minio. This is a great opportunity to try out many of the best features Dremio offers.
You'll learn how to:
- Read and write Apache Iceberg tables on your object storage, cataloged by Nessie,
- Create views in the semantic layer,
- And much more
GDW Community Edition Workshop Description:
This hands-on workshop, participants will embark on a journey to construct their very own data lakehouse platform using their laptops. The workshop is designed to introduce and guide participants through the setup and utilization of three pivotal tools in the data lakehouse architecture: Dremio, Nessie, and Apache Iceberg. Each of these tools plays a crucial role in enabling the flexibility of data lakes with the efficiency and ease of use of data warehouses aiming to simplify and economize data management.
You will start by setting up a Docker environment to run all necessary services, including a notebook server, Nessie for catalog tracking with Git-like versioning, Minio as an S3-compatible storage layer, and Dremio as the core lakehouse platform. The workshop will provide a practical, step-by-step guide to federating data sources, organizing and documenting data, and performing queries with Dremio; tracking table changes and branching with Nessie; and creating, querying, and managing Apache Iceberg tables for an ACID-compliant data lakehouse.
Prerequisites for the workshop include having Docker installed on your laptop. You will be taken through the process of creating a docker-compose file to spin up the required services, configuring Dremio to connect with Nessie and Minio, and finally, executing SQL queries to manipulate and query data within their lakehouse.
This immersive session aims to not just educate but to empower attendees with the knowledge and tools needed to experiment with and implement their data lakehouse solutions. By the end of the workshop, participants will have a functional data lakehouse environment on their laptops, enabling them to explore further and apply what they have learned to real-world scenarios. Whether you're looking to improve your data management strategies or curious about the data lakehouse architecture, this workshop will provide a solid foundation and practical experience.
Data leaders are navigating the challenging landscape of enabling data-driven customer experiences and enhancing operational efficiency through analytics insights, all while meticulously managing budgets. Organizations leveraging cloud data warehouses, like Snowflake, often grapple with the complexities of unifying data analytics across diverse cloud and on-premise applications. The process involves significant costs, resources, and time to extract, rebuild, and integrate data for consumability.
Enter the data lakehouse - offering the potential to drastically reduce the total cost of ownership (TCO) associated with analytics.
In this video, you will gain insights into:
Key distinctions between traditional data warehouses and the innovative data lakehouse model.
How Dremio empowers organizations to slash analytics TCO by over 50%.
Uncovering hidden costs associated with data ingestion, storage, compute, business intelligence, and labor.
Simplifying self-service analytics through Dremio's unified lakehouse platform.
Watch Alex Merced, Developer Advocate at Dremio to explore the future of data management and discover how Dremio can revolutionize your analytics TCO, enabling you to do more with less.
Organizations aim to increase data access and lower the time it takes to gain insights, all while managing governance and controlling rising data costs.
Dremio’s unified lakehouse platform for self-service analytics enables data consumers to move fast while also reducing manual repetitive tasks and ticket overload for data engineers.
In this Gnarly Data Waves episode, you will learn:
- Overview of Dremio, what is it and why is it growing rapidly
- Proven use cases by some of the most demanding customers in the world
- Demonstration for how to rapidly get started and try it out
Ready to Get-Started: https://www.dremio.com/get-started/?u...
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa...
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm...
Events: https://www.dremio.com/events/?utm_me...#datalakehouse #analytics #datawarehouse #datalake #dataengineers #dataarchitects #governance #dremiocloud #opendatalakehouse #apachei
Traditional ETL processes are notorious for their complexity and cost inefficiencies. Join us as we introduce a game-changing virtual data pipeline approach with Dremio's next-gen DataOps, aimed at streamlining, simplifying, and fortifying your data pipelines to save time and reduce cost.
In this webinar, you'll gain insights into:
- Simplified Data Pipeline Management: How to use Dremio for data source branching, merging, and pipeline automation.
- Mastering Data Ingestion and Access: Learn how to curate data using virtual data marts accessed through a universal Semantic layer.
- Better Orchestration with dbt: Discover the benefits of orchestrating DML and view logic, optimizing data workflows.
- Elevating Data Quality: Learn techniques to automate lakehouse maintenance and improve data integrity.
S&P Global is a leading global financial services company headquartered in New York. It provides credit ratings, benchmarks, analytics, and workflow solutions in the global capital, commodity, and automotive markets. As a company, data is an essential asset across all of S&P Global’s solutions offerings.
Watch Tian de Klerk, Director of Business Intelligence, as he shares how they built a data lakehouse for FinOps analysis with Dremio Cloud on Microsoft Azure.
Tian will share about:
- The hidden costs of extracting operational data into BI cubes
- Simplifying traditional data engineering processes with Dremio’s zero-ETL lakehouse
- How Dremio’s semantic layer and query acceleration make self-service analytics easy for end users
In this session, Dremio and Microsoft will delve into the exciting developments surrounding the public preview launch of Dremio Cloud on Microsoft Azure. This presentation will provide a comprehensive exploration of how businesses are strategically operationalizing their data lakes, with a particular focus on unlocking the vast potential residing within Azure Storage. Attendees will gain valuable insights into the transformative journey toward harnessing the full benefits of a data lakehouse.
The discussion will guide participants through the myriad possibilities that emerge when leveraging Dremio Cloud seamlessly on Azure, offering a holistic approach to executing analytics pipelines. This integration eliminates the need for costly data warehouses, presenting a revolutionary paradigm shift. A step-by-step walkthrough will illuminate the process of landing data within the lakehouse, followed by seamlessly progressing data through a virtual semantic layer. This strategic approach adds significant business meaning and value, enhancing the overall utility of the data before it is surfaced to end users.
The session will also shed light on the noteworthy performance improvements and cost savings achieved by reducing data extract expenses associated with Power BI workloads. By embracing Dremio Cloud on Azure, organizations can elevate their analytical capabilities while optimizing operational costs, marking a pivotal advancement in the realm of data management and analytics. Join us as we explore the forefront of innovation in data lake operationalization and witness the tangible benefits of this dynamic integration.
Watch and learn how Jonny Dixon, Sr. Product Manager at Dremio and Hanno Borns, Principal Product Manager at Microsoft Azure will look into:
- Problems companies face with existing analytical architectures
- How Dremio and Microsoft Azure work together
- What Dremio Cloud on Azure is, and the value it provides
- How the Dremio Cloud on Azure solution works, with a demo
Dremio delivers no compromise lakehouse analytics for all of your data - and recent launches are making Dremio faster, more reliable, and more flexible than ever. Watch Mark Shainman, Product Marketing Manager at Dremio, Colleen Quinn, Product Marketing Manager at Dremio share and learn what's new in Dremio.
- New Gen-AI capabilities for automated data descriptions and labeling
- Dremio Cloud SaaS service now available on Microsoft Azure
- Advances to ensure 100% query reliability with no memory failures
- Expanded Apache Iceberg capabilities to streamline Iceberg adoption and improve performance
Ready to Get-Started: https://www.dremio.com/get-started/?u...
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa...
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm...
Events: https://www.dremio.com/events/?utm_me...#datalakehouse #analytics #datawarehouse #datalake #dataengineers #dataarchitects #governance #dremiocloud #opendatalakehouse #apacheiceberg #selfservice #enterprisedata #multitables #tableformat #microsoftazure #saas #automateddata #query #whatsnew #colleenquinn #markshainman #getstarted #etl #pipelines #genai #generativeai #parquet #json #tableau
Comments
Top Podcasts
The Best New Comedy Podcast Right Now – June 2024The Best News Podcast Right Now – June 2024The Best New Business Podcast Right Now – June 2024The Best New Sports Podcast Right Now – June 2024The Best New True Crime Podcast Right Now – June 2024The Best New Joe Rogan Experience Podcast Right Now – June 20The Best New Dan Bongino Show Podcast Right Now – June 20The Best New Mark Levin Podcast – June 2024
United States