Abstract: As organizations attempt to optimize their technology investments and capabilities in cloud infrastructure and data processing frameworks, the potential for an isolated approach to solving data problems (such as streaming analytics) can arise. Open source to the rescue! In this session, we demonstrate how to get started with a streaming architecture using Delta Lake and Kafka on Azure. However, rather than ingesting data into Delta Lake via Spark jobs, we are going to use a couple of open-source libraries (yeah, no Spark required...some restrictions apply). By the end of this session you'll walk away with a playbook for up-leveling your streaming analytics architecture that includes:
· Lakehouse implementation via Delta Lake
· Kafka streaming via kafka-delta-ingest
· python binding via delta-rs
· Data storage on Azure via blob and ADLS
Bio: Gary Nakanelua is a professional technologist with over 17 years of experience and the author of Experiment or Expire. Gary is the Managing Director of Innovation at Blueprint, a data intelligence company based in Bellevue, WA. He’s responsible for the experimentation and creation of Blueprint’s transformative solutions and accelerators. With his diverse background, Gary brings a different perspective to problems that businesses are facing today to create quantifiable solutions driven through a high level of collaborative thought processing, strategic planning, and cannibalization.