Brought to you by
Building Modern Data Pipelines with Mage, Confluent, ClickHouse, and Snowflake
This comprehensive 3-4 hour hands-on workshop is designed to take participants of all skill levels through the journey of building modern data pipelines. We'll start with batch processing and gradually move to real-time streaming, using real-world NSW Transport incident data. Learning Objectives By the end of this workshop, you will: ✅ Understand the difference between batch and streaming data processing ✅ Build a batch pipeline using Mage AI and Apache Iceberg ✅ Query data lakes using Starburst (replacing Athena) ✅ Create real-time streaming pipelines with Confluent Kafka ✅ Store and analyze streaming data in ClickHouse ✅ Build interactive dashboards with Snowflake and Streamlit ✅ Integrate AI/ML using Claude API and Snowflake Cortex AI ✅ Deploy production-ready data solutions 🏗️ What We'll Build Today Module 1: Batch Pipeline [NSW Transport API] → [Mage Batch Pipeline] → [Apache Iceberg] → [Starburst Analytics] Module 2: AI-Powered Analytics [Iceberg + Streaming Data] → [Snowflake] → [Streamlit App] → [Claude AI Assistant] Module 3: Real-Time Streaming [NSW Transport API] → [Mage Streaming] → [Confluent Kafka] → [ClickHouse] → [Real-time Dashboard]