Join us for Module 3: SQL and the Transaction Log - Tuesday, June 14
-Delta Lake SQL
-Time Travel
Transaction Log Fundamentals
This 3-part workshop is intended to teach you what Delta Lake is and how to use Apache Spark™ and Delta Lake in your data architectures for reliable large-scale distributed data pipelines. This course will show the features of Delta Lake that, alongside Spark SQL and Spark Structured Streaming, introduce ACID transactions and time travel (data versioning) to your ETL batch and streaming workloads. Slides, demos, exercises, and Q&A sessions should all together help you understand the concepts of the modern data lakehouse architecture.
Requirements
-Sign up for Databricks Community Edition
-Participants are recommended to have experience with Apache Spark SQL and Python (PySpark)
IT Freelancer for Apache Spark, Delta Lake, Apache Kafka & Kafka Streams
Jacek is an IT freelancer specializing in Apache Spark, Delta Lake, Apache Kafka (with brief forays into a wider data engineering space, e.g. Trino and ksqlDB, mostly during Warsaw Data Engineering meetups).
Jacek offers software development and consultancy services with very hands-on in-depth workshops and mentoring. He is best known by "The Internals Of" online books available free of …
Delta Lake
Developer Advocate
Marketing Manager