Главная
Study mode:
on
1
Intro
2
Life as a data professional.
3
What is a Live Table?
4
Development vs Production
5
Declare LIVE Dependencies
6
Choosing pipeline boundaries
7
Pitfall: hard-code sources & destinations
8
Ensure correctness with Expectations
9
Expectations using the power of SQL
10
Using Python
11
Installing libraries with pip
12
Metaprogramming in python
13
Best Practice: Integrate using the event log
14
DLT Automates Failure Recovery
15
What is SparkTM Structured Streaming?
16
Using Spark Structured Streaming for ingestion
17
Use Delta for infinite retention
18
Partition recomputation
Description:
Explore a comprehensive talk on Delta Live Tables (DLT), a revolutionary ETL framework that simplifies data transformation and pipeline management. Learn how DLT incorporates modern software engineering practices to deliver reliable and trusted data pipelines at scale. Discover techniques for rapid innovation in pipeline development and maintenance, automation of administrative tasks, and improved visibility into pipeline operations. Gain insights into built-in quality controls and monitoring for accurate BI, data science, and ML. Understand how to implement simplified batch and streaming with self-optimizing and auto-scaling data pipelines. Delve into topics such as live table dependencies, pipeline boundaries, SQL expectations, Python integration, metaprogramming, event log integration, failure recovery automation, Spark Structured Streaming for ingestion, and Delta for infinite retention.

Delta Live Tables: Modern Software Engineering for ETL Pipelines

Databricks
Add to list
0:00 / 0:00