Главная
Study mode:
on
1
What to expect
2
Project background
3
Azure Solution
4
ELT Framework
5
Don't move the ETL
6
Pause/Resume / Scale
7
Backup/Restore (granular)
8
Fun with HEAPS
9
Rebuild your HEAPS
10
Retry is unavoidable
11
Retry cos of random errors like these
12
04: Random flakiness is inherent
13
05: Idempotent (in Azure Data Factory...)
14
05: ADF Singleton Pattern
15
Concurrency is almost everything
16
Concurrency (one size fits all is bad)
17
Concurrency (use params in ADF)
18
Index and Stats Maintenance
19
Netezza has hidden columns
20
XX:And the rest
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Discover essential insights for building a large-scale data warehouse on Azure Dedicated SQL Pool in this 41-minute SQLBits conference talk. Learn from Steve Powell's experience in constructing a 60TB data warehouse, exploring crucial aspects of Azure Data Factory (ADF), data management, and automation. Gain valuable knowledge about the limitations of backups, hidden features of dedicated pools, and strategies to balance throughput and concurrency. Delve into topics such as ELT frameworks, data loading patterns, HEAP management, and the importance of retry mechanisms. Understand how to optimize concurrency, maintain indexes and statistics, and navigate the challenges of migrating from Netezza. Benefit from practical lessons on Azure Synapse Analytics, database engine management, and successful project delivery in big data environments.

7 Things I Wish I'd Known Before Building a 60TB Data Warehouse on a Dedicated Pool

SQLBits
Add to list