Главная
Study mode:
on
1
Intro
2
Data at the Large Hadron Collider
3
Analytics Platform @CERN
4
Hadoop and Spark Clusters at CERN
5
Performance Troubleshooting Goals
6
Performance Methodologies and Anti-Patterns Typical benchmark graph
7
Workload and Performance Data
8
Measuring Spark
9
Spark Instrumentation - Metrics
10
How to Gather Spark Task Metrics
11
Spark Metrics in REST API
12
Task Metrics in the Event Log
13
SparkMeasure - Getting Started
14
SparkMeasure, Usage Modes
15
Instrument Code with Spark Measure
16
Spark Metrics System • Spark is also instrumented using the Dropwizard/Codahale metrics library • Multiple sources (data providers)
17
Ingredients for a Spark Performance Dashboard
18
Assemble Dashboard Components
19
Spark Dashboard - Examples Graph: "number of active tasks" vs. time
20
Dashboard - Memory
21
Dashboard - Executor CPU Utilization Graph: "CPU utilization by executors' JVM" vs. time
22
Executor Plugins Extend Metrics • User-defined executor metrics, SPARK-28091, target Spark 3.0.0
23
Metrics from OS Monitoring
24
Data + Context = Insights
Description:
Explore performance troubleshooting techniques for Apache Spark in this 40-minute conference talk by Luca Canali from CERN. Dive into Spark's extensive metrics and instrumentation, including executor task metrics and the Dropwizard-based system. Learn how CERN's Hadoop and Spark service leverages these metrics for troubleshooting and measuring production workloads. Discover how to deploy a performance dashboard for Spark workloads and utilize sparkMeasure, a tool based on the Spark Listener interface. Gain insights into lessons learned and upcoming improvements in Apache Spark 3.0. Cover topics such as data analytics at the Large Hadron Collider, CERN's analytics platform, performance methodologies, and anti-patterns. Examine various ways to gather and analyze Spark metrics, including REST API and event logs. Explore the components of a Spark performance dashboard, including memory usage, executor CPU utilization, and user-defined metrics. Understand the importance of combining data with context to derive meaningful insights for optimizing Spark-based applications. Read more

Performance Troubleshooting Using Apache Spark Metrics - Databricks Talk

Databricks
Add to list
0:00 / 0:00