Package org.apache.spark.scheduler
Class StatsReportListener
Object
org.apache.spark.scheduler.SparkListener
org.apache.spark.scheduler.StatsReportListener
- All Implemented Interfaces:
org.apache.spark.internal.Logging
,SparkListenerInterface
:: DeveloperApi ::
Simple SparkListener that logs a few summary statistics when each stage completes.
-
Nested Class Summary
Nested classes/interfaces inherited from interface org.apache.spark.internal.Logging
org.apache.spark.internal.Logging.SparkShellLoggingFilter
-
Constructor Summary
-
Method Summary
Modifier and TypeMethodDescriptionstatic scala.Option<org.apache.spark.util.Distribution>
extractDoubleDistribution
(scala.collection.Seq<scala.Tuple2<TaskInfo, org.apache.spark.executor.TaskMetrics>> taskInfoMetrics, scala.Function2<TaskInfo, org.apache.spark.executor.TaskMetrics, Object> getMetric) static scala.Option<org.apache.spark.util.Distribution>
extractLongDistribution
(scala.collection.Seq<scala.Tuple2<TaskInfo, org.apache.spark.executor.TaskMetrics>> taskInfoMetrics, scala.Function2<TaskInfo, org.apache.spark.executor.TaskMetrics, Object> getMetric) static long
hours()
static String
millisToString
(long ms) Reformat a time interval in milliseconds to a prettier format for outputstatic long
minutes()
void
onStageCompleted
(SparkListenerStageCompleted stageCompleted) Called when a stage completes successfully or fails, with information on the completed stage.void
onTaskEnd
(SparkListenerTaskEnd taskEnd) Called when a task endsstatic org.slf4j.Logger
static void
org$apache$spark$internal$Logging$$log__$eq
(org.slf4j.Logger x$1) static int[]
static String
static double[]
static long
seconds()
static void
showBytesDistribution
(String heading, org.apache.spark.util.Distribution dist) static void
showBytesDistribution
(String heading, scala.Function2<TaskInfo, org.apache.spark.executor.TaskMetrics, Object> getMetric, scala.collection.Seq<scala.Tuple2<TaskInfo, org.apache.spark.executor.TaskMetrics>> taskInfoMetrics) static void
showBytesDistribution
(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt) static void
showDistribution
(String heading, String format, scala.Function2<TaskInfo, org.apache.spark.executor.TaskMetrics, Object> getMetric, scala.collection.Seq<scala.Tuple2<TaskInfo, org.apache.spark.executor.TaskMetrics>> taskInfoMetrics) static void
showDistribution
(String heading, org.apache.spark.util.Distribution d, scala.Function1<Object, String> formatNumber) static void
showDistribution
(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt, String format) static void
showDistribution
(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt, scala.Function1<Object, String> formatNumber) static void
showMillisDistribution
(String heading, scala.Function2<TaskInfo, org.apache.spark.executor.TaskMetrics, Object> getMetric, scala.collection.Seq<scala.Tuple2<TaskInfo, org.apache.spark.executor.TaskMetrics>> taskInfoMetrics) static void
showMillisDistribution
(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt) Methods inherited from class org.apache.spark.scheduler.SparkListener
onApplicationEnd, onApplicationStart, onBlockManagerAdded, onBlockManagerRemoved, onBlockUpdated, onEnvironmentUpdate, onExecutorAdded, onExecutorBlacklisted, onExecutorBlacklistedForStage, onExecutorExcluded, onExecutorExcludedForStage, onExecutorMetricsUpdate, onExecutorRemoved, onExecutorUnblacklisted, onExecutorUnexcluded, onJobEnd, onJobStart, onNodeBlacklisted, onNodeBlacklistedForStage, onNodeExcluded, onNodeExcludedForStage, onNodeUnblacklisted, onNodeUnexcluded, onOtherEvent, onResourceProfileAdded, onSpeculativeTaskSubmitted, onStageExecutorMetrics, onStageSubmitted, onTaskGettingResult, onTaskStart, onUnpersistRDD, onUnschedulableTaskSetAdded, onUnschedulableTaskSetRemoved
Methods inherited from class java.lang.Object
equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
Methods inherited from interface org.apache.spark.internal.Logging
initializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, org$apache$spark$internal$Logging$$log_, org$apache$spark$internal$Logging$$log__$eq
-
Constructor Details
-
StatsReportListener
public StatsReportListener()
-
-
Method Details
-
percentiles
public static int[] percentiles() -
probabilities
public static double[] probabilities() -
percentilesHeader
-
extractDoubleDistribution
-
extractLongDistribution
-
showDistribution
-
showDistribution
-
showDistribution
-
showDistribution
-
showBytesDistribution
-
showBytesDistribution
public static void showBytesDistribution(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt) -
showBytesDistribution
-
showMillisDistribution
public static void showMillisDistribution(String heading, scala.Option<org.apache.spark.util.Distribution> dOpt) -
showMillisDistribution
-
seconds
public static long seconds() -
minutes
public static long minutes() -
hours
public static long hours() -
millisToString
Reformat a time interval in milliseconds to a prettier format for output- Parameters:
ms
- (undocumented)- Returns:
- (undocumented)
-
org$apache$spark$internal$Logging$$log_
public static org.slf4j.Logger org$apache$spark$internal$Logging$$log_() -
org$apache$spark$internal$Logging$$log__$eq
public static void org$apache$spark$internal$Logging$$log__$eq(org.slf4j.Logger x$1) -
onTaskEnd
Description copied from interface:SparkListenerInterface
Called when a task ends- Specified by:
onTaskEnd
in interfaceSparkListenerInterface
- Overrides:
onTaskEnd
in classSparkListener
- Parameters:
taskEnd
- (undocumented)
-
onStageCompleted
Description copied from interface:SparkListenerInterface
Called when a stage completes successfully or fails, with information on the completed stage.- Specified by:
onStageCompleted
in interfaceSparkListenerInterface
- Overrides:
onStageCompleted
in classSparkListener
- Parameters:
stageCompleted
- (undocumented)
-