Package org.apache.spark.scheduler
Class StageInfo
Object
org.apache.spark.scheduler.StageInfo
:: DeveloperApi ::
Stores information about a stage to pass from the scheduler to SparkListeners.
-
Constructor Summary
ConstructorDescriptionStageInfo
(int stageId, int attemptId, String name, int numTasks, scala.collection.Seq<RDDInfo> rddInfos, scala.collection.Seq<Object> parentIds, String details, org.apache.spark.executor.TaskMetrics taskMetrics, scala.collection.Seq<scala.collection.Seq<TaskLocation>> taskLocalityPreferences, scala.Option<Object> shuffleDepId, int resourceProfileId, boolean isShufflePushEnabled, int shuffleMergerCount) -
Method Summary
Modifier and TypeMethodDescriptionscala.collection.mutable.HashMap<Object,
AccumulableInfo> Terminal values of accumulables updated during this stage, including all the user-defined accumulators.int
scala.Option<Object>
Time when the stage completed or when the stage was cancelled.details()
scala.Option<String>
If the stage failed, the reason why.static StageInfo
fromStage
(org.apache.spark.scheduler.Stage stage, int attemptId, scala.Option<Object> numTasks, org.apache.spark.executor.TaskMetrics taskMetrics, scala.collection.Seq<scala.collection.Seq<TaskLocation>> taskLocalityPreferences, int resourceProfileId) Construct a StageInfo from a Stage.name()
int
numTasks()
scala.collection.Seq<Object>
scala.collection.Seq<RDDInfo>
rddInfos()
int
void
stageFailed
(String reason) int
stageId()
scala.Option<Object>
When this stage was submitted from the DAGScheduler to a TaskScheduler.org.apache.spark.executor.TaskMetrics
-
Constructor Details
-
StageInfo
public StageInfo(int stageId, int attemptId, String name, int numTasks, scala.collection.Seq<RDDInfo> rddInfos, scala.collection.Seq<Object> parentIds, String details, org.apache.spark.executor.TaskMetrics taskMetrics, scala.collection.Seq<scala.collection.Seq<TaskLocation>> taskLocalityPreferences, scala.Option<Object> shuffleDepId, int resourceProfileId, boolean isShufflePushEnabled, int shuffleMergerCount)
-
-
Method Details
-
fromStage
public static StageInfo fromStage(org.apache.spark.scheduler.Stage stage, int attemptId, scala.Option<Object> numTasks, org.apache.spark.executor.TaskMetrics taskMetrics, scala.collection.Seq<scala.collection.Seq<TaskLocation>> taskLocalityPreferences, int resourceProfileId) Construct a StageInfo from a Stage.Each Stage is associated with one or many RDDs, with the boundary of a Stage marked by shuffle dependencies. Therefore, all ancestor RDDs related to this Stage's RDD through a sequence of narrow dependencies should also be associated with this Stage.
- Parameters:
stage
- (undocumented)attemptId
- (undocumented)numTasks
- (undocumented)taskMetrics
- (undocumented)taskLocalityPreferences
- (undocumented)resourceProfileId
- (undocumented)- Returns:
- (undocumented)
-
stageId
public int stageId() -
name
-
numTasks
public int numTasks() -
rddInfos
-
parentIds
-
details
-
taskMetrics
public org.apache.spark.executor.TaskMetrics taskMetrics() -
resourceProfileId
public int resourceProfileId() -
submissionTime
When this stage was submitted from the DAGScheduler to a TaskScheduler. -
completionTime
Time when the stage completed or when the stage was cancelled. -
failureReason
If the stage failed, the reason why. -
accumulables
Terminal values of accumulables updated during this stage, including all the user-defined accumulators.- Returns:
- (undocumented)
-
stageFailed
-
attemptNumber
public int attemptNumber()
-