case class FetchFailed(bmAddress: BlockManagerId, shuffleId: Int, mapId: Long, mapIndex: Int, reduceId: Int, message: String) extends TaskFailedReason with Product with Serializable
Task failed to fetch shuffle data from a remote node. Probably means we have lost the remote executors the task is trying to fetch from, and thus need to rerun the previous stage.
- Annotations
- @DeveloperApi()
- Source
- TaskEndReason.scala
- Alphabetic
- By Inheritance
- FetchFailed
- Serializable
- Serializable
- Product
- Equals
- TaskFailedReason
- TaskEndReason
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
- new FetchFailed(bmAddress: BlockManagerId, shuffleId: Int, mapId: Long, mapIndex: Int, reduceId: Int, message: String)
Value Members
- val bmAddress: BlockManagerId
-
def
countTowardsTaskFailures: Boolean
Fetch failures lead to a different failure handling path: (1) we don't abort the stage after 4 task failures, instead we immediately go back to the stage which generated the map output, and regenerate the missing data.
Fetch failures lead to a different failure handling path: (1) we don't abort the stage after 4 task failures, instead we immediately go back to the stage which generated the map output, and regenerate the missing data. (2) we don't count fetch failures from executors excluded due to too many task failures, since presumably its not the fault of the executor where the task ran, but the executor which stored the data. This is especially important because we might rack up a bunch of fetch-failures in rapid succession, on all nodes of the cluster, due to one bad node.
- Definition Classes
- FetchFailed → TaskFailedReason
- val mapId: Long
- val mapIndex: Int
- val message: String
- val reduceId: Int
- val shuffleId: Int
-
def
toErrorString: String
Error message displayed in the web UI.
Error message displayed in the web UI.
- Definition Classes
- FetchFailed → TaskFailedReason