https://github.com/apache/spark
Revision 4259a28588a4dceb55d7bf1bf9327065dd751863 authored by Josh Rosen on 03 June 2016, 00:47:31 UTC, committed by Andrew Or on 03 June 2016, 00:47:31 UTC
If an RDD partition is cached on disk and the DiskStore file is lost, then reads of that cached partition will fail and the missing partition is supposed to be recomputed by a new task attempt. In the current BlockManager implementation, however, the missing file does not trigger any metadata updates / does not invalidate the cache, so subsequent task attempts will be scheduled on the same executor and the doomed read will be repeatedly retried, leading to repeated task failures and eventually a total job failure.

In order to fix this problem, the executor with the missing file needs to properly mark the corresponding block as missing so that it stops advertising itself as a cache location for that block.

This patch fixes this bug and adds an end-to-end regression test (in `FailureSuite`) and a set of unit tests (`in BlockManagerSuite`).

This is a branch-1.6 backport of #13473.

Author: Josh Rosen <joshrosen@databricks.com>

Closes #13479 from JoshRosen/handle-missing-cache-files-branch-1.6.
1 parent 0a13e4c
History
Tip revision: 4259a28588a4dceb55d7bf1bf9327065dd751863 authored by Josh Rosen on 03 June 2016, 00:47:31 UTC
[SPARK-15736][CORE][BRANCH-1.6] Gracefully handle loss of DiskStore files
Tip revision: 4259a28

README.md

back to top