https://github.com/apache/spark
Revision 371be22b1e56e6b56ad59130bcd9381a2ee4a014 authored by Santiago Saavedra on 10 November 2017, 18:57:58 UTC, committed by Shixiong Zhu on 10 November 2017, 18:58:10 UTC
## What changes were proposed in this pull request?

It seems that recovering from a checkpoint can replace the old
driver and executor IP addresses, as the workload can now be taking
place in a different cluster configuration. It follows that the
bindAddress for the master may also have changed. Thus we should not be
keeping the old one, and instead be added to the list of properties to
reset and recreate from the new environment.

## How was this patch tested?

This patch was tested via manual testing on AWS, using the experimental (not yet merged) Kubernetes scheduler, which uses bindAddress to bind to a Kubernetes service (and thus was how I first encountered the bug too), but it is not a code-path related to the scheduler and this may have slipped through when merging SPARK-4563.

Author: Santiago Saavedra <ssaavedra@openshine.com>

Closes #19427 from ssaavedra/fix-checkpointing-master.

(cherry picked from commit 5ebdcd185f2108a90e37a1aa4214c3b6c69a97a4)
Signed-off-by: Shixiong Zhu <zsxwing@gmail.com>
1 parent eb49c32
History
Tip revision: 371be22b1e56e6b56ad59130bcd9381a2ee4a014 authored by Santiago Saavedra on 10 November 2017, 18:57:58 UTC
[SPARK-22294][DEPLOY] Reset spark.driver.bindAddress when starting a Checkpoint
Tip revision: 371be22

README.md

back to top