https://github.com/apache/spark
Revision 498fb70522f9320632eb2dffd8e62ce767278387 authored by Sean Owen on 05 March 2019, 14:26:30 UTC, committed by Sean Owen on 05 March 2019, 14:27:03 UTC
## What changes were proposed in this pull request?

Spark apps do not need to package Spark. In fact it can cause problems in some cases. Our examples should show depending on Spark as a 'provided' dependency.

Packaging Spark makes the app much bigger by tens of megabytes. It can also bring in conflicting dependencies that wouldn't otherwise be a problem. https://issues.apache.org/jira/browse/SPARK-26146 was what reminded me of this.

## How was this patch tested?

Doc build

Closes #23938 from srowen/Provided.

Authored-by: Sean Owen <sean.owen@databricks.com>
Signed-off-by: Sean Owen <sean.owen@databricks.com>
(cherry picked from commit 39092236819da097e9c8a3b2fa975105f08ae5b9)
Signed-off-by: Sean Owen <sean.owen@databricks.com>
1 parent ae462b1
History
Tip revision: 498fb70522f9320632eb2dffd8e62ce767278387 authored by Sean Owen on 05 March 2019, 14:26:30 UTC
[MINOR][DOCS] Clarify that Spark apps should mark Spark as a 'provided' dependency, not package it
Tip revision: 498fb70

README.md

back to top