[
https://issues.apache.org/jira/browse/SPARK-18810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15736230#comment-15736230
]
Felix Cheung commented on SPARK-18810:
--------------------------------------
Also to expand on the earlier note above, I think the main thing to be able to
run existing tests, build vignettes and so on
- without having to change any code
or
- without having to manually call install.spark in a separate session first to
cache the spark jar
this is why I think it makes sense to have an environment override instead of
an API parameter switch.
> SparkR install.spark does not work for RCs, snapshots
> -----------------------------------------------------
>
> Key: SPARK-18810
> URL: https://issues.apache.org/jira/browse/SPARK-18810
> Project: Spark
> Issue Type: Bug
> Components: SparkR
> Affects Versions: 2.0.2, 2.1.0
> Reporter: Shivaram Venkataraman
>
> We publish source archives of the SparkR package now in RCs and in nightly
> snapshot builds. One of the problems that still remains is that
> `install.spark` does not work for these as it looks for the final Spark
> version to be present in the apache download mirrors.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]