[ https://issues.apache.org/jira/browse/SPARK-16579?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15384725#comment-15384725 ]
Shivaram Venkataraman commented on SPARK-16579: ----------------------------------------------- Yeah - i think we want to put Spark in the cache dir and use that as SPARK_HOME. Also we can take in a hadoop version as an optional flag in sparkR.session and if the JARs dont match we can fix the download. > Add a spark install function > ---------------------------- > > Key: SPARK-16579 > URL: https://issues.apache.org/jira/browse/SPARK-16579 > Project: Spark > Issue Type: Sub-task > Components: SparkR > Reporter: Shivaram Venkataraman > Assignee: Junyang Qian > > As described in the design doc we need to introduce a function to install > Spark in case the user directly downloads SparkR from CRAN. > To do that we can introduce a install_spark function that takes in the > following arguments > {code} > hadoop_version > url_to_use # defaults to apache > local_dir # defaults to a cache dir > {code} > Further more I think we can automatically run this from sparkR.init if we > find Spark home and the JARs missing. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org