So you created an EC2 instance with RStudio installed first, then installed 
Spark under that same username?  That makes sense, I just want to verify your 
work flow. 


Thank you again for your willingness to help! 




On Fri, Jun 26, 2015 at 10:13 AM -0700, "Shivaram Venkataraman" 
<shiva...@eecs.berkeley.edu> wrote:










I was using RStudio on the master node of the same cluster in the demo. However 
I had installed Spark under the user `rstudio` (i.e. /home/rstudio) and that 
will make the permissions work correctly. You will need to copy the config 
files from /root/spark/conf after installing Spark though and it might need 
some more manual tweaks.
ThanksShivaram  
On Fri, Jun 26, 2015 at 9:59 AM, Mark Stephenson <m...@redoakstrategic.com> 
wrote:
Thanks!
In your demo video, were you using RStudio to hit a separate EC2 Spark cluster? 
 I noticed that it appeared your browser that you were using EC2 at that time, 
so I was just curious.  It appears that might be one of the possible 
workarounds - fire up a separate EC2 instance with RStudio Server that 
initializes the spark context against a separate Spark cluster.

On Jun 26, 2015, at 11:46 AM, Shivaram Venkataraman 
<shiva...@eecs.berkeley.edu> wrote:
We don't have a documented way to use RStudio on EC2 right now. We have a 
ticket open at https://issues.apache.org/jira/browse/SPARK-8596 to discuss 
work-arounds and potential solutions for this. 
ThanksShivaram
On Fri, Jun 26, 2015 at 6:27 AM, RedOakMark <m...@redoakstrategic.com> wrote:
Good morning,



I am having a bit of trouble finalizing the installation and usage of the

newest Spark version 1.4.0, deploying to an Amazon EC2 instance and using

RStudio to run on top of it.



Using these instructions (

http://spark.apache.org/docs/latest/ec2-scripts.html

<http://spark.apache.org/docs/latest/ec2-scripts.html>  ) we can fire up an

EC2 instance (which we have been successful doing - we have gotten the

cluster to launch from the command line without an issue).  Then, I

installed RStudio Server on the same EC2 instance (the master) and

successfully logged into it (using the test/test user) through the web

browser.



This is where I get stuck - within RStudio, when I try to reference/find the

folder that SparkR was installed, to load the SparkR library and initialize

a SparkContext, I get permissions errors on the folders, or the library

cannot be found because I cannot find the folder in which the library is

sitting.



Has anyone successfully launched and utilized SparkR 1.4.0 in this way, with

RStudio Server running on top of the master instance?  Are we on the right

track, or should we manually launch a cluster and attempt to connect to it

from another instance running R?



Thank you in advance!



Mark







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-4-0-Using-SparkR-on-EC2-Instance-tp23506.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



---------------------------------------------------------------------

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to