Ok, once I downloaded the pre built version, I created a directory for it and named Spark
When I try ./bin/start-all.sh It comes back with : no such file or directory When I try ./bin/spark-shell --master local[2] I get: no such file or directory Failed to find spark assembly, you need to build Spark before running this program Sent from my iPhone > On 8 Mar 2016, at 21:50, Cody Koeninger <c...@koeninger.org> wrote: > > That's what I'm saying, there is no "installing" necessary for > pre-built packages. Just unpack it and change directory into it. > > What happens when you do > > ./bin/spark-shell --master local[2] > > or > > ./bin/start-all.sh > > > >> On Tue, Mar 8, 2016 at 3:45 PM, Aida Tefera <aida1.tef...@gmail.com> wrote: >> Hi Cody, thanks for your reply >> >> I tried "sbt/sbt clean assembly" in the Terminal; somehow I still end up >> with errors. >> >> I have looked at the below links, doesn't give much detail on how to install >> it before executing "./sbin/start-master.sh" >> >> Thanks, >> >> Aida >> Sent from my iPhone >> >>> On 8 Mar 2016, at 19:02, Cody Koeninger <c...@koeninger.org> wrote: >>> >>> You said you downloaded a prebuilt version. >>> >>> You shouldn't have to mess with maven or building spark at all. All >>> you need is a jvm, which it looks like you already have installed. >>> >>> You should be able to follow the instructions at >>> >>> http://spark.apache.org/docs/latest/ >>> >>> and >>> >>> http://spark.apache.org/docs/latest/spark-standalone.html >>> >>> If you want standalone mode (master and several worker processes on >>> your machine) rather than local mode (single process on your machine), >>> you need to set up passwordless ssh to localhost >>> >>> http://stackoverflow.com/questions/7134535/setup-passphraseless-ssh-to-localhost-on-os-x >>> >>> >>> >>> On Tue, Mar 8, 2016 at 12:45 PM, Eduardo Costa Alfaia >>> <e.costaalf...@unibs.it> wrote: >>>> Hi Aida, >>>> The installation has detected a maven version 3.0.3. Update to 3.3.3 and >>>> try >>>> again. >>>> >>>> Il 08/Mar/2016 14:06, "Aida" <aida1.tef...@gmail.com> ha scritto: >>>>> >>>>> Hi all, >>>>> >>>>> Thanks everyone for your responses; really appreciate it. >>>>> >>>>> Eduardo - I tried your suggestions but ran into some issues, please see >>>>> below: >>>>> >>>>> ukdrfs01:Spark aidatefera$ cd spark-1.6.0 >>>>> ukdrfs01:spark-1.6.0 aidatefera$ build/mvn -DskipTests clean package >>>>> Using `mvn` from path: /usr/bin/mvn >>>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option >>>>> MaxPermSize=512M; >>>>> support was removed in 8.0 >>>>> [INFO] Scanning for projects... >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] Reactor Build Order: >>>>> [INFO] >>>>> [INFO] Spark Project Parent POM >>>>> [INFO] Spark Project Test Tags >>>>> [INFO] Spark Project Launcher >>>>> [INFO] Spark Project Networking >>>>> [INFO] Spark Project Shuffle Streaming Service >>>>> [INFO] Spark Project Unsafe >>>>> [INFO] Spark Project Core >>>>> [INFO] Spark Project Bagel >>>>> [INFO] Spark Project GraphX >>>>> [INFO] Spark Project Streaming >>>>> [INFO] Spark Project Catalyst >>>>> [INFO] Spark Project SQL >>>>> [INFO] Spark Project ML Library >>>>> [INFO] Spark Project Tools >>>>> [INFO] Spark Project Hive >>>>> [INFO] Spark Project Docker Integration Tests >>>>> [INFO] Spark Project REPL >>>>> [INFO] Spark Project Assembly >>>>> [INFO] Spark Project External Twitter >>>>> [INFO] Spark Project External Flume Sink >>>>> [INFO] Spark Project External Flume >>>>> [INFO] Spark Project External Flume Assembly >>>>> [INFO] Spark Project External MQTT >>>>> [INFO] Spark Project External MQTT Assembly >>>>> [INFO] Spark Project External ZeroMQ >>>>> [INFO] Spark Project External Kafka >>>>> [INFO] Spark Project Examples >>>>> [INFO] Spark Project External Kafka Assembly >>>>> [INFO] >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] Building Spark Project Parent POM 1.6.0 >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] >>>>> [INFO] --- maven-clean-plugin:2.6.1:clean (default-clean) @ >>>>> spark-parent_2.10 --- >>>>> [INFO] >>>>> [INFO] --- maven-enforcer-plugin:1.4:enforce (enforce-versions) @ >>>>> spark-parent_2.10 --- >>>>> [WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion >>>>> failed with message: >>>>> Detected Maven Version: 3.0.3 is not in the allowed range 3.3.3. >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] Reactor Summary: >>>>> [INFO] >>>>> [INFO] Spark Project Parent POM .......................... FAILURE >>>>> [0.821s] >>>>> [INFO] Spark Project Test Tags ........................... SKIPPED >>>>> [INFO] Spark Project Launcher ............................ SKIPPED >>>>> [INFO] Spark Project Networking .......................... SKIPPED >>>>> [INFO] Spark Project Shuffle Streaming Service ........... SKIPPED >>>>> [INFO] Spark Project Unsafe .............................. SKIPPED >>>>> [INFO] Spark Project Core ................................ SKIPPED >>>>> [INFO] Spark Project Bagel ............................... SKIPPED >>>>> [INFO] Spark Project GraphX .............................. SKIPPED >>>>> [INFO] Spark Project Streaming ........................... SKIPPED >>>>> [INFO] Spark Project Catalyst ............................ SKIPPED >>>>> [INFO] Spark Project SQL ................................. SKIPPED >>>>> [INFO] Spark Project ML Library .......................... SKIPPED >>>>> [INFO] Spark Project Tools ............................... SKIPPED >>>>> [INFO] Spark Project Hive ................................ SKIPPED >>>>> [INFO] Spark Project Docker Integration Tests ............ SKIPPED >>>>> [INFO] Spark Project REPL ................................ SKIPPED >>>>> [INFO] Spark Project Assembly ............................ SKIPPED >>>>> [INFO] Spark Project External Twitter .................... SKIPPED >>>>> [INFO] Spark Project External Flume Sink ................. SKIPPED >>>>> [INFO] Spark Project External Flume ...................... SKIPPED >>>>> [INFO] Spark Project External Flume Assembly ............. SKIPPED >>>>> [INFO] Spark Project External MQTT ....................... SKIPPED >>>>> [INFO] Spark Project External MQTT Assembly .............. SKIPPED >>>>> [INFO] Spark Project External ZeroMQ ..................... SKIPPED >>>>> [INFO] Spark Project External Kafka ...................... SKIPPED >>>>> [INFO] Spark Project Examples ............................ SKIPPED >>>>> [INFO] Spark Project External Kafka Assembly ............. SKIPPED >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] BUILD FAILURE >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [INFO] Total time: 1.745s >>>>> [INFO] Finished at: Tue Mar 08 18:01:48 GMT 2016 >>>>> [INFO] Final Memory: 19M/183M >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> [ERROR] Failed to execute goal >>>>> org.apache.maven.plugins:maven-enforcer-plugin:1.4:enforce >>>>> (enforce-versions) on project spark-parent_2.10: Some Enforcer rules have >>>>> failed. Look above for specific messages explaining why the rule failed. >>>>> -> >>>>> [Help 1] >>>>> [ERROR] >>>>> [ERROR] To see the full stack trace of the errors, re-run Maven with the >>>>> -e >>>>> switch. >>>>> [ERROR] Re-run Maven using the -X switch to enable full debug logging. >>>>> [ERROR] >>>>> [ERROR] For more information about the errors and possible solutions, >>>>> please >>>>> read the following articles: >>>>> [ERROR] [Help 1] >>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException >>>>> ukdrfs01:spark-1.6.0 aidatefera$ >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Installing-Spark-on-Mac-tp26397p26431.html >>>>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>> >>>> Informativa sulla Privacy: http://www.unibs.it/node/8155 --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org