@Diana - you can set sbt manually for your project by following the instructions here.
http://www.scala-sbt.org/release/docs/Getting-Started/Setup.html Manual Installation¶<http://www.scala-sbt.org/release/docs/Getting-Started/Setup.html#manual-installation> Manual installation requires downloading sbt-launch.jar<http://repo.typesafe.com/typesafe/ivy-releases/org.scala-sbt/sbt-launch/0.13.1/sbt-launch.jar> and creating a script to start it. Unix¶<http://www.scala-sbt.org/release/docs/Getting-Started/Setup.html#unix> Put sbt-launch.jar<http://repo.typesafe.com/typesafe/ivy-releases/org.scala-sbt/sbt-launch/0.13.1/sbt-launch.jar> in ~/bin. Create a script to run the jar, by creating ~/bin/sbt with these contents: SBT_OPTS="-Xms512M -Xmx1536M -Xss1M -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=256M"java $SBT_OPTS -jar `dirname $0`/sbt-launch.jar "$@" Make the script executable: $ chmod u+x ~/bin/sbt On Mon, Mar 24, 2014 at 4:30 PM, Diana Carroll <dcarr...@cloudera.com>wrote: > Yeah, that's exactly what I did. Unfortunately it doesn't work: > > $SPARK_HOME/sbt/sbt package > awk: cmd. line:1: fatal: cannot open file `./project/build.properties' for > reading (No such file or directory) > Attempting to fetch sbt > /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or > directory > /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or > directory > Our attempt to download sbt locally to sbt/sbt-launch-.jar failed. Please > install sbt manually from http://www.scala-sbt.org/ > > > > On Mon, Mar 24, 2014 at 4:25 PM, Ognen Duzlevski < > og...@plainvanillagames.com> wrote: > >> You can use any sbt on your machine, including the one that comes with >> spark. For example, try: >> >> ~/path_to_spark/sbt/sbt compile >> ~/path_to_spark/sbt/sbt run <arguments> >> >> Or you can just add that to your PATH by: >> >> export $PATH=$PATH:~/path_to_spark/sbt >> >> To make it permanent, you can add it to your ~/.bashrc or ~/.bash_profile >> or ??? depending on the system you are using. If you are on Windows, sorry, >> I can't offer any help there ;) >> >> Ognen >> >> >> On 3/24/14, 3:16 PM, Diana Carroll wrote: >> >> Thanks Ongen. >> >> Unfortunately I'm not able to follow your instructions either. In >> particular: >> >>> >>> sbt compile >>> sbt run <arguments if any> >> >> >> This doesn't work for me because there's no program on my path called >> "sbt". The instructions in the Quick Start guide are specific that I >> should call "$SPARK_HOME/sbt/sbt". I don't have any other executable on my >> system called "sbt". >> >> Did you download and install sbt separately? In following the Quick >> Start guide, that was not stated as a requirement, and I'm trying to run >> through the guide word for word. >> >> Diana >> >> >> On Mon, Mar 24, 2014 at 4:12 PM, Ognen Duzlevski < >> og...@plainvanillagames.com> wrote: >> >>> Diana, >>> >>> Anywhere on the filesystem you have read/write access (you need not be >>> in your spark home directory): >>> >>> mkdir myproject >>> cd myproject >>> mkdir project >>> mkdir target >>> mkdir -p src/main/scala >>> cp $mypath/$mymysource.scala src/main/scala/ >>> cp $mypath/myproject.sbt . >>> >>> Make sure that myproject.sbt has the following in it: >>> >>> name := "I NEED A NAME!" >>> >>> version := "I NEED A VERSION!" >>> >>> scalaVersion := "2.10.3" >>> >>> libraryDependencies += "org.apache.spark" % "spark-core_2.10" % >>> "0.9.0-incubating" >>> >>> If you will be using Hadoop/HDFS functionality you will need the below >>> line also >>> >>> libraryDependencies += "org.apache.hadoop" % "hadoop-client" % "2.2.0" >>> >>> The above assumes you are using Spark 0.9 and Scala 2.10.3. If you are >>> using 0.8.1 - adjust appropriately. >>> >>> That's it. Now you can do >>> >>> sbt compile >>> sbt run <arguments if any> >>> >>> You can also do >>> sbt package to produce a jar file of your code which you can then add to >>> the SparkContext at runtime. >>> >>> In a more complicated project you may need to have a bit more involved >>> hierarchy like com.github.dianacarroll which will then translate to >>> src/main/scala/com/github/dianacarroll/ where you can put your multiple >>> .scala files which will then have to be a part of a package >>> com.github.dianacarroll (you can just put that as your first line in each >>> of these scala files). I am new to Java/Scala so this is how I do it. More >>> educated Java/Scala programmers may tell you otherwise ;) >>> >>> You can get more complicated with the sbt project subrirectory but you >>> can read independently about sbt and what it can do, above is the bare >>> minimum. >>> >>> Let me know if that helped. >>> Ognen >>> >>> >>> On 3/24/14, 2:44 PM, Diana Carroll wrote: >>> >>>> Has anyone successfully followed the instructions on the Quick Start >>>> page of the Spark home page to run a "standalone" Scala application? I >>>> can't, and I figure I must be missing something obvious! >>>> >>>> I'm trying to follow the instructions here as close to "word for word" >>>> as possible: >>>> >>>> http://spark.apache.org/docs/latest/quick-start.html#a-standalone-app-in-scala >>>> >>>> 1. The instructions don't say what directory to create my test >>>> application in, but later I'm instructed to run "sbt/sbt" so I conclude >>>> that my working directory must be $SPARK_HOME. (Temporarily ignoring that >>>> it is a little weird to be working directly in the Spark distro.) >>>> >>>> 2. Create $SPARK_HOME/mysparktest/src/main/scala/SimpleApp.scala. >>>> Copy&paste in the code from the instructions exactly, replacing >>>> YOUR_SPARK_HOME with my spark home path. >>>> >>>> 3. Create $SPARK_HOME/mysparktest/simple.sbt. Copy&paste in the sbt >>>> file from the instructions >>>> >>>> 4. From the $SPARK_HOME I run "sbt/sbt package". It runs through the >>>> ENTIRE Spark project! This takes several minutes, and at the end, it says >>>> "Done packaging". unfortunately, there's nothing in the >>>> $SPARK_HOME/mysparktest/ folder other than what I already had there. >>>> >>>> (Just for fun, I also did what I thought was more logical, which is set >>>> my working directory to $SPARK_HOME/mysparktest, and but >>>> $SPARK_HOME/sbt/sbt package, but that was even less successful: I got an >>>> error: >>>> awk: cmd. line:1: fatal: cannot open file `./project/build.properties' >>>> for reading (No such file or directory) >>>> Attempting to fetch sbt >>>> /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or >>>> directory >>>> /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or >>>> directory >>>> Our attempt to download sbt locally to sbt/sbt-launch-.jar failed. >>>> Please install sbt manually from http://www.scala-sbt.org/ >>>> >>>> >>>> So, help? I'm sure these instructions work because people are >>>> following them every day, but I can't tell what they are supposed to do. >>>> >>>> Thanks! >>>> Diana >>>> >>> >>> >> >> -- >> "A distributed system is one in which the failure of a computer you didn't >> even know existed can render your own computer unusable" >> -- Leslie Lamport >> >> >