Hey Ashish, Thanks a lot for your update. I am probably using the old script, and was unaware of a version option. Much props.
On Mon, Nov 25, 2013 at 9:41 PM, Ashish Rangole <[email protected]> wrote: > Hi Walrus theCat, > > We have been successfully using Spark 0.8 on EC2 ever since it was > released and we do this > several times a day. > > We use spark-ec2.py with the new version option (--spark-version=0.8.0), > to spin-up the Spark 0.8 cluster on ec2. > The key is to use the new spark-ec2.py and not the old one. > > The only change we had to make was to modify our imports to point to the > new apache > package names, again as indicated in Spark 0.8 release notes. > > What exactly are the errors you are seeing? > > > On Mon, Nov 25, 2013 at 9:52 PM, Walrus theCat <[email protected]>wrote: > >> Thanks Paco, but I have no problems running my application on EC2 when >> it's using Spark 0.7.3. What I'm looking for is a way to use Spark 0.8 on >> EC2. >> >> Cheers, >> >> >> On Mon, Nov 25, 2013 at 5:01 PM, Paco Nathan <[email protected]> wrote: >> >>> Not the answer to your specific question (official solution) but >>> https://elastic.mesosphere.io/ is a 3-step service to run Apache Mesos >>> atop EC2. Then there's a quick tutorial about running Apache Spark on that >>> cluster http://mesosphere.io/learn/run-spark-on-mesos/ >>> >>> >>> On Mon, Nov 25, 2013 at 4:57 PM, Walrus theCat >>> <[email protected]>wrote: >>> >>>> Hi, >>>> >>>> I just updated my imports and tried to run my app using Spark 0.8, but >>>> it breaks. The AMI's spark-shell says it's 0.7.3 or thereabouts, which is >>>> what my app previously used. What is the official, step-by-step solution >>>> to using Spark 0.8 on EC2? >>>> >>>> Thanks >>>> >>> >>> >> >
