You can add it to the spark packages i guess http://spark-packages.org/

Thanks
Best Regards

On Fri, Aug 14, 2015 at 1:45 PM, pishen tsai <pishe...@gmail.com> wrote:

> Sorry for previous line-breaking format, try to resend the mail again.
>
> I have written a sbt plugin called spark-deployer, which is able to deploy
> a standalone spark cluster on aws ec2 and submit jobs to it.
> https://github.com/pishen/spark-deployer
>
> Compared to current spark-ec2 script, this design may have several
> benefits (features):
> 1. All the code are written in Scala.
> 2. Just add one line in your project/plugins.sbt and you are ready to go.
> (You don't have to download the python code and store it at someplace.)
> 3. The whole development flow (write code for spark job, compile the code,
> launch the cluster, assembly and submit the job to master, terminate the
> cluster when the job is finished) can be done in sbt.
> 4. Support parallel deployment of the worker machines by Scala's Future.
> 5. Allow dynamically add or remove worker machines to/from the current
> cluster.
> 6. All the configurations are stored in a typesafe config file. You don't
> need to store it elsewhere and map the settings into spark-ec2's command
> line arguments.
> 7. The core library is separated from sbt plugin, hence it's possible to
> execute the deployment from an environment without sbt (only JVM is
> required).
> 8. Support adjustable ec2 root disk size, custom security groups, custom
> ami (can run on default Amazon ami), custom spark tarball, and VPC. (Well,
> most of these are also supported in spark-ec2 in slightly different form,
> just mention it anyway.)
>
> Since this project is still in its early stage, it lacks some features of
> spark-ec2 such as self-installed HDFS (we use s3 directly), stoppable
> cluster, ganglia, and the copy script.
> However, it's already usable for our company and we are trying to move our
> production spark projects from spark-ec2 to spark-deployer.
>
> Any suggestion, testing help, or pull request are highly appreciated.
>
> On top of that, I would like to contribute this project to Spark, maybe as
> another choice (suggestion link) alongside spark-ec2 on Spark's official
> documentation.
> Of course, before that, I have to make this project stable enough (strange
> errors just happen on aws api from time to time).
> I'm wondering if this kind of contribution is possible and is there any
> rule to follow or anyone to contact?
> (Maybe the source code will not be merged into spark's main repository,
> since I've noticed that spark-ec2 is also planning to move out.)
>
> Regards,
> Pishen Tsai
>
>

Reply via email to