King, consider trying the Spark Kernel ( https://github.com/ibm-et/spark-kernel) which will install Spark etc and provide you with a Spark/Scala Notebook in which you can develop your algorithm. The Vagrant installation described in https://github.com/ibm-et/spark-kernel/wiki/Vagrant-Development-Environment will have you quickly up and running on a single machine without having to manage the details of the system installations. There is a Docker version, https://github.com/ibm-et/spark-kernel/wiki/Using-the-Docker-Container-for-the-Spark-Kernel , if you prefer Docker. Regards, David
King sami <kgsam...@gmail.com> wrote on 02/06/2015 08:09:39 AM: > From: King sami <kgsam...@gmail.com> > To: user@spark.apache.org > Date: 02/06/2015 08:11 AM > Subject: Beginner in Spark > > Hi, > > I'm new in Spark, I'd like to install Spark with Scala. The aim is > to build a data processing system foor door events. > > the first step is install spark, scala, hdfs and other required tools. > the second is build the algorithm programm in Scala which can treat > a file of my data logs (events). > > Could you please help me to install the required tools: Spark, > Scala, HDF and tell me how can I execute my programm treating the entry file. > > Best regards,