Re: Beginner in Spark

2015-02-10 Thread Akhil Das
You may also go through these posts
https://docs.sigmoidanalytics.com/index.php/Spark_Installation

Thanks
Best Regards

On Fri, Feb 6, 2015 at 9:39 PM, King sami  wrote:

> Hi,
>
> I'm new in Spark, I'd like to install Spark with Scala. The aim is to
> build a data processing system foor door events.
>
> the first step is install spark, scala, hdfs and other required tools.
> the second is build the algorithm programm in Scala which can treat a file
> of my data logs (events).
>
> Could you please help me to install the required tools: Spark, Scala, HDF
> and tell me how can I execute my programm treating the entry file.
>
>
> Best regards,
>
>
>
>
>


Re: Beginner in Spark

2015-02-10 Thread prabeesh k
Refer this blog
<http://blog.prabeeshk.com/blog/2014/10/31/install-apache-spark-on-ubuntu-14-dot-04/>
for step by step installation of Spark on Ubuntu

On 7 February 2015 at 03:12, Matei Zaharia  wrote:

> You don't need HDFS or virtual machines to run Spark. You can just
> download it, unzip it and run it on your laptop. See
> http://spark.apache.org/docs/latest/index.html.
>
> Matei
>
>
> On Feb 6, 2015, at 2:58 PM, David Fallside  wrote:
>
> King, consider trying the Spark Kernel (
> https://github.com/ibm-et/spark-kernel) which will install Spark etc and
> provide you with a Spark/Scala Notebook in which you can develop your
> algorithm. The Vagrant installation described in
> https://github.com/ibm-et/spark-kernel/wiki/Vagrant-Development-Environment 
> will
> have you quickly up and running on a single machine without having to
> manage the details of the system installations. There is a Docker version,
> https://github.com/ibm-et/spark-kernel/wiki/Using-the-Docker-Container-for-the-Spark-Kernel,
> if you prefer Docker.
> Regards,
> David
>
>
> King sami  wrote on 02/06/2015 08:09:39 AM:
>
> > From: King sami 
> > To: user@spark.apache.org
> > Date: 02/06/2015 08:11 AM
> > Subject: Beginner in Spark
> >
> > Hi,
> >
> > I'm new in Spark, I'd like to install Spark with Scala. The aim is
> > to build a data processing system foor door events.
> >
> > the first step is install spark, scala, hdfs and other required tools.
> > the second is build the algorithm programm in Scala which can treat
> > a file of my data logs (events).
> >
> > Could you please help me to install the required tools: Spark,
> > Scala, HDF and tell me how can I execute my programm treating the entry
> file.
> >
> > Best regards,
>
>
>


Re: Beginner in Spark

2015-02-10 Thread King sami
2015-02-06 17:28 GMT+00:00 King sami :

> The purpose is to build a data processing system for door events. An event
> will describe a door unlocking
> with a badge system. This event will differentiate unlocking by somebody
> from the inside and by somebody
> from the outside.
>
> *Producing the events*:
> You will need a simulator capable of producing events at random intervals.
> Simulating 200 doors seems like
> a good number, but adapt it as you see fit to get relevant results. Make
> sure different doors have different
> patterns to make the analysis interesting.
>
> *Processing the events:*
> After having accumulated a certain amount of events (for example: a day),
> you will calculate statistics. To do
> this, you will use spark for your batch processing. You will extract:
> • most used door, less used door, door with most exits, door with most
> entrances
> • most and less busy moment (when people entered and exited a lot, or not
> at all)
> • less busy moment of the day
>
> *Hints:*
> • Spark is required: http://spark.apache.org
> • Coding in Scala is required.
> • Using HDFS for file storage is a plus.
>
> 2015-02-06 17:00 GMT+00:00 Nagesh sarvepalli 
> :
>
>> Hi,
>>
>> Here is the sequence I suggest. Feel free if you need further help.
>>
>> 1) You need to decide if you want to go with any particular distribution
>> of Hadoop (Cloudera / Hortonworks / MapR) or want to go for apache version
>> . Downloading Hadoop from Apache and integrating with various projects is
>> laborious (compared to distributions).  Also, you need to take care of
>> maintenance including version compatibility of various projects. Cloudera
>> Manager is the best when it comes to cluster installation and maintenance
>> but it is memory intensive. Cloud offerings (ex: from Microsoft) are even
>> much more simpler and hassle free when it comes to installation and
>> maintenance.
>>
>> 2) Depending on the server resources and the data size, you need to
>> decide on the HDFS cluster size (number of nodes). Ensure you have  the
>> right JDK version installed if you are installing Hadoop on your own.
>>
>> 3) Once Hadoop is installed, you need to download Scala from
>> scala-lang.org and then
>>
>> 4) Download and install spark from http://spark.apache.org/downloads.html
>>
>> Hope this helps to kick-start.
>>
>> Thanks & Regards
>> Nagesh
>> Cloudera Certified Hadoop Developer
>>
>> On Fri, Feb 6, 2015 at 4:09 PM, King sami  wrote:
>>
>>> Hi,
>>>
>>> I'm new in Spark, I'd like to install Spark with Scala. The aim is to
>>> build a data processing system foor door events.
>>>
>>> the first step is install spark, scala, hdfs and other required tools.
>>> the second is build the algorithm programm in Scala which can treat a
>>> file of my data logs (events).
>>>
>>> Could you please help me to install the required tools: Spark, Scala,
>>> HDF and tell me how can I execute my programm treating the entry file.
>>>
>>>
>>> Best regards,
>>>
>>>
>>>
>>>
>>>
>>
>


Re: Beginner in Spark

2015-02-06 Thread Matei Zaharia
You don't need HDFS or virtual machines to run Spark. You can just download it, 
unzip it and run it on your laptop. See 
http://spark.apache.org/docs/latest/index.html 
<http://spark.apache.org/docs/latest/index.html>.

Matei


> On Feb 6, 2015, at 2:58 PM, David Fallside  wrote:
> 
> King, consider trying the Spark Kernel 
> (https://github.com/ibm-et/spark-kernel 
> <https://github.com/ibm-et/spark-kernel>) which will install Spark etc and 
> provide you with a Spark/Scala Notebook in which you can develop your 
> algorithm. The Vagrant installation described in 
> https://github.com/ibm-et/spark-kernel/wiki/Vagrant-Development-Environment 
> <https://github.com/ibm-et/spark-kernel/wiki/Vagrant-Development-Environment> 
> will have you quickly up and running on a single machine without having to 
> manage the details of the system installations. There is a Docker version, 
> https://github.com/ibm-et/spark-kernel/wiki/Using-the-Docker-Container-for-the-Spark-Kernel
>  
> <https://github.com/ibm-et/spark-kernel/wiki/Using-the-Docker-Container-for-the-Spark-Kernel>,
>  if you prefer Docker.
> Regards,
> David
> 
> 
> King sami  wrote on 02/06/2015 08:09:39 AM:
> 
> > From: King sami 
> > To: user@spark.apache.org
> > Date: 02/06/2015 08:11 AM
> > Subject: Beginner in Spark
> > 
> > Hi,
> > 
> > I'm new in Spark, I'd like to install Spark with Scala. The aim is 
> > to build a data processing system foor door events. 
> > 
> > the first step is install spark, scala, hdfs and other required tools.
> > the second is build the algorithm programm in Scala which can treat 
> > a file of my data logs (events).
> > 
> > Could you please help me to install the required tools: Spark, 
> > Scala, HDF and tell me how can I execute my programm treating the entry 
> > file.
> > 
> > Best regards,
> 



Re: Beginner in Spark

2015-02-06 Thread David Fallside


King, consider trying the Spark Kernel (
https://github.com/ibm-et/spark-kernel) which will install Spark etc and
provide you with a Spark/Scala Notebook in which you can develop your
algorithm. The Vagrant installation described in
https://github.com/ibm-et/spark-kernel/wiki/Vagrant-Development-Environment
will have you quickly up and running on a single machine without having to
manage the details of the system installations. There is a Docker version,
https://github.com/ibm-et/spark-kernel/wiki/Using-the-Docker-Container-for-the-Spark-Kernel
, if you prefer Docker.
Regards,
David


King sami  wrote on 02/06/2015 08:09:39 AM:

> From: King sami 
> To: user@spark.apache.org
> Date: 02/06/2015 08:11 AM
> Subject: Beginner in Spark
>
> Hi,
>
> I'm new in Spark, I'd like to install Spark with Scala. The aim is
> to build a data processing system foor door events.
>
> the first step is install spark, scala, hdfs and other required tools.
> the second is build the algorithm programm in Scala which can treat
> a file of my data logs (events).
>
> Could you please help me to install the required tools: Spark,
> Scala, HDF and tell me how can I execute my programm treating the entry
file.
>
> Best regards,

Beginner in Spark

2015-02-06 Thread King sami
Hi,

I'm new in Spark, I'd like to install Spark with Scala. The aim is to build
a data processing system foor door events.

the first step is install spark, scala, hdfs and other required tools.
the second is build the algorithm programm in Scala which can treat a file
of my data logs (events).

Could you please help me to install the required tools: Spark, Scala, HDF
and tell me how can I execute my programm treating the entry file.


Best regards,