On Feb 18, 2009, at 10:37 AM, Amin Astaneh wrote:

Lukáš-

Well, we have a graduate student that is using our facilities for a Masters' thesis in Map/Reduce. You guys are generating topics in computer science research.

What do we need to do in order to get our documentation on the Hadoop pages?


You have a couple of options:
a) Put it on the Hadoop wiki (http://wiki.apache.org/hadoop/), for e.g. look at the ones which have docs on using Hadoop on EC2/S3. b) Open a jira (Create New Issue at https://issues.apache.org/jira/browse/HADOOP) and attach forrest-based documentation.

Arun

-Amin
Thanks guys,it is good to head that Hadoop is spreading... :-)
Regards,
Lukas

On Wed, Feb 18, 2009 at 5:24 PM, Steve Loughran <[email protected]> wrote:


Amin Astaneh wrote:


Lukáš-


Hi Amin,
I am not familiar with SGE, do you think you could tell me what did you
get
from this combination? What is the benefit of running Hadoop on SGE?



Sun Grid Engine is a distributed resource management platform for
supercomputing centers. We use it to allocate resources to a supercomputing task, such as requesting 32 processors to run a particular simulation. This mechanism is analogous to the scheduler on a multi-user OS. What I was able to accomplish was to turn Hadoop into an as-needed service. When you submit a job request to run Hadoop as the documentation describes, a Hadoop cluster of arbitrary size is instantiated depending on how many nodes were requested by generating a cluster configuration specific to that job request. This allows the Hadoop cluster to be deployed within the context of Gridengine, as well as being able to coexist with other running simulations on the
cluster.

To the researcher or user needing to run a mapreduce code, all they need to worry about is telling Hadoop to execute it as well as determining how many machines should be dedicated to the task. This benefit makes Hadoop very accessible to people since they don't need to worry about configuring a
cluster, SGE and it's helper scripts do it for them.

As Steve Loughran accurately commented, as of now we can only run one set of Hadoop slave processes per machine, due to the network binding issue. That problem is mitigated by configuring SGE to spread the slaves one per
machine automatically to avoid failures.


Only the Namenode and JobTracker need hard-coded/well-known port numbers,
the rest could all be done dynamically.

One thing SGE does offer over Xen-hosted images is better performance than
virtual machines, for both CPU  and storage, as  virtualised disk
performance can be awful, and even on the latest x86 parts, there is a
measurable hit from VM overheads.








Reply via email to