Mark,
There is a setup price when using Hadoop, for each task a new JVM must
be spawned. On such a small scale, you won't see any good using MR.
J-D
On Mon, Apr 20, 2009 at 12:26 AM, Mark Kerzner markkerz...@gmail.com wrote:
Hi,
I ran a Hadoop MapReduce task in the local mode, reading and
that this
is the guideline - each task should take minutes.
Thank you,
Mark
On Mon, Apr 20, 2009 at 7:42 AM, Jean-Daniel Cryans
jdcry...@apache.orgwrote:
Mark,
There is a setup price when using Hadoop, for each task a new JVM must
be spawned. On such a small scale, you won't see any good using MR
Kevin,
I'm glad it worked for you.
We talked a bit about 5114 yesterday, any chance of trying 0.18 branch
on that same cluster without the socket timeout thing?
Thx,
J-D
On Wed, Apr 8, 2009 at 9:24 AM, Kevin Eppinger
keppin...@adknowledge.com wrote:
FYI: Problem fixed. It was apparently a
Sandhya,
You can specify the file to use for slaves so instead of start-all you
can start-dfs with the normal slave file and start-mapred with a
specified file on the command line.
J-D
On Wed, Apr 1, 2009 at 3:58 AM, Sandhya E sandhyabhas...@gmail.com wrote:
Hi
When the host is listed in
I found the solution here :
http://pero.blogs.aprilmayjune.org/2009/01/22/hadoop-and-linux-kernel-2627-epoll-limits/
J-D
On Fri, Mar 6, 2009 at 6:08 PM, Jean-Daniel Cryans jdcry...@apache.org wrote:
I know this one may be weird, but I'll give it a try. Thanks to anyone
reading this through
I know this one may be weird, but I'll give it a try. Thanks to anyone
reading this through.
Setup : hadoop-0.19.0 with hbase-0.19.0 on 10 nodes, quads with 8GB RAM, 2
disks. nofile limit is set at 30 000, xceivers at 1023,
dfs.datanode.socket.write.timeout at 0, dfs.datanode.handler.count at 9.
Philipp,
For HBase-related questions, please post to hbase-u...@hadoop.apache.org
Try importing commons-cli-2.0-SNAPSHOT.jar as well as any other jar in the
lib folder just to be sure you won't get any other missing class def error.
J-D
On Thu, Feb 12, 2009 at 6:32 PM, Philipp Dobrigkeit
not using EBS, just HDFS between the machines. As for
tasks,
there are 4 mappers and 0 reducers.
Richard J. Zak
-Original Message-
From: jdcry...@gmail.com [mailto:jdcry...@gmail.com] On Behalf Of
Jean-Daniel Cryans
Sent: Friday, January 23, 2009 13:24
To: core-user@hadoop.apache.org
Sid,
For such a small cluster, just put the Jobtracker and Namenode on the same
machine and the Tasktrackers and Datanodes in pairs on the other machines. I
can't think of anything else that would have an impact on performance for
you.
J-D
On Thu, Dec 11, 2008 at 6:20 PM, Siddharth Malhotra
Alex,
It's a HBase design goal to be able to answer to live non-relational
queries. True, up to 0.18, performance was not the priority but 0.19 will be
MUCH faster. Also, more and more websites use HBase in a production
environment, see http://wiki.apache.org/hadoop/Hbase/PoweredBy
Regards the
to do to eliminate NN
SPOF?
Thanks,
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
From: Jean-Daniel Cryans [EMAIL PROTECTED]
To: core-user@hadoop.apache.org
Sent: Tuesday, October 28, 2008 8:14:44 PM
Subject: Re: SecondaryNameNode
Tomislav.
Contrary to popular belief the secondary namenode does not provide failover,
it's only used to do what is described here :
http://hadoop.apache.org/core/docs/r0.18.1/hdfs_user_guide.html#Secondary+NameNode
So the term secondary does not mean a second one but is more like a
second part
Amit,
dfs.replication defines how many time each block of data will be replicated.
In your setup, if you're planning on keeping only one datanode, a value of 1
will reduce the overhead since keeping 2 or more copies of each block would
be useless if you lose your node. More info on how
Please use the HBase mailing list for HBase-related questions:
http://hadoop.apache.org/hbase/mailing_lists.html#Users
Regards your question, have you looked at
http://wiki.apache.org/hadoop/Hbase/HbaseRest ?
J-D
On Mon, Oct 6, 2008 at 12:05 AM, Trinh Tuan Cuong [EMAIL PROTECTED]
wrote:
Camilo,
See http://wiki.apache.org/hadoop/NameNode
And see the discussion NameNode Hardware specs started here:
http://www.mail-archive.com/core-user@hadoop.apache.org/msg04109.html
This should give you the basics.
Regards,
J-D
On Fri, Sep 5, 2008 at 10:31 PM, Camilo Gonzalez [EMAIL
Ryan,
I currently have a Hadoop/HBase setup that uses EBS. It works but using EBS
implied an additional overhead of configuration (too bad you can't spawn
instances with volumes already attached to it tho I'm sure that'll come).
Shutting down instances and bringing others up also requires more
Kevin,
Did you try changing the
dfs.datanode.dns.interface/dfs.datanode.dns.nameserver/mapred.tasktracker.dns.interface/mapred.tasktracker.dns.nameserver
parameters?
J-D
On Fri, Sep 5, 2008 at 8:14 PM, Kevin [EMAIL PROTECTED] wrote:
Hi,
The machines I am using each has multiple network
Hi,
I suggest that you read how data is stored in HDFS, see
http://hadoop.apache.org/core/docs/r0.18.0/hdfs_design.html
J-D
On Sat, Sep 6, 2008 at 12:11 AM, ZhiHong Fu [EMAIL PROTECTED] wrote:
hello. I'm a new user to hadoop. and Now I hava a problem in understanding
Hdfs.
In such a scene.
Hi wangxiaowei,
Just chmod the file to get execution rights. You should also use
hadoop-0.16.4 because this will be your next problem.
Finally, problems regarding HBase should be sent to it`s mailing list :
http://hadoop.apache.org/hbase/mailing_lists.html
Regards,
Jean-Daniel
2008/5/19
At the root of the source and it's called build.xml
Jean-Daniel
2008/4/9, Khalil Honsali [EMAIL PROTECTED]:
Mr. Jean-Daniel,
where is the ant script please?
On 10/04/2008, Jean-Daniel Cryans [EMAIL PROTECTED] wrote:
The ANT script works well also.
Jean-Daniel
2008/4/9, Khalil
The ANT script works well also.
Jean-Daniel
2008/4/9, Khalil Honsali [EMAIL PROTECTED]:
Hi,
With eclise it's easy, you just have to add it as a new project, make sure
you add all libraries in folder lib and should compile fine
There is also an eclipse plugin for running hadoop jobs directly
Hi,
Have you read : http://wiki.apache.org/hadoop/QuickStart
Stage 3, second dot?
Regards,
jdcryans
2008/2/15, Chandran, Sathish [EMAIL PROTECTED]:
Hi all,
Can you help me out the following?
Normally Hadoop takes the free disk spaces available from the machine.
But I want to
Hi,
I guess you're using the SVN version. Are you running your test on the
master node or remotely?
jdcryans
2008/2/3, Cass Costello [EMAIL PROTECTED]:
Hey all,
I'm just starting with both Hadoop and HBase. I've created a 3-node
cluster
- 1 master and 2 slaves. I've had some fun in the
23 matches
Mail list logo