Running balancer throws exceptions

2009-04-12 Thread Stas Oskin
Hi. When I run hadoop balancer, I get the following error: 09/04/12 10:28:46 INFO dfs.Balancer: Will move 3.02 GBbytes in this iteration Apr 12, 2009 10:28:46 AM 0 0 KB19.02 GB3.02 GB 09/04/12 10:28:46 INFO dfs.Balancer: Decided to move block

API: FSDataOutputStream create(Path f, boolean overwrite)

2009-04-12 Thread javateck javateck
Hi: I'm trying to use FSDataOutputStream create(Path f, boolean overwrite), I'm calling create(new Path(somePath), false), but creation still fails with IOException even when the file does not exist, can someone explain the behavior? thanks,

Exception Problem while running map-Reduce jobs :-Wrong FileSystem Exception

2009-04-12 Thread snehal nagmote
Hi, I am trying to create the Har files in hadoop 0.19.0, but on my hadoop cluster mapreduce jobs are not running , It gives very known exception 09/04/12 09:54:07 INFO mapred.FileInputFormat: Total input paths to process : 1 09/04/12 09:54:08 INFO mapred.JobClient: Running job:

Changing block size of hadoop

2009-04-12 Thread Rakhi Khatwani
Hi, I would like to know if it is feasbile to change the blocksize of Hadoop while map reduce jobs are executing? and if not would the following work? 1. stop map-reduce 2. stop-hbase 3. stop hadoop 4. change hadoop-sites.xml to reduce the blocksize 5. restart all whether the data in the

Cannot access Jobtracker and namenode

2009-04-12 Thread halilibrahimcakir
Hi I am new at hadoop. I downloaded Hadoop-0.19.0 and followed the instructions in the quick start manual(http://hadoop.apache.org/core/docs/r0.19.1/quickstart.html). When I came to Pseudo-Distributed Operation section there was no problem but localhost:50070 and localhost:50030 couldn't be

Re: Cannot access Jobtracker and namenode

2009-04-12 Thread Rasit OZDAS
Does your system request a password when you ssh to localhost outside hadoop? 12 Nisan 2009 Pazar 20:51 tarihinde halilibrahimcakir halilibrahimca...@mynet.com yazdı: Hi I am new at hadoop. I downloaded Hadoop-0.19.0 and followed the instructions in the quick start

Ynt: Re: Cannot access Jobtracker and namenode

2009-04-12 Thread halilibrahimcakir
Yes, usually when I type ssh localhost in terminal. - Özgün İleti - Kimden : core-user@hadoop.apache.org Kime : core-user@hadoop.apache.org Gönderme tarihi : 12/04/2009 20:58 Konu : Re: Cannot access Jobtracker and namenode Does your system request a password when you ssh to localhost

Re: Ynt: Re: Cannot access Jobtracker and namenode

2009-04-12 Thread Rasit OZDAS
There are two commands in hadoop quick start, used for passwordless ssh. Try those. $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa $ cat ~/.ssh/id_dsa.pub ~/.ssh/authorized_keys http://hadoop.apache.org/core/docs/current/quickstart.html -- M. Raşit ÖZDAŞ

Ynt: Re: Ynt: Re: Cannot access Jobtracker and namenode

2009-04-12 Thread halilibrahimcakir
I typed: $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa $ cat ~/.ssh/id_dsa.pub gt;gt; ~/.ssh/authorized_keys Deleted this directory: $ rm -ri /tmp/hadoop-root Formatted namenode again: $ /bin/hadoop namenode -format Stopped: $ /bin/stop-all.sh then typed: $ ssh localhost so it didn't

Re: Ynt: Re: Ynt: Re: Cannot access Jobtracker and namenode

2009-04-12 Thread Mithila Nagendra
You have to stop the cluster before you reformat. Also restarting the master might help. Mithila 2009/4/12 halilibrahimcakir halilibrahimca...@mynet.com I typed: $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa $ cat ~/.ssh/id_dsa.pub gt;gt; ~/.ssh/authorized_keys Deleted this directory: $ rm

Re: Changing block size of hadoop

2009-04-12 Thread Aaron Kimball
Blocks already written to HDFS will remain their current size. Blocks are immutable objects. That procedure would set the size used for all subsequently-written blocks. I don't think you can change the block size while the cluster is running, because that would require the NameNode and DataNodes

Re: Map-Reduce Slow Down

2009-04-12 Thread Aaron Kimball
Virtually none of the examples that ship with Hadoop are designed to showcase its speed. Hadoop's speedup comes from its ability to process very large volumes of data (starting around, say, tens of GB per job, and going up in orders of magnitude from there). So if you are timing the pi calculator

Re: Map-Reduce Slow Down

2009-04-12 Thread Mithila Nagendra
Aaron That could be the issue, my data is just 516MB - wouldn't this see a bit of speed up? Could you guide me to the example? I ll run my cluster on it and see what I get. Also for my program I had a java timer running to record the time taken to complete execution. Does Hadoop have an inbuilt

Re: Changing block size of hadoop

2009-04-12 Thread Raghu Angadi
Aaron Kimball wrote: Blocks already written to HDFS will remain their current size. Blocks are immutable objects. That procedure would set the size used for all subsequently-written blocks. I don't think you can change the block size while the cluster is running, because that would require the

Re: Interesting Hadoop/FUSE-DFS access patterns

2009-04-12 Thread Brian Bockelman
Ok, here's something perhaps even more strange. I removed the seek part out of my timings, so I was only timing the read instead of the seek + read as in the first case. I also turned the read-ahead down to 1-byte (aka, off). The jump *always* occurs at 128KB, exactly. I'm a bit