Hi.
When I run hadoop balancer, I get the following error:
09/04/12 10:28:46 INFO dfs.Balancer: Will move 3.02 GBbytes in this
iteration
Apr 12, 2009 10:28:46 AM 0 0 KB19.02
GB3.02 GB
09/04/12 10:28:46 INFO dfs.Balancer: Decided to move block
Hi:
I'm trying to use FSDataOutputStream create(Path f, boolean overwrite),
I'm calling create(new Path(somePath), false), but creation still fails
with IOException even when the file does not exist, can someone explain the
behavior?
thanks,
Hi,
I am trying to create the Har files in hadoop 0.19.0, but on my hadoop
cluster mapreduce jobs are not running , It gives very known exception
09/04/12 09:54:07 INFO mapred.FileInputFormat: Total input paths to process
: 1
09/04/12 09:54:08 INFO mapred.JobClient: Running job:
Hi,
I would like to know if it is feasbile to change the blocksize of Hadoop
while map reduce jobs are executing? and if not would the following work? 1.
stop map-reduce 2. stop-hbase 3. stop hadoop 4. change hadoop-sites.xml
to reduce the blocksize 5. restart all
whether the data in the
Hi
I am new at hadoop. I downloaded Hadoop-0.19.0 and followed the
instructions in the quick start
manual(http://hadoop.apache.org/core/docs/r0.19.1/quickstart.html). When I
came to Pseudo-Distributed Operation section there was no problem but
localhost:50070 and localhost:50030 couldn't be
Does your system request a password when you ssh to localhost outside hadoop?
12 Nisan 2009 Pazar 20:51 tarihinde halilibrahimcakir
halilibrahimca...@mynet.com yazdı:
Hi
I am new at hadoop. I downloaded Hadoop-0.19.0 and followed the
instructions in the quick start
Yes, usually when I type ssh localhost in terminal.
- Özgün İleti -
Kimden : core-user@hadoop.apache.org
Kime : core-user@hadoop.apache.org
Gönderme tarihi : 12/04/2009 20:58
Konu : Re: Cannot access Jobtracker and namenode
Does your system request a password when you ssh to localhost
There are two commands in hadoop quick start, used for passwordless ssh.
Try those.
$ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
$ cat ~/.ssh/id_dsa.pub ~/.ssh/authorized_keys
http://hadoop.apache.org/core/docs/current/quickstart.html
--
M. Raşit ÖZDAŞ
I typed:
$ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
$ cat ~/.ssh/id_dsa.pub gt;gt; ~/.ssh/authorized_keys
Deleted this directory:
$ rm -ri /tmp/hadoop-root
Formatted namenode again:
$ /bin/hadoop namenode -format
Stopped:
$ /bin/stop-all.sh
then typed:
$ ssh localhost
so it didn't
You have to stop the cluster before you reformat. Also restarting the master
might help.
Mithila
2009/4/12 halilibrahimcakir halilibrahimca...@mynet.com
I typed:
$ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
$ cat ~/.ssh/id_dsa.pub gt;gt; ~/.ssh/authorized_keys
Deleted this directory:
$ rm
Blocks already written to HDFS will remain their current size. Blocks are
immutable objects. That procedure would set the size used for all
subsequently-written blocks. I don't think you can change the block size
while the cluster is running, because that would require the NameNode and
DataNodes
Virtually none of the examples that ship with Hadoop are designed to
showcase its speed. Hadoop's speedup comes from its ability to process very
large volumes of data (starting around, say, tens of GB per job, and going
up in orders of magnitude from there). So if you are timing the pi
calculator
Aaron
That could be the issue, my data is just 516MB - wouldn't this see a bit of
speed up?
Could you guide me to the example? I ll run my cluster on it and see what I
get. Also for my program I had a java timer running to record the time taken
to complete execution. Does Hadoop have an inbuilt
Aaron Kimball wrote:
Blocks already written to HDFS will remain their current size. Blocks are
immutable objects. That procedure would set the size used for all
subsequently-written blocks. I don't think you can change the block size
while the cluster is running, because that would require the
Ok, here's something perhaps even more strange. I removed the seek
part out of my timings, so I was only timing the read instead of the
seek + read as in the first case. I also turned the read-ahead down
to 1-byte (aka, off).
The jump *always* occurs at 128KB, exactly.
I'm a bit
15 matches
Mail list logo