Re: Versions of Jetty Log4j in CDHu3

2012-04-29 Thread Nikhil
Oh ok. Thanks Chaidy.

I was wondering if I can just use log4j compression facility along with
timebased rolling as in terms of gzip and then use less disk space.
This seems to be feature available in 1.3 (not sure if it is also available
in log4j-1.2.15), I think I need to give it a try and see.
But on the other hand, if it does not work, any process suggestion that I
need to follow to upgrade log4j 1.3 and expect hadoop to be compatible with
new log4j library (version changes?)

Thanks,
Nikhil


On Sun, Apr 29, 2012 at 11:07 AM, CHAIDY cha...@nsccsz.gov.cn wrote:

 Hi, Nikhil!


 FYI: jetty-6.1.26, log4j-1.2.15



 At 2012-04-29 13:03:44,Nikhil mnik...@gmail.com wrote:
 Hi,
 
 I was wondering about the release versions of both Jetty and log4j
 components released as part of CDHu3 release package.
 Can someone please let me know.
 
 Thanks.



Re: Can’t stop hadoop daemons

2012-04-29 Thread Harsh J
Hey Barry,

How did you start these daemons in the first place?

On Sun, Apr 29, 2012 at 1:16 AM, Barry, Sean F sean.f.ba...@intel.com wrote:
 hduser@master:~ /usr/java/jdk1.7.0/bin/jps

 20907 TaskTracker

 20629 SecondaryNameNode

 25863 Jps

 20777 JobTracker

 20383 NameNode

 20507 DataNode

 hduser@master:~ stop-

 stop-all.sh       stop-balancer.sh  stop-dfs.sh       stop-mapred.sh

 hduser@master:~ stop-all.sh

 no jobtracker to stop

 master: no tasktracker to stop

 slave: no tasktracker to stop

 no namenode to stop

 master: no datanode to stop

 slave: no datanode to stop

 master: no secondarynamenode to stop

 hduser@master:~

 as you can see jps shows that the daemons are running but I cant stop them 
 with the stop-all.sh command.

 Does anyone have an idea for why this is happening ?

 -SB



-- 
Harsh J


RE: Can’t stop hadoop daemons

2012-04-29 Thread Barry, Sean F
I just restarted my machines and it works fine now.

-SB

-Original Message-
From: Harsh J [mailto:ha...@cloudera.com] 
Sent: Sunday, April 29, 2012 5:55 AM
To: common-user@hadoop.apache.org
Subject: Re: Can’t stop hadoop daemons

Hey Barry,

How did you start these daemons in the first place?

On Sun, Apr 29, 2012 at 1:16 AM, Barry, Sean F sean.f.ba...@intel.com wrote:
 hduser@master:~ /usr/java/jdk1.7.0/bin/jps

 20907 TaskTracker

 20629 SecondaryNameNode

 25863 Jps

 20777 JobTracker

 20383 NameNode

 20507 DataNode

 hduser@master:~ stop-

 stop-all.sh       stop-balancer.sh  stop-dfs.sh       stop-mapred.sh

 hduser@master:~ stop-all.sh

 no jobtracker to stop

 master: no tasktracker to stop

 slave: no tasktracker to stop

 no namenode to stop

 master: no datanode to stop

 slave: no datanode to stop

 master: no secondarynamenode to stop

 hduser@master:~

 as you can see jps shows that the daemons are running but I cant stop them 
 with the stop-all.sh command.

 Does anyone have an idea for why this is happening ?

 -SB



-- 
Harsh J


Re: Can’t stop hadoop daemons

2012-04-29 Thread Harsh J
Sean,

One cause I can think of is that your PID directory is on /tmp or so,
and the original saved PID files got cleared away by tmpwatch, leading
to this state.

To fix such a flaw, export HADOOP_PID_DIR in hadoop-env.sh to a more
persistent location (such as within HADOOP_HOME/pids itself, say).

What version of Hadoop are you using though?

On Mon, Apr 30, 2012 at 12:58 AM, Barry, Sean F sean.f.ba...@intel.com wrote:
 I just restarted my machines and it works fine now.

 -SB

 -Original Message-
 From: Harsh J [mailto:ha...@cloudera.com]
 Sent: Sunday, April 29, 2012 5:55 AM
 To: common-user@hadoop.apache.org
 Subject: Re: Can’t stop hadoop daemons

 Hey Barry,

 How did you start these daemons in the first place?

 On Sun, Apr 29, 2012 at 1:16 AM, Barry, Sean F sean.f.ba...@intel.com wrote:
 hduser@master:~ /usr/java/jdk1.7.0/bin/jps

 20907 TaskTracker

 20629 SecondaryNameNode

 25863 Jps

 20777 JobTracker

 20383 NameNode

 20507 DataNode

 hduser@master:~ stop-

 stop-all.sh       stop-balancer.sh  stop-dfs.sh       stop-mapred.sh

 hduser@master:~ stop-all.sh

 no jobtracker to stop

 master: no tasktracker to stop

 slave: no tasktracker to stop

 no namenode to stop

 master: no datanode to stop

 slave: no datanode to stop

 master: no secondarynamenode to stop

 hduser@master:~

 as you can see jps shows that the daemons are running but I cant stop them 
 with the stop-all.sh command.

 Does anyone have an idea for why this is happening ?

 -SB



 --
 Harsh J



-- 
Harsh J


Nightly Builds

2012-04-29 Thread nosql

Hi everyone,

i'm want to run hadoop (hbase) in an IBM JVM. I've seen that there were several 
patches for that reason. I am not a developer so my knowleges in building java 
jars fromsources are very limited and the link with the nightly builds do not 
work.
I only need hadoop-core-1.0.3.jar. Where can i find it even when it's unstable?

Thanks and bye,
Stephan

Re: DFSClient error

2012-04-29 Thread Harsh J
It sounds to me like you're running out of DN xceivers. Try the
solution offered at
http://hbase.apache.org/book.html#dfs.datanode.max.xcievers

I.e., add:

property
namedfs.datanode.max.xcievers/name
value4096/value
 /property

To your DNs' config/hdfs-site.xml and restart the DNs.

On Mon, Apr 30, 2012 at 1:35 AM, Mohit Anchlia mohitanch...@gmail.com wrote:
 I even tried to lower number of parallel jobs even further but I still get
 these errors. Any suggestion on how to troubleshoot this issue would be
 very helpful. Should I run hadoop fsck? How do people troubleshoot such
 issues?? Does it sound like a bug?

 2012-04-27 14:37:42,921 [main] INFO
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
 - 1 map-reduce job(s) waiting for submission.
 2012-04-27 14:37:42,931 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.199:50010java.io.EOFException
 2012-04-27 14:37:42,932 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_6343044536824463287_24619
 2012-04-27 14:37:42,932 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.199:50010
 2012-04-27 14:37:42,935 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.204:50010java.io.EOFException
 2012-04-27 14:37:42,935 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_2837215798109471362_24620
 2012-04-27 14:37:42,936 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.204:50010
 2012-04-27 14:37:42,937 [main] INFO
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
 - 1 map-reduce job(s) waiting for submission.
 2012-04-27 14:37:42,939 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.198:50010java.io.EOFException
 2012-04-27 14:37:42,939 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_2223489090936415027_24620
 2012-04-27 14:37:42,940 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.198:50010
 2012-04-27 14:37:42,943 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.197:50010java.io.EOFException
 2012-04-27 14:37:42,943 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_1265169201875643059_24620
 2012-04-27 14:37:42,944 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.197:50010
 2012-04-27 14:37:42,945 [Thread-5] WARN  org.apache.hadoop.hdfs.DFSClient -
 DataStreamer Exception: java.io.IOException: Unable to create new block.
        at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3446)
        at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2100(DFSClient.java:2627)
        at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2822)
 2012-04-27 14:37:42,945 [Thread-5] WARN  org.apache.hadoop.hdfs.DFSClient -
 Error Recovery for block blk_1265169201875643059_24620 bad datanode[0]
 nodes == null
 2012-04-27 14:37:42,945 [Thread-5] WARN  org.apache.hadoop.hdfs.DFSClient -
 Could not get block locations. Source file
 /tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201204261707_0411/job.jar
 - Aborting...
 2012-04-27 14:37:42,945 [Thread-4] INFO  org.apache.hadoop.mapred.JobClient
 - Cleaning up the staging area
 hdfs://dsdb1:54310/tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201204261707_0411
 2012-04-27 14:37:42,945 [Thread-4] ERROR
 org.apache.hadoop.security.UserGroupInformation -
 PriviledgedActionException as:hadoop (auth:SIMPLE)
 cause:java.io.EOFException
 2012-04-27 14:37:42,996 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream
 125.18.62.200:50010java.io.IOException: Bad connect ack with
 firstBadLink as
 125.18.62.198:50010
 2012-04-27 14:37:42,996 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_-7583284266913502018_24621
 2012-04-27 14:37:42,997 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.198:50010java.io.EOFException
 2012-04-27 14:37:42,997 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_4207260385919079785_24622
 2012-04-27 14:37:42,998 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.198:50010
 2012-04-27 14:37:43,000 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 125.18.62.198:50010
 2012-04-27 14:37:43,002 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Exception in createBlockOutputStream 125.18.62.197:50010java.io.EOFException
 2012-04-27 14:37:43,002 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Abandoning block blk_-2859304645525022496_24624
 2012-04-27 14:37:43,003 [Thread-5] INFO  org.apache.hadoop.hdfs.DFSClient -
 Excluding datanode 

Nightly Build

2012-04-29 Thread nosql
Hi everyone,

i'm want to run hadoop (hbase) in an IBM JVM. I've seen that there were several 
patches for that reason. I am not a developer so my knowleges in building java 
jars fromsources are very limited and the link with the nightly builds do not 
work.
I only need hadoop-core-1.0.3.jar. Where can i find it even when it's unstable?

Thanks and bye,
Stephan

Re: DFSClient error

2012-04-29 Thread Mohit Anchlia
Thanks for the quick response, appreciate it. It looks like this might be
the issue. But I am still trying to understand what is causing so many
threads in my situation? Is this thread per block that gets created or per
file? Because if it's per file then it should not be more than 15.
My second question, I read around 5 .gz files in 5 separate processed. This
is constant and also the size of those 5 is roughly equivalent. So then why
does it fail only halfway and not right in the begining. I am reading
around 400 files and it always fails when I reach around 180th file.

What's the default value of xceivers? Is 4096 consume too much of stack
size?

Thanks
On Sun, Apr 29, 2012 at 1:14 PM, Harsh J ha...@cloudera.com wrote:

 It sounds to me like you're running out of DN xceivers. Try the
 solution offered at
 http://hbase.apache.org/book.html#dfs.datanode.max.xcievers

 I.e., add:

 property
namedfs.datanode.max.xcievers/name
value4096/value
  /property

 To your DNs' config/hdfs-site.xml and restart the DNs.

 On Mon, Apr 30, 2012 at 1:35 AM, Mohit Anchlia mohitanch...@gmail.com
 wrote:
  I even tried to lower number of parallel jobs even further but I still
 get
  these errors. Any suggestion on how to troubleshoot this issue would be
  very helpful. Should I run hadoop fsck? How do people troubleshoot such
  issues?? Does it sound like a bug?
 
  2012-04-27 14:37:42,921 [main] INFO
 
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
  - 1 map-reduce job(s) waiting for submission.
  2012-04-27 14:37:42,931 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Exception in createBlockOutputStream 125.18.62.199:50010
 java.io.EOFException
  2012-04-27 14:37:42,932 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Abandoning block blk_6343044536824463287_24619
  2012-04-27 14:37:42,932 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Excluding datanode 125.18.62.199:50010
  2012-04-27 14:37:42,935 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Exception in createBlockOutputStream 125.18.62.204:50010
 java.io.EOFException
  2012-04-27 14:37:42,935 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Abandoning block blk_2837215798109471362_24620
  2012-04-27 14:37:42,936 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Excluding datanode 125.18.62.204:50010
  2012-04-27 14:37:42,937 [main] INFO
 
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
  - 1 map-reduce job(s) waiting for submission.
  2012-04-27 14:37:42,939 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Exception in createBlockOutputStream 125.18.62.198:50010
 java.io.EOFException
  2012-04-27 14:37:42,939 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Abandoning block blk_2223489090936415027_24620
  2012-04-27 14:37:42,940 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Excluding datanode 125.18.62.198:50010
  2012-04-27 14:37:42,943 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Exception in createBlockOutputStream 125.18.62.197:50010
 java.io.EOFException
  2012-04-27 14:37:42,943 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Abandoning block blk_1265169201875643059_24620
  2012-04-27 14:37:42,944 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Excluding datanode 125.18.62.197:50010
  2012-04-27 14:37:42,945 [Thread-5] WARN
  org.apache.hadoop.hdfs.DFSClient -
  DataStreamer Exception: java.io.IOException: Unable to create new block.
 at
 
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3446)
 at
 
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2100(DFSClient.java:2627)
 at
 
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2822)
  2012-04-27 14:37:42,945 [Thread-5] WARN
  org.apache.hadoop.hdfs.DFSClient -
  Error Recovery for block blk_1265169201875643059_24620 bad datanode[0]
  nodes == null
  2012-04-27 14:37:42,945 [Thread-5] WARN
  org.apache.hadoop.hdfs.DFSClient -
  Could not get block locations. Source file
 
 /tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201204261707_0411/job.jar
  - Aborting...
  2012-04-27 14:37:42,945 [Thread-4] INFO
  org.apache.hadoop.mapred.JobClient
  - Cleaning up the staging area
 
 hdfs://dsdb1:54310/tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201204261707_0411
  2012-04-27 14:37:42,945 [Thread-4] ERROR
  org.apache.hadoop.security.UserGroupInformation -
  PriviledgedActionException as:hadoop (auth:SIMPLE)
  cause:java.io.EOFException
  2012-04-27 14:37:42,996 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Exception in createBlockOutputStream
  125.18.62.200:50010java.io.IOException: Bad connect ack with
  firstBadLink as
   125.18.62.198:50010
  2012-04-27 14:37:42,996 [Thread-5] INFO
  org.apache.hadoop.hdfs.DFSClient -
  Abandoning block blk_-7583284266913502018_24621
  2012-04-27 14:37:42,997 [Thread-5] INFO
  

Getting job progress in java application

2012-04-29 Thread Ondřej Klimpera
Hello I'd like to ask you what is the preferred way of getting running 
jobs progress from Java application, that has executed them.


Im using Hadoop 0.20.203, tried job.end.notification.url property that 
works well, but as the property name says, it sends only job end 
notifications.


What I need is to get updates on map() and reduce() progress.

Please help how to do this.

Thanks.
Ondrej Klimpera



Hadoop Security 2011/2012

2012-04-29 Thread Atul Thapliyal
Hi Hadoop users,

Has anyone attended a Hadoop conference where there were talks about any
new features in Hadoop security ?

I am trying to figure out if  any new features have been added to Hadoop
security after Kerberos .

Thanks

-- 
Cheers
Atul


Re: Getting job progress in java application

2012-04-29 Thread Bill Graham
Take a look at the JobClient API. You can use that to get the current
progress of a running job.

On Sunday, April 29, 2012, Ondřej Klimpera wrote:

 Hello I'd like to ask you what is the preferred way of getting running
 jobs progress from Java application, that has executed them.

 Im using Hadoop 0.20.203, tried job.end.notification.url property that
 works well, but as the property name says, it sends only job end
 notifications.

 What I need is to get updates on map() and reduce() progress.

 Please help how to do this.

 Thanks.
 Ondrej Klimpera



-- 
*Note that I'm no longer using my Yahoo! email address. Please email me at
billgra...@gmail.com going forward.*


Mass SocketTimeoutException - 0.20.203

2012-04-29 Thread Shi Yu
Tons of errors seen after Map 100% Reduce 50%, but the job 
still struggles to finish. What is the possible reason? Is 
this issue fixed in any of the version?

java.net.SocketTimeoutException: 69000 millis timeout while 
waiting for channel to be ready for read. ch : 
java.nio.channels.SocketChannel[connected 
local=/10.105.19.35:58539 remote=/10.105.19.35:50010]
at 
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTim
eout.java:164)
at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream
.java:155)
at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream
.java:128)
at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream
.java:116)
at 
java.io.DataInputStream.readShort(DataInputStream.java:295)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOu
tputStream(DFSClient.java:3060)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutp
utStream(DFSClient.java:2983)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(D
FSClient.java:2255)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.
run(DFSClient.java:2446)


EMR Hadoop

2012-04-29 Thread Jay Vyas
Hi guys :

1) Does anybody know if there is a VM out there which runs EMR hadoop ?  I
would like to have a
local vm for dev purposes that mirrored the EMR hadoop instances.

2) How does EMR's hadoop differ from apache hadoop and Cloudera's hadoop ?

-- 
Jay Vyas
MMSB/UCHC