Re: Jobtracker history logs missing

2012-04-05 Thread Nitin Khandelwal
Hi Prashant,

The userlogs for job are deleted after time specified by  *
mapred.userlog.retain.hours*  property defined in mapred-site.xml (default
is 24 Hrs).

Thanks,
Nitin

On 5 April 2012 14:26, Prashant Kommireddi prash1...@gmail.com wrote:

 I am noticing something strange with JobTracker history logs on my cluster.
 I see configuration files (*_conf.xml) under /logs/history/ but none of the
 actual job logs. Anyone has ideas on what might be happening?

 Thanks,




-- 


Nitin Khandelwal


Re: How do i set number of tasktrackers per node ?

2011-12-14 Thread Nitin Khandelwal
Hi ,

You can set using properties mapred.tasktracker.map.tasks.maximum,
mapred.tasktracker.reduce.tasks.maximum in mapred-site.xml

Thanks,
Nition

On 15 December 2011 10:44, ArunKumar arunk...@gmail.com wrote:

 Hi guys !

 I see that we can set number of map/reduce tasks per tasktracker but how do
 i set number of tasktrackers per node in a cluster ?



 Arun

 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/How-do-i-set-number-of-tasktrackers-per-node-tp3587672p3587672.html
 Sent from the Hadoop lucene-users mailing list archive at Nabble.com.




-- 


Nitin Khandelwal


Hadoop 0.22.0 Documentation: Config Files default point to diff(current) version Config

2011-12-12 Thread Nitin Khandelwal
Hi All,

I was going through Hadoop 0.22.0 Documentation at
http://hadoop.apache.org/common/docs/r0.22.0/cluster_setup.html#Configuration+Files.
The default config files (core-default, mapred-default, hdfs-default) are
pointing to* CURRENT Hadoop release and not 0.22.0. *

*For Example :*
*core-default.xml* points to *
http://hadoop.apache.org/common/docs/current/core-default.html* instead of
: *http://hadoop.apache.org/common/docs/r0.22.0/core-default.html*

Also, link to 
mapred-queues.xml.templatehttp://hadoop.apache.org/mapreduce/docs/current/mapred_queues.xmlis
broken.

Please change the links (if it is wrong) as it is pretty confusing.

Thanks,
-- 
Nitin Khandelwal


Re: Not able to post a job in Hadoop 0.23.0

2011-12-08 Thread Nitin Khandelwal
Hi Arun,
Thanks for your reply.


There is one NodeManager running ; Following is the NodeManager UI :

Rack
Node State
Node Address
Node HTTP Address
Health-status
Last health-update
Health-report
Containers
Mem Used
Mem Avail
 /default-rack RUNNING germinait93:50033 germinait93: Healthy 9-Dec-2011
13:03:33 Healthy 0 0 KB 1 GB
Also, I get to see only following Logs relevant to the job posting :

2011-12-09 13:10:57,300 INFO  fifo.FifoScheduler (FifoScheduler.java:
addApplication(288)) - Application Submission:
application_1323416004722_0002 from minal.kothari, currently
 active: 1
2011-12-09 13:10:57,300 INFO  attempt.RMAppAttemptImpl
(RMAppAttemptImpl.java:handle(464)) - Processing event for
appattempt_1323416004722_0002_01 of type APP_ACCEPTED
2011-12-09 13:10:57,317 INFO  attempt.RMAppAttemptImpl
(RMAppAttemptImpl.java:handle(476)) - appattempt_1323416004722_0002_01
State change from SUBMITTED to SCHEDULED
2011-12-09 13:10:57,318 INFO  rmapp.RMAppImpl (RMAppImpl.java:handle(416))
- Processing event for application_1323416004722_0002 of type APP_ACCEPTED
2011-12-09 13:10:57,318 INFO  rmapp.RMAppImpl (RMAppImpl.java:handle(428))
- application_1323416004722_0002 State change from SUBMITTED to ACCEPTED
2011-12-09 13:10:57,320 INFO  resourcemanager.RMAuditLogger
(RMAuditLogger.java:logSuccess(140)) - USER=minal.kothari   IP=192.168.0.93
OPERATION=Submit Application Request
TARGET=ClientRMService  RESULT=SUCCESS  APPID=application_1323416004722_0002


Please let me know if you need some other logs .

Thanks,
Nitin




On 9 December 2011 12:44, Arun C Murthy a...@hortonworks.com wrote:

 Moving to mapreduce-user@, bcc common-user@.

 Can you see any errors in the logs? Typically this happens when you have
 no NodeManagers.

 Check the 'nodes' link and then RM logs.

 Arun

 On Nov 29, 2011, at 8:36 PM, Nitin Khandelwal wrote:

  HI ,
 
  I have successfully setup Hadoop 0.23.0 in a single m/c. When i post a
 job,
  it gets posted successfully (i can see the job in UI), but the job is
 never
  ASSIGNED and waits forever.
  Here are details of what i see for that Job in UI
 
 
  Name: random-writer  State: ACCEPTED  FinalStatus: UNDEFINED
  Started: 30-Nov-2011
  10:08:55  Elapsed: 49sec  Tracking URL:
  UNASSIGNED
 http://192.168.0.93:8900/cluster/app/application_1322627869620_0001#
  Diagnostics:
  AM container logs: AM not yet registered with RM  Cluster ID:
 1322627869620
  ResourceManager state: STARTED  ResourceManager started on: 30-Nov-2011
  10:07:49  ResourceManager version: 0.23.0 from
  722cd694fc4ab6d040c0a34f9fb5b476e330ee60 by hortonmu source checksum
  4975bf112aa7faa5673f604045ced798 on Thu Nov 3 09:07:31 UTC 2011  Hadoop
  version: 0.23.0 from d4fee83ec1462ab9824add6449320617caa7c605 by hortonmu
  source checksum 4e42b2d96c899a98a8ab8c7cc23f27ae on Thu Nov 3 08:59:12
 UTC
  2011
  Can some one tell where am i going wrong??
 
  Thanks,
  --
  Nitin Khandelwal




-- 


Nitin Khandelwal


Re: Not able to post a job in Hadoop 0.23.0

2011-12-08 Thread Nitin Khandelwal
CC : mapreduce-user

On 9 December 2011 13:14, Nitin Khandelwal
nitin.khandel...@germinait.comwrote:

 Hi Arun,
 Thanks for your reply.


 There is one NodeManager running ; Following is the NodeManager UI :

 Rack
 Node State
 Node Address
 Node HTTP Address
 Health-status
 Last health-update
 Health-report
 Containers
 Mem Used
 Mem Avail
  /default-rack RUNNING germinait93:50033 germinait93: Healthy 9-Dec-2011
 13:03:33 Healthy 0 0 KB 1 GB
 Also, I get to see only following Logs relevant to the job posting :

 2011-12-09 13:10:57,300 INFO  fifo.FifoScheduler (FifoScheduler.java:
 addApplication(288)) - Application Submission:
 application_1323416004722_0002 from minal.kothari, currently
  active: 1
 2011-12-09 13:10:57,300 INFO  attempt.RMAppAttemptImpl
 (RMAppAttemptImpl.java:handle(464)) - Processing event for
 appattempt_1323416004722_0002_01 of type APP_ACCEPTED
 2011-12-09 13:10:57,317 INFO  attempt.RMAppAttemptImpl
 (RMAppAttemptImpl.java:handle(476)) - appattempt_1323416004722_0002_01
 State change from SUBMITTED to SCHEDULED
 2011-12-09 13:10:57,318 INFO  rmapp.RMAppImpl (RMAppImpl.java:handle(416))
 - Processing event for application_1323416004722_0002 of type APP_ACCEPTED
 2011-12-09 13:10:57,318 INFO  rmapp.RMAppImpl (RMAppImpl.java:handle(428))
 - application_1323416004722_0002 State change from SUBMITTED to ACCEPTED
 2011-12-09 13:10:57,320 INFO  resourcemanager.RMAuditLogger
 (RMAuditLogger.java:logSuccess(140)) - USER=minal.kothari   IP=192.168.0.93
 OPERATION=Submit Application Request
 TARGET=ClientRMService  RESULT=SUCCESS
 APPID=application_1323416004722_0002


 Please let me know if you need some other logs .

 Thanks,
 Nitin




 On 9 December 2011 12:44, Arun C Murthy a...@hortonworks.com wrote:

 Moving to mapreduce-user@, bcc common-user@.

 Can you see any errors in the logs? Typically this happens when you have
 no NodeManagers.

 Check the 'nodes' link and then RM logs.

 Arun

 On Nov 29, 2011, at 8:36 PM, Nitin Khandelwal wrote:

  HI ,
 
  I have successfully setup Hadoop 0.23.0 in a single m/c. When i post a
 job,
  it gets posted successfully (i can see the job in UI), but the job is
 never
  ASSIGNED and waits forever.
  Here are details of what i see for that Job in UI
 
 
  Name: random-writer  State: ACCEPTED  FinalStatus: UNDEFINED
  Started: 30-Nov-2011
  10:08:55  Elapsed: 49sec  Tracking URL:
  UNASSIGNED
 http://192.168.0.93:8900/cluster/app/application_1322627869620_0001#
  Diagnostics:
  AM container logs: AM not yet registered with RM  Cluster ID:
 1322627869620
  ResourceManager state: STARTED  ResourceManager started on: 30-Nov-2011
  10:07:49  ResourceManager version: 0.23.0 from
  722cd694fc4ab6d040c0a34f9fb5b476e330ee60 by hortonmu source checksum
  4975bf112aa7faa5673f604045ced798 on Thu Nov 3 09:07:31 UTC 2011  Hadoop
  version: 0.23.0 from d4fee83ec1462ab9824add6449320617caa7c605 by
 hortonmu
  source checksum 4e42b2d96c899a98a8ab8c7cc23f27ae on Thu Nov 3 08:59:12
 UTC
  2011
  Can some one tell where am i going wrong??
 
  Thanks,
  --
  Nitin Khandelwal




 --


 Nitin Khandelwal





-- 


Nitin Khandelwal


Re: Re: [help]how to stop HDFS

2011-11-29 Thread Nitin Khandelwal
Hi,

Even i am facing the same problem. There may be some issue with script .
The doc says to start namenode type :
bin/hdfs namenode start

But start is not recognized. There is a hack to start namenode with
command bin/hdfs namenode  , but no idea how to stop.
If it had been a issue with config , the later also should not have worked.

Thanks,
Nitin


2011/11/30 cat fa boost.subscrib...@gmail.com

 In fact it's me to say sorry. I used the word install which was
 misleading.

 In fact I downloaded a tar file and extracted it to /usr/bin/hadoop

 Could you please tell me where to point those variables?

 2011/11/30, Prashant Sharma prashant.ii...@gmail.com:
  I am sorry, I had no idea you have done a rpm install, my suggestion was
  based on the assumption that you have done a tar extract install where
 all
  three distribution have to extracted and then export variables.
  Also I have no experience with rpm based installs - so no comments about
  what went wrong in your case.
 
  Basically from the error i can say that it is not able to find the jars
  needed  on classpath which is referred by scripts through
  HADOOP_COMMON_HOME. I would say check with the access permission as in
  which user was it installed with and which user is it running with ?
 
  On Tue, Nov 29, 2011 at 10:48 PM, cat fa boost.subscrib...@gmail.com
 wrote:
 
  Thank you for your help, but I'm still a little confused.
  Suppose I installed hadoop in /usr/bin/hadoop/ .Should I
  point HADOOP_COMMON_HOME to /usr/bin/hadoop ? Where should I
  point HADOOP_HDFS_HOME? Also to /usr/bin/hadoop/ ?
 
  2011/11/30 Prashant Sharma prashant.ii...@gmail.com
 
   I mean, you have to export the variables
  
   export HADOOP_CONF_DIR=/path/to/your/configdirectory.
  
   also export HADOOP_HDFS_HOME ,HADOOP_COMMON_HOME. before your run your
   command. I suppose this should fix the problem.
   -P
  
   On Tue, Nov 29, 2011 at 6:23 PM, cat fa boost.subscrib...@gmail.com
   wrote:
  
it didn't work. It gave me the Usage information.
   
2011/11/29 hailong.yang1115 hailong.yang1...@gmail.com
   
 Try $HADOOP_PREFIX_HOME/bin/hdfs namenode stop --config
   $HADOOP_CONF_DIR
 and $HADOOP_PREFIX_HOME/bin/hdfs datanode stop --config
   $HADOOP_CONF_DIR.
 It would stop namenode and datanode separately.
 The HADOOP_CONF_DIR is the directory where you store your
  configuration
 files.
 Hailong




 ***
 * Hailong Yang, PhD. Candidate
 * Sino-German Joint Software Institute,
 * School of Computer ScienceEngineering, Beihang University
 * Phone: (86-010)82315908
 * Email: hailong.yang1...@gmail.com
 * Address: G413, New Main Building in Beihang University,
 *  No.37 XueYuan Road,HaiDian District,
 *  Beijing,P.R.China,100191
 ***

 From: cat fa
 Date: 2011-11-29 20:22
 To: common-user
 Subject: Re: [help]how to stop HDFS
 use $HADOOP_CONF or $HADOOP_CONF_DIR ? I'm using hadoop 0.23.

 you mean which class? the class of hadoop or of java?

 2011/11/29 Prashant Sharma prashant.ii...@gmail.com

  Try making $HADOOP_CONF point to right classpath including your
  configuration folder.
 
 
  On Tue, Nov 29, 2011 at 3:58 PM, cat fa 
  boost.subscrib...@gmail.com
   
  wrote:
 
   I used the command :
  
   $HADOOP_PREFIX_HOME/bin/hdfs start namenode --config
   $HADOOP_CONF_DIR
  
   to sart HDFS.
  
   This command is in Hadoop document (here
   
  
 

   
  
 
 http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/ClusterSetup.html
   )
  
   However, I got errors as
  
   Exception in thread main
 java.lang.NoClassDefFoundError:start
  
   Could anyone tell me how to start and stop HDFS?
  
   By the way, how to set Gmail so that it doesn't top post my
  reply?
  
 

   
  
 
 




-- 


Nitin Khandelwal


Re: Re: [help]how to stop HDFS

2011-11-29 Thread Nitin Khandelwal
I am using Hadoop 0.23.0
There is no hadoop-daemon.sh in bin directory..

Thanks,
Nitin

On 30 November 2011 09:49, Harsh J ha...@cloudera.com wrote:

 I simply use the /sbin/hadoop-daemon.sh {start|stop} {service} script
 to control daemons at my end.

 Does this not work for you? Or perhaps this thread is more about
 documenting that?

 2011/11/30 Nitin Khandelwal nitin.khandel...@germinait.com:
  Hi,
 
  Even i am facing the same problem. There may be some issue with script .
  The doc says to start namenode type :
  bin/hdfs namenode start
 
  But start is not recognized. There is a hack to start namenode with
  command bin/hdfs namenode  , but no idea how to stop.
  If it had been a issue with config , the later also should not have
 worked.
 
  Thanks,
  Nitin
 
 
  2011/11/30 cat fa boost.subscrib...@gmail.com
 
  In fact it's me to say sorry. I used the word install which was
  misleading.
 
  In fact I downloaded a tar file and extracted it to /usr/bin/hadoop
 
  Could you please tell me where to point those variables?
 
  2011/11/30, Prashant Sharma prashant.ii...@gmail.com:
   I am sorry, I had no idea you have done a rpm install, my suggestion
 was
   based on the assumption that you have done a tar extract install where
  all
   three distribution have to extracted and then export variables.
   Also I have no experience with rpm based installs - so no comments
 about
   what went wrong in your case.
  
   Basically from the error i can say that it is not able to find the
 jars
   needed  on classpath which is referred by scripts through
   HADOOP_COMMON_HOME. I would say check with the access permission as in
   which user was it installed with and which user is it running with ?
  
   On Tue, Nov 29, 2011 at 10:48 PM, cat fa boost.subscrib...@gmail.com
  wrote:
  
   Thank you for your help, but I'm still a little confused.
   Suppose I installed hadoop in /usr/bin/hadoop/ .Should I
   point HADOOP_COMMON_HOME to /usr/bin/hadoop ? Where should I
   point HADOOP_HDFS_HOME? Also to /usr/bin/hadoop/ ?
  
   2011/11/30 Prashant Sharma prashant.ii...@gmail.com
  
I mean, you have to export the variables
   
export HADOOP_CONF_DIR=/path/to/your/configdirectory.
   
also export HADOOP_HDFS_HOME ,HADOOP_COMMON_HOME. before your run
 your
command. I suppose this should fix the problem.
-P
   
On Tue, Nov 29, 2011 at 6:23 PM, cat fa 
 boost.subscrib...@gmail.com
wrote:
   
 it didn't work. It gave me the Usage information.

 2011/11/29 hailong.yang1115 hailong.yang1...@gmail.com

  Try $HADOOP_PREFIX_HOME/bin/hdfs namenode stop --config
$HADOOP_CONF_DIR
  and $HADOOP_PREFIX_HOME/bin/hdfs datanode stop --config
$HADOOP_CONF_DIR.
  It would stop namenode and datanode separately.
  The HADOOP_CONF_DIR is the directory where you store your
   configuration
  files.
  Hailong
 
 
 
 
  ***
  * Hailong Yang, PhD. Candidate
  * Sino-German Joint Software Institute,
  * School of Computer ScienceEngineering, Beihang University
  * Phone: (86-010)82315908
  * Email: hailong.yang1...@gmail.com
  * Address: G413, New Main Building in Beihang University,
  *  No.37 XueYuan Road,HaiDian District,
  *  Beijing,P.R.China,100191
  ***
 
  From: cat fa
  Date: 2011-11-29 20:22
  To: common-user
  Subject: Re: [help]how to stop HDFS
  use $HADOOP_CONF or $HADOOP_CONF_DIR ? I'm using hadoop 0.23.
 
  you mean which class? the class of hadoop or of java?
 
  2011/11/29 Prashant Sharma prashant.ii...@gmail.com
 
   Try making $HADOOP_CONF point to right classpath including
 your
   configuration folder.
  
  
   On Tue, Nov 29, 2011 at 3:58 PM, cat fa 
   boost.subscrib...@gmail.com

   wrote:
  
I used the command :
   
$HADOOP_PREFIX_HOME/bin/hdfs start namenode --config
$HADOOP_CONF_DIR
   
to sart HDFS.
   
This command is in Hadoop document (here

   
  
 

   
  
 
 http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/ClusterSetup.html
)
   
However, I got errors as
   
Exception in thread main
  java.lang.NoClassDefFoundError:start
   
Could anyone tell me how to start and stop HDFS?
   
By the way, how to set Gmail so that it doesn't top post my
   reply?
   
  
 

   
  
  
 
 
 
 
  --
 
 
  Nitin Khandelwal
 



 --
 Harsh J




-- 


Nitin Khandelwal


Re: Re: [help]how to stop HDFS

2011-11-29 Thread Nitin Khandelwal
Thanks,
I missed the sbin directory, was using the normal bin directory.
Thanks,
Nitin

On 30 November 2011 09:54, Harsh J ha...@cloudera.com wrote:

 Like I wrote earlier, its in the $HADOOP_HOME/sbin directory. Not the
 regular bin/ directory.

 On Wed, Nov 30, 2011 at 9:52 AM, Nitin Khandelwal
 nitin.khandel...@germinait.com wrote:
  I am using Hadoop 0.23.0
  There is no hadoop-daemon.sh in bin directory..
 
  Thanks,
  Nitin
 
  On 30 November 2011 09:49, Harsh J ha...@cloudera.com wrote:
 
  I simply use the /sbin/hadoop-daemon.sh {start|stop} {service} script
  to control daemons at my end.
 
  Does this not work for you? Or perhaps this thread is more about
  documenting that?
 
  2011/11/30 Nitin Khandelwal nitin.khandel...@germinait.com:
   Hi,
  
   Even i am facing the same problem. There may be some issue with
 script .
   The doc says to start namenode type :
   bin/hdfs namenode start
  
   But start is not recognized. There is a hack to start namenode with
   command bin/hdfs namenode  , but no idea how to stop.
   If it had been a issue with config , the later also should not have
  worked.
  
   Thanks,
   Nitin
  
  
   2011/11/30 cat fa boost.subscrib...@gmail.com
  
   In fact it's me to say sorry. I used the word install which was
   misleading.
  
   In fact I downloaded a tar file and extracted it to /usr/bin/hadoop
  
   Could you please tell me where to point those variables?
  
   2011/11/30, Prashant Sharma prashant.ii...@gmail.com:
I am sorry, I had no idea you have done a rpm install, my
 suggestion
  was
based on the assumption that you have done a tar extract install
 where
   all
three distribution have to extracted and then export variables.
Also I have no experience with rpm based installs - so no comments
  about
what went wrong in your case.
   
Basically from the error i can say that it is not able to find the
  jars
needed  on classpath which is referred by scripts through
HADOOP_COMMON_HOME. I would say check with the access permission
 as in
which user was it installed with and which user is it running with
 ?
   
On Tue, Nov 29, 2011 at 10:48 PM, cat fa 
 boost.subscrib...@gmail.com
   wrote:
   
Thank you for your help, but I'm still a little confused.
Suppose I installed hadoop in /usr/bin/hadoop/ .Should I
point HADOOP_COMMON_HOME to /usr/bin/hadoop ? Where should I
point HADOOP_HDFS_HOME? Also to /usr/bin/hadoop/ ?
   
2011/11/30 Prashant Sharma prashant.ii...@gmail.com
   
 I mean, you have to export the variables

 export HADOOP_CONF_DIR=/path/to/your/configdirectory.

 also export HADOOP_HDFS_HOME ,HADOOP_COMMON_HOME. before your
 run
  your
 command. I suppose this should fix the problem.
 -P

 On Tue, Nov 29, 2011 at 6:23 PM, cat fa 
  boost.subscrib...@gmail.com
 wrote:

  it didn't work. It gave me the Usage information.
 
  2011/11/29 hailong.yang1115 hailong.yang1...@gmail.com
 
   Try $HADOOP_PREFIX_HOME/bin/hdfs namenode stop --config
 $HADOOP_CONF_DIR
   and $HADOOP_PREFIX_HOME/bin/hdfs datanode stop --config
 $HADOOP_CONF_DIR.
   It would stop namenode and datanode separately.
   The HADOOP_CONF_DIR is the directory where you store your
configuration
   files.
   Hailong
  
  
  
  
   ***
   * Hailong Yang, PhD. Candidate
   * Sino-German Joint Software Institute,
   * School of Computer ScienceEngineering, Beihang University
   * Phone: (86-010)82315908
   * Email: hailong.yang1...@gmail.com
   * Address: G413, New Main Building in Beihang University,
   *  No.37 XueYuan Road,HaiDian District,
   *  Beijing,P.R.China,100191
   ***
  
   From: cat fa
   Date: 2011-11-29 20:22
   To: common-user
   Subject: Re: [help]how to stop HDFS
   use $HADOOP_CONF or $HADOOP_CONF_DIR ? I'm using hadoop
 0.23.
  
   you mean which class? the class of hadoop or of java?
  
   2011/11/29 Prashant Sharma prashant.ii...@gmail.com
  
Try making $HADOOP_CONF point to right classpath including
  your
configuration folder.
   
   
On Tue, Nov 29, 2011 at 3:58 PM, cat fa 
boost.subscrib...@gmail.com
 
wrote:
   
 I used the command :

 $HADOOP_PREFIX_HOME/bin/hdfs start namenode --config
 $HADOOP_CONF_DIR

 to sart HDFS.

 This command is in Hadoop document (here
 

   
  
 

   
  
 
 http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/ClusterSetup.html
 )

 However, I got errors as

 Exception in thread main
   java.lang.NoClassDefFoundError:start

 Could anyone tell me how to start

Not able to post a job in Hadoop 0.23.0

2011-11-29 Thread Nitin Khandelwal
HI ,

I have successfully setup Hadoop 0.23.0 in a single m/c. When i post a job,
it gets posted successfully (i can see the job in UI), but the job is never
ASSIGNED and waits forever.
Here are details of what i see for that Job in UI


 Name: random-writer  State: ACCEPTED  FinalStatus: UNDEFINED
Started: 30-Nov-2011
10:08:55  Elapsed: 49sec  Tracking URL:
UNASSIGNEDhttp://192.168.0.93:8900/cluster/app/application_1322627869620_0001#
Diagnostics:
 AM container logs: AM not yet registered with RM  Cluster ID: 1322627869620
 ResourceManager state: STARTED  ResourceManager started on: 30-Nov-2011
10:07:49  ResourceManager version: 0.23.0 from
722cd694fc4ab6d040c0a34f9fb5b476e330ee60 by hortonmu source checksum
4975bf112aa7faa5673f604045ced798 on Thu Nov 3 09:07:31 UTC 2011  Hadoop
version: 0.23.0 from d4fee83ec1462ab9824add6449320617caa7c605 by hortonmu
source checksum 4e42b2d96c899a98a8ab8c7cc23f27ae on Thu Nov 3 08:59:12 UTC
2011
Can some one tell where am i going wrong??

Thanks,
-- 
Nitin Khandelwal


Problem running Hadoop 0.23.0

2011-11-28 Thread Nitin Khandelwal
Hi,

I was trying to setup Hadoop 0.23.0 with help of
http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/SingleCluster.html.
After starting resourcemanager and nodemanager, I get following error when
i try to hit Hadoop UI *���)org**.apache.hadoop.**ipc.RPC$Version**
Mismatch���Ser**ver IPC version 5 cannot communicate with client version 47
*.
There is no significant error in Hadoop logs (it shows everything started
successfully).

Do you have any idea about this error?

Thanks,

-- 

Nitin Khandelwal


Re: Problem running Hadoop 0.23.0

2011-11-28 Thread Nitin Khandelwal
Thanks, Tom
It worked..

Thanks,
Nitin

On 28 November 2011 22:12, Tom White t...@cloudera.com wrote:

 Hi Nitin,

 It looks like you may be using the wrong port number - try 8088 for
 the resource manager UI.

 Cheers,
 Tom

 On Mon, Nov 28, 2011 at 4:02 AM, Nitin Khandelwal
 nitin.khandel...@germinait.com wrote:
  Hi,
 
  I was trying to setup Hadoop 0.23.0 with help of
 
 http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/SingleCluster.html
 .
  After starting resourcemanager and nodemanager, I get following error
 when
  i try to hit Hadoop UI
 *���)org**.apache.hadoop.**ipc.RPC$Version**
  Mismatch���Ser**ver IPC version 5 cannot communicate with client
 version 47
  *.
  There is no significant error in Hadoop logs (it shows everything started
  successfully).
 
  Do you have any idea about this error?
 
  Thanks,
 
  --
 
  Nitin Khandelwal
 




-- 


Nitin Khandelwal


Problem in Getting Job Status

2011-09-05 Thread Nitin Khandelwal
Hi All,

We are migrating from Hadoop 0.19.2 to Hadoop 0.20.203. The normal flow (
job posting / running ) works fine.  In one of the flow, we are required to
check Task Reports for a running Job to fetch its state.

In the older code (0.19.2)  we had following piece of code which does the
required tasks

runningJob is an instance of Job

TaskReport[] reports = jobClient.getMapTaskReports(
runningJob.getID());

Wanted to know how it will change in the newer version. I tried with

TaskReport[] reports = jobClient.getMapTaskReports((JobID)
runningJob.getJobID());

This did not work as there is no JobID is set in Job instance.

Requesting your assistance in this.

Let me know if any other details are required.

Thanks.
-- 

Nitin Khandelwal


Re: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Nitin Khandelwal
Plz change ur* fs.default.name* to hdfs://192.168.1.101:9000
Thanks,
Nitin

On 28 July 2011 17:46, Doan Ninh uitnetw...@gmail.com wrote:

 In the first time, i use *hadoop-cluster-1* for 192.168.1.101.
 That is the hostname of the master node.
 But, the same error occurs
 How can i fix it?

 On Thu, Jul 28, 2011 at 7:07 PM, madhu phatak phatak@gmail.com
 wrote:

  I had issue using IP address in XML files . You can try to use host names
  in
  the place of IP address .
 
  On Thu, Jul 28, 2011 at 5:22 PM, Doan Ninh uitnetw...@gmail.com wrote:
 
   Hi,
  
   I run Hadoop in 4 Ubuntu 11.04 on VirtualBox.
   On the master node (192.168.1.101), I configure fs.default.name =
  hdfs://
   127.0.0.1:9000. Then i configure everything on 3 other node
   When i start the cluster by entering $HADOOP_HOME/bin/start-all.sh on
  the
   master node
   Everything is ok, but the slave can't connect to the master on 9000,
 9001
   port.
   I manually telnet to 192.168.1.101 in 9000, 9001. And the result is
   connection refused
   Then, i'm on the master node, telnet to localhost, 127.0.0.1:9000. The
   result is connected.
   But, on the master node, i telnet to 192.168.1.101:9000 = Connection
   Refused
  
   Can somebody help me?
  
 




-- 


Nitin Khandelwal


Hadoop Question

2011-07-27 Thread Nitin Khandelwal
Hi All,

How can I determine if a file is being written to (by any thread) in HDFS. I
have a continuous process on the master node, which is tracking a particular
folder in HDFS for files to process. On the slave nodes, I am creating files
in the same folder using the following code :

At the slave node:

import org.apache.commons.io.IOUtils;
import org.apache.hadoop.fs.FileSystem;
import java.io.OutputStream;

OutputStream oStream = fileSystem.create(path);
IOUtils.write(Some String, oStream);
IOUtils.closeQuietly(oStream);


At the master node,
I am getting the earliest modified file in the folder. At times when I try
reading the file, I get nothing in the file, mostly because the slave might
be still finishing writing to the file. Is there any way, to somehow tell
the master, that the slave is still writing to the file and to check the
file sometime later for actual content.

Thanks,
-- 


Nitin Khandelwal


Job Priority Hadoop 0.20.203

2011-07-06 Thread Nitin Khandelwal
Hi,

I am using Hadoop 0.20.203 with the new API ( mapreduce package) . I want to
use Jobpriority, but unfortunately there is no option to set that in Job (
the option is there in 0.21.0). Can somebody plz tell me is there is a
walkaround to set job priority?

Thanks,

-- 

Nitin Khandelwal


Isssue with Job Scheduling

2011-04-13 Thread Nitin Khandelwal
Hi,

I want ot use Capacity Scheduler for my Hadoop Jobs. I have currently three
Queues defined and are configured and working properly. I am using Hadoop
0.20.2 And in the new library, we are not supposed to use JobConf. So, I
need to set Queue name as a property in Configuration (
mapred.job.queue.name ) . I want to use a single instance of Configuration
from Spring ( since i  have some other properties like plugin info. ebbedded
in Configuration) . I tried DeepCloning ( throes an exception as it tries to
make clone of Logger too) and Shalloe Cloning ( not working properly, may be
beacuse properties in Configuration are loaded in object when called for)  .
Could you tell me what is the best approach to use Capacity Scheduler??

Thanks And Regards,

-- 

Nitin Khandelwal


Re: Lost Task Tracker because of no heartbeat

2011-03-16 Thread Nitin Khandelwal
Hi,
Just do context.progress() after small interval of time inside Your
Map/reduce. That will do. If you are using Older package then, you can use
reporter.progress().

Thanks  Regards,
Nitin Khandelwal

On 16 March 2011 21:30, Baran_Cakici barancak...@gmail.com wrote:


 Hi Everyone,

 I make a Project with Hadoop-MapRedeuce for my master-Thesis. I have a
 strange problem on my System.

 First of all, I use Hadoop-0.20.2 on Windows XP Pro with Eclipse Plug-In.
 When I start a job with big Input(4GB - it`s may be not to big, but
 algorithm require some time), then i lose my Task Tracker in several
 minutes
 or seconds. I mean, Seconds since heartbeat increase
 and then after 600 Seconds I lose TaskTracker.

 I read somewhere, that can be occured because of small number of open files
 (ulimit -n). I try to increase this value, but i can write as max value in
 Cygwin 3200.(ulimit -n 3200) and default value is 256. Actually I don`t
 know, is it helps or not.

 In my job and task tracker.log have I some Errors, I posted those to.

 Jobtracker.log

 -Call to localhost/127.0.0.1:9000 failed on local exception:
 java.io.IOException: An existing connection was forcibly closed by the
 remote host

 another :
 -
 2011-03-15 12:13:30,718 INFO org.apache.hadoop.mapred.JobTracker:
 attempt_201103151143_0002_m_91_0 is 97125 ms debug.
 2011-03-15 12:16:50,718 INFO org.apache.hadoop.mapred.JobTracker:
 attempt_201103151143_0002_m_91_0 is 297125 ms debug.
 2011-03-15 12:20:10,718 INFO org.apache.hadoop.mapred.JobTracker:
 attempt_201103151143_0002_m_91_0 is 497125 ms debug.
 2011-03-15 12:23:30,718 INFO org.apache.hadoop.mapred.JobTracker:
 attempt_201103151143_0002_m_91_0 is 697125 ms debug.

 Error launching task
 Lost tracker 'tracker_apple:localhost/127.0.0.1:2654'

 there are my logs(jobtracker.log, tasktracker.log ...) in attachment

 I need really Help, I don`t have so much time for my Thessis.

 Thanks a lot for your Helps,

 Baran

 http://old.nabble.com/file/p31164785/logs.rar logs.rar
 --
 View this message in context:
 http://old.nabble.com/Lost-Task-Tracker-because-of-no-heartbeat-tp31164785p31164785.html
 Sent from the Hadoop core-user mailing list archive at Nabble.com.




-- 


Nitin Khandelwal


Re: How to make zip files as Hadoop input

2011-03-01 Thread Nitin Khandelwal
Hi,
You can actually make your own input format and reader which will read one
file from a directory and give it to a node. If You are using hadoop 0.19
then extending MultiFilesplit format can do this task for you . If you are
using Hadoop 0.20 or greater then your  your inputformat can extend
fileInputformat and yor reader can extend recordreader.
Thanks and Regards,
Nitin

On 2 March 2011 11:15, Mark Kerzner markkerz...@gmail.com wrote:

 Hi,

 I have a bunch of zip files that I want to serve as input to a MapReduce
 job. My initial design was to list them in a text file and then give this
 list file as input. The list file would be read, and each line would be
 handed off to a node to process, which would pick up the corresponding zip
 file and work on it.

 But I feel that a better design is possible, and that my way is redundant.
 Can I just give the input directory as input? How do I make sure each node
 gets a file to process?

 Thank you,
 Mark




-- 


Nitin Khandelwal


Re: Map Task Fails.........

2011-02-20 Thread Nitin Khandelwal
Hi,
Thanks for taking time to  read my mail. Actually i get error only in master
which is both master and slave . If i try to run only on one machine (
master and slave itself)  then also i get error

On 16 February 2011 20:00, madhu phatak phatak@gmail.com wrote:

 tasktracker log *
 On Wed, Feb 16, 2011 at 8:00 PM, madhu phatak phatak@gmail.com
 wrote:

  See the tasklog  of the slave to see why the task attempt is failing...



 
 
  On Wed, Feb 16, 2011 at 7:29 PM, Nitin Khandelwal 
  nitin.khandel...@germinait.com wrote:
 
  Hi,
  I am using Hadoop 0.21.0. I am getting Exception as
  java.lang.Throwable: Child Error at
  org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:249) Caused by:
  java.io.IOException: Task process exit with nonzero status of 1. at
  org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:236)
  when i am trying to run Map Red. This error comes in one of the slaves (
  which is also master in my case) . Cam any body tell why i may be
 getting
  this?
  Thanks,
 
 
  --
 
 
  Nitin Khandelwal
 
 
 




-- 


Nitin Khandelwal


Map Task Fails.........

2011-02-16 Thread Nitin Khandelwal
Hi,
I am using Hadoop 0.21.0. I am getting Exception as
java.lang.Throwable: Child Error at
org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:249) Caused by:
java.io.IOException: Task process exit with nonzero status of 1. at
org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:236)
when i am trying to run Map Red. This error comes in one of the slaves (
which is also master in my case) . Cam any body tell why i may be getting
this?
Thanks,


-- 


Nitin Khandelwal


Custom Input format...

2011-02-11 Thread Nitin Khandelwal
Hi,
I want to give a folder as input path to Map Red. Each Task should read one
file out of that folder at once . i was using it before in 0.19 using
multiFileSplit Format and my own Input format extending it. can u plz tell
how to do the same in 0.21.0.

Thanks,
-- 


Nitin Khandelwal