Re: Modeling WordCount in a different way

2009-04-13 Thread Pankil Doshi
Hey

Did u find any class or way out for storing results of Job1 map/reduce in
memory and using that as an input to job2 map/Reduce?I am facing a situation
where I need to do similar thing.If anyone can help me out..

Pankil

On Wed, Apr 8, 2009 at 12:51 AM, Sharad Agarwal shara...@yahoo-inc.comwrote:


  I have confusion how would I start the next job after finishing the one,
  could you just make it clear by some rough example.
 See JobControl class to chain the jobs. You can specify dependencies as
 well. You can checkout the TestJobControl class  for example code.
 
  Also do I need to use
  SequenceFileInputFormat to maintain the results in the memory and then
  accessing it.
 
 Not really. You have to use the corresponding reader to read the data. For
 example if you have written it using TextOutputFormat(default), you can then
 read it using TextInputFormat. The reader can be created in the reducer
 initialization code. In the new api (org.apache.hadoop.mapreduce.Reducer) it
 can be done in setup method. Here you can load the word,count mappings in
 a HashMap.
 In case you don't want to load all data in memory, you can create the
 reader in setup method and keep on doing the next
 (LineRecordReader#nextKeyValue())  in the reduce function if the reduce key
 is greater than the current key from the reader.

 - Sharad



Re: Modeling WordCount in a different way

2009-04-14 Thread Pankil Doshi
Hey,

I am trying complex queries on hadoop and in which i require more than one
job to run to get final result..results of job one captures few joins of the
query and I want to pass those results as input to 2nd job and again do
processing so that I can get final results.queries are such that I cant do
all types of joins and filterin in job1 and so I require two jobs.

right now I write results of job 1 to hdfs and read dem for job2..but thats
take unecessary IO time.So was looking for something that I can store my
results of job1 in memory and use them as input for job 2.

do let me know if you need any  more details.
Pankil

On Mon, Apr 13, 2009 at 9:51 PM, sharad agarwal shara...@yahoo-inc.comwrote:

 Pankil Doshi wrote:


 Hey

 Did u find any class or way out for storing results of Job1 map/reduce in
 memory and using that as an input to job2 map/Reduce?I am facing a
 situation
 where I need to do similar thing.If anyone can help me out..

  Normally you would write the job output to a file and input that to the
 next job.
 Any reason why you want to store the map reduce output in memory ? If you
 can describe your problem, perhaps it could be solved in more mapreduce-ish
 way.

 - Sharad



Directory /tmp/hadoop-hadoop/dfs/name is in an inconsistent state: storage directory does not exist

2009-04-14 Thread Pankil Doshi
Hello Everyone,

At time I get following error,when i restart my cluster desktops.(Before
that I shutdown mapred and dfs properly though).
Temp folder contains of the directory its looking for.Still I get this
error.
Only solution I found to get rid with this error is I have to format my dfs
entirely and then load the data again. and start whole process.

But in that I loose my data on HDFS and I have to reload it.

Does anyone has any clue abt it?

Error from log fil e:-

2009-04-14 19:40:29,963 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG:
/
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = Semantic002/192.168.1.133
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.18.3
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r 736250;
compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009
/
2009-04-14 19:40:30,958 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
Initializing RPC Metrics with hostName=NameNode, port=9000
2009-04-14 19:40:30,996 INFO org.apache.hadoop.dfs.NameNode: Namenode up at:
Semantic002/192.168.1.133:9000
2009-04-14 19:40:31,007 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
Initializing JVM Metrics with processName=NameNode, sessionId=null
2009-04-14 19:40:31,014 INFO org.apache.hadoop.dfs.NameNodeMetrics:
Initializing NameNodeMeterics using context
object:org.apache.hadoop.metrics.spi.NullCont
ext
2009-04-14 19:40:31,160 INFO org.apache.hadoop.fs.FSNamesystem:
fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,plugdev,scanner,fuse,admin
2009-04-14 19:40:31,161 INFO org.apache.hadoop.fs.FSNamesystem:
supergroup=supergroup
2009-04-14 19:40:31,161 INFO org.apache.hadoop.fs.FSNamesystem:
isPermissionEnabled=true
2009-04-14 19:40:31,183 INFO org.apache.hadoop.dfs.FSNamesystemMetrics:
Initializing FSNamesystemMeterics using context
object:org.apache.hadoop.metrics.spi.
NullContext
2009-04-14 19:40:31,184 INFO org.apache.hadoop.fs.FSNamesystem: Registered
FSNamesystemStatusMBean
2009-04-14 19:40:31,248 INFO org.apache.hadoop.dfs.Storage: Storage
directory /tmp/hadoop-hadoop/dfs/name does not exist.
2009-04-14 19:40:31,251 ERROR org.apache.hadoop.fs.FSNamesystem:
FSNamesystem initialization failed.
org.apache.hadoop.dfs.InconsistentFSStateException: Directory
/tmp/hadoop-hadoop/dfs/name is in an inconsistent state: storage directory
does not exist or is
 not accessible.
at
org.apache.hadoop.dfs.FSImage.recoverTransitionRead(FSImage.java:211)
at
org.apache.hadoop.dfs.FSDirectory.loadFSImage(FSDirectory.java:80)
at
org.apache.hadoop.dfs.FSNamesystem.initialize(FSNamesystem.java:294)
at org.apache.hadoop.dfs.FSNamesystem.init(FSNamesystem.java:273)
at org.apache.hadoop.dfs.NameNode.initialize(NameNode.java:148)
at org.apache.hadoop.dfs.NameNode.init(NameNode.java:193)
at org.apache.hadoop.dfs.NameNode.init(NameNode.java:179)
at org.apache.hadoop.dfs.NameNode.createNameNode(NameNode.java:830)
at org.apache.hadoop.dfs.NameNode.main(NameNode.java:839)
2009-04-14 19:40:31,261 INFO org.apache.hadoop.ipc.Server: Stopping server
on 9000
2009-04-14 19:40:31,262 ERROR org.apache.hadoop.dfs.NameNode:
org.apache.hadoop.dfs.InconsistentFSStateException: Directory
/tmp/hadoop-hadoop/dfs/name is in
 an inconsistent state: storage directory does not exist or is not
accessible.
at
org.apache.hadoop.dfs.FSImage.recoverTransitionRead(FSImage.java:211)
at
org.apache.hadoop.dfs.FSDirectory.loadFSImage(FSDirectory.java:80)
at
org.apache.hadoop.dfs.FSNamesystem.initialize(FSNamesystem.java:294)
at org.apache.hadoop.dfs.FSNamesystem.init(FSNamesystem.java:273)
at org.apache.hadoop.dfs.NameNode.initialize(NameNode.java:148)
at org.apache.hadoop.dfs.NameNode.init(NameNode.java:193)
at org.apache.hadoop.dfs.NameNode.init(NameNode.java:179)
at org.apache.hadoop.dfs.NameNode.createNameNode(NameNode.java:830)
at org.apache.hadoop.dfs.NameNode.main(NameNode.java:839)

2009-04-14 19:40:31,267 INFO org.apache.hadoop.dfs.NameNode: SHUTDOWN_MSG:
/
:

Thanks

Pankil


Re: Modeling WordCount in a different way

2009-04-15 Thread Pankil Doshi
On Wed, Apr 15, 2009 at 1:26 AM, Sharad Agarwal shara...@yahoo-inc.comwrote:



  I am trying complex queries on hadoop and in which i require more than
 one
  job to run to get final result..results of job one captures few joins of
 the
  query and I want to pass those results as input to 2nd job and again do
  processing so that I can get final results.queries are such that I cant
 do
  all types of joins and filterin in job1 and so I require two jobs.
 
  right now I write results of job 1 to hdfs and read dem for job2..but
 thats
  take unecessary IO time.So was looking for something that I can store my
  results of job1 in memory and use them as input for job 2.
 
  do let me know if you need any  more details.
 How big is your input and output data ?

And my total data is of 7.8 gb out of which for Job 1 i use around 3
gb.output of job1 is of about 1gb and I use this output as input to job 2.


 How many nodes you are using?

Well Right now due to lack of Resources I have only 4 nodes each dual core
processors with 1GB og ram and about 80gb hard
disk in  each..


 What is your job runtime?

My first jobs takes long time after reaching 90% of reduce phase as it does
in-memory merge sort and so that is also an big issue.I will have to arrange
for more memory for my clusters I suppose.

I will have look at jvm reuse feature. thanks



 Pankil


Re: hadoop-a small doubt

2009-04-15 Thread Pankil Doshi

Hey ,
You can do that.That system should have same usrname like those of cluster
and ofcourse it  should be able to ssh name node.Also it should have hadoop
and its hadoop-site.xml should be similar .Then u can access namenode,hdfs
etc.

if you are willing to see the web interface that can be done easily using
any system.

deepya wrote:
 
 Hi,
I am SreeDeepya doing MTech in IIIT.I am working on a project named
 cost effective and scalable storage server.I configured a small hadoop
 cluster with only two nodes one namenode and one datanode.I am new to
 hadoop.
 I have a small doubt.
 
 Can a system not in the hadoop cluster access the namenode or the
 datanodeIf yes,then can you please tell me the necessary
 configurations that has to be done.
 
 Thanks in advance.
 
 SreeDeepya
 

-- 
View this message in context: 
http://www.nabble.com/hadoop-a-small-doubt-tp22764615p23061794.html
Sent from the Hadoop lucene-users mailing list archive at Nabble.com.



Re: Directory /tmp/hadoop-hadoop/dfs/name is in an inconsistent state: storage directory does not exist

2009-04-15 Thread Pankil Doshi
Thanks

Pankil

On Wed, Apr 15, 2009 at 5:09 PM, Alex Loddengaard a...@cloudera.com wrote:

 Data stored to /tmp has no consistency / reliability guarantees.  Your OS
 can delete that data at any time.

 Configure hadoop-site.xml to store data elsewhere.  Grep for /tmp in
 hadoop-default.xml to see all the configuration options you'll have to
 change.  Here's the list I came up with:

 hadoop.tmp.dir
 fs.checkpoint.dir
 dfs.name.dir
 dfs.data.dir
 mapred.local.dir
 mapred.system.dir
 mapred.temp.dir

 Again, you need to be storing your data somewhere other than /tmp.

 Alex

 On Tue, Apr 14, 2009 at 6:06 PM, Pankil Doshi forpan...@gmail.com wrote:

  Hello Everyone,
 
  At time I get following error,when i restart my cluster desktops.(Before
  that I shutdown mapred and dfs properly though).
  Temp folder contains of the directory its looking for.Still I get this
  error.
  Only solution I found to get rid with this error is I have to format my
 dfs
  entirely and then load the data again. and start whole process.
 
  But in that I loose my data on HDFS and I have to reload it.
 
  Does anyone has any clue abt it?
 
  Error from log fil e:-
 
  2009-04-14 19:40:29,963 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG:
  /
  STARTUP_MSG: Starting NameNode
  STARTUP_MSG:   host = Semantic002/192.168.1.133
  STARTUP_MSG:   args = []
  STARTUP_MSG:   version = 0.18.3
  STARTUP_MSG:   build =
  https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r
  736250;
  compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009
  /
  2009-04-14 19:40:30,958 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
  Initializing RPC Metrics with hostName=NameNode, port=9000
  2009-04-14 19:40:30,996 INFO org.apache.hadoop.dfs.NameNode: Namenode up
  at:
  Semantic002/192.168.1.133:9000
  2009-04-14 19:40:31,007 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
  Initializing JVM Metrics with processName=NameNode, sessionId=null
  2009-04-14 19:40:31,014 INFO org.apache.hadoop.dfs.NameNodeMetrics:
  Initializing NameNodeMeterics using context
  object:org.apache.hadoop.metrics.spi.NullCont
  ext
  2009-04-14 19:40:31,160 INFO org.apache.hadoop.fs.FSNamesystem:
 
 
 fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,plugdev,scanner,fuse,admin
  2009-04-14 19:40:31,161 INFO org.apache.hadoop.fs.FSNamesystem:
  supergroup=supergroup
  2009-04-14 19:40:31,161 INFO org.apache.hadoop.fs.FSNamesystem:
  isPermissionEnabled=true
  2009-04-14 19:40:31,183 INFO org.apache.hadoop.dfs.FSNamesystemMetrics:
  Initializing FSNamesystemMeterics using context
  object:org.apache.hadoop.metrics.spi.
  NullContext
  2009-04-14 19:40:31,184 INFO org.apache.hadoop.fs.FSNamesystem:
 Registered
  FSNamesystemStatusMBean
  2009-04-14 19:40:31,248 INFO org.apache.hadoop.dfs.Storage: Storage
  directory /tmp/hadoop-hadoop/dfs/name does not exist.
  2009-04-14 19:40:31,251 ERROR org.apache.hadoop.fs.FSNamesystem:
  FSNamesystem initialization failed.
  org.apache.hadoop.dfs.InconsistentFSStateException: Directory
  /tmp/hadoop-hadoop/dfs/name is in an inconsistent state: storage
 directory
  does not exist or is
   not accessible.
 at
  org.apache.hadoop.dfs.FSImage.recoverTransitionRead(FSImage.java:211)
 at
  org.apache.hadoop.dfs.FSDirectory.loadFSImage(FSDirectory.java:80)
 at
  org.apache.hadoop.dfs.FSNamesystem.initialize(FSNamesystem.java:294)
 at
 org.apache.hadoop.dfs.FSNamesystem.init(FSNamesystem.java:273)
 at org.apache.hadoop.dfs.NameNode.initialize(NameNode.java:148)
 at org.apache.hadoop.dfs.NameNode.init(NameNode.java:193)
 at org.apache.hadoop.dfs.NameNode.init(NameNode.java:179)
 at
 org.apache.hadoop.dfs.NameNode.createNameNode(NameNode.java:830)
 at org.apache.hadoop.dfs.NameNode.main(NameNode.java:839)
  2009-04-14 19:40:31,261 INFO org.apache.hadoop.ipc.Server: Stopping
 server
  on 9000
  2009-04-14 19:40:31,262 ERROR org.apache.hadoop.dfs.NameNode:
  org.apache.hadoop.dfs.InconsistentFSStateException: Directory
  /tmp/hadoop-hadoop/dfs/name is in
   an inconsistent state: storage directory does not exist or is not
  accessible.
 at
  org.apache.hadoop.dfs.FSImage.recoverTransitionRead(FSImage.java:211)
 at
  org.apache.hadoop.dfs.FSDirectory.loadFSImage(FSDirectory.java:80)
 at
  org.apache.hadoop.dfs.FSNamesystem.initialize(FSNamesystem.java:294)
 at
 org.apache.hadoop.dfs.FSNamesystem.init(FSNamesystem.java:273)
 at org.apache.hadoop.dfs.NameNode.initialize(NameNode.java:148)
 at org.apache.hadoop.dfs.NameNode.init(NameNode.java:193)
 at org.apache.hadoop.dfs.NameNode.init(NameNode.java:179)
 at
 org.apache.hadoop.dfs.NameNode.createNameNode(NameNode.java:830)
 at org.apache.hadoop.dfs.NameNode.main(NameNode.java:839)
 
  2009-04-14 19:40:31,267 INFO

Re: No space left on device Exception

2009-04-16 Thread Pankil Doshi
Hey

what's your input size?

from the info you gave it seems you have used 4.2GB and so probably if thats
your input size your intermediate results mostly is less then your input.but
that too depends on your map function. Make sure about the size of
intermediate results.

Pankil

On Thu, Apr 16, 2009 at 3:25 AM, Rakhi Khatwani rakhi.khatw...@gmail.comwrote:

 Thanks,
  I will check tht

 Regards,
 Raakhi

 On Thu, Apr 16, 2009 at 1:42 PM, Miles Osborne mi...@inf.ed.ac.uk wrote:

  it may be that intermediate results are filling your disks and when
  the jobs crash, this all gets deleted.  so it would look like you have
  spare space when in reality you don't.
 
  i would check on the file system as your jobs run and see if indeed
  they are filling-up.
 
  Miles
 
  2009/4/16 Rakhi Khatwani rakhi.khatw...@gmail.com:
   Hi,
  following is the output on the df command
   [r...@domu-12-31-39-00-e5-d2 conf]# df -h
   FilesystemSize  Used Avail Use% Mounted on
   /dev/sda1 9.9G  4.2G  5.2G  45% /
   /dev/sdb  414G  924M  392G   1% /mnt
  
   from the o/p it seems that i have quite an amount of memory available.
  but i
   still get the exception :(
  
   Thanks
   Raakhi
  
   On Thu, Apr 16, 2009 at 1:18 PM, Desai, Milind B milind.de...@hp.com
  wrote:
  
   From the exception it appears that there is no space left on machine.
  You
   can check using 'df'
  
   Thanks
   Milind
  
   -Original Message-
   From: Rakhi Khatwani [mailto:rakhi.khatw...@gmail.com]
   Sent: Thursday, April 16, 2009 1:15 PM
   To: hbase-u...@hadoop.apache.org; core-user@hadoop.apache.org
   Subject: No space left on device Exception
  
   Hi,
   I am running a map-reduce program on 6-Node ec2 cluster. and after
 a
   couple of hours all my tasks gets hanged.
  
   so i started digging into the logs
  
   there were no logs for regionserver
   no logs for tasktracker.
   However for jobtracker i get the following:
  
   2009-04-16 03:00:29,691 INFO org.apache.hadoop.ipc.Server: IPC Server
   handler 9 on 50002, call
   heartbeat(org.apache.hadoop.mapred.tasktrackersta...@2eed7d11, false,
   true,
   10745) from 10.254.27.79:44222: error: java.io.IOException:
   org.apache.hadoop.fs.FSError: java.io.IOException: No space left on
  device
   java.io.IOException: org.apache.hadoop.fs.FSError:
 java.io.IOException:
  No
   space left on device
 at
  
  
 
 org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:199)
 at
   java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
 at
  java.io.BufferedOutputStream.write(BufferedOutputStream.java:109)
 at
  
  
 
 org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47)
 at java.io.DataOutputStream.write(DataOutputStream.java:90)
 at
  
  
 
 org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.writeChunk(ChecksumFileSystem.java:346)
 at
  
  
 
 org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:150)
 at
   org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:100)
 at
  org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:86)
 at
  
  
 
 org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47)
 at java.io.DataOutputStream.write(DataOutputStream.java:90)
 at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:202)
 at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:297)
 at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:130)
 at java.io.OutputStreamWriter.close(OutputStreamWriter.java:216)
 at java.io.BufferedWriter.close(BufferedWriter.java:248)
 at java.io.PrintWriter.close(PrintWriter.java:295)
 at
  
  
 
 org.apache.hadoop.mapred.JobHistory$JobInfo.logFinished(JobHistory.java:1024)
 at
  
 
 org.apache.hadoop.mapred.JobInProgress.jobComplete(JobInProgress.java:1906)
 at org.apache.hadoop.mapred.JobInProgress.comp
  
  
  
   following are the disk information on dfs UI
   domU-12-31-39-00-0C-A1
  
 
 http://domu-12-31-39-00-0c-a1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070dir=%2F
   0In
   Service413.380.8321.19391.360.2
   94.672353 domU-12-31-39-00-16-F1
  
 
 http://domu-12-31-39-00-16-f1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070dir=%2F
   1In
   Service413.380.4621.24391.670.11
   94.752399 domU-12-31-39-00-45-71
  
 
 http://domu-12-31-39-00-45-71.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070dir=%2F
   1In
   Service413.380.6421.34391.40.16
   94.682303 domU-12-31-39-00-E5-D2
  
 
 http://domu-12-31-39-00-e5-d2.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070dir=%2F
   0In
   Service413.380.6621.53391.180.16
   94.632319 domU-12-31-39-01-64-12
  
 
 

Re: HDFS files naming convention

2009-04-25 Thread Pankil Doshi
hey

you can surely do that using MulipleFileOutputFormat. We have already
implemented that

Pankil

On Fri, Apr 24, 2009 at 8:58 PM, Aaron Kimball aa...@cloudera.com wrote:

 Alternatively, just use FileSystem.rename() on the normal output files
 after
 reducing is complete?

 On Sat, Apr 25, 2009 at 10:57 AM, Aaron Kimball aa...@cloudera.com
 wrote:

  I think you can do this with MultiFileOutputFormat; specifically override
  the getLeafFileName() method.
  - Aaron
 
 
  On Sat, Apr 25, 2009 at 4:29 AM, Parul Kudtarkar 
  parul_kudtar...@hms.harvard.edu wrote:
 
 
  The HDFS files generated after mapreduce run are strored in HDFS as
  part-0 and so on.part-n
 
  Is it possible to name these output files stored in HDFS as per my own
  convention i.e. I would like to name these files my_file_1 and so
  onmy_files_n
 
  Please advice how this can be achieved?
 
  Thanks,
  Parul V. Kudtarkar
  --
  View this message in context:
 
 http://www.nabble.com/HDFS-files-naming-convention-tp23223348p23223348.html
  Sent from the Hadoop core-user mailing list archive at Nabble.com.
 
 
 



Datanodes fail to start

2009-05-14 Thread Pankil Doshi
Hello Everyone,

Actually I had a cluster which was up.

But i stopped the cluster as i  wanted to format it.But cant start it back.

1)when i give start-dfs.sh I get following on screen

starting namenode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-namenode-hadoopmaster.out
slave1.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave1.out
slave3.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave3.out
slave4.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave4.out
slave2.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave2.out
slave5.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave5.out
slave6.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave6.out
slave9.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave9.out
slave8.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave8.out
slave7.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave7.out
slave10.local: starting datanode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave10.out
hadoopmaster.local: starting secondarynamenode, logging to
/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-secondarynamenode-hadoopmaster.out


2) from log file named hadoop-hadoop-namenode-hadoopmaster.log I get
following



2009-05-14 20:28:23,515 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG:
/
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = hadoopmaster/127.0.0.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.18.3
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r 736250;
compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009
/
2009-05-14 20:28:23,717 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
Initializing RPC Metrics with hostName=NameNode, port=9000
2009-05-14 20:28:23,728 INFO org.apache.hadoop.dfs.NameNode: Namenode up at:
hadoopmaster.local/192.168.0.1:9000
2009-05-14 20:28:23,733 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
Initializing JVM Metrics with processName=NameNode, sessionId=null
2009-05-14 20:28:23,743 INFO org.apache.hadoop.dfs.NameNodeMetrics:
Initializing NameNodeMeterics using context
object:org.apache.hadoop.metrics.spi.NullContext
2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,video,plugdev,fuse,lpadmin,admin,sambashare
2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
supergroup=supergroup
2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
isPermissionEnabled=true
2009-05-14 20:28:23,883 INFO org.apache.hadoop.dfs.FSNamesystemMetrics:
Initializing FSNamesystemMeterics using context
object:org.apache.hadoop.metrics.spi.NullContext
2009-05-14 20:28:23,885 INFO org.apache.hadoop.fs.FSNamesystem: Registered
FSNamesystemStatusMBean
2009-05-14 20:28:23,964 INFO org.apache.hadoop.dfs.Storage: Number of files
= 1
2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Number of files
under construction = 0
2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Image file of
size 80 loaded in 0 seconds.
2009-05-14 20:28:23,972 INFO org.apache.hadoop.dfs.Storage: Edits file edits
of size 4 edits # 0 loaded in 0 seconds.
2009-05-14 20:28:23,974 INFO org.apache.hadoop.fs.FSNamesystem: Finished
loading FSImage in 155 msecs
2009-05-14 20:28:23,976 INFO org.apache.hadoop.fs.FSNamesystem: Total number
of blocks = 0
2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of
invalid blocks = 0
2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of
under-replicated blocks = 0
2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of
over-replicated blocks = 0
2009-05-14 20:28:23,988 INFO org.apache.hadoop.dfs.StateChange: STATE*
Leaving safe mode after 0 secs.
*2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE*
Network topology has 0 racks and 0 datanodes*
2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE*
UnderReplicatedBlocks has 0 blocks
2009-05-14 20:28:29,128 INFO org.mortbay.util.Credential: Checking Resource
aliases
2009-05-14 20:28:29,243 INFO org.mortbay.http.HttpServer: Version
Jetty/5.1.4
2009-05-14 20:28:29,244 INFO org.mortbay.util.Container: Started
HttpContext[/static,/static]
2009-05-14 20:28:29,245 INFO org.mortbay.util.Container: Started
HttpContext[/logs,/logs]
2009-05-14 20:28:29,750 INFO org.mortbay.util.Container: Started
org.mortbay.jetty.servlet.webapplicationhand...@7fcebc9f
2009-05-14 20:28:29,838 INFO 

Re: Datanodes fail to start

2009-05-14 Thread Pankil Doshi
Can u guide me where can I find datanode log files? As I cannot find it in
$hadoop/logs and so.

I can only find  following files in logs folder :-

hadoop-hadoop-namenode-hadoopmaster.log
   hadoop-hadoop-namenode-hadoopmaster.out
hadoop-hadoop-namenode-hadoopmaster.out.1
   hadoop-hadoop-secondarynamenode-hadoopmaster.log
hadoop-hadoop-secondarynamenode-hadoopmaster.out
hadoop-hadoop-secondarynamenode-hadoopmaster.out.1
history


Thanks
Pankil

On Thu, May 14, 2009 at 11:27 PM, jason hadoop jason.had...@gmail.comwrote:

 You have to examine the datanode log files
 the namenode does not start the datanodes, the start script does.
 The name node passively waits for the datanodes to connect to it.

 On Thu, May 14, 2009 at 6:43 PM, Pankil Doshi forpan...@gmail.com wrote:

  Hello Everyone,
 
  Actually I had a cluster which was up.
 
  But i stopped the cluster as i  wanted to format it.But cant start it
 back.
 
  1)when i give start-dfs.sh I get following on screen
 
  starting namenode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-namenode-hadoopmaster.out
  slave1.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave1.out
  slave3.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave3.out
  slave4.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave4.out
  slave2.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave2.out
  slave5.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave5.out
  slave6.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave6.out
  slave9.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave9.out
  slave8.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave8.out
  slave7.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave7.out
  slave10.local: starting datanode, logging to
  /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave10.out
  hadoopmaster.local: starting secondarynamenode, logging to
 
 
 /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-secondarynamenode-hadoopmaster.out
 
 
  2) from log file named hadoop-hadoop-namenode-hadoopmaster.log I get
  following
 
 
 
  2009-05-14 20:28:23,515 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG:
  /
  STARTUP_MSG: Starting NameNode
  STARTUP_MSG:   host = hadoopmaster/127.0.0.1
  STARTUP_MSG:   args = []
  STARTUP_MSG:   version = 0.18.3
  STARTUP_MSG:   build =
  https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r
  736250;
  compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009
  /
  2009-05-14 20:28:23,717 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
  Initializing RPC Metrics with hostName=NameNode, port=9000
  2009-05-14 20:28:23,728 INFO org.apache.hadoop.dfs.NameNode: Namenode up
  at:
  hadoopmaster.local/192.168.0.1:9000
  2009-05-14 20:28:23,733 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
  Initializing JVM Metrics with processName=NameNode, sessionId=null
  2009-05-14 20:28:23,743 INFO org.apache.hadoop.dfs.NameNodeMetrics:
  Initializing NameNodeMeterics using context
  object:org.apache.hadoop.metrics.spi.NullContext
  2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
 
 
 fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,video,plugdev,fuse,lpadmin,admin,sambashare
  2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
  supergroup=supergroup
  2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
  isPermissionEnabled=true
  2009-05-14 20:28:23,883 INFO org.apache.hadoop.dfs.FSNamesystemMetrics:
  Initializing FSNamesystemMeterics using context
  object:org.apache.hadoop.metrics.spi.NullContext
  2009-05-14 20:28:23,885 INFO org.apache.hadoop.fs.FSNamesystem:
 Registered
  FSNamesystemStatusMBean
  2009-05-14 20:28:23,964 INFO org.apache.hadoop.dfs.Storage: Number of
 files
  = 1
  2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Number of
 files
  under construction = 0
  2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Image file of
  size 80 loaded in 0 seconds.
  2009-05-14 20:28:23,972 INFO org.apache.hadoop.dfs.Storage: Edits file
  edits
  of size 4 edits # 0 loaded in 0 seconds.
  2009-05-14 20:28:23,974 INFO org.apache.hadoop.fs.FSNamesystem: Finished
  loading FSImage in 155 msecs
  2009-05-14 20:28:23,976 INFO org.apache.hadoop.fs.FSNamesystem: Total
  number
  of blocks = 0
  2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of
  invalid blocks = 0
  2009-05-14 20:28:23,988 INFO

Re: Datanodes fail to start

2009-05-14 Thread Pankil Doshi
This is log from datanode.


2009-05-14 00:36:14,559 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 01:36:15,768 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 8 msecs
2009-05-14 02:36:13,975 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 03:36:15,189 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 04:36:13,384 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 05:36:14,592 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 06:36:15,806 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 07:36:14,008 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 08:36:15,204 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 09:36:13,430 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 10:36:14,642 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 11:36:15,850 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 12:36:14,193 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 11 msecs
2009-05-14 13:36:15,454 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 14:36:13,662 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 15:36:14,930 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 13 msecs
2009-05-14 16:36:16,151 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 12 msecs
2009-05-14 17:36:14,407 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 9 msecs
2009-05-14 18:36:15,659 INFO org.apache.hadoop.dfs.DataNode: BlockReport of
82 blocks got processed in 10 msecs
2009-05-14 19:27:02,188 WARN org.apache.hadoop.dfs.DataNode:
java.io.IOException: Call to
hadoopmaster.utdallas.edu/10.110.95.61:9000failed on local except$
at org.apache.hadoop.ipc.Client.wrapException(Client.java:751)
at org.apache.hadoop.ipc.Client.call(Client.java:719)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
at org.apache.hadoop.dfs.$Proxy4.sendHeartbeat(Unknown Source)
at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:690)
at org.apache.hadoop.dfs.DataNode.run(DataNode.java:2967)
at java.lang.Thread.run(Thread.java:619)
Caused by: java.io.EOFException
at java.io.DataInputStream.readInt(DataInputStream.java:375)
at
org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:500)
at org.apache.hadoop.ipc.Client$Connection.run(Client.java:442)

2009-05-14 19:27:06,198 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: hadoopmaster.utdallas.edu/10.110.95.61:9000. Already tried 0
time(s).
2009-05-14 19:27:06,436 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG:
/
SHUTDOWN_MSG: Shutting down DataNode at Slave1/127.0.1.1
/
2009-05-14 19:27:21,737 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
/
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = Slave1/127.0.1.1


On Thu, May 14, 2009 at 11:43 PM, jason hadoop jason.had...@gmail.comwrote:

 The data node logs are on the datanode machines in the log directory.
 You may wish to buy my book and read chapter 4 on hdfs management.

 On Thu, May 14, 2009 at 9:39 PM, Pankil Doshi forpan...@gmail.com wrote:

  Can u guide me where can I find datanode log files? As I cannot find it
 in
  $hadoop/logs and so.
 
  I can only find  following files in logs folder :-
 
  hadoop-hadoop-namenode-hadoopmaster.log
 hadoop-hadoop-namenode-hadoopmaster.out
  hadoop-hadoop-namenode-hadoopmaster.out.1
hadoop-hadoop-secondarynamenode-hadoopmaster.log
  hadoop-hadoop-secondarynamenode-hadoopmaster.out
  hadoop-hadoop-secondarynamenode-hadoopmaster.out.1
 history
 
 
  Thanks
  Pankil
 
  On Thu, May 14, 2009 at 11:27 PM, jason hadoop jason.had...@gmail.com
  wrote:
 
   You have to examine the datanode log files
   the namenode does not start the datanodes, the start script does.
   The name node passively waits for the datanodes to connect to it.
  
   On Thu, May 14, 2009 at 6:43 PM, Pankil Doshi forpan...@gmail.com
  wrote:
  
Hello Everyone,
   
Actually I had a cluster which was up.
   
But i stopped the cluster as i  wanted to format it.But cant start it
   back.
   
1)when i give start-dfs.sh I get following on screen

Re: Datanodes fail to start

2009-05-15 Thread Pankil Doshi
I got the solution..

Namespace IDs where some how incompatible.So I had to clean data dir and
temp dir ,format the cluster and make a fresh start

Pankil

On Fri, May 15, 2009 at 2:25 AM, jason hadoop jason.had...@gmail.comwrote:

 There should be a few more lines at the end.
 We only want the part from last the STARTUP_MSG to the end

 On one of mine a successfull start looks like this:
 STARTUP_MSG: Starting DataNode
 STARTUP_MSG:   host = at/192.168.1.119
 STARTUP_MSG:   args = []
 STARTUP_MSG:   version = 0.19.1-dev
 STARTUP_MSG:   build =  -r ; compiled by 'jason' on Tue Mar 17 04:03:57 PDT
 2009
 /
 2009-03-17 03:08:11,884 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
 FSDatasetStatusMBean
 2009-03-17 03:08:11,886 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at
 50010
 2009-03-17 03:08:11,889 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
 1048576 bytes/s
 2009-03-17 03:08:12,142 INFO org.mortbay.http.HttpServer: Version
 Jetty/5.1.4
 2009-03-17 03:08:12,155 INFO org.mortbay.util.Credential: Checking Resource
 aliases
 2009-03-17 03:08:12,518 INFO org.mortbay.util.Container: Started
 org.mortbay.jetty.servlet.webapplicationhand...@1e184cb
 2009-03-17 03:08:12,578 INFO org.mortbay.util.Container: Started
 WebApplicationContext[/static,/static]
 2009-03-17 03:08:12,721 INFO org.mortbay.util.Container: Started
 org.mortbay.jetty.servlet.webapplicationhand...@1d9e282
 2009-03-17 03:08:12,722 INFO org.mortbay.util.Container: Started
 WebApplicationContext[/logs,/logs]
 2009-03-17 03:08:12,878 INFO org.mortbay.util.Container: Started
 org.mortbay.jetty.servlet.webapplicationhand...@14a75bb
 2009-03-17 03:08:12,884 INFO org.mortbay.util.Container: Started
 WebApplicationContext[/,/]
 2009-03-17 03:08:12,951 INFO org.mortbay.http.SocketListener: Started
 SocketListener on 0.0.0.0:50075
 2009-03-17 03:08:12,951 INFO org.mortbay.util.Container: Started
 org.mortbay.jetty.ser...@1358f03
 2009-03-17 03:08:12,957 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
 Initializing JVM Metrics with processName=DataNode, sessionId=null
 2009-03-17 03:08:13,242 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
 Initializing RPC Metrics with hostName=DataNode, port=50020
 2009-03-17 03:08:13,264 INFO org.apache.hadoop.ipc.Server: IPC Server
 Responder: starting
 2009-03-17 03:08:13,304 INFO org.apache.hadoop.ipc.Server: IPC Server
 listener on 50020: starting
 2009-03-17 03:08:13,343 INFO org.apache.hadoop.ipc.Server: IPC Server
 handler 0 on 50020: starting
 2009-03-17 03:08:13,343 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
 DatanodeRegistration(192.168.1.119:50010,
 storageID=DS-540597485-192.168.1.119-50010-1237022386925, infoPort=50075,
 ipcPort=50020)
 2009-03-17 03:08:13,344 INFO org.apache.hadoop.ipc.Server: IPC Server
 handler 1 on 50020: starting
 2009-03-17 03:08:13,344 INFO org.apache.hadoop.ipc.Server: IPC Server
 handler 2 on 50020: starting
 2009-03-17 03:08:13,351 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
 192.168.1.119:50010,
 storageID=DS-540597485-192.168.1.119-50010-1237022386925, infoPort=50075,
 ipcPort=50020)In DataNode.run, data =
 FSDataset{dirpath='/tmp/hadoop-0.19.0-jason/dfs/data/current'}
 2009-03-17 03:08:13,352 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: using BLOCKREPORT_INTERVAL
 of 360msec Initial delay: 0msec
 2009-03-17 03:08:13,391 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 14 blocks
 got processed in 27 msecs
 2009-03-17 03:08:13,392 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block
 scanner.



 On Thu, May 14, 2009 at 9:51 PM, Pankil Doshi forpan...@gmail.com wrote:

  This is log from datanode.
 
 
  2009-05-14 00:36:14,559 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 12 msecs
  2009-05-14 01:36:15,768 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 8 msecs
  2009-05-14 02:36:13,975 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 9 msecs
  2009-05-14 03:36:15,189 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 12 msecs
  2009-05-14 04:36:13,384 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 9 msecs
  2009-05-14 05:36:14,592 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 9 msecs
  2009-05-14 06:36:15,806 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 12 msecs
  2009-05-14 07:36:14,008 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 12 msecs
  2009-05-14 08:36:15,204 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks got processed in 9 msecs
  2009-05-14 09:36:13,430 INFO org.apache.hadoop.dfs.DataNode: BlockReport
 of
  82 blocks

ssh issues

2009-05-21 Thread Pankil Doshi
Hello everyone,

I got hint how to solve the problem where clusters have different
usernames.but now other problem I face is that i can ssh a machine by using
-i path/to key/ ..I cant ssh them directly but I will have to always pass
the key.

Now i face problem in ssh-ing my machines.Does anyone have any ideas how to
deal with that??

Regards
Pankil


Re: ssh issues

2009-05-21 Thread Pankil Doshi
The problem is that it also prompts for the pass phrase.

On Thu, May 21, 2009 at 2:14 PM, Brian Bockelman bbock...@cse.unl.eduwrote:

 Hey Pankil,

 Use ~/.ssh/config to set the default key location to the proper place for
 each host, if you're going down that route.

 I'd remind you that SSH is only used as a convenient method to launch
 daemons.  If you have a preferred way to start things up on your cluster,
 you can use that (I think most large clusters don't use ssh... could be
 wrong).

 Brian


 On May 21, 2009, at 2:07 PM, Pankil Doshi wrote:

  Hello everyone,

 I got hint how to solve the problem where clusters have different
 usernames.but now other problem I face is that i can ssh a machine by
 using
 -i path/to key/ ..I cant ssh them directly but I will have to always pass
 the key.

 Now i face problem in ssh-ing my machines.Does anyone have any ideas how
 to
 deal with that??

 Regards
 Pankil





Re: ssh issues

2009-05-22 Thread Pankil Doshi
Well i made ssh with passphares. as the system in which i need to login
requires ssh with pass phrases and those systems have to be part of my
cluster. and so I need a way where I can specify -i path/to key/ and
passphrase to hadoop in before hand.

Pankil

On Thu, May 21, 2009 at 9:35 PM, Aaron Kimball aa...@cloudera.com wrote:

 Pankil,

 That means that either you're using the wrong ssh key and it's falling back
 to password authentication, or else you created your ssh keys with
 passphrases attached; try making new ssh keys with ssh-keygen and
 distributing those to start again?

 - Aaron

 On Thu, May 21, 2009 at 3:49 PM, Pankil Doshi forpan...@gmail.com wrote:

  The problem is that it also prompts for the pass phrase.
 
  On Thu, May 21, 2009 at 2:14 PM, Brian Bockelman bbock...@cse.unl.edu
  wrote:
 
   Hey Pankil,
  
   Use ~/.ssh/config to set the default key location to the proper place
 for
   each host, if you're going down that route.
  
   I'd remind you that SSH is only used as a convenient method to launch
   daemons.  If you have a preferred way to start things up on your
 cluster,
   you can use that (I think most large clusters don't use ssh... could be
   wrong).
  
   Brian
  
  
   On May 21, 2009, at 2:07 PM, Pankil Doshi wrote:
  
Hello everyone,
  
   I got hint how to solve the problem where clusters have different
   usernames.but now other problem I face is that i can ssh a machine by
   using
   -i path/to key/ ..I cant ssh them directly but I will have to always
  pass
   the key.
  
   Now i face problem in ssh-ing my machines.Does anyone have any ideas
 how
   to
   deal with that??
  
   Regards
   Pankil
  
  
  
 



Re: org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread Pankil Doshi
make sure u can ping that data node and ssh it.


On Thu, May 28, 2009 at 12:02 PM, ashish pareek pareek...@gmail.com wrote:

 HI ,
 I am trying to step up a hadoop cluster on 512 MB machine and using
 hadoop 0.18 and  have followed procedure given in  apache hadoop site for
 hadoop cluster.
 I included  in conf/slaves two datanode i.e including the namenode
 vitrual machine and other machine virtual machine  . and have set up
 passwordless ssh between both virtual machines. But now problem is when
 is run command 

 bin/hadoop start-all.sh

 It start only one datanode on the same namenode vitrual machine but it
 doesn't start the datanode on other machine.

 in logs/hadoop-datanode  i get message


  INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already
  tried 1 time(s).
  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).

 .
 .
 .
 .
 .
 .
 .
 .
 .

 .
 .

 .


 So can any one help in solving this problem. :)

 Thanks

 Regards
 Ashish Pareek



Re: Hadoop scheduling question

2009-06-04 Thread Pankil Doshi
Hello Kristi,

I am Research Assistant at University of Texas at Dallas. We are working of
RDF data and we come across many joins in our queries. But We are not able
to carry out all joins in a single job..we also tried our hadoop code using
Pig scripts and found that for each join in PIG script new job is used.So
basically what i think its a sequential process to handle typesof join where
output of one job is required s an input to other one.

do let me know what you think about my view point.

Thanks
Pankil

On Thu, Jun 4, 2009 at 7:12 PM, Kristi Morton kmor...@cs.washington.eduwrote:

 Hi,

 I'm a Hadoop 17 user who is doing research with Prof. Magda Balazinska at
 the University of Washington on an improved progress indicator for Pig
 Latin.  We have a question regarding how Hadoop schedules Pig Latin queries
 with JOIN operators.  Does Hadoop schedule all MapReduce jobs in a script
 sequentially or does it ever schedule two MapReduce jobs in parallel.  For
 example, if the output of two Map-Reduce jobs is later joined and each of
 these jobs only needs a subset of the cluster resources, would they be
 scheduled in parallel or in series?

 I apologize if I sent this to the wrong list, but please let me know which
 list is most appropriate for this type of question.

 Thanks,
 Kristi




Re: Making sure the tmp directory is cleaned?

2009-06-22 Thread Pankil Doshi
Yes, If your job gets completed successfully .possibly it removes after
completion of both map and reduce tasks.

Pankil

On Mon, Jun 22, 2009 at 3:15 PM, Qin Gao q...@cs.cmu.edu wrote:

 Hi All,

 Do you know if the tmp directory on every map/reduce task will be deleted
 automatically after the map task finishes or will do I have to delete them?

 I mean the tmp directory that automatically created by on current
 directory.

 Thanks a lot
 --Q



Re: Problem in viewing WEB UI

2009-06-22 Thread Pankil Doshi
I am not sure but sometimes you might see that datanodes are working from
cmd prompt..
But actually when you look at the logs you find sme kind of error in
that..Check the logs of datanode..

Pankil

On Wed, Jun 17, 2009 at 1:42 AM, ashish pareek pareek...@gmail.com wrote:

 Hi,

  When I run command *bin/hadoop dfsadmin -report *it shows that 2
 datanodes are alive but when i try to http://hadoopmster:50070/ but the
 problem is that it opens doesnot opne
 http://hadoopmaster:50070/dfshealth.jsp page and throws *error HTTP: 404 .
 So why is't happening like this?
 *
 Regards,
 Ashish Pareek


  On Wed, Jun 17, 2009 at 10:06 AM, Sugandha Neaolekar 
 sugandha@gmail.com wrote:

  Well, You just have to specify the address in the URL address bar as::
  http://hadoopmaster:50070 U'll be able to see the web UI..!
 
 
  On Tue, Jun 16, 2009 at 7:17 PM, ashish pareek pareek...@gmail.com
 wrote:
 
  HI Sugandha,
 Hmmm your suggestion helped and Now I am able
  to run two datanode one on the same machine as name node and other
 on
  the different machine Thanks a lot :)
 
   But the problem is now I am not able to see web UI
 .
  for  both datanode and as well as name node
  should I have to consider some more things in the site.xml ? if so
 please
  help...
 
  Thanking you again,
  regards,
  Ashish Pareek.
 
  On Tue, Jun 16, 2009 at 3:10 PM, Sugandha Naolekar 
  sugandha@gmail.com wrote:
 
  hi,,!
 
 
  First of all, get your concepts clear of hadoop.
  You can refer to the following
 
  site::
 
 http://www.google.co.in/url?sa=tsource=webct=rescd=1url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster)ei=lGU3Spv2FZbLjAe19KmiDQusg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzAsig2=t2AJ_nf24SFtveN4PHS_TAhttp://www.google.co.in/url?sa=tsource=webct=rescd=1url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29ei=lGU3Spv2FZbLjAe19KmiDQusg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzAsig2=t2AJ_nf24SFtveN4PHS_TA
 
 http://www.google.co.in/url?sa=tsource=webct=rescd=1url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29ei=lGU3Spv2FZbLjAe19KmiDQusg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzAsig2=t2AJ_nf24SFtveN4PHS_TA
 
 
 
  I have small doubt whether in the mater.xml and slave.xml we can have
  same port numbers to both of them like
 
 
  for slave :
 
  property
  namefs.default.name/name
  valuehdfs://hadoopslave:
 
  9000/value
/property
 
 
   for master:::
 
  property
  namefs.default.name/name
  valuehdfs://hadoopmaster:9000/value
/property
 
 
 
  Well, any  two daemons or services can run on the same port unless,
 they
  are not run on the same machine.If you wish to run DN and NN on the
 same
  machine, their port numbers have to be different.
 
 
 
 
  On Tue, Jun 16, 2009 at 2:55 PM, ashish pareek pareek...@gmail.com
 wrote:
 
  HI sugandha,
 
 
 
  and one more thing can we have in slave:::
 
  property
namedfs.datanode.address/
 
  name
valuehadoopmaster:9000/value
  valuehadoopslave:9001/value
/property
 
 
 
  Also, fs,default.name is the tag which specifies the default
 filesystem.
  And generaLLY, it is run on namenode. So, it;s value has to be a
 namenode's
  address only and not slave's.
 
 
 
  Else if you have complete procedure for installing and running Hadoop
 in
  cluster can you please send it to me .. I need to step up hadoop
 with in
  two days and show it to my guide.Currently I am doing my masters.
 
  Thanks for your spending time
 
 
  Try for the above, and this should work!
 
 
 
  regards,
  Ashish Pareek
 
 
  On Tue, Jun 16, 2009 at 2:33 PM, Sugandha Naolekar 
  sugandha@gmail.com wrote:
 
  Following changes are to be done::
 
  Under master folder::
 
  - put slaves address as well under the values of
  tag(dfs.datanode.address)
 
  - You want to make namenode as datanode as well. As per your config
  file, you have specified hadoopmaster  in your slave file. If you
 don't want
  that, remove ti from slaves file.
 
  UNder slave folder::
 
  - put only slave's (the m/c where you intend to run your datanode)'s
  address.under datanode.address tag. Else
  it should go as such::
 
  property
namedfs.datanode.address/name
valuehadoopmaster:9000/value
  valuehadoopslave:9001/value
/property
 
  Also, your port numbers hould be different. the daemons NN,DN,JT,TT
  should run independently on different ports.
 
 
  On Tue, Jun 16, 2009 at 2:05 PM, Sugandha Naolekar 
  sugandha@gmail.com wrote:
 
 
 
  -- Forwarded message --
  From: ashish pareek pareek...@gmail.com
  Date: Tue, Jun 16, 2009 at 2:00 PM
  Subject: Re: org.apache.hadoop.ipc.client : trying connect to server
  failed
  To: Sugandha Naolekar sugandha@gmail.com
 
 
 
 
  On Tue, Jun 16, 2009 at 1:58 PM, ashish pareek pareek...@gmail.com
 wrote:
 
  

Re: Disk Usage Overhead of Hadoop Upgrade

2009-06-22 Thread Pankil Doshi
hi Stu,

which block conversion are you talking about? If you are talking abt block
size of data then it remains same in upgrade unless and until you change it.

Pankil

On Tue, Jun 16, 2009 at 5:16 PM, Stu Hood stuart.h...@rackspace.com wrote:

 Hey gang,

 We're preparing to upgrade our cluster from Hadoop 0.15.3 to 0.18.3.

 How much disk usage overhead can we expect from the block conversion before
 we finalize the upgrade? In the worst case, will the upgrade cause our disk
 usage to double?

 Thanks,

 Stu Hood
 Search Team Technical Lead
 Email  Apps Division, Rackspace Hosting




Re: HDFS out of space

2009-06-22 Thread Pankil Doshi
Hey Alex,

Will Hadoop balancer utility work in this case?

Pankil

On Mon, Jun 22, 2009 at 4:30 PM, Alex Loddengaard a...@cloudera.com wrote:

 Are you seeing any exceptions because of the disk being at 99% capacity?

 Hadoop should do something sane here and write new data to the disk with
 more capacity.  That said, it is ideal to be balanced.  As far as I know,
 there is no way to balance an individual DataNode's hard drives (Hadoop
 does
 round-robin scheduling when writing data).

 Alex

 On Mon, Jun 22, 2009 at 10:12 AM, Kris Jirapinyo kjirapi...@biz360.com
 wrote:

  Hi all,
 How does one handle a mount running out of space for HDFS?  We have
 two
  disks mounted on /mnt and /mnt2 respectively on one of the machines that
  are
  used for HDFS, and /mnt is at 99% while /mnt2 is at 30%.  Is there a way
 to
  tell the machine to balance itself out?  I know for the cluster, you can
  balance it using start-balancer.sh but I don't think that it will tell
 the
  individual machine to balance itself out.  Our hack right now would be
  just to delete the data on /mnt, since we have replication of 3x, we
 should
  be OK.  But I'd prefer not to do that.  Any thoughts?
 



Re: Making sure the tmp directory is cleaned?

2009-06-22 Thread Pankil Doshi
No..If your job gets killed or failed.Temp wont clean up.. and In that case
you will have to carefully clean that on your own. If you dont clean it up
yourself it will eat up your disk space.

Pankil

On Mon, Jun 22, 2009 at 4:24 PM, Qin Gao q...@cs.cmu.edu wrote:

 Thanks!

 But what if the jobs get killed or failed? Does hadoop try to clean it? we
 are considering bad situations - if job gets killed, will the tmp dirs sit
 on local disks forever and eats up all the diskspace?

 I guess this should be considered in distributed cache, but those files are
 read-only, and our program will generate new temporary files.


 --Q


 On Mon, Jun 22, 2009 at 4:19 PM, Pankil Doshi forpan...@gmail.com wrote:

  Yes, If your job gets completed successfully .possibly it removes after
  completion of both map and reduce tasks.
 
  Pankil
 
  On Mon, Jun 22, 2009 at 3:15 PM, Qin Gao q...@cs.cmu.edu wrote:
 
   Hi All,
  
   Do you know if the tmp directory on every map/reduce task will be
 deleted
   automatically after the map task finishes or will do I have to delete
  them?
  
   I mean the tmp directory that automatically created by on current
   directory.
  
   Thanks a lot
   --Q
  
 



Re: HDFS out of space

2009-06-22 Thread Pankil Doshi
Matt.

Kris can give that info..
I am one of the users from mailing list.

PAnkil

On Mon, Jun 22, 2009 at 4:37 PM, Matt Massie m...@cloudera.com wrote:

 Pankil-

 I'd be interested to know the size of the /mnt and /mnt2 partitions.  Are
 they the same?  Can you run the following and report the output...

 % df -h /mnt /mnt2

 Thanks.

 -Matt


 On Jun 22, 2009, at 1:32 PM, Pankil Doshi wrote:

  Hey Alex,

 Will Hadoop balancer utility work in this case?

 Pankil

 On Mon, Jun 22, 2009 at 4:30 PM, Alex Loddengaard a...@cloudera.com
 wrote:

  Are you seeing any exceptions because of the disk being at 99% capacity?

 Hadoop should do something sane here and write new data to the disk with
 more capacity.  That said, it is ideal to be balanced.  As far as I know,
 there is no way to balance an individual DataNode's hard drives (Hadoop
 does
 round-robin scheduling when writing data).

 Alex

 On Mon, Jun 22, 2009 at 10:12 AM, Kris Jirapinyo kjirapi...@biz360.com

 wrote:


  Hi all,
  How does one handle a mount running out of space for HDFS?  We have

 two

 disks mounted on /mnt and /mnt2 respectively on one of the machines that
 are
 used for HDFS, and /mnt is at 99% while /mnt2 is at 30%.  Is there a way

 to

 tell the machine to balance itself out?  I know for the cluster, you can
 balance it using start-balancer.sh but I don't think that it will tell

 the

 individual machine to balance itself out.  Our hack right now would be
 just to delete the data on /mnt, since we have replication of 3x, we

 should

 be OK.  But I'd prefer not to do that.  Any thoughts?