Can you please change below property and restart your cluster again? FROM:
<name>dfs.http.address</name> TO: <name>dfs.namenode.http-addres</name> Thanks Jitendra On Fri, Jan 31, 2014 at 7:07 PM, Stuti Awasthi <[email protected]> wrote: > Hi Jitendra, > > > > I realized that some days back ,my cluster was down due to power failure > after which nn/current directory has : edits, edits.new file and now SNN is > not rolling these edits due to HTTP error. > > Also currently my NN and SNN are operating on same machine > > > > > > DFSadmin report : > > > > Configured Capacity: 659494076416 (614.2 GB) > > Present Capacity: 535599210496 (498.82 GB) > > DFS Remaining: 497454006272 (463.29 GB) > > DFS Used: 38145204224 (35.53 GB) > > DFS Used%: 7.12% > > Under replicated blocks: 283 > > Blocks with corrupt replicas: 3 > > Missing blocks: 3 > > > > ------------------------------------------------- > > Datanodes available: 8 (8 total, 0 dead) > > > > Name: 10.139.9.238:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 4302274560 (4.01 GB) > > Non DFS Used: 8391843840 (7.82 GB) > > DFS Remaining: 69742641152(64.95 GB) > > DFS Used%: 5.22% > > DFS Remaining%: 84.6% > > Last contact: Fri Jan 31 18:55:18 IST 2014 > > > > > > Name: 10.139.9.233:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 5774745600 (5.38 GB) > > Non DFS Used: 13409488896 (12.49 GB) > > DFS Remaining: 63252525056(58.91 GB) > > DFS Used%: 7.01% > > DFS Remaining%: 76.73% > > Last contact: Fri Jan 31 18:55:19 IST 2014 > > > > > > Name: 10.139.9.232:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 8524451840 (7.94 GB) > > Non DFS Used: 24847884288 (23.14 GB) > > DFS Remaining: 49064423424(45.69 GB) > > DFS Used%: 10.34% > > DFS Remaining%: 59.52% > > Last contact: Fri Jan 31 18:55:21 IST 2014 > > > > > > Name: 10.139.9.236:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 4543819776 (4.23 GB) > > Non DFS Used: 8669548544 (8.07 GB) > > DFS Remaining: 69223391232(64.47 GB) > > DFS Used%: 5.51% > > DFS Remaining%: 83.97% > > Last contact: Fri Jan 31 18:55:19 IST 2014 > > > > > > Name: 10.139.9.235:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 5092986880 (4.74 GB) > > Non DFS Used: 8669454336 (8.07 GB) > > DFS Remaining: 68674318336(63.96 GB) > > DFS Used%: 6.18% > > DFS Remaining%: 83.31% > > Last contact: Fri Jan 31 18:55:19 IST 2014 > > > > > > Name: 10.139.9.237:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 4604301312 (4.29 GB) > > Non DFS Used: 11005788160 (10.25 GB) > > DFS Remaining: 66826670080(62.24 GB) > > DFS Used%: 5.59% > > DFS Remaining%: 81.06% > > Last contact: Fri Jan 31 18:55:18 IST 2014 > > > > > > Name: 10.139.9.234:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 4277760000 (3.98 GB) > > Non DFS Used: 12124221440 (11.29 GB) > > DFS Remaining: 66034778112(61.5 GB) > > DFS Used%: 5.19% > > DFS Remaining%: 80.1% > > Last contact: Fri Jan 31 18:55:18 IST 2014 > > > > > > Name: 10.139.9.231:50010 > > Decommission Status : Normal > > Configured Capacity: 82436759552 (76.78 GB) > > DFS Used: 1024864256 (977.39 MB) > > Non DFS Used: 36776636416 (34.25 GB) > > DFS Remaining: 44635258880(41.57 GB) > > DFS Used%: 1.24% > > DFS Remaining%: 54.14% > > Last contact: Fri Jan 31 18:55:20 IST 2014 > > > > > > > > *From:* Jitendra Yadav [mailto:[email protected]] > *Sent:* Friday, January 31, 2014 6:58 PM > *To:* user > *Subject:* Re: java.io.FileNotFoundException: > http://HOSTNAME:50070/getimage?getimage=1 > > > > Hi, > > > > Please post the output of dfs report command, this could help us to > understand cluster health. > > > > # *hadoop dfsadmin -report* > > > > Thanks > > Jitendra > > > > On Fri, Jan 31, 2014 at 6:44 PM, Stuti Awasthi <[email protected]> > wrote: > > Hi All, > > > > I am suddenly started facing issue on Hadoop Cluster. Seems like HTTP > request at port 50070 on dfs is not working properly. > > I have an Hadoop cluster which is operating from several days. Recently we > are also not able to see dfshealth.jsp page from webconsole. > > > > Problems : > > 1. http://<Hostname>:50070/dfshealth.jsp shows following error > > > > HTTP ERROR: 404 > > Problem accessing /. Reason: > > NOT_FOUND > > > > 2. SNN is not able to roll edits : > > ERROR in SecondaryNameNode Log > > java.io.FileNotFoundException: http://HOSTNAME:50070/getimage?getimage=1 > > at > sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1401) > > at > org.apache.hadoop.hdfs.server.namenode.TransferFsImage.getFileClient(TransferFsImage.java:160) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$3.run(SecondaryNameNode.java:347) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$3.run(SecondaryNameNode.java:336) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:416) > > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.downloadCheckpointFiles(SecondaryNameNode.java:336) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:411) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doWork(SecondaryNameNode.java:312) > > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNode.java:275) > > > > ERROR in Namenode Log > > 2014-01-31 18:15:12,046 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from > 10.139.9.231 > > 2014-01-31 18:15:12,046 WARN > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Cannot roll edit log, > edits.new files already exists in all healthy directories: > > /usr/lib/hadoop/storage/dfs/nn/current/edits.new > > > > > > > > Namenode logs which suggest that webserver is started on 50070 > successfully: > > 2014-01-31 14:42:35,208 INFO org.apache.hadoop.http.HttpServer: Port > returned by webServer.getConnectors()[0].getLocalPort() before open() is > -1. Opening the listener on 50070 > > 2014-01-31 14:42:35,209 INFO org.apache.hadoop.http.HttpServer: > listener.getLocalPort() returned 50070 > webServer.getConnectors()[0].getLocalPort() returned 50070 > > 2014-01-31 14:42:35,209 INFO org.apache.hadoop.http.HttpServer: Jetty > bound to port 50070 > > 2014-01-31 14:42:35,378 INFO > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: > HOSTNAME:50070 > > > > > > Hdfs-site.xml > > <configuration> > > <property> > > <name>dfs.replication</name> > > <value>2</value> > > </property> > > > > <property> > > <name>dfs.name.dir</name> > > <value>/usr/lib/hadoop/storage/dfs/nn</value> > > </property> > > > > <property> > > <name>dfs.data.dir</name> > > <value>/usr/lib/hadoop/storage/dfs/dn</value> > > </property> > > > > <property> > > <name>dfs.permissions</name> > > <value>false</value> > > </property> > > <property> > > <name>dfs.webhdfs.enabled</name> > > <value>true</value> > > </property> > > > > <property> > > <name>dfs.http.address</name> > > <value>HOSTNAME:50070</value> > > </property> > > > > <property> > > <name>dfs.secondary.http.address</name> > > <value>HOSTNAME:50090</value> > > </property> > > > > <property> > > <name>fs.checkpoint.dir</name> > > <value>/usr/lib/hadoop/storage/dfs/snn</value> > > </property> > > > > </configuration> > > > > > > */etc/hosts* (Note I have also tried by commenting 127.0.0.1 entry in > host file but the issue was not resolved) > > > > 127.0.0.1 localhost > > > > IP1 Hostname1 # Namenode- vm01 - itself > > IP2 Hostname2 # DataNode- vm02 > > ........ > > > > # The following lines are desirable for IPv6 capable hosts > > ::1 ip6-localhost ip6-loopback > > fe00::0 ip6-localnet > > ff00::0 ip6-mcastprefix > > ff02::1 ip6-allnodes > > ff02::2 ip6-allrouters > > > > > > Note : All Hadoop daemons are executing fine and the jobs are running > properly. > > > > How to resolve this issue, I have tried many options provided on different > forums but still facing the same issue. > > I belive that this can cause a major problem later as my edits are not > getting rolled into fsimage.. This can cause me a data loss in case of > failure. > > > > Please suggest > > > > Thanks > > Stuti > > > > > > > > > > ::DISCLAIMER:: > > ---------------------------------------------------------------------------------------------------------------------------------------------------- > > The contents of this e-mail and any attachment(s) are confidential and > intended for the named recipient(s) only. > E-mail transmission is not guaranteed to be secure or error-free as > information could be intercepted, corrupted, > lost, destroyed, arrive late or incomplete, or may contain viruses in > transmission. The e mail and its contents > (with or without referred errors) shall therefore not attach any liability > on the originator or HCL or its affiliates. > Views or opinions, if any, presented in this email are solely those of the > author and may not necessarily reflect the > views or opinions of HCL or its affiliates. Any form of reproduction, > dissemination, copying, disclosure, modification, > distribution and / or publication of this message without the prior > written consent of authorized representative of > HCL is strictly prohibited. If you have received this email in error > please delete it and notify the sender immediately. > Before opening any email and/or attachments, please check them for viruses > and other defects. > > > ---------------------------------------------------------------------------------------------------------------------------------------------------- > > >
