2014-07-24 17:33:04,783 WARN org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:hdfs (auth:SIMPLE) cause:org.apache.hadoop.ipc.StandbyException: Operation category READ is not supported in state standby 2014-07-24 17:33:05,742 WARN org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:hdfs (auth:SIMPLE) cause:org.apache.hadoop.ipc.StandbyException: Operation category READ is not supported in state standby 2014-07-24 17:33:33,179 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Triggering log roll on remote NameNode hz24/192.168.10.24:8020 2014-07-24 17:33:33,442 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@67698344 expecting start txid #62525 2014-07-24 17:33:33,442 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Start loading edits file http://hz24:8480/getJournal?jid=develop&segmentTxId=62525&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c, http://hz23:8480/getJournal?jid=develop&segmentTxId=62525&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c 2014-07-24 17:33:33,442 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz24:8480/getJournal?jid=develop&segmentTxId=62525&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c, http://hz23:8480/getJournal?jid=develop&segmentTxId=62525&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62525 2014-07-24 17:33:33,442 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz24:8480/getJournal?jid=develop&segmentTxId=62525&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62525 2014-07-24 17:33:33,480 INFO BlockStateChange: BLOCK* addToInvalidates: blk_1073753268_12641 192.168.10.51:50010 192.168.10.49:50010 192.168.10.50:50010 2014-07-24 17:33:33,482 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.50:50010 is added to blk_1073753337_12710{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,482 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.51:50010 is added to blk_1073753337_12710{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,482 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.49:50010 is added to blk_1073753337_12710{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,484 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.51:50010 is added to blk_1073753338_12711{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,484 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.49:50010 is added to blk_1073753338_12711{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,484 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.50:50010 is added to blk_1073753338_12711{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW]]} size 0 2014-07-24 17:33:33,485 INFO BlockStateChange: BLOCK* addToInvalidates: blk_1073753338_12711 192.168.10.50:50010 192.168.10.49:50010 192.168.10.51:50010 2014-07-24 17:33:33,485 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.49:50010 is added to blk_1073753339_12712{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW]]} size 0 .................................
2014-07-24 17:35:33,573 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Triggering log roll on remote NameNode hz24/192.168.10.24:8020 2014-07-24 17:35:33,826 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@3a7ff649 expecting start txid #62721 2014-07-24 17:35:33,826 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Start loading edits file http://hz23:8480/getJournal?jid=develop&segmentTxId=62721&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c 2014-07-24 17:35:33,826 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz23:8480/getJournal?jid=develop&segmentTxId=62721&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62721 2014-07-24 17:35:33,826 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz23:8480/getJournal?jid=develop&segmentTxId=62721&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62721 2014-07-24 17:35:33,868 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.49:50010 is added to blk_1073753367_12740{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW]]} size 0 2014-07-24 17:35:33,868 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.51:50010 is added to blk_1073753367_12740{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW]]} size 0 2014-07-24 17:35:33,868 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: 192.168.10.50:50010 is added to blk_1073753367_12740{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-a4cfa75c-28f4-4e73-9e17-b6e3f129864f:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-23f57228-24d8-4e51-afe9-c13a8b47a0a5:NORMAL|RBW], ReplicaUnderConstruction[[DISK]DS-7496d6a7-2a8f-4884-8a8f-f3a0f3037c0e:NORMAL|RBW]]} size 0 2014-07-24 17:35:33,869 INFO BlockStateChange: BLOCK* addToInvalidates: blk_1073753270_12643 192.168.10.49:50010 192.168.10.51:50010 192.168.10.50:50010 2014-07-24 17:35:33,871 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Edits file http://hz23:8480/getJournal?jid=develop&segmentTxId=62721&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c of size 1385 edits # 16 loaded in 0 seconds 2014-07-24 17:35:33,872 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Loaded 16 edits starting from txid 62720 2014-07-24 17:35:34,042 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.49:50010 to delete [blk_1073753270_12643] 2014-07-24 17:35:37,043 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.50:50010 to delete [blk_1073753270_12643] 2014-07-24 17:35:40,043 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.51:50010 to delete [blk_1073753270_12643] 2014-07-24 17:37:33,915 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Triggering log roll on remote NameNode hz24/192.168.10.24:8020 2014-07-24 17:37:34,194 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5ed5ecda expecting start txid #62737 2014-07-24 17:37:34,195 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Start loading edits file http://hz24:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c, http://hz23:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c 2014-07-24 17:37:34,195 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz24:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c, http://hz23:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62737 2014-07-24 17:37:34,195 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream ' http://hz24:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0%3ACID-a140fb1a-ac10-4053-8b91-8f19f2809b7c' to transaction ID 62737 2014-07-24 17:37:34,223 INFO BlockStateChange: BLOCK* addToInvalidates: blk_1073753271_12644 192.168.10.51:50010 192.168.10.49:50010 192.168.10.50:50010 2014-07-24 17:37:34,224 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Edits file http://hz24:8480/getJournal?jid=develop&segmentTxId=62737&storageInfo=-55%3A466484546%3A0 : 2014-07-24 17:37:34,225 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Loaded 3 edits starting from txid 62736 2014-07-24 17:37:37,050 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.51:50010 to delete [blk_1073753271_12644] 2014-07-24 17:37:40,050 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.49:50010 to delete [blk_1073753271_12644] 2014-07-24 17:37:43,051 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* InvalidateBlocks: ask 192.168.10.50:50010 to delete [blk_1073753271_12644] 2014-07-24 17:39:34,255 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Triggering log roll on remote NameNode hz24/192.168.10.24:8020 On Fri, Jul 25, 2014 at 10:25 AM, Stanley Shi <[email protected]> wrote: > Would you please also past the corresponding namenode log? > > Regards, > *Stanley Shi,* > > > > On Fri, Jul 25, 2014 at 9:15 AM, ch huang <[email protected]> wrote: > >> hi,maillist: >> i try to copy data from my old cluster to new cluster,i get >> error ,how to handle this? >> >> 14/07/24 18:35:58 INFO mapreduce.Job: Task Id : >> attempt_1406182801379_0004_m_000000_1, Status : FAILED >> Error: java.io.IOException: File copy failed: >> webhdfs://CH22:50070/mytest/pipe_url_bak/part-m-00001 --> >> webhdfs://develop/tmp/pipe_url_bak/part-m-00001 >> at >> org.apache.hadoop.tools.mapred.CopyMapper.copyFileWithRetry(CopyMapper.java:262) >> at >> org.apache.hadoop.tools.mapred.CopyMapper.map(CopyMapper.java:229) >> at >> org.apache.hadoop.tools.mapred.CopyMapper.map(CopyMapper.java:45) >> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) >> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548) >> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) >> Caused by: java.io.IOException: Couldn't run retriable-command: Copying >> webhdfs://CH22:50070/mytest/pipe_url_bak/part-m-00001 to >> webhdfs://develop/tmp/pipe_url_bak/part-m-00001 >> at >> org.apache.hadoop.tools.util.RetriableCommand.execute(RetriableCommand.java:101) >> at >> org.apache.hadoop.tools.mapred.CopyMapper.copyFileWithRetry(CopyMapper.java:258) >> ... 10 more >> Caused by: java.io.IOException: Error writing request body to server >> at >> sun.net.www.protocol.http.HttpURLConnection$StreamingOutputStream.checkError(HttpURLConnection.java:3192) >> at >> sun.net.www.protocol.http.HttpURLConnection$StreamingOutputStream.write(HttpURLConnection.java:3175) >> at >> java.io.BufferedOutputStream.write(BufferedOutputStream.java:122) >> at >> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:58) >> at java.io.DataOutputStream.write(DataOutputStream.java:107) >> at >> java.io.BufferedOutputStream.write(BufferedOutputStream.java:122) >> at >> org.apache.hadoop.tools.mapred.RetriableFileCopyCommand.copyBytes(RetriableFileCopyCommand.java:231) >> at >> org.apache.hadoop.tools.mapred.RetriableFileCopyCommand.copyToTmpFile(RetriableFileCopyCommand.java:164) >> at >> org.apache.hadoop.tools.mapred.RetriableFileCopyCommand.doCopy(RetriableFileCopyCommand.java:118) >> at >> org.apache.hadoop.tools.mapred.RetriableFileCopyCommand.doExecute(RetriableFileCopyCommand.java:95) >> at >> org.apache.hadoop.tools.util.RetriableCommand.execute(RetriableCommand.java:87) >> ... 11 more >> 14/07/24 18:35:59 INFO mapreduce.Job: map 16% reduce 0% >> 14/07/24 18:39:39 INFO mapreduce.Job: map 17% reduce 0% >> 14/07/24 19:04:27 INFO mapreduce.Job: Task Id : >> attempt_1406182801379_0004_m_000000_2, Status : FAILED >> Error: java.io.IOException: File copy failed: >> webhdfs://CH22:50070/mytest/pipe_url_bak/part-m-00001 --> >> webhdfs://develop/tmp/pipe_url_bak/part-m-00001 >> at >> org.apache.hadoop.tools.mapred.CopyMapper.copyFileWithRetry(CopyMapper.java:262) >> at >> org.apache.hadoop.tools.mapred.CopyMapper.map(CopyMapper.java:229) >> at >> org.apache.hadoop.tools.mapred.CopyMapper.map(CopyMapper.java:45) >> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) >> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548) >> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) >> Caused by: java.io.IOException: Couldn't run retriable-command: Copying >> webhdfs://CH22:50070/mytest/pipe_url_bak/part-m-00001 to >> webhdfs://develop/tmp/pipe_url_bak/part-m-00001 >> at >> org.apache.hadoop.tools.util.RetriableCommand.execute(RetriableCommand.java:101) >> > >
