Hi,

I am running a M/R with an implementation similar to AvroMultipleOutputs for 
dynamic output location. I am getting the following error in the reduce phase. 
I have checked that I have enough disk space with only 40% DFS storage used. 
The job run fines with smaller dataset, its only with large input data, I get 
the exception. I am using Oozie to run the Pig job in IBM bigInsight 
environment on Amazon AWS. Any hints what might go wrong here?

org.apache.pig.backend.executionengine.ExecException: ERROR 2135: Received 
error from store function.File 
/data/1/output/mt/2014/01/15/13/1406937600/_attempt_201408270926_0003_r_000000_0/part-r-00000
 could only be replicated to 0 nodes instead of minReplication (=1). There are 
2 datanode(s) running and no node(s) are excluded in this operation. at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2477)
 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928) at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048) at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044) at 
java.security.AccessController.doPrivileged(AccessController.java:310) at 
javax.security.auth.Subject.doAs(Subject.java:573) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042) at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POStore.getNextTuple(POStore.java:165)
 at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSplit.runPipeline(POSplit.java:255)
 at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSplit.processPlan(POSplit.java:237)
 at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSplit.getNextTuple(POSplit.java:229)
 at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.PODemux.runPipeline(PODemux.java:220)
 at 
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.PODemux.getNextTuple(PODemux.java:210)
 at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.runPipeline(PigGenericMapReduce.java:464)
 at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.processOnePackageOutput(PigGenericMapReduce.java:432)
 at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:412)
 at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:256)
 at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:166) at 
org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:652) at 
org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:420) at 
org.apache.hadoop.mapred.Child$4.run(Child.java:255) at 
java.security.AccessController.doPrivileged(AccessController.java:310) at 
javax.security.auth.Subject.doAs(Subject.java:573) at

Regards,

Pankaj Agrawal

Reply via email to