Hi All,
We will monitor memory usages meanwhile below are the error:-
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File
/hbase/WALs/quickstart.cloudera,60020,1469713295051/quickstart.cloudera%2C60020%2C1469713295051..meta.1469766981975.meta
could only be replicated to 0 nodes instead of minReplication (=1). There
are 0 datanode(s) running and no node(s) are excluded in this operation.
at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1595)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3287)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:677)
at
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:213)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:485)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)
at org.apache.hadoop.ipc.Client.call(Client.java:1471)
at org.apache.hadoop.ipc.Client.call(Client.java:1408)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230)
at com.sun.proxy.$Proxy22.addBlock(Unknown Source)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:404)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)
at com.sun.proxy.$Proxy23.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)
at com.sun.proxy.$Proxy24.addBlock(Unknown Source)
at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1704)
at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1500)
at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:668)
2016-07-28 21:37:49,143 WARN
org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter: Failed to write
trailer, non-fatal, continuing...
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File
/hbase/WALs/quickstart.cloudera,60020,1469713295051/quickstart.cloudera%2C60020%2C1469713295051.null0.1469766981975
could only be replicated to 0 nodes instead of minReplication (=1). There
are 0 datanode(s) running and no node(s) are excluded in this operation.
at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1595)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3287)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:677)
at
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:213)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:485)
Best Regards,
ANKIT BEOHAR
On Fri, Jul 29, 2016 at 9:10 AM, Du, Jingcheng <[email protected]>
wrote:
> Hi Ankit,
>
> It seems like a memory issue in region servers. Did you monitor the memory
> usage in region servers during the run? How about to increase the heap size?
> Do you get exceptions after the region server goes down? Mind sharing it
> here?
>
> Regards,
> Jingcheng
>
> -----Original Message-----
> From: Dima Spivak [mailto:[email protected]]
> Sent: Friday, July 29, 2016 11:32 AM
> To: [email protected]
> Subject: Re: Hbase USERT
>
> Hey Ankit,
>
> Moving the dev list to bcc and adding the user mailing list as the
> recipient. Maybe a fellow user can offer some suggestions.
>
> All the best,
> Dima
>
> On Thursday, July 28, 2016, ankit beohar <[email protected]> wrote:
>
> > Hi Hbase,
> >
> > My use case is :- I am getting files and I want to insert the records
> > in hbase with rowkey if rowkey available I have to update the values
> > with
> > old+new values.
> >
> > For this I wrote MR job and get the values of each rowkey and in If
> > else I manage my update and insert but with only 0.1 millions records
> > hbase region server goes down.
> >
> > Any idea on this?
> >
> > I tried to incorporate Phoenix upsert also but with this same error
> occurs.
> >
> > Please help me out this.
> >
> > Best Regards,
> > ANKIT BEOHAR
> >
>