Hi Jia, That's all how hbase works ;)
When regions are bigger than the configured value, hbase will split them. Default is 10GB but you can configure that per table. So with 77GB you should have at least 8 regions. For performances, don't forget to pre-split before you load... JM Le 2013-08-13 22:16, <[email protected]> a écrit : > Hi jean-marc,**** > > ** ** > > The hdfs is running all the time ; I guess hbase occurs splitting during > putting large data , and the original hfile is splitted into new hfiles?** > ** > > Is that possible?**** > > ** ** > > ** ** > > ** ** > > Hi Jia,**** > > ** ** > > How is you HDFS running?**** > > ** ** > > "Caused by: org.apache.hadoop.ipc.**** > > RemoteException(java.io.IOException): File**** > > > /apps/hbase/data/lbc_zte_1_imei_index/4469e6b0500bf3f5ed0ac1247d249537/.tmp/e7bb489662344b26bc6de1e72c122eec > **** > > could only be replicated to 0 nodes instead of minReplication (=1). There > **** > > are 3 datanode(s) running and no node(s) are excluded in this operation.** > ** > > "**** > > ** ** > > Sound like there is some issues on the datanode. Have you checked its logs? > **** > > ** ** >
