Did you configure "kylin.hbase.cluster.fs", pointing to your HBase HDFS?
Check this blog for more: https://kylin.apache.org/blog/2016/06/10/standalone-hbase-cluster/ 2017-10-16 9:51 GMT+08:00 yu feng <[email protected]>: > yes, hbase is running on another HDFS, and in a very big BulkLoad, the HDFS > is blocking (network or disk I/O), which block Hbase. > > 2017-10-15 9:38 GMT+08:00 ShaoFeng Shi <[email protected]>: > > > The generation of HFile is happened in the "Convert to HFile" step, which > > is an MR job, won't block HBase normal tasks. > > > > The HBase BulkLoad on HDFS should be very fast (second level), as it is > > just a move operation. > > > > For your case, is your HBase running with another HDFS other than the > > default HDFS? > > > > > > 2017-10-13 16:16 GMT+08:00 yu feng <[email protected]>: > > > > > A very big cube, such as cube size id bigger than 1TB will block > hbase's > > > normal operation when doing the BulkLoad job (because the job will > write > > to > > > much data to HDFS), such as kylin metadata operation/ query. especially > > > when the cube's merge job maybe write to hbase N TB in a mr job. > > > > > > Has anyone met the problem? > > > > > > > > > > > -- > > Best regards, > > > > Shaofeng Shi 史少锋 > > > -- Best regards, Shaofeng Shi 史少锋
