Agreed with 梁猛. Actually we found the same issue, the number of reducers is too small in step ‘convert to hfile’, which is same as the region count.
I think we could increase the number of reducers, to improve performance. If anyone has interesting in this, we could discuss more about the solution. > 在 2016年1月15日,09:46,13802880...@139.com 写道: > > actually,I found the last step " convert to hfile" take too much time, more > than 40 minutes for single region(use small, and result file about 5GB) > > > > 中国移动广东有限公司 网管中心 梁猛 > 13802880...@139.com > > From: ShaoFeng Shi > Date: 2016-01-15 09:40 > To: dev > Subject: Re: beg suggestions to speed up the Kylin cube build > The cube build performance is much determined by your Hadoop cluster's > capacity. You can do some inspection with the MR job's statistics to > analysis the potential bottlenecks. > > > > 2016-01-15 7:19 GMT+08:00 zhong zhang <zzaco...@gmail.com>: > >> Hi All, >> >> We are trying to build a nine-dimension cube: >> eight mandatory dimensions and one hierarchy >> dimension. The fact table is like 20G. Two lookup >> tables are 1.3M and 357k separately. It takes like >> 3 hours to go to 30% progress which is kind of slow. >> >> We'd like to know are there suggestions to speed up >> the Kylin cube build. We got a suggestion from >> a slide said that sort the dimension based on the >> cardinality. Are there any other ways we can try? >> >> We also noticed that only half of the memory and >> half of the CPU are used during the cube build. >> Are there any ways to fully utilize the resource? >> >> Looking forward to hear from you. >> >> Best regards, >> Zhong >> > > > > -- > Best regards, > > Shaofeng Shi