I think making the data model right is the first thing. Expand the multi-value into multiple rows is the right approach. The concern that the result will be too big is then a secondary issue. There are plenty ways to handle a big table. E.g. it can be a view that only temporarily exists during cube build and is deleted right after build complete.
On Thu, Aug 18, 2016 at 7:50 PM, 张天生 <[email protected]> wrote: > You perhaps don't understand my question. My question is: original column > value is '1_3_12_15_27_35', but it can't directly be used to dimension > value, so it must be splited to 6 values [1, 3, 12, 15, 27, 35], and this > values will be used to construct the rowkey, and origianl record row will > be expanded to 6 times, it is too big. Is there a way to read ' > 1_3_12_15_27_35' and automate split it to 6 values in distinct column and > other step, use this values to create dimension dictionary and rowkey, and > don't need to preprocess orignal data. > > Li Yang <[email protected]>于2016年8月18日周四 下午6:47写道: > >> Depends on how you query/process the multi-value field, the answer will >> be different. >> >> Could you share some query sample? >> >> On Wed, Aug 17, 2016 at 2:35 PM, 张天生 <[email protected]> wrote: >> >>> Can someone help me to answer this question? I was still waiting for >>> answer. >>> >>> 张天生 <[email protected]>于2016年8月15日周一 上午11:28写道: >>> >>>> I have a dimension user_tags, it is a multi-value column, for example >>>> the value is "1_3_12_15_27_35_...", it was seperated by "_". As i known, >>>> kylin don't directly propress this multi-value column, it must preprocess >>>> it to a single value column, but it will increase record count to 50~100 >>>> times, the data is too big.So is there a way to deal with multi-value >>>> dimension, it don't need to split the value to many record, in calculate >>>> dimension cardinality, it can read original data and automate split the >>>> value to multi-value and process, and it will save disk i/o and cpu >>>> spending. >>>> >>> >>
