Did you figure out the exception of "No enum constant
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES” ? Is it still be
thrown in the logs? In the last step, Kylin need to parse the MR counters
to update cube size; Please refer to
https://issues.apache.org/jira/browse/MAPREDUCE-5831 for that error.

On 2/27/15, 5:04 PM, "Santoshakhilesh" <[email protected]> wrote:

>Hi Shaofeng ,
>          Cube building is failed at last step while loading Hfile to
>Hbase with exception "Can't get cube segment size.
>".  What could be reason ?
>
>parameter : -input
>/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/
>-htablename KYLIN_K27LDMX63W -cubename RetailCube
>
>Log:
>
>Start to execute command:
> -input /tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/
>-htablename KYLIN_K27LDMX63W -cubename RetailCube
>Command execute return code 0
>Failed with Exception:java.lang.RuntimeException: Can't get cube segment
>size.
> at 
>com.kylinolap.job.flow.JobFlowListener.updateCubeSegmentInfoOnSucceed(JobF
>lowListener.java:247)
> at 
>com.kylinolap.job.flow.JobFlowListener.jobWasExecuted(JobFlowListener.java
>:101)
> at 
>org.quartz.core.QuartzScheduler.notifyJobListenersWasExecuted(QuartzSchedu
>ler.java:1985)
> at 
>org.quartz.core.JobRunShell.notifyJobListenersComplete(JobRunShell.java:34
>0)
> at org.quartz.core.JobRunShell.run(JobRunShell.java:224)
> at 
>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5
>73)
>
>I have checked in hbase shell and following are the tables in hbase;
>hbase(main):001:0> list
>TABLE
>
>KYLIN_K27LDMX63W
>kylin_metadata_qa
>kylin_metadata_qa_acl
>kylin_metadata_qa_cube
>kylin_metadata_qa_dict
>kylin_metadata_qa_invertedindex
>kylin_metadata_qa_job
>kylin_metadata_qa_job_output
>kylin_metadata_qa_proj
>kylin_metadata_qa_table_snapshot
>kylin_metadata_qa_user
>11 row(s) in 0.8990 seconds
>
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Santoshakhilesh
>Sent: Friday, February 27, 2015 2:15 PM
>To: [email protected]
>Subject: RE: Error while making cube & Measure option is not responding
>on GUI
>
>I have manually copied the jar to /tmp/kylin , now satge 2 is done ,
>thanks.
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Shi, Shaofeng [[email protected]]
>Sent: Friday, February 27, 2015 1:00 PM
>To: [email protected]
>Cc: Kulbhushan Rana
>Subject: Re: Error while making cube & Measure option is not responding
>on GUI
>
>In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we
>renamed the package to “org.apache.kylin.xxx”; When you downgrade to 0.6,
>did you also replace the jar location with 0.6 ones in kylin.properties?
>
>On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]>
>wrote:
>
>>Hi Shaofeng ,
>>         I have added my fact and dimension tables under default database
>>of hive.
>>         Now stage 1 of Cube Build is ok. And there is failure at step2.
>>        The map reduce job for the finding distinct columns of fact table
>>is error. Yarn log is as below.
>>        Strangely this is class not found error. I have checked the
>>Kylin.properties and the jar is already set as below.
>>kylin. log has one exception connecting to linux/10.19.93.68 to
>>0.0.0.0:10020
>> Please help me to give a clue , I am also trying to check meanwhile
>>
>>Thanks.
>>kylin property
>># Temp folder in hdfs
>>kylin.hdfs.working.dir=/tmp
>># Path to the local(relative to job engine) job jar, job engine will use
>>this jar
>>kylin.job.jar=/tmp/kylin/kylin-job-latest.jar
>>
>>Map Reduce error
>>----------------------------
>>2015-02-27 20:24:25,262 FATAL [main] org.apache.hadoop.mapred.YarnChild:
>>Error running child : java.lang.NoClassDefFoundError:
>>com/kylinolap/common/mr/KylinMapper
>> at java.lang.ClassLoader.defineClass1(Native Method)
>> at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>> at
>>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>> at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>> at java.lang.Class.forName0(Native Method)
>> at java.lang.Class.forName(Class.java:274)
>> at
>>org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.j
>>a
>>va:2013)
>>
>>Kylin.log
>>QuartzScheduler_Worker-20]:[2015-02-27
>>20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFetch
>>e
>>r.java:60)] - 0 pending jobs
>>[QuartzScheduler_Worker-19]:[2015-02-27
>>20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobC
>>o
>>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException:
>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>>failed on connection exception: java.net.ConnectException: Connection
>>refused; For more details see:
>>http://wiki.apache.org/hadoop/ConnectionRefused
>>com.kylinolap.job.exception.JobException: java.io.IOException:
>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>>failed on connection exception: java.net.ConnectException: Connection
>>refused; For more details see:
>>http://wiki.apache.org/hadoop/ConnectionRefused
>> at
>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.
>>j
>>ava:289)
>> at
>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdO
>>u
>>tput.java:162)
>> at
>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.j
>>a
>>va:85)
>> at
>>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86)
>> at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
>> at
>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:
>>5
>>73)
>>Caused by: java.io.IOException: java.net.ConnectException: Call From
>>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception:
>>java.net.ConnectException: Connection refused; For more details see:
>>http://wiki.apache.org/hadoop/ConnectionRefused
>> at
>>org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelega
>>t
>>e.java:331)
>> at
>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServi
>>c
>>eDelegate.java:368)
>> at
>>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:415)
>> at
>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation
>>.
>>java:1491)
>> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>> at
>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.
>>j
>>ava:287)
>> ... 5 more
>>
>>Regards,
>>Santosh Akhilesh
>>Bangalore R&D
>>HUAWEI TECHNOLOGIES CO.,LTD.
>>
>>www.huawei.com
>>-------------------------------------------------------------------------
>>-
>>-----------------------------------------------------------
>>This e-mail and its attachments contain confidential information from
>>HUAWEI, which
>>is intended only for the person or entity whose address is listed above.
>>Any use of the
>>information contained herein in any way (including, but not limited to,
>>total or partial
>>disclosure, reproduction, or dissemination) by persons other than the
>>intended
>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>notify the sender by
>>phone or email immediately and delete it!
>>
>>________________________________________
>>From: Shi, Shaofeng [[email protected]]
>>Sent: Friday, February 27, 2015 8:01 AM
>>To: [email protected]
>>Subject: Re: Error while making cube & Measure option is not responding
>>on GUI
>>
>>In 0.6.x it only support tables in default database, this is a
>>limitation;
>>The support for non-default tables will be released in 0.7;
>>
>>To bypass this issue for now, please copy the table to default database
>>as
>>a workaround;
>>
>>On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]>
>>wrote:
>>
>>>@Jason
>>>thanks , but now as suggested by Saofeng I m not using the inverted
>>>index
>>>brach as its not stable.
>>>I have switched back to 0.6 branch , in this branch yesterday night I
>>>could
>>>crete the cube successfully but there is issue while building it. I feel
>>>that at step 1 of cube build  while creating flat table when command is
>>>issued to hive if the tables are not under default datbase flat table
>>>creation is failed and cube build fails. my fact and dimension tables
>>>are
>>>under a database called retail.
>>>
>>>@Saofeng - Can you please confirm this behavior ? Do I need to create
>>>the
>>>hive tables under default database?
>>>
>>>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]>
>>>wrote:
>>>
>>>> @Santoshakhilesh
>>>>
>>>> 1. When I go to measure section and click on measure option , there is
>>>>no
>>>> response , I want add measure on qty and price with sum
>>>>          --bug fixed on inverted-index branch
>>>>
>>>>
>>>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh <
>>>> [email protected]
>>>> > wrote:
>>>>
>>>> > Hi Shaofeng ,
>>>> >      I have build the 0.6 version and now able to create the cube
>>>> > successfully.
>>>> >      While building the cube , it fails at step1 with following
>>>>error.
>>>> > Table not found 'DIM_ITEM'
>>>> >      the table exists , but its under retail data base and not under
>>>> > default database.
>>>> >      does kylin require hive taables to be under default database ?
>>>>I
>>>>see
>>>> > the flat table being created under default database.
>>>> >
>>>> > Logging initialized using configuration in
>>>> >
>>>> >
>>>>
>>>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-commo
>>>>n
>>>>-
>>>>1.0.0.jar!/hive-log4j.properties
>>>> > SLF4J: Class path contains multiple SLF4J bindings.
>>>> > SLF4J: Found binding in
>>>> >
>>>> >
>>>>
>>>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/commo
>>>>n
>>>>/
>>>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>> > SLF4J: Found binding in
>>>> >
>>>> >
>>>>
>>>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jdbc
>>>>-
>>>>1
>>>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>> > explanation.
>>>> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>> > OK
>>>> > Time taken: 0.964 seconds
>>>> > OK
>>>> > Time taken: 0.948 seconds
>>>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found
>>>> > 'DIM_ITEM'
>>>> >
>>>> >
>>>> >
>>>> > Command is as below.
>>>> >
>>>> > hive -e "DROP TABLE IF EXISTS
>>>> >
>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44
>>>>;
>>>> > CREATE EXTERNAL TABLE IF NOT EXISTS
>>>> >
>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44
>>>> > (
>>>> > STOREID int
>>>> > ,ITEMID int
>>>> > ,CUSTID int
>>>> > ,QTY int
>>>> > ,AMOUNT double
>>>> > )
>>>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177'
>>>> > STORED AS SEQUENCEFILE
>>>> > LOCATION
>>>> >
>>>>
>>>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_tes
>>>>t
>>>>_
>>>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44';
>>>> > SET hive.exec.compress.output=true;
>>>> > SET hive.auto.convert.join.noconditionaltask = true;
>>>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000;
>>>> > INSERT OVERWRITE TABLE
>>>> >
>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44
>>>> > SELECT
>>>> > FACT_SALES.STOREID
>>>> > ,FACT_SALES.ITEMID
>>>> > ,FACT_SALES.CUSTID
>>>> > ,FACT_SALES.QTY
>>>> > ,FACT_SALES.AMOUNT
>>>> > FROM FACT_SALES
>>>> > INNER JOIN DIM_STORE
>>>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID
>>>> > INNER JOIN DIM_ITEM
>>>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID
>>>> > INNER JOIN DIM_CUSTOMER
>>>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID
>>>> > ;
>>>> > "
>>>> >
>>>> >
>>>> >
>>>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]>
>>>>wrote:
>>>> >
>>>> > > The 0.7.1 is test version, its package contains the “snapshot”
>>>>suffix;
>>>> we
>>>> > > will upload a new package there; Luke will also add a message
>>>>there
>>>>to
>>>> > > avoid this confusion;
>>>> > >
>>>> > > Regarding the problem that you encountered, could you please open
>>>>a
>>>> JIRA
>>>> > > ticket for tracking? Here is link of Apache JIRA:
>>>> > >
>>>> > > https://issues.apache.org/jira/secure/Dashboard.jspa
>>>> > >
>>>> > >
>>>> > > Thanks for the feedback!
>>>> > >
>>>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh"
>>>><[email protected]>
>>>> > > wrote:
>>>> > >
>>>> > > >Actually I see this being published on kylin webpage.
>>>> > > >http://kylin.incubator.apache.org/download/
>>>> > > >I am using 0.7.1 inverted index branch binary distribution.
>>>> > > >If this is not stable please give me the link of stable branch I
>>>>would
>>>> > try
>>>> > > >building and testing tonight.
>>>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]>
>>>> wrote:
>>>> > > >
>>>> > > >> Hi Santosh, it is not recommended to use the dev code branch
>>>> > (actually I
>>>> > > >> don’t know how you get the v0.7.x build and what’s the detail
>>>> version
>>>> > of
>>>> > > >> that; each day we submit many changes to that);
>>>> > > >>
>>>> > > >> The options are 1) switch back to latest release v0.6.5; or 2)
>>>>wait
>>>> > for
>>>> > > >> the formal release of 0.7, that should be in March; Otherwise,
>>>>we
>>>> > > >>couldn’t
>>>> > > >> ensure there is no new problems come out in your next steps;
>>>> > > >>
>>>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh"
>>>><[email protected]>
>>>> > > >>wrote:
>>>> > > >>
>>>> > > >> >Hi Shaofeng
>>>> > > >> >So what do you suggest , how should I proceed further with
>>>>this
>>>> > > >>release?
>>>> > > >> >Will there be a patch? Any alternate way I can create cube?
>>>> > > >> >Please suggest.
>>>> > > >> >Regards
>>>> > > >> >Santosh
>>>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng
>>>><[email protected]>
>>>> > > wrote:
>>>> > > >> >
>>>> > > >> >> Hi Santosh,
>>>> > > >> >>
>>>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we
>>>>have
>>>> > > >> >>metadata
>>>> > > >> >> structure change; While, the web UI (cube wizard) for this
>>>>change
>>>> > > >>hasn’t
>>>> > > >> >> been stabilized; So it is not strange that you got trouble
>>>>when
>>>> > > >>saving
>>>> > > >> >>the
>>>> > > >> >> cube;
>>>> > > >> >>
>>>> > > >> >> @Jason, any idea about the JS error?
>>>> > > >> >>
>>>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <
>>>> [email protected]
>>>> > >
>>>> > > >> >>wrote:
>>>> > > >> >>
>>>> > > >> >> >Hi Shaofeng,
>>>> > > >> >> >
>>>> > > >> >> >I am using the binary distribution 0.7.1. I have not been
>>>>able
>>>> to
>>>> > > >>save
>>>> > > >> >> >cube
>>>> > > >> >> >even once. I have tried creating new project and from local
>>>> > machine
>>>> > > >>and
>>>> > > >> >> >server machine. But I am always stuck with this error. I am
>>>> never
>>>> > > >> >>allowed
>>>> > > >> >> >to add measures and never been able to save the cube. I
>>>>also
>>>>see
>>>> > the
>>>> > > >> >> >kylin.log and it always tries to save cube with append
>>>>mode.
>>>>One
>>>> > > >>thing
>>>> > > >> >>I
>>>> > > >> >> >need to tell that at partition stage since I don't have a
>>>>big
>>>> fact
>>>> > > >> >>table
>>>> > > >> >> >now I have not partititioned the fact table and I skip this
>>>> step.
>>>> > > >>Does
>>>> > > >> >> >this
>>>> > > >> >> >have affect in saving the cube. Is this because some
>>>>metadata is
>>>> > > >> >>available
>>>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop
>>>>2.6.6.
>>>> > Yes
>>>> > > >> >>kylin
>>>> > > >> >> >propert I have not added the jar. I will add them and
>>>>check.
>>>>But
>>>> > > >>cube
>>>> > > >> >> >creation failure is really puzzling me. I could see no
>>>>error
>>>> logs
>>>> > in
>>>> > > >> >> >kylin.log.
>>>> > > >> >> >Regards
>>>> > > >> >> >Santosh
>>>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng
>>>><[email protected]
>>>> >
>>>> > > >> wrote:
>>>> > > >> >> >
>>>> > > >> >> >> Which version or code branch are you using? I assume
>>>>you’re
>>>> > using
>>>> > > >>the
>>>> > > >> >> >> stable version from master; Seems you’re trying to edit
>>>>an
>>>> > > >>existing
>>>> > > >> >>cube
>>>> > > >> >> >> to add new measurement, try refresh your browser's cache;
>>>>if
>>>> it
>>>> > > >>still
>>>> > > >> >> >> couldn’t be saved, try to create a new cube;
>>>> > > >> >> >>
>>>> > > >> >> >> The two error traces in tomcat need be taken care:
>>>> > > >> >> >>
>>>> > > >> >> >> 1) java.lang.NoClassDefFoundError:
>>>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper
>>>> > > >> >> >>         Please check kylin.properties file, making sure
>>>>the
>>>> > > >> >> >>“kylin.job.jar”
>>>> > > >> >> >> points to a right jar file; It will be loaded in
>>>>Map-reduce;
>>>> > > >> >> >>
>>>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant
>>>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>>> > > >> >> >>         This indicates your hadoop version might be old;
>>>> Please
>>>> > > >>check
>>>> > > >> >> >>and
>>>> > > >> >> >> ensure
>>>> > > >> >> >> hadoop version is 2.2 or above.
>>>> > > >> >> >>
>>>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh"
>>>> > > >><[email protected]>
>>>> > > >> >> >> wrote:
>>>> > > >> >> >>
>>>> > > >> >> >> >Hi Shaofeng ,
>>>> > > >> >> >> >
>>>> > > >> >> >> >   I am using chrome , When I click on button to add
>>>> measures ,
>>>> > > >> >> >>following
>>>> > > >> >> >> >is error on chrome console. When I try to save the cube
>>>>there
>>>> > is
>>>> > > >>no
>>>> > > >> >> >>error
>>>> > > >> >> >> >in console. I just get a error dialog saying failed to
>>>>take
>>>> > > >>action
>>>> > > >> >>and
>>>> > > >> >> >> >gives me the JSON cube schema.
>>>> > > >> >> >> >
>>>> > > >> >> >> >Error on chrome debug console is as below;
>>>> > > >> >> >> >
>>>> > > >> >> >> > ReferenceError: CubeDescModel is not defined
>>>> > > >> >> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>>>> > > >> >> >> >    at scripts.min.0.js:180
>>>> > > >> >> >> >    at scripts.min.0.js:197
>>>> > > >> >> >> >    at h.$eval (scripts.min.0.js:119)
>>>> > > >> >> >> >    at h.$apply (scripts.min.0.js:119)
>>>> > > >> >> >> >    at HTMLButtonElement.<anonymous>
>>>>(scripts.min.0.js:197)
>>>> > > >> >> >> >    at HTMLButtonElement.m.event.dispatch
>>>> (scripts.min.0.js:3)
>>>> > > >> >> >> >    at HTMLButtonElement.r.handle
>>>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>>>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous
>>>> > > >> >>function)scripts.min.0.js:119
>>>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous
>>>> > > >>function)scripts.min.0.js:3
>>>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle
>>>> > > >> >> >> >
>>>> > > >> >> >> >   About the hive table import , I got pass the run
>>>>shell
>>>> > command
>>>> > > >> >> >> >exception but it still fails the hadoop log is;
>>>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main]
>>>>org.apache.hadoop.mapred.
>>>> > > >> >> YarnChild:
>>>> > > >> >> >> >mapreduce.cluster.local.dir for child:
>>>> > > >> >> >>
>>>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
>>>> > > >> >> >> cation_14249530
>>>> > > >> >> >> >91340_0002
>>>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main]
>>>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation:
>>>>session.id
>>>> > is
>>>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id
>>>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main]
>>>> > > >>org.apache.hadoop.mapred.Task:
>>>> > > >> >> >>Using
>>>> > > >> >> >> >ResourceCalculatorProcessTree : [ ]
>>>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main]
>>>> > > >> >> >>org.apache.hadoop.mapred.YarnChild:
>>>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError:
>>>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper
>>>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method)
>>>> > > >> >> >> > at
>>>>java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>>>> > > >> >> >> > at
>>>> > > >> >> >> >java.security.SecureClassLoader.defineClass(
>>>> > > >> >> SecureClassLoader.java:142)
>>>> > > >> >> >> > at
>>>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>>>> > > >> >> >> > at
>>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>>> > > >> >> >> > at
>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>>>> > > >> >> >> > at
>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
>>>> Method)
>>>> > > >> >> >> > at
>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>> > > >> >> >> > at
>>>>java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>> > > >> >> >> > at
>>>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>> > > >> >> >> >
>>>> > > >> >> >> >tomcat logs:
>>>> > > >> >> >> >usage: HiveColumnCardinalityJob
>>>> > > >> >> >> > -output <path>        Output path
>>>> > > >> >> >> > -table <table name>   The hive table name
>>>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26
>>>> > > >> >> >>
>>>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
>>>> > > >> >> >> lExecutable.doW
>>>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute
>>>> > > >> >> >>
>>>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>>>> > > >> >> >> >name=null, state=RUNNING}
>>>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant
>>>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>> > >
>>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
>>>> > > >> >> >> Framewo
>>>> > > >> >> >> >rkCounterGroup.java:148)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
>>>> > > >> >> >> findCounter(Fra
>>>> > > >> >> >> >meworkCounterGroup.java:182)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
>>>> > > >> >> >> ounter(Abstract
>>>> > > >> >> >> >Counters.java:154)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
>>>> > > >> >> >> erter.java:240)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
>>>> > > >> >> >> rs(ClientServic
>>>> > > >> >> >> >eDelegate.java:370)
>>>> > > >> >> >> > at
>>>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(
>>>> > > >> >> YARNRunner.java:511)
>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
>>>> Method)
>>>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
>>>> > > >> >> >> oupInformation.
>>>> > > >> >> >> >java:1491)
>>>> > > >> >> >> > at
>>>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>>>> > > >> >> >> > at
>>>> > > >>
>>>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>>>> > > >> >> >> > at
>>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.
>>>> > > >> java:1289)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
>>>> > > >> >> >> tion(AbstractHa
>>>> > > >> >> >> >doopJob.java:134)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>> > > >>
>>>> >
>>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run
>>>>>>>(
>>>> > > >> >> >> HiveC
>>>> > > >> >> >> >olumnCardinalityJob.java:114)
>>>> > > >> >> >> > at
>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>>>> > > >> >> >> > at
>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
>>>> > > >> >> >> doopShellExecut
>>>> > > >> >> >> >able.java:62)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>>> > > >> >> >> bstractExecutab
>>>> > > >> >> >> >le.java:99)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW
>>>> > > >> >> >> ork(DefaultChai
>>>> > > >> >> >> >nedExecutable.java:50)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>>> > > >> >> >> bstractExecutab
>>>> > > >> >> >> >le.java:99)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
>>>> > > >> >> >> nner.run(Defaul
>>>> > > >> >> >> >tScheduler.java:132)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
>>>> > > >> >> >> lExecutor.java:
>>>> > > >> >> >> >1145)
>>>> > > >> >> >> > at
>>>> > > >> >> >>
>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
>>>> > > >> >> >> olExecutor.java
>>>> > > >> >> >> >:615)
>>>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745)
>>>> > > >> >> >> >
>>>> > > >> >> >> >Regards,
>>>> > > >> >> >> >Santosh Akhilesh
>>>> > > >> >> >> >Bangalore R&D
>>>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD.
>>>> > > >> >> >> >
>>>> > > >> >> >> >www.huawei.com
>>>> > > >> >> >>
>>>>>-----------------------------------------------------------
>>>> > > >> >> >> ---------------
>>>> > > >> >> >>
>>>>>-----------------------------------------------------------
>>>> > > >> >> >> >This e-mail and its attachments contain confidential
>>>> > information
>>>> > > >> >>from
>>>> > > >> >> >> >HUAWEI, which
>>>> > > >> >> >> >is intended only for the person or entity whose address
>>>>is
>>>> > listed
>>>> > > >> >> >>above.
>>>> > > >> >> >> >Any use of the
>>>> > > >> >> >> >information contained herein in any way (including, but
>>>>not
>>>> > > >>limited
>>>> > > >> >>to,
>>>> > > >> >> >> >total or partial
>>>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons
>>>>other
>>>> > than
>>>> > > >> >>the
>>>> > > >> >> >> >intended
>>>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail
>>>>in
>>>> > error,
>>>> > > >> >>please
>>>> > > >> >> >> >notify the sender by
>>>> > > >> >> >> >phone or email immediately and delete it!
>>>> > > >> >> >> >
>>>> > > >> >> >> >________________________________________
>>>> > > >> >> >> >From: Shi, Shaofeng [[email protected]]
>>>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM
>>>> > > >> >> >> >To: [email protected]
>>>> > > >> >> >> >Cc: Kulbhushan Rana
>>>> > > >> >> >> >Subject: Re: Error while making cube & Measure option is
>>>>not
>>>> > > >> >>responding
>>>> > > >> >> >> >on GUI
>>>> > > >> >> >> >
>>>> > > >> >> >> >Hi Santosh, hive table importing issue should not impact
>>>>on
>>>> > cube
>>>> > > >> >> >>saving.
>>>> > > >> >> >> >
>>>> > > >> >> >> >If you couldn’t save the cube, firstly please check
>>>>whether
>>>> > > >>there is
>>>> > > >> >> >>error
>>>> > > >> >> >> >in the tomcat’s log; If not please check your web
>>>>browser; We
>>>> > > >> >>suggest
>>>> > > >> >> >>use
>>>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS
>>>>console
>>>> > > >>(press
>>>> > > >> >> >>F12)
>>>> > > >> >> >> >and then operate web UI, check whether there is any
>>>>error
>>>> > > >>reported
>>>> > > >> >>in
>>>> > > >> >> >> >browser.
>>>> > > >> >> >> >
>>>> > > >> >> >> >
>>>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
>>>> > > >><[email protected]
>>>> > > >> >
>>>> > > >> >> >> >wrote:
>>>> > > >> >> >> >
>>>> > > >> >> >> >>Hi Shaofeng ,
>>>> > > >> >> >> >>   Thanks for replying.
>>>> > > >> >> >> >>   Yes I am checking the yarn exception, But I find
>>>>that
>>>> this
>>>> > > >>error
>>>> > > >> >> >>comes
>>>> > > >> >> >> >>while importing the hive table to kylin.
>>>> > > >> >> >> >>   Even if this error comes , hive tables is exported
>>>> > > >>successfully
>>>> > > >> >>in
>>>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ?
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>   Next step when I go on creating the cube for
>>>>following
>>>> > > >>schema  I
>>>> > > >> >> >>get
>>>> > > >> >> >> >>error at last step while saving and  I am unable to add
>>>>any
>>>> > > >> >>measures ,
>>>> > > >> >> >> >>clicking on measure option just dont pop up any dialog,
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>I am using star schema with fact_sales as fact table
>>>>and
>>>> dim_*
>>>> > > >>as
>>>> > > >> >> >> >>dimension tables.
>>>> > > >> >> >> >>
>>>> > > >> >> >> >> fact_sales:
>>>> > > >> >> >> >> storeid                 int
>>>> > > >> >> >> >> itemid                  int
>>>> > > >> >> >> >> custid                  int
>>>> > > >> >> >> >> qty                     int
>>>> > > >> >> >> >> price                   double
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>dim_customer
>>>> > > >> >> >> >> custid                  int
>>>> > > >> >> >> >> name                    string
>>>> > > >> >> >> >>
>>>> > > >> >> >> >> dim_item
>>>> > > >> >> >> >> itemid                  int
>>>> > > >> >> >> >> category                string
>>>> > > >> >> >> >> brand                   string
>>>> > > >> >> >> >> color                   string
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>dim_store
>>>> > > >> >> >> >> storeid                 int
>>>> > > >> >> >> >> city                    string
>>>> > > >> >> >> >> state                   string
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>JSON is as below.
>>>> > > >> >> >> >> The JSON is as below.
>>>> > > >> >> >> >>
>>>> > > >> >> >> >> {
>>>> > > >> >> >> >>   "name": "Retail_Cube",
>>>> > > >> >> >> >>   "description": "",
>>>> > > >> >> >> >>   "dimensions": [
>>>> > > >> >> >> >>     {
>>>> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
>>>> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>       "hierarchy": false,
>>>> > > >> >> >> >>       "derived": null,
>>>> > > >> >> >> >>       "column": [
>>>> > > >> >> >> >>         "STOREID"
>>>> > > >> >> >> >>       ],
>>>> > > >> >> >> >>       "id": 1
>>>> > > >> >> >> >>     },
>>>> > > >> >> >> >>     {
>>>> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
>>>> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>       "hierarchy": false,
>>>> > > >> >> >> >>       "derived": null,
>>>> > > >> >> >> >>       "column": [
>>>> > > >> >> >> >>         "ITEMID"
>>>> > > >> >> >> >>       ],
>>>> > > >> >> >> >>       "id": 2
>>>> > > >> >> >> >>     },
>>>> > > >> >> >> >>     {
>>>> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
>>>> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>       "hierarchy": false,
>>>> > > >> >> >> >>       "derived": null,
>>>> > > >> >> >> >>       "column": [
>>>> > > >> >> >> >>         "CUSTID"
>>>> > > >> >> >> >>       ],
>>>> > > >> >> >> >>       "id": 3
>>>> > > >> >> >> >>     }
>>>> > > >> >> >> >>   ],
>>>> > > >> >> >> >>   "measures": [
>>>> > > >> >> >> >>     {
>>>> > > >> >> >> >>       "id": 1,
>>>> > > >> >> >> >>       "name": "_COUNT_",
>>>> > > >> >> >> >>       "function": {
>>>> > > >> >> >> >>         "expression": "COUNT",
>>>> > > >> >> >> >>         "returntype": "bigint",
>>>> > > >> >> >> >>         "parameter": {
>>>> > > >> >> >> >>           "type": "constant",
>>>> > > >> >> >> >>           "value": 1
>>>> > > >> >> >> >>         }
>>>> > > >> >> >> >>       }
>>>> > > >> >> >> >>     }
>>>> > > >> >> >> >>   ],
>>>> > > >> >> >> >>   "rowkey": {
>>>> > > >> >> >> >>     "rowkey_columns": [
>>>> > > >> >> >> >>       {
>>>> > > >> >> >> >>         "column": "STOREID",
>>>> > > >> >> >> >>         "length": 0,
>>>> > > >> >> >> >>         "dictionary": "true",
>>>> > > >> >> >> >>         "mandatory": false
>>>> > > >> >> >> >>       },
>>>> > > >> >> >> >>       {
>>>> > > >> >> >> >>         "column": "ITEMID",
>>>> > > >> >> >> >>         "length": 0,
>>>> > > >> >> >> >>         "dictionary": "true",
>>>> > > >> >> >> >>         "mandatory": false
>>>> > > >> >> >> >>       },
>>>> > > >> >> >> >>       {
>>>> > > >> >> >> >>         "column": "CUSTID",
>>>> > > >> >> >> >>         "length": 0,
>>>> > > >> >> >> >>         "dictionary": "true",
>>>> > > >> >> >> >>         "mandatory": false
>>>> > > >> >> >> >>       }
>>>> > > >> >> >> >>     ],
>>>> > > >> >> >> >>     "aggregation_groups": [
>>>> > > >> >> >> >>       [
>>>> > > >> >> >> >>         "STOREID",
>>>> > > >> >> >> >>         "ITEMID",
>>>> > > >> >> >> >>         "CUSTID"
>>>> > > >> >> >> >>       ]
>>>> > > >> >> >> >>     ]
>>>> > > >> >> >> >>   },
>>>> > > >> >> >> >>   "notify_list": [],
>>>> > > >> >> >> >>   "capacity": "",
>>>> > > >> >> >> >>   "hbase_mapping": {
>>>> > > >> >> >> >>     "column_family": [
>>>> > > >> >> >> >>       {
>>>> > > >> >> >> >>         "name": "f1",
>>>> > > >> >> >> >>         "columns": [
>>>> > > >> >> >> >>           {
>>>> > > >> >> >> >>             "qualifier": "m",
>>>> > > >> >> >> >>             "measure_refs": [
>>>> > > >> >> >> >>               "_COUNT_"
>>>> > > >> >> >> >>             ]
>>>> > > >> >> >> >>           }
>>>> > > >> >> >> >>         ]
>>>> > > >> >> >> >>       }
>>>> > > >> >> >> >>     ]
>>>> > > >> >> >> >>   },
>>>> > > >> >> >> >>   "project": "RetailProject",
>>>> > > >> >> >> >>   "model_name": "Retail_Cube"
>>>> > > >> >> >> >> }
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>Regards,
>>>> > > >> >> >> >>Santosh Akhilesh
>>>> > > >> >> >> >>Bangalore R&D
>>>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>www.huawei.com
>>>> > > >> >> >>
>>>>>>----------------------------------------------------------
>>>> > > >> >> >> ---------------
>>>> > > >> >> >> >>-
>>>> > > >> >> >>
>>>>>>-----------------------------------------------------------
>>>> > > >> >> >> >>This e-mail and its attachments contain confidential
>>>> > information
>>>> > > >> >>from
>>>> > > >> >> >> >>HUAWEI, which
>>>> > > >> >> >> >>is intended only for the person or entity whose address
>>>>is
>>>> > > >>listed
>>>> > > >> >> >>above.
>>>> > > >> >> >> >>Any use of the
>>>> > > >> >> >> >>information contained herein in any way (including, but
>>>>not
>>>> > > >>limited
>>>> > > >> >> >>to,
>>>> > > >> >> >> >>total or partial
>>>> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons
>>>>other
>>>> > > >>than
>>>> > > >> >>the
>>>> > > >> >> >> >>intended
>>>> > > >> >> >> >>recipient(s) is prohibited. If you receive this e-mail
>>>>in
>>>> > error,
>>>> > > >> >> >>please
>>>> > > >> >> >> >>notify the sender by
>>>> > > >> >> >> >>phone or email immediately and delete it!
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>________________________________________
>>>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]]
>>>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM
>>>> > > >> >> >> >>To: [email protected]
>>>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option 
>>>>is
>>>>not
>>>> > > >> >> >>responding
>>>> > > >> >> >> >>on GUI
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>Hi Santosh,
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>It looks like hadoop failed to execute some shell
>>>>command in
>>>> > the
>>>> > > >> >> >> >>container; You need dive into hadoop to see what¹s the
>>>> > concrete
>>>> > > >> >>error.
>>>> > > >> >> >> >>You
>>>> > > >> >> >> >>can use yarn logs command to fetch all logs:
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>yarn logs -applicationId <app_id>
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
>>>> > > >><[email protected]
>>>> > > >> >
>>>> > > >> >> >> >>wrote:
>>>> > > >> >> >> >>
>>>> > > >> >> >> >>>Hi Luke / Shaofeng ,
>>>> > > >> >> >> >>>           Can you please help me to check this issue.
>>>> > > >> >> >> >>>Regards,
>>>> > > >> >> >> >>>Santosh Akhilesh
>>>> > > >> >> >> >>>
>>>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>>> > > >> >> >> >>>[email protected]> wrote:
>>>> > > >> >> >> >>>
>>>> > > >> >> >> >>>> Hi All ,
>>>> > > >> >> >> >>>>         is it because of following error in map
>>>>reduce
>>>> job
>>>> > ?
>>>> > > >> >>what
>>>> > > >> >> >> >>>>could
>>>> > > >> >> >> >>>>be
>>>> > > >> >> >> >>>> way to resolve this , a google search says that its
>>>>issue
>>>> > of
>>>> > > >> >>Yarn
>>>> > > >> >> >> >>>>class
>>>> > > >> >> >> >>>> path , but I am not sure what it is ?
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job
>>>>table=RETAIL.FACT_SALES
>>>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2
>>>>times
>>>> > due
>>>> > > >> >>to AM
>>>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002
>>>>exited
>>>> > > >>with
>>>> > > >> >> >> >>>>exitCode: 1
>>>> > > >> >> >> >>>> For more detailed output, check application tracking
>>>> page:
>>>> > > >> >> >> >>>>
>>>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then,
>>>> > > >> >> >>click
>>>> > > >> >> >> >>>>on
>>>> > > >> >> >> >>>> links to logs of each attempt.
>>>> > > >> >> >> >>>> Diagnostics: Exception from container-launch.
>>>> > > >> >> >> >>>> Container id: container_1424791969399_0008_02_000001
>>>> > > >> >> >> >>>> Exit code: 1
>>>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1:
>>>> > > >> >> >> >>>> at
>>>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>>> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
>>>> > > >> >> >> e(Shell.java:71
>>>> > > >> >> >> >>>>5
>>>> > > >> >> >> >>>>)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
>>>> > > >> >> >> erExecutor.laun
>>>> > > >> >> >> >>>>c
>>>> > > >> >> >> >>>>h
>>>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>>> > > >> >> >> er.launcher.Con
>>>> > > >> >> >> >>>>t
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>>> > > >> >> >> er.launcher.Con
>>>> > > >> >> >> >>>>t
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
>>>> > > >> >> >> >>>> at
>>>> java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>>> > > >> >> >> PoolExecutor.ja
>>>> > > >> >> >> >>>>v
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>:1145)
>>>> > > >> >> >> >>>> at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>>> > > >> >> >> dPoolExecutor.j
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>v
>>>> > > >> >> >> >>>>a:615)
>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745)
>>>> > > >> >> >> >>>> Container exited with a non-zero exit code 1
>>>> > > >> >> >> >>>> Failing this attempt. Failing the application.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> ---------- Forwarded message ----------
>>>> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]>
>>>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure 
>>>>option
>>>>is
>>>> > not
>>>> > > >> >> >> >>>>responding
>>>> > > >> >> >> >>>>on
>>>> > > >> >> >> >>>> GUI
>>>> > > >> >> >> >>>> To: "[email protected]"
>>>> > > >> >> >><[email protected]>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> hi ,
>>>> > > >> >> >> >>>>    please someone give me a hand to resolve this
>>>>issue ,
>>>> > > >>thanks.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Regards,
>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>> > > >> >> >> >>>> Bangalore R&D
>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> www.huawei.com
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>--------------------------------------------------------
>>>> > > >> >> >> ---------------
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >>
>>>> >>>>------------------------------------------------------------
>>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential
>>>> > > >>information
>>>> > > >> >> >>from
>>>> > > >> >> >> >>>> HUAWEI, which
>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>address
>>>> is
>>>> > > >> >>listed
>>>> > > >> >> >> >>>>above.
>>>> > > >> >> >> >>>> Any use of the
>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>but
>>>> not
>>>> > > >> >>limited
>>>> > > >> >> >> >>>>to,
>>>> > > >> >> >> >>>> total or partial
>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>persons
>>>> > other
>>>> > > >> >>than
>>>> > > >> >> >>the
>>>> > > >> >> >> >>>> intended
>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>e-mail in
>>>> > > >>error,
>>>> > > >> >> >> >>>>please
>>>> > > >> >> >> >>>> notify the sender by
>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> ________________________________________
>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]]
>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>>> > > >> >> >> >>>> To: [email protected]
>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana
>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure 
>>>>option
>>>>is
>>>> > not
>>>> > > >> >> >> >>>>responding
>>>> > > >> >> >> >>>>on
>>>> > > >> >> >> >>>> GUI
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> 2. If  I ignore and continue and try to save the 
>>>>cube
>>>>I
>>>> get
>>>> > > >>an
>>>> > > >> >> >> >>>>exception
>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
>>>>correctly
>>>> and
>>>> > > >> >> >> >>>>HCatInputFormat
>>>> > > >> >> >> >>>> this file is present in 
>>>>hive-hcatalog-core-0.14.0.jar
>>>>.
>>>> > > >>Please
>>>> > > >> >>let
>>>> > > >> >> >>me
>>>> > > >> >> >> >>>>know
>>>> > > >> >> >> >>>> what can I do to resolve this ?
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>  -- This was path issue , now no more exception in
>>>> > kylin.log
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> But saveing cube still fails with error. And still
>>>>can't
>>>> > add
>>>> > > >> >> >>measures.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Error Message
>>>> > > >> >> >> >>>> Failed to take action.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> In log I can find no exception. Following is the 
>>>>last
>>>>log
>>>> > in
>>>> > > >> >> >>kylin.log
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
>>>> > > >> >> >> .DefaultSchedul
>>>> > > >> >> >> >>>>e
>>>> > > >> >> >> >>>>r
>>>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready,
>>>>6
>>>> > others
>>>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
>>>> > > >> >> >> beController.de
>>>> > > >> >> >> >>>>s
>>>> > > >> >> >> >>>>e
>>>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)]
>>>> > > >> >> >> >>>> - Saving cube {
>>>> > > >> >> >> >>>>   "name": "",
>>>> > > >> >> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>>>   "lookups": [],
>>>> > > >> >> >> >>>>   "filter_condition": "",
>>>> > > >> >> >> >>>>   "capacity": "SMALL",
>>>> > > >> >> >> >>>>   "partition_desc": {
>>>> > > >> >> >> >>>>     "partition_date_column": "",
>>>> > > >> >> >> >>>>     "partition_date_start": 0,
>>>> > > >> >> >> >>>>     "partition_type": "APPEND"
>>>> > > >> >> >> >>>>   },
>>>> > > >> >> >> >>>>   "last_modified": 0
>>>> > > >> >> >> >>>> }
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> local access logs all with 200 , so seems ok.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
>>>> > > >> >> >>/kylin/api/cubes
>>>> > > >> >> >> >>>> HTTP/1.1" 200 701
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Regards,
>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>> > > >> >> >> >>>> Bangalore R&D
>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> www.huawei.com
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>--------------------------------------------------------
>>>> > > >> >> >> ---------------
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >>
>>>> >>>>------------------------------------------------------------
>>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential
>>>> > > >>information
>>>> > > >> >> >>from
>>>> > > >> >> >> >>>> HUAWEI, which
>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>address
>>>> is
>>>> > > >> >>listed
>>>> > > >> >> >> >>>>above.
>>>> > > >> >> >> >>>> Any use of the
>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>but
>>>> not
>>>> > > >> >>limited
>>>> > > >> >> >> >>>>to,
>>>> > > >> >> >> >>>> total or partial
>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>persons
>>>> > other
>>>> > > >> >>than
>>>> > > >> >> >>the
>>>> > > >> >> >> >>>> intended
>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>e-mail in
>>>> > > >>error,
>>>> > > >> >> >> >>>>please
>>>> > > >> >> >> >>>> notify the sender by
>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> ________________________________________
>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]]
>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>>> > > >> >> >> >>>> To: [email protected]
>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana
>>>> > > >> >> >> >>>> Subject: Error while making cube & Measure option is
>>>>not
>>>> > > >> >> >>responding on
>>>> > > >> >> >> >>>>GUI
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Hi All ,
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>     I am building a simple cube for test and using
>>>>the
>>>> > binary
>>>> > > >> >>build
>>>> > > >> >> >> >>>>0.7.1
>>>> > > >> >> >> >>>> . I have following hive tables with columns.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> fact_sales:
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> storeid                 int
>>>> > > >> >> >> >>>> itemid                  int
>>>> > > >> >> >> >>>> custid                  int
>>>> > > >> >> >> >>>> qty                     int
>>>> > > >> >> >> >>>> price                   double
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> dim_customer
>>>> > > >> >> >> >>>> custid                  int
>>>> > > >> >> >> >>>> name                    string
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> dim_item
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> itemid                  int
>>>> > > >> >> >> >>>> category                string
>>>> > > >> >> >> >>>> brand                   string
>>>> > > >> >> >> >>>> color                   string
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> dim_store
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> storeid                 int
>>>> > > >> >> >> >>>> city                    string
>>>> > > >> >> >> >>>> state                   string
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Please help me to answer following issues;
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> 1. When I go to measure section and click on measure
>>>> > option ,
>>>> > > >> >> >>there is
>>>> > > >> >> >> >>>>no
>>>> > > >> >> >> >>>> response , I want add measure on qty and price with
>>>>sum
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> 2. If  I ignore and continue and try to save the 
>>>>cube
>>>>I
>>>> get
>>>> > > >>an
>>>> > > >> >> >> >>>>exception
>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
>>>>correctly
>>>> and
>>>> > > >> >> >> >>>>HCatInputFormat
>>>> > > >> >> >> >>>> this file is present in 
>>>>hive-hcatalog-core-0.14.0.jar
>>>>.
>>>> > > >>Please
>>>> > > >> >>let
>>>> > > >> >> >>me
>>>> > > >> >> >> >>>>know
>>>> > > >> >> >> >>>> what can I do to resolve this ?
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> 3. Also I have another question since this is a test
>>>>and
>>>> > > >>data is
>>>> > > >> >> >>small
>>>> > > >> >> >> >>>>I
>>>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to
>>>>skip
>>>> > > >>partition
>>>> > > >> >> >>stage
>>>> > > >> >> >> >>>> while cube build ?
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Exception
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
>>>> > > >> >> >> l.DefaultSchedu
>>>> > > >> >> >> >>>>l
>>>> > > >> >> >> >>>>e
>>>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>>> > > >> >> >> >>>> - ExecuteException
>>>> job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError:
>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>> > > >> >> >> >>>>  at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut
>>>> > > >> >> >> e(AbstractExecu
>>>> > > >> >> >> >>>>t
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>ble.java:102)
>>>> > > >> >> >> >>>>  at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
>>>> > > >> >> >> bRunner.run(Def
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>u
>>>> > > >> >> >> >>>>ltScheduler.java:132)
>>>> > > >> >> >> >>>>  at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>>> > > >> >> >> PoolExecutor.ja
>>>> > > >> >> >> >>>>v
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>:1145)
>>>> > > >> >> >> >>>>  at
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>>> > > >> >> >> dPoolExecutor.j
>>>> > > >> >> >> >>>>a
>>>> > > >> >> >> >>>>v
>>>> > > >> >> >> >>>>a:615)
>>>> > > >> >> >> >>>>  at java.lang.Thread.run(Thread.java:745)
>>>> > > >> >> >> >>>> Caused by:
>>>> org.apache.kylin.job.exception.ExecuteException:
>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError:
>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> The JSON is as below.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> {
>>>> > > >> >> >> >>>>   "name": "Retail_Cube",
>>>> > > >> >> >> >>>>   "description": "",
>>>> > > >> >> >> >>>>   "dimensions": [
>>>> > > >> >> >> >>>>     {
>>>> > > >> >> >> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
>>>> > > >> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>>>       "hierarchy": false,
>>>> > > >> >> >> >>>>       "derived": null,
>>>> > > >> >> >> >>>>       "column": [
>>>> > > >> >> >> >>>>         "STOREID"
>>>> > > >> >> >> >>>>       ],
>>>> > > >> >> >> >>>>       "id": 1
>>>> > > >> >> >> >>>>     },
>>>> > > >> >> >> >>>>     {
>>>> > > >> >> >> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>>> > > >> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>>>       "hierarchy": false,
>>>> > > >> >> >> >>>>       "derived": null,
>>>> > > >> >> >> >>>>       "column": [
>>>> > > >> >> >> >>>>         "ITEMID"
>>>> > > >> >> >> >>>>       ],
>>>> > > >> >> >> >>>>       "id": 2
>>>> > > >> >> >> >>>>     },
>>>> > > >> >> >> >>>>     {
>>>> > > >> >> >> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>>> > > >> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>>>> > > >> >> >> >>>>       "hierarchy": false,
>>>> > > >> >> >> >>>>       "derived": null,
>>>> > > >> >> >> >>>>       "column": [
>>>> > > >> >> >> >>>>         "CUSTID"
>>>> > > >> >> >> >>>>       ],
>>>> > > >> >> >> >>>>       "id": 3
>>>> > > >> >> >> >>>>     }
>>>> > > >> >> >> >>>>   ],
>>>> > > >> >> >> >>>>   "measures": [
>>>> > > >> >> >> >>>>     {
>>>> > > >> >> >> >>>>       "id": 1,
>>>> > > >> >> >> >>>>       "name": "_COUNT_",
>>>> > > >> >> >> >>>>       "function": {
>>>> > > >> >> >> >>>>         "expression": "COUNT",
>>>> > > >> >> >> >>>>         "returntype": "bigint",
>>>> > > >> >> >> >>>>         "parameter": {
>>>> > > >> >> >> >>>>           "type": "constant",
>>>> > > >> >> >> >>>>           "value": 1
>>>> > > >> >> >> >>>>         }
>>>> > > >> >> >> >>>>       }
>>>> > > >> >> >> >>>>     }
>>>> > > >> >> >> >>>>   ],
>>>> > > >> >> >> >>>>   "rowkey": {
>>>> > > >> >> >> >>>>     "rowkey_columns": [
>>>> > > >> >> >> >>>>       {
>>>> > > >> >> >> >>>>         "column": "STOREID",
>>>> > > >> >> >> >>>>         "length": 0,
>>>> > > >> >> >> >>>>         "dictionary": "true",
>>>> > > >> >> >> >>>>         "mandatory": false
>>>> > > >> >> >> >>>>       },
>>>> > > >> >> >> >>>>       {
>>>> > > >> >> >> >>>>         "column": "ITEMID",
>>>> > > >> >> >> >>>>         "length": 0,
>>>> > > >> >> >> >>>>         "dictionary": "true",
>>>> > > >> >> >> >>>>         "mandatory": false
>>>> > > >> >> >> >>>>       },
>>>> > > >> >> >> >>>>       {
>>>> > > >> >> >> >>>>         "column": "CUSTID",
>>>> > > >> >> >> >>>>         "length": 0,
>>>> > > >> >> >> >>>>         "dictionary": "true",
>>>> > > >> >> >> >>>>         "mandatory": false
>>>> > > >> >> >> >>>>       }
>>>> > > >> >> >> >>>>     ],
>>>> > > >> >> >> >>>>     "aggregation_groups": [
>>>> > > >> >> >> >>>>       [
>>>> > > >> >> >> >>>>         "STOREID",
>>>> > > >> >> >> >>>>         "ITEMID",
>>>> > > >> >> >> >>>>         "CUSTID"
>>>> > > >> >> >> >>>>       ]
>>>> > > >> >> >> >>>>     ]
>>>> > > >> >> >> >>>>   },
>>>> > > >> >> >> >>>>   "notify_list": [],
>>>> > > >> >> >> >>>>   "capacity": "",
>>>> > > >> >> >> >>>>   "hbase_mapping": {
>>>> > > >> >> >> >>>>     "column_family": [
>>>> > > >> >> >> >>>>       {
>>>> > > >> >> >> >>>>         "name": "f1",
>>>> > > >> >> >> >>>>         "columns": [
>>>> > > >> >> >> >>>>           {
>>>> > > >> >> >> >>>>             "qualifier": "m",
>>>> > > >> >> >> >>>>             "measure_refs": [
>>>> > > >> >> >> >>>>               "_COUNT_"
>>>> > > >> >> >> >>>>             ]
>>>> > > >> >> >> >>>>           }
>>>> > > >> >> >> >>>>         ]
>>>> > > >> >> >> >>>>       }
>>>> > > >> >> >> >>>>     ]
>>>> > > >> >> >> >>>>   },
>>>> > > >> >> >> >>>>   "project": "RetailProject",
>>>> > > >> >> >> >>>>   "model_name": "Retail_Cube"
>>>> > > >> >> >> >>>> }
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> Regards,
>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>> > > >> >> >> >>>> Bangalore R&D
>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> www.huawei.com
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >>
>>>>>>>>--------------------------------------------------------
>>>> > > >> >> >> ---------------
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >> >>>>-
>>>> > > >> >> >>
>>>> >>>>------------------------------------------------------------
>>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential
>>>> > > >>information
>>>> > > >> >> >>from
>>>> > > >> >> >> >>>> HUAWEI, which
>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>address
>>>> is
>>>> > > >> >>listed
>>>> > > >> >> >> >>>>above.
>>>> > > >> >> >> >>>> Any use of the
>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>but
>>>> not
>>>> > > >> >>limited
>>>> > > >> >> >> >>>>to,
>>>> > > >> >> >> >>>> total or partial
>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>persons
>>>> > other
>>>> > > >> >>than
>>>> > > >> >> >>the
>>>> > > >> >> >> >>>> intended
>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>e-mail in
>>>> > > >>error,
>>>> > > >> >> >> >>>>please
>>>> > > >> >> >> >>>> notify the sender by
>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>> --
>>>> > > >> >> >> >>>> Regards,
>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>> > > >> >> >> >>>> +91-0-9845482201
>>>> > > >> >> >> >>>>
>>>> > > >> >> >> >>>
>>>> > > >> >> >> >>>
>>>> > > >> >> >> >>>
>>>> > > >> >> >> >>>--
>>>> > > >> >> >> >>>Regards,
>>>> > > >> >> >> >>>Santosh Akhilesh
>>>> > > >> >> >> >>>+91-0-9845482201
>>>> > > >> >> >>
>>>> > > >> >> >>
>>>> > > >> >>
>>>> > > >> >>
>>>> > > >>
>>>> > > >>
>>>> > >
>>>> > >
>>>> >
>>>> >
>>>> > --
>>>> > Regards,
>>>> > Santosh Akhilesh
>>>> > +91-0-9845482201
>>>> >
>>>>
>>>
>>>
>>>
>>>--
>>>Regards,
>>>Santosh Akhilesh
>>>+91-0-9845482201

Reply via email to