Hmm… please use the same level client jars; In Kylin’s pom.xml, it compiles with 2.6.0 jars: https://github.com/KylinOLAP/Kylin/blob/master/pom.xml#L19
On 2/27/15, 8:53 PM, "Santoshakhilesh" <[email protected]> wrote: >Hi Shaofeng , > I checked the hbase libs , I am using hbas 0.98.10-hadoop2 its using >hadoop-mapreduce-client-app-2.2.0.jar >but hadoop is using 2.6.0 > >Is this the issue ? > >I checked kylin POM its using the 0.98.4-hadoop2 > >Is this problem due to this mismatch ? do you sugegst me to try with >changing my hbase version ? > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Santoshakhilesh [[email protected]] >Sent: Friday, February 27, 2015 4:49 PM >To: [email protected] >Cc: Kulbhushan Rana >Subject: RE: Cube Build Failed at Last Step//RE: Error while making cube >& Measure option is not responding on GUI > >Hi Shaofeng , > I configured job histroy server and no more connection exception. now >I get the MR counter exception which we were suspecting. > My haddop version is indeed 2.6.0 , So any idea what can be done for >this ? > >QuartzScheduler_Worker-8]:[2015-02-28 >00:36:26,507][DEBUG][com.kylinolap.job.tools.HadoopStatusChecker.checkStat >us(HadoopStatusChecker.java:74)] - State of Hadoop job: >job_1424957178195_0031:FINISHED-SUCCEEDED >[QuartzScheduler_Worker-8]:[2015-02-28 >00:36:27,204][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCo >unter(JavaHadoopCmdOutput.java:176)] - No enum constant >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES >java.lang.IllegalArgumentException: No enum constant >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES > at java.lang.Enum.valueOf(Enum.java:236) > at >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(Framewo >rkCounterGroup.java:148) > at >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(Fra >meworkCounterGroup.java:182) > at >org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(Abstract >Counters.java:154) > at >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240) > at >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServic >eDelegate.java:370) > at >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at >org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation. >java:1491) > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) > at >com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.j >ava:287) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOu >tput.java:162) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.ja >va:85) > at >com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86) > at org.quartz.core.JobRunShell.run(JobRunShell.java:202) > at >org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5 >73) > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Shi, Shaofeng [[email protected]] >Sent: Friday, February 27, 2015 3:10 PM >To: [email protected] >Subject: Re: Cube Build Failed at Last Step//RE: Error while making cube >& Measure option is not responding on GUI > >0.0.0.0:10020 isn’t a valid network address I think; please check the >“mapreduce.jobhistory.address” in your mapred-site.xml; it should be >something like: > > ><property> > <name>mapreduce.jobhistory.address</name> > <value>sandbox.hortonworks.com:10020</value> ></property> > > > >On 2/27/15, 5:29 PM, "Santoshakhilesh" <[email protected]> >wrote: > >>Hi Shaofeng , >> No I have not found MR counter exception. I get following exception >>frequently. I think this is related LogHistory server of hadoop. >> >>[QuartzScheduler_Worker-23]:[2015-02-27 >>22:18:37,299][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobC >>o >>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>failed on connection exception: java.net.ConnectException: Connection >>refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >>com.kylinolap.job.exception.JobException: java.io.IOException: >>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>failed on connection exception: java.net.ConnectException: Connection >>refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >> at >>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob. >>j >>ava:289) >> at >>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdO >>u >>tput.java:162) >> at >>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.j >>a >>va:85) >> at >>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86) >> at org.quartz.core.JobRunShell.run(JobRunShell.java:202) >> at >>rg.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java: >>5 >>73) >>Caused by: java.io.IOException: java.net.ConnectException: Call From >>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >>java.net.ConnectException: Connection refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >> >>Regards, >>Santosh Akhilesh >>Bangalore R&D >>HUAWEI TECHNOLOGIES CO.,LTD. >> >>www.huawei.com >>------------------------------------------------------------------------- >>- >>----------------------------------------------------------- >>This e-mail and its attachments contain confidential information from >>HUAWEI, which >>is intended only for the person or entity whose address is listed above. >>Any use of the >>information contained herein in any way (including, but not limited to, >>total or partial >>disclosure, reproduction, or dissemination) by persons other than the >>intended >>recipient(s) is prohibited. If you receive this e-mail in error, please >>notify the sender by >>phone or email immediately and delete it! >> >>________________________________________ >>From: Shi, Shaofeng [[email protected]] >>Sent: Friday, February 27, 2015 2:47 PM >>To: [email protected] >>Cc: Kulbhushan Rana >>Subject: Re: Cube Build Failed at Last Step//RE: Error while making cube >>& Measure option is not responding on GUI >> >>Did you figure out the exception of "No enum constant >>org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES” ? Is it still >>be >>thrown in the logs? In the last step, Kylin need to parse the MR counters >>to update cube size; Please refer to >>https://issues.apache.org/jira/browse/MAPREDUCE-5831 for that error. >> >>On 2/27/15, 5:04 PM, "Santoshakhilesh" <[email protected]> >>wrote: >> >>>Hi Shaofeng , >>> Cube building is failed at last step while loading Hfile to >>>Hbase with exception "Can't get cube segment size. >>>". What could be reason ? >>> >>>parameter : -input >>>/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >>>-htablename KYLIN_K27LDMX63W -cubename RetailCube >>> >>>Log: >>> >>>Start to execute command: >>> -input >>>/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >>>-htablename KYLIN_K27LDMX63W -cubename RetailCube >>>Command execute return code 0 >>>Failed with Exception:java.lang.RuntimeException: Can't get cube segment >>>size. >>> at >>>com.kylinolap.job.flow.JobFlowListener.updateCubeSegmentInfoOnSucceed(Jo >>>b >>>F >>>lowListener.java:247) >>> at >>>com.kylinolap.job.flow.JobFlowListener.jobWasExecuted(JobFlowListener.ja >>>v >>>a >>>:101) >>> at >>>org.quartz.core.QuartzScheduler.notifyJobListenersWasExecuted(QuartzSche >>>d >>>u >>>ler.java:1985) >>> at >>>org.quartz.core.JobRunShell.notifyJobListenersComplete(JobRunShell.java: >>>3 >>>4 >>>0) >>> at org.quartz.core.JobRunShell.run(JobRunShell.java:224) >>> at >>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java >>>: >>>5 >>>73) >>> >>>I have checked in hbase shell and following are the tables in hbase; >>>hbase(main):001:0> list >>>TABLE >>> >>>KYLIN_K27LDMX63W >>>kylin_metadata_qa >>>kylin_metadata_qa_acl >>>kylin_metadata_qa_cube >>>kylin_metadata_qa_dict >>>kylin_metadata_qa_invertedindex >>>kylin_metadata_qa_job >>>kylin_metadata_qa_job_output >>>kylin_metadata_qa_proj >>>kylin_metadata_qa_table_snapshot >>>kylin_metadata_qa_user >>>11 row(s) in 0.8990 seconds >>> >>> >>>Regards, >>>Santosh Akhilesh >>>Bangalore R&D >>>HUAWEI TECHNOLOGIES CO.,LTD. >>> >>>www.huawei.com >>>------------------------------------------------------------------------ >>>- >>>- >>>----------------------------------------------------------- >>>This e-mail and its attachments contain confidential information from >>>HUAWEI, which >>>is intended only for the person or entity whose address is listed above. >>>Any use of the >>>information contained herein in any way (including, but not limited to, >>>total or partial >>>disclosure, reproduction, or dissemination) by persons other than the >>>intended >>>recipient(s) is prohibited. If you receive this e-mail in error, please >>>notify the sender by >>>phone or email immediately and delete it! >>> >>>________________________________________ >>>From: Santoshakhilesh >>>Sent: Friday, February 27, 2015 2:15 PM >>>To: [email protected] >>>Subject: RE: Error while making cube & Measure option is not responding >>>on GUI >>> >>>I have manually copied the jar to /tmp/kylin , now satge 2 is done , >>>thanks. >>> >>>Regards, >>>Santosh Akhilesh >>>Bangalore R&D >>>HUAWEI TECHNOLOGIES CO.,LTD. >>> >>>www.huawei.com >>>------------------------------------------------------------------------ >>>- >>>- >>>----------------------------------------------------------- >>>This e-mail and its attachments contain confidential information from >>>HUAWEI, which >>>is intended only for the person or entity whose address is listed above. >>>Any use of the >>>information contained herein in any way (including, but not limited to, >>>total or partial >>>disclosure, reproduction, or dissemination) by persons other than the >>>intended >>>recipient(s) is prohibited. If you receive this e-mail in error, please >>>notify the sender by >>>phone or email immediately and delete it! >>> >>>________________________________________ >>>From: Shi, Shaofeng [[email protected]] >>>Sent: Friday, February 27, 2015 1:00 PM >>>To: [email protected] >>>Cc: Kulbhushan Rana >>>Subject: Re: Error while making cube & Measure option is not responding >>>on GUI >>> >>>In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we >>>renamed the package to “org.apache.kylin.xxx”; When you downgrade to >>>0.6, >>>did you also replace the jar location with 0.6 ones in kylin.properties? >>> >>>On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]> >>>wrote: >>> >>>>Hi Shaofeng , >>>> I have added my fact and dimension tables under default >>>>database >>>>of hive. >>>> Now stage 1 of Cube Build is ok. And there is failure at >>>>step2. >>>> The map reduce job for the finding distinct columns of fact >>>>table >>>>is error. Yarn log is as below. >>>> Strangely this is class not found error. I have checked the >>>>Kylin.properties and the jar is already set as below. >>>>kylin. log has one exception connecting to linux/10.19.93.68 to >>>>0.0.0.0:10020 >>>> Please help me to give a clue , I am also trying to check meanwhile >>>> >>>>Thanks. >>>>kylin property >>>># Temp folder in hdfs >>>>kylin.hdfs.working.dir=/tmp >>>># Path to the local(relative to job engine) job jar, job engine will >>>>use >>>>this jar >>>>kylin.job.jar=/tmp/kylin/kylin-job-latest.jar >>>> >>>>Map Reduce error >>>>---------------------------- >>>>2015-02-27 20:24:25,262 FATAL [main] >>>>org.apache.hadoop.mapred.YarnChild: >>>>Error running child : java.lang.NoClassDefFoundError: >>>>com/kylinolap/common/mr/KylinMapper >>>> at java.lang.ClassLoader.defineClass1(Native Method) >>>> at java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>>> at >>>>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>>> at java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >>>> at java.lang.Class.forName0(Native Method) >>>> at java.lang.Class.forName(Class.java:274) >>>> at >>>>org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration >>>>. >>>>j >>>>a >>>>va:2013) >>>> >>>>Kylin.log >>>>QuartzScheduler_Worker-20]:[2015-02-27 >>>>20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFet >>>>c >>>>h >>>>e >>>>r.java:60)] - 0 pending jobs >>>>[QuartzScheduler_Worker-19]:[2015-02-27 >>>>20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJo >>>>b >>>>C >>>>o >>>>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >>>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>>>failed on connection exception: java.net.ConnectException: Connection >>>>refused; For more details see: >>>>http://wiki.apache.org/hadoop/ConnectionRefused >>>>com.kylinolap.job.exception.JobException: java.io.IOException: >>>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>>>failed on connection exception: java.net.ConnectException: Connection >>>>refused; For more details see: >>>>http://wiki.apache.org/hadoop/ConnectionRefused >>>> at >>>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJo >>>>b >>>>. >>>>j >>>>ava:289) >>>> at >>>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCm >>>>d >>>>O >>>>u >>>>tput.java:162) >>>> at >>>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput >>>>. >>>>j >>>>a >>>>va:85) >>>> at >>>>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:8 >>>>6 >>>>) >>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:202) >>>> at >>>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.jav >>>>a >>>>: >>>>5 >>>>73) >>>>Caused by: java.io.IOException: java.net.ConnectException: Call From >>>>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >>>>java.net.ConnectException: Connection refused; For more details see: >>>>http://wiki.apache.org/hadoop/ConnectionRefused >>>> at >>>>org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDele >>>>g >>>>a >>>>t >>>>e.java:331) >>>> at >>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientSer >>>>v >>>>i >>>>c >>>>eDelegate.java:368) >>>> at >>>>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) >>>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>> at >>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformati >>>>o >>>>n >>>>. >>>>java:1491) >>>> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>>> at >>>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJo >>>>b >>>>. >>>>j >>>>ava:287) >>>> ... 5 more >>>> >>>>Regards, >>>>Santosh Akhilesh >>>>Bangalore R&D >>>>HUAWEI TECHNOLOGIES CO.,LTD. >>>> >>>>www.huawei.com >>>>----------------------------------------------------------------------- >>>>- >>>>- >>>>- >>>>----------------------------------------------------------- >>>>This e-mail and its attachments contain confidential information from >>>>HUAWEI, which >>>>is intended only for the person or entity whose address is listed >>>>above. >>>>Any use of the >>>>information contained herein in any way (including, but not limited to, >>>>total or partial >>>>disclosure, reproduction, or dissemination) by persons other than the >>>>intended >>>>recipient(s) is prohibited. If you receive this e-mail in error, please >>>>notify the sender by >>>>phone or email immediately and delete it! >>>> >>>>________________________________________ >>>>From: Shi, Shaofeng [[email protected]] >>>>Sent: Friday, February 27, 2015 8:01 AM >>>>To: [email protected] >>>>Subject: Re: Error while making cube & Measure option is not responding >>>>on GUI >>>> >>>>In 0.6.x it only support tables in default database, this is a >>>>limitation; >>>>The support for non-default tables will be released in 0.7; >>>> >>>>To bypass this issue for now, please copy the table to default database >>>>as >>>>a workaround; >>>> >>>>On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]> >>>>wrote: >>>> >>>>>@Jason >>>>>thanks , but now as suggested by Saofeng I m not using the inverted >>>>>index >>>>>brach as its not stable. >>>>>I have switched back to 0.6 branch , in this branch yesterday night I >>>>>could >>>>>crete the cube successfully but there is issue while building it. I >>>>>feel >>>>>that at step 1 of cube build while creating flat table when command >>>>>is >>>>>issued to hive if the tables are not under default datbase flat table >>>>>creation is failed and cube build fails. my fact and dimension tables >>>>>are >>>>>under a database called retail. >>>>> >>>>>@Saofeng - Can you please confirm this behavior ? Do I need to create >>>>>the >>>>>hive tables under default database? >>>>> >>>>>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]> >>>>>wrote: >>>>> >>>>>> @Santoshakhilesh >>>>>> >>>>>> 1. When I go to measure section and click on measure option , there >>>>>>is >>>>>>no >>>>>> response , I want add measure on qty and price with sum >>>>>> --bug fixed on inverted-index branch >>>>>> >>>>>> >>>>>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh < >>>>>> [email protected] >>>>>> > wrote: >>>>>> >>>>>> > Hi Shaofeng , >>>>>> > I have build the 0.6 version and now able to create the cube >>>>>> > successfully. >>>>>> > While building the cube , it fails at step1 with following >>>>>>error. >>>>>> > Table not found 'DIM_ITEM' >>>>>> > the table exists , but its under retail data base and not >>>>>>under >>>>>> > default database. >>>>>> > does kylin require hive taables to be under default database >>>>>>? >>>>>>I >>>>>>see >>>>>> > the flat table being created under default database. >>>>>> > >>>>>> > Logging initialized using configuration in >>>>>> > >>>>>> > >>>>>> >>>>>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-com >>>>>>m >>>>>>o >>>>>>n >>>>>>- >>>>>>1.0.0.jar!/hive-log4j.properties >>>>>> > SLF4J: Class path contains multiple SLF4J bindings. >>>>>> > SLF4J: Found binding in >>>>>> > >>>>>> > >>>>>> >>>>>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/com >>>>>>m >>>>>>o >>>>>>n >>>>>>/ >>>>>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>> > SLF4J: Found binding in >>>>>> > >>>>>> > >>>>>> >>>>>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jd >>>>>>b >>>>>>c >>>>>>- >>>>>>1 >>>>>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for >>>>>>an >>>>>> > explanation. >>>>>> > SLF4J: Actual binding is of type >>>>>>[org.slf4j.impl.Log4jLoggerFactory] >>>>>> > OK >>>>>> > Time taken: 0.964 seconds >>>>>> > OK >>>>>> > Time taken: 0.948 seconds >>>>>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not >>>>>>found >>>>>> > 'DIM_ITEM' >>>>>> > >>>>>> > >>>>>> > >>>>>> > Command is as below. >>>>>> > >>>>>> > hive -e "DROP TABLE IF EXISTS >>>>>> > >>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd >>>>>>4 >>>>>>4 >>>>>>; >>>>>> > CREATE EXTERNAL TABLE IF NOT EXISTS >>>>>> > >>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd >>>>>>4 >>>>>>4 >>>>>> > ( >>>>>> > STOREID int >>>>>> > ,ITEMID int >>>>>> > ,CUSTID int >>>>>> > ,QTY int >>>>>> > ,AMOUNT double >>>>>> > ) >>>>>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' >>>>>> > STORED AS SEQUENCEFILE >>>>>> > LOCATION >>>>>> > >>>>>> >>>>>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_t >>>>>>e >>>>>>s >>>>>>t >>>>>>_ >>>>>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44'; >>>>>> > SET hive.exec.compress.output=true; >>>>>> > SET hive.auto.convert.join.noconditionaltask = true; >>>>>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000; >>>>>> > INSERT OVERWRITE TABLE >>>>>> > >>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd >>>>>>4 >>>>>>4 >>>>>> > SELECT >>>>>> > FACT_SALES.STOREID >>>>>> > ,FACT_SALES.ITEMID >>>>>> > ,FACT_SALES.CUSTID >>>>>> > ,FACT_SALES.QTY >>>>>> > ,FACT_SALES.AMOUNT >>>>>> > FROM FACT_SALES >>>>>> > INNER JOIN DIM_STORE >>>>>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID >>>>>> > INNER JOIN DIM_ITEM >>>>>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID >>>>>> > INNER JOIN DIM_CUSTOMER >>>>>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID >>>>>> > ; >>>>>> > " >>>>>> > >>>>>> > >>>>>> > >>>>>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]> >>>>>>wrote: >>>>>> > >>>>>> > > The 0.7.1 is test version, its package contains the “snapshot” >>>>>>suffix; >>>>>> we >>>>>> > > will upload a new package there; Luke will also add a message >>>>>>there >>>>>>to >>>>>> > > avoid this confusion; >>>>>> > > >>>>>> > > Regarding the problem that you encountered, could you please >>>>>>open >>>>>>a >>>>>> JIRA >>>>>> > > ticket for tracking? Here is link of Apache JIRA: >>>>>> > > >>>>>> > > https://issues.apache.org/jira/secure/Dashboard.jspa >>>>>> > > >>>>>> > > >>>>>> > > Thanks for the feedback! >>>>>> > > >>>>>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh" >>>>>><[email protected]> >>>>>> > > wrote: >>>>>> > > >>>>>> > > >Actually I see this being published on kylin webpage. >>>>>> > > >http://kylin.incubator.apache.org/download/ >>>>>> > > >I am using 0.7.1 inverted index branch binary distribution. >>>>>> > > >If this is not stable please give me the link of stable branch >>>>>>I >>>>>>would >>>>>> > try >>>>>> > > >building and testing tonight. >>>>>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng >>>>>><[email protected]> >>>>>> wrote: >>>>>> > > > >>>>>> > > >> Hi Santosh, it is not recommended to use the dev code branch >>>>>> > (actually I >>>>>> > > >> don’t know how you get the v0.7.x build and what’s the detail >>>>>> version >>>>>> > of >>>>>> > > >> that; each day we submit many changes to that); >>>>>> > > >> >>>>>> > > >> The options are 1) switch back to latest release v0.6.5; or >>>>>>2) >>>>>>wait >>>>>> > for >>>>>> > > >> the formal release of 0.7, that should be in March; >>>>>>Otherwise, >>>>>>we >>>>>> > > >>couldn’t >>>>>> > > >> ensure there is no new problems come out in your next steps; >>>>>> > > >> >>>>>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh" >>>>>><[email protected]> >>>>>> > > >>wrote: >>>>>> > > >> >>>>>> > > >> >Hi Shaofeng >>>>>> > > >> >So what do you suggest , how should I proceed further with >>>>>>this >>>>>> > > >>release? >>>>>> > > >> >Will there be a patch? Any alternate way I can create cube? >>>>>> > > >> >Please suggest. >>>>>> > > >> >Regards >>>>>> > > >> >Santosh >>>>>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng >>>>>><[email protected]> >>>>>> > > wrote: >>>>>> > > >> > >>>>>> > > >> >> Hi Santosh, >>>>>> > > >> >> >>>>>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x >>>>>>we >>>>>>have >>>>>> > > >> >>metadata >>>>>> > > >> >> structure change; While, the web UI (cube wizard) for this >>>>>>change >>>>>> > > >>hasn’t >>>>>> > > >> >> been stabilized; So it is not strange that you got trouble >>>>>>when >>>>>> > > >>saving >>>>>> > > >> >>the >>>>>> > > >> >> cube; >>>>>> > > >> >> >>>>>> > > >> >> @Jason, any idea about the JS error? >>>>>> > > >> >> >>>>>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" < >>>>>> [email protected] >>>>>> > > >>>>>> > > >> >>wrote: >>>>>> > > >> >> >>>>>> > > >> >> >Hi Shaofeng, >>>>>> > > >> >> > >>>>>> > > >> >> >I am using the binary distribution 0.7.1. I have not been >>>>>>able >>>>>> to >>>>>> > > >>save >>>>>> > > >> >> >cube >>>>>> > > >> >> >even once. I have tried creating new project and from >>>>>>local >>>>>> > machine >>>>>> > > >>and >>>>>> > > >> >> >server machine. But I am always stuck with this error. I >>>>>>am >>>>>> never >>>>>> > > >> >>allowed >>>>>> > > >> >> >to add measures and never been able to save the cube. I >>>>>>also >>>>>>see >>>>>> > the >>>>>> > > >> >> >kylin.log and it always tries to save cube with append >>>>>>mode. >>>>>>One >>>>>> > > >>thing >>>>>> > > >> >>I >>>>>> > > >> >> >need to tell that at partition stage since I don't have a >>>>>>big >>>>>> fact >>>>>> > > >> >>table >>>>>> > > >> >> >now I have not partititioned the fact table and I skip >>>>>>this >>>>>> step. >>>>>> > > >>Does >>>>>> > > >> >> >this >>>>>> > > >> >> >have affect in saving the cube. Is this because some >>>>>>metadata is >>>>>> > > >> >>available >>>>>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop >>>>>>2.6.6. >>>>>> > Yes >>>>>> > > >> >>kylin >>>>>> > > >> >> >propert I have not added the jar. I will add them and >>>>>>check. >>>>>>But >>>>>> > > >>cube >>>>>> > > >> >> >creation failure is really puzzling me. I could see no >>>>>>error >>>>>> logs >>>>>> > in >>>>>> > > >> >> >kylin.log. >>>>>> > > >> >> >Regards >>>>>> > > >> >> >Santosh >>>>>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng >>>>>><[email protected] >>>>>> > >>>>>> > > >> wrote: >>>>>> > > >> >> > >>>>>> > > >> >> >> Which version or code branch are you using? I assume >>>>>>you’re >>>>>> > using >>>>>> > > >>the >>>>>> > > >> >> >> stable version from master; Seems you’re trying to edit >>>>>>an >>>>>> > > >>existing >>>>>> > > >> >>cube >>>>>> > > >> >> >> to add new measurement, try refresh your browser's >>>>>>cache; >>>>>>if >>>>>> it >>>>>> > > >>still >>>>>> > > >> >> >> couldn’t be saved, try to create a new cube; >>>>>> > > >> >> >> >>>>>> > > >> >> >> The two error traces in tomcat need be taken care: >>>>>> > > >> >> >> >>>>>> > > >> >> >> 1) java.lang.NoClassDefFoundError: >>>>>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper >>>>>> > > >> >> >> Please check kylin.properties file, making sure >>>>>>the >>>>>> > > >> >> >>“kylin.job.jar” >>>>>> > > >> >> >> points to a right jar file; It will be loaded in >>>>>>Map-reduce; >>>>>> > > >> >> >> >>>>>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant >>>>>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>>>> > > >> >> >> This indicates your hadoop version might be >>>>>>old; >>>>>> Please >>>>>> > > >>check >>>>>> > > >> >> >>and >>>>>> > > >> >> >> ensure >>>>>> > > >> >> >> hadoop version is 2.2 or above. >>>>>> > > >> >> >> >>>>>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" >>>>>> > > >><[email protected]> >>>>>> > > >> >> >> wrote: >>>>>> > > >> >> >> >>>>>> > > >> >> >> >Hi Shaofeng , >>>>>> > > >> >> >> > >>>>>> > > >> >> >> > I am using chrome , When I click on button to add >>>>>> measures , >>>>>> > > >> >> >>following >>>>>> > > >> >> >> >is error on chrome console. When I try to save the >>>>>>cube >>>>>>there >>>>>> > is >>>>>> > > >>no >>>>>> > > >> >> >>error >>>>>> > > >> >> >> >in console. I just get a error dialog saying failed to >>>>>>take >>>>>> > > >>action >>>>>> > > >> >>and >>>>>> > > >> >> >> >gives me the JSON cube schema. >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >Error on chrome debug console is as below; >>>>>> > > >> >> >> > >>>>>> > > >> >> >> > ReferenceError: CubeDescModel is not defined >>>>>> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >>>>>> > > >> >> >> > at scripts.min.0.js:180 >>>>>> > > >> >> >> > at scripts.min.0.js:197 >>>>>> > > >> >> >> > at h.$eval (scripts.min.0.js:119) >>>>>> > > >> >> >> > at h.$apply (scripts.min.0.js:119) >>>>>> > > >> >> >> > at HTMLButtonElement.<anonymous> >>>>>>(scripts.min.0.js:197) >>>>>> > > >> >> >> > at HTMLButtonElement.m.event.dispatch >>>>>> (scripts.min.0.js:3) >>>>>> > > >> >> >> > at HTMLButtonElement.r.handle >>>>>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >>>>>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous >>>>>> > > >> >>function)scripts.min.0.js:119 >>>>>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous >>>>>> > > >>function)scripts.min.0.js:3 >>>>>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle >>>>>> > > >> >> >> > >>>>>> > > >> >> >> > About the hive table import , I got pass the run >>>>>>shell >>>>>> > command >>>>>> > > >> >> >> >exception but it still fails the hadoop log is; >>>>>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main] >>>>>>org.apache.hadoop.mapred. >>>>>> > > >> >> YarnChild: >>>>>> > > >> >> >> >mapreduce.cluster.local.dir for child: >>>>>> > > >> >> >> >>>>>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >>>>>> > > >> >> >> cation_14249530 >>>>>> > > >> >> >> >91340_0002 >>>>>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main] >>>>>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation: >>>>>>session.id >>>>>> > is >>>>>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id >>>>>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main] >>>>>> > > >>org.apache.hadoop.mapred.Task: >>>>>> > > >> >> >>Using >>>>>> > > >> >> >> >ResourceCalculatorProcessTree : [ ] >>>>>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main] >>>>>> > > >> >> >>org.apache.hadoop.mapred.YarnChild: >>>>>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError: >>>>>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper >>>>>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method) >>>>>> > > >> >> >> > at >>>>>>java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >java.security.SecureClassLoader.defineClass( >>>>>> > > >> >> SecureClassLoader.java:142) >>>>>> > > >> >> >> > at >>>>>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>>>>> > > >> >> >> > at >>>>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>>>>> > > >> >> >> > at >>>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>>>> > > >> >> >> > at >>>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>>>> Method) >>>>>> > > >> >> >> > at >>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>>>> > > >> >> >> > at >>>>>>java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>>>> > > >> >> >> > at >>>>>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >tomcat logs: >>>>>> > > >> >> >> >usage: HiveColumnCardinalityJob >>>>>> > > >> >> >> > -output <path> Output path >>>>>> > > >> >> >> > -table <table name> The hive table name >>>>>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26 >>>>>> > > >> >> >> >>>>>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >>>>>> > > >> >> >> lExecutable.doW >>>>>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute >>>>>> > > >> >> >> >>>>>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >>>>>> > > >> >> >> >name=null, state=RUNNING} >>>>>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant >>>>>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>>>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>> > > >>>>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf >>>>>>>>>( >>>>>> > > >> >> >> Framewo >>>>>> > > >> >> >> >rkCounterGroup.java:148) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >>>>>> > > >> >> >> findCounter(Fra >>>>>> > > >> >> >> >meworkCounterGroup.java:182) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >>>>>> > > >> >> >> ounter(Abstract >>>>>> > > >> >> >> >Counters.java:154) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >>>>>> > > >> >> >> erter.java:240) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >>>>>> > > >> >> >> rs(ClientServic >>>>>> > > >> >> >> >eDelegate.java:370) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters( >>>>>> > > >> >> YARNRunner.java:511) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>>>> Method) >>>>>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >>>>>> > > >> >> >> oupInformation. >>>>>> > > >> >> >> >java:1491) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>>>>> > > >> >> >> > at >>>>>> > > >> >>>>>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job. >>>>>> > > >> java:1289) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >>>>>> > > >> >> >> tion(AbstractHa >>>>>> > > >> >> >> >doopJob.java:134) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>> > > >> >>>>>> > >>>>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.r >>>>>>>>>u >>>>>>>>>n >>>>>>>>>( >>>>>> > > >> >> >> HiveC >>>>>> > > >> >> >> >olumnCardinalityJob.java:114) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>>>> > > >> >> >> > at >>>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >>>>>> > > >> >> >> doopShellExecut >>>>>> > > >> >> >> >able.java:62) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>>>> > > >> >> >> bstractExecutab >>>>>> > > >> >> >> >le.java:99) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW >>>>>> > > >> >> >> ork(DefaultChai >>>>>> > > >> >> >> >nedExecutable.java:50) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>>>> > > >> >> >> bstractExecutab >>>>>> > > >> >> >> >le.java:99) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >>>>>> > > >> >> >> nner.run(Defaul >>>>>> > > >> >> >> >tScheduler.java:132) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >>>>>> > > >> >> >> lExecutor.java: >>>>>> > > >> >> >> >1145) >>>>>> > > >> >> >> > at >>>>>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >>>>>> > > >> >> >> olExecutor.java >>>>>> > > >> >> >> >:615) >>>>>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745) >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >Regards, >>>>>> > > >> >> >> >Santosh Akhilesh >>>>>> > > >> >> >> >Bangalore R&D >>>>>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD. >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >www.huawei.com >>>>>> > > >> >> >> >>>>>>>----------------------------------------------------------- >>>>>> > > >> >> >> --------------- >>>>>> > > >> >> >> >>>>>>>----------------------------------------------------------- >>>>>> > > >> >> >> >This e-mail and its attachments contain confidential >>>>>> > information >>>>>> > > >> >>from >>>>>> > > >> >> >> >HUAWEI, which >>>>>> > > >> >> >> >is intended only for the person or entity whose >>>>>>address >>>>>>is >>>>>> > listed >>>>>> > > >> >> >>above. >>>>>> > > >> >> >> >Any use of the >>>>>> > > >> >> >> >information contained herein in any way (including, >>>>>>but >>>>>>not >>>>>> > > >>limited >>>>>> > > >> >>to, >>>>>> > > >> >> >> >total or partial >>>>>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons >>>>>>other >>>>>> > than >>>>>> > > >> >>the >>>>>> > > >> >> >> >intended >>>>>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail >>>>>>in >>>>>> > error, >>>>>> > > >> >>please >>>>>> > > >> >> >> >notify the sender by >>>>>> > > >> >> >> >phone or email immediately and delete it! >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >________________________________________ >>>>>> > > >> >> >> >From: Shi, Shaofeng [[email protected]] >>>>>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM >>>>>> > > >> >> >> >To: [email protected] >>>>>> > > >> >> >> >Cc: Kulbhushan Rana >>>>>> > > >> >> >> >Subject: Re: Error while making cube & Measure option >>>>>>is >>>>>>not >>>>>> > > >> >>responding >>>>>> > > >> >> >> >on GUI >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >Hi Santosh, hive table importing issue should not >>>>>>impact >>>>>>on >>>>>> > cube >>>>>> > > >> >> >>saving. >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >If you couldn’t save the cube, firstly please check >>>>>>whether >>>>>> > > >>there is >>>>>> > > >> >> >>error >>>>>> > > >> >> >> >in the tomcat’s log; If not please check your web >>>>>>browser; We >>>>>> > > >> >>suggest >>>>>> > > >> >> >>use >>>>>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS >>>>>>console >>>>>> > > >>(press >>>>>> > > >> >> >>F12) >>>>>> > > >> >> >> >and then operate web UI, check whether there is any >>>>>>error >>>>>> > > >>reported >>>>>> > > >> >>in >>>>>> > > >> >> >> >browser. >>>>>> > > >> >> >> > >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" >>>>>> > > >><[email protected] >>>>>> > > >> > >>>>>> > > >> >> >> >wrote: >>>>>> > > >> >> >> > >>>>>> > > >> >> >> >>Hi Shaofeng , >>>>>> > > >> >> >> >> Thanks for replying. >>>>>> > > >> >> >> >> Yes I am checking the yarn exception, But I find >>>>>>that >>>>>> this >>>>>> > > >>error >>>>>> > > >> >> >>comes >>>>>> > > >> >> >> >>while importing the hive table to kylin. >>>>>> > > >> >> >> >> Even if this error comes , hive tables is exported >>>>>> > > >>successfully >>>>>> > > >> >>in >>>>>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed >>>>>>? >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >> Next step when I go on creating the cube for >>>>>>following >>>>>> > > >>schema I >>>>>> > > >> >> >>get >>>>>> > > >> >> >> >>error at last step while saving and I am unable to >>>>>>add >>>>>>any >>>>>> > > >> >>measures , >>>>>> > > >> >> >> >>clicking on measure option just dont pop up any >>>>>>dialog, >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>I am using star schema with fact_sales as fact table >>>>>>and >>>>>> dim_* >>>>>> > > >>as >>>>>> > > >> >> >> >>dimension tables. >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >> fact_sales: >>>>>> > > >> >> >> >> storeid int >>>>>> > > >> >> >> >> itemid int >>>>>> > > >> >> >> >> custid int >>>>>> > > >> >> >> >> qty int >>>>>> > > >> >> >> >> price double >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>dim_customer >>>>>> > > >> >> >> >> custid int >>>>>> > > >> >> >> >> name string >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >> dim_item >>>>>> > > >> >> >> >> itemid int >>>>>> > > >> >> >> >> category string >>>>>> > > >> >> >> >> brand string >>>>>> > > >> >> >> >> color string >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>dim_store >>>>>> > > >> >> >> >> storeid int >>>>>> > > >> >> >> >> city string >>>>>> > > >> >> >> >> state string >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>JSON is as below. >>>>>> > > >> >> >> >> The JSON is as below. >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "name": "Retail_Cube", >>>>>> > > >> >> >> >> "description": "", >>>>>> > > >> >> >> >> "dimensions": [ >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID", >>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >> "hierarchy": false, >>>>>> > > >> >> >> >> "derived": null, >>>>>> > > >> >> >> >> "column": [ >>>>>> > > >> >> >> >> "STOREID" >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "id": 1 >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID", >>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >> "hierarchy": false, >>>>>> > > >> >> >> >> "derived": null, >>>>>> > > >> >> >> >> "column": [ >>>>>> > > >> >> >> >> "ITEMID" >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "id": 2 >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID", >>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >> "hierarchy": false, >>>>>> > > >> >> >> >> "derived": null, >>>>>> > > >> >> >> >> "column": [ >>>>>> > > >> >> >> >> "CUSTID" >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "id": 3 >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "measures": [ >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "id": 1, >>>>>> > > >> >> >> >> "name": "_COUNT_", >>>>>> > > >> >> >> >> "function": { >>>>>> > > >> >> >> >> "expression": "COUNT", >>>>>> > > >> >> >> >> "returntype": "bigint", >>>>>> > > >> >> >> >> "parameter": { >>>>>> > > >> >> >> >> "type": "constant", >>>>>> > > >> >> >> >> "value": 1 >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "rowkey": { >>>>>> > > >> >> >> >> "rowkey_columns": [ >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "column": "STOREID", >>>>>> > > >> >> >> >> "length": 0, >>>>>> > > >> >> >> >> "dictionary": "true", >>>>>> > > >> >> >> >> "mandatory": false >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "column": "ITEMID", >>>>>> > > >> >> >> >> "length": 0, >>>>>> > > >> >> >> >> "dictionary": "true", >>>>>> > > >> >> >> >> "mandatory": false >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "column": "CUSTID", >>>>>> > > >> >> >> >> "length": 0, >>>>>> > > >> >> >> >> "dictionary": "true", >>>>>> > > >> >> >> >> "mandatory": false >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> ], >>>>>> > > >> >> >> >> "aggregation_groups": [ >>>>>> > > >> >> >> >> [ >>>>>> > > >> >> >> >> "STOREID", >>>>>> > > >> >> >> >> "ITEMID", >>>>>> > > >> >> >> >> "CUSTID" >>>>>> > > >> >> >> >> ] >>>>>> > > >> >> >> >> ] >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> "notify_list": [], >>>>>> > > >> >> >> >> "capacity": "", >>>>>> > > >> >> >> >> "hbase_mapping": { >>>>>> > > >> >> >> >> "column_family": [ >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "name": "f1", >>>>>> > > >> >> >> >> "columns": [ >>>>>> > > >> >> >> >> { >>>>>> > > >> >> >> >> "qualifier": "m", >>>>>> > > >> >> >> >> "measure_refs": [ >>>>>> > > >> >> >> >> "_COUNT_" >>>>>> > > >> >> >> >> ] >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> ] >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> ] >>>>>> > > >> >> >> >> }, >>>>>> > > >> >> >> >> "project": "RetailProject", >>>>>> > > >> >> >> >> "model_name": "Retail_Cube" >>>>>> > > >> >> >> >> } >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>Regards, >>>>>> > > >> >> >> >>Santosh Akhilesh >>>>>> > > >> >> >> >>Bangalore R&D >>>>>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD. >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>www.huawei.com >>>>>> > > >> >> >> >>>>>>>>---------------------------------------------------------- >>>>>> > > >> >> >> --------------- >>>>>> > > >> >> >> >>- >>>>>> > > >> >> >> >>>>>>>>----------------------------------------------------------- >>>>>> > > >> >> >> >>This e-mail and its attachments contain confidential >>>>>> > information >>>>>> > > >> >>from >>>>>> > > >> >> >> >>HUAWEI, which >>>>>> > > >> >> >> >>is intended only for the person or entity whose >>>>>>address >>>>>>is >>>>>> > > >>listed >>>>>> > > >> >> >>above. >>>>>> > > >> >> >> >>Any use of the >>>>>> > > >> >> >> >>information contained herein in any way (including, >>>>>>but >>>>>>not >>>>>> > > >>limited >>>>>> > > >> >> >>to, >>>>>> > > >> >> >> >>total or partial >>>>>> > > >> >> >> >>disclosure, reproduction, or dissemination) by >>>>>>persons >>>>>>other >>>>>> > > >>than >>>>>> > > >> >>the >>>>>> > > >> >> >> >>intended >>>>>> > > >> >> >> >>recipient(s) is prohibited. If you receive this >>>>>>e-mail >>>>>>in >>>>>> > error, >>>>>> > > >> >> >>please >>>>>> > > >> >> >> >>notify the sender by >>>>>> > > >> >> >> >>phone or email immediately and delete it! >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>________________________________________ >>>>>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]] >>>>>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM >>>>>> > > >> >> >> >>To: [email protected] >>>>>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option >>>>>>is >>>>>>not >>>>>> > > >> >> >>responding >>>>>> > > >> >> >> >>on GUI >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>Hi Santosh, >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>It looks like hadoop failed to execute some shell >>>>>>command in >>>>>> > the >>>>>> > > >> >> >> >>container; You need dive into hadoop to see what¹s >>>>>>the >>>>>> > concrete >>>>>> > > >> >>error. >>>>>> > > >> >> >> >>You >>>>>> > > >> >> >> >>can use yarn logs command to fetch all logs: >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>yarn logs -applicationId <app_id> >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" >>>>>> > > >><[email protected] >>>>>> > > >> > >>>>>> > > >> >> >> >>wrote: >>>>>> > > >> >> >> >> >>>>>> > > >> >> >> >>>Hi Luke / Shaofeng , >>>>>> > > >> >> >> >>> Can you please help me to check this >>>>>>issue. >>>>>> > > >> >> >> >>>Regards, >>>>>> > > >> >> >> >>>Santosh Akhilesh >>>>>> > > >> >> >> >>> >>>>>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >>>>>> > > >> >> >> >>>[email protected]> wrote: >>>>>> > > >> >> >> >>> >>>>>> > > >> >> >> >>>> Hi All , >>>>>> > > >> >> >> >>>> is it because of following error in map >>>>>>reduce >>>>>> job >>>>>> > ? >>>>>> > > >> >>what >>>>>> > > >> >> >> >>>>could >>>>>> > > >> >> >> >>>>be >>>>>> > > >> >> >> >>>> way to resolve this , a google search says that >>>>>>its >>>>>>issue >>>>>> > of >>>>>> > > >> >>Yarn >>>>>> > > >> >> >> >>>>class >>>>>> > > >> >> >> >>>> path , but I am not sure what it is ? >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job >>>>>>table=RETAIL.FACT_SALES >>>>>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Application application_1424791969399_0008 failed >>>>>>2 >>>>>>times >>>>>> > due >>>>>> > > >> >>to AM >>>>>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002 >>>>>>exited >>>>>> > > >>with >>>>>> > > >> >> >> >>>>exitCode: 1 >>>>>> > > >> >> >> >>>> For more detailed output, check application >>>>>>tracking >>>>>> page: >>>>>> > > >> >> >> >>>> >>>>>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then, >>>>>> > > >> >> >>click >>>>>> > > >> >> >> >>>>on >>>>>> > > >> >> >> >>>> links to logs of each attempt. >>>>>> > > >> >> >> >>>> Diagnostics: Exception from container-launch. >>>>>> > > >> >> >> >>>> Container id: >>>>>>container_1424791969399_0008_02_000001 >>>>>> > > >> >> >> >>>> Exit code: 1 >>>>>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1: >>>>>> > > >> >> >> >>>> at >>>>>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >>>>>> > > >> >> >> >>>> at >>>>>>org.apache.hadoop.util.Shell.run(Shell.java:455) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >>>>>> > > >> >> >> e(Shell.java:71 >>>>>> > > >> >> >> >>>>5 >>>>>> > > >> >> >> >>>>) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >>>>>> > > >> >> >> erExecutor.laun >>>>>> > > >> >> >> >>>>c >>>>>> > > >> >> >> >>>>h >>>>>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>>>> > > >> >> >> er.launcher.Con >>>>>> > > >> >> >> >>>>t >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>>>> > > >> >> >> er.launcher.Con >>>>>> > > >> >> >> >>>>t >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82) >>>>>> > > >> >> >> >>>> at >>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>>>> > > >> >> >> PoolExecutor.ja >>>>>> > > >> >> >> >>>>v >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>:1145) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>>>> > > >> >> >> dPoolExecutor.j >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>v >>>>>> > > >> >> >> >>>>a:615) >>>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>>>> > > >> >> >> >>>> Container exited with a non-zero exit code 1 >>>>>> > > >> >> >> >>>> Failing this attempt. Failing the application. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> ---------- Forwarded message ---------- >>>>>> > > >> >> >> >>>> From: Santoshakhilesh >>>>>><[email protected]> >>>>>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >>>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>>>option >>>>>>is >>>>>> > not >>>>>> > > >> >> >> >>>>responding >>>>>> > > >> >> >> >>>>on >>>>>> > > >> >> >> >>>> GUI >>>>>> > > >> >> >> >>>> To: "[email protected]" >>>>>> > > >> >> >><[email protected]> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> hi , >>>>>> > > >> >> >> >>>> please someone give me a hand to resolve this >>>>>>issue , >>>>>> > > >>thanks. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Regards, >>>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>>> > > >> >> >> >>>> Bangalore R&D >>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> www.huawei.com >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>-------------------------------------------------------- >>>>>> > > >> >> >> --------------- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>>> >>>>------------------------------------------------------------ >>>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>>confidential >>>>>> > > >>information >>>>>> > > >> >> >>from >>>>>> > > >> >> >> >>>> HUAWEI, which >>>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>>address >>>>>> is >>>>>> > > >> >>listed >>>>>> > > >> >> >> >>>>above. >>>>>> > > >> >> >> >>>> Any use of the >>>>>> > > >> >> >> >>>> information contained herein in any way >>>>>>(including, >>>>>>but >>>>>> not >>>>>> > > >> >>limited >>>>>> > > >> >> >> >>>>to, >>>>>> > > >> >> >> >>>> total or partial >>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>>persons >>>>>> > other >>>>>> > > >> >>than >>>>>> > > >> >> >>the >>>>>> > > >> >> >> >>>> intended >>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>>e-mail in >>>>>> > > >>error, >>>>>> > > >> >> >> >>>>please >>>>>> > > >> >> >> >>>> notify the sender by >>>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> ________________________________________ >>>>>> > > >> >> >> >>>> From: Santoshakhilesh >>>>>>[[email protected]] >>>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >>>>>> > > >> >> >> >>>> To: [email protected] >>>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>>>option >>>>>>is >>>>>> > not >>>>>> > > >> >> >> >>>>responding >>>>>> > > >> >> >> >>>>on >>>>>> > > >> >> >> >>>> GUI >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>>>cube >>>>>>I >>>>>> get >>>>>> > > >>an >>>>>> > > >> >> >> >>>>exception >>>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>>>correctly >>>>>> and >>>>>> > > >> >> >> >>>>HCatInputFormat >>>>>> > > >> >> >> >>>> this file is present in >>>>>>hive-hcatalog-core-0.14.0.jar >>>>>>. >>>>>> > > >>Please >>>>>> > > >> >>let >>>>>> > > >> >> >>me >>>>>> > > >> >> >> >>>>know >>>>>> > > >> >> >> >>>> what can I do to resolve this ? >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> -- This was path issue , now no more exception in >>>>>> > kylin.log >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> But saveing cube still fails with error. And still >>>>>>can't >>>>>> > add >>>>>> > > >> >> >>measures. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Error Message >>>>>> > > >> >> >> >>>> Failed to take action. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> In log I can find no exception. Following is the >>>>>>last >>>>>>log >>>>>> > in >>>>>> > > >> >> >>kylin.log >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24 >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >>>>>> > > >> >> >> .DefaultSchedul >>>>>> > > >> >> >> >>>>e >>>>>> > > >> >> >> >>>>r >>>>>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >>>>>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 >>>>>>ready, >>>>>>6 >>>>>> > others >>>>>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >>>>>> > > >> >> >> beController.de >>>>>> > > >> >> >> >>>>s >>>>>> > > >> >> >> >>>>e >>>>>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)] >>>>>> > > >> >> >> >>>> - Saving cube { >>>>>> > > >> >> >> >>>> "name": "", >>>>>> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >>>> "lookups": [], >>>>>> > > >> >> >> >>>> "filter_condition": "", >>>>>> > > >> >> >> >>>> "capacity": "SMALL", >>>>>> > > >> >> >> >>>> "partition_desc": { >>>>>> > > >> >> >> >>>> "partition_date_column": "", >>>>>> > > >> >> >> >>>> "partition_date_start": 0, >>>>>> > > >> >> >> >>>> "partition_type": "APPEND" >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> "last_modified": 0 >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> local access logs all with 200 , so seems ok. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] >>>>>>"GET >>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] >>>>>>"POST >>>>>> > > >> >> >>/kylin/api/cubes >>>>>> > > >> >> >> >>>> HTTP/1.1" 200 701 >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Regards, >>>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>>> > > >> >> >> >>>> Bangalore R&D >>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> www.huawei.com >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>-------------------------------------------------------- >>>>>> > > >> >> >> --------------- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>>> >>>>------------------------------------------------------------ >>>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>>confidential >>>>>> > > >>information >>>>>> > > >> >> >>from >>>>>> > > >> >> >> >>>> HUAWEI, which >>>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>>address >>>>>> is >>>>>> > > >> >>listed >>>>>> > > >> >> >> >>>>above. >>>>>> > > >> >> >> >>>> Any use of the >>>>>> > > >> >> >> >>>> information contained herein in any way >>>>>>(including, >>>>>>but >>>>>> not >>>>>> > > >> >>limited >>>>>> > > >> >> >> >>>>to, >>>>>> > > >> >> >> >>>> total or partial >>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>>persons >>>>>> > other >>>>>> > > >> >>than >>>>>> > > >> >> >>the >>>>>> > > >> >> >> >>>> intended >>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>>e-mail in >>>>>> > > >>error, >>>>>> > > >> >> >> >>>>please >>>>>> > > >> >> >> >>>> notify the sender by >>>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> ________________________________________ >>>>>> > > >> >> >> >>>> From: Santoshakhilesh >>>>>>[[email protected]] >>>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >>>>>> > > >> >> >> >>>> To: [email protected] >>>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>>>> > > >> >> >> >>>> Subject: Error while making cube & Measure option >>>>>>is >>>>>>not >>>>>> > > >> >> >>responding on >>>>>> > > >> >> >> >>>>GUI >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Hi All , >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> I am building a simple cube for test and using >>>>>>the >>>>>> > binary >>>>>> > > >> >>build >>>>>> > > >> >> >> >>>>0.7.1 >>>>>> > > >> >> >> >>>> . I have following hive tables with columns. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> fact_sales: >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> storeid int >>>>>> > > >> >> >> >>>> itemid int >>>>>> > > >> >> >> >>>> custid int >>>>>> > > >> >> >> >>>> qty int >>>>>> > > >> >> >> >>>> price double >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> dim_customer >>>>>> > > >> >> >> >>>> custid int >>>>>> > > >> >> >> >>>> name string >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> dim_item >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> itemid int >>>>>> > > >> >> >> >>>> category string >>>>>> > > >> >> >> >>>> brand string >>>>>> > > >> >> >> >>>> color string >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> dim_store >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> storeid int >>>>>> > > >> >> >> >>>> city string >>>>>> > > >> >> >> >>>> state string >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Please help me to answer following issues; >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> 1. When I go to measure section and click on >>>>>>measure >>>>>> > option , >>>>>> > > >> >> >>there is >>>>>> > > >> >> >> >>>>no >>>>>> > > >> >> >> >>>> response , I want add measure on qty and price >>>>>>with >>>>>>sum >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>>>cube >>>>>>I >>>>>> get >>>>>> > > >>an >>>>>> > > >> >> >> >>>>exception >>>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>>>correctly >>>>>> and >>>>>> > > >> >> >> >>>>HCatInputFormat >>>>>> > > >> >> >> >>>> this file is present in >>>>>>hive-hcatalog-core-0.14.0.jar >>>>>>. >>>>>> > > >>Please >>>>>> > > >> >>let >>>>>> > > >> >> >>me >>>>>> > > >> >> >> >>>>know >>>>>> > > >> >> >> >>>> what can I do to resolve this ? >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> 3. Also I have another question since this is a >>>>>>test >>>>>>and >>>>>> > > >>data is >>>>>> > > >> >> >>small >>>>>> > > >> >> >> >>>>I >>>>>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to >>>>>>skip >>>>>> > > >>partition >>>>>> > > >> >> >>stage >>>>>> > > >> >> >> >>>> while cube build ? >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Exception >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24 >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >>>>>> > > >> >> >> l.DefaultSchedu >>>>>> > > >> >> >> >>>>l >>>>>> > > >> >> >> >>>>e >>>>>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >>>>>> > > >> >> >> >>>> - ExecuteException >>>>>> job:c3532a6f-97ea-474a-b36a-218dd517cedb >>>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut >>>>>> > > >> >> >> e(AbstractExecu >>>>>> > > >> >> >> >>>>t >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>ble.java:102) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >>>>>> > > >> >> >> bRunner.run(Def >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>u >>>>>> > > >> >> >> >>>>ltScheduler.java:132) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>>>> > > >> >> >> PoolExecutor.ja >>>>>> > > >> >> >> >>>>v >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>:1145) >>>>>> > > >> >> >> >>>> at >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>>>> > > >> >> >> dPoolExecutor.j >>>>>> > > >> >> >> >>>>a >>>>>> > > >> >> >> >>>>v >>>>>> > > >> >> >> >>>>a:615) >>>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>>>> > > >> >> >> >>>> Caused by: >>>>>> org.apache.kylin.job.exception.ExecuteException: >>>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> The JSON is as below. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "name": "Retail_Cube", >>>>>> > > >> >> >> >>>> "description": "", >>>>>> > > >> >> >> >>>> "dimensions": [ >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >>>> "hierarchy": false, >>>>>> > > >> >> >> >>>> "derived": null, >>>>>> > > >> >> >> >>>> "column": [ >>>>>> > > >> >> >> >>>> "STOREID" >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "id": 1 >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >>>> "hierarchy": false, >>>>>> > > >> >> >> >>>> "derived": null, >>>>>> > > >> >> >> >>>> "column": [ >>>>>> > > >> >> >> >>>> "ITEMID" >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "id": 2 >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>>> > > >> >> >> >>>> "hierarchy": false, >>>>>> > > >> >> >> >>>> "derived": null, >>>>>> > > >> >> >> >>>> "column": [ >>>>>> > > >> >> >> >>>> "CUSTID" >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "id": 3 >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "measures": [ >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "id": 1, >>>>>> > > >> >> >> >>>> "name": "_COUNT_", >>>>>> > > >> >> >> >>>> "function": { >>>>>> > > >> >> >> >>>> "expression": "COUNT", >>>>>> > > >> >> >> >>>> "returntype": "bigint", >>>>>> > > >> >> >> >>>> "parameter": { >>>>>> > > >> >> >> >>>> "type": "constant", >>>>>> > > >> >> >> >>>> "value": 1 >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "rowkey": { >>>>>> > > >> >> >> >>>> "rowkey_columns": [ >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "column": "STOREID", >>>>>> > > >> >> >> >>>> "length": 0, >>>>>> > > >> >> >> >>>> "dictionary": "true", >>>>>> > > >> >> >> >>>> "mandatory": false >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "column": "ITEMID", >>>>>> > > >> >> >> >>>> "length": 0, >>>>>> > > >> >> >> >>>> "dictionary": "true", >>>>>> > > >> >> >> >>>> "mandatory": false >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "column": "CUSTID", >>>>>> > > >> >> >> >>>> "length": 0, >>>>>> > > >> >> >> >>>> "dictionary": "true", >>>>>> > > >> >> >> >>>> "mandatory": false >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> ], >>>>>> > > >> >> >> >>>> "aggregation_groups": [ >>>>>> > > >> >> >> >>>> [ >>>>>> > > >> >> >> >>>> "STOREID", >>>>>> > > >> >> >> >>>> "ITEMID", >>>>>> > > >> >> >> >>>> "CUSTID" >>>>>> > > >> >> >> >>>> ] >>>>>> > > >> >> >> >>>> ] >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> "notify_list": [], >>>>>> > > >> >> >> >>>> "capacity": "", >>>>>> > > >> >> >> >>>> "hbase_mapping": { >>>>>> > > >> >> >> >>>> "column_family": [ >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "name": "f1", >>>>>> > > >> >> >> >>>> "columns": [ >>>>>> > > >> >> >> >>>> { >>>>>> > > >> >> >> >>>> "qualifier": "m", >>>>>> > > >> >> >> >>>> "measure_refs": [ >>>>>> > > >> >> >> >>>> "_COUNT_" >>>>>> > > >> >> >> >>>> ] >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> ] >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> ] >>>>>> > > >> >> >> >>>> }, >>>>>> > > >> >> >> >>>> "project": "RetailProject", >>>>>> > > >> >> >> >>>> "model_name": "Retail_Cube" >>>>>> > > >> >> >> >>>> } >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> Regards, >>>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>>> > > >> >> >> >>>> Bangalore R&D >>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> www.huawei.com >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>>>>>>>>-------------------------------------------------------- >>>>>> > > >> >> >> --------------- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>- >>>>>> > > >> >> >> >>>>>> >>>>------------------------------------------------------------ >>>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>>confidential >>>>>> > > >>information >>>>>> > > >> >> >>from >>>>>> > > >> >> >> >>>> HUAWEI, which >>>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>>address >>>>>> is >>>>>> > > >> >>listed >>>>>> > > >> >> >> >>>>above. >>>>>> > > >> >> >> >>>> Any use of the >>>>>> > > >> >> >> >>>> information contained herein in any way >>>>>>(including, >>>>>>but >>>>>> not >>>>>> > > >> >>limited >>>>>> > > >> >> >> >>>>to, >>>>>> > > >> >> >> >>>> total or partial >>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>>persons >>>>>> > other >>>>>> > > >> >>than >>>>>> > > >> >> >>the >>>>>> > > >> >> >> >>>> intended >>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>>e-mail in >>>>>> > > >>error, >>>>>> > > >> >> >> >>>>please >>>>>> > > >> >> >> >>>> notify the sender by >>>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>>> -- >>>>>> > > >> >> >> >>>> Regards, >>>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>>> > > >> >> >> >>>> +91-0-9845482201 >>>>>> > > >> >> >> >>>> >>>>>> > > >> >> >> >>> >>>>>> > > >> >> >> >>> >>>>>> > > >> >> >> >>> >>>>>> > > >> >> >> >>>-- >>>>>> > > >> >> >> >>>Regards, >>>>>> > > >> >> >> >>>Santosh Akhilesh >>>>>> > > >> >> >> >>>+91-0-9845482201 >>>>>> > > >> >> >> >>>>>> > > >> >> >> >>>>>> > > >> >> >>>>>> > > >> >> >>>>>> > > >> >>>>>> > > >> >>>>>> > > >>>>>> > > >>>>>> > >>>>>> > >>>>>> > -- >>>>>> > Regards, >>>>>> > Santosh Akhilesh >>>>>> > +91-0-9845482201 >>>>>> > >>>>>> >>>>> >>>>> >>>>> >>>>>-- >>>>>Regards, >>>>>Santosh Akhilesh >>>>>+91-0-9845482201
