0.0.0.0:10020 isn’t a valid network address I think; please check the “mapreduce.jobhistory.address” in your mapred-site.xml; it should be something like:
<property> <name>mapreduce.jobhistory.address</name> <value>sandbox.hortonworks.com:10020</value> </property> On 2/27/15, 5:29 PM, "Santoshakhilesh" <[email protected]> wrote: >Hi Shaofeng , > No I have not found MR counter exception. I get following exception >frequently. I think this is related LogHistory server of hadoop. > >[QuartzScheduler_Worker-23]:[2015-02-27 >22:18:37,299][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCo >unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >failed on connection exception: java.net.ConnectException: Connection >refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused >com.kylinolap.job.exception.JobException: java.io.IOException: >java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >failed on connection exception: java.net.ConnectException: Connection >refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused > at >com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.j >ava:289) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOu >tput.java:162) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.ja >va:85) > at >com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86) > at org.quartz.core.JobRunShell.run(JobRunShell.java:202) > at >org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5 >73) >Caused by: java.io.IOException: java.net.ConnectException: Call From >linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >java.net.ConnectException: Connection refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Shi, Shaofeng [[email protected]] >Sent: Friday, February 27, 2015 2:47 PM >To: [email protected] >Cc: Kulbhushan Rana >Subject: Re: Cube Build Failed at Last Step//RE: Error while making cube >& Measure option is not responding on GUI > >Did you figure out the exception of "No enum constant >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES” ? Is it still be >thrown in the logs? In the last step, Kylin need to parse the MR counters >to update cube size; Please refer to >https://issues.apache.org/jira/browse/MAPREDUCE-5831 for that error. > >On 2/27/15, 5:04 PM, "Santoshakhilesh" <[email protected]> >wrote: > >>Hi Shaofeng , >> Cube building is failed at last step while loading Hfile to >>Hbase with exception "Can't get cube segment size. >>". What could be reason ? >> >>parameter : -input >>/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >>-htablename KYLIN_K27LDMX63W -cubename RetailCube >> >>Log: >> >>Start to execute command: >> -input /tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >>-htablename KYLIN_K27LDMX63W -cubename RetailCube >>Command execute return code 0 >>Failed with Exception:java.lang.RuntimeException: Can't get cube segment >>size. >> at >>com.kylinolap.job.flow.JobFlowListener.updateCubeSegmentInfoOnSucceed(Job >>F >>lowListener.java:247) >> at >>com.kylinolap.job.flow.JobFlowListener.jobWasExecuted(JobFlowListener.jav >>a >>:101) >> at >>org.quartz.core.QuartzScheduler.notifyJobListenersWasExecuted(QuartzSched >>u >>ler.java:1985) >> at >>org.quartz.core.JobRunShell.notifyJobListenersComplete(JobRunShell.java:3 >>4 >>0) >> at org.quartz.core.JobRunShell.run(JobRunShell.java:224) >> at >>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java: >>5 >>73) >> >>I have checked in hbase shell and following are the tables in hbase; >>hbase(main):001:0> list >>TABLE >> >>KYLIN_K27LDMX63W >>kylin_metadata_qa >>kylin_metadata_qa_acl >>kylin_metadata_qa_cube >>kylin_metadata_qa_dict >>kylin_metadata_qa_invertedindex >>kylin_metadata_qa_job >>kylin_metadata_qa_job_output >>kylin_metadata_qa_proj >>kylin_metadata_qa_table_snapshot >>kylin_metadata_qa_user >>11 row(s) in 0.8990 seconds >> >> >>Regards, >>Santosh Akhilesh >>Bangalore R&D >>HUAWEI TECHNOLOGIES CO.,LTD. >> >>www.huawei.com >>------------------------------------------------------------------------- >>- >>----------------------------------------------------------- >>This e-mail and its attachments contain confidential information from >>HUAWEI, which >>is intended only for the person or entity whose address is listed above. >>Any use of the >>information contained herein in any way (including, but not limited to, >>total or partial >>disclosure, reproduction, or dissemination) by persons other than the >>intended >>recipient(s) is prohibited. If you receive this e-mail in error, please >>notify the sender by >>phone or email immediately and delete it! >> >>________________________________________ >>From: Santoshakhilesh >>Sent: Friday, February 27, 2015 2:15 PM >>To: [email protected] >>Subject: RE: Error while making cube & Measure option is not responding >>on GUI >> >>I have manually copied the jar to /tmp/kylin , now satge 2 is done , >>thanks. >> >>Regards, >>Santosh Akhilesh >>Bangalore R&D >>HUAWEI TECHNOLOGIES CO.,LTD. >> >>www.huawei.com >>------------------------------------------------------------------------- >>- >>----------------------------------------------------------- >>This e-mail and its attachments contain confidential information from >>HUAWEI, which >>is intended only for the person or entity whose address is listed above. >>Any use of the >>information contained herein in any way (including, but not limited to, >>total or partial >>disclosure, reproduction, or dissemination) by persons other than the >>intended >>recipient(s) is prohibited. If you receive this e-mail in error, please >>notify the sender by >>phone or email immediately and delete it! >> >>________________________________________ >>From: Shi, Shaofeng [[email protected]] >>Sent: Friday, February 27, 2015 1:00 PM >>To: [email protected] >>Cc: Kulbhushan Rana >>Subject: Re: Error while making cube & Measure option is not responding >>on GUI >> >>In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we >>renamed the package to “org.apache.kylin.xxx”; When you downgrade to 0.6, >>did you also replace the jar location with 0.6 ones in kylin.properties? >> >>On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]> >>wrote: >> >>>Hi Shaofeng , >>> I have added my fact and dimension tables under default >>>database >>>of hive. >>> Now stage 1 of Cube Build is ok. And there is failure at step2. >>> The map reduce job for the finding distinct columns of fact >>>table >>>is error. Yarn log is as below. >>> Strangely this is class not found error. I have checked the >>>Kylin.properties and the jar is already set as below. >>>kylin. log has one exception connecting to linux/10.19.93.68 to >>>0.0.0.0:10020 >>> Please help me to give a clue , I am also trying to check meanwhile >>> >>>Thanks. >>>kylin property >>># Temp folder in hdfs >>>kylin.hdfs.working.dir=/tmp >>># Path to the local(relative to job engine) job jar, job engine will use >>>this jar >>>kylin.job.jar=/tmp/kylin/kylin-job-latest.jar >>> >>>Map Reduce error >>>---------------------------- >>>2015-02-27 20:24:25,262 FATAL [main] org.apache.hadoop.mapred.YarnChild: >>>Error running child : java.lang.NoClassDefFoundError: >>>com/kylinolap/common/mr/KylinMapper >>> at java.lang.ClassLoader.defineClass1(Native Method) >>> at java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>> at >>>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>> at java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >>> at java.lang.Class.forName0(Native Method) >>> at java.lang.Class.forName(Class.java:274) >>> at >>>org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration. >>>j >>>a >>>va:2013) >>> >>>Kylin.log >>>QuartzScheduler_Worker-20]:[2015-02-27 >>>20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFetc >>>h >>>e >>>r.java:60)] - 0 pending jobs >>>[QuartzScheduler_Worker-19]:[2015-02-27 >>>20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJob >>>C >>>o >>>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>>failed on connection exception: java.net.ConnectException: Connection >>>refused; For more details see: >>>http://wiki.apache.org/hadoop/ConnectionRefused >>>com.kylinolap.job.exception.JobException: java.io.IOException: >>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>>failed on connection exception: java.net.ConnectException: Connection >>>refused; For more details see: >>>http://wiki.apache.org/hadoop/ConnectionRefused >>> at >>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob >>>. >>>j >>>ava:289) >>> at >>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmd >>>O >>>u >>>tput.java:162) >>> at >>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput. >>>j >>>a >>>va:85) >>> at >>>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86 >>>) >>> at org.quartz.core.JobRunShell.run(JobRunShell.java:202) >>> at >>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java >>>: >>>5 >>>73) >>>Caused by: java.io.IOException: java.net.ConnectException: Call From >>>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >>>java.net.ConnectException: Connection refused; For more details see: >>>http://wiki.apache.org/hadoop/ConnectionRefused >>> at >>>org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDeleg >>>a >>>t >>>e.java:331) >>> at >>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServ >>>i >>>c >>>eDelegate.java:368) >>> at >>>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) >>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:415) >>> at >>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformatio >>>n >>>. >>>java:1491) >>> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>> at >>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob >>>. >>>j >>>ava:287) >>> ... 5 more >>> >>>Regards, >>>Santosh Akhilesh >>>Bangalore R&D >>>HUAWEI TECHNOLOGIES CO.,LTD. >>> >>>www.huawei.com >>>------------------------------------------------------------------------ >>>- >>>- >>>----------------------------------------------------------- >>>This e-mail and its attachments contain confidential information from >>>HUAWEI, which >>>is intended only for the person or entity whose address is listed above. >>>Any use of the >>>information contained herein in any way (including, but not limited to, >>>total or partial >>>disclosure, reproduction, or dissemination) by persons other than the >>>intended >>>recipient(s) is prohibited. If you receive this e-mail in error, please >>>notify the sender by >>>phone or email immediately and delete it! >>> >>>________________________________________ >>>From: Shi, Shaofeng [[email protected]] >>>Sent: Friday, February 27, 2015 8:01 AM >>>To: [email protected] >>>Subject: Re: Error while making cube & Measure option is not responding >>>on GUI >>> >>>In 0.6.x it only support tables in default database, this is a >>>limitation; >>>The support for non-default tables will be released in 0.7; >>> >>>To bypass this issue for now, please copy the table to default database >>>as >>>a workaround; >>> >>>On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]> >>>wrote: >>> >>>>@Jason >>>>thanks , but now as suggested by Saofeng I m not using the inverted >>>>index >>>>brach as its not stable. >>>>I have switched back to 0.6 branch , in this branch yesterday night I >>>>could >>>>crete the cube successfully but there is issue while building it. I >>>>feel >>>>that at step 1 of cube build while creating flat table when command is >>>>issued to hive if the tables are not under default datbase flat table >>>>creation is failed and cube build fails. my fact and dimension tables >>>>are >>>>under a database called retail. >>>> >>>>@Saofeng - Can you please confirm this behavior ? Do I need to create >>>>the >>>>hive tables under default database? >>>> >>>>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]> >>>>wrote: >>>> >>>>> @Santoshakhilesh >>>>> >>>>> 1. When I go to measure section and click on measure option , there >>>>>is >>>>>no >>>>> response , I want add measure on qty and price with sum >>>>> --bug fixed on inverted-index branch >>>>> >>>>> >>>>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh < >>>>> [email protected] >>>>> > wrote: >>>>> >>>>> > Hi Shaofeng , >>>>> > I have build the 0.6 version and now able to create the cube >>>>> > successfully. >>>>> > While building the cube , it fails at step1 with following >>>>>error. >>>>> > Table not found 'DIM_ITEM' >>>>> > the table exists , but its under retail data base and not >>>>>under >>>>> > default database. >>>>> > does kylin require hive taables to be under default database ? >>>>>I >>>>>see >>>>> > the flat table being created under default database. >>>>> > >>>>> > Logging initialized using configuration in >>>>> > >>>>> > >>>>> >>>>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-comm >>>>>o >>>>>n >>>>>- >>>>>1.0.0.jar!/hive-log4j.properties >>>>> > SLF4J: Class path contains multiple SLF4J bindings. >>>>> > SLF4J: Found binding in >>>>> > >>>>> > >>>>> >>>>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/comm >>>>>o >>>>>n >>>>>/ >>>>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>> > SLF4J: Found binding in >>>>> > >>>>> > >>>>> >>>>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jdb >>>>>c >>>>>- >>>>>1 >>>>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >>>>> > explanation. >>>>> > SLF4J: Actual binding is of type >>>>>[org.slf4j.impl.Log4jLoggerFactory] >>>>> > OK >>>>> > Time taken: 0.964 seconds >>>>> > OK >>>>> > Time taken: 0.948 seconds >>>>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found >>>>> > 'DIM_ITEM' >>>>> > >>>>> > >>>>> > >>>>> > Command is as below. >>>>> > >>>>> > hive -e "DROP TABLE IF EXISTS >>>>> > >>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4 >>>>>4 >>>>>; >>>>> > CREATE EXTERNAL TABLE IF NOT EXISTS >>>>> > >>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4 >>>>>4 >>>>> > ( >>>>> > STOREID int >>>>> > ,ITEMID int >>>>> > ,CUSTID int >>>>> > ,QTY int >>>>> > ,AMOUNT double >>>>> > ) >>>>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' >>>>> > STORED AS SEQUENCEFILE >>>>> > LOCATION >>>>> > >>>>> >>>>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_te >>>>>s >>>>>t >>>>>_ >>>>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44'; >>>>> > SET hive.exec.compress.output=true; >>>>> > SET hive.auto.convert.join.noconditionaltask = true; >>>>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000; >>>>> > INSERT OVERWRITE TABLE >>>>> > >>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4 >>>>>4 >>>>> > SELECT >>>>> > FACT_SALES.STOREID >>>>> > ,FACT_SALES.ITEMID >>>>> > ,FACT_SALES.CUSTID >>>>> > ,FACT_SALES.QTY >>>>> > ,FACT_SALES.AMOUNT >>>>> > FROM FACT_SALES >>>>> > INNER JOIN DIM_STORE >>>>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID >>>>> > INNER JOIN DIM_ITEM >>>>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID >>>>> > INNER JOIN DIM_CUSTOMER >>>>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID >>>>> > ; >>>>> > " >>>>> > >>>>> > >>>>> > >>>>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]> >>>>>wrote: >>>>> > >>>>> > > The 0.7.1 is test version, its package contains the “snapshot” >>>>>suffix; >>>>> we >>>>> > > will upload a new package there; Luke will also add a message >>>>>there >>>>>to >>>>> > > avoid this confusion; >>>>> > > >>>>> > > Regarding the problem that you encountered, could you please open >>>>>a >>>>> JIRA >>>>> > > ticket for tracking? Here is link of Apache JIRA: >>>>> > > >>>>> > > https://issues.apache.org/jira/secure/Dashboard.jspa >>>>> > > >>>>> > > >>>>> > > Thanks for the feedback! >>>>> > > >>>>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh" >>>>><[email protected]> >>>>> > > wrote: >>>>> > > >>>>> > > >Actually I see this being published on kylin webpage. >>>>> > > >http://kylin.incubator.apache.org/download/ >>>>> > > >I am using 0.7.1 inverted index branch binary distribution. >>>>> > > >If this is not stable please give me the link of stable branch I >>>>>would >>>>> > try >>>>> > > >building and testing tonight. >>>>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]> >>>>> wrote: >>>>> > > > >>>>> > > >> Hi Santosh, it is not recommended to use the dev code branch >>>>> > (actually I >>>>> > > >> don’t know how you get the v0.7.x build and what’s the detail >>>>> version >>>>> > of >>>>> > > >> that; each day we submit many changes to that); >>>>> > > >> >>>>> > > >> The options are 1) switch back to latest release v0.6.5; or 2) >>>>>wait >>>>> > for >>>>> > > >> the formal release of 0.7, that should be in March; Otherwise, >>>>>we >>>>> > > >>couldn’t >>>>> > > >> ensure there is no new problems come out in your next steps; >>>>> > > >> >>>>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh" >>>>><[email protected]> >>>>> > > >>wrote: >>>>> > > >> >>>>> > > >> >Hi Shaofeng >>>>> > > >> >So what do you suggest , how should I proceed further with >>>>>this >>>>> > > >>release? >>>>> > > >> >Will there be a patch? Any alternate way I can create cube? >>>>> > > >> >Please suggest. >>>>> > > >> >Regards >>>>> > > >> >Santosh >>>>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng >>>>><[email protected]> >>>>> > > wrote: >>>>> > > >> > >>>>> > > >> >> Hi Santosh, >>>>> > > >> >> >>>>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we >>>>>have >>>>> > > >> >>metadata >>>>> > > >> >> structure change; While, the web UI (cube wizard) for this >>>>>change >>>>> > > >>hasn’t >>>>> > > >> >> been stabilized; So it is not strange that you got trouble >>>>>when >>>>> > > >>saving >>>>> > > >> >>the >>>>> > > >> >> cube; >>>>> > > >> >> >>>>> > > >> >> @Jason, any idea about the JS error? >>>>> > > >> >> >>>>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" < >>>>> [email protected] >>>>> > > >>>>> > > >> >>wrote: >>>>> > > >> >> >>>>> > > >> >> >Hi Shaofeng, >>>>> > > >> >> > >>>>> > > >> >> >I am using the binary distribution 0.7.1. I have not been >>>>>able >>>>> to >>>>> > > >>save >>>>> > > >> >> >cube >>>>> > > >> >> >even once. I have tried creating new project and from >>>>>local >>>>> > machine >>>>> > > >>and >>>>> > > >> >> >server machine. But I am always stuck with this error. I >>>>>am >>>>> never >>>>> > > >> >>allowed >>>>> > > >> >> >to add measures and never been able to save the cube. I >>>>>also >>>>>see >>>>> > the >>>>> > > >> >> >kylin.log and it always tries to save cube with append >>>>>mode. >>>>>One >>>>> > > >>thing >>>>> > > >> >>I >>>>> > > >> >> >need to tell that at partition stage since I don't have a >>>>>big >>>>> fact >>>>> > > >> >>table >>>>> > > >> >> >now I have not partititioned the fact table and I skip >>>>>this >>>>> step. >>>>> > > >>Does >>>>> > > >> >> >this >>>>> > > >> >> >have affect in saving the cube. Is this because some >>>>>metadata is >>>>> > > >> >>available >>>>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop >>>>>2.6.6. >>>>> > Yes >>>>> > > >> >>kylin >>>>> > > >> >> >propert I have not added the jar. I will add them and >>>>>check. >>>>>But >>>>> > > >>cube >>>>> > > >> >> >creation failure is really puzzling me. I could see no >>>>>error >>>>> logs >>>>> > in >>>>> > > >> >> >kylin.log. >>>>> > > >> >> >Regards >>>>> > > >> >> >Santosh >>>>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng >>>>><[email protected] >>>>> > >>>>> > > >> wrote: >>>>> > > >> >> > >>>>> > > >> >> >> Which version or code branch are you using? I assume >>>>>you’re >>>>> > using >>>>> > > >>the >>>>> > > >> >> >> stable version from master; Seems you’re trying to edit >>>>>an >>>>> > > >>existing >>>>> > > >> >>cube >>>>> > > >> >> >> to add new measurement, try refresh your browser's >>>>>cache; >>>>>if >>>>> it >>>>> > > >>still >>>>> > > >> >> >> couldn’t be saved, try to create a new cube; >>>>> > > >> >> >> >>>>> > > >> >> >> The two error traces in tomcat need be taken care: >>>>> > > >> >> >> >>>>> > > >> >> >> 1) java.lang.NoClassDefFoundError: >>>>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper >>>>> > > >> >> >> Please check kylin.properties file, making sure >>>>>the >>>>> > > >> >> >>“kylin.job.jar” >>>>> > > >> >> >> points to a right jar file; It will be loaded in >>>>>Map-reduce; >>>>> > > >> >> >> >>>>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant >>>>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>>> > > >> >> >> This indicates your hadoop version might be old; >>>>> Please >>>>> > > >>check >>>>> > > >> >> >>and >>>>> > > >> >> >> ensure >>>>> > > >> >> >> hadoop version is 2.2 or above. >>>>> > > >> >> >> >>>>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" >>>>> > > >><[email protected]> >>>>> > > >> >> >> wrote: >>>>> > > >> >> >> >>>>> > > >> >> >> >Hi Shaofeng , >>>>> > > >> >> >> > >>>>> > > >> >> >> > I am using chrome , When I click on button to add >>>>> measures , >>>>> > > >> >> >>following >>>>> > > >> >> >> >is error on chrome console. When I try to save the cube >>>>>there >>>>> > is >>>>> > > >>no >>>>> > > >> >> >>error >>>>> > > >> >> >> >in console. I just get a error dialog saying failed to >>>>>take >>>>> > > >>action >>>>> > > >> >>and >>>>> > > >> >> >> >gives me the JSON cube schema. >>>>> > > >> >> >> > >>>>> > > >> >> >> >Error on chrome debug console is as below; >>>>> > > >> >> >> > >>>>> > > >> >> >> > ReferenceError: CubeDescModel is not defined >>>>> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >>>>> > > >> >> >> > at scripts.min.0.js:180 >>>>> > > >> >> >> > at scripts.min.0.js:197 >>>>> > > >> >> >> > at h.$eval (scripts.min.0.js:119) >>>>> > > >> >> >> > at h.$apply (scripts.min.0.js:119) >>>>> > > >> >> >> > at HTMLButtonElement.<anonymous> >>>>>(scripts.min.0.js:197) >>>>> > > >> >> >> > at HTMLButtonElement.m.event.dispatch >>>>> (scripts.min.0.js:3) >>>>> > > >> >> >> > at HTMLButtonElement.r.handle >>>>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >>>>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous >>>>> > > >> >>function)scripts.min.0.js:119 >>>>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous >>>>> > > >>function)scripts.min.0.js:3 >>>>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle >>>>> > > >> >> >> > >>>>> > > >> >> >> > About the hive table import , I got pass the run >>>>>shell >>>>> > command >>>>> > > >> >> >> >exception but it still fails the hadoop log is; >>>>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main] >>>>>org.apache.hadoop.mapred. >>>>> > > >> >> YarnChild: >>>>> > > >> >> >> >mapreduce.cluster.local.dir for child: >>>>> > > >> >> >> >>>>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >>>>> > > >> >> >> cation_14249530 >>>>> > > >> >> >> >91340_0002 >>>>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main] >>>>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation: >>>>>session.id >>>>> > is >>>>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id >>>>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main] >>>>> > > >>org.apache.hadoop.mapred.Task: >>>>> > > >> >> >>Using >>>>> > > >> >> >> >ResourceCalculatorProcessTree : [ ] >>>>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main] >>>>> > > >> >> >>org.apache.hadoop.mapred.YarnChild: >>>>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError: >>>>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper >>>>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method) >>>>> > > >> >> >> > at >>>>>java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >java.security.SecureClassLoader.defineClass( >>>>> > > >> >> SecureClassLoader.java:142) >>>>> > > >> >> >> > at >>>>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>>>> > > >> >> >> > at >>>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>>>> > > >> >> >> > at >>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>>> > > >> >> >> > at >>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>>> Method) >>>>> > > >> >> >> > at >>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>>> > > >> >> >> > at >>>>>java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>>> > > >> >> >> > at >>>>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>>> > > >> >> >> > >>>>> > > >> >> >> >tomcat logs: >>>>> > > >> >> >> >usage: HiveColumnCardinalityJob >>>>> > > >> >> >> > -output <path> Output path >>>>> > > >> >> >> > -table <table name> The hive table name >>>>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26 >>>>> > > >> >> >> >>>>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >>>>> > > >> >> >> lExecutable.doW >>>>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute >>>>> > > >> >> >> >>>>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >>>>> > > >> >> >> >name=null, state=RUNNING} >>>>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant >>>>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>> > > >>>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf( >>>>> > > >> >> >> Framewo >>>>> > > >> >> >> >rkCounterGroup.java:148) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >>>>> > > >> >> >> findCounter(Fra >>>>> > > >> >> >> >meworkCounterGroup.java:182) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >>>>> > > >> >> >> ounter(Abstract >>>>> > > >> >> >> >Counters.java:154) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >>>>> > > >> >> >> erter.java:240) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >>>>> > > >> >> >> rs(ClientServic >>>>> > > >> >> >> >eDelegate.java:370) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters( >>>>> > > >> >> YARNRunner.java:511) >>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>>> Method) >>>>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >>>>> > > >> >> >> oupInformation. >>>>> > > >> >> >> >java:1491) >>>>> > > >> >> >> > at >>>>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>>>> > > >> >> >> > at >>>>> > > >> >>>>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >>>>> > > >> >> >> > at >>>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job. >>>>> > > >> java:1289) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >>>>> > > >> >> >> tion(AbstractHa >>>>> > > >> >> >> >doopJob.java:134) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>> > > >> >>>>> > >>>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.ru >>>>>>>>n >>>>>>>>( >>>>> > > >> >> >> HiveC >>>>> > > >> >> >> >olumnCardinalityJob.java:114) >>>>> > > >> >> >> > at >>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>>> > > >> >> >> > at >>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >>>>> > > >> >> >> doopShellExecut >>>>> > > >> >> >> >able.java:62) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>>> > > >> >> >> bstractExecutab >>>>> > > >> >> >> >le.java:99) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW >>>>> > > >> >> >> ork(DefaultChai >>>>> > > >> >> >> >nedExecutable.java:50) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>>> > > >> >> >> bstractExecutab >>>>> > > >> >> >> >le.java:99) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >>>>> > > >> >> >> nner.run(Defaul >>>>> > > >> >> >> >tScheduler.java:132) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >>>>> > > >> >> >> lExecutor.java: >>>>> > > >> >> >> >1145) >>>>> > > >> >> >> > at >>>>> > > >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >>>>> > > >> >> >> olExecutor.java >>>>> > > >> >> >> >:615) >>>>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745) >>>>> > > >> >> >> > >>>>> > > >> >> >> >Regards, >>>>> > > >> >> >> >Santosh Akhilesh >>>>> > > >> >> >> >Bangalore R&D >>>>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD. >>>>> > > >> >> >> > >>>>> > > >> >> >> >www.huawei.com >>>>> > > >> >> >> >>>>>>----------------------------------------------------------- >>>>> > > >> >> >> --------------- >>>>> > > >> >> >> >>>>>>----------------------------------------------------------- >>>>> > > >> >> >> >This e-mail and its attachments contain confidential >>>>> > information >>>>> > > >> >>from >>>>> > > >> >> >> >HUAWEI, which >>>>> > > >> >> >> >is intended only for the person or entity whose address >>>>>is >>>>> > listed >>>>> > > >> >> >>above. >>>>> > > >> >> >> >Any use of the >>>>> > > >> >> >> >information contained herein in any way (including, but >>>>>not >>>>> > > >>limited >>>>> > > >> >>to, >>>>> > > >> >> >> >total or partial >>>>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons >>>>>other >>>>> > than >>>>> > > >> >>the >>>>> > > >> >> >> >intended >>>>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail >>>>>in >>>>> > error, >>>>> > > >> >>please >>>>> > > >> >> >> >notify the sender by >>>>> > > >> >> >> >phone or email immediately and delete it! >>>>> > > >> >> >> > >>>>> > > >> >> >> >________________________________________ >>>>> > > >> >> >> >From: Shi, Shaofeng [[email protected]] >>>>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM >>>>> > > >> >> >> >To: [email protected] >>>>> > > >> >> >> >Cc: Kulbhushan Rana >>>>> > > >> >> >> >Subject: Re: Error while making cube & Measure option >>>>>is >>>>>not >>>>> > > >> >>responding >>>>> > > >> >> >> >on GUI >>>>> > > >> >> >> > >>>>> > > >> >> >> >Hi Santosh, hive table importing issue should not >>>>>impact >>>>>on >>>>> > cube >>>>> > > >> >> >>saving. >>>>> > > >> >> >> > >>>>> > > >> >> >> >If you couldn’t save the cube, firstly please check >>>>>whether >>>>> > > >>there is >>>>> > > >> >> >>error >>>>> > > >> >> >> >in the tomcat’s log; If not please check your web >>>>>browser; We >>>>> > > >> >>suggest >>>>> > > >> >> >>use >>>>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS >>>>>console >>>>> > > >>(press >>>>> > > >> >> >>F12) >>>>> > > >> >> >> >and then operate web UI, check whether there is any >>>>>error >>>>> > > >>reported >>>>> > > >> >>in >>>>> > > >> >> >> >browser. >>>>> > > >> >> >> > >>>>> > > >> >> >> > >>>>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" >>>>> > > >><[email protected] >>>>> > > >> > >>>>> > > >> >> >> >wrote: >>>>> > > >> >> >> > >>>>> > > >> >> >> >>Hi Shaofeng , >>>>> > > >> >> >> >> Thanks for replying. >>>>> > > >> >> >> >> Yes I am checking the yarn exception, But I find >>>>>that >>>>> this >>>>> > > >>error >>>>> > > >> >> >>comes >>>>> > > >> >> >> >>while importing the hive table to kylin. >>>>> > > >> >> >> >> Even if this error comes , hive tables is exported >>>>> > > >>successfully >>>>> > > >> >>in >>>>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ? >>>>> > > >> >> >> >> >>>>> > > >> >> >> >> Next step when I go on creating the cube for >>>>>following >>>>> > > >>schema I >>>>> > > >> >> >>get >>>>> > > >> >> >> >>error at last step while saving and I am unable to >>>>>add >>>>>any >>>>> > > >> >>measures , >>>>> > > >> >> >> >>clicking on measure option just dont pop up any >>>>>dialog, >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>I am using star schema with fact_sales as fact table >>>>>and >>>>> dim_* >>>>> > > >>as >>>>> > > >> >> >> >>dimension tables. >>>>> > > >> >> >> >> >>>>> > > >> >> >> >> fact_sales: >>>>> > > >> >> >> >> storeid int >>>>> > > >> >> >> >> itemid int >>>>> > > >> >> >> >> custid int >>>>> > > >> >> >> >> qty int >>>>> > > >> >> >> >> price double >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>dim_customer >>>>> > > >> >> >> >> custid int >>>>> > > >> >> >> >> name string >>>>> > > >> >> >> >> >>>>> > > >> >> >> >> dim_item >>>>> > > >> >> >> >> itemid int >>>>> > > >> >> >> >> category string >>>>> > > >> >> >> >> brand string >>>>> > > >> >> >> >> color string >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>dim_store >>>>> > > >> >> >> >> storeid int >>>>> > > >> >> >> >> city string >>>>> > > >> >> >> >> state string >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>JSON is as below. >>>>> > > >> >> >> >> The JSON is as below. >>>>> > > >> >> >> >> >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "name": "Retail_Cube", >>>>> > > >> >> >> >> "description": "", >>>>> > > >> >> >> >> "dimensions": [ >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID", >>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >> "hierarchy": false, >>>>> > > >> >> >> >> "derived": null, >>>>> > > >> >> >> >> "column": [ >>>>> > > >> >> >> >> "STOREID" >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "id": 1 >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID", >>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >> "hierarchy": false, >>>>> > > >> >> >> >> "derived": null, >>>>> > > >> >> >> >> "column": [ >>>>> > > >> >> >> >> "ITEMID" >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "id": 2 >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID", >>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >> "hierarchy": false, >>>>> > > >> >> >> >> "derived": null, >>>>> > > >> >> >> >> "column": [ >>>>> > > >> >> >> >> "CUSTID" >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "id": 3 >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "measures": [ >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "id": 1, >>>>> > > >> >> >> >> "name": "_COUNT_", >>>>> > > >> >> >> >> "function": { >>>>> > > >> >> >> >> "expression": "COUNT", >>>>> > > >> >> >> >> "returntype": "bigint", >>>>> > > >> >> >> >> "parameter": { >>>>> > > >> >> >> >> "type": "constant", >>>>> > > >> >> >> >> "value": 1 >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "rowkey": { >>>>> > > >> >> >> >> "rowkey_columns": [ >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "column": "STOREID", >>>>> > > >> >> >> >> "length": 0, >>>>> > > >> >> >> >> "dictionary": "true", >>>>> > > >> >> >> >> "mandatory": false >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "column": "ITEMID", >>>>> > > >> >> >> >> "length": 0, >>>>> > > >> >> >> >> "dictionary": "true", >>>>> > > >> >> >> >> "mandatory": false >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "column": "CUSTID", >>>>> > > >> >> >> >> "length": 0, >>>>> > > >> >> >> >> "dictionary": "true", >>>>> > > >> >> >> >> "mandatory": false >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> ], >>>>> > > >> >> >> >> "aggregation_groups": [ >>>>> > > >> >> >> >> [ >>>>> > > >> >> >> >> "STOREID", >>>>> > > >> >> >> >> "ITEMID", >>>>> > > >> >> >> >> "CUSTID" >>>>> > > >> >> >> >> ] >>>>> > > >> >> >> >> ] >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> "notify_list": [], >>>>> > > >> >> >> >> "capacity": "", >>>>> > > >> >> >> >> "hbase_mapping": { >>>>> > > >> >> >> >> "column_family": [ >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "name": "f1", >>>>> > > >> >> >> >> "columns": [ >>>>> > > >> >> >> >> { >>>>> > > >> >> >> >> "qualifier": "m", >>>>> > > >> >> >> >> "measure_refs": [ >>>>> > > >> >> >> >> "_COUNT_" >>>>> > > >> >> >> >> ] >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> ] >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> ] >>>>> > > >> >> >> >> }, >>>>> > > >> >> >> >> "project": "RetailProject", >>>>> > > >> >> >> >> "model_name": "Retail_Cube" >>>>> > > >> >> >> >> } >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>Regards, >>>>> > > >> >> >> >>Santosh Akhilesh >>>>> > > >> >> >> >>Bangalore R&D >>>>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD. >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>www.huawei.com >>>>> > > >> >> >> >>>>>>>---------------------------------------------------------- >>>>> > > >> >> >> --------------- >>>>> > > >> >> >> >>- >>>>> > > >> >> >> >>>>>>>----------------------------------------------------------- >>>>> > > >> >> >> >>This e-mail and its attachments contain confidential >>>>> > information >>>>> > > >> >>from >>>>> > > >> >> >> >>HUAWEI, which >>>>> > > >> >> >> >>is intended only for the person or entity whose >>>>>address >>>>>is >>>>> > > >>listed >>>>> > > >> >> >>above. >>>>> > > >> >> >> >>Any use of the >>>>> > > >> >> >> >>information contained herein in any way (including, >>>>>but >>>>>not >>>>> > > >>limited >>>>> > > >> >> >>to, >>>>> > > >> >> >> >>total or partial >>>>> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons >>>>>other >>>>> > > >>than >>>>> > > >> >>the >>>>> > > >> >> >> >>intended >>>>> > > >> >> >> >>recipient(s) is prohibited. If you receive this e-mail >>>>>in >>>>> > error, >>>>> > > >> >> >>please >>>>> > > >> >> >> >>notify the sender by >>>>> > > >> >> >> >>phone or email immediately and delete it! >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>________________________________________ >>>>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]] >>>>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM >>>>> > > >> >> >> >>To: [email protected] >>>>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option >>>>>is >>>>>not >>>>> > > >> >> >>responding >>>>> > > >> >> >> >>on GUI >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>Hi Santosh, >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>It looks like hadoop failed to execute some shell >>>>>command in >>>>> > the >>>>> > > >> >> >> >>container; You need dive into hadoop to see what¹s the >>>>> > concrete >>>>> > > >> >>error. >>>>> > > >> >> >> >>You >>>>> > > >> >> >> >>can use yarn logs command to fetch all logs: >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>yarn logs -applicationId <app_id> >>>>> > > >> >> >> >> >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" >>>>> > > >><[email protected] >>>>> > > >> > >>>>> > > >> >> >> >>wrote: >>>>> > > >> >> >> >> >>>>> > > >> >> >> >>>Hi Luke / Shaofeng , >>>>> > > >> >> >> >>> Can you please help me to check this >>>>>issue. >>>>> > > >> >> >> >>>Regards, >>>>> > > >> >> >> >>>Santosh Akhilesh >>>>> > > >> >> >> >>> >>>>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >>>>> > > >> >> >> >>>[email protected]> wrote: >>>>> > > >> >> >> >>> >>>>> > > >> >> >> >>>> Hi All , >>>>> > > >> >> >> >>>> is it because of following error in map >>>>>reduce >>>>> job >>>>> > ? >>>>> > > >> >>what >>>>> > > >> >> >> >>>>could >>>>> > > >> >> >> >>>>be >>>>> > > >> >> >> >>>> way to resolve this , a google search says that its >>>>>issue >>>>> > of >>>>> > > >> >>Yarn >>>>> > > >> >> >> >>>>class >>>>> > > >> >> >> >>>> path , but I am not sure what it is ? >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job >>>>>table=RETAIL.FACT_SALES >>>>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2 >>>>>times >>>>> > due >>>>> > > >> >>to AM >>>>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002 >>>>>exited >>>>> > > >>with >>>>> > > >> >> >> >>>>exitCode: 1 >>>>> > > >> >> >> >>>> For more detailed output, check application >>>>>tracking >>>>> page: >>>>> > > >> >> >> >>>> >>>>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then, >>>>> > > >> >> >>click >>>>> > > >> >> >> >>>>on >>>>> > > >> >> >> >>>> links to logs of each attempt. >>>>> > > >> >> >> >>>> Diagnostics: Exception from container-launch. >>>>> > > >> >> >> >>>> Container id: >>>>>container_1424791969399_0008_02_000001 >>>>> > > >> >> >> >>>> Exit code: 1 >>>>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1: >>>>> > > >> >> >> >>>> at >>>>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >>>>> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >>>>> > > >> >> >> e(Shell.java:71 >>>>> > > >> >> >> >>>>5 >>>>> > > >> >> >> >>>>) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >>>>> > > >> >> >> erExecutor.laun >>>>> > > >> >> >> >>>>c >>>>> > > >> >> >> >>>>h >>>>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>>> > > >> >> >> er.launcher.Con >>>>> > > >> >> >> >>>>t >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>>> > > >> >> >> er.launcher.Con >>>>> > > >> >> >> >>>>t >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82) >>>>> > > >> >> >> >>>> at >>>>> java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>>> > > >> >> >> PoolExecutor.ja >>>>> > > >> >> >> >>>>v >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>:1145) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>>> > > >> >> >> dPoolExecutor.j >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>v >>>>> > > >> >> >> >>>>a:615) >>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>>> > > >> >> >> >>>> Container exited with a non-zero exit code 1 >>>>> > > >> >> >> >>>> Failing this attempt. Failing the application. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> ---------- Forwarded message ---------- >>>>> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]> >>>>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>>option >>>>>is >>>>> > not >>>>> > > >> >> >> >>>>responding >>>>> > > >> >> >> >>>>on >>>>> > > >> >> >> >>>> GUI >>>>> > > >> >> >> >>>> To: "[email protected]" >>>>> > > >> >> >><[email protected]> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> hi , >>>>> > > >> >> >> >>>> please someone give me a hand to resolve this >>>>>issue , >>>>> > > >>thanks. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Regards, >>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>> > > >> >> >> >>>> Bangalore R&D >>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> www.huawei.com >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>-------------------------------------------------------- >>>>> > > >> >> >> --------------- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>> >>>>------------------------------------------------------------ >>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>confidential >>>>> > > >>information >>>>> > > >> >> >>from >>>>> > > >> >> >> >>>> HUAWEI, which >>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>address >>>>> is >>>>> > > >> >>listed >>>>> > > >> >> >> >>>>above. >>>>> > > >> >> >> >>>> Any use of the >>>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>>but >>>>> not >>>>> > > >> >>limited >>>>> > > >> >> >> >>>>to, >>>>> > > >> >> >> >>>> total or partial >>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>persons >>>>> > other >>>>> > > >> >>than >>>>> > > >> >> >>the >>>>> > > >> >> >> >>>> intended >>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>e-mail in >>>>> > > >>error, >>>>> > > >> >> >> >>>>please >>>>> > > >> >> >> >>>> notify the sender by >>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> ________________________________________ >>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >>>>> > > >> >> >> >>>> To: [email protected] >>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>>option >>>>>is >>>>> > not >>>>> > > >> >> >> >>>>responding >>>>> > > >> >> >> >>>>on >>>>> > > >> >> >> >>>> GUI >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>>cube >>>>>I >>>>> get >>>>> > > >>an >>>>> > > >> >> >> >>>>exception >>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>>correctly >>>>> and >>>>> > > >> >> >> >>>>HCatInputFormat >>>>> > > >> >> >> >>>> this file is present in >>>>>hive-hcatalog-core-0.14.0.jar >>>>>. >>>>> > > >>Please >>>>> > > >> >>let >>>>> > > >> >> >>me >>>>> > > >> >> >> >>>>know >>>>> > > >> >> >> >>>> what can I do to resolve this ? >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> -- This was path issue , now no more exception in >>>>> > kylin.log >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> But saveing cube still fails with error. And still >>>>>can't >>>>> > add >>>>> > > >> >> >>measures. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Error Message >>>>> > > >> >> >> >>>> Failed to take action. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> In log I can find no exception. Following is the >>>>>last >>>>>log >>>>> > in >>>>> > > >> >> >>kylin.log >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24 >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >>>>> > > >> >> >> .DefaultSchedul >>>>> > > >> >> >> >>>>e >>>>> > > >> >> >> >>>>r >>>>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >>>>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 >>>>>ready, >>>>>6 >>>>> > others >>>>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >>>>> > > >> >> >> beController.de >>>>> > > >> >> >> >>>>s >>>>> > > >> >> >> >>>>e >>>>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)] >>>>> > > >> >> >> >>>> - Saving cube { >>>>> > > >> >> >> >>>> "name": "", >>>>> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >>>> "lookups": [], >>>>> > > >> >> >> >>>> "filter_condition": "", >>>>> > > >> >> >> >>>> "capacity": "SMALL", >>>>> > > >> >> >> >>>> "partition_desc": { >>>>> > > >> >> >> >>>> "partition_date_column": "", >>>>> > > >> >> >> >>>> "partition_date_start": 0, >>>>> > > >> >> >> >>>> "partition_type": "APPEND" >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> "last_modified": 0 >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> local access logs all with 200 , so seems ok. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET >>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] >>>>>"POST >>>>> > > >> >> >>/kylin/api/cubes >>>>> > > >> >> >> >>>> HTTP/1.1" 200 701 >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Regards, >>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>> > > >> >> >> >>>> Bangalore R&D >>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> www.huawei.com >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>-------------------------------------------------------- >>>>> > > >> >> >> --------------- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>> >>>>------------------------------------------------------------ >>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>confidential >>>>> > > >>information >>>>> > > >> >> >>from >>>>> > > >> >> >> >>>> HUAWEI, which >>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>address >>>>> is >>>>> > > >> >>listed >>>>> > > >> >> >> >>>>above. >>>>> > > >> >> >> >>>> Any use of the >>>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>>but >>>>> not >>>>> > > >> >>limited >>>>> > > >> >> >> >>>>to, >>>>> > > >> >> >> >>>> total or partial >>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>persons >>>>> > other >>>>> > > >> >>than >>>>> > > >> >> >>the >>>>> > > >> >> >> >>>> intended >>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>e-mail in >>>>> > > >>error, >>>>> > > >> >> >> >>>>please >>>>> > > >> >> >> >>>> notify the sender by >>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> ________________________________________ >>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >>>>> > > >> >> >> >>>> To: [email protected] >>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>>> > > >> >> >> >>>> Subject: Error while making cube & Measure option >>>>>is >>>>>not >>>>> > > >> >> >>responding on >>>>> > > >> >> >> >>>>GUI >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Hi All , >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> I am building a simple cube for test and using >>>>>the >>>>> > binary >>>>> > > >> >>build >>>>> > > >> >> >> >>>>0.7.1 >>>>> > > >> >> >> >>>> . I have following hive tables with columns. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> fact_sales: >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> storeid int >>>>> > > >> >> >> >>>> itemid int >>>>> > > >> >> >> >>>> custid int >>>>> > > >> >> >> >>>> qty int >>>>> > > >> >> >> >>>> price double >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> dim_customer >>>>> > > >> >> >> >>>> custid int >>>>> > > >> >> >> >>>> name string >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> dim_item >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> itemid int >>>>> > > >> >> >> >>>> category string >>>>> > > >> >> >> >>>> brand string >>>>> > > >> >> >> >>>> color string >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> dim_store >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> storeid int >>>>> > > >> >> >> >>>> city string >>>>> > > >> >> >> >>>> state string >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Please help me to answer following issues; >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> 1. When I go to measure section and click on >>>>>measure >>>>> > option , >>>>> > > >> >> >>there is >>>>> > > >> >> >> >>>>no >>>>> > > >> >> >> >>>> response , I want add measure on qty and price with >>>>>sum >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>>cube >>>>>I >>>>> get >>>>> > > >>an >>>>> > > >> >> >> >>>>exception >>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>>correctly >>>>> and >>>>> > > >> >> >> >>>>HCatInputFormat >>>>> > > >> >> >> >>>> this file is present in >>>>>hive-hcatalog-core-0.14.0.jar >>>>>. >>>>> > > >>Please >>>>> > > >> >>let >>>>> > > >> >> >>me >>>>> > > >> >> >> >>>>know >>>>> > > >> >> >> >>>> what can I do to resolve this ? >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> 3. Also I have another question since this is a >>>>>test >>>>>and >>>>> > > >>data is >>>>> > > >> >> >>small >>>>> > > >> >> >> >>>>I >>>>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to >>>>>skip >>>>> > > >>partition >>>>> > > >> >> >>stage >>>>> > > >> >> >> >>>> while cube build ? >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Exception >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24 >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >>>>> > > >> >> >> l.DefaultSchedu >>>>> > > >> >> >> >>>>l >>>>> > > >> >> >> >>>>e >>>>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >>>>> > > >> >> >> >>>> - ExecuteException >>>>> job:c3532a6f-97ea-474a-b36a-218dd517cedb >>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut >>>>> > > >> >> >> e(AbstractExecu >>>>> > > >> >> >> >>>>t >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>ble.java:102) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >>>>> > > >> >> >> bRunner.run(Def >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>u >>>>> > > >> >> >> >>>>ltScheduler.java:132) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>>> > > >> >> >> PoolExecutor.ja >>>>> > > >> >> >> >>>>v >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>:1145) >>>>> > > >> >> >> >>>> at >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>>> > > >> >> >> dPoolExecutor.j >>>>> > > >> >> >> >>>>a >>>>> > > >> >> >> >>>>v >>>>> > > >> >> >> >>>>a:615) >>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>>> > > >> >> >> >>>> Caused by: >>>>> org.apache.kylin.job.exception.ExecuteException: >>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> The JSON is as below. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "name": "Retail_Cube", >>>>> > > >> >> >> >>>> "description": "", >>>>> > > >> >> >> >>>> "dimensions": [ >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >>>> "hierarchy": false, >>>>> > > >> >> >> >>>> "derived": null, >>>>> > > >> >> >> >>>> "column": [ >>>>> > > >> >> >> >>>> "STOREID" >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "id": 1 >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >>>> "hierarchy": false, >>>>> > > >> >> >> >>>> "derived": null, >>>>> > > >> >> >> >>>> "column": [ >>>>> > > >> >> >> >>>> "ITEMID" >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "id": 2 >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>>> > > >> >> >> >>>> "hierarchy": false, >>>>> > > >> >> >> >>>> "derived": null, >>>>> > > >> >> >> >>>> "column": [ >>>>> > > >> >> >> >>>> "CUSTID" >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "id": 3 >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "measures": [ >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "id": 1, >>>>> > > >> >> >> >>>> "name": "_COUNT_", >>>>> > > >> >> >> >>>> "function": { >>>>> > > >> >> >> >>>> "expression": "COUNT", >>>>> > > >> >> >> >>>> "returntype": "bigint", >>>>> > > >> >> >> >>>> "parameter": { >>>>> > > >> >> >> >>>> "type": "constant", >>>>> > > >> >> >> >>>> "value": 1 >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "rowkey": { >>>>> > > >> >> >> >>>> "rowkey_columns": [ >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "column": "STOREID", >>>>> > > >> >> >> >>>> "length": 0, >>>>> > > >> >> >> >>>> "dictionary": "true", >>>>> > > >> >> >> >>>> "mandatory": false >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "column": "ITEMID", >>>>> > > >> >> >> >>>> "length": 0, >>>>> > > >> >> >> >>>> "dictionary": "true", >>>>> > > >> >> >> >>>> "mandatory": false >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "column": "CUSTID", >>>>> > > >> >> >> >>>> "length": 0, >>>>> > > >> >> >> >>>> "dictionary": "true", >>>>> > > >> >> >> >>>> "mandatory": false >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> ], >>>>> > > >> >> >> >>>> "aggregation_groups": [ >>>>> > > >> >> >> >>>> [ >>>>> > > >> >> >> >>>> "STOREID", >>>>> > > >> >> >> >>>> "ITEMID", >>>>> > > >> >> >> >>>> "CUSTID" >>>>> > > >> >> >> >>>> ] >>>>> > > >> >> >> >>>> ] >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> "notify_list": [], >>>>> > > >> >> >> >>>> "capacity": "", >>>>> > > >> >> >> >>>> "hbase_mapping": { >>>>> > > >> >> >> >>>> "column_family": [ >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "name": "f1", >>>>> > > >> >> >> >>>> "columns": [ >>>>> > > >> >> >> >>>> { >>>>> > > >> >> >> >>>> "qualifier": "m", >>>>> > > >> >> >> >>>> "measure_refs": [ >>>>> > > >> >> >> >>>> "_COUNT_" >>>>> > > >> >> >> >>>> ] >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> ] >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> ] >>>>> > > >> >> >> >>>> }, >>>>> > > >> >> >> >>>> "project": "RetailProject", >>>>> > > >> >> >> >>>> "model_name": "Retail_Cube" >>>>> > > >> >> >> >>>> } >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> Regards, >>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>> > > >> >> >> >>>> Bangalore R&D >>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> www.huawei.com >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>>>>>>>-------------------------------------------------------- >>>>> > > >> >> >> --------------- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>- >>>>> > > >> >> >> >>>>> >>>>------------------------------------------------------------ >>>>> > > >> >> >> >>>> This e-mail and its attachments contain >>>>>confidential >>>>> > > >>information >>>>> > > >> >> >>from >>>>> > > >> >> >> >>>> HUAWEI, which >>>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>>address >>>>> is >>>>> > > >> >>listed >>>>> > > >> >> >> >>>>above. >>>>> > > >> >> >> >>>> Any use of the >>>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>>but >>>>> not >>>>> > > >> >>limited >>>>> > > >> >> >> >>>>to, >>>>> > > >> >> >> >>>> total or partial >>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>>persons >>>>> > other >>>>> > > >> >>than >>>>> > > >> >> >>the >>>>> > > >> >> >> >>>> intended >>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>>e-mail in >>>>> > > >>error, >>>>> > > >> >> >> >>>>please >>>>> > > >> >> >> >>>> notify the sender by >>>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>>> -- >>>>> > > >> >> >> >>>> Regards, >>>>> > > >> >> >> >>>> Santosh Akhilesh >>>>> > > >> >> >> >>>> +91-0-9845482201 >>>>> > > >> >> >> >>>> >>>>> > > >> >> >> >>> >>>>> > > >> >> >> >>> >>>>> > > >> >> >> >>> >>>>> > > >> >> >> >>>-- >>>>> > > >> >> >> >>>Regards, >>>>> > > >> >> >> >>>Santosh Akhilesh >>>>> > > >> >> >> >>>+91-0-9845482201 >>>>> > > >> >> >> >>>>> > > >> >> >> >>>>> > > >> >> >>>>> > > >> >> >>>>> > > >> >>>>> > > >> >>>>> > > >>>>> > > >>>>> > >>>>> > >>>>> > -- >>>>> > Regards, >>>>> > Santosh Akhilesh >>>>> > +91-0-9845482201 >>>>> > >>>>> >>>> >>>> >>>> >>>>-- >>>>Regards, >>>>Santosh Akhilesh >>>>+91-0-9845482201
