Hi Shaofeng ,
I configured job histroy server and no more connection exception. now I get
the MR counter exception which we were suspecting.
My haddop version is indeed 2.6.0 , So any idea what can be done for this ?
QuartzScheduler_Worker-8]:[2015-02-28
00:36:26,507][DEBUG][com.kylinolap.job.tools.HadoopStatusChecker.checkStatus(HadoopStatusChecker.java:74)]
- State of Hadoop job: job_1424957178195_0031:FINISHED-SUCCEEDED
[QuartzScheduler_Worker-8]:[2015-02-28
00:36:27,204][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOutput.java:176)]
- No enum constant org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES
java.lang.IllegalArgumentException: No enum constant
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES
at java.lang.Enum.valueOf(Enum.java:236)
at
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
at
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
at
org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
at org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
at
org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
at org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
at
com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.java:287)
at
com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOutput.java:162)
at
com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.java:85)
at com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86)
at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
at
org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:573)
Regards,
Santosh Akhilesh
Bangalore R&D
HUAWEI TECHNOLOGIES CO.,LTD.
www.huawei.com
-------------------------------------------------------------------------------------------------------------------------------------
This e-mail and its attachments contain confidential information from HUAWEI,
which
is intended only for the person or entity whose address is listed above. Any
use of the
information contained herein in any way (including, but not limited to, total
or partial
disclosure, reproduction, or dissemination) by persons other than the intended
recipient(s) is prohibited. If you receive this e-mail in error, please notify
the sender by
phone or email immediately and delete it!
________________________________________
From: Shi, Shaofeng [[email protected]]
Sent: Friday, February 27, 2015 3:10 PM
To: [email protected]
Subject: Re: Cube Build Failed at Last Step//RE: Error while making cube &
Measure option is not responding on GUI
0.0.0.0:10020 isn’t a valid network address I think; please check the
“mapreduce.jobhistory.address” in your mapred-site.xml; it should be
something like:
<property>
<name>mapreduce.jobhistory.address</name>
<value>sandbox.hortonworks.com:10020</value>
</property>
On 2/27/15, 5:29 PM, "Santoshakhilesh" <[email protected]> wrote:
>Hi Shaofeng ,
> No I have not found MR counter exception. I get following exception
>frequently. I think this is related LogHistory server of hadoop.
>
>[QuartzScheduler_Worker-23]:[2015-02-27
>22:18:37,299][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCo
>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException:
>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>failed on connection exception: java.net.ConnectException: Connection
>refused; For more details see:
>http://wiki.apache.org/hadoop/ConnectionRefused
>com.kylinolap.job.exception.JobException: java.io.IOException:
>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>failed on connection exception: java.net.ConnectException: Connection
>refused; For more details see:
>http://wiki.apache.org/hadoop/ConnectionRefused
> at
>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.j
>ava:289)
> at
>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOu
>tput.java:162)
> at
>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.ja
>va:85)
> at
>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86)
> at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
> at
>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5
>73)
>Caused by: java.io.IOException: java.net.ConnectException: Call From
>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception:
>java.net.ConnectException: Connection refused; For more details see:
>http://wiki.apache.org/hadoop/ConnectionRefused
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Shi, Shaofeng [[email protected]]
>Sent: Friday, February 27, 2015 2:47 PM
>To: [email protected]
>Cc: Kulbhushan Rana
>Subject: Re: Cube Build Failed at Last Step//RE: Error while making cube
>& Measure option is not responding on GUI
>
>Did you figure out the exception of "No enum constant
>org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES” ? Is it still be
>thrown in the logs? In the last step, Kylin need to parse the MR counters
>to update cube size; Please refer to
>https://issues.apache.org/jira/browse/MAPREDUCE-5831 for that error.
>
>On 2/27/15, 5:04 PM, "Santoshakhilesh" <[email protected]>
>wrote:
>
>>Hi Shaofeng ,
>> Cube building is failed at last step while loading Hfile to
>>Hbase with exception "Can't get cube segment size.
>>". What could be reason ?
>>
>>parameter : -input
>>/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/
>>-htablename KYLIN_K27LDMX63W -cubename RetailCube
>>
>>Log:
>>
>>Start to execute command:
>> -input /tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/
>>-htablename KYLIN_K27LDMX63W -cubename RetailCube
>>Command execute return code 0
>>Failed with Exception:java.lang.RuntimeException: Can't get cube segment
>>size.
>> at
>>com.kylinolap.job.flow.JobFlowListener.updateCubeSegmentInfoOnSucceed(Job
>>F
>>lowListener.java:247)
>> at
>>com.kylinolap.job.flow.JobFlowListener.jobWasExecuted(JobFlowListener.jav
>>a
>>:101)
>> at
>>org.quartz.core.QuartzScheduler.notifyJobListenersWasExecuted(QuartzSched
>>u
>>ler.java:1985)
>> at
>>org.quartz.core.JobRunShell.notifyJobListenersComplete(JobRunShell.java:3
>>4
>>0)
>> at org.quartz.core.JobRunShell.run(JobRunShell.java:224)
>> at
>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:
>>5
>>73)
>>
>>I have checked in hbase shell and following are the tables in hbase;
>>hbase(main):001:0> list
>>TABLE
>>
>>KYLIN_K27LDMX63W
>>kylin_metadata_qa
>>kylin_metadata_qa_acl
>>kylin_metadata_qa_cube
>>kylin_metadata_qa_dict
>>kylin_metadata_qa_invertedindex
>>kylin_metadata_qa_job
>>kylin_metadata_qa_job_output
>>kylin_metadata_qa_proj
>>kylin_metadata_qa_table_snapshot
>>kylin_metadata_qa_user
>>11 row(s) in 0.8990 seconds
>>
>>
>>Regards,
>>Santosh Akhilesh
>>Bangalore R&D
>>HUAWEI TECHNOLOGIES CO.,LTD.
>>
>>www.huawei.com
>>-------------------------------------------------------------------------
>>-
>>-----------------------------------------------------------
>>This e-mail and its attachments contain confidential information from
>>HUAWEI, which
>>is intended only for the person or entity whose address is listed above.
>>Any use of the
>>information contained herein in any way (including, but not limited to,
>>total or partial
>>disclosure, reproduction, or dissemination) by persons other than the
>>intended
>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>notify the sender by
>>phone or email immediately and delete it!
>>
>>________________________________________
>>From: Santoshakhilesh
>>Sent: Friday, February 27, 2015 2:15 PM
>>To: [email protected]
>>Subject: RE: Error while making cube & Measure option is not responding
>>on GUI
>>
>>I have manually copied the jar to /tmp/kylin , now satge 2 is done ,
>>thanks.
>>
>>Regards,
>>Santosh Akhilesh
>>Bangalore R&D
>>HUAWEI TECHNOLOGIES CO.,LTD.
>>
>>www.huawei.com
>>-------------------------------------------------------------------------
>>-
>>-----------------------------------------------------------
>>This e-mail and its attachments contain confidential information from
>>HUAWEI, which
>>is intended only for the person or entity whose address is listed above.
>>Any use of the
>>information contained herein in any way (including, but not limited to,
>>total or partial
>>disclosure, reproduction, or dissemination) by persons other than the
>>intended
>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>notify the sender by
>>phone or email immediately and delete it!
>>
>>________________________________________
>>From: Shi, Shaofeng [[email protected]]
>>Sent: Friday, February 27, 2015 1:00 PM
>>To: [email protected]
>>Cc: Kulbhushan Rana
>>Subject: Re: Error while making cube & Measure option is not responding
>>on GUI
>>
>>In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we
>>renamed the package to “org.apache.kylin.xxx”; When you downgrade to 0.6,
>>did you also replace the jar location with 0.6 ones in kylin.properties?
>>
>>On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]>
>>wrote:
>>
>>>Hi Shaofeng ,
>>> I have added my fact and dimension tables under default
>>>database
>>>of hive.
>>> Now stage 1 of Cube Build is ok. And there is failure at step2.
>>> The map reduce job for the finding distinct columns of fact
>>>table
>>>is error. Yarn log is as below.
>>> Strangely this is class not found error. I have checked the
>>>Kylin.properties and the jar is already set as below.
>>>kylin. log has one exception connecting to linux/10.19.93.68 to
>>>0.0.0.0:10020
>>> Please help me to give a clue , I am also trying to check meanwhile
>>>
>>>Thanks.
>>>kylin property
>>># Temp folder in hdfs
>>>kylin.hdfs.working.dir=/tmp
>>># Path to the local(relative to job engine) job jar, job engine will use
>>>this jar
>>>kylin.job.jar=/tmp/kylin/kylin-job-latest.jar
>>>
>>>Map Reduce error
>>>----------------------------
>>>2015-02-27 20:24:25,262 FATAL [main] org.apache.hadoop.mapred.YarnChild:
>>>Error running child : java.lang.NoClassDefFoundError:
>>>com/kylinolap/common/mr/KylinMapper
>>> at java.lang.ClassLoader.defineClass1(Native Method)
>>> at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>>> at
>>>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>>> at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>> at java.security.AccessController.doPrivileged(Native Method)
>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>> at java.lang.Class.forName0(Native Method)
>>> at java.lang.Class.forName(Class.java:274)
>>> at
>>>org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.
>>>j
>>>a
>>>va:2013)
>>>
>>>Kylin.log
>>>QuartzScheduler_Worker-20]:[2015-02-27
>>>20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFetc
>>>h
>>>e
>>>r.java:60)] - 0 pending jobs
>>>[QuartzScheduler_Worker-19]:[2015-02-27
>>>20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJob
>>>C
>>>o
>>>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException:
>>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>>>failed on connection exception: java.net.ConnectException: Connection
>>>refused; For more details see:
>>>http://wiki.apache.org/hadoop/ConnectionRefused
>>>com.kylinolap.job.exception.JobException: java.io.IOException:
>>>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020
>>>failed on connection exception: java.net.ConnectException: Connection
>>>refused; For more details see:
>>>http://wiki.apache.org/hadoop/ConnectionRefused
>>> at
>>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob
>>>.
>>>j
>>>ava:289)
>>> at
>>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmd
>>>O
>>>u
>>>tput.java:162)
>>> at
>>>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.
>>>j
>>>a
>>>va:85)
>>> at
>>>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86
>>>)
>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
>>> at
>>>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java
>>>:
>>>5
>>>73)
>>>Caused by: java.io.IOException: java.net.ConnectException: Call From
>>>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception:
>>>java.net.ConnectException: Connection refused; For more details see:
>>>http://wiki.apache.org/hadoop/ConnectionRefused
>>> at
>>>org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDeleg
>>>a
>>>t
>>>e.java:331)
>>> at
>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServ
>>>i
>>>c
>>>eDelegate.java:368)
>>> at
>>>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
>>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>>> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>>> at java.security.AccessController.doPrivileged(Native Method)
>>> at javax.security.auth.Subject.doAs(Subject.java:415)
>>> at
>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformatio
>>>n
>>>.
>>>java:1491)
>>> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>>> at
>>>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob
>>>.
>>>j
>>>ava:287)
>>> ... 5 more
>>>
>>>Regards,
>>>Santosh Akhilesh
>>>Bangalore R&D
>>>HUAWEI TECHNOLOGIES CO.,LTD.
>>>
>>>www.huawei.com
>>>------------------------------------------------------------------------
>>>-
>>>-
>>>-----------------------------------------------------------
>>>This e-mail and its attachments contain confidential information from
>>>HUAWEI, which
>>>is intended only for the person or entity whose address is listed above.
>>>Any use of the
>>>information contained herein in any way (including, but not limited to,
>>>total or partial
>>>disclosure, reproduction, or dissemination) by persons other than the
>>>intended
>>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>>notify the sender by
>>>phone or email immediately and delete it!
>>>
>>>________________________________________
>>>From: Shi, Shaofeng [[email protected]]
>>>Sent: Friday, February 27, 2015 8:01 AM
>>>To: [email protected]
>>>Subject: Re: Error while making cube & Measure option is not responding
>>>on GUI
>>>
>>>In 0.6.x it only support tables in default database, this is a
>>>limitation;
>>>The support for non-default tables will be released in 0.7;
>>>
>>>To bypass this issue for now, please copy the table to default database
>>>as
>>>a workaround;
>>>
>>>On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]>
>>>wrote:
>>>
>>>>@Jason
>>>>thanks , but now as suggested by Saofeng I m not using the inverted
>>>>index
>>>>brach as its not stable.
>>>>I have switched back to 0.6 branch , in this branch yesterday night I
>>>>could
>>>>crete the cube successfully but there is issue while building it. I
>>>>feel
>>>>that at step 1 of cube build while creating flat table when command is
>>>>issued to hive if the tables are not under default datbase flat table
>>>>creation is failed and cube build fails. my fact and dimension tables
>>>>are
>>>>under a database called retail.
>>>>
>>>>@Saofeng - Can you please confirm this behavior ? Do I need to create
>>>>the
>>>>hive tables under default database?
>>>>
>>>>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]>
>>>>wrote:
>>>>
>>>>> @Santoshakhilesh
>>>>>
>>>>> 1. When I go to measure section and click on measure option , there
>>>>>is
>>>>>no
>>>>> response , I want add measure on qty and price with sum
>>>>> --bug fixed on inverted-index branch
>>>>>
>>>>>
>>>>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh <
>>>>> [email protected]
>>>>> > wrote:
>>>>>
>>>>> > Hi Shaofeng ,
>>>>> > I have build the 0.6 version and now able to create the cube
>>>>> > successfully.
>>>>> > While building the cube , it fails at step1 with following
>>>>>error.
>>>>> > Table not found 'DIM_ITEM'
>>>>> > the table exists , but its under retail data base and not
>>>>>under
>>>>> > default database.
>>>>> > does kylin require hive taables to be under default database ?
>>>>>I
>>>>>see
>>>>> > the flat table being created under default database.
>>>>> >
>>>>> > Logging initialized using configuration in
>>>>> >
>>>>> >
>>>>>
>>>>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-comm
>>>>>o
>>>>>n
>>>>>-
>>>>>1.0.0.jar!/hive-log4j.properties
>>>>> > SLF4J: Class path contains multiple SLF4J bindings.
>>>>> > SLF4J: Found binding in
>>>>> >
>>>>> >
>>>>>
>>>>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/comm
>>>>>o
>>>>>n
>>>>>/
>>>>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> > SLF4J: Found binding in
>>>>> >
>>>>> >
>>>>>
>>>>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jdb
>>>>>c
>>>>>-
>>>>>1
>>>>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>> > explanation.
>>>>> > SLF4J: Actual binding is of type
>>>>>[org.slf4j.impl.Log4jLoggerFactory]
>>>>> > OK
>>>>> > Time taken: 0.964 seconds
>>>>> > OK
>>>>> > Time taken: 0.948 seconds
>>>>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found
>>>>> > 'DIM_ITEM'
>>>>> >
>>>>> >
>>>>> >
>>>>> > Command is as below.
>>>>> >
>>>>> > hive -e "DROP TABLE IF EXISTS
>>>>> >
>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4
>>>>>4
>>>>>;
>>>>> > CREATE EXTERNAL TABLE IF NOT EXISTS
>>>>> >
>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4
>>>>>4
>>>>> > (
>>>>> > STOREID int
>>>>> > ,ITEMID int
>>>>> > ,CUSTID int
>>>>> > ,QTY int
>>>>> > ,AMOUNT double
>>>>> > )
>>>>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177'
>>>>> > STORED AS SEQUENCEFILE
>>>>> > LOCATION
>>>>> >
>>>>>
>>>>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_te
>>>>>s
>>>>>t
>>>>>_
>>>>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44';
>>>>> > SET hive.exec.compress.output=true;
>>>>> > SET hive.auto.convert.join.noconditionaltask = true;
>>>>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000;
>>>>> > INSERT OVERWRITE TABLE
>>>>> >
>>>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd4
>>>>>4
>>>>> > SELECT
>>>>> > FACT_SALES.STOREID
>>>>> > ,FACT_SALES.ITEMID
>>>>> > ,FACT_SALES.CUSTID
>>>>> > ,FACT_SALES.QTY
>>>>> > ,FACT_SALES.AMOUNT
>>>>> > FROM FACT_SALES
>>>>> > INNER JOIN DIM_STORE
>>>>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID
>>>>> > INNER JOIN DIM_ITEM
>>>>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID
>>>>> > INNER JOIN DIM_CUSTOMER
>>>>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID
>>>>> > ;
>>>>> > "
>>>>> >
>>>>> >
>>>>> >
>>>>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]>
>>>>>wrote:
>>>>> >
>>>>> > > The 0.7.1 is test version, its package contains the “snapshot”
>>>>>suffix;
>>>>> we
>>>>> > > will upload a new package there; Luke will also add a message
>>>>>there
>>>>>to
>>>>> > > avoid this confusion;
>>>>> > >
>>>>> > > Regarding the problem that you encountered, could you please open
>>>>>a
>>>>> JIRA
>>>>> > > ticket for tracking? Here is link of Apache JIRA:
>>>>> > >
>>>>> > > https://issues.apache.org/jira/secure/Dashboard.jspa
>>>>> > >
>>>>> > >
>>>>> > > Thanks for the feedback!
>>>>> > >
>>>>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh"
>>>>><[email protected]>
>>>>> > > wrote:
>>>>> > >
>>>>> > > >Actually I see this being published on kylin webpage.
>>>>> > > >http://kylin.incubator.apache.org/download/
>>>>> > > >I am using 0.7.1 inverted index branch binary distribution.
>>>>> > > >If this is not stable please give me the link of stable branch I
>>>>>would
>>>>> > try
>>>>> > > >building and testing tonight.
>>>>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]>
>>>>> wrote:
>>>>> > > >
>>>>> > > >> Hi Santosh, it is not recommended to use the dev code branch
>>>>> > (actually I
>>>>> > > >> don’t know how you get the v0.7.x build and what’s the detail
>>>>> version
>>>>> > of
>>>>> > > >> that; each day we submit many changes to that);
>>>>> > > >>
>>>>> > > >> The options are 1) switch back to latest release v0.6.5; or 2)
>>>>>wait
>>>>> > for
>>>>> > > >> the formal release of 0.7, that should be in March; Otherwise,
>>>>>we
>>>>> > > >>couldn’t
>>>>> > > >> ensure there is no new problems come out in your next steps;
>>>>> > > >>
>>>>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh"
>>>>><[email protected]>
>>>>> > > >>wrote:
>>>>> > > >>
>>>>> > > >> >Hi Shaofeng
>>>>> > > >> >So what do you suggest , how should I proceed further with
>>>>>this
>>>>> > > >>release?
>>>>> > > >> >Will there be a patch? Any alternate way I can create cube?
>>>>> > > >> >Please suggest.
>>>>> > > >> >Regards
>>>>> > > >> >Santosh
>>>>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng
>>>>><[email protected]>
>>>>> > > wrote:
>>>>> > > >> >
>>>>> > > >> >> Hi Santosh,
>>>>> > > >> >>
>>>>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we
>>>>>have
>>>>> > > >> >>metadata
>>>>> > > >> >> structure change; While, the web UI (cube wizard) for this
>>>>>change
>>>>> > > >>hasn’t
>>>>> > > >> >> been stabilized; So it is not strange that you got trouble
>>>>>when
>>>>> > > >>saving
>>>>> > > >> >>the
>>>>> > > >> >> cube;
>>>>> > > >> >>
>>>>> > > >> >> @Jason, any idea about the JS error?
>>>>> > > >> >>
>>>>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <
>>>>> [email protected]
>>>>> > >
>>>>> > > >> >>wrote:
>>>>> > > >> >>
>>>>> > > >> >> >Hi Shaofeng,
>>>>> > > >> >> >
>>>>> > > >> >> >I am using the binary distribution 0.7.1. I have not been
>>>>>able
>>>>> to
>>>>> > > >>save
>>>>> > > >> >> >cube
>>>>> > > >> >> >even once. I have tried creating new project and from
>>>>>local
>>>>> > machine
>>>>> > > >>and
>>>>> > > >> >> >server machine. But I am always stuck with this error. I
>>>>>am
>>>>> never
>>>>> > > >> >>allowed
>>>>> > > >> >> >to add measures and never been able to save the cube. I
>>>>>also
>>>>>see
>>>>> > the
>>>>> > > >> >> >kylin.log and it always tries to save cube with append
>>>>>mode.
>>>>>One
>>>>> > > >>thing
>>>>> > > >> >>I
>>>>> > > >> >> >need to tell that at partition stage since I don't have a
>>>>>big
>>>>> fact
>>>>> > > >> >>table
>>>>> > > >> >> >now I have not partititioned the fact table and I skip
>>>>>this
>>>>> step.
>>>>> > > >>Does
>>>>> > > >> >> >this
>>>>> > > >> >> >have affect in saving the cube. Is this because some
>>>>>metadata is
>>>>> > > >> >>available
>>>>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop
>>>>>2.6.6.
>>>>> > Yes
>>>>> > > >> >>kylin
>>>>> > > >> >> >propert I have not added the jar. I will add them and
>>>>>check.
>>>>>But
>>>>> > > >>cube
>>>>> > > >> >> >creation failure is really puzzling me. I could see no
>>>>>error
>>>>> logs
>>>>> > in
>>>>> > > >> >> >kylin.log.
>>>>> > > >> >> >Regards
>>>>> > > >> >> >Santosh
>>>>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng
>>>>><[email protected]
>>>>> >
>>>>> > > >> wrote:
>>>>> > > >> >> >
>>>>> > > >> >> >> Which version or code branch are you using? I assume
>>>>>you’re
>>>>> > using
>>>>> > > >>the
>>>>> > > >> >> >> stable version from master; Seems you’re trying to edit
>>>>>an
>>>>> > > >>existing
>>>>> > > >> >>cube
>>>>> > > >> >> >> to add new measurement, try refresh your browser's
>>>>>cache;
>>>>>if
>>>>> it
>>>>> > > >>still
>>>>> > > >> >> >> couldn’t be saved, try to create a new cube;
>>>>> > > >> >> >>
>>>>> > > >> >> >> The two error traces in tomcat need be taken care:
>>>>> > > >> >> >>
>>>>> > > >> >> >> 1) java.lang.NoClassDefFoundError:
>>>>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper
>>>>> > > >> >> >> Please check kylin.properties file, making sure
>>>>>the
>>>>> > > >> >> >>“kylin.job.jar”
>>>>> > > >> >> >> points to a right jar file; It will be loaded in
>>>>>Map-reduce;
>>>>> > > >> >> >>
>>>>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant
>>>>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>>>> > > >> >> >> This indicates your hadoop version might be old;
>>>>> Please
>>>>> > > >>check
>>>>> > > >> >> >>and
>>>>> > > >> >> >> ensure
>>>>> > > >> >> >> hadoop version is 2.2 or above.
>>>>> > > >> >> >>
>>>>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh"
>>>>> > > >><[email protected]>
>>>>> > > >> >> >> wrote:
>>>>> > > >> >> >>
>>>>> > > >> >> >> >Hi Shaofeng ,
>>>>> > > >> >> >> >
>>>>> > > >> >> >> > I am using chrome , When I click on button to add
>>>>> measures ,
>>>>> > > >> >> >>following
>>>>> > > >> >> >> >is error on chrome console. When I try to save the cube
>>>>>there
>>>>> > is
>>>>> > > >>no
>>>>> > > >> >> >>error
>>>>> > > >> >> >> >in console. I just get a error dialog saying failed to
>>>>>take
>>>>> > > >>action
>>>>> > > >> >>and
>>>>> > > >> >> >> >gives me the JSON cube schema.
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >Error on chrome debug console is as below;
>>>>> > > >> >> >> >
>>>>> > > >> >> >> > ReferenceError: CubeDescModel is not defined
>>>>> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>>>>> > > >> >> >> > at scripts.min.0.js:180
>>>>> > > >> >> >> > at scripts.min.0.js:197
>>>>> > > >> >> >> > at h.$eval (scripts.min.0.js:119)
>>>>> > > >> >> >> > at h.$apply (scripts.min.0.js:119)
>>>>> > > >> >> >> > at HTMLButtonElement.<anonymous>
>>>>>(scripts.min.0.js:197)
>>>>> > > >> >> >> > at HTMLButtonElement.m.event.dispatch
>>>>> (scripts.min.0.js:3)
>>>>> > > >> >> >> > at HTMLButtonElement.r.handle
>>>>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>>>>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous
>>>>> > > >> >>function)scripts.min.0.js:119
>>>>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous
>>>>> > > >>function)scripts.min.0.js:3
>>>>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle
>>>>> > > >> >> >> >
>>>>> > > >> >> >> > About the hive table import , I got pass the run
>>>>>shell
>>>>> > command
>>>>> > > >> >> >> >exception but it still fails the hadoop log is;
>>>>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main]
>>>>>org.apache.hadoop.mapred.
>>>>> > > >> >> YarnChild:
>>>>> > > >> >> >> >mapreduce.cluster.local.dir for child:
>>>>> > > >> >> >>
>>>>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
>>>>> > > >> >> >> cation_14249530
>>>>> > > >> >> >> >91340_0002
>>>>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main]
>>>>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation:
>>>>>session.id
>>>>> > is
>>>>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id
>>>>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main]
>>>>> > > >>org.apache.hadoop.mapred.Task:
>>>>> > > >> >> >>Using
>>>>> > > >> >> >> >ResourceCalculatorProcessTree : [ ]
>>>>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main]
>>>>> > > >> >> >>org.apache.hadoop.mapred.YarnChild:
>>>>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError:
>>>>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper
>>>>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method)
>>>>> > > >> >> >> > at
>>>>>java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >> >java.security.SecureClassLoader.defineClass(
>>>>> > > >> >> SecureClassLoader.java:142)
>>>>> > > >> >> >> > at
>>>>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>>>>> > > >> >> >> > at
>>>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>>>> > > >> >> >> > at
>>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>>>>> > > >> >> >> > at
>>>>>java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
>>>>> Method)
>>>>> > > >> >> >> > at
>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>>> > > >> >> >> > at
>>>>>java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>>> > > >> >> >> > at
>>>>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >tomcat logs:
>>>>> > > >> >> >> >usage: HiveColumnCardinalityJob
>>>>> > > >> >> >> > -output <path> Output path
>>>>> > > >> >> >> > -table <table name> The hive table name
>>>>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26
>>>>> > > >> >> >>
>>>>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
>>>>> > > >> >> >> lExecutable.doW
>>>>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute
>>>>> > > >> >> >>
>>>>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>>>>> > > >> >> >> >name=null, state=RUNNING}
>>>>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant
>>>>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>>>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>> > >
>>>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
>>>>> > > >> >> >> Framewo
>>>>> > > >> >> >> >rkCounterGroup.java:148)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
>>>>> > > >> >> >> findCounter(Fra
>>>>> > > >> >> >> >meworkCounterGroup.java:182)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
>>>>> > > >> >> >> ounter(Abstract
>>>>> > > >> >> >> >Counters.java:154)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
>>>>> > > >> >> >> erter.java:240)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
>>>>> > > >> >> >> rs(ClientServic
>>>>> > > >> >> >> >eDelegate.java:370)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(
>>>>> > > >> >> YARNRunner.java:511)
>>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>>>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>>>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
>>>>> Method)
>>>>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
>>>>> > > >> >> >> oupInformation.
>>>>> > > >> >> >> >java:1491)
>>>>> > > >> >> >> > at
>>>>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>>>>> > > >> >> >> > at
>>>>> > > >>
>>>>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>>>>> > > >> >> >> > at
>>>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.
>>>>> > > >> java:1289)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
>>>>> > > >> >> >> tion(AbstractHa
>>>>> > > >> >> >> >doopJob.java:134)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>> > > >>
>>>>> >
>>>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.ru
>>>>>>>>n
>>>>>>>>(
>>>>> > > >> >> >> HiveC
>>>>> > > >> >> >> >olumnCardinalityJob.java:114)
>>>>> > > >> >> >> > at
>>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>>>>> > > >> >> >> > at
>>>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
>>>>> > > >> >> >> doopShellExecut
>>>>> > > >> >> >> >able.java:62)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>>>> > > >> >> >> bstractExecutab
>>>>> > > >> >> >> >le.java:99)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW
>>>>> > > >> >> >> ork(DefaultChai
>>>>> > > >> >> >> >nedExecutable.java:50)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>>>> > > >> >> >> bstractExecutab
>>>>> > > >> >> >> >le.java:99)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
>>>>> > > >> >> >> nner.run(Defaul
>>>>> > > >> >> >> >tScheduler.java:132)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
>>>>> > > >> >> >> lExecutor.java:
>>>>> > > >> >> >> >1145)
>>>>> > > >> >> >> > at
>>>>> > > >> >> >>
>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
>>>>> > > >> >> >> olExecutor.java
>>>>> > > >> >> >> >:615)
>>>>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745)
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >Regards,
>>>>> > > >> >> >> >Santosh Akhilesh
>>>>> > > >> >> >> >Bangalore R&D
>>>>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD.
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >www.huawei.com
>>>>> > > >> >> >>
>>>>>>-----------------------------------------------------------
>>>>> > > >> >> >> ---------------
>>>>> > > >> >> >>
>>>>>>-----------------------------------------------------------
>>>>> > > >> >> >> >This e-mail and its attachments contain confidential
>>>>> > information
>>>>> > > >> >>from
>>>>> > > >> >> >> >HUAWEI, which
>>>>> > > >> >> >> >is intended only for the person or entity whose address
>>>>>is
>>>>> > listed
>>>>> > > >> >> >>above.
>>>>> > > >> >> >> >Any use of the
>>>>> > > >> >> >> >information contained herein in any way (including, but
>>>>>not
>>>>> > > >>limited
>>>>> > > >> >>to,
>>>>> > > >> >> >> >total or partial
>>>>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons
>>>>>other
>>>>> > than
>>>>> > > >> >>the
>>>>> > > >> >> >> >intended
>>>>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail
>>>>>in
>>>>> > error,
>>>>> > > >> >>please
>>>>> > > >> >> >> >notify the sender by
>>>>> > > >> >> >> >phone or email immediately and delete it!
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >________________________________________
>>>>> > > >> >> >> >From: Shi, Shaofeng [[email protected]]
>>>>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM
>>>>> > > >> >> >> >To: [email protected]
>>>>> > > >> >> >> >Cc: Kulbhushan Rana
>>>>> > > >> >> >> >Subject: Re: Error while making cube & Measure option
>>>>>is
>>>>>not
>>>>> > > >> >>responding
>>>>> > > >> >> >> >on GUI
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >Hi Santosh, hive table importing issue should not
>>>>>impact
>>>>>on
>>>>> > cube
>>>>> > > >> >> >>saving.
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >If you couldn’t save the cube, firstly please check
>>>>>whether
>>>>> > > >>there is
>>>>> > > >> >> >>error
>>>>> > > >> >> >> >in the tomcat’s log; If not please check your web
>>>>>browser; We
>>>>> > > >> >>suggest
>>>>> > > >> >> >>use
>>>>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS
>>>>>console
>>>>> > > >>(press
>>>>> > > >> >> >>F12)
>>>>> > > >> >> >> >and then operate web UI, check whether there is any
>>>>>error
>>>>> > > >>reported
>>>>> > > >> >>in
>>>>> > > >> >> >> >browser.
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
>>>>> > > >><[email protected]
>>>>> > > >> >
>>>>> > > >> >> >> >wrote:
>>>>> > > >> >> >> >
>>>>> > > >> >> >> >>Hi Shaofeng ,
>>>>> > > >> >> >> >> Thanks for replying.
>>>>> > > >> >> >> >> Yes I am checking the yarn exception, But I find
>>>>>that
>>>>> this
>>>>> > > >>error
>>>>> > > >> >> >>comes
>>>>> > > >> >> >> >>while importing the hive table to kylin.
>>>>> > > >> >> >> >> Even if this error comes , hive tables is exported
>>>>> > > >>successfully
>>>>> > > >> >>in
>>>>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ?
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >> Next step when I go on creating the cube for
>>>>>following
>>>>> > > >>schema I
>>>>> > > >> >> >>get
>>>>> > > >> >> >> >>error at last step while saving and I am unable to
>>>>>add
>>>>>any
>>>>> > > >> >>measures ,
>>>>> > > >> >> >> >>clicking on measure option just dont pop up any
>>>>>dialog,
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>I am using star schema with fact_sales as fact table
>>>>>and
>>>>> dim_*
>>>>> > > >>as
>>>>> > > >> >> >> >>dimension tables.
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >> fact_sales:
>>>>> > > >> >> >> >> storeid int
>>>>> > > >> >> >> >> itemid int
>>>>> > > >> >> >> >> custid int
>>>>> > > >> >> >> >> qty int
>>>>> > > >> >> >> >> price double
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>dim_customer
>>>>> > > >> >> >> >> custid int
>>>>> > > >> >> >> >> name string
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >> dim_item
>>>>> > > >> >> >> >> itemid int
>>>>> > > >> >> >> >> category string
>>>>> > > >> >> >> >> brand string
>>>>> > > >> >> >> >> color string
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>dim_store
>>>>> > > >> >> >> >> storeid int
>>>>> > > >> >> >> >> city string
>>>>> > > >> >> >> >> state string
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>JSON is as below.
>>>>> > > >> >> >> >> The JSON is as below.
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "name": "Retail_Cube",
>>>>> > > >> >> >> >> "description": "",
>>>>> > > >> >> >> >> "dimensions": [
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID",
>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >> "hierarchy": false,
>>>>> > > >> >> >> >> "derived": null,
>>>>> > > >> >> >> >> "column": [
>>>>> > > >> >> >> >> "STOREID"
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "id": 1
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID",
>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >> "hierarchy": false,
>>>>> > > >> >> >> >> "derived": null,
>>>>> > > >> >> >> >> "column": [
>>>>> > > >> >> >> >> "ITEMID"
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "id": 2
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID",
>>>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >> "hierarchy": false,
>>>>> > > >> >> >> >> "derived": null,
>>>>> > > >> >> >> >> "column": [
>>>>> > > >> >> >> >> "CUSTID"
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "id": 3
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "measures": [
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "id": 1,
>>>>> > > >> >> >> >> "name": "_COUNT_",
>>>>> > > >> >> >> >> "function": {
>>>>> > > >> >> >> >> "expression": "COUNT",
>>>>> > > >> >> >> >> "returntype": "bigint",
>>>>> > > >> >> >> >> "parameter": {
>>>>> > > >> >> >> >> "type": "constant",
>>>>> > > >> >> >> >> "value": 1
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "rowkey": {
>>>>> > > >> >> >> >> "rowkey_columns": [
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "column": "STOREID",
>>>>> > > >> >> >> >> "length": 0,
>>>>> > > >> >> >> >> "dictionary": "true",
>>>>> > > >> >> >> >> "mandatory": false
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "column": "ITEMID",
>>>>> > > >> >> >> >> "length": 0,
>>>>> > > >> >> >> >> "dictionary": "true",
>>>>> > > >> >> >> >> "mandatory": false
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "column": "CUSTID",
>>>>> > > >> >> >> >> "length": 0,
>>>>> > > >> >> >> >> "dictionary": "true",
>>>>> > > >> >> >> >> "mandatory": false
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> ],
>>>>> > > >> >> >> >> "aggregation_groups": [
>>>>> > > >> >> >> >> [
>>>>> > > >> >> >> >> "STOREID",
>>>>> > > >> >> >> >> "ITEMID",
>>>>> > > >> >> >> >> "CUSTID"
>>>>> > > >> >> >> >> ]
>>>>> > > >> >> >> >> ]
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> "notify_list": [],
>>>>> > > >> >> >> >> "capacity": "",
>>>>> > > >> >> >> >> "hbase_mapping": {
>>>>> > > >> >> >> >> "column_family": [
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "name": "f1",
>>>>> > > >> >> >> >> "columns": [
>>>>> > > >> >> >> >> {
>>>>> > > >> >> >> >> "qualifier": "m",
>>>>> > > >> >> >> >> "measure_refs": [
>>>>> > > >> >> >> >> "_COUNT_"
>>>>> > > >> >> >> >> ]
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> ]
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >> ]
>>>>> > > >> >> >> >> },
>>>>> > > >> >> >> >> "project": "RetailProject",
>>>>> > > >> >> >> >> "model_name": "Retail_Cube"
>>>>> > > >> >> >> >> }
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>Regards,
>>>>> > > >> >> >> >>Santosh Akhilesh
>>>>> > > >> >> >> >>Bangalore R&D
>>>>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>www.huawei.com
>>>>> > > >> >> >>
>>>>>>>----------------------------------------------------------
>>>>> > > >> >> >> ---------------
>>>>> > > >> >> >> >>-
>>>>> > > >> >> >>
>>>>>>>-----------------------------------------------------------
>>>>> > > >> >> >> >>This e-mail and its attachments contain confidential
>>>>> > information
>>>>> > > >> >>from
>>>>> > > >> >> >> >>HUAWEI, which
>>>>> > > >> >> >> >>is intended only for the person or entity whose
>>>>>address
>>>>>is
>>>>> > > >>listed
>>>>> > > >> >> >>above.
>>>>> > > >> >> >> >>Any use of the
>>>>> > > >> >> >> >>information contained herein in any way (including,
>>>>>but
>>>>>not
>>>>> > > >>limited
>>>>> > > >> >> >>to,
>>>>> > > >> >> >> >>total or partial
>>>>> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons
>>>>>other
>>>>> > > >>than
>>>>> > > >> >>the
>>>>> > > >> >> >> >>intended
>>>>> > > >> >> >> >>recipient(s) is prohibited. If you receive this e-mail
>>>>>in
>>>>> > error,
>>>>> > > >> >> >>please
>>>>> > > >> >> >> >>notify the sender by
>>>>> > > >> >> >> >>phone or email immediately and delete it!
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>________________________________________
>>>>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]]
>>>>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM
>>>>> > > >> >> >> >>To: [email protected]
>>>>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option
>>>>>is
>>>>>not
>>>>> > > >> >> >>responding
>>>>> > > >> >> >> >>on GUI
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>Hi Santosh,
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>It looks like hadoop failed to execute some shell
>>>>>command in
>>>>> > the
>>>>> > > >> >> >> >>container; You need dive into hadoop to see what¹s the
>>>>> > concrete
>>>>> > > >> >>error.
>>>>> > > >> >> >> >>You
>>>>> > > >> >> >> >>can use yarn logs command to fetch all logs:
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>yarn logs -applicationId <app_id>
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
>>>>> > > >><[email protected]
>>>>> > > >> >
>>>>> > > >> >> >> >>wrote:
>>>>> > > >> >> >> >>
>>>>> > > >> >> >> >>>Hi Luke / Shaofeng ,
>>>>> > > >> >> >> >>> Can you please help me to check this
>>>>>issue.
>>>>> > > >> >> >> >>>Regards,
>>>>> > > >> >> >> >>>Santosh Akhilesh
>>>>> > > >> >> >> >>>
>>>>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>>>> > > >> >> >> >>>[email protected]> wrote:
>>>>> > > >> >> >> >>>
>>>>> > > >> >> >> >>>> Hi All ,
>>>>> > > >> >> >> >>>> is it because of following error in map
>>>>>reduce
>>>>> job
>>>>> > ?
>>>>> > > >> >>what
>>>>> > > >> >> >> >>>>could
>>>>> > > >> >> >> >>>>be
>>>>> > > >> >> >> >>>> way to resolve this , a google search says that its
>>>>>issue
>>>>> > of
>>>>> > > >> >>Yarn
>>>>> > > >> >> >> >>>>class
>>>>> > > >> >> >> >>>> path , but I am not sure what it is ?
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job
>>>>>table=RETAIL.FACT_SALES
>>>>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2
>>>>>times
>>>>> > due
>>>>> > > >> >>to AM
>>>>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002
>>>>>exited
>>>>> > > >>with
>>>>> > > >> >> >> >>>>exitCode: 1
>>>>> > > >> >> >> >>>> For more detailed output, check application
>>>>>tracking
>>>>> page:
>>>>> > > >> >> >> >>>>
>>>>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then,
>>>>> > > >> >> >>click
>>>>> > > >> >> >> >>>>on
>>>>> > > >> >> >> >>>> links to logs of each attempt.
>>>>> > > >> >> >> >>>> Diagnostics: Exception from container-launch.
>>>>> > > >> >> >> >>>> Container id:
>>>>>container_1424791969399_0008_02_000001
>>>>> > > >> >> >> >>>> Exit code: 1
>>>>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1:
>>>>> > > >> >> >> >>>> at
>>>>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>>>> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
>>>>> > > >> >> >> e(Shell.java:71
>>>>> > > >> >> >> >>>>5
>>>>> > > >> >> >> >>>>)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
>>>>> > > >> >> >> erExecutor.laun
>>>>> > > >> >> >> >>>>c
>>>>> > > >> >> >> >>>>h
>>>>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>>>> > > >> >> >> er.launcher.Con
>>>>> > > >> >> >> >>>>t
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>>>> > > >> >> >> er.launcher.Con
>>>>> > > >> >> >> >>>>t
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
>>>>> > > >> >> >> >>>> at
>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>>>> > > >> >> >> PoolExecutor.ja
>>>>> > > >> >> >> >>>>v
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>:1145)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>>>> > > >> >> >> dPoolExecutor.j
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>v
>>>>> > > >> >> >> >>>>a:615)
>>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745)
>>>>> > > >> >> >> >>>> Container exited with a non-zero exit code 1
>>>>> > > >> >> >> >>>> Failing this attempt. Failing the application.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> ---------- Forwarded message ----------
>>>>> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]>
>>>>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure
>>>>>option
>>>>>is
>>>>> > not
>>>>> > > >> >> >> >>>>responding
>>>>> > > >> >> >> >>>>on
>>>>> > > >> >> >> >>>> GUI
>>>>> > > >> >> >> >>>> To: "[email protected]"
>>>>> > > >> >> >><[email protected]>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> hi ,
>>>>> > > >> >> >> >>>> please someone give me a hand to resolve this
>>>>>issue ,
>>>>> > > >>thanks.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Regards,
>>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>>> > > >> >> >> >>>> Bangalore R&D
>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> www.huawei.com
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>--------------------------------------------------------
>>>>> > > >> >> >> ---------------
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >>
>>>>> >>>>------------------------------------------------------------
>>>>> > > >> >> >> >>>> This e-mail and its attachments contain
>>>>>confidential
>>>>> > > >>information
>>>>> > > >> >> >>from
>>>>> > > >> >> >> >>>> HUAWEI, which
>>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>>address
>>>>> is
>>>>> > > >> >>listed
>>>>> > > >> >> >> >>>>above.
>>>>> > > >> >> >> >>>> Any use of the
>>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>>but
>>>>> not
>>>>> > > >> >>limited
>>>>> > > >> >> >> >>>>to,
>>>>> > > >> >> >> >>>> total or partial
>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>>persons
>>>>> > other
>>>>> > > >> >>than
>>>>> > > >> >> >>the
>>>>> > > >> >> >> >>>> intended
>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>>e-mail in
>>>>> > > >>error,
>>>>> > > >> >> >> >>>>please
>>>>> > > >> >> >> >>>> notify the sender by
>>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> ________________________________________
>>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]]
>>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>>>> > > >> >> >> >>>> To: [email protected]
>>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana
>>>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure
>>>>>option
>>>>>is
>>>>> > not
>>>>> > > >> >> >> >>>>responding
>>>>> > > >> >> >> >>>>on
>>>>> > > >> >> >> >>>> GUI
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the
>>>>>cube
>>>>>I
>>>>> get
>>>>> > > >>an
>>>>> > > >> >> >> >>>>exception
>>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
>>>>>correctly
>>>>> and
>>>>> > > >> >> >> >>>>HCatInputFormat
>>>>> > > >> >> >> >>>> this file is present in
>>>>>hive-hcatalog-core-0.14.0.jar
>>>>>.
>>>>> > > >>Please
>>>>> > > >> >>let
>>>>> > > >> >> >>me
>>>>> > > >> >> >> >>>>know
>>>>> > > >> >> >> >>>> what can I do to resolve this ?
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> -- This was path issue , now no more exception in
>>>>> > kylin.log
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> But saveing cube still fails with error. And still
>>>>>can't
>>>>> > add
>>>>> > > >> >> >>measures.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Error Message
>>>>> > > >> >> >> >>>> Failed to take action.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> In log I can find no exception. Following is the
>>>>>last
>>>>>log
>>>>> > in
>>>>> > > >> >> >>kylin.log
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
>>>>> > > >> >> >> .DefaultSchedul
>>>>> > > >> >> >> >>>>e
>>>>> > > >> >> >> >>>>r
>>>>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>>>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0
>>>>>ready,
>>>>>6
>>>>> > others
>>>>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
>>>>> > > >> >> >> beController.de
>>>>> > > >> >> >> >>>>s
>>>>> > > >> >> >> >>>>e
>>>>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)]
>>>>> > > >> >> >> >>>> - Saving cube {
>>>>> > > >> >> >> >>>> "name": "",
>>>>> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >>>> "lookups": [],
>>>>> > > >> >> >> >>>> "filter_condition": "",
>>>>> > > >> >> >> >>>> "capacity": "SMALL",
>>>>> > > >> >> >> >>>> "partition_desc": {
>>>>> > > >> >> >> >>>> "partition_date_column": "",
>>>>> > > >> >> >> >>>> "partition_date_start": 0,
>>>>> > > >> >> >> >>>> "partition_type": "APPEND"
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> "last_modified": 0
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> local access logs all with 200 , so seems ok.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800]
>>>>>"POST
>>>>> > > >> >> >>/kylin/api/cubes
>>>>> > > >> >> >> >>>> HTTP/1.1" 200 701
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Regards,
>>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>>> > > >> >> >> >>>> Bangalore R&D
>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> www.huawei.com
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>--------------------------------------------------------
>>>>> > > >> >> >> ---------------
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >>
>>>>> >>>>------------------------------------------------------------
>>>>> > > >> >> >> >>>> This e-mail and its attachments contain
>>>>>confidential
>>>>> > > >>information
>>>>> > > >> >> >>from
>>>>> > > >> >> >> >>>> HUAWEI, which
>>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>>address
>>>>> is
>>>>> > > >> >>listed
>>>>> > > >> >> >> >>>>above.
>>>>> > > >> >> >> >>>> Any use of the
>>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>>but
>>>>> not
>>>>> > > >> >>limited
>>>>> > > >> >> >> >>>>to,
>>>>> > > >> >> >> >>>> total or partial
>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>>persons
>>>>> > other
>>>>> > > >> >>than
>>>>> > > >> >> >>the
>>>>> > > >> >> >> >>>> intended
>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>>e-mail in
>>>>> > > >>error,
>>>>> > > >> >> >> >>>>please
>>>>> > > >> >> >> >>>> notify the sender by
>>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> ________________________________________
>>>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]]
>>>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>>>> > > >> >> >> >>>> To: [email protected]
>>>>> > > >> >> >> >>>> Cc: Kulbhushan Rana
>>>>> > > >> >> >> >>>> Subject: Error while making cube & Measure option
>>>>>is
>>>>>not
>>>>> > > >> >> >>responding on
>>>>> > > >> >> >> >>>>GUI
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Hi All ,
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> I am building a simple cube for test and using
>>>>>the
>>>>> > binary
>>>>> > > >> >>build
>>>>> > > >> >> >> >>>>0.7.1
>>>>> > > >> >> >> >>>> . I have following hive tables with columns.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> fact_sales:
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> storeid int
>>>>> > > >> >> >> >>>> itemid int
>>>>> > > >> >> >> >>>> custid int
>>>>> > > >> >> >> >>>> qty int
>>>>> > > >> >> >> >>>> price double
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> dim_customer
>>>>> > > >> >> >> >>>> custid int
>>>>> > > >> >> >> >>>> name string
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> dim_item
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> itemid int
>>>>> > > >> >> >> >>>> category string
>>>>> > > >> >> >> >>>> brand string
>>>>> > > >> >> >> >>>> color string
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> dim_store
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> storeid int
>>>>> > > >> >> >> >>>> city string
>>>>> > > >> >> >> >>>> state string
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Please help me to answer following issues;
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> 1. When I go to measure section and click on
>>>>>measure
>>>>> > option ,
>>>>> > > >> >> >>there is
>>>>> > > >> >> >> >>>>no
>>>>> > > >> >> >> >>>> response , I want add measure on qty and price with
>>>>>sum
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the
>>>>>cube
>>>>>I
>>>>> get
>>>>> > > >>an
>>>>> > > >> >> >> >>>>exception
>>>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
>>>>>correctly
>>>>> and
>>>>> > > >> >> >> >>>>HCatInputFormat
>>>>> > > >> >> >> >>>> this file is present in
>>>>>hive-hcatalog-core-0.14.0.jar
>>>>>.
>>>>> > > >>Please
>>>>> > > >> >>let
>>>>> > > >> >> >>me
>>>>> > > >> >> >> >>>>know
>>>>> > > >> >> >> >>>> what can I do to resolve this ?
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> 3. Also I have another question since this is a
>>>>>test
>>>>>and
>>>>> > > >>data is
>>>>> > > >> >> >>small
>>>>> > > >> >> >> >>>>I
>>>>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to
>>>>>skip
>>>>> > > >>partition
>>>>> > > >> >> >>stage
>>>>> > > >> >> >> >>>> while cube build ?
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Exception
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
>>>>> > > >> >> >> l.DefaultSchedu
>>>>> > > >> >> >> >>>>l
>>>>> > > >> >> >> >>>>e
>>>>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>>>> > > >> >> >> >>>> - ExecuteException
>>>>> job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError:
>>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut
>>>>> > > >> >> >> e(AbstractExecu
>>>>> > > >> >> >> >>>>t
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>ble.java:102)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
>>>>> > > >> >> >> bRunner.run(Def
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>u
>>>>> > > >> >> >> >>>>ltScheduler.java:132)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>>>> > > >> >> >> PoolExecutor.ja
>>>>> > > >> >> >> >>>>v
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>:1145)
>>>>> > > >> >> >> >>>> at
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>>>> > > >> >> >> dPoolExecutor.j
>>>>> > > >> >> >> >>>>a
>>>>> > > >> >> >> >>>>v
>>>>> > > >> >> >> >>>>a:615)
>>>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745)
>>>>> > > >> >> >> >>>> Caused by:
>>>>> org.apache.kylin.job.exception.ExecuteException:
>>>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError:
>>>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> The JSON is as below.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "name": "Retail_Cube",
>>>>> > > >> >> >> >>>> "description": "",
>>>>> > > >> >> >> >>>> "dimensions": [
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID",
>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >>>> "hierarchy": false,
>>>>> > > >> >> >> >>>> "derived": null,
>>>>> > > >> >> >> >>>> "column": [
>>>>> > > >> >> >> >>>> "STOREID"
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "id": 1
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID",
>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >>>> "hierarchy": false,
>>>>> > > >> >> >> >>>> "derived": null,
>>>>> > > >> >> >> >>>> "column": [
>>>>> > > >> >> >> >>>> "ITEMID"
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "id": 2
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID",
>>>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES",
>>>>> > > >> >> >> >>>> "hierarchy": false,
>>>>> > > >> >> >> >>>> "derived": null,
>>>>> > > >> >> >> >>>> "column": [
>>>>> > > >> >> >> >>>> "CUSTID"
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "id": 3
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "measures": [
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "id": 1,
>>>>> > > >> >> >> >>>> "name": "_COUNT_",
>>>>> > > >> >> >> >>>> "function": {
>>>>> > > >> >> >> >>>> "expression": "COUNT",
>>>>> > > >> >> >> >>>> "returntype": "bigint",
>>>>> > > >> >> >> >>>> "parameter": {
>>>>> > > >> >> >> >>>> "type": "constant",
>>>>> > > >> >> >> >>>> "value": 1
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "rowkey": {
>>>>> > > >> >> >> >>>> "rowkey_columns": [
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "column": "STOREID",
>>>>> > > >> >> >> >>>> "length": 0,
>>>>> > > >> >> >> >>>> "dictionary": "true",
>>>>> > > >> >> >> >>>> "mandatory": false
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "column": "ITEMID",
>>>>> > > >> >> >> >>>> "length": 0,
>>>>> > > >> >> >> >>>> "dictionary": "true",
>>>>> > > >> >> >> >>>> "mandatory": false
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "column": "CUSTID",
>>>>> > > >> >> >> >>>> "length": 0,
>>>>> > > >> >> >> >>>> "dictionary": "true",
>>>>> > > >> >> >> >>>> "mandatory": false
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> ],
>>>>> > > >> >> >> >>>> "aggregation_groups": [
>>>>> > > >> >> >> >>>> [
>>>>> > > >> >> >> >>>> "STOREID",
>>>>> > > >> >> >> >>>> "ITEMID",
>>>>> > > >> >> >> >>>> "CUSTID"
>>>>> > > >> >> >> >>>> ]
>>>>> > > >> >> >> >>>> ]
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> "notify_list": [],
>>>>> > > >> >> >> >>>> "capacity": "",
>>>>> > > >> >> >> >>>> "hbase_mapping": {
>>>>> > > >> >> >> >>>> "column_family": [
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "name": "f1",
>>>>> > > >> >> >> >>>> "columns": [
>>>>> > > >> >> >> >>>> {
>>>>> > > >> >> >> >>>> "qualifier": "m",
>>>>> > > >> >> >> >>>> "measure_refs": [
>>>>> > > >> >> >> >>>> "_COUNT_"
>>>>> > > >> >> >> >>>> ]
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> ]
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>> ]
>>>>> > > >> >> >> >>>> },
>>>>> > > >> >> >> >>>> "project": "RetailProject",
>>>>> > > >> >> >> >>>> "model_name": "Retail_Cube"
>>>>> > > >> >> >> >>>> }
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> Regards,
>>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>>> > > >> >> >> >>>> Bangalore R&D
>>>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> www.huawei.com
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >>
>>>>>>>>>--------------------------------------------------------
>>>>> > > >> >> >> ---------------
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >> >>>>-
>>>>> > > >> >> >>
>>>>> >>>>------------------------------------------------------------
>>>>> > > >> >> >> >>>> This e-mail and its attachments contain
>>>>>confidential
>>>>> > > >>information
>>>>> > > >> >> >>from
>>>>> > > >> >> >> >>>> HUAWEI, which
>>>>> > > >> >> >> >>>> is intended only for the person or entity whose
>>>>>address
>>>>> is
>>>>> > > >> >>listed
>>>>> > > >> >> >> >>>>above.
>>>>> > > >> >> >> >>>> Any use of the
>>>>> > > >> >> >> >>>> information contained herein in any way (including,
>>>>>but
>>>>> not
>>>>> > > >> >>limited
>>>>> > > >> >> >> >>>>to,
>>>>> > > >> >> >> >>>> total or partial
>>>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
>>>>>persons
>>>>> > other
>>>>> > > >> >>than
>>>>> > > >> >> >>the
>>>>> > > >> >> >> >>>> intended
>>>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
>>>>>e-mail in
>>>>> > > >>error,
>>>>> > > >> >> >> >>>>please
>>>>> > > >> >> >> >>>> notify the sender by
>>>>> > > >> >> >> >>>> phone or email immediately and delete it!
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>> --
>>>>> > > >> >> >> >>>> Regards,
>>>>> > > >> >> >> >>>> Santosh Akhilesh
>>>>> > > >> >> >> >>>> +91-0-9845482201
>>>>> > > >> >> >> >>>>
>>>>> > > >> >> >> >>>
>>>>> > > >> >> >> >>>
>>>>> > > >> >> >> >>>
>>>>> > > >> >> >> >>>--
>>>>> > > >> >> >> >>>Regards,
>>>>> > > >> >> >> >>>Santosh Akhilesh
>>>>> > > >> >> >> >>>+91-0-9845482201
>>>>> > > >> >> >>
>>>>> > > >> >> >>
>>>>> > > >> >>
>>>>> > > >> >>
>>>>> > > >>
>>>>> > > >>
>>>>> > >
>>>>> > >
>>>>> >
>>>>> >
>>>>> > --
>>>>> > Regards,
>>>>> > Santosh Akhilesh
>>>>> > +91-0-9845482201
>>>>> >
>>>>>
>>>>
>>>>
>>>>
>>>>--
>>>>Regards,
>>>>Santosh Akhilesh
>>>>+91-0-9845482201