Did you figure out the exception of "No enum constant org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES” ? Is it still be thrown in the logs? In the last step, Kylin need to parse the MR counters to update cube size; Please refer to https://issues.apache.org/jira/browse/MAPREDUCE-5831 for that error.
On 2/27/15, 5:04 PM, "Santoshakhilesh" <[email protected]> wrote: >Hi Shaofeng , > Cube building is failed at last step while loading Hfile to >Hbase with exception "Can't get cube segment size. >". What could be reason ? > >parameter : -input >/tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >-htablename KYLIN_K27LDMX63W -cubename RetailCube > >Log: > >Start to execute command: > -input /tmp/kylin-17a4606f-905b-4ea1-922a-27c2bfb5c68b/RetailCube/hfile/ >-htablename KYLIN_K27LDMX63W -cubename RetailCube >Command execute return code 0 >Failed with Exception:java.lang.RuntimeException: Can't get cube segment >size. > at >com.kylinolap.job.flow.JobFlowListener.updateCubeSegmentInfoOnSucceed(JobF >lowListener.java:247) > at >com.kylinolap.job.flow.JobFlowListener.jobWasExecuted(JobFlowListener.java >:101) > at >org.quartz.core.QuartzScheduler.notifyJobListenersWasExecuted(QuartzSchedu >ler.java:1985) > at >org.quartz.core.JobRunShell.notifyJobListenersComplete(JobRunShell.java:34 >0) > at org.quartz.core.JobRunShell.run(JobRunShell.java:224) > at >org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5 >73) > >I have checked in hbase shell and following are the tables in hbase; >hbase(main):001:0> list >TABLE > >KYLIN_K27LDMX63W >kylin_metadata_qa >kylin_metadata_qa_acl >kylin_metadata_qa_cube >kylin_metadata_qa_dict >kylin_metadata_qa_invertedindex >kylin_metadata_qa_job >kylin_metadata_qa_job_output >kylin_metadata_qa_proj >kylin_metadata_qa_table_snapshot >kylin_metadata_qa_user >11 row(s) in 0.8990 seconds > > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Santoshakhilesh >Sent: Friday, February 27, 2015 2:15 PM >To: [email protected] >Subject: RE: Error while making cube & Measure option is not responding >on GUI > >I have manually copied the jar to /tmp/kylin , now satge 2 is done , >thanks. > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Shi, Shaofeng [[email protected]] >Sent: Friday, February 27, 2015 1:00 PM >To: [email protected] >Cc: Kulbhushan Rana >Subject: Re: Error while making cube & Measure option is not responding >on GUI > >In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we >renamed the package to “org.apache.kylin.xxx”; When you downgrade to 0.6, >did you also replace the jar location with 0.6 ones in kylin.properties? > >On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]> >wrote: > >>Hi Shaofeng , >> I have added my fact and dimension tables under default database >>of hive. >> Now stage 1 of Cube Build is ok. And there is failure at step2. >> The map reduce job for the finding distinct columns of fact table >>is error. Yarn log is as below. >> Strangely this is class not found error. I have checked the >>Kylin.properties and the jar is already set as below. >>kylin. log has one exception connecting to linux/10.19.93.68 to >>0.0.0.0:10020 >> Please help me to give a clue , I am also trying to check meanwhile >> >>Thanks. >>kylin property >># Temp folder in hdfs >>kylin.hdfs.working.dir=/tmp >># Path to the local(relative to job engine) job jar, job engine will use >>this jar >>kylin.job.jar=/tmp/kylin/kylin-job-latest.jar >> >>Map Reduce error >>---------------------------- >>2015-02-27 20:24:25,262 FATAL [main] org.apache.hadoop.mapred.YarnChild: >>Error running child : java.lang.NoClassDefFoundError: >>com/kylinolap/common/mr/KylinMapper >> at java.lang.ClassLoader.defineClass1(Native Method) >> at java.lang.ClassLoader.defineClass(ClassLoader.java:800) >> at >>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >> at java.net.URLClassLoader.access$100(URLClassLoader.java:71) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >> at java.lang.Class.forName0(Native Method) >> at java.lang.Class.forName(Class.java:274) >> at >>org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.j >>a >>va:2013) >> >>Kylin.log >>QuartzScheduler_Worker-20]:[2015-02-27 >>20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFetch >>e >>r.java:60)] - 0 pending jobs >>[QuartzScheduler_Worker-19]:[2015-02-27 >>20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobC >>o >>unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>failed on connection exception: java.net.ConnectException: Connection >>refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >>com.kylinolap.job.exception.JobException: java.io.IOException: >>java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >>failed on connection exception: java.net.ConnectException: Connection >>refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >> at >>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob. >>j >>ava:289) >> at >>com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdO >>u >>tput.java:162) >> at >>com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.j >>a >>va:85) >> at >>com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86) >> at org.quartz.core.JobRunShell.run(JobRunShell.java:202) >> at >>org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java: >>5 >>73) >>Caused by: java.io.IOException: java.net.ConnectException: Call From >>linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >>java.net.ConnectException: Connection refused; For more details see: >>http://wiki.apache.org/hadoop/ConnectionRefused >> at >>org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelega >>t >>e.java:331) >> at >>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServi >>c >>eDelegate.java:368) >> at >>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) >> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation >>. >>java:1491) >> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >> at >>com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob. >>j >>ava:287) >> ... 5 more >> >>Regards, >>Santosh Akhilesh >>Bangalore R&D >>HUAWEI TECHNOLOGIES CO.,LTD. >> >>www.huawei.com >>------------------------------------------------------------------------- >>- >>----------------------------------------------------------- >>This e-mail and its attachments contain confidential information from >>HUAWEI, which >>is intended only for the person or entity whose address is listed above. >>Any use of the >>information contained herein in any way (including, but not limited to, >>total or partial >>disclosure, reproduction, or dissemination) by persons other than the >>intended >>recipient(s) is prohibited. If you receive this e-mail in error, please >>notify the sender by >>phone or email immediately and delete it! >> >>________________________________________ >>From: Shi, Shaofeng [[email protected]] >>Sent: Friday, February 27, 2015 8:01 AM >>To: [email protected] >>Subject: Re: Error while making cube & Measure option is not responding >>on GUI >> >>In 0.6.x it only support tables in default database, this is a >>limitation; >>The support for non-default tables will be released in 0.7; >> >>To bypass this issue for now, please copy the table to default database >>as >>a workaround; >> >>On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]> >>wrote: >> >>>@Jason >>>thanks , but now as suggested by Saofeng I m not using the inverted >>>index >>>brach as its not stable. >>>I have switched back to 0.6 branch , in this branch yesterday night I >>>could >>>crete the cube successfully but there is issue while building it. I feel >>>that at step 1 of cube build while creating flat table when command is >>>issued to hive if the tables are not under default datbase flat table >>>creation is failed and cube build fails. my fact and dimension tables >>>are >>>under a database called retail. >>> >>>@Saofeng - Can you please confirm this behavior ? Do I need to create >>>the >>>hive tables under default database? >>> >>>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]> >>>wrote: >>> >>>> @Santoshakhilesh >>>> >>>> 1. When I go to measure section and click on measure option , there is >>>>no >>>> response , I want add measure on qty and price with sum >>>> --bug fixed on inverted-index branch >>>> >>>> >>>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh < >>>> [email protected] >>>> > wrote: >>>> >>>> > Hi Shaofeng , >>>> > I have build the 0.6 version and now able to create the cube >>>> > successfully. >>>> > While building the cube , it fails at step1 with following >>>>error. >>>> > Table not found 'DIM_ITEM' >>>> > the table exists , but its under retail data base and not under >>>> > default database. >>>> > does kylin require hive taables to be under default database ? >>>>I >>>>see >>>> > the flat table being created under default database. >>>> > >>>> > Logging initialized using configuration in >>>> > >>>> > >>>> >>>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-commo >>>>n >>>>- >>>>1.0.0.jar!/hive-log4j.properties >>>> > SLF4J: Class path contains multiple SLF4J bindings. >>>> > SLF4J: Found binding in >>>> > >>>> > >>>> >>>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/commo >>>>n >>>>/ >>>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>> > SLF4J: Found binding in >>>> > >>>> > >>>> >>>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jdbc >>>>- >>>>1 >>>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >>>> > explanation. >>>> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] >>>> > OK >>>> > Time taken: 0.964 seconds >>>> > OK >>>> > Time taken: 0.948 seconds >>>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found >>>> > 'DIM_ITEM' >>>> > >>>> > >>>> > >>>> > Command is as below. >>>> > >>>> > hive -e "DROP TABLE IF EXISTS >>>> > >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>>>; >>>> > CREATE EXTERNAL TABLE IF NOT EXISTS >>>> > >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>>> > ( >>>> > STOREID int >>>> > ,ITEMID int >>>> > ,CUSTID int >>>> > ,QTY int >>>> > ,AMOUNT double >>>> > ) >>>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' >>>> > STORED AS SEQUENCEFILE >>>> > LOCATION >>>> > >>>> >>>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_tes >>>>t >>>>_ >>>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44'; >>>> > SET hive.exec.compress.output=true; >>>> > SET hive.auto.convert.join.noconditionaltask = true; >>>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000; >>>> > INSERT OVERWRITE TABLE >>>> > >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>>> > SELECT >>>> > FACT_SALES.STOREID >>>> > ,FACT_SALES.ITEMID >>>> > ,FACT_SALES.CUSTID >>>> > ,FACT_SALES.QTY >>>> > ,FACT_SALES.AMOUNT >>>> > FROM FACT_SALES >>>> > INNER JOIN DIM_STORE >>>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID >>>> > INNER JOIN DIM_ITEM >>>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID >>>> > INNER JOIN DIM_CUSTOMER >>>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID >>>> > ; >>>> > " >>>> > >>>> > >>>> > >>>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]> >>>>wrote: >>>> > >>>> > > The 0.7.1 is test version, its package contains the “snapshot” >>>>suffix; >>>> we >>>> > > will upload a new package there; Luke will also add a message >>>>there >>>>to >>>> > > avoid this confusion; >>>> > > >>>> > > Regarding the problem that you encountered, could you please open >>>>a >>>> JIRA >>>> > > ticket for tracking? Here is link of Apache JIRA: >>>> > > >>>> > > https://issues.apache.org/jira/secure/Dashboard.jspa >>>> > > >>>> > > >>>> > > Thanks for the feedback! >>>> > > >>>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh" >>>><[email protected]> >>>> > > wrote: >>>> > > >>>> > > >Actually I see this being published on kylin webpage. >>>> > > >http://kylin.incubator.apache.org/download/ >>>> > > >I am using 0.7.1 inverted index branch binary distribution. >>>> > > >If this is not stable please give me the link of stable branch I >>>>would >>>> > try >>>> > > >building and testing tonight. >>>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]> >>>> wrote: >>>> > > > >>>> > > >> Hi Santosh, it is not recommended to use the dev code branch >>>> > (actually I >>>> > > >> don’t know how you get the v0.7.x build and what’s the detail >>>> version >>>> > of >>>> > > >> that; each day we submit many changes to that); >>>> > > >> >>>> > > >> The options are 1) switch back to latest release v0.6.5; or 2) >>>>wait >>>> > for >>>> > > >> the formal release of 0.7, that should be in March; Otherwise, >>>>we >>>> > > >>couldn’t >>>> > > >> ensure there is no new problems come out in your next steps; >>>> > > >> >>>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh" >>>><[email protected]> >>>> > > >>wrote: >>>> > > >> >>>> > > >> >Hi Shaofeng >>>> > > >> >So what do you suggest , how should I proceed further with >>>>this >>>> > > >>release? >>>> > > >> >Will there be a patch? Any alternate way I can create cube? >>>> > > >> >Please suggest. >>>> > > >> >Regards >>>> > > >> >Santosh >>>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng >>>><[email protected]> >>>> > > wrote: >>>> > > >> > >>>> > > >> >> Hi Santosh, >>>> > > >> >> >>>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we >>>>have >>>> > > >> >>metadata >>>> > > >> >> structure change; While, the web UI (cube wizard) for this >>>>change >>>> > > >>hasn’t >>>> > > >> >> been stabilized; So it is not strange that you got trouble >>>>when >>>> > > >>saving >>>> > > >> >>the >>>> > > >> >> cube; >>>> > > >> >> >>>> > > >> >> @Jason, any idea about the JS error? >>>> > > >> >> >>>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" < >>>> [email protected] >>>> > > >>>> > > >> >>wrote: >>>> > > >> >> >>>> > > >> >> >Hi Shaofeng, >>>> > > >> >> > >>>> > > >> >> >I am using the binary distribution 0.7.1. I have not been >>>>able >>>> to >>>> > > >>save >>>> > > >> >> >cube >>>> > > >> >> >even once. I have tried creating new project and from local >>>> > machine >>>> > > >>and >>>> > > >> >> >server machine. But I am always stuck with this error. I am >>>> never >>>> > > >> >>allowed >>>> > > >> >> >to add measures and never been able to save the cube. I >>>>also >>>>see >>>> > the >>>> > > >> >> >kylin.log and it always tries to save cube with append >>>>mode. >>>>One >>>> > > >>thing >>>> > > >> >>I >>>> > > >> >> >need to tell that at partition stage since I don't have a >>>>big >>>> fact >>>> > > >> >>table >>>> > > >> >> >now I have not partititioned the fact table and I skip this >>>> step. >>>> > > >>Does >>>> > > >> >> >this >>>> > > >> >> >have affect in saving the cube. Is this because some >>>>metadata is >>>> > > >> >>available >>>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop >>>>2.6.6. >>>> > Yes >>>> > > >> >>kylin >>>> > > >> >> >propert I have not added the jar. I will add them and >>>>check. >>>>But >>>> > > >>cube >>>> > > >> >> >creation failure is really puzzling me. I could see no >>>>error >>>> logs >>>> > in >>>> > > >> >> >kylin.log. >>>> > > >> >> >Regards >>>> > > >> >> >Santosh >>>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng >>>><[email protected] >>>> > >>>> > > >> wrote: >>>> > > >> >> > >>>> > > >> >> >> Which version or code branch are you using? I assume >>>>you’re >>>> > using >>>> > > >>the >>>> > > >> >> >> stable version from master; Seems you’re trying to edit >>>>an >>>> > > >>existing >>>> > > >> >>cube >>>> > > >> >> >> to add new measurement, try refresh your browser's cache; >>>>if >>>> it >>>> > > >>still >>>> > > >> >> >> couldn’t be saved, try to create a new cube; >>>> > > >> >> >> >>>> > > >> >> >> The two error traces in tomcat need be taken care: >>>> > > >> >> >> >>>> > > >> >> >> 1) java.lang.NoClassDefFoundError: >>>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper >>>> > > >> >> >> Please check kylin.properties file, making sure >>>>the >>>> > > >> >> >>“kylin.job.jar” >>>> > > >> >> >> points to a right jar file; It will be loaded in >>>>Map-reduce; >>>> > > >> >> >> >>>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant >>>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>> > > >> >> >> This indicates your hadoop version might be old; >>>> Please >>>> > > >>check >>>> > > >> >> >>and >>>> > > >> >> >> ensure >>>> > > >> >> >> hadoop version is 2.2 or above. >>>> > > >> >> >> >>>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" >>>> > > >><[email protected]> >>>> > > >> >> >> wrote: >>>> > > >> >> >> >>>> > > >> >> >> >Hi Shaofeng , >>>> > > >> >> >> > >>>> > > >> >> >> > I am using chrome , When I click on button to add >>>> measures , >>>> > > >> >> >>following >>>> > > >> >> >> >is error on chrome console. When I try to save the cube >>>>there >>>> > is >>>> > > >>no >>>> > > >> >> >>error >>>> > > >> >> >> >in console. I just get a error dialog saying failed to >>>>take >>>> > > >>action >>>> > > >> >>and >>>> > > >> >> >> >gives me the JSON cube schema. >>>> > > >> >> >> > >>>> > > >> >> >> >Error on chrome debug console is as below; >>>> > > >> >> >> > >>>> > > >> >> >> > ReferenceError: CubeDescModel is not defined >>>> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >>>> > > >> >> >> > at scripts.min.0.js:180 >>>> > > >> >> >> > at scripts.min.0.js:197 >>>> > > >> >> >> > at h.$eval (scripts.min.0.js:119) >>>> > > >> >> >> > at h.$apply (scripts.min.0.js:119) >>>> > > >> >> >> > at HTMLButtonElement.<anonymous> >>>>(scripts.min.0.js:197) >>>> > > >> >> >> > at HTMLButtonElement.m.event.dispatch >>>> (scripts.min.0.js:3) >>>> > > >> >> >> > at HTMLButtonElement.r.handle >>>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >>>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous >>>> > > >> >>function)scripts.min.0.js:119 >>>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous >>>> > > >>function)scripts.min.0.js:3 >>>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle >>>> > > >> >> >> > >>>> > > >> >> >> > About the hive table import , I got pass the run >>>>shell >>>> > command >>>> > > >> >> >> >exception but it still fails the hadoop log is; >>>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main] >>>>org.apache.hadoop.mapred. >>>> > > >> >> YarnChild: >>>> > > >> >> >> >mapreduce.cluster.local.dir for child: >>>> > > >> >> >> >>>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >>>> > > >> >> >> cation_14249530 >>>> > > >> >> >> >91340_0002 >>>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main] >>>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation: >>>>session.id >>>> > is >>>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id >>>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main] >>>> > > >>org.apache.hadoop.mapred.Task: >>>> > > >> >> >>Using >>>> > > >> >> >> >ResourceCalculatorProcessTree : [ ] >>>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main] >>>> > > >> >> >>org.apache.hadoop.mapred.YarnChild: >>>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError: >>>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper >>>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method) >>>> > > >> >> >> > at >>>>java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>>> > > >> >> >> > at >>>> > > >> >> >> >java.security.SecureClassLoader.defineClass( >>>> > > >> >> SecureClassLoader.java:142) >>>> > > >> >> >> > at >>>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>>> > > >> >> >> > at >>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>>> > > >> >> >> > at >>>>java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>> > > >> >> >> > at >>>>java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>> Method) >>>> > > >> >> >> > at >>>> java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>> > > >> >> >> > at >>>>java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>> > > >> >> >> > at >>>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>> > > >> >> >> > >>>> > > >> >> >> >tomcat logs: >>>> > > >> >> >> >usage: HiveColumnCardinalityJob >>>> > > >> >> >> > -output <path> Output path >>>> > > >> >> >> > -table <table name> The hive table name >>>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26 >>>> > > >> >> >> >>>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >>>> > > >> >> >> lExecutable.doW >>>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute >>>> > > >> >> >> >>>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >>>> > > >> >> >> >name=null, state=RUNNING} >>>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant >>>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>> > > >>>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf( >>>> > > >> >> >> Framewo >>>> > > >> >> >> >rkCounterGroup.java:148) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >>>> > > >> >> >> findCounter(Fra >>>> > > >> >> >> >meworkCounterGroup.java:182) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >>>> > > >> >> >> ounter(Abstract >>>> > > >> >> >> >Counters.java:154) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >>>> > > >> >> >> erter.java:240) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >>>> > > >> >> >> rs(ClientServic >>>> > > >> >> >> >eDelegate.java:370) >>>> > > >> >> >> > at >>>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters( >>>> > > >> >> YARNRunner.java:511) >>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>>> Method) >>>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >>>> > > >> >> >> oupInformation. >>>> > > >> >> >> >java:1491) >>>> > > >> >> >> > at >>>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>>> > > >> >> >> > at >>>> > > >> >>>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >>>> > > >> >> >> > at >>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job. >>>> > > >> java:1289) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >>>> > > >> >> >> tion(AbstractHa >>>> > > >> >> >> >doopJob.java:134) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>> > > >> >>>> > >>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run >>>>>>>( >>>> > > >> >> >> HiveC >>>> > > >> >> >> >olumnCardinalityJob.java:114) >>>> > > >> >> >> > at >>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> > > >> >> >> > at >>>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >>>> > > >> >> >> doopShellExecut >>>> > > >> >> >> >able.java:62) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>> > > >> >> >> bstractExecutab >>>> > > >> >> >> >le.java:99) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW >>>> > > >> >> >> ork(DefaultChai >>>> > > >> >> >> >nedExecutable.java:50) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>>> > > >> >> >> bstractExecutab >>>> > > >> >> >> >le.java:99) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >>>> > > >> >> >> nner.run(Defaul >>>> > > >> >> >> >tScheduler.java:132) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >>>> > > >> >> >> lExecutor.java: >>>> > > >> >> >> >1145) >>>> > > >> >> >> > at >>>> > > >> >> >> >>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >>>> > > >> >> >> olExecutor.java >>>> > > >> >> >> >:615) >>>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745) >>>> > > >> >> >> > >>>> > > >> >> >> >Regards, >>>> > > >> >> >> >Santosh Akhilesh >>>> > > >> >> >> >Bangalore R&D >>>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD. >>>> > > >> >> >> > >>>> > > >> >> >> >www.huawei.com >>>> > > >> >> >> >>>>>----------------------------------------------------------- >>>> > > >> >> >> --------------- >>>> > > >> >> >> >>>>>----------------------------------------------------------- >>>> > > >> >> >> >This e-mail and its attachments contain confidential >>>> > information >>>> > > >> >>from >>>> > > >> >> >> >HUAWEI, which >>>> > > >> >> >> >is intended only for the person or entity whose address >>>>is >>>> > listed >>>> > > >> >> >>above. >>>> > > >> >> >> >Any use of the >>>> > > >> >> >> >information contained herein in any way (including, but >>>>not >>>> > > >>limited >>>> > > >> >>to, >>>> > > >> >> >> >total or partial >>>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons >>>>other >>>> > than >>>> > > >> >>the >>>> > > >> >> >> >intended >>>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail >>>>in >>>> > error, >>>> > > >> >>please >>>> > > >> >> >> >notify the sender by >>>> > > >> >> >> >phone or email immediately and delete it! >>>> > > >> >> >> > >>>> > > >> >> >> >________________________________________ >>>> > > >> >> >> >From: Shi, Shaofeng [[email protected]] >>>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM >>>> > > >> >> >> >To: [email protected] >>>> > > >> >> >> >Cc: Kulbhushan Rana >>>> > > >> >> >> >Subject: Re: Error while making cube & Measure option is >>>>not >>>> > > >> >>responding >>>> > > >> >> >> >on GUI >>>> > > >> >> >> > >>>> > > >> >> >> >Hi Santosh, hive table importing issue should not impact >>>>on >>>> > cube >>>> > > >> >> >>saving. >>>> > > >> >> >> > >>>> > > >> >> >> >If you couldn’t save the cube, firstly please check >>>>whether >>>> > > >>there is >>>> > > >> >> >>error >>>> > > >> >> >> >in the tomcat’s log; If not please check your web >>>>browser; We >>>> > > >> >>suggest >>>> > > >> >> >>use >>>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS >>>>console >>>> > > >>(press >>>> > > >> >> >>F12) >>>> > > >> >> >> >and then operate web UI, check whether there is any >>>>error >>>> > > >>reported >>>> > > >> >>in >>>> > > >> >> >> >browser. >>>> > > >> >> >> > >>>> > > >> >> >> > >>>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" >>>> > > >><[email protected] >>>> > > >> > >>>> > > >> >> >> >wrote: >>>> > > >> >> >> > >>>> > > >> >> >> >>Hi Shaofeng , >>>> > > >> >> >> >> Thanks for replying. >>>> > > >> >> >> >> Yes I am checking the yarn exception, But I find >>>>that >>>> this >>>> > > >>error >>>> > > >> >> >>comes >>>> > > >> >> >> >>while importing the hive table to kylin. >>>> > > >> >> >> >> Even if this error comes , hive tables is exported >>>> > > >>successfully >>>> > > >> >>in >>>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ? >>>> > > >> >> >> >> >>>> > > >> >> >> >> Next step when I go on creating the cube for >>>>following >>>> > > >>schema I >>>> > > >> >> >>get >>>> > > >> >> >> >>error at last step while saving and I am unable to add >>>>any >>>> > > >> >>measures , >>>> > > >> >> >> >>clicking on measure option just dont pop up any dialog, >>>> > > >> >> >> >> >>>> > > >> >> >> >>I am using star schema with fact_sales as fact table >>>>and >>>> dim_* >>>> > > >>as >>>> > > >> >> >> >>dimension tables. >>>> > > >> >> >> >> >>>> > > >> >> >> >> fact_sales: >>>> > > >> >> >> >> storeid int >>>> > > >> >> >> >> itemid int >>>> > > >> >> >> >> custid int >>>> > > >> >> >> >> qty int >>>> > > >> >> >> >> price double >>>> > > >> >> >> >> >>>> > > >> >> >> >>dim_customer >>>> > > >> >> >> >> custid int >>>> > > >> >> >> >> name string >>>> > > >> >> >> >> >>>> > > >> >> >> >> dim_item >>>> > > >> >> >> >> itemid int >>>> > > >> >> >> >> category string >>>> > > >> >> >> >> brand string >>>> > > >> >> >> >> color string >>>> > > >> >> >> >> >>>> > > >> >> >> >>dim_store >>>> > > >> >> >> >> storeid int >>>> > > >> >> >> >> city string >>>> > > >> >> >> >> state string >>>> > > >> >> >> >> >>>> > > >> >> >> >>JSON is as below. >>>> > > >> >> >> >> The JSON is as below. >>>> > > >> >> >> >> >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "name": "Retail_Cube", >>>> > > >> >> >> >> "description": "", >>>> > > >> >> >> >> "dimensions": [ >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID", >>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >> "hierarchy": false, >>>> > > >> >> >> >> "derived": null, >>>> > > >> >> >> >> "column": [ >>>> > > >> >> >> >> "STOREID" >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "id": 1 >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID", >>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >> "hierarchy": false, >>>> > > >> >> >> >> "derived": null, >>>> > > >> >> >> >> "column": [ >>>> > > >> >> >> >> "ITEMID" >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "id": 2 >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID", >>>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >> "hierarchy": false, >>>> > > >> >> >> >> "derived": null, >>>> > > >> >> >> >> "column": [ >>>> > > >> >> >> >> "CUSTID" >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "id": 3 >>>> > > >> >> >> >> } >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "measures": [ >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "id": 1, >>>> > > >> >> >> >> "name": "_COUNT_", >>>> > > >> >> >> >> "function": { >>>> > > >> >> >> >> "expression": "COUNT", >>>> > > >> >> >> >> "returntype": "bigint", >>>> > > >> >> >> >> "parameter": { >>>> > > >> >> >> >> "type": "constant", >>>> > > >> >> >> >> "value": 1 >>>> > > >> >> >> >> } >>>> > > >> >> >> >> } >>>> > > >> >> >> >> } >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "rowkey": { >>>> > > >> >> >> >> "rowkey_columns": [ >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "column": "STOREID", >>>> > > >> >> >> >> "length": 0, >>>> > > >> >> >> >> "dictionary": "true", >>>> > > >> >> >> >> "mandatory": false >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "column": "ITEMID", >>>> > > >> >> >> >> "length": 0, >>>> > > >> >> >> >> "dictionary": "true", >>>> > > >> >> >> >> "mandatory": false >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "column": "CUSTID", >>>> > > >> >> >> >> "length": 0, >>>> > > >> >> >> >> "dictionary": "true", >>>> > > >> >> >> >> "mandatory": false >>>> > > >> >> >> >> } >>>> > > >> >> >> >> ], >>>> > > >> >> >> >> "aggregation_groups": [ >>>> > > >> >> >> >> [ >>>> > > >> >> >> >> "STOREID", >>>> > > >> >> >> >> "ITEMID", >>>> > > >> >> >> >> "CUSTID" >>>> > > >> >> >> >> ] >>>> > > >> >> >> >> ] >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> "notify_list": [], >>>> > > >> >> >> >> "capacity": "", >>>> > > >> >> >> >> "hbase_mapping": { >>>> > > >> >> >> >> "column_family": [ >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "name": "f1", >>>> > > >> >> >> >> "columns": [ >>>> > > >> >> >> >> { >>>> > > >> >> >> >> "qualifier": "m", >>>> > > >> >> >> >> "measure_refs": [ >>>> > > >> >> >> >> "_COUNT_" >>>> > > >> >> >> >> ] >>>> > > >> >> >> >> } >>>> > > >> >> >> >> ] >>>> > > >> >> >> >> } >>>> > > >> >> >> >> ] >>>> > > >> >> >> >> }, >>>> > > >> >> >> >> "project": "RetailProject", >>>> > > >> >> >> >> "model_name": "Retail_Cube" >>>> > > >> >> >> >> } >>>> > > >> >> >> >> >>>> > > >> >> >> >>Regards, >>>> > > >> >> >> >>Santosh Akhilesh >>>> > > >> >> >> >>Bangalore R&D >>>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD. >>>> > > >> >> >> >> >>>> > > >> >> >> >>www.huawei.com >>>> > > >> >> >> >>>>>>---------------------------------------------------------- >>>> > > >> >> >> --------------- >>>> > > >> >> >> >>- >>>> > > >> >> >> >>>>>>----------------------------------------------------------- >>>> > > >> >> >> >>This e-mail and its attachments contain confidential >>>> > information >>>> > > >> >>from >>>> > > >> >> >> >>HUAWEI, which >>>> > > >> >> >> >>is intended only for the person or entity whose address >>>>is >>>> > > >>listed >>>> > > >> >> >>above. >>>> > > >> >> >> >>Any use of the >>>> > > >> >> >> >>information contained herein in any way (including, but >>>>not >>>> > > >>limited >>>> > > >> >> >>to, >>>> > > >> >> >> >>total or partial >>>> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons >>>>other >>>> > > >>than >>>> > > >> >>the >>>> > > >> >> >> >>intended >>>> > > >> >> >> >>recipient(s) is prohibited. If you receive this e-mail >>>>in >>>> > error, >>>> > > >> >> >>please >>>> > > >> >> >> >>notify the sender by >>>> > > >> >> >> >>phone or email immediately and delete it! >>>> > > >> >> >> >> >>>> > > >> >> >> >>________________________________________ >>>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]] >>>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM >>>> > > >> >> >> >>To: [email protected] >>>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option >>>>is >>>>not >>>> > > >> >> >>responding >>>> > > >> >> >> >>on GUI >>>> > > >> >> >> >> >>>> > > >> >> >> >>Hi Santosh, >>>> > > >> >> >> >> >>>> > > >> >> >> >>It looks like hadoop failed to execute some shell >>>>command in >>>> > the >>>> > > >> >> >> >>container; You need dive into hadoop to see what¹s the >>>> > concrete >>>> > > >> >>error. >>>> > > >> >> >> >>You >>>> > > >> >> >> >>can use yarn logs command to fetch all logs: >>>> > > >> >> >> >> >>>> > > >> >> >> >>yarn logs -applicationId <app_id> >>>> > > >> >> >> >> >>>> > > >> >> >> >> >>>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" >>>> > > >><[email protected] >>>> > > >> > >>>> > > >> >> >> >>wrote: >>>> > > >> >> >> >> >>>> > > >> >> >> >>>Hi Luke / Shaofeng , >>>> > > >> >> >> >>> Can you please help me to check this issue. >>>> > > >> >> >> >>>Regards, >>>> > > >> >> >> >>>Santosh Akhilesh >>>> > > >> >> >> >>> >>>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >>>> > > >> >> >> >>>[email protected]> wrote: >>>> > > >> >> >> >>> >>>> > > >> >> >> >>>> Hi All , >>>> > > >> >> >> >>>> is it because of following error in map >>>>reduce >>>> job >>>> > ? >>>> > > >> >>what >>>> > > >> >> >> >>>>could >>>> > > >> >> >> >>>>be >>>> > > >> >> >> >>>> way to resolve this , a google search says that its >>>>issue >>>> > of >>>> > > >> >>Yarn >>>> > > >> >> >> >>>>class >>>> > > >> >> >> >>>> path , but I am not sure what it is ? >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job >>>>table=RETAIL.FACT_SALES >>>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2 >>>>times >>>> > due >>>> > > >> >>to AM >>>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002 >>>>exited >>>> > > >>with >>>> > > >> >> >> >>>>exitCode: 1 >>>> > > >> >> >> >>>> For more detailed output, check application tracking >>>> page: >>>> > > >> >> >> >>>> >>>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then, >>>> > > >> >> >>click >>>> > > >> >> >> >>>>on >>>> > > >> >> >> >>>> links to logs of each attempt. >>>> > > >> >> >> >>>> Diagnostics: Exception from container-launch. >>>> > > >> >> >> >>>> Container id: container_1424791969399_0008_02_000001 >>>> > > >> >> >> >>>> Exit code: 1 >>>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1: >>>> > > >> >> >> >>>> at >>>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >>>> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >>>> > > >> >> >> e(Shell.java:71 >>>> > > >> >> >> >>>>5 >>>> > > >> >> >> >>>>) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >>>> > > >> >> >> erExecutor.laun >>>> > > >> >> >> >>>>c >>>> > > >> >> >> >>>>h >>>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>> > > >> >> >> er.launcher.Con >>>> > > >> >> >> >>>>t >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>>> > > >> >> >> er.launcher.Con >>>> > > >> >> >> >>>>t >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82) >>>> > > >> >> >> >>>> at >>>> java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>> > > >> >> >> PoolExecutor.ja >>>> > > >> >> >> >>>>v >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>:1145) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>> > > >> >> >> dPoolExecutor.j >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>v >>>> > > >> >> >> >>>>a:615) >>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>> > > >> >> >> >>>> Container exited with a non-zero exit code 1 >>>> > > >> >> >> >>>> Failing this attempt. Failing the application. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> ---------- Forwarded message ---------- >>>> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]> >>>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>option >>>>is >>>> > not >>>> > > >> >> >> >>>>responding >>>> > > >> >> >> >>>>on >>>> > > >> >> >> >>>> GUI >>>> > > >> >> >> >>>> To: "[email protected]" >>>> > > >> >> >><[email protected]> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> hi , >>>> > > >> >> >> >>>> please someone give me a hand to resolve this >>>>issue , >>>> > > >>thanks. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Regards, >>>> > > >> >> >> >>>> Santosh Akhilesh >>>> > > >> >> >> >>>> Bangalore R&D >>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> www.huawei.com >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>-------------------------------------------------------- >>>> > > >> >> >> --------------- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>> >>>>------------------------------------------------------------ >>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>>> > > >>information >>>> > > >> >> >>from >>>> > > >> >> >> >>>> HUAWEI, which >>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>address >>>> is >>>> > > >> >>listed >>>> > > >> >> >> >>>>above. >>>> > > >> >> >> >>>> Any use of the >>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>but >>>> not >>>> > > >> >>limited >>>> > > >> >> >> >>>>to, >>>> > > >> >> >> >>>> total or partial >>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>persons >>>> > other >>>> > > >> >>than >>>> > > >> >> >>the >>>> > > >> >> >> >>>> intended >>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>e-mail in >>>> > > >>error, >>>> > > >> >> >> >>>>please >>>> > > >> >> >> >>>> notify the sender by >>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> ________________________________________ >>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >>>> > > >> >> >> >>>> To: [email protected] >>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>>>option >>>>is >>>> > not >>>> > > >> >> >> >>>>responding >>>> > > >> >> >> >>>>on >>>> > > >> >> >> >>>> GUI >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>cube >>>>I >>>> get >>>> > > >>an >>>> > > >> >> >> >>>>exception >>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>correctly >>>> and >>>> > > >> >> >> >>>>HCatInputFormat >>>> > > >> >> >> >>>> this file is present in >>>>hive-hcatalog-core-0.14.0.jar >>>>. >>>> > > >>Please >>>> > > >> >>let >>>> > > >> >> >>me >>>> > > >> >> >> >>>>know >>>> > > >> >> >> >>>> what can I do to resolve this ? >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> -- This was path issue , now no more exception in >>>> > kylin.log >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> But saveing cube still fails with error. And still >>>>can't >>>> > add >>>> > > >> >> >>measures. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Error Message >>>> > > >> >> >> >>>> Failed to take action. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> In log I can find no exception. Following is the >>>>last >>>>log >>>> > in >>>> > > >> >> >>kylin.log >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24 >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >>>> > > >> >> >> .DefaultSchedul >>>> > > >> >> >> >>>>e >>>> > > >> >> >> >>>>r >>>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >>>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, >>>>6 >>>> > others >>>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >>>> > > >> >> >> beController.de >>>> > > >> >> >> >>>>s >>>> > > >> >> >> >>>>e >>>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)] >>>> > > >> >> >> >>>> - Saving cube { >>>> > > >> >> >> >>>> "name": "", >>>> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >>>> "lookups": [], >>>> > > >> >> >> >>>> "filter_condition": "", >>>> > > >> >> >> >>>> "capacity": "SMALL", >>>> > > >> >> >> >>>> "partition_desc": { >>>> > > >> >> >> >>>> "partition_date_column": "", >>>> > > >> >> >> >>>> "partition_date_start": 0, >>>> > > >> >> >> >>>> "partition_type": "APPEND" >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> "last_modified": 0 >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> local access logs all with 200 , so seems ok. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET >>>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST >>>> > > >> >> >>/kylin/api/cubes >>>> > > >> >> >> >>>> HTTP/1.1" 200 701 >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Regards, >>>> > > >> >> >> >>>> Santosh Akhilesh >>>> > > >> >> >> >>>> Bangalore R&D >>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> www.huawei.com >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>-------------------------------------------------------- >>>> > > >> >> >> --------------- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>> >>>>------------------------------------------------------------ >>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>>> > > >>information >>>> > > >> >> >>from >>>> > > >> >> >> >>>> HUAWEI, which >>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>address >>>> is >>>> > > >> >>listed >>>> > > >> >> >> >>>>above. >>>> > > >> >> >> >>>> Any use of the >>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>but >>>> not >>>> > > >> >>limited >>>> > > >> >> >> >>>>to, >>>> > > >> >> >> >>>> total or partial >>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>persons >>>> > other >>>> > > >> >>than >>>> > > >> >> >>the >>>> > > >> >> >> >>>> intended >>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>e-mail in >>>> > > >>error, >>>> > > >> >> >> >>>>please >>>> > > >> >> >> >>>> notify the sender by >>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> ________________________________________ >>>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >>>> > > >> >> >> >>>> To: [email protected] >>>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>>> > > >> >> >> >>>> Subject: Error while making cube & Measure option is >>>>not >>>> > > >> >> >>responding on >>>> > > >> >> >> >>>>GUI >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Hi All , >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> I am building a simple cube for test and using >>>>the >>>> > binary >>>> > > >> >>build >>>> > > >> >> >> >>>>0.7.1 >>>> > > >> >> >> >>>> . I have following hive tables with columns. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> fact_sales: >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> storeid int >>>> > > >> >> >> >>>> itemid int >>>> > > >> >> >> >>>> custid int >>>> > > >> >> >> >>>> qty int >>>> > > >> >> >> >>>> price double >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> dim_customer >>>> > > >> >> >> >>>> custid int >>>> > > >> >> >> >>>> name string >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> dim_item >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> itemid int >>>> > > >> >> >> >>>> category string >>>> > > >> >> >> >>>> brand string >>>> > > >> >> >> >>>> color string >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> dim_store >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> storeid int >>>> > > >> >> >> >>>> city string >>>> > > >> >> >> >>>> state string >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Please help me to answer following issues; >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> 1. When I go to measure section and click on measure >>>> > option , >>>> > > >> >> >>there is >>>> > > >> >> >> >>>>no >>>> > > >> >> >> >>>> response , I want add measure on qty and price with >>>>sum >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>>>cube >>>>I >>>> get >>>> > > >>an >>>> > > >> >> >> >>>>exception >>>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>>correctly >>>> and >>>> > > >> >> >> >>>>HCatInputFormat >>>> > > >> >> >> >>>> this file is present in >>>>hive-hcatalog-core-0.14.0.jar >>>>. >>>> > > >>Please >>>> > > >> >>let >>>> > > >> >> >>me >>>> > > >> >> >> >>>>know >>>> > > >> >> >> >>>> what can I do to resolve this ? >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> 3. Also I have another question since this is a test >>>>and >>>> > > >>data is >>>> > > >> >> >>small >>>> > > >> >> >> >>>>I >>>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to >>>>skip >>>> > > >>partition >>>> > > >> >> >>stage >>>> > > >> >> >> >>>> while cube build ? >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Exception >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24 >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >>>> > > >> >> >> l.DefaultSchedu >>>> > > >> >> >> >>>>l >>>> > > >> >> >> >>>>e >>>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >>>> > > >> >> >> >>>> - ExecuteException >>>> job:c3532a6f-97ea-474a-b36a-218dd517cedb >>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut >>>> > > >> >> >> e(AbstractExecu >>>> > > >> >> >> >>>>t >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>ble.java:102) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >>>> > > >> >> >> bRunner.run(Def >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>u >>>> > > >> >> >> >>>>ltScheduler.java:132) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>>> > > >> >> >> PoolExecutor.ja >>>> > > >> >> >> >>>>v >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>:1145) >>>> > > >> >> >> >>>> at >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>>> > > >> >> >> dPoolExecutor.j >>>> > > >> >> >> >>>>a >>>> > > >> >> >> >>>>v >>>> > > >> >> >> >>>>a:615) >>>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>>> > > >> >> >> >>>> Caused by: >>>> org.apache.kylin.job.exception.ExecuteException: >>>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> The JSON is as below. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "name": "Retail_Cube", >>>> > > >> >> >> >>>> "description": "", >>>> > > >> >> >> >>>> "dimensions": [ >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >>>> "hierarchy": false, >>>> > > >> >> >> >>>> "derived": null, >>>> > > >> >> >> >>>> "column": [ >>>> > > >> >> >> >>>> "STOREID" >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "id": 1 >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >>>> "hierarchy": false, >>>> > > >> >> >> >>>> "derived": null, >>>> > > >> >> >> >>>> "column": [ >>>> > > >> >> >> >>>> "ITEMID" >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "id": 2 >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >>>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>>> > > >> >> >> >>>> "hierarchy": false, >>>> > > >> >> >> >>>> "derived": null, >>>> > > >> >> >> >>>> "column": [ >>>> > > >> >> >> >>>> "CUSTID" >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "id": 3 >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "measures": [ >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "id": 1, >>>> > > >> >> >> >>>> "name": "_COUNT_", >>>> > > >> >> >> >>>> "function": { >>>> > > >> >> >> >>>> "expression": "COUNT", >>>> > > >> >> >> >>>> "returntype": "bigint", >>>> > > >> >> >> >>>> "parameter": { >>>> > > >> >> >> >>>> "type": "constant", >>>> > > >> >> >> >>>> "value": 1 >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "rowkey": { >>>> > > >> >> >> >>>> "rowkey_columns": [ >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "column": "STOREID", >>>> > > >> >> >> >>>> "length": 0, >>>> > > >> >> >> >>>> "dictionary": "true", >>>> > > >> >> >> >>>> "mandatory": false >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "column": "ITEMID", >>>> > > >> >> >> >>>> "length": 0, >>>> > > >> >> >> >>>> "dictionary": "true", >>>> > > >> >> >> >>>> "mandatory": false >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "column": "CUSTID", >>>> > > >> >> >> >>>> "length": 0, >>>> > > >> >> >> >>>> "dictionary": "true", >>>> > > >> >> >> >>>> "mandatory": false >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> ], >>>> > > >> >> >> >>>> "aggregation_groups": [ >>>> > > >> >> >> >>>> [ >>>> > > >> >> >> >>>> "STOREID", >>>> > > >> >> >> >>>> "ITEMID", >>>> > > >> >> >> >>>> "CUSTID" >>>> > > >> >> >> >>>> ] >>>> > > >> >> >> >>>> ] >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> "notify_list": [], >>>> > > >> >> >> >>>> "capacity": "", >>>> > > >> >> >> >>>> "hbase_mapping": { >>>> > > >> >> >> >>>> "column_family": [ >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "name": "f1", >>>> > > >> >> >> >>>> "columns": [ >>>> > > >> >> >> >>>> { >>>> > > >> >> >> >>>> "qualifier": "m", >>>> > > >> >> >> >>>> "measure_refs": [ >>>> > > >> >> >> >>>> "_COUNT_" >>>> > > >> >> >> >>>> ] >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> ] >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> ] >>>> > > >> >> >> >>>> }, >>>> > > >> >> >> >>>> "project": "RetailProject", >>>> > > >> >> >> >>>> "model_name": "Retail_Cube" >>>> > > >> >> >> >>>> } >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> Regards, >>>> > > >> >> >> >>>> Santosh Akhilesh >>>> > > >> >> >> >>>> Bangalore R&D >>>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> www.huawei.com >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>>>>>>-------------------------------------------------------- >>>> > > >> >> >> --------------- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>>- >>>> > > >> >> >> >>>> >>>>------------------------------------------------------------ >>>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>>> > > >>information >>>> > > >> >> >>from >>>> > > >> >> >> >>>> HUAWEI, which >>>> > > >> >> >> >>>> is intended only for the person or entity whose >>>>address >>>> is >>>> > > >> >>listed >>>> > > >> >> >> >>>>above. >>>> > > >> >> >> >>>> Any use of the >>>> > > >> >> >> >>>> information contained herein in any way (including, >>>>but >>>> not >>>> > > >> >>limited >>>> > > >> >> >> >>>>to, >>>> > > >> >> >> >>>> total or partial >>>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>>persons >>>> > other >>>> > > >> >>than >>>> > > >> >> >>the >>>> > > >> >> >> >>>> intended >>>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>>e-mail in >>>> > > >>error, >>>> > > >> >> >> >>>>please >>>> > > >> >> >> >>>> notify the sender by >>>> > > >> >> >> >>>> phone or email immediately and delete it! >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>>> -- >>>> > > >> >> >> >>>> Regards, >>>> > > >> >> >> >>>> Santosh Akhilesh >>>> > > >> >> >> >>>> +91-0-9845482201 >>>> > > >> >> >> >>>> >>>> > > >> >> >> >>> >>>> > > >> >> >> >>> >>>> > > >> >> >> >>> >>>> > > >> >> >> >>>-- >>>> > > >> >> >> >>>Regards, >>>> > > >> >> >> >>>Santosh Akhilesh >>>> > > >> >> >> >>>+91-0-9845482201 >>>> > > >> >> >> >>>> > > >> >> >> >>>> > > >> >> >>>> > > >> >> >>>> > > >> >>>> > > >> >>>> > > >>>> > > >>>> > >>>> > >>>> > -- >>>> > Regards, >>>> > Santosh Akhilesh >>>> > +91-0-9845482201 >>>> > >>>> >>> >>> >>> >>>-- >>>Regards, >>>Santosh Akhilesh >>>+91-0-9845482201
