Discard old job and re-launch a new build? The old step 1 may created something out-dated after config change.
On Wed, Oct 21, 2015 at 10:14 AM, LIU Ze (刘则) <[email protected]> wrote: > > Thanks for all the help! I used the > https://issues.apache.org/jira/browse/KYLIN-1021, but it make another > error: > > kylin.log: > Caused by: org.apache.hadoop.ipc.RemoteException(java.io.IOException): > Unknown Job job_1444723293631_16444 > at > org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.verifyAndGetJob(HistoryClientService.java:218) > at > org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.getCounters(HistoryClientService.java:232) > at > org.apache.hadoop.mapreduce.v2.api.impl.pb.service.MRClientProtocolPBServiceImpl.getCounters(MRClie > > > yarn logs -applicationId application_1444723293631_16444: > > 2015-10-21 09:08:08,593 WARN [main] org.apache.hadoop.mapred.YarnChild: > Exception running child : java.io.IOException: Deserialization error: > invalid stream header: FED007E8 > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:119) > at > org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42) > at > org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:754) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: java.io.StreamCorruptedException: invalid stream header: > FED007E8 > at > java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:804) > at java.io.ObjectInputStream.<init>(ObjectInputStream.java:299) > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:116) > ... 11 more > ________________________________ > > 发件人: Shi, Shaofeng<mailto:[email protected]> > 发送时间: 2015-10-19 10:42 > 收件人: [email protected]<mailto:[email protected]>; > LIU Ze (刘则)<mailto:[email protected]> > 主题: Re: Error in Step 2 "Extract Fact Table Distinct Columns" > > good share Vadim; Kylin is trying to formally support this case (hive jars > are not in hadoop nodes), please look into: > https://issues.apache.org/jira/browse/KYLIN-1021 > > On 10/19/15, 10:31 AM, "Vadim Semenov" <[email protected]> wrote: > > >You can make changes in job/pom.xml for hive-hcatalog package, so it will > >be included in the job jar, and rebuild kylin using scripts/package.sh. > >I.e. in job/pom.xml change this > > <version>${hive-hcatalog.version}</version> > > <scope>provided</scope> > >to: > > <version>${hive-hcatalog.version}</version> > > <!--<scope>provided</scope>--> > > > > > >On October 18, 2015 at 10:15:51 PM, yu feng ([email protected]) wrote: > > > >From error log I think it was caused by being short of jars such as > >hive-hcatalog, I think it is because you did not deploy your hive env in > >every node of your hadoop cluster as we did, We solved this problem by > >modify the source code , upload or add tmpjars of those dependency jars > >and > >files before every mapreduce job was submitted. > > > >JIRA ticket is here : https://issues.apache.org/jira/browse/KYLIN-1021 > > > >hope it can help you ~ > > > >2015-10-16 15:52 GMT+08:00 LIU Ze (刘则) <[email protected]>: > > > >> hi all: > >> it make a error in step 2 > >> > >> kylin.log: > >> ________________________________ > >> [pool-5-thread-9]:[2015-10-16 > >> > >>15:47:06,398][ERROR][org.apache.kylin.job.common.HadoopCmdOutput.updateJo > >>bCounter(HadoopCmdOutput.java:100)] > >> - java.io.IOException: Unknown Job job_1444723293631_9288 > >> at > >> > >>org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHa > >>ndler.verifyAndGetJob(HistoryClientService.java:218) > >> at > >> > >>org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHa > >>ndler.getCounters(HistoryClientService.java:232) > >> at > >> > >>org.apache.hadoop.mapreduce.v2.api.impl.pb.service.MRClientProtocolPBServ > >>iceImpl.getCounters(MRClientProtocolPBServiceImpl.java:159) > >> at > >> > >>org.apache.hadoop.yarn.proto.MRClientProtocol$MRClientProtocolService$2.c > >>allBlockingMethod(MRClientProtocol.java:281) > >> at > >> > >>org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(Pr > >>otobufRpcEngine.java:616) > >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > >> at java.security.AccessController.doPrivileged(Native Method) > >> at javax.security.auth.Subject.doAs(Subject.java:415) > >> at > >> > >>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation > >>.java:1657) > >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > >> > >> > >> yarn logs -applicationId application_1444723293631_9286 : > >> > >> 2015-10-16 15:43:07,543 INFO [main] > >> org.apache.hadoop.conf.Configuration.deprecation: session.id is > >> deprecated. Instead, use dfs.metrics.session-id > >> 2015-10-16 15:43:07,826 INFO [main] > >> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output > >> > >> Committer Algorithm version is 1 > >> 2015-10-16 15:43:07,834 INFO [main] org.apache.hadoop.mapred.Task: > >>Using > >> ResourceCalculatorProcessTree : [ ] > >> 2015-10-16 15:43:07,901 WARN [main] org.apache.hadoop.mapred.YarnChild: > >> > >> Exception running child : java.lang.RuntimeException: > >> java.lang.ClassNotFoundException: Class > >> org.apache.hive.hcatalog.mapreduce.HCatInputFormat not found > >> at > >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2195) > >> at > >> > >>org.apache.hadoop.mapreduce.task.JobContextImpl.getInputFormatClass(JobCo > >>ntextImpl.java:174) > >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:749) > >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > >> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) > >> at java.security.AccessController.doPrivileged(Native Method) > >> at javax.security.auth.Subject.doAs(Subject.java:415) > >> at > >> > >>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation > >>.java:1657) > >> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > >> Caused by: java.lang.ClassNotFoundException: Class > >> org.apache.hive.hcatalog.mapreduce.HCatInputFormat not found > >> at > >> > >>org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:21 > >>01) > >> at > >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2193) > >> ... 8 more > >> > >> > >
