This error indicates there might be different version jars in your
environment; As the class is packaged in hive-metastore-*.jar, please
search your machine to see whether there are different versions, and then
remove the redundant one;

2015-05-28 15:33 GMT+08:00 Mohit Bhatnagar <[email protected]>:

> Hi,
> Can someone help me out with this. I am getting the below error while
> building the cube in stage 2 :
> Here are the logs
>
> 2015-05-27 19:04:58,108 FATAL [IPC Server handler 1 on 44765]
> org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task:
> attempt_1432706588203_2125_m_000000_0 - exited : java.io.IOException:
> Deserialization error: org.apache.hadoop.hive.metastore.api.Table; local
> class incompatible: stream classdesc serialVersionUID = 398473631015277182,
> local class serialVersionUID = -946662244473213550
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42)
>         at
> org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> Caused by: java.io.InvalidClassException:
> org.apache.hadoop.hive.metastore.api.Table; local class incompatible:
> stream classdesc serialVersionUID = 398473631015277182, local class
> serialVersionUID = -946662244473213550
>         at
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>         at
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>         at
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115)
>         ... 11 more
>
> 2015-05-27 19:04:58,108 INFO [IPC Server handler 1 on 44765]
> org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report from
> attempt_1432706588203_2125_m_000000_0: Error: java.io.IOException:
> Deserialization error: org.apache.hadoop.hive.metastore.api.Table; local
> class incompatible: stream classdesc serialVersionUID = 398473631015277182,
> local class serialVersionUID = -946662244473213550
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42)
>         at
> org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> Caused by: java.io.InvalidClassException:
> org.apache.hadoop.hive.metastore.api.Table; local class incompatible:
> stream classdesc serialVersionUID = 398473631015277182, local class
> serialVersionUID = -946662244473213550
>         at
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>         at
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>         at
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115)
>         ... 11 more
>
> 2015-05-27 19:04:58,110 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> report from attempt_1432706588203_2125_m_000000_0: Error:
> java.io.IOException: Deserialization error:
> org.apache.hadoop.hive.metastore.api.Table; local class incompatible:
> stream classdesc serialVersionUID = 398473631015277182, local class
> serialVersionUID = -946662244473213550
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71)
>         at
> org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42)
>         at
> org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> Caused by: java.io.InvalidClassException:
> org.apache.hadoop.hive.metastore.api.Table; local class incompatible:
> stream classdesc serialVersionUID = 398473631015277182, local class
> serialVersionUID = -946662244473213550
>         at
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>         at
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>         at
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115)
>         ... 11 more
>
> 2015-05-27 19:04:58,112 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from RUNNING
> to FAIL_CONTAINER_CLEANUP
> 2015-05-27 19:04:58,114 INFO [ContainerLauncher #1]
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl:
> Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container
> container_e38_1432706588203_2125_01_000002 taskAttempt
> attempt_1432706588203_2125_m_000000_0
> 2015-05-27 19:04:58,114 INFO [ContainerLauncher #1]
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING
> attempt_1432706588203_2125_m_000000_0
> 2015-05-27 19:04:58,115 INFO [ContainerLauncher #1]
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy:
> Opening proxy : sj1dra054.corp.adobe.com:8041
> 2015-05-27 19:04:58,138 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from
> FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
> 2015-05-27 19:04:58,139 INFO [CommitterEvent Processor #1]
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing
> the event EventType: TASK_ABORT
> 2015-05-27 19:04:58,148 WARN [CommitterEvent Processor #1]
> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not
> delete
> hdfs://nameservice1/tmp/kylin-ffb0e3f9-4720-418a-a8d9-924f47fe1eea/mmmmmm/fact_distinct_columns/_temporary/1/_temporary/attempt_1432706588203_2125_m_000000_0
> 2015-05-27 19:04:58,150 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from
> FAIL_TASK_CLEANUP to FAILED
> 2015-05-27 19:04:58,161 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.yarn.util.RackResolver: Resolved
> sj1dra054.corp.adobe.com to /sb
> 2015-05-27 19:04:58,161 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.yarn.util.RackResolver: Resolved
> sj1dra151.corp.adobe.com to /sa
> 2015-05-27 19:04:58,161 INFO [Thread-51]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on
> node sj1dra054.corp.adobe.com
> 2015-05-27 19:04:58,164 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl:
> attempt_1432706588203_2125_m_000000_1 TaskAttempt Transitioned from NEW to
> UNASSIGNED
> 2015-05-27 19:04:58,165 INFO [Thread-51]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added
> attempt_1432706588203_2125_m_000000_1 to list of failed maps
> 2015-05-27 19:04:58,365 INFO [RMCommunicator Allocator]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before
> Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1
> AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0
> HostLocal:1 RackLocal:0
> 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources()
> for application_1432706588203_2125: ask=1 release= 0 newContainers=0
> finishedContainers=0 resourcelimit=<memory:321024, vCores:262> knownNMs=10
> 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating
> schedule, headroom=<memory:321024, vCores:262>
> 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow
> start threshold not met. completedMapsForReduceSlowstart 1
> 2015-05-27 19:04:59,378 INFO [RMCommunicator Allocator]
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received
> completed container container_e38_1432706588203_2125_01_000002
> 2015-05-27 19:04:59,379 INFO [AsyncDispatcher event handler]
> org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics
> report from attempt_1432706588203_2125_m_000000_0: Container killed by the
> ApplicationMaster.
> Container killed on request. Exit code is 143
>
>
>
>
> Thanks,
> Mohit
>

Reply via email to