Hi, We have 2 jars hive-metastore-0.13.1-cdh5.3.3.jar and hive-metastore.jar. Both of these have org.apache.hadoop.hive.metastore.api.Table class. But both their checksums are same. i.e same version. Can there be any other issue?
Thanks, Mohit -----Original Message----- From: ShaoFeng Shi [mailto:[email protected]] Sent: Thursday, May 28, 2015 2:03 PM To: [email protected] Subject: Re: Issue while building CUBE: java.io.IOException: Deserialization error: This error indicates there might be different version jars in your environment; As the class is packaged in hive-metastore-*.jar, please search your machine to see whether there are different versions, and then remove the redundant one; 2015-05-28 15:33 GMT+08:00 Mohit Bhatnagar <[email protected]>: > Hi, > Can someone help me out with this. I am getting the below error while > building the cube in stage 2 : > Here are the logs > > 2015-05-27 19:04:58,108 FATAL [IPC Server handler 1 on 44765] > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: > attempt_1432706588203_2125_m_000000_0 - exited : java.io.IOException: > Deserialization error: org.apache.hadoop.hive.metastore.api.Table; > local class incompatible: stream classdesc serialVersionUID = > 398473631015277182, local class serialVersionUID = -946662244473213550 > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117) > at > org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42) > at > org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) > Caused by: java.io.InvalidClassException: > org.apache.hadoop.hive.metastore.api.Table; local class incompatible: > stream classdesc serialVersionUID = 398473631015277182, local class > serialVersionUID = -946662244473213550 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115) > ... 11 more > > 2015-05-27 19:04:58,108 INFO [IPC Server handler 1 on 44765] > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report > from > attempt_1432706588203_2125_m_000000_0: Error: java.io.IOException: > Deserialization error: org.apache.hadoop.hive.metastore.api.Table; > local class incompatible: stream classdesc serialVersionUID = > 398473631015277182, local class serialVersionUID = -946662244473213550 > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117) > at > org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42) > at > org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) > Caused by: java.io.InvalidClassException: > org.apache.hadoop.hive.metastore.api.Table; local class incompatible: > stream classdesc serialVersionUID = 398473631015277182, local class > serialVersionUID = -946662244473213550 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115) > ... 11 more > > 2015-05-27 19:04:58,110 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > Diagnostics report from attempt_1432706588203_2125_m_000000_0: Error: > java.io.IOException: Deserialization error: > org.apache.hadoop.hive.metastore.api.Table; local class incompatible: > stream classdesc serialVersionUID = 398473631015277182, local class > serialVersionUID = -946662244473213550 > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:117) > at > org.apache.hive.hcatalog.mapreduce.HCatSplit.readFields(HCatSplit.java:139) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:71) > at > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:42) > at > org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:372) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:751) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) > Caused by: java.io.InvalidClassException: > org.apache.hadoop.hive.metastore.api.Table; local class incompatible: > stream classdesc serialVersionUID = 398473631015277182, local class > serialVersionUID = -946662244473213550 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > org.apache.hive.hcatalog.common.HCatUtil.deserialize(HCatUtil.java:115) > ... 11 more > > 2015-05-27 19:04:58,112 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from > RUNNING to FAIL_CONTAINER_CLEANUP > 2015-05-27 19:04:58,114 INFO [ContainerLauncher #1] > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container > container_e38_1432706588203_2125_01_000002 taskAttempt > attempt_1432706588203_2125_m_000000_0 > 2015-05-27 19:04:58,114 INFO [ContainerLauncher #1] > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > KILLING > attempt_1432706588203_2125_m_000000_0 > 2015-05-27 19:04:58,115 INFO [ContainerLauncher #1] > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > Opening proxy : sj1dra054.corp.adobe.com:8041 > 2015-05-27 19:04:58,138 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP > 2015-05-27 19:04:58,139 INFO [CommitterEvent Processor #1] > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > Processing the event EventType: TASK_ABORT > 2015-05-27 19:04:58,148 WARN [CommitterEvent Processor #1] > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not > delete > hdfs://nameservice1/tmp/kylin-ffb0e3f9-4720-418a-a8d9-924f47fe1eea/mmm > mmm/fact_distinct_columns/_temporary/1/_temporary/attempt_143270658820 > 3_2125_m_000000_0 > 2015-05-27 19:04:58,150 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > attempt_1432706588203_2125_m_000000_0 TaskAttempt Transitioned from > FAIL_TASK_CLEANUP to FAILED > 2015-05-27 19:04:58,161 INFO [AsyncDispatcher event handler] > org.apache.hadoop.yarn.util.RackResolver: Resolved > sj1dra054.corp.adobe.com to /sb > 2015-05-27 19:04:58,161 INFO [AsyncDispatcher event handler] > org.apache.hadoop.yarn.util.RackResolver: Resolved > sj1dra151.corp.adobe.com to /sa > 2015-05-27 19:04:58,161 INFO [Thread-51] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures > on node sj1dra054.corp.adobe.com > 2015-05-27 19:04:58,164 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > attempt_1432706588203_2125_m_000000_1 TaskAttempt Transitioned from > NEW to UNASSIGNED > 2015-05-27 19:04:58,165 INFO [Thread-51] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added > attempt_1432706588203_2125_m_000000_1 to list of failed maps > 2015-05-27 19:04:58,365 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 > AssignedMaps:1 > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 > HostLocal:1 RackLocal:0 > 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > getResources() for application_1432706588203_2125: ask=1 release= 0 > newContainers=0 > finishedContainers=0 resourcelimit=<memory:321024, vCores:262> > knownNMs=10 > 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating schedule, headroom=<memory:321024, vCores:262> > 2015-05-27 19:04:58,368 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce > slow start threshold not met. completedMapsForReduceSlowstart 1 > 2015-05-27 19:04:59,378 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received > completed container container_e38_1432706588203_2125_01_000002 > 2015-05-27 19:04:59,379 INFO [AsyncDispatcher event handler] > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > Diagnostics report from attempt_1432706588203_2125_m_000000_0: > Container killed by the ApplicationMaster. > Container killed on request. Exit code is 143 > > > > > Thanks, > Mohit >
