Gabriel Reid created CRUNCH-539: ----------------------------------- Summary: Use of TupleWritable.setConf fails in mapper/reducer Key: CRUNCH-539 URL: https://issues.apache.org/jira/browse/CRUNCH-539 Project: Crunch Issue Type: Bug Affects Versions: 0.12.0 Reporter: Gabriel Reid
In (at least) more recent versions of Hadoop 2, the implicit call to TupleWritable.setConf that happens when using TupleWritables fails with a ClassNotFoundException for (ironically) the TupleWritable class. This appears to be due to the way that ObjectInputStream resolves classes in its [resolveClass method|https://docs.oracle.com/javase/7/docs/api/java/io/ObjectInputStream.html#resolveClass(java.io.ObjectStreamClass)], together with the way that the context classloader is set within a hadoop mapper or reducer. This is similar to PIG-2532. This can be reproduced in the local job tracker (at least) in Hadoop 2.7.0, but it can't be reproduced in Crunch integration tests (due to classloading setup). It appears that this issue is only present in Crunch 0.12. The following code within a simple pipeline will cause this issue to occur: {code} PTable<String, Integer> yearTemperatures = ... /* Writable-based PTable */ PTable<String, Integer> maxTemps = yearTemperatures .groupByKey() .combineValues(Aggregators.MAX_INTS()) .top(1); //LINE THAT CAUSES THE ERROR {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)