Thanks. https://issues.apache.org/jira/browse/AVRO-1506
I can take it up. On Mon, May 12, 2014 at 7:22 AM, Lewis John Mcgibbney < [email protected]> wrote: > My guess is that this is Avro side. We've seen similar traces with Nutch. > This looks like a JIRA ticket. > On May 11, 2014 4:53 PM, "Deepak" <[email protected]> wrote: > >> >> >> On 07-May-2014, at 7:35 am, ÐΞ€ρ@Ҝ (๏̯͡๏) <[email protected]> wrote: >> >> Exception: >> >> jjava.lang.Exception: java.lang.IncompatibleClassChangeError: Found >> interface org.apache.hadoop.mapreduce.TaskAttemptContext, but class was >> expected >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:522) >> >> Caused by: java.lang.IncompatibleClassChangeError: Found interface >> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected >> >> at >> org.apache.avro.mapreduce.AvroRecordReaderBase.initialize(AvroRecordReaderBase.java:86) >> >> at >> com.tracking.sdk.pig.load.format.AggregateRecordReader.initialize(AggregateRecordReader.java:41) >> >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.initialize(PigRecordReader.java:192) >> >> at >> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.initialize(MapTask.java:525) >> >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) >> >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243) >> >> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> >> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >> >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> >> at java.lang.Thread.run(Thread.java:744) >> >> >> Imports used in my recordreader class. >> >> import org.apache.avro.Schema; >> >> import org.apache.avro.mapreduce.AvroKeyValueRecordReader; >> >> import org.apache.hadoop.mapreduce.InputSplit; >> >> import org.apache.hadoop.mapreduce.TaskAttemptContext; >> >> Any suggestions ? Or does this require a fix from Avro ? >> >> Regards, >> >> Deepak >> >> -- Deepak
