Jeff,
I'll check it out this weekend.

-D

On Sat, Jan 9, 2010 at 3:47 PM, Jeff Dalton <[email protected]> wrote:
> I downloaded the version of PiggyBank from the 0.6 branch, compiled,
> and deployed it.  However, I still get the same error message:
>
> java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast
> to org.apache.pig.impl.PigContext
>        at 
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168)
>        at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333)
>        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>        at org.apache.hadoop.mapred.Child.main(Child.java:159)
>
> I'll try again later, but if anyone has any insights, I would
> appreciate the help.
>
> Thanks,
>
> - Jeff
>
> On Sat, Jan 9, 2010 at 6:33 PM, Jeff Dalton <[email protected]> wrote:
>> Ahh, the PiggyBank version was the latest from Trunk.  I probably need
>> to go track down the version from the 0.6 branch.
>>
>> On Sat, Jan 9, 2010 at 6:26 PM, Dmitriy Ryaboy <[email protected]> wrote:
>>> When you say that the code is from SVN, do you mean trunk, or the 0.6 
>>> branch?
>>>
>>>
>>> On Sat, Jan 9, 2010 at 3:22 PM, Jeff Dalton <[email protected]> wrote:
>>>> A cluster I'm using was recently upgraded to PIG 0.6.  Since then,
>>>> I've been having problems with scripts that use PiggyBank functions.
>>>> All the map jobs for the script fail with:
>>>> WARN org.apache.hadoop.mapred.Child: Error running child
>>>> java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast
>>>> to org.apache.pig.impl.PigContext
>>>>        at 
>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168)
>>>>        at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333)
>>>>        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>>>>        at org.apache.hadoop.mapred.Child.main(Child.java:159)
>>>> INFO org.apache.hadoop.mapred.Task: Runnning cleanup for the task
>>>>
>>>> I compiled the PiggyBank jar using the latest code from SVN (as of Jan
>>>> 9) and Pig 0.6.  Below I've included a simple example program that
>>>> caused the error that simply reads a text file of words and lower
>>>> cases them.
>>>>
>>>> register ./piggybank.jar
>>>> DEFINE ToLower org.apache.pig.piggybank.evaluation.string.LOWER();
>>>> words = LOAD './data/headwords_sample' USING PigStorage() as 
>>>> (word:charArray);
>>>> lowerCaseWords = FOREACH words GENERATE ToLower(word) as word;
>>>> STORE lowerCaseWords into './tmp/cooc3' USING PigStorage();
>>>>
>>>> The Hadoop error isn't very informative about what is going on.  Am I
>>>> using a compatible version of PiggyBank?  What should I be doing
>>>> differently?
>>>>
>>>> Thanks,
>>>>
>>>> - Jeff
>>>>
>>>
>>
>

Reply via email to