Hi Folks,

I am fairly new to Hadoop MR2 environment.  I have Hadoop0.23.1 (single
node) on Linux box and can successfully execute YARN test job, but when I
execute PIG (Ver. 12) to access and process files at HDFS, generate error
msg as follows.  I recompiled PIG with Hadoop23, but still no luck.

ant clean jar-withouthadoop -Dhadoopversion=23

Can someone please help and advise what I am missing here !!

Note: PIG in local mode works without any issue.

================

*grunt> A = load 'NYSE_dividends' as (exch, symb, dt, div);*
2013-11-26 13:20:23,151 [main] WARN  org.apache.hadoop.conf.
Configuration - fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:23,151 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:23,323 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:23,323 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
*grunt> dump A ;*
2013-11-26 13:20:31,805 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:31,806 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:31,841 [main] INFO
org.apache.pig.tools.pigstats.ScriptState - Pig features used in the
script: UNKNOWN
2013-11-26 13:20:31,913 [main] INFO
org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer -
{RULES_ENABLED=[AddForEach, ColumnMapKeyPrune,
DuplicateForEachColumnRewrite, GroupByConstParallelSetter,
ImplicitSplitInserter, LimitOptimizer, LoadTypeCastInserter, MergeFilter,
MergeForEach, NewPartitionFilterOptimizer, PartitionFilterOptimizer,
PushDownForEachFlatten, PushUpFilter, SplitFilter, StreamTypeCastInserter],
RULES_DISABLED=[FilterLogicExpressionSimplifier]}
2013-11-26 13:20:32,159 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,160 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,266 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler -
File concatenation threshold: 100 optimistic? false
2013-11-26 13:20:32,304 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,304 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,318 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
- MR plan size before optimization: 1
2013-11-26 13:20:32,319 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
- MR plan size after optimization: 1
2013-11-26 13:20:32,322 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,322 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,498 [main] INFO
org.apache.pig.tools.pigstats.ScriptState - Pig script settings are added
to the job
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use
mapreduce.reduce.markreset.buffer.percent
2013-11-26 13:20:32,504 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- mapred.job.reduce.markreset.buffer.percent is not set, set to default 0.3
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use
mapreduce.reduce.markreset.buffer.percent
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.output.compress is deprecated. Instead, use
mapreduce.output.fileoutputformat.compress
2013-11-26 13:20:32,506 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- creating jar file Job1088179330840878220.jar
2013-11-26 13:20:35,473 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- jar file Job1088179330840878220.jar created
2013-11-26 13:20:35,473 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.jar is deprecated. Instead, use mapreduce.job.jar
2013-11-26 13:20:35,488 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- Setting up single store job
2013-11-26 13:20:35,492 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- Map only job, skipping reducer estimation
2013-11-26 13:20:35,513 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 1 map-reduce job(s) waiting for submission.
2013-11-26 13:20:35,514 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker.http.address is deprecated. Instead, use
mapreduce.jobtracker.http.address
2013-11-26 13:20:35,954 [JobControl] INFO
org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths
to process : 1
2013-11-26 13:20:35,954 [JobControl] INFO
org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input
paths to process : 1
2013-11-26 13:20:35,971 [JobControl] WARN
org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
2013-11-26 13:20:35,971 [JobControl] WARN
org.apache.hadoop.io.compress.snappy.LoadSnappy - Snappy native library not
loaded
2013-11-26 13:20:35,973 [JobControl] INFO
org.apache.hadoop.mapreduce.JobSubmitter - Cleaning up the staging area
null/hadoop/.staging/job_1385489868206_0001

*2013-11-26 13:20:35,975 [JobControl] ERROR
org.apache.pig.backend.hadoop23.PigJobControl - Error while trying to run
jobs.java.lang.RuntimeException:
java.lang.reflect.InvocationTargetException*
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
        at
org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
        at java.lang.Thread.run(Thread.java:619)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:270)


*Caused by: java.lang.reflect.InvocationTargetException        at
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)*
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
        ... 3 more
Caused by: java.lang.NoSuchMethodError:
org.apache.hadoop.fs.FileSystem.getDefaultBlockSize(Lorg/apache/hadoop/fs/Path;)J
        at
org.apache.pig.backend.hadoop.executionengine.shims.HadoopShims.getDefaultBlockSize(HadoopShims.java:108)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:277)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:451)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:468)
        at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:360)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1221)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1218)
        at
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:336)
        ... 8 more
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- HadoopJobId: job_1385489868206_0001
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Processing aliases A
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- detailed locations: M: A[1,4] C:  R:
2013-11-26 13:20:35,979 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 0% complete
2013-11-26 13:20:35,983 [main] WARN
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to
stop immediately on failure.
2013-11-26 13:20:35,983 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- job job_1385489868206_0001 has failed! Stop running all dependent jobs
2013-11-26 13:20:35,983 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 100% complete
2013-11-26 13:20:36,104 [main] WARN
org.apache.hadoop.mapred.ClientServiceDelegate - Exception thrown by remote
end.
RemoteTrace:
 at LocalTrace:

org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl: Trying
to get information for an absent application application_1385489868206_0001
        at
org.apache.hadoop.yarn.ipc.ProtoOverHadoopRpcEngine$Invoker.invoke(ProtoOverHadoopRpcEngine.java:150)
        at $Proxy7.getApplicationReport(Unknown Source)
        at
org.apache.hadoop.yarn.api.impl.pb.client.ClientRMProtocolPBClientImpl.getApplicationReport(ClientRMProtocolPBClientImpl.java:116)
        at
org.apache.hadoop.mapred.ResourceMgrDelegate.getApplicationReport(ResourceMgrDelegate.java:343)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.getProxy(ClientServiceDelegate.java:142)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:288)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.getJobStatus(ClientServiceDelegate.java:365)
        at
org.apache.hadoop.mapred.YARNRunner.getJobStatus(YARNRunner.java:483)
        at org.apache.hadoop.mapreduce.Cluster.getJob(Cluster.java:186)
        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:627)
        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:625)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at
org.apache.hadoop.mapred.JobClient.getJobUsingCluster(JobClient.java:625)
        at
org.apache.hadoop.mapred.JobClient.getTaskReports(JobClient.java:679)
        at
org.apache.hadoop.mapred.JobClient.getMapTaskReports(JobClient.java:673)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.Launcher.getStats(Launcher.java:150)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.launchPig(MapReduceLauncher.java:428)
        at org.apache.pig.PigServer.launchPlan(PigServer.java:1322)
        at
org.apache.pig.PigServer.executeCompiledLogicalPlan(PigServer.java:1307)
        at org.apache.pig.PigServer.storeEx(PigServer.java:978)
        at org.apache.pig.PigServer.store(PigServer.java:942)
        at org.apache.pig.PigServer.openIterator(PigServer.java:855)
        at
org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:774)
        at
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:372)
        at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:198)
        at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:173)
        at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
        at org.apache.pig.Main.run(Main.java:541)
        at org.apache.pig.Main.main(Main.java:156)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.util.RunJar.main(RunJar.java:200)
2013-11-26 13:20:36,105 [main] ERROR
org.apache.hadoop.security.UserGroupInformation -
PriviledgedActionException as:hadoop (auth:SIMPLE)
cause:org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl:
Trying to get information for an absent application
application_1385489868206_0001
2013-11-26 13:20:36,105 [main] ERROR
org.apache.pig.tools.pigstats.SimplePigStats - ERROR: Trying to get
information for an absent application application_1385489868206_0001
2013-11-26 13:20:36,105 [main] ERROR
org.apache.pig.tools.pigstats.PigStatsUtil - 1 map reduce job(s) failed!
2013-11-26 13:20:36,107 [main] INFO
org.apache.pig.tools.pigstats.SimplePigStats - Script Statistics:

HadoopVersion   PigVersion      UserId  StartedAt       FinishedAt
Features
0.23.1  0.12.1-SNAPSHOT hadoop  2013-11-26 13:20:32     2013-11-26
13:20:36     UNKNOWN

Failed!

Failed Jobs:
JobId   Alias   Feature Message Outputs
job_1385489868206_0001  A       MAP_ONLY        Message: Unexpected System
Error Occured: java.lang.RuntimeException:
java.lang.reflect.InvocationTargetException
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
        at
org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
        at java.lang.Thread.run(Thread.java:619)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:270)
Caused by: java.lang.reflect.InvocationTargetException
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
        ... 3 more
Caused by: java.lang.NoSuchMethodError:
org.apache.hadoop.fs.FileSystem.getDefaultBlockSize(Lorg/apache/hadoop/fs/Path;)J
        at
org.apache.pig.backend.hadoop.executionengine.shims.HadoopShims.getDefaultBlockSize(HadoopShims.java:108)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:277)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:451)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:468)
        at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:360)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1221)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1218)
        at
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:336)
        ... 8 more
        hdfs://sasserver05:54310/tmp/temp601167191/tmp-1605231029,

Input(s):
Failed to read data from
"hdfs://sasserver05:54310/user/hadoop/NYSE_dividends"

Output(s):
Failed to produce result in
"hdfs://sasserver05:54310/tmp/temp601167191/tmp-1605231029"

Counters:
Total records written : 0
Total bytes written : 0
Spillable Memory Manager spill count : 0
Total bags proactively spilled: 0
Total records proactively spilled: 0

Job DAG:
job_1385489868206_0001


2013-11-26 13:20:36,107 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Failed!
2013-11-26 13:20:36,112 [main] ERROR org.apache.pig.tools.grunt.Grunt -
ERROR 1066: Unable to open iterator for alias A. Backend error : Trying to
get information for an absent application application_1385489868206_0001
Details at logfile: /home/hadoop/pig_1385489876211.log
grunt>

=============



On Tue, Nov 26, 2013 at 3:32 PM, Claudio Romo Otto <
claudio.r...@namictech.com> wrote:

> Hi Juan,
>
>     In a nutshell, you must pay attention to memory settings inside
> mapred-site.xml, yarn-site.xml, hadoop-env.sh and yarn-env.sh, so you have
> to design a memory distribution strategy according to your performance
> requirements. In this way you will have, among other things, enough memory
> for the Scheduler.
>
> Remember to reserve at least 600 - 800 mb for the operative system to
> avoid OOM errors.
>
> Best regards
> El 26/11/13 16:07, Juan Martin Pampliega escribió:
>
>> Hi Claudio,
>>
>> It would be nice to know which were the settings that you had to tune to
>> get this. I am having a similar issue with some jobs that I am running.
>> Thanks,
>> Juan.
>>
>>
>> On Wed, Oct 30, 2013 at 7:40 PM, Claudio Romo Otto <
>> claudio.r...@namictech.com> wrote:
>>
>>  Jarcec, finally I got solved this problem by learning more on hadoop 2
>>> (lot of reading), and then tuning some settings to let the work move from
>>> the SCHEDULED state. With this said, the last problem was only concerning
>>> on hadoop.
>>>
>>> Thanks for your support!
>>>
>>> El 30/10/13 18:03, Jarek Jarcec Cecho escribió:
>>>
>>>   Hi Claudio,
>>>
>>>> it's hard to guess from the limited information. I would suggest to take
>>>> a look into logs to see what is happening.
>>>>
>>>> One guess though - You've mentioned that the task was "running" for 30
>>>> minutes, but it still seems to be in SCHEDULED time - are your node
>>>> managers correctly running?
>>>>
>>>> Jarcec
>>>>
>>>> On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
>>>>
>>>>  You got it!
>>>>>
>>>>> The solution was to compile with  -Dhadoopversion=23 option. After
>>>>> your message I tried another test removing Cassandra from the chain
>>>>> and Pig sent successfully the job to hadoop.
>>>>>
>>>>> BUT! the problem changed, now the Map task remains forever stuck on
>>>>> Hadoop (30 minutes waiting, no other jobs running):
>>>>>
>>>>> Task
>>>>>
>>>>> Progress
>>>>>
>>>>> State
>>>>>
>>>>> Start Time
>>>>>
>>>>> Finish Time
>>>>>
>>>>> Elapsed Time
>>>>> task_1382631533263_0012_m_000000 <http://topgps-test-3.
>>>>> dnsalias.com:8088/proxy/application_1382631533263_
>>>>> 0012/mapreduce/task/task_1382631533263_0012_m_000000>
>>>>>
>>>>>          SCHEDULED       Fri, 25 Oct 2013 18:18:32 GMT   N/A     0sec
>>>>>
>>>>>
>>>>>
>>>>> Attempt
>>>>>
>>>>> Progress
>>>>>
>>>>> State
>>>>>
>>>>> Node
>>>>>
>>>>> Logs
>>>>>
>>>>> Started
>>>>>
>>>>> Finished
>>>>>
>>>>> Elapsed
>>>>>
>>>>> Note
>>>>> attempt_1382631533263_0012_m_000000_0   0,00    STARTING        N/A
>>>>> N/A     N/A
>>>>> N/A     0sec
>>>>>
>>>>>
>>>>> Don't know if this is a Hadoop problem or Pig, what do you think?
>>>>>
>>>>>
>>>>> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
>>>>>
>>>>>  It seems that Pig was correctly compiled against Hadoop 23, but the
>>>>>> Cassandra piece was not, check out the where the exception is coming
>>>>>> from:
>>>>>>
>>>>>>   Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>>
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>>
>>>>>>>  So, I would say that you also need to get Hadoop 2 compatible
>>>>>> Cassandra
>>>>>> connector first.
>>>>>>
>>>>>> Jarcec
>>>>>>
>>>>>> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>>>>>>
>>>>>>  After change from hadoop20 to hadoop23 the warning dissapeared but I
>>>>>>> got the same exception (Found interface
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>>>>>>
>>>>>>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>>>>>>> compiled by me, no other product nor configuration, just two
>>>>>>> servers, one master with ResourceManager and NameNode, one slave
>>>>>>> with DataNode and NodeManager.
>>>>>>>
>>>>>>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>>>>>>> is the new trace:
>>>>>>>
>>>>>>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>>>>>>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>>>>>>> to run jobs.
>>>>>>> java.lang.RuntimeException: java.lang.reflect.
>>>>>>> InvocationTargetException
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>>> PigJobControl.java:130)
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.run(
>>>>>>> PigJobControl.java:191)
>>>>>>>       at java.lang.Thread.run(Thread.java:724)
>>>>>>>       at org.apache.pig.backend.hadoop.executionengine.
>>>>>>> mapReduceLayer.
>>>>>>> MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>>>>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke(
>>>>>>> NativeMethodAccessorImpl.java:57)
>>>>>>>       at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>>>> DelegatingMethodAccessorImpl.java:43)
>>>>>>>       at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>>> PigJobControl.java:128)
>>>>>>>       ... 3 more
>>>>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>>       at org.apache.pig.backend.hadoop.executionengine.
>>>>>>> mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(
>>>>>>> JobSubmitter.java:491)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(
>>>>>>> JobSubmitter.java:508)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
>>>>>>> JobSubmitter.java:392)
>>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>>>>>>       at java.security.AccessController.doPrivileged(Native Method)
>>>>>>>       at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>>       at org.apache.hadoop.security.UserGroupInformation.doAs(
>>>>>>> UserGroupInformation.java:1491)
>>>>>>>       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>>>>>>       at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.
>>>>>>> submit(ControlledJob.java:335)
>>>>>>>       ... 8 more
>>>>>>>
>>>>>>>
>>>>>>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>>>>>>
>>>>>>>  Yes it does. You need to recompile Pig for hadoop 2
>>>>>>>>
>>>>>>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>>>>>>> claudio.r...@namictech.com> wrote:
>>>>>>>>
>>>>>>>>   Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I
>>>>>>>>
>>>>>>>>> get an
>>>>>>>>> error even with simple operations like
>>>>>>>>>
>>>>>>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>>>>>>> dump data;
>>>>>>>>>
>>>>>>>>> I only got a warning first and then and exception:
>>>>>>>>>
>>>>>>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**
>>>>>>>>> hadoop20.PigJobControl
>>>>>>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>>>>>>> java.lang.**NoSuchFieldException: runnerState
>>>>>>>>>       at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>>>>>>       at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>>>>>>> clinit>(PigJobControl.java:51)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>>>>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**
>>>>>>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**
>>>>>>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>>>>>>       at org.apache.pig.PigServer.**launchPlan(PigServer.java:**
>>>>>>>>> 1264)
>>>>>>>>>       at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>>>>>>> PigServer.java:1249)
>>>>>>>>>       at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>>>>>>       at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>>>>>>       at org.apache.pig.PigServer.**openIterator(PigServer.java:**
>>>>>>>>> 811)
>>>>>>>>>       at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>>>> GruntParser.java:696)
>>>>>>>>>       at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>>> PigScriptParser.parse(**
>>>>>>>>> PigScriptParser.java:320)
>>>>>>>>>       at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:194)
>>>>>>>>>       at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:170)
>>>>>>>>>       at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>>       at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>>       at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>>       at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>>       at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>>       at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>>>
>>>>>>>>> ------------------------------**--------
>>>>>>>>>
>>>>>>>>> Backend error message during job submission
>>>>>>>>> ------------------------------**-------------
>>>>>>>>> Unexpected System Error Occured: java.lang.**
>>>>>>>>> IncompatibleClassChangeError:
>>>>>>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but
>>>>>>>>> class
>>>>>>>>> was
>>>>>>>>> expected
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>>>>>>> PigOutputFormat.java:225)
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>>>>>>> PigOutputFormat.java:186)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> JobSubmitter.checkSpecs(**
>>>>>>>>> JobSubmitter.java:456)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**JobSubmitter.**
>>>>>>>>> submitJobInternal(
>>>>>>>>> **JobSubmitter.java:342)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job$10.run(Job.java:1268)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job$10.run(Job.java:1265)
>>>>>>>>>           at java.security.**AccessController.doPrivileged(
>>>>>>>>> **Native
>>>>>>>>> Method)
>>>>>>>>>           at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>>>>>>           at org.apache.hadoop.security.**
>>>>>>>>> UserGroupInformation.doAs(**
>>>>>>>>> UserGroupInformation.java:**1491)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job.submit(Job.java:1265)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> lib.jobcontrol.ControlledJob.**
>>>>>>>>> submit(ControlledJob.java:335)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> lib.jobcontrol.JobControl.run(**
>>>>>>>>> JobControl.java:240)
>>>>>>>>>           at org.apache.pig.backend.**
>>>>>>>>> hadoop20.PigJobControl.run(**
>>>>>>>>> PigJobControl.java:121)
>>>>>>>>>           at java.lang.Thread.run(Thread.**java:724)
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**MapReduceLauncher$1.run(**
>>>>>>>>> MapReduceLauncher.java:257)
>>>>>>>>>
>>>>>>>>> Pig Stack Trace
>>>>>>>>> ---------------
>>>>>>>>> ERROR 1066: Unable to open iterator for alias data
>>>>>>>>>
>>>>>>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR
>>>>>>>>> 1066:
>>>>>>>>> Unable to open iterator for alias data
>>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>>> openIterator(PigServer.java:**836)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.processDump(**
>>>>>>>>> GruntParser.java:696)
>>>>>>>>>           at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:194)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:170)
>>>>>>>>>           at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>>           at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>>           at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**
>>>>>>>>> invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>>           at sun.reflect.**DelegatingMethodAccessorImpl.*
>>>>>>>>> *invoke(**
>>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>>           at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>>           at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>>> Caused by: java.io.IOException: Job terminated with anomalous
>>>>>>>>> status
>>>>>>>>> FAILED
>>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>>> openIterator(PigServer.java:**828)
>>>>>>>>>           ... 12 more
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>


-- 
Regards,
Uttam Kumar

Reply via email to