Looks like version incompatible issue, please check the hadoop version you
installed, and hadoop version you use in your eclipse. Make sure they are
the same.

On Tue, Dec 30, 2014 at 5:38 PM, 李运田 <cumt...@163.com> wrote:

> my eclipse and pig are in same linux.
> this is my pig configuration in  eclipse:
>  props.setProperty("fs.defaultFS", "hdfs://10.210.90.101:8020");
>      props.setProperty("hadoop.job.user", "hadoop");
>      props.setProperty("mapreduce.framework.name", "yarn");
>      props.setProperty("yarn.resourcemanager.hostname", "10.210.90.101");
>      props.setProperty("yarn.resourcemanager.admin.address", "
> 10.210.90.101:8141");
>         props.setProperty("yarn.resourcemanager.address", "
> 10.210.90.101:8050");
>      props.setProperty("yarn.resourcemanager.resource-tracker.address", "
> 10.210.90.101:8025");
>      props.setProperty("yarn.resourcemanager.scheduler.address", "
> 10.210.90.101:8030");
> I have added core-site.xml、 yarn-site.xml、。。。。。into eclipse project.
> I can run pig script in " grunt> "
> but,when I run
> pigServer = new PigServer( ExecType.MAPREDUCE, props);
>    pigServer.registerQuery("tmp= LOAD '/user/hadoop/aa.txt';");
>    pigServer.registerQuery("tmp_table_limit = order tmp by $0;");
>    pigServer.store("tmp_table_limit", "/user/hadoop/shi.txt");
> I always get error:
> 14/12/30 17:28:33 WARN hadoop20.PigJobControl: falling back to default
> JobControl (not using hadoop 0.20 ?)
> java.lang.NoSuchFieldException: runnerState
>  at java.lang.Class.getDeclaredField(Class.java:1948)
>  at
> org.apache.pig.backend.hadoop20.PigJobControl.<clinit>(PigJobControl.java:51)
>
>
>
>
>
>
>
>         help!!!!!!




-- 
Best Regards

Jeff Zhang

Reply via email to