Kevin, What version of Pig are you using?
Have you tried setting the right MR home directory to point Pig to the local MR configuration for YARN? $ HADOOP_MAPRED_HOME=/usr/lib/hadoop-mapreduce $PIG_HOME/bin/pig Usually does it for me, so long as I have /usr/lib/hadoop-mapreduce/conf configured properly for YARN+MR (and considering that my YARN libs, etc. are all inside /usr/lib/hadoop-mapreduce). On Thu, May 3, 2012 at 12:11 AM, Kevin <kevin.macksa...@gmail.com> wrote: > Hi, > > I have a cluster running YARN, and mapreduce jobs run as expected when they > are executed from one of the nodes. However, when I run Pig scripts from a > remote client, Pig connects to HDFS and HBase but runs its MapReduce job > using the LocalJobRunner. Jobs finish successfully, but they aren't using > the YARN architecture. I have placed all the configuration files in the Pig > configuration directory, and this must be right otherwise Pig wouldn't > connect to my cluster's HDFS and HBase. > > I have even put "mapreduce.framework.name=yarn" in the pig.properties file. > > Any ideas to get jobs submitted to a remote Hadoop cluster to work in > distributed mode? > > -Kevin -- Harsh J