Depends on what the script does? Of it's launching a job on a remote cluster then yes.
Bit of script does something more and needs to run for longer than no. But if script it on a remote system, not what you asked but an alternative see https://stackoverflow.com/questions/39574653/error-executing-pigserver-in-java On Sat, 15 Feb, 2020, 19:16 Daniel Santos, <daniel.d...@gmail.com> wrote: > Hello, > > What I was thinking was : launching the pig script on my laptop, the > hadoop cluster would be left executing it, and I could shut down the laptop. > > Is this possible ? > > Thanks, > Regards > > On 12 Feb 2020, at 02:06, Shashwat Shriparv <dwivedishash...@gmail.com> > wrote: > > nohup <your pig command> & > > > *Warm Regards,* > *Shashwat Shriparv* > *http://bit.ly/14cHpad <http://bit.ly/14cHpad> * > > *http://goo.gl/rxz0z8 <http://goo.gl/rxz0z8>* > *http://goo.gl/RKyqO8 <http://goo.gl/RKyqO8>* > http://helpmetocode.blogspot.in/ > http://photoinfinity.blogspot.in/ > http://writingishabit.blogspot.in/ > http://realiq.blogspot.in/ > http://sshriparv.blogspot.in/ > https://goo.gl/M8Us3B > https://goo.gl/nrI2mv > https://500px.com/shriparv > https://www.flickr.com/photos/55141469@N02/ > https://about.me/shriparv > ISBN - 10: 1783985941 > ISBN - 13: 9781783985944 > [image: https://www.linkedin.com/pub/shashwat-shriparv/19/214/2a9] > <https://www.linkedin.com/pub/shashwat-shriparv/19/214/2a9>[image: > https://twitter.com/shriparv] <https://twitter.com/shriparv>[image: > https://www.facebook.com/shriparv] <https://www.facebook.com/shriparv>[image: > http://google.com/+ShashwatShriparv] > <http://google.com/+ShashwatShriparv>[image: > http://www.youtube.com/user/sShriparv/videos] > <http://www.youtube.com/user/sShriparv/videos>[image: > http://profile.yahoo.com/SWXSTW3DVSDTF2HHSRM47AV6DI/] <shrip...@yahoo.com> > > > > On Wed, 12 Feb 2020 at 04:48, Daniel Santos <daniel.d...@gmail.com> wrote: > >> Hello, >> >> I managed to create a properties file with the following contents : >> >> fs.defaultFS=hdfs://hadoopnamenode:9000 >> mapreduce.framework.name=yarn >> yarn.resourcemanager.address=hadoopresourcemanager:8032 >> >> It is now submitting the jobs to the cluster. I also set the HADOOP_HOME >> on my laptop to point to the same version of hadoop that is running on the >> cluster (2.7.0). I am running pig version 0.17 >> >> Then a main class not found error happened on the yarn nodes where the >> job was scheduled to run. I had to add the following to yarn-site.xml and >> restart yarn and the nodes : >> >> <property> >> <name>mapreduce.application.classpath</name> >> >> <value>/home/hadoop/hadoop-2.7.0/share/hadoop/mapreduce/*,/home/hadoop/hadoop-2.7.0/share/hadoop/mapreduce/lib/*,/home/hadoop/hadoop-2.7.0/share/hadoop/common/*,/home/hadoop/hadoop-2.7.0/share/hadoop/common/lib/*,/home/hadoop/hadoop-2.7.0/share/hadoop/yarn/*,/home/hadoop/hadoop-2.7.0/share/hadoop/yarn/lib/*,/home/hadoop/hadoop-2.7.0/share/hadoop/hdfs/*,/home/hadoop/hadoop-2.7.0/share/hadoop/hdfs/lib/*</value> >> </property> >> >> After this change, the script ran. But the pig command only returned >> after the job finished. >> Does anyone know how to launch the script and exit immediately to the >> shell ? >> If the job takes a long time I will have to keep the terminal open. >> >> Thanks, >> Regards >> >> >> > On 11 Feb 2020, at 05:25, Vinod Kumar Vavilapalli <vino...@apache.org> >> wrote: >> > >> > It’s running the job in local mode (LocalJobRunner), that’s why. Please >> check your configuration files and make sure that the right directories are >> on the classpath. Also look in mapred-site.xml for >> mapreduce.framework.name (should be yarn). >> > >> > Thanks >> > +Vinod >> > >> >> On Feb 11, 2020, at 2:09 AM, Daniel Santos <daniel.d...@gmail.com> >> wrote: >> >> >> >> Hello all, >> >> >> >> I have developed a script in my laptop. The script is now ready to be >> unleashed in a non secured cluster. >> >> But when I do : pig -x mapreduce <script name> it doesn’t return to >> the shell immediately. It prints stuff like [LocalJobRunner Map Task >> Executor #0] >> >> >> >> I have exported the PIG_CLASSPATH shell variable to point to a >> directory with the cluster’s configuration and its operating on the files >> located there. >> >> But I would expect the job to be launched, the shell prompt returned >> and the job would be left executing independently on the cluster. >> >> >> >> Another thing I noticed while developing the script and running it >> both locally and on the cluster, is that the web page for there resource >> manager does not show the map reduce jobs that pig generates. What do I >> have to do to be able to see them ? >> >> >> >> Thanks, >> >> Regards >> >> --------------------------------------------------------------------- >> >> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org >> >> For additional commands, e-mail: user-h...@hadoop.apache.org >> >> >> > >> >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org >> For additional commands, e-mail: user-h...@hadoop.apache.org >> >> >