Pardon me, for a YARN cluster, that should actually be "yarn rmadmin -refreshNodes".
See this page for usage details: http://hadoop.apache.org/docs/r2.3.0/hadoop-yarn/hadoop-yarn-site/YarnCommands.html On Tue, May 20, 2014 at 6:28 PM, Gary Helmling <[email protected]> wrote: > Hi Srinivas, > > If you're trying to remove a node from your YARN cluster, you'll need to > run "hadoop mradmin -refreshNodes". The command that you ran ("hadoop > dfsadmin") is for adding/removing nodes from the HDFS service. > > > On Tue, May 20, 2014 at 4:56 PM, Srinivas Reddy Kancharla < > [email protected]> wrote: > >> Thanks Terence for your clarification. >> I tried to remove the node from the cluster by removing an entry from >> "slaves" file and then ran "hadoop dfsadmin -refreshNodes" but looks like >> this is not the right command. >> Is there any specific command I need to use to remove or add a node , >> *without >> restarting the services* ? >> >> I coded in such a way that new Runnables should get launched if they see >> new nodes but I am stuck with basic yarn command. >> >> Thanks and regards, >> Srini >> >> >> On Tue, May 20, 2014 at 12:02 AM, Terence Yim <[email protected]> wrote: >> >> > Hi Srinivas, >> > >> > Sorry for the late reply. BTW, I just noticed that this discussion is >> > not on the dev@ mailing list, hence I CC my reply to the mailing list. >> > You could subscribe to the list by sending an email to >> > [email protected] >> > >> > To your question about rebalancing, currently Twill won't stop >> > executing Runnable and move it to run on newly available resource, as >> > it doesn't know what the Runnable is doing and whether it is close to >> > finish or not. After you added a new node to the cluster, only newly >> > launched runnable (either a new application run or increase number of >> > instances of existing runnable) may runs on the new node (up to YARN >> > to allocate). >> > >> > Terence >> > >> > >> > On Fri, May 16, 2014 at 1:31 PM, Srinivas Reddy Kancharla >> > <[email protected]> wrote: >> > > HI Terence, >> > > >> > > Thanks for the information you have provided and now I could execute >> my >> > > programs. I am trying to experiment on re-balance behavior, your input >> > will >> > > really help me to test further: >> > > >> > > - I created a my own TwillApplication which launches 3 >> > > AbstractTwillRunnables (say this program is time consuming job). >> > > - I have a setup of 3 nodes (one master and 2 slave). When I launch my >> > > program, I could see that : >> > > > First slave node has launched ApplicationMaster and one >> Runnable. >> > > > Second Slave node has taken care of launch other 2 runnables. >> > > >> > > - During execution of above application, If I add 3rd slave node to >> > cluster, >> > > and configure it for re-balance, will this re-balance process will >> take >> > care >> > > of re-distributing of runnables again? i.e. now in this scenario the >> > > second slave node will have only one runnable and third new slave node >> > > should take care of one of the runnable. This way the load is >> > distributed. >> > > >> > > Thanks and regards, >> > > Srini >> > > >> > > >> > > >> > > >> > > On Fri, May 9, 2014 at 12:15 AM, Terence Yim <[email protected]> >> wrote: >> > >> >> > >> Hi Srinivas, >> > >> >> > >> First of all, through I never tried, I won't expect a YARN app could >> > >> work correctly on local cluster after computer sleep and wake. >> > >> >> > >> The exception is about RM tries to restart the AM after wake up >> (maybe >> > >> it though the AM is dead, as it has't been heartbeating when the >> > >> computer sleep, and RM uses wall clock to check), however the restart >> > >> failed due to token expiration (when someone asked RM for a >> container, >> > >> it comes with a timed token). The expiration time is governed by the >> > >> setting >> > "yarn.resourcemanager.rm.container-allocation.expiry-interval-ms" >> > >> and default is 600 seconds. >> > >> >> > >> Terence >> > >> >> > >> On Thu, May 8, 2014 at 11:45 AM, Srinivas Reddy Kancharla >> > >> <[email protected]> wrote: >> > >> > HI Terence, >> > >> > >> > >> > Yesterda the same program was working. Today when I opened my >> MacBook >> > >> > and so >> > >> > my 3 VM nodes are running back, I am seeing below exception as >> shown: >> > >> > >> > >> > I am getting below exception, is there any configuration which can >> > >> > ignore >> > >> > such exception??: >> > >> > >> > >> > Got exception: org.apache.hadoop.yarn.exceptions.YarnException: >> > >> > Unauthorized >> > >> > request to start container. >> > >> > This token is expired. current time is 1399573775978 found >> > 1399573627677 >> > >> > >> > >> > 2014-05-08 11:17:07,682 INFO >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: >> > >> > Command >> > >> > to launch container container_1399572736534_0002_02_000001 : >> > >> > $JAVA_HOME/bin/java -Djava.io.tmpdir=tmp >> -Dyarn.appId=$YARN_APP_ID_STR >> > >> > -Dtwill.app=$TWILL_APP_NAME -cp launcher.jar:$HADOOP_CONF_DIR >> -Xmx362m >> > >> > org.apache.twill.launcher.TwillLauncher appMaster.jar >> > >> > org.apache.twill.internal.appmaster.ApplicationMasterMain false >> > >> > 1><LOG_DIR>/stdout 2><LOG_DIR>/stderr >> > >> > 2014-05-08 11:17:07,694 INFO >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: >> > >> > Error >> > >> > launching appattempt_1399572736534_0002_000002. Got exception: >> > >> > org.apache.hadoop.yarn.exceptions.YarnException: Unauthorized >> request >> > to >> > >> > start container. >> > >> > This token is expired. current time is 1399573775978 found >> > 1399573627677 >> > >> > at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> > >> > Method) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> > >> > at >> java.lang.reflect.Constructor.newInstance(Constructor.java: >> > >> > 534) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:152) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:122) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:249) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1146) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> > >> > at java.lang.Thread.run(Thread.java:701) >> > >> > >> > >> > 2014-05-08 11:17:07,695 INFO >> > >> > >> > org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService: >> > >> > Unregistering app attempt : appattempt_1399572736534_0002_000002 >> > >> > 2014-05-08 11:17:07,695 INFO >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: >> > >> > appattempt_1399572736534_0002_000002 State change from ALLOCATED to >> > >> > FAILED >> > >> > 2014-05-08 11:17:07,695 INFO >> > >> > org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: >> > >> > Application >> > >> > application_1399572736534_0002 failed 2 times due to Error >> launching >> > >> > appattempt_1399572736534_0002_000002. Got exception: >> > >> > org.apache.hadoop.yarn.exceptions.YarnException: Unauthorized >> request >> > to >> > >> > start container. >> > >> > This token is expired. current time is 1399573775978 found >> > 1399573627677 >> > >> > at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> > >> > Method) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> > >> > at >> > java.lang.reflect.Constructor.newInstance(Constructor.java:534) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:152) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:122) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:249) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1146) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> > >> > at java.lang.Thread.run(Thread.java:701) >> > >> > . Failing the application. >> > >> > 2014-05-08 11:17:07,695 INFO >> > >> > org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: >> > >> > application_1399572736534_0002 State change from ACCEPTED to FAILED >> > >> > 2014-05-08 11:17:07,695 WARN >> > >> > org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: >> > >> > USER=srini OPERATION=Application Finished - >> > >> > Failed TARGET=RMAppManager RESULT=FAILURE >> DESCRIPTION=App >> > >> > failed with state: >> > >> > FAILED PERMISSIONS=Application >> application_1399572736534_0002 >> > >> > failed 2 times >> > >> > due to Error launching appattempt_1399572736534_0002_000002. Got >> > >> > exception: >> > >> > org.apache.hadoop.yarn.exceptions.YarnException: Unauthorized >> request >> > to >> > >> > start container. >> > >> > This token is expired. current time is 1399573775978 found >> > 1399573627677 >> > >> > at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> > >> > Method) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> > >> > at >> > >> > >> > >> > >> > >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> > >> > at >> > java.lang.reflect.Constructor.newInstance(Constructor.java:534) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:152) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:122) >> > >> > at >> > >> > >> > >> > >> > >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:249) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1146) >> > >> > at >> > >> > >> > >> > >> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> > >> > at java.lang.Thread.run(Thread.java:701) >> > >> > >> > >> > >> > >> > >> > >> > On Wed, May 7, 2014 at 1:35 PM, Srinivas Reddy Kancharla >> > >> > <[email protected]> wrote: >> > >> >> >> > >> >> I got answer for one my own question: >> > >> >> Can I expect "Hello world" on master node where I launched the >> > program >> > >> >> ? >> > >> >> >> > >> >> After I copied jopt-simple.jar of proper version, it worked and I >> can >> > >> >> see >> > >> >> "Hello world" output on master node. Sorry for the spam. >> > >> >> >> > >> >> Srini >> > >> >> >> > >> >> >> > >> >> On Wed, May 7, 2014 at 1:12 PM, Srinivas Reddy Kancharla >> > >> >> <[email protected]> wrote: >> > >> >>> >> > >> >>> Exciting.. it worked after I got all required jars. Advantage of >> not >> > >> >>> using maven project is , faced all these issues and exposed to >> all >> > the >> > >> >>> required jars and exceptions. >> > >> >>> >> > >> >>> Now when I launched my program, it got executed in one of my >> slave >> > >> >>> node. >> > >> >>> Both application master and task ran on the same node and I could >> > see >> > >> >>> "Hello >> > >> >>> world" in "stdout" log. >> > >> >>> >> > >> >>> Can I expect "Hello world" on master node where I launched the >> > program >> > >> >>> ? >> > >> >>> >> > >> >>> Thanks again for all your help. From here I will try different >> > >> >>> programs >> > >> >>> with different options and will see how it goes. >> > >> >>> >> > >> >>> Is there any particular forum where I can ask questions or >> should be >> > >> >>> fine >> > >> >>> to send you questions ? It was a great help from you. >> > >> >>> >> > >> >>> I am doing all this during my free time (i.e. after office >> hours). I >> > >> >>> would like to try more and so if possible please let me know if I >> > can >> > >> >>> be >> > >> >>> helpful in anyway. >> > >> >>> >> > >> >>> Regards, >> > >> >>> Srini >> > >> >>> >> > >> >>> >> > >> >>> >> > >> >>> On Wed, May 7, 2014 at 1:06 AM, Terence Yim <[email protected]> >> > wrote: >> > >> >>>> >> > >> >>>> Hi Srinivas, >> > >> >>>> >> > >> >>>> It’s the ASM library version issue. Try to include the >> > >> >>>> asm-4.0-all.jar >> > >> >>>> in your classpath before the hadoop classpath. >> > >> >>>> >> > >> >>>> http://mvnrepository.com/artifact/org.ow2.asm/asm-all/4.0 >> > >> >>>> >> > >> >>>> Terence >> > >> >>>> >> > >> >>>> On May 6, 2014, at 4:22 PM, Srinivas Reddy Kancharla >> > >> >>>> <[email protected]> wrote: >> > >> >>>> >> > >> >>>> Hi Terence, >> > >> >>>> >> > >> >>>> After all step-by-step downloading of required jar files (b'cos >> I >> > am >> > >> >>>> not >> > >> >>>> using maven for now), I am able to pass through the zookeeper >> issue >> > >> >>>> (have a >> > >> >>>> setup of 3 nodes i.e. one leader and 2 followers) and now I am >> > >> >>>> seeing below >> > >> >>>> exception: (Any pointer for this would be helpful for me). >> > >> >>>> >> > >> >>>> I suspect on the hadoop libraries I am using, b'cos from the pom >> > >> >>>> files >> > >> >>>> which you have created for hello world examples is referring to >> > >> >>>> hadoop 2.3 >> > >> >>>> ... whereas I am using Hadoop 2.2 .. Do you think below >> exception >> > is >> > >> >>>> due to >> > >> >>>> that reason? >> > >> >>>> >> > >> >>>> >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:java.io.tmpdir=/tmp >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:java.compiler=<NA> >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:os.name=Linux >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:os.arch=amd64 >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:os.version=3.11.0-12-generic >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:user.name=srini >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:user.home=/home/srini >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Client >> > >> >>>> environment:user.dir=/home/srini/twill/twilljars >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ZooKeeper: Initiating client >> > >> >>>> connection, connectString=localhost:2181 sessionTimeout=10000 >> > >> >>>> watcher=ServiceDelegate [STARTING] >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>> connection >> > >> >>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > authenticate >> > >> >>>> using >> > >> >>>> SASL (unknown error) >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ClientCnxn: Socket connection >> > >> >>>> established to localhost/127.0.0.1:2181, initiating session >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.ClientCnxn: Session >> establishment >> > >> >>>> complete on server localhost/127.0.0.1:2181, sessionid = >> > >> >>>> 0x145d3a544bd0006, >> > >> >>>> negotiated timeout = 10000 >> > >> >>>> 14/05/06 15:53:39 INFO zookeeper.DefaultZKClientService: >> Connected >> > to >> > >> >>>> ZooKeeper: localhost:2181 >> > >> >>>> Exception in thread " STARTING" >> > >> >>>> java.lang.IncompatibleClassChangeError: >> > >> >>>> class >> > >> >>>> >> org.apache.twill.internal.utils.Dependencies$DependencyClassVisitor >> > >> >>>> has interface org.objectweb.asm.ClassVisitor as super class >> > >> >>>> at java.lang.ClassLoader.defineClass1(Native Method) >> > >> >>>> at java.lang.ClassLoader.defineClass(ClassLoader.java:643) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> > >> >>>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >> > >> >>>> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >> > >> >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >> > >> >>>> at java.security.AccessController.doPrivileged(Native Method) >> > >> >>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >> > >> >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:323) >> > >> >>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:294) >> > >> >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:268) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.utils.Dependencies.findClassDependencies(Dependencies.java:102) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.ApplicationBundler.findDependencies(ApplicationBundler.java:179) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.ApplicationBundler.createBundle(ApplicationBundler.java:136) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.ApplicationBundler.createBundle(ApplicationBundler.java:106) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.yarn.YarnTwillPreparer.createAppMasterJar(YarnTwillPreparer.java:366) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.yarn.YarnTwillPreparer.access$2(YarnTwillPreparer.java:350) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.yarn.YarnTwillPreparer$1.call(YarnTwillPreparer.java:263) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > org.apache.twill.yarn.YarnTwillPreparer$1.call(YarnTwillPreparer.java:1) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.yarn.YarnTwillController.doStartUp(YarnTwillController.java:98) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.AbstractZKServiceController.startUp(AbstractZKServiceController.java:82) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> org.apache.twill.internal.AbstractExecutionServiceController$ServiceDelegate.startUp(AbstractExecutionServiceController.java:109) >> > >> >>>> at >> > >> >>>> >> > >> >>>> >> > >> com.google.common.util.concurrent.AbstractIdleService$1$1.run(AbstractIdleService.java:43) >> > >> >>>> at java.lang.Thread.run(Thread.java:701) >> > >> >>>> >> > >> >>>> >> > >> >>>> Thanks and regards, >> > >> >>>> Srini >> > >> >>>> >> > >> >>>> >> > >> >>>> On Tue, May 6, 2014 at 2:40 PM, Srinivas Reddy Kancharla >> > >> >>>> <[email protected]> wrote: >> > >> >>>>> >> > >> >>>>> Got it. I will do that and will update you. Earlier my >> assumption >> > >> >>>>> was >> > >> >>>>> my hadoop cluster would be starting zookeeper as part of >> Namenode, >> > >> >>>>> Datanode, >> > >> >>>>> resourcemanager , Nodemanager initialization. Seems like I was >> > wrong >> > >> >>>>> and I >> > >> >>>>> have to start zookeeper as a separate process. >> > >> >>>>> >> > >> >>>>> Thanks again for this information. >> > >> >>>>> >> > >> >>>>> Regards, >> > >> >>>>> Srini >> > >> >>>>> >> > >> >>>>> >> > >> >>>>> >> > >> >>>>> On Tue, May 6, 2014 at 2:32 PM, Terence Yim <[email protected]> >> > >> >>>>> wrote: >> > >> >>>>>> >> > >> >>>>>> Hi Srinivas, >> > >> >>>>>> >> > >> >>>>>> Yes you'll need to start zookeeper manually before executing >> the >> > >> >>>>>> twill >> > >> >>>>>> program. The assumption is that zookeeper is a long running >> > service >> > >> >>>>>> in the >> > >> >>>>>> cluster >> > >> >>>>>> >> > >> >>>>>> Terence >> > >> >>>>>> >> > >> >>>>>> Sent from my iPhone >> > >> >>>>>> >> > >> >>>>>> On May 6, 2014, at 2:14 PM, Srinivas Reddy Kancharla >> > >> >>>>>> <[email protected]> wrote: >> > >> >>>>>> >> > >> >>>>>> HI Terence, >> > >> >>>>>> >> > >> >>>>>> Thank you very much for the pointer. So i have used "hadoop >> > >> >>>>>> classpath" >> > >> >>>>>> command and copied that list to my "java" command and atleast >> > now I >> > >> >>>>>> am out >> > >> >>>>>> of classpath issues. So this shows that I am fine with my >> current >> > >> >>>>>> version of >> > >> >>>>>> Hadoop 2.2 jars. >> > >> >>>>>> >> > >> >>>>>> Now as I asked in my previous mail, Do I need to start >> > "zookeeper" >> > >> >>>>>> separately or its a part of my existing hadoop running >> cluster ?? >> > >> >>>>>> b'cos I am >> > >> >>>>>> getting below exception for my "Hello world" example (I have >> > taken >> > >> >>>>>> your >> > >> >>>>>> example of "localhost:2181" for ZKServer string: >> > >> >>>>>> >> > >> >>>>>> >> > >> >>>>>> >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> >> > >> >>>>>> >> > >> environment:java.library.path=/usr/lib/jvm/java-6-openjdk-amd64/jre/lib/amd64/server:/usr/lib/jvm/java-6-openjdk-amd64/jre/lib/amd64:/usr/lib/jvm/java-6-openjdk-amd64/jre/../lib/amd64:/usr/java/packages/lib/amd64:/usr/lib/jni:/lib:/usr/lib >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:java.io.tmpdir=/tmp >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:java.compiler=<NA> >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:os.name=Linux >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:os.arch=amd64 >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:os.version=3.11.0-12-generic >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:user.name=srini >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:user.home=/home/srini >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Client >> > >> >>>>>> environment:user.dir=/home/srini/twill/twilljars >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ZooKeeper: Initiating client >> > >> >>>>>> connection, connectString=localhost:2181 sessionTimeout=10000 >> > >> >>>>>> watcher=ServiceDelegate [STARTING] >> > >> >>>>>> 14/05/06 14:08:11 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>>>> connection >> > >> >>>>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > >> >>>>>> authenticate using >> > >> >>>>>> SASL (unknown error) >> > >> >>>>>> 14/05/06 14:08:11 WARN zookeeper.ClientCnxn: Session 0x0 for >> > server >> > >> >>>>>> null, unexpected error, closing socket connection and >> attempting >> > >> >>>>>> reconnect >> > >> >>>>>> java.net.ConnectException: Connection refused >> > >> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:601) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > >> >>>>>> 14/05/06 14:08:12 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>>>> connection >> > >> >>>>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > >> >>>>>> authenticate using >> > >> >>>>>> SASL (unknown error) >> > >> >>>>>> 14/05/06 14:08:12 WARN zookeeper.ClientCnxn: Session 0x0 for >> > server >> > >> >>>>>> null, unexpected error, closing socket connection and >> attempting >> > >> >>>>>> reconnect >> > >> >>>>>> java.net.ConnectException: Connection refused >> > >> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:601) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > >> >>>>>> 14/05/06 14:08:13 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>>>> connection >> > >> >>>>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > >> >>>>>> authenticate using >> > >> >>>>>> SASL (unknown error) >> > >> >>>>>> 14/05/06 14:08:13 WARN zookeeper.ClientCnxn: Session 0x0 for >> > server >> > >> >>>>>> null, unexpected error, closing socket connection and >> attempting >> > >> >>>>>> reconnect >> > >> >>>>>> java.net.ConnectException: Connection refused >> > >> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:601) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > >> >>>>>> 14/05/06 14:08:14 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>>>> connection >> > >> >>>>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > >> >>>>>> authenticate using >> > >> >>>>>> SASL (unknown error) >> > >> >>>>>> 14/05/06 14:08:14 WARN zookeeper.ClientCnxn: Session 0x0 for >> > server >> > >> >>>>>> null, unexpected error, closing socket connection and >> attempting >> > >> >>>>>> reconnect >> > >> >>>>>> java.net.ConnectException: Connection refused >> > >> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:601) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > >> >>>>>> 14/05/06 14:08:15 INFO zookeeper.ClientCnxn: Opening socket >> > >> >>>>>> connection >> > >> >>>>>> to server localhost/127.0.0.1:2181. Will not attempt to >> > >> >>>>>> authenticate using >> > >> >>>>>> SASL (unknown error) >> > >> >>>>>> 14/05/06 14:08:15 WARN zookeeper.ClientCnxn: Session 0x0 for >> > server >> > >> >>>>>> null, unexpected error, closing socket connection and >> attempting >> > >> >>>>>> reconnect >> > >> >>>>>> java.net.ConnectException: Connection refused >> > >> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:601) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > >> >>>>>> at >> > >> >>>>>> >> > >> >>>>>> >> > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > >> >>>>>> >> > >> >>>>>> >> > >> >>>>>> Thank you again for your help and hopefully once I am out of >> this >> > >> >>>>>> initial setup issues, I will not bother you much unless its >> very >> > >> >>>>>> technical. >> > >> >>>>>> >> > >> >>>>>> Thanks and regards, >> > >> >>>>>> Srini >> > >> >>>>>> >> > >> >>>>>> >> > >> >>>>>> On Mon, May 5, 2014 at 10:34 PM, Terence Yim < >> [email protected]> >> > >> >>>>>> wrote: >> > >> >>>>>>> >> > >> >>>>>>> Hi Srinivas, >> > >> >>>>>>> >> > >> >>>>>>> Looks like you missed some hadoop classes in your classpath. >> > >> >>>>>>> You’ll >> > >> >>>>>>> need the hadoop classpath in your classpath. Have you try >> > running >> > >> >>>>>>> like this? >> > >> >>>>>>> >> > >> >>>>>>> HDCP=`hadoop classpath`; java -cp >> > >> >>>>>>> >> > >> >>>>>>> >> > >> ./SriniTwillYarnClasses.jar:twill-api-0.3.0-incubating-SNAPSHOT.jar:……:$HDCP >> > >> >>>>>>> com.srini.hadoopTwill.HelloTwill >> > >> >>>>>>> >> > >> >>>>>>> Terence >> > >> >>>>>>> >> > >> >>>>>>> On May 5, 2014, at 9:07 PM, Srinivas Reddy Kancharla >> > >> >>>>>>> <[email protected]> wrote: >> > >> >>>>>>> >> > >> >>>>>>> Hello Terence, >> > >> >>>>>>> >> > >> >>>>>>> I am Srini and new to twill. I am very sorry for sending you >> > email >> > >> >>>>>>> like this, b'cos I could not find any other discussion forum >> to >> > >> >>>>>>> post this >> > >> >>>>>>> message. My bad, please let me know if there is a forum exist >> > and >> > >> >>>>>>> I can get >> > >> >>>>>>> some help in future instead of direct mails to you. Below is >> the >> > >> >>>>>>> issue I am >> > >> >>>>>>> facing while executing my first Twill program: >> > >> >>>>>>> >> > >> >>>>>>> - I have a setup of hadoop-2.2.0 which has total 3 nodes. one >> > >> >>>>>>> master >> > >> >>>>>>> and 2 slave. >> > >> >>>>>>> - I could execute the DistributedShell program successfully. >> > >> >>>>>>> - Now I downloaded twill project, generated required jar >> files >> > >> >>>>>>> using >> > >> >>>>>>> mvn commands. >> > >> >>>>>>> - I replicated Helloworld sample program and during >> execution, I >> > >> >>>>>>> am >> > >> >>>>>>> getting below exception: >> > >> >>>>>>> >> > >> >>>>>>> srini@ubuntu:~/twill/twilljars$ java -classpath >> > >> >>>>>>> >> > >> >>>>>>> >> > >> ./SriniTwillYarnClasses.jar:twill-api-0.3.0-incubating-SNAPSHOT.jar:guava-13.0.1.jar:slf4j-api-1.7.7.jar:twill-yarn-0.3.0-incubating-SNAPSHOT.jar:hadoop-common-2.2.0.jar:hadoop-yarn-api-2.2.0.jar:twill-ext-0.3.0-incubating-SNAPSHOT.jar:twill-core-0.3.0-incubating-SNAPSHOT.jar:commons-logging-1.1.1.jar:commons-configuration-1.6.jar:commons-lang-2.5.jar:twill-common-0.3.0-incubating-SNAPSHOT.jar:twill-zookeeper-0.3.0-incubating-SNAPSHOT.jar:hadoop-auth-2.2.0.jar >> > >> >>>>>>> com.srini.hadoopTwill.HelloTwill >> > >> >>>>>>> >> > >> >>>>>>> SLF4J: Failed to load class >> "org.slf4j.impl.StaticLoggerBinder". >> > >> >>>>>>> SLF4J: Defaulting to no-operation (NOP) logger implementation >> > >> >>>>>>> SLF4J: See >> http://www.slf4j.org/codes.html#StaticLoggerBinderfor >> > >> >>>>>>> further details. >> > >> >>>>>>> May 5, 2014 8:49:53 PM >> org.apache.hadoop.util.NativeCodeLoader >> > >> >>>>>>> <clinit> >> > >> >>>>>>> WARNING: Unable to load native-hadoop library for your >> > platform... >> > >> >>>>>>> using builtin-java classes where applicable >> > >> >>>>>>> Exception in thread "main" java.lang.RuntimeException: >> > >> >>>>>>> java.lang.reflect.InvocationTargetException >> > >> >>>>>>> at >> > >> >>>>>>> >> com.google.common.base.Throwables.propagate(Throwables.java:160) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> org.apache.twill.internal.yarn.VersionDetectYarnAppClientFactory.create(VersionDetectYarnAppClientFactory.java:47) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> org.apache.twill.yarn.YarnTwillRunnerService.<init>(YarnTwillRunnerService.java:143) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> org.apache.twill.yarn.YarnTwillRunnerService.<init>(YarnTwillRunnerService.java:138) >> > >> >>>>>>> at com.srini.hadoopTwill.HelloTwill.main(HelloTwill.java:37) >> > >> >>>>>>> Caused by: java.lang.reflect.InvocationTargetException >> > >> >>>>>>> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> > >> >>>>>>> Method) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> > >> >>>>>>> at >> > java.lang.reflect.Constructor.newInstance(Constructor.java:534) >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> org.apache.twill.internal.yarn.VersionDetectYarnAppClientFactory.create(VersionDetectYarnAppClientFactory.java:44) >> > >> >>>>>>> ... 3 more >> > >> >>>>>>> Caused by: java.lang.Error: Unresolved compilation problems: >> > >> >>>>>>> The import org.apache.hadoop.yarn.api.records.DelegationToken >> > >> >>>>>>> cannot >> > >> >>>>>>> be resolved >> > >> >>>>>>> The import org.apache.hadoop.yarn.client.YarnClient cannot be >> > >> >>>>>>> resolved >> > >> >>>>>>> The import org.apache.hadoop.yarn.client.YarnClientImpl >> cannot >> > be >> > >> >>>>>>> resolved >> > >> >>>>>>> The import >> org.apache.hadoop.yarn.exceptions.YarnRemoteException >> > >> >>>>>>> cannot be resolved >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClientImpl cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> The method setUser(String) is undefined for the type >> > >> >>>>>>> ApplicationSubmissionContext >> > >> >>>>>>> The method getUser() is undefined for the type >> > >> >>>>>>> ApplicationSubmissionContext >> > >> >>>>>>> The method setResource(Resource) is undefined for the type >> > >> >>>>>>> ContainerLaunchContext >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnRemoteException cannot be resolved to a type >> > >> >>>>>>> The method getMinimumResourceCapability() is undefined for >> the >> > >> >>>>>>> type >> > >> >>>>>>> GetNewApplicationResponse >> > >> >>>>>>> The method getContainerTokens() is undefined for the type >> > >> >>>>>>> ContainerLaunchContext >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> The method setContainerTokens(ByteBuffer) is undefined for >> the >> > >> >>>>>>> type >> > >> >>>>>>> ContainerLaunchContext >> > >> >>>>>>> DelegationToken cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnRemoteException cannot be resolved to a type >> > >> >>>>>>> YarnClient cannot be resolved to a type >> > >> >>>>>>> YarnRemoteException cannot be resolved to a type >> > >> >>>>>>> >> > >> >>>>>>> at >> > >> >>>>>>> >> > >> >>>>>>> >> > >> org.apache.twill.internal.yarn.Hadoop20YarnAppClient.<init>(Hadoop20YarnAppClient.java:33) >> > >> >>>>>>> ... 8 more >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> Could you please let me know if I am missing anything here to >> > >> >>>>>>> execute >> > >> >>>>>>> this program. In my program, "localhost:2181" is hard coded >> for >> > >> >>>>>>> zookeeper >> > >> >>>>>>> string. >> > >> >>>>>>> >> > >> >>>>>>> My suspect: >> > >> >>>>>>> - My setup is having hadoop-2.2.0 , to execute this program, >> do >> > I >> > >> >>>>>>> need to provide hadoop-2.0 libraries instead of 2.2 . >> > >> >>>>>>> - Do I need to start zookeeper server separately ? >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> Thanks for your any help, >> > >> >>>>>>> >> > >> >>>>>>> Srini >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>>> >> > >> >>>>>> >> > >> >>>>> >> > >> >>>> >> > >> >>>> >> > >> >>> >> > >> >> >> > >> > >> > > >> > > >> > >> > >
