Well that is unexpected. Is that consistently failing for you?
On Fri, Apr 19, 2013 at 6:29 PM, Jim Donahue <[email protected]> wrote: > Also the example seems to have stopped working … That's not a serious > problem (I can just ignore it) but it did work last week as far as I can > remember. :-) > > Waiting 5 seconds for it to start. . . . . . > Alright, now let's run the "wordcount" example via: > > $ ./bin/hadoop jar hadoop-examples-0.20.205.0.jar wordcount > src/contrib/mesos/src/java/org/apache/hadoop/mapred out > > Hit enter to continue. > 13/04/19 23:31:54 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 0 time(s). > 13/04/19 23:31:55 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 1 time(s). > 13/04/19 23:31:56 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 2 time(s). > 13/04/19 23:31:57 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 3 time(s). > 13/04/19 23:31:58 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 4 time(s). > 13/04/19 23:31:59 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 5 time(s). > 13/04/19 23:32:00 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 6 time(s). > 13/04/19 23:32:01 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 7 time(s). > 13/04/19 23:32:02 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 8 time(s). > 13/04/19 23:32:03 INFO ipc.Client: Retrying connect to server: localhost/ > 127.0.0.1:54311. Already tried 9 time(s). > java.net.ConnectException: Call to localhost/127.0.0.1:54311 failed on > connection exception: java.net.ConnectException: Connection refused > at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095) > at org.apache.hadoop.ipc.Client.call(Client.java:1071) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) > at org.apache.hadoop.mapred.$Proxy1.getProtocolVersion(Unknown Source) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379) > at org.apache.hadoop.mapred.JobClient.createRPCProxy(JobClient.java:478) > at org.apache.hadoop.mapred.JobClient.init(JobClient.java:472) > at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:455) > at org.apache.hadoop.mapreduce.Job$1.run(Job.java:478) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:416) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059) > at org.apache.hadoop.mapreduce.Job.connect(Job.java:476) > at org.apache.hadoop.mapreduce.Job.submit(Job.java:464) > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:495) > at org.apache.hadoop.examples.WordCount.main(WordCount.java:67) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:616) > at > org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68) > at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139) > at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:64) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:616) > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > Caused by: java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:604) > at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434) > at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560) > at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184) > at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202) > at org.apache.hadoop.ipc.Client.call(Client.java:1046) > ... 27 more > > Oh no, it failed! Try running the JobTracker and wordcount > example manually ... it might be an issue with your environment that > this tutorial didn't cover (if you find this to be the case, please > create a JIRA for us and/or send us a code review). > > ./TUTORIAL.sh: line 662: kill: (1522) - No such process > make: *** [hadoop-0.20.205.0] Error 1 > > > On 4/19/13 4:29 PM, "Benjamin Mahler" <[email protected]<mailto: > [email protected]>> wrote: > > Brenden: It looks like Maven isn't required when building > hadoop-0.20.205.0, can you send a patch to fix your change to only check > for Maven when building the CDH releases? > > Jim: Thanks for the report. > > I committed a recent change by Brenden here, which enforces that both 'ant' > and 'mvn' are present when building the hadoop port: > https://reviews.apache.org/r/10558/ > > > On Fri, Apr 19, 2013 at 3:51 PM, Jim Donahue <[email protected]<mailto: > [email protected]>> wrote: > > I was -- the last build I did was ten days ago. Somebody broke the build > scripts that I've been using for quite a while. > > Jim > > > > On 4/19/13 3:48 PM, "Benjamin Mahler" <[email protected]<mailto: > [email protected]>> wrote: > > >You can fix this by installing Maven. > > > >However, I was under the assumption that we required Maven in order to run > >the Hadoop tutorial. You were successfully building hadoop without Maven > >installed? > > > > > >On Fri, Apr 19, 2013 at 3:44 PM, Jim Donahue <[email protected]<mailto: > [email protected]>> wrote: > > > >> I'm trying to build Mesos on Amazon Linux and it appears that the Hadoop > >> build script has changed. It worked just fine a few days ago, but now > >>I'm > >> getting: > >> > >> sudo make hadoop-0.20.205.0 > >> if test ".." != ".."; then \ > >> cp -p ./TUTORIAL.sh .; \ > >> cp -p ./hadoop-gridmix.patch .; \ > >> cp -p ./hadoop-7698-1.patch .; \ > >> cp -p ./hadoop-0.20.205.0_hadoop-env.sh.patch .; \ > >> cp -p ./hadoop-0.20.205.0_mesos.patch .; \ > >> cp -p ./mapred-site.xml.patch .; \ > >> cp -rp ./mesos .; \ > >> cp -p ./mesos-executor .; \ > >> fi > >> rm -rf hadoop-0.20.205.0 > >> which: no mvn in (/sbin:/bin:/usr/sbin:/usr/bin) > >> > >> We seem to be missing mvn from the path. Please install > >> mvn and re-run this tutorial. If you still have troubles, please report > >> this to: > >> > >> [email protected]<mailto:[email protected]> > >> > >> (Remember to include as much debug information as possible.) > >> > >> Help, please! > >> > >> > >> Jim > >> > > > >
