This was committed, as for your TUTORIAL.sh issues, can you confirm those
are consistently failing?

'make hadoop-0.20.205.0' is working correctly on my end.


On Fri, Apr 19, 2013 at 7:18 PM, Benjamin Mahler
<[email protected]>wrote:

> Alright, I'm seeing a separate issue with the patch:
>
> Patch conf/hadoop-env.sh? [N]
>
>   $ patch -p1 <../hadoop-0.20.205.0_hadoop-env.sh.patch
>
> Hit enter to continue.
> patching file conf/hadoop-env.sh
> Hunk #1 FAILED at 9.
> 1 out of 1 hunk FAILED -- saving rejects to file conf/hadoop-env.sh.rej
>
> I have a fix that I will be committing shortly:
> https://reviews.apache.org/r/10668/
>
>
> On Fri, Apr 19, 2013 at 6:45 PM, Jim Donahue <[email protected]> wrote:
>
>> Yup, I tried it on both a 32-bit and 64-bit Amazon Linux instance and got
>> the same behavior.  It is possible that I inadvertently made some change
>> in my build scripts that caused it to fail, but the scripts are pretty
>> stable and I can't think of any change that I made to both (32 and 64-bit)
>> of them that would cause a problem.
>>
>> Jim
>>
>> On 4/19/13 6:39 PM, "Benjamin Mahler" <[email protected]> wrote:
>>
>> >Well that is unexpected. Is that consistently failing for you?
>> >
>> >
>> >On Fri, Apr 19, 2013 at 6:29 PM, Jim Donahue <[email protected]> wrote:
>> >
>> >> Also the example seems to have stopped working Š  That's not a serious
>> >> problem (I can just ignore it) but it did work last week as far as I
>> can
>> >> remember. :-)
>> >>
>> >> Waiting 5 seconds for it to start. . . . . .
>> >> Alright, now let's run the "wordcount" example via:
>> >>
>> >> $ ./bin/hadoop jar hadoop-examples-0.20.205.0.jar wordcount
>> >> src/contrib/mesos/src/java/org/apache/hadoop/mapred out
>> >>
>> >> Hit enter to continue.
>> >> 13/04/19 23:31:54 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 0 time(s).
>> >> 13/04/19 23:31:55 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 1 time(s).
>> >> 13/04/19 23:31:56 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 2 time(s).
>> >> 13/04/19 23:31:57 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 3 time(s).
>> >> 13/04/19 23:31:58 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 4 time(s).
>> >> 13/04/19 23:31:59 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 5 time(s).
>> >> 13/04/19 23:32:00 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 6 time(s).
>> >> 13/04/19 23:32:01 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 7 time(s).
>> >> 13/04/19 23:32:02 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 8 time(s).
>> >> 13/04/19 23:32:03 INFO ipc.Client: Retrying connect to server:
>> >>localhost/
>> >> 127.0.0.1:54311. Already tried 9 time(s).
>> >> java.net.ConnectException: Call to localhost/127.0.0.1:54311 failed on
>> >> connection exception: java.net.ConnectException: Connection refused
>> >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095)
>> >> at org.apache.hadoop.ipc.Client.call(Client.java:1071)
>> >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>> >> at org.apache.hadoop.mapred.$Proxy1.getProtocolVersion(Unknown Source)
>> >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
>> >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
>> >> at
>> org.apache.hadoop.mapred.JobClient.createRPCProxy(JobClient.java:478)
>> >> at org.apache.hadoop.mapred.JobClient.init(JobClient.java:472)
>> >> at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:455)
>> >> at org.apache.hadoop.mapreduce.Job$1.run(Job.java:478)
>> >> at java.security.AccessController.doPrivileged(Native Method)
>> >> at javax.security.auth.Subject.doAs(Subject.java:416)
>> >> at
>> >>
>>
>> >>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation
>> >>.java:1059)
>> >> at org.apache.hadoop.mapreduce.Job.connect(Job.java:476)
>> >> at org.apache.hadoop.mapreduce.Job.submit(Job.java:464)
>> >> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:495)
>> >> at org.apache.hadoop.examples.WordCount.main(WordCount.java:67)
>> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> >> at
>> >>
>>
>> >>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
>> >>:57)
>> >> at
>> >>
>>
>> >>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
>> >>mpl.java:43)
>> >> at java.lang.reflect.Method.invoke(Method.java:616)
>> >> at
>> >>
>>
>> >>org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDri
>> >>ver.java:68)
>> >> at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139)
>> >> at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:64)
>> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> >> at
>> >>
>>
>> >>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
>> >>:57)
>> >> at
>> >>
>>
>> >>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
>> >>mpl.java:43)
>> >> at java.lang.reflect.Method.invoke(Method.java:616)
>> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
>> >> Caused by: java.net.ConnectException: Connection refused
>> >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>> >> at
>> >>sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592)
>> >> at
>> >>
>>
>> >>org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.jav
>> >>a:206)
>> >> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:604)
>> >> at
>> >>org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434)
>> >> at
>> >>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
>> >> at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
>> >> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202)
>> >> at org.apache.hadoop.ipc.Client.call(Client.java:1046)
>> >> ... 27 more
>> >>
>> >> Oh no, it failed! Try running the JobTracker and wordcount
>> >> example manually ... it might be an issue with your environment that
>> >> this tutorial didn't cover (if you find this to be the case, please
>> >> create a JIRA for us and/or send us a code review).
>> >>
>> >> ./TUTORIAL.sh: line 662: kill: (1522) - No such process
>> >> make: *** [hadoop-0.20.205.0] Error 1
>> >>
>> >>
>> >> On 4/19/13 4:29 PM, "Benjamin Mahler" <[email protected]
>> <mailto:
>> >> [email protected]>> wrote:
>> >>
>> >> Brenden: It looks like Maven isn't required when building
>> >> hadoop-0.20.205.0, can you send a patch to fix your change to only
>> check
>> >> for Maven when building the CDH releases?
>> >>
>> >> Jim: Thanks for the report.
>> >>
>> >> I committed a recent change by Brenden here, which enforces that both
>> >>'ant'
>> >> and 'mvn' are present when building the hadoop port:
>> >> https://reviews.apache.org/r/10558/
>> >>
>> >>
>> >> On Fri, Apr 19, 2013 at 3:51 PM, Jim Donahue <[email protected]
>> <mailto:
>> >> [email protected]>> wrote:
>> >>
>> >> I was -- the last build I did was ten days ago.  Somebody broke the
>> >>build
>> >> scripts that I've been using for quite a while.
>> >>
>> >> Jim
>> >>
>> >>
>> >>
>> >> On 4/19/13 3:48 PM, "Benjamin Mahler" <[email protected]
>> <mailto:
>> >> [email protected]>> wrote:
>> >>
>> >> >You can fix this by installing Maven.
>> >> >
>> >> >However, I was under the assumption that we required Maven in order to
>> >>run
>> >> >the Hadoop tutorial. You were successfully building hadoop without
>> >>Maven
>> >> >installed?
>> >> >
>> >> >
>> >> >On Fri, Apr 19, 2013 at 3:44 PM, Jim Donahue
>> >><[email protected]<mailto:
>> >> [email protected]>> wrote:
>> >> >
>> >> >> I'm trying to build Mesos on Amazon Linux and it appears that the
>> >>Hadoop
>> >> >> build script has changed.  It worked just fine a few days ago, but
>> >>now
>> >> >>I'm
>> >> >> getting:
>> >> >>
>> >> >> sudo make hadoop-0.20.205.0
>> >> >> if test ".." != ".."; then \
>> >> >> cp -p ./TUTORIAL.sh .; \
>> >> >> cp -p ./hadoop-gridmix.patch .; \
>> >> >> cp -p ./hadoop-7698-1.patch .; \
>> >> >> cp -p ./hadoop-0.20.205.0_hadoop-env.sh.patch .; \
>> >> >> cp -p ./hadoop-0.20.205.0_mesos.patch .; \
>> >> >> cp -p ./mapred-site.xml.patch .; \
>> >> >> cp -rp ./mesos .; \
>> >> >> cp -p ./mesos-executor .; \
>> >> >> fi
>> >> >> rm -rf hadoop-0.20.205.0
>> >> >> which: no mvn in (/sbin:/bin:/usr/sbin:/usr/bin)
>> >> >>
>> >> >> We seem to be missing mvn from the path. Please install
>> >> >> mvn and re-run this tutorial. If you still have troubles, please
>> >>report
>> >> >> this to:
>> >> >>
>> >> >> [email protected]<mailto:
>> [email protected]>
>> >> >>
>> >> >> (Remember to include as much debug information as possible.)
>> >> >>
>> >> >> Help, please!
>> >> >>
>> >> >>
>> >> >> Jim
>> >> >>
>> >>
>> >>
>> >>
>> >>
>>
>>
>

Reply via email to