Try using fully-qualified domain names in your xml file.

On 5/23/07 11:22 AM, "yu-yang chen" <[EMAIL PROTECTED]> wrote:

> I know this question pop up so mant times, but I couldn't find a right
> answer to my case....
> 
> hello everyone, i just started to use hadoop and encountered problem on
> setting up the single node system:
> 
> 
> This is what inside my hadoop-site.xml:
> 
> 
> ------------------------------------------------------------------------------
> ------
> camelot01% cat hadoop-site.xml
> 
> 
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> 
> <!-- Put site-specific property overrides in this file. -->
> 
> <configuration>
>         <property>
>                 <name>fs.default.name</name>
>                 <value>localhost:9000</value>
>         </property>
>         <property>
>                 <name>mapred.job.tracker</name>
>                 <value>localhost:9001</value>
>         </property>
>         <property>
>                 <name>dfs.replication</name>
>                 <value>1</value>
>         </property>
> </configuration>
> 
> ------------------------------------------------------------------------------
> -------
> 
> After:
> 
> bin/hadoop namenode -format
> Re-format filesystem in /tmp/hadoop-yyc04/dfs/name ? (Y or N) Y
> Formatted /tmp/hadoop-yyc04/dfs/name
> 
> camelot01% bin/start-all.sh<---works fine, passwrdless ssh also works
> 
> 
> And this is the error i obtain:
> 
> 
> camelot01% bin/hadoop jar hadoop-0.12.3-examples.jar pi 10 20
> 
> Number of Maps = 10 Samples per Map = 20
> 07/05/23 19:14:11 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 1 time(s).
> 07/05/23 19:14:12 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 2 time(s).
> 07/05/23 19:14:13 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 3 time(s).
> 07/05/23 19:14:14 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 4 time(s).
> 07/05/23 19:14:15 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 5 time(s).
> 07/05/23 19:14:16 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 6 time(s).
> 07/05/23 19:14:17 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 7 time(s).
> 07/05/23 19:14:18 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 8 time(s).
> 07/05/23 19:14:19 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 9 time(s).
> 07/05/23 19:14:20 INFO ipc.Client: Retrying connect to server:
> localhost/146.169.2.131:9000. Already tried 10 time(s).
> java.net.ConnectException: Connection refused
>         at java.net.PlainSocketImpl.socketConnect(Native Method)
>         at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
>         at 
> java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
>         at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
>         at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:364)
>         at java.net.Socket.connect(Socket.java:507)
>         at 
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:149)
>         at org.apache.hadoop.ipc.Client.getConnection(Client.java:529)
>         at org.apache.hadoop.ipc.Client.call(Client.java:458)
>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:163)
>         at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown Source)
>         at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:247)
>         at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:105)
>         at 
> org.apache.hadoop.dfs.DistributedFileSystem$RawDistributedFileSystem.initializ
> e(DistributedFileSystem.java:67)
>         at 
> org.apache.hadoop.fs.FilterFileSystem.initialize(FilterFileSystem.java:57)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:160)
>         at org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:119)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:91)
>         at 
> org.apache.hadoop.examples.PiEstimator.launch(PiEstimator.java:169)
>         at org.apache.hadoop.examples.PiEstimator.main(PiEstimator.java:226)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.j
> ava:25)
>         at java.lang.reflect.Method.invoke(Method.java:585)
>         at 
> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.j
> ava:71)
>         at 
> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:143)
>         at 
> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:40)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.j
> ava:25)
>         at java.lang.reflect.Method.invoke(Method.java:585)
>         at org.apache.hadoop.util.RunJar.main(RunJar.java:155)
> 
> 
> 
> 
> 
> can anyone please tell me why? I did not change any other files apart
> from hadoop-env.xml (set JAVA_HOME) and hadoop-site.xml
> 
> 
> 
> please help....thank you so much!
> 
> yu-yang

Reply via email to