hey mohammad but its already 127.0.0.1 i guess

On Fri, Jun 15, 2012 at 4:24 PM, Mohammad Tariq <[email protected]> wrote:
> All looks fine to me..change the line "127.0.1.1" in your hosts file
> to "127.0.0.1" and see if it works for you.
>
> Regards,
>     Mohammad Tariq
>
>
> On Fri, Jun 15, 2012 at 4:14 PM, soham sardar <[email protected]> 
> wrote:
>> configuration in the sense i have given the following configs
>>
>> hdfs-site
>>
>> <property>
>>  <name>dfs.replication</name>
>>  <value>1</value>
>>  <description>Default block replication.
>>  The actual number of replications can be specified when the file is created.
>>  The default is used if replication is not specified in create time.
>>  </description>
>> </property>
>>
>> core-site
>>
>> <property>
>>  <name>hadoop.tmp.dir</name>
>>  <value>/app/hadoop/tmp</value>
>>  <description>A base for other temporary directories.</description>
>> </property>
>>
>> <property>
>>  <name>fs.default.name</name>
>>  <value>hdfs://localhost:54310</value>
>>  <description>The name of the default file system.  A URI whose
>>  scheme and authority determine the FileSystem implementation.  The
>>  uri's scheme determines the config property (fs.SCHEME.impl) naming
>>  the FileSystem implementation class.  The uri's authority is used to
>>  determine the host, port, etc. for a filesystem.</description>
>> </property>
>>
>> and yarn-site
>>
>> <property>
>>    <name>yarn.resourcemanager.resource-tracker.address</name>
>>    <value>localhost:8031</value>
>>    <description>host is the hostname of the resource manager and
>>    port is the port on which the NodeManagers contact the Resource Manager.
>>    </description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.resourcemanager.scheduler.address</name>
>>    <value>localhost:8030</value>
>>    <description>host is the hostname of the resourcemanager and port
>> is the port
>>    on which the Applications in the cluster talk to the Resource Manager.
>>    </description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.resourcemanager.scheduler.class</name>
>>    
>> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
>>    <description>In case you do not want to use the default
>> scheduler</description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.resourcemanager.address</name>
>>    <value>localhost:8032</value>
>>    <description>the host is the hostname of the ResourceManager and
>> the port is the port on
>>    which the clients can talk to the Resource Manager. </description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.nodemanager.local-dirs</name>
>>    <value></value>
>>    <description>the local directories used by the nodemanager</description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.nodemanager.address</name>
>>    <value>127.0.0.1:8041</value>
>>    <description>the nodemanagers bind to this port</description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.nodemanager.resource.memory-mb</name>
>>    <value>10240</value>
>>    <description>the amount of memory on the NodeManager in GB</description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.nodemanager.remote-app-log-dir</name>
>>    <value>/app-logs</value>
>>    <description>directory on hdfs where the application logs are
>> moved to </description>
>>  </property>
>>
>>   <property>
>>    <name>yarn.nodemanager.log-dirs</name>
>>    <value></value>
>>    <description>the directories used by Nodemanagers as log
>> directories</description>
>>  </property>
>>
>>  <property>
>>    <name>yarn.nodemanager.aux-services</name>
>>    <value>mapreduce.shuffle</value>
>>    <description>shuffle service that needs to be set for Map Reduce
>> to run </description>
>>  </property>
>>
>> is there i need to make any other changes ????
>>
>>
>> On Fri, Jun 15, 2012 at 4:10 PM, Mohammad Tariq <[email protected]> wrote:
>>> Hi Soham,
>>>
>>>      Have you mentioned all the necessary properties in the
>>> configuration files??Also make sure your hosts file is ok.
>>>
>>> Regards,
>>>     Mohammad Tariq
>>>
>>>
>>> On Fri, Jun 15, 2012 at 3:53 PM, soham sardar <[email protected]> 
>>> wrote:
>>>> hey friends !!
>>>>
>>>> I have downloaded the cdh4 tarballs and kept in a folder and try to
>>>> run the hadoop nodes and other subsequent tools
>>>> I have also set each of the home paths in my bashrc
>>>>
>>>> now the problem is
>>>> when i try
>>>>
>>>>
>>>> hadoop fs -ls
>>>>
>>>>
>>>> 12/06/15 15:51:35 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 0 time(s).
>>>> 12/06/15 15:51:36 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 1 time(s).
>>>> 12/06/15 15:51:37 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 2 time(s).
>>>> 12/06/15 15:51:38 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 3 time(s).
>>>> 12/06/15 15:51:39 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 4 time(s).
>>>> 12/06/15 15:51:40 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 5 time(s).
>>>> 12/06/15 15:51:41 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 6 time(s).
>>>> 12/06/15 15:51:42 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 7 time(s).
>>>> 12/06/15 15:51:43 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 8 time(s).
>>>> 12/06/15 15:51:44 INFO ipc.Client: Retrying connect to server:
>>>> localhost/127.0.0.1:54310. Already tried 9 time(s).
>>>> ls: Call From XPS-L501X/127.0.1.1 to localhost:54310 failed on
>>>> connection exception: java.net.ConnectException: Connection refused;
>>>> For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
>>>>
>>>>
>>>> this is the error can someone help me as to why this error is occuring 
>>>> ????????

Reply via email to