hey when i tried that it says that command not found .. I wanna tell u that i have installed via tarball ( cdh4) so is there some changes i need to imply because of the tarball ??? I badly need start the nodes ....
On Fri, Jun 15, 2012 at 5:05 PM, Mohammad Tariq <[email protected]> wrote: > to start the hdfs use - > > $ for service in /etc/init.d/hadoop-hdfs-* >> do >> sudo $service start >> done > > and to start mapreduce do - > > $ for service in /etc/init.d/hadoop-0.20-mapreduce-* >> do >> sudo $service start >> done > > Regards, > Mohammad Tariq > > > On Fri, Jun 15, 2012 at 4:54 PM, soham sardar <[email protected]> > wrote: >> hey mohammad >> i wanna knw how to start all the nodes of hadoop like in cdh3 there >> was a script /bin/start-all.sh >> but in the cdh4 tarballs i dont find any such script?? >> >> On Fri, Jun 15, 2012 at 4:39 PM, Mohammad Tariq <[email protected]> wrote: >>> in both the lines???i mean your hosts file should look something like this - >>> >>> 127.0.0.1 localhost >>> 127.0.0.1 ubuntu.ubuntu-domain ubuntu >>> >>> # The following lines are desirable for IPv6 capable hosts >>> ::1 ip6-localhost ip6-loopback >>> fe00::0 ip6-localnet >>> ff00::0 ip6-mcastprefix >>> ff02::1 ip6-allnodes >>> ff02::2 ip6-allrouters >>> >>> Regards, >>> Mohammad Tariq >>> >>> >>> On Fri, Jun 15, 2012 at 4:32 PM, soham sardar <[email protected]> >>> wrote: >>>> hey mohammad but its already 127.0.0.1 i guess >>>> >>>> >>>> On Fri, Jun 15, 2012 at 4:24 PM, Mohammad Tariq <[email protected]> wrote: >>>>> All looks fine to me..change the line "127.0.1.1" in your hosts file >>>>> to "127.0.0.1" and see if it works for you. >>>>> >>>>> Regards, >>>>> Mohammad Tariq >>>>> >>>>> >>>>> On Fri, Jun 15, 2012 at 4:14 PM, soham sardar <[email protected]> >>>>> wrote: >>>>>> configuration in the sense i have given the following configs >>>>>> >>>>>> hdfs-site >>>>>> >>>>>> <property> >>>>>> <name>dfs.replication</name> >>>>>> <value>1</value> >>>>>> <description>Default block replication. >>>>>> The actual number of replications can be specified when the file is >>>>>> created. >>>>>> The default is used if replication is not specified in create time. >>>>>> </description> >>>>>> </property> >>>>>> >>>>>> core-site >>>>>> >>>>>> <property> >>>>>> <name>hadoop.tmp.dir</name> >>>>>> <value>/app/hadoop/tmp</value> >>>>>> <description>A base for other temporary directories.</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>fs.default.name</name> >>>>>> <value>hdfs://localhost:54310</value> >>>>>> <description>The name of the default file system. A URI whose >>>>>> scheme and authority determine the FileSystem implementation. The >>>>>> uri's scheme determines the config property (fs.SCHEME.impl) naming >>>>>> the FileSystem implementation class. The uri's authority is used to >>>>>> determine the host, port, etc. for a filesystem.</description> >>>>>> </property> >>>>>> >>>>>> and yarn-site >>>>>> >>>>>> <property> >>>>>> <name>yarn.resourcemanager.resource-tracker.address</name> >>>>>> <value>localhost:8031</value> >>>>>> <description>host is the hostname of the resource manager and >>>>>> port is the port on which the NodeManagers contact the Resource >>>>>> Manager. >>>>>> </description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.resourcemanager.scheduler.address</name> >>>>>> <value>localhost:8030</value> >>>>>> <description>host is the hostname of the resourcemanager and port >>>>>> is the port >>>>>> on which the Applications in the cluster talk to the Resource Manager. >>>>>> </description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.resourcemanager.scheduler.class</name> >>>>>> >>>>>> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value> >>>>>> <description>In case you do not want to use the default >>>>>> scheduler</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.resourcemanager.address</name> >>>>>> <value>localhost:8032</value> >>>>>> <description>the host is the hostname of the ResourceManager and >>>>>> the port is the port on >>>>>> which the clients can talk to the Resource Manager. </description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.local-dirs</name> >>>>>> <value></value> >>>>>> <description>the local directories used by the >>>>>> nodemanager</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.address</name> >>>>>> <value>127.0.0.1:8041</value> >>>>>> <description>the nodemanagers bind to this port</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.resource.memory-mb</name> >>>>>> <value>10240</value> >>>>>> <description>the amount of memory on the NodeManager in >>>>>> GB</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.remote-app-log-dir</name> >>>>>> <value>/app-logs</value> >>>>>> <description>directory on hdfs where the application logs are >>>>>> moved to </description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.log-dirs</name> >>>>>> <value></value> >>>>>> <description>the directories used by Nodemanagers as log >>>>>> directories</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>yarn.nodemanager.aux-services</name> >>>>>> <value>mapreduce.shuffle</value> >>>>>> <description>shuffle service that needs to be set for Map Reduce >>>>>> to run </description> >>>>>> </property> >>>>>> >>>>>> is there i need to make any other changes ???? >>>>>> >>>>>> >>>>>> On Fri, Jun 15, 2012 at 4:10 PM, Mohammad Tariq <[email protected]> >>>>>> wrote: >>>>>>> Hi Soham, >>>>>>> >>>>>>> Have you mentioned all the necessary properties in the >>>>>>> configuration files??Also make sure your hosts file is ok. >>>>>>> >>>>>>> Regards, >>>>>>> Mohammad Tariq >>>>>>> >>>>>>> >>>>>>> On Fri, Jun 15, 2012 at 3:53 PM, soham sardar >>>>>>> <[email protected]> wrote: >>>>>>>> hey friends !! >>>>>>>> >>>>>>>> I have downloaded the cdh4 tarballs and kept in a folder and try to >>>>>>>> run the hadoop nodes and other subsequent tools >>>>>>>> I have also set each of the home paths in my bashrc >>>>>>>> >>>>>>>> now the problem is >>>>>>>> when i try >>>>>>>> >>>>>>>> >>>>>>>> hadoop fs -ls >>>>>>>> >>>>>>>> >>>>>>>> 12/06/15 15:51:35 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 0 time(s). >>>>>>>> 12/06/15 15:51:36 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 1 time(s). >>>>>>>> 12/06/15 15:51:37 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 2 time(s). >>>>>>>> 12/06/15 15:51:38 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 3 time(s). >>>>>>>> 12/06/15 15:51:39 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 4 time(s). >>>>>>>> 12/06/15 15:51:40 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 5 time(s). >>>>>>>> 12/06/15 15:51:41 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 6 time(s). >>>>>>>> 12/06/15 15:51:42 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 7 time(s). >>>>>>>> 12/06/15 15:51:43 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 8 time(s). >>>>>>>> 12/06/15 15:51:44 INFO ipc.Client: Retrying connect to server: >>>>>>>> localhost/127.0.0.1:54310. Already tried 9 time(s). >>>>>>>> ls: Call From XPS-L501X/127.0.1.1 to localhost:54310 failed on >>>>>>>> connection exception: java.net.ConnectException: Connection refused; >>>>>>>> For more details see: http://wiki.apache.org/hadoop/ConnectionRefused >>>>>>>> >>>>>>>> >>>>>>>> this is the error can someone help me as to why this error is occuring >>>>>>>> ????????
