Try.. *hadoop fs -ls /*
** Thanks On Tue, Jul 23, 2013 at 10:27 PM, Ashish Umrani <[email protected]>wrote: > Thanks Jitendra, Bejoy and Yexi, > > I got past that. And now the ls command says it can not access the > directory. I am sure this is a permissions issue. I am just wondering > which directory and I missing permissions on. > > Any pointers? > > And once again, thanks a lot > > Regards > ashish > > *hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ > hadoop fs -ls* > *Warning: $HADOOP_HOME is deprecated.* > * > * > *ls: Cannot access .: No such file or directory.* > > > > On Tue, Jul 23, 2013 at 9:42 AM, Jitendra Yadav < > [email protected]> wrote: > >> Hi Ashish, >> >> Please check <property></property> in hdfs-site.xml. >> >> It is missing. >> >> Thanks. >> On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani >> <[email protected]>wrote: >> >>> Hey thanks for response. I have changed 4 files during installation >>> >>> core-site.xml >>> mapred-site.xml >>> hdfs-site.xml and >>> hadoop-env.sh >>> >>> >>> I could not find any issues except that all params in the hadoop-env.sh >>> are commented out. Only java_home is un commented. >>> >>> If you have a quick minute can you please browse through these files in >>> email and let me know where could be the issue. >>> >>> Regards >>> ashish >>> >>> >>> >>> I am listing those files below. >>> *core-site.xml * >>> <?xml version="1.0"?> >>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>> >>> <!-- Put site-specific property overrides in this file. --> >>> >>> <configuration> >>> <property> >>> <name>hadoop.tmp.dir</name> >>> <value>/app/hadoop/tmp</value> >>> <description>A base for other temporary directories.</description> >>> </property> >>> >>> <property> >>> <name>fs.default.name</name> >>> <value>hdfs://localhost:54310</value> >>> <description>The name of the default file system. A URI whose >>> scheme and authority determine the FileSystem implementation. The >>> uri's scheme determines the config property (fs.SCHEME.impl) naming >>> the FileSystem implementation class. The uri's authority is used to >>> determine the host, port, etc. for a filesystem.</description> >>> </property> >>> </configuration> >>> >>> >>> >>> *mapred-site.xml* >>> <?xml version="1.0"?> >>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>> >>> <!-- Put site-specific property overrides in this file. --> >>> >>> <configuration> >>> <property> >>> <name>mapred.job.tracker</name> >>> <value>localhost:54311</value> >>> <description>The host and port that the MapReduce job tracker runs >>> at. If "local", then jobs are run in-process as a single map >>> and reduce task. >>> </description> >>> </property> >>> </configuration> >>> >>> >>> >>> *hdfs-site.xml and* >>> <?xml version="1.0"?> >>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>> >>> <!-- Put site-specific property overrides in this file. --> >>> >>> <configuration> >>> <name>dfs.replication</name> >>> <value>1</value> >>> <description>Default block replication. >>> The actual number of replications can be specified when the file is >>> created. >>> The default is used if replication is not specified in create time. >>> </description> >>> </configuration> >>> >>> >>> >>> *hadoop-env.sh* >>> # Set Hadoop-specific environment variables here. >>> >>> # The only required environment variable is JAVA_HOME. All others are >>> # optional. When running a distributed configuration it is best to >>> # set JAVA_HOME in this file, so that it is correctly defined on >>> # remote nodes. >>> >>> # The java implementation to use. Required. >>> export JAVA_HOME=/usr/lib/jvm/jdk1.7.0_25 >>> >>> # Extra Java CLASSPATH elements. Optional. >>> # export HADOOP_CLASSPATH= >>> >>> >>> All pther params in hadoop-env.sh are commented >>> >>> >>> >>> >>> >>> >>> >>> >>> On Tue, Jul 23, 2013 at 8:38 AM, Jitendra Yadav < >>> [email protected]> wrote: >>> >>>> Hi, >>>> >>>> You might have missed some configuration (XML tags ), Please check all >>>> the Conf files. >>>> >>>> Thanks >>>> On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani <[email protected] >>>> > wrote: >>>> >>>>> Hi There, >>>>> >>>>> First of all, sorry if I am asking some stupid question. Myself being >>>>> new to the Hadoop environment , am finding it a bit difficult to figure >>>>> out >>>>> why its failing >>>>> >>>>> I have installed hadoop 1.2, based on instructions given in the >>>>> folllowing link >>>>> >>>>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >>>>> >>>>> All went well and I could do the start-all.sh and the jps command does >>>>> show all 5 process to be present. >>>>> >>>>> However when I try to do >>>>> >>>>> hadoop fs -ls >>>>> >>>>> I get the following error >>>>> >>>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>>> hadoop fs -ls >>>>> Warning: $HADOOP_HOME is deprecated. >>>>> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>>> <property> >>>>> ls: Cannot access .: No such file or directory. >>>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>>> >>>>> >>>>> >>>>> Can someone help me figure out whats the issue in my installation >>>>> >>>>> >>>>> Regards >>>>> ashish >>>>> >>>> >>>> >>> >> >
