Hi Vijay, It sounds like your user is seeing a different configuration than the hadoop user, such that fs.default.name isn't getting set. Did you set a HADOOP_CONF_DIR environment variable in the bashrc of the user that runs the hadoop process? If so, you should instead set that variable in hadoop-env.sh
-Todd On Fri, Jul 17, 2009 at 11:09 AM, Vijay Kumar Adhikari <[email protected]>wrote: > I have an installation of hadoop where when I log into the account > that runs the hadoop process, everything looks fine. I can copy local > files to the dfs, I can view the files inside dfs etc etc... > > When I log into a different user, I can still run all the dfs commands > but dfs shows files from my current local directory. If I am inside my > home directory in the local machine, and issue "hadoop dfs -ls" it > lists all the files inside my local home directory. If I change to > some other local directory and issue the same command all files from > that directory are shown. When I issue "hadoop dfs -copyFromLocal > <filename>", it says the uri already exists. > > What is wrong here? What do I need to fix? > > -- > Vijay >
