sorry clicked send too soon, but basically changing that did not produce any result, still seeing the same message.So I guess my question is what is the property that is responsible for that?
Thanks Oleg On Fri, Oct 5, 2012 at 10:40 AM, Oleg Zhurakousky < [email protected]> wrote: > Yes I understand that and I guess I am trying to find that 'right > property' > I did find one reference to it in hdfs-defaul.xml > > <name>dfs.datanode.address</name> > > <value>0.0.0.0:50010</value> > > so i changed that in my hdfs-site.xml to > > <name>dfs.datanode.address</name> > > <value>192.168.15.20:50010</value> > > > But > > > On Fri, Oct 5, 2012 at 10:33 AM, Bertrand Dechoux <[email protected]>wrote: > >> Indeed, you are connecting to localhost and you said it was a remote >> connection so I guess there is nothing there which is relevant for you. >> The main idea is that you need to provide the configuration files. They >> are read by default from the classpath. Any place where you have a >> Configuration/JobConf you could also set up the right properties which >> would be the location of the HDFS master (and mapred if you want to do >> something about it). >> >> Regards >> >> Bertrand >> >> >> On Fri, Oct 5, 2012 at 4:15 PM, Oleg Zhurakousky < >> [email protected]> wrote: >> >>> So now I am passed it and able to RunAs 'hduser', but when I attempt to >>> read from FSDataInputStream i see this message in my console >>> >>> 10:12:10,065 WARN main hdfs.DFSClient:2106 - Failed to connect to / >>> 127.0.0.1:50010, add to deadNodes and continuejava.net.ConnectException: >>> Connection refused >>> >>> 10:12:10,072 INFO main hdfs.DFSClient:2272 - Could not obtain block >>> blk_-4047236896256451627_1003 from any node: java.io.IOException: No >>> live nodes contain current block. Will get new block locations from >>> namenode and retry... >>> >>> >>> I am obviously missing a configuration setting somewhere. . . any idea? >>> >>> Thanks >>> >>> Oleg >>> >>> On Fri, Oct 5, 2012 at 9:37 AM, Oleg Zhurakousky < >>> [email protected]> wrote: >>> >>>> After i clicked send I found the same link ;), but thank you anyway. >>>> >>>> Oleg >>>> >>>> >>>> On Fri, Oct 5, 2012 at 9:34 AM, Bertrand Dechoux <[email protected]>wrote: >>>> >>>>> Hi, >>>>> >>>>> You might be looking for something like : >>>>> UserGroupInformation.createRemoteUser(user).doAs( >>>>> >>>>> see >>>>> >>>>> http://hadoop.apache.org/docs/r1.0.3/api/org/apache/hadoop/security/UserGroupInformation.html >>>>> >>>>> It is a JAAS wrapper for Hadoop. >>>>> >>>>> Regards >>>>> >>>>> Bertrand >>>>> >>>>> >>>>> >>>>> >>>>> On Fri, Oct 5, 2012 at 3:19 PM, Oleg Zhurakousky < >>>>> [email protected]> wrote: >>>>> >>>>>> I am working on some samples where I want to write to HDFS running on >>>>>> another machine (different OS etc.) >>>>>> The identity of my client process is just whatever my OS says it is >>>>>> (e.g., 'oleg') hence: >>>>>> >>>>>> 08:56:49,240 DEBUG IPC Client (47) connection to >>>>>> /192.168.15.20:54310from oleg ipc.Client:803 - IPC Client (47) >>>>>> connection to / >>>>>> 192.168.15.20:54310 from oleg got value #2 >>>>>> >>>>>> But there is no 'oleg' where the hadoop is running. Instead there is >>>>>> 'hduser'. >>>>>> >>>>>> Is there a way or an equivalent of "RunAs" in hadoop? >>>>>> >>>>>> Thanks >>>>>> Oleg >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Bertrand Dechoux >>>>> >>>> >>>> >>> >> >> >> -- >> Bertrand Dechoux >> > >
