Harsh,

Thanks for sharing. Recently, I ran into multiple user issue on cluster
because all the time we were using same user to run MRs so I found below
work around for it.

I am curious to know, How do we separate HDFS processes for users?



On Fri, Jul 1, 2011 at 1:39 PM, Harsh J <ha...@cloudera.com> wrote:

> Shahnawaz,
>
> If required, the MR and the HDFS processes can be user separated too.
> Its generally a good thing to do in practice (so that MR daemons and
> user don't get "superuser" access to HDFS files).
>
> FWIW, I usually flip open the cluster setup guide on ccp.cloudera.com
> and everything comes up right when I follow that (its got footnotes
> too, yay!).
>
> But yes, managing local permissions is quite a pain to do manually,
> and I believe the new bigtop incubator entrant is gonna help in
> setting up clusters painlessly as we go forward.
>
> On Fri, Jul 1, 2011 at 12:35 PM, Shahnawaz Saifi <shahsa...@gmail.com>
> wrote:
> > You can create multiple users on Cluster because it follows LSM (Linux
> > security model). But Hadoop related process NN, DN, JT, TT need to be
> > started with one user e.g. hadoop.
> >
> > Thanks,
> > Shah
> >
> > On Fri, Jul 1, 2011 at 4:57 AM, Mitra Kaseebhotla <
> > mitra.kaseebho...@gmail.com> wrote:
> >
> >> Not to divert the question here. I would like to know how we should
> manage
> >> the cluster users (I am a newbie)? insead of having one user(hadoop)
> creds
> >> shared?
> >>
> >> Thanks
> >> Mitra.
> >>
> >> On Thu, Jun 30, 2011 at 4:23 PM, Paul Rimba <paul.ri...@gmail.com>
> wrote:
> >>
> >> > sudo chown -R hadoop:hadoop /usr/local/hadoop.
> >> > That will give the directory ownership over to your hadoop account.
> >> >
> >> > On Fri, Jul 1, 2011 at 5:07 AM, Dhruv Kumar <dku...@ecs.umass.edu>
> >> wrote:
> >> >
> >> > > It is a permission issue. Are you sure that the account "hadoop" has
> >> read
> >> > > and write access to /usr/local/* directories?
> >> > >
> >> > > The installation of Hadoop has always been effortless for me. Just
> >> follow
> >> > > the instructions step by step given on:
> >> > > http://hadoop.apache.org/common/docs/stable/single_node_setup.html
> >> > >
> >> > >
> >> > >
> >> > >
> >> > > On Thu, Jun 30, 2011 at 11:08 AM, Ashish Tamrakar
> >> > > <ashishtamra...@gmail.com>wrote:
> >> > >
> >> > > > I am having a problem starting my hadoop. i setup my for my
> >> single-node
> >> > > > cluster. please help me to solve this out!!!!
> >> > > >
> >> > > > hadoop@ashishpc:~$ /usr/local/hadoop/bin/start-all.sh
> >> > > > starting namenode, logging to
> >> > > > /usr/local/hadoop/bin/../logs/hadoop-hadoop-namenode-ashishpc.out
> >> > > > /usr/local/hadoop/bin/hadoop-daemon.sh: line 117:
> >> > > > /usr/local/hadoop/bin/../logs/hadoop-hadoop-namenode-ashishpc.out:
> >> > > > Permission denied
> >> > > > head: cannot open
> >> > > >
> `/usr/local/hadoop/bin/../logs/hadoop-hadoop-namenode-ashishpc.out'
> >> for
> >> > > > reading: No such file or directory
> >> > > > localhost: starting datanode, logging to
> >> > > > /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-ashishpc.out
> >> > > > localhost: /usr/local/hadoop/bin/hadoop-daemon.sh: line 117:
> >> > > > /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-ashishpc.out:
> >> > > > Permission denied
> >> > > > localhost: head: cannot open
> >> > > >
> `/usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-ashishpc.out'
> >> for
> >> > > > reading: No such file or directory
> >> > > > localhost: starting secondarynamenode, logging to
> >> > > >
> >> > >
> >> >
> >>
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-secondarynamenode-ashishpc.out
> >> > > > localhost: /usr/local/hadoop/bin/hadoop-daemon.sh: line 117:
> >> > > >
> >> > >
> >> >
> >>
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-secondarynamenode-ashishpc.out:
> >> > > > Permission denied
> >> > > > localhost: head: cannot open
> >> > > >
> >> > > >
> >> > >
> >> >
> >>
> `/usr/local/hadoop/bin/../logs/hadoop-hadoop-secondarynamenode-ashishpc.out'
> >> > > > for reading: No such file or directory
> >> > > > starting jobtracker, logging to
> >> > > >
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-jobtracker-ashishpc.out
> >> > > > /usr/local/hadoop/bin/hadoop-daemon.sh: line 117:
> >> > > >
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-jobtracker-ashishpc.out:
> >> > > > Permission denied
> >> > > > head: cannot open
> >> > > >
> `/usr/local/hadoop/bin/../logs/hadoop-hadoop-jobtracker-ashishpc.out'
> >> > for
> >> > > > reading: No such file or directory
> >> > > > localhost: starting tasktracker, logging to
> >> > > >
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-ashishpc.out
> >> > > > localhost: /usr/local/hadoop/bin/hadoop-daemon.sh: line 117:
> >> > > >
> /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-ashishpc.out:
> >> > > > Permission denied
> >> > > > localhost: head: cannot open
> >> > > >
> >> `/usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-ashishpc.out'
> >> > > for
> >> > > > reading: No such file or directory
> >> > > >
> >> > >
> >> >
> >>
> >>
> >>
> >> --
> >>  ________________________________________
> >>
> >>  To Understand Recursion,
> >>  You must first Understand Recursion
> >>  ________________________________________
> >>
> >
>
>
>
> --
> Harsh J
>



-- 
Thanks,
Shah

Reply via email to