Good morning everyone,

I have a question about correct setup for hadoop. I have 14 Dell computers in a lab. Each connected to the internet and each independent of each other. All run CentOS. Logins are handled by NIS. If userA logs into the master and starts the daemons and UserB logs into the master and wants to run a job while the daemons from UserA are still running the following error occurs:

copyFromLocal: org.apache.hadoop.security.AccessControlException: Permission denied: user=UserB, access=WRITE, inode="user":UserA:supergroup:rwxr-xr-x

what needs to be changed to allow UserB-UserZ to run their jobs? Does there need to be a local user the everyone logs into as and run from there? Should Hadoop be ran in an actual cluster instead of independent computers? Any ideas what is the correct configuration settings that allow it?

I followed Ravi Phulari suggestions and followed:

http://hadoop.apache.org/core/docs/current/quickstart.html
http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Multi-Node_Cluster) <http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Multi-Node_Cluster%29> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster) <http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29>

These allowed me to get Hadoop running on the 14 computers when I login and everything works fine, thank you Ravi. The problem occurs when additional people attempt to run jobs simultaneously.

Thank you,

Brian

Reply via email to