Regarding Sqoop, you can install it wherever you would have access to your database and HDFS cluster, you could e.g. install it on the namenode if you want it as long as it has access to the database that is the source or target of your data transfer.
On Thu, Aug 29, 2013 at 3:11 PM, Raj Hadoop <[email protected]> wrote: > Hi, > > I am trying to setup a multi node hadoop cluster. I am trying to > understand where hadoop clients like (Hive,Pig,Sqoop) would be installed in > the Hadoop Cluster. > > Say - I have three Linux machines- > > Node 1 - Master - (Name Node , Job Tracker and Secondary Name Node) > Node 2 - Slave (Task Tracker,Data Node) > Node 3 - Slave (Task Tracker,Data Node) > > On which machines should I install Hive? Should it be installed or Can it > be installed on a separate machine? What user and privileges are required ? > On which machines should I install Pig? Should it be installed or Can it > be installed on a separate machine? What user and privileges are required ? > On which machines should I install Sqoop? Should it be installed or Can it > be installed on a separate machine? What user and privileges are required ? > > Thanks, > Raj >
