[ 
https://issues.apache.org/jira/browse/HADOOP-3835?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12617804#action_12617804
 ] 

dhruba borthakur commented on HADOOP-3835:
------------------------------------------

My thinking is that all (core and contrib) jars, docs, javadocs, bin, lib, etc 
get packaged into a single rpm. This gets installed in a default location, 
probably /var/opt/hadoop. This path can be overridden at rpm install time. This 
package does not have any configuration files.

Then, there will be a separate rpm package that contains the configuration 
(hadoop*.xml, metrics.properties, log4j.properties, and all file sin the conf 
directory). This will be installed by default at /var/opt/hadoop/conf. 

No NFS mounting is necessary. These two packages will most likely be installed 
on local directories on each cluster machine. The installation of a new package 
will not start/stop services. It is possible that an "install" might check to 
see if any hadoop processes are running, and of so, refuse to install. This 
will be for RehHat Linux. Once the scripts are made public, anyone can extend 
it to work on other Linux distributions.








> Develop scripts to create rpm package to facilitate deployment of hadoop on 
> Linux machines
> ------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-3835
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3835
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: build
>            Reporter: dhruba borthakur
>            Priority: Minor
>
> A rpm-like packing scheme to package and then install hadoop binaries is very 
> helpful, especially when the number of machines in the cluster is huge. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to