[ 
https://issues.apache.org/jira/browse/MAHOUT-482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joe Prasanna Kumar updated MAHOUT-482:
--------------------------------------

    Status: Patch Available  (was: Open)

Patch available as specified in the issue description

> Defaulting $HADOOP_CONF_DIR to $HADOOP_HOME/conf
> ------------------------------------------------
>
>                 Key: MAHOUT-482
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-482
>             Project: Mahout
>          Issue Type: Bug
>            Reporter: Joe Prasanna Kumar
>            Priority: Trivial
>         Attachments: MAHOUT-482.patch
>
>   Original Estimate: 0.33h
>  Remaining Estimate: 0.33h
>
> Hi all,
> while trying to understand the clusterdump utility, I encountered a message 
> "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally". As per some of the 
> discussions in Apr'10 
> (http://lucene.472066.n3.nabble.com/Dealing-with-kmean-and-meanshift-output-td708824.html#a709066),
>  it would be good to default the hadoop conf directory to hadoop_home/conf. 
> So I've changed the script to accommodate this.
> Existing code:
> if [ "$HADOOP_CONF_DIR" = "" ] || [ "$HADOOP_HOME" = "" ] || [ 
> "$MAHOUT_LOCAL" != "" ] ; then
>   if [ "$HADOOP_CONF_DIR" = "" ] || [ "$HADOOP_HOME" = "" ] ; then
>     echo "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally"
>   elif [ "$MAHOUT_LOCAL" != "" ] ; then 
>     echo "MAHOUT_LOCAL is set, running locally"
>   fi
>   exec "$JAVA" $JAVA_HEAP_MAX $MAHOUT_OPTS -classpath "$CLASSPATH" $CLASS "$@"
> else
>   echo "running on hadoop, using HADOOP_HOME=$HADOOP_HOME and 
> HADOOP_CONF_DIR=$HADOOP_CONF_DIR"
>  
> New code:
> if [ "$HADOOP_HOME" = "" ] || [ "$MAHOUT_LOCAL" != "" ] ; then
>   if [ "$HADOOP_HOME" = "" ] ; then
>     echo "no HADOOP_HOME set, running locally"
>   elif [ "$MAHOUT_LOCAL" != "" ] ; then 
>     echo "MAHOUT_LOCAL is set, running locally"
>   fi
>   exec "$JAVA" $JAVA_HEAP_MAX $MAHOUT_OPTS -classpath "$CLASSPATH" $CLASS "$@"
> else
>   if [ "$HADOOP_CONF_DIR" = "" ] ; then
>     HADOOP_CONF_DIR=$HADOOP_HOME/conf
>     echo "no HADOOP_CONF_DIR set. so defaulting HADOOP_CONF_DIR to 
> $HADOOP_HOME/conf "
>   fi
> echo "running on hadoop, using HADOOP_HOME=$HADOOP_HOME and 
> HADOOP_CONF_DIR=$HADOOP_CONF_DIR"
> In my local machine, I've just set HADOOP_HOME and the script defaults the 
> $HADOOP_CONF_DIR directory and runs on hadoop.
> If this issue / fix looks valid, I'll open a JIRA issue and attach the patch.
> Appreciate your thoughts / feedbacks,
> Joe.
> Drew Farris to dev
> show details 8:53 AM (12 hours ago)
> On Tue, Aug 17, 2010 at 12:26 AM, Joe Kumar <[email protected]> wrote:
> > Hi all,
> >
> > while trying to understand the clusterdump utility, I encountered a message
> > "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally". As per some of the
> > discussions in Apr'10 (
> > http://lucene.472066.n3.nabble.com/Dealing-with-kmean-and-meanshift-output-td708824.html#a709066),
> > it would be good to default the hadoop conf directory to hadoop_home/conf.
> Agreed, and your solution looks good. I'll keep an eye out for the patch.
> Jake Mannix to dev
> show details 3:46 PM (5 hours ago)
> +1
> Good idea.  I thought we already did this...
>  -jake

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to