[ 
https://issues.apache.org/jira/browse/HADOOP-785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12517354
 ] 

Arun C Murthy commented on HADOOP-785:
--------------------------------------

bq. Currently, hadoop-default.xml is not supposed to be changed by users. Would 
you relax this convention in your proposal? There might be a few variables that 
I'd like to set for client and server at the same time (eg. namenode address).

Hmm... how about letting both *server* and *client* values for 
{{fs.default.name}}'s {{context}} tag to let people know to it can be specified 
in both hadoop-server.xml and hadoop-client.xml, and will be used 
appropriately? Would that help? I'd rather keep hadoop-default.xml sacrosanct, 
though we don't prevent you from editing it even today - thus it serves as a 
gold-standard for everyone.

bq. Why don't you want to split up namenode vs. jobtracker and datanode vs. 
tasktracker? 

I did think about this, and I really don't see what value a 
{HDFS|MR}ServerConfiguration and {HDFS|MR}ClientConfiguration will provide, 
which is why I didn't take this route... but I'm open to arguments. Just 
separation of physical files doesn't seem enough to warrant  4 classes rather 
than 2.

bq. This division could be done with xml comments - I don't think it needs to 
be so formal as to need a new field.

I agree, yet it's my take that it is better to institutionalise this by adding 
another tag, same with the {{context}} tag. Again this depends on whether or 
not we can reach a common ground...

> Divide the server and client configurations
> -------------------------------------------
>
>                 Key: HADOOP-785
>                 URL: https://issues.apache.org/jira/browse/HADOOP-785
>             Project: Hadoop
>          Issue Type: Improvement
>          Components: conf
>    Affects Versions: 0.9.0
>            Reporter: Owen O'Malley
>            Assignee: Arun C Murthy
>             Fix For: 0.15.0
>
>
> The configuration system is easy to misconfigure and I think we need to 
> strongly divide the server from client configs. 
> An example of the problem was a configuration where the task tracker has a 
> hadoop-site.xml that set mapred.reduce.tasks to 1. Therefore, the job tracker 
> had the right number of reduces, but the map task thought there was a single 
> reduce. This lead to a hard to find diagnose failure.
> Therefore, I propose separating out the configuration types as:
> class Configuration;
> // reads site-default.xml, hadoop-default.xml
> class ServerConf extends Configuration;
> // reads hadoop-server.xml, $super
> class DfsServerConf extends ServerConf;
> // reads dfs-server.xml, $super
> class MapRedServerConf extends ServerConf;
> // reads mapred-server.xml, $super
> class ClientConf extends Configuration;
> // reads hadoop-client.xml, $super
> class JobConf extends ClientConf;
> // reads job.xml, $super
> Note in particular, that nothing corresponds to hadoop-site.xml, which 
> overrides both client and server configs. Furthermore, the properties from 
> the *-default.xml files should never be saved into the job.xml.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to