[ https://issues.apache.org/jira/browse/HADOOP-785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12517354 ]
Arun C Murthy commented on HADOOP-785: -------------------------------------- bq. Currently, hadoop-default.xml is not supposed to be changed by users. Would you relax this convention in your proposal? There might be a few variables that I'd like to set for client and server at the same time (eg. namenode address). Hmm... how about letting both *server* and *client* values for {{fs.default.name}}'s {{context}} tag to let people know to it can be specified in both hadoop-server.xml and hadoop-client.xml, and will be used appropriately? Would that help? I'd rather keep hadoop-default.xml sacrosanct, though we don't prevent you from editing it even today - thus it serves as a gold-standard for everyone. bq. Why don't you want to split up namenode vs. jobtracker and datanode vs. tasktracker? I did think about this, and I really don't see what value a {HDFS|MR}ServerConfiguration and {HDFS|MR}ClientConfiguration will provide, which is why I didn't take this route... but I'm open to arguments. Just separation of physical files doesn't seem enough to warrant 4 classes rather than 2. bq. This division could be done with xml comments - I don't think it needs to be so formal as to need a new field. I agree, yet it's my take that it is better to institutionalise this by adding another tag, same with the {{context}} tag. Again this depends on whether or not we can reach a common ground... > Divide the server and client configurations > ------------------------------------------- > > Key: HADOOP-785 > URL: https://issues.apache.org/jira/browse/HADOOP-785 > Project: Hadoop > Issue Type: Improvement > Components: conf > Affects Versions: 0.9.0 > Reporter: Owen O'Malley > Assignee: Arun C Murthy > Fix For: 0.15.0 > > > The configuration system is easy to misconfigure and I think we need to > strongly divide the server from client configs. > An example of the problem was a configuration where the task tracker has a > hadoop-site.xml that set mapred.reduce.tasks to 1. Therefore, the job tracker > had the right number of reduces, but the map task thought there was a single > reduce. This lead to a hard to find diagnose failure. > Therefore, I propose separating out the configuration types as: > class Configuration; > // reads site-default.xml, hadoop-default.xml > class ServerConf extends Configuration; > // reads hadoop-server.xml, $super > class DfsServerConf extends ServerConf; > // reads dfs-server.xml, $super > class MapRedServerConf extends ServerConf; > // reads mapred-server.xml, $super > class ClientConf extends Configuration; > // reads hadoop-client.xml, $super > class JobConf extends ClientConf; > // reads job.xml, $super > Note in particular, that nothing corresponds to hadoop-site.xml, which > overrides both client and server configs. Furthermore, the properties from > the *-default.xml files should never be saved into the job.xml. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.