[ https://issues.apache.org/jira/browse/HADOOP-5670?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12833504#action_12833504 ]
Tomer Shiran commented on HADOOP-5670: -------------------------------------- Today, many users put the configuration files on an NFS server (e.g., NetApp) and all daemons and clients read the configuration from there. I can see two goals for this JIRA: # Remove the external dependency. That is, allow people to deploy Hadoop without an NFS server. # Remove the single point of failure. The NFS server might not have HA, in which case there is a single point of failure. Are there other issues with the current NFS-based architecture? What are we trying to solve here? > Hadoop configurations should be read from a distributed system > -------------------------------------------------------------- > > Key: HADOOP-5670 > URL: https://issues.apache.org/jira/browse/HADOOP-5670 > Project: Hadoop Common > Issue Type: New Feature > Components: conf > Reporter: Allen Wittenauer > > Rather than distributing the hadoop configuration files to every data node, > compute node, etc, Hadoop should be able to read configuration information > (dynamically!) from LDAP, ZooKeeper, whatever. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.