[
https://issues.apache.org/jira/browse/HADOOP-4631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12650504#action_12650504
]
Sharad Agarwal commented on HADOOP-4631:
----------------------------------------
bq. When you first call FileSystem.get("hdfs:///", conf), your configuration
won't yet have hdfs-specific default values, but, in the course of this call,
they will be loaded, before any hdfs code references them. I think this is
fine, since application code should not be directly referencing hdfs-specific
values.
For the Configuration objects created prior to loading of HDFS defaults, the
properties won't be reloaded. Currently, Configuration objects are created and
same being passed around. These objects if used to load hdfs specific values,
they won't be able to. no?
For example: DistributedFileSystem uses the Configuration object created prior
to loading of itself.
> Split the default configurations into 3 parts
> ---------------------------------------------
>
> Key: HADOOP-4631
> URL: https://issues.apache.org/jira/browse/HADOOP-4631
> Project: Hadoop Core
> Issue Type: Improvement
> Components: conf
> Reporter: Owen O'Malley
> Assignee: Sharad Agarwal
> Fix For: 0.20.0
>
>
> We need to split hadoop-default.xml into core-default.xml, hdfs-default.xml
> and mapreduce-default.xml. That will enable us to split the project into 3
> parts that have the defaults distributed with each component.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.