[ 
https://issues.apache.org/jira/browse/HADOOP-10813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martin Bukatovic updated HADOOP-10813:
--------------------------------------
    Labels: hcfs  (was: )

> Define general filesystem exceptions (usable by any HCFS)
> ---------------------------------------------------------
>
>                 Key: HADOOP-10813
>                 URL: https://issues.apache.org/jira/browse/HADOOP-10813
>             Project: Hadoop Common
>          Issue Type: Improvement
>          Components: fs
>    Affects Versions: 2.2.0
>            Reporter: Martin Bukatovic
>            Priority: Minor
>              Labels: hcfs
>
> While Hadoop defines filesystem API which makes possible to use different
> filesystem implementation than HDFS (aka HCFS), we are missing HCFS
> exceptions for some failures wrt to namenode federation.
> For namenode federation, one can specify different namenode like this:
> {{hdfs://namenode_hostname/some/path}}. So when the given namenode doesn't
> exist, {{UnknownHostException}} is thrown:
> {noformat}
> $ hadoop fs -mkdir -p hdfs://bugcheck/foo/bar
> -mkdir: java.net.UnknownHostException: bugcheck
> Usage: hadoop fs [generic options] -mkdir [-p] <path> ...
> {noformat}
> Which is ok for HDFS, but there are other hadoop filesystem with different
> implementation and raising {{UnknownHostException}} doesn't make sense for
> them. For example the following path: {{glusterfs://bugcheck/foo/bar}} points
> to file {{/foo/bar}} on GlusterFS volume named {{bugcheck}}. That said, the
> meaning is the same compared to HDFS, both namenode hostname and glusterfs
> volume specifies different filesystem tree available for Hadoop.
> Would it make sense to define general HCFS exception which would wrap such
> cases so that it would be possible to fail in the same way when given
> filesystem tree is not available/defined, not matter which hadoop filesystem
> is used?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to