[ https://issues.apache.org/jira/browse/ACCUMULO-3943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14642974#comment-14642974 ]
Josh Elser commented on ACCUMULO-3943: -------------------------------------- First, I wonder if the master could do some unwrapping of the client-provided URI automatically. Second, my point was that knowing when somehost (w/o port) can be correctly treated as somehost:someport. We have {{fs.defaultFS}} in the Hadoop configuration to know how HDFS was configured but that's insufficient info. For example, HDFS configs could only have {{hdfs://somehost}} too. I'm not sure how this breaks down when dealing with multiple HDFS instances (and have multiple HDFS instances to encode). Ultimately, this is very open-ended and I can see it just sitting on JIRA indefinitely. How can we better scope the problem down. Automatically handle the default HDFS port of 8020 in volume URIs? > volumn definition agreement with default settings > ------------------------------------------------- > > Key: ACCUMULO-3943 > URL: https://issues.apache.org/jira/browse/ACCUMULO-3943 > Project: Accumulo > Issue Type: Bug > Components: gc, master, tserver > Reporter: Eric Newton > Priority: Minor > Fix For: 1.8.0 > > > I was helping a new user trying to use Accumulo. They managed to set up HDFS, > running on hdfs://localhost:8020. But they didn't set it up with specific > settings, and just used the default port. Accumulo worked initially, but > would not allow a bulk import. > During the bulk import process, the servers need to move the files into the > accumulo volumes, but keeping the volume the same. This makes the move > efficient, since nothing is copied between namespaces. In this case it > refused the import because it could not find the correct volume. > Accumulo needs to be more nuanced when comparing hdfs://localhost:8020, and > hdfs://localhost. -- This message was sent by Atlassian JIRA (v6.3.4#6332)