New installation with Hadoop 2.7 and Drill 1.0 on 4 nodes, it appears text files need to be on all nodes in a cluster?

Using the dfs config below, I am only able to query if a csv file is on all 4 nodes. If the file is only on the local node and not others, I get errors in the form of:

~~~
0: jdbc:drill:zk=es05:2181> select * from root.`customer_reviews_1998.csv`; Error: PARSE ERROR: From line 1, column 15 to line 1, column 18: Table 'root.customer_reviews_1998.csv' not found
~~~

~~~
{
  "type": "file",
  "enabled": true,
  "connection": "file:///",
  "workspaces": {
    "root": {
      "location": "/localdata/hadoop/stage",
      "writable": false,
      "defaultInputFormat": null
    },
~~~

On 25 May 2015, at 20:39, Kristine Hahn wrote:

The storage plugin "location" needs to be the full path to the localdata directory. This partial storage plugin definition works for the user named
mapr:

{
"type": "file",
"enabled": true,
"connection": "file:///",
"workspaces": {
 "root": {
   "location": "/home/mapr/localdata",
   "writable": false,
   "defaultInputFormat": null
 },
. . .

Here's a working query for the data in localdata:

0: jdbc:drill:> SELECT COLUMNS[0] AS Ngram,
. . . . . . . > COLUMNS[1] AS Publication_Date,
. . . . . . . > COLUMNS[2] AS Frequency
. . . . . . . > FROM dfs.root.`mydata.csv`
. . . . . . . > WHERE ((columns[0] = 'Zoological Journal of the Linnean')
. . . . . . . > AND (columns[2] > 250)) LIMIT 10;

An complete example, not yet published on the Drill site, shows in detail
the steps involved:
http://tshiran.github.io/drill/docs/querying-plain-text-files/#example-of-querying-a-tsv-file


Kristine Hahn
Sr. Technical Writer
415-497-8107 @krishahn


On Sun, May 24, 2015 at 1:56 PM, Matt <bsg...@gmail.com> wrote:

I have used a single node install (unzip and run) to query local text / csv files, but on a 3 node cluster (installed via MapR CE), a query with
local files results in:

~~~
sqlline version 1.1.6
0: jdbc:drill:> select * from dfs.`testdata.csv`;
Query failed: PARSE ERROR: From line 1, column 15 to line 1, column 17:
Table 'dfs./localdata/testdata.csv' not found

0: jdbc:drill:> select * from dfs.`/localdata/testdata.csv`;
Query failed: PARSE ERROR: From line 1, column 15 to line 1, column 17:
Table 'dfs./localdata/testdata.csv' not found
~~~

Is there a special config for local file querying? An initial doc search
did not point me to a solution, but I may simply not have found the
relevant sections.

I have tried modifying the default dfs config to no avail:

~~~
"type": "file",
"enabled": true,
"connection": "file:///",
"workspaces": {
 "root": {
   "location": "/localdata",
   "writable": false,
   "defaultInputFormat": null
 }
~~~

Reply via email to