Hbase SQL

2016-06-16 Thread Krishnaprasad A S
Hi, I have a doubt on Hbase storage in Drill. My HBase table 'X' has a column family 'CF' and a column name which is an integer. we store the data in hbase with column name as bytes array, using the following code to create the byte[] (as normal in Hbase). byte[] column =

Re: Question about converting to parquet data

2016-06-16 Thread Ted Dunning
Can you process this file when it is stored on a local file system? On Thu, Jun 16, 2016 at 11:29 AM, Tanmay Solanki < tsolank...@yahoo.in.invalid> wrote: > Hello, > I am currently running Apache Drill on a 20 node cluster and was running > into some errors that I was wondering if you would

Re: Dynamic UDFs support

2016-06-16 Thread Ted Dunning
Yes. A UDF could be built that constructs row objects and sends them to an external function. There are some issues since you need to somehow know what columns and types are being passed. A more promising approach I think is dependent on Arrow succeeding. At that point, you can define a batch UDF

Re: Dynamic UDFs support

2016-06-16 Thread Jacques Nadeau
Two quick thoughts: - (user) In the design document I didn't see any discussion of ownership/conflicts or unloading. Would be helpful to see the thinking there - (dev) There is a row oriented facade via the FieldReader/FieldWriter/ComplexWriter classes. That would be a good place to start when

Creating Parquet Files from tsv .gz Data_Read Error

2016-06-16 Thread Tanmay Solanki
Hello, I am currently running Apache Drill on a 20 node cluster and was running into some errors that I was wondering if you would be able to help me with this. I am attempting to run the following query to create a parquet table in a new S3 bucket from another table that is in a tsv format:

Question about querying parquet data

2016-06-16 Thread Tanmay Solanki
Hello, I am currently running Apache Drill on a 20 node cluster and was running into some errors that I was wondering if you would be able to help me with this. I am attempting to run the following query to create a parquet table in a new S3 bucket from another table that is in a tsv format:

Re: DRILL-4199: Add Support for HBase 1.X - planning to merge

2016-06-16 Thread Aman Sinha
Qiang/Aditya can you create a JIRA for this and mark it for 1.7. thanks. On Thu, Jun 16, 2016 at 11:25 AM, Aditya wrote: > Thanks for reporting, I'm looking into it and will post a patch soon. > > On Wed, Jun 15, 2016 at 7:27 PM, qiang li wrote:

Question about converting to parquet data

2016-06-16 Thread Tanmay Solanki
Hello, I am currently running Apache Drill on a 20 node cluster and was running into some errors that I was wondering if you would be able to help me with this. I am attempting to run the following query to create a parquet table in a new S3 bucket from another table that is in a tsv format:

Re: Dynamic UDFs support

2016-06-16 Thread John Omernik
Honestly, I don't see it as a priority issue. I think some of the ideas around community java UDFs could be a better approach. I'd hate to take away from other work to hack in something like this. On Thu, Jun 16, 2016 at 1:19 PM, Paul Rogers wrote: > Ted refers to source

Re: apache-drill-1.6.0: failed to connect to hdfs 1.0.4

2016-06-16 Thread Jinfeng Ni
The error seems to be thrown by hdfs, when Drill tries to get list of file status. This might be caused by incompatibility of hdfs version. Per Drill's doc [1] , HDFS 2.3 + API is required. Did you say you are running on hdfs 1.0.4? " Hadoop: All Hadoop distributions (HDFS API 2.3+), "

Re: Dynamic UDFs support

2016-06-16 Thread Paul Rogers
Ted refers to source code transformation. Drill gains its speed from value vectors. However, VVs are a far cry from the row-based interface that most mere mortals are accustomed to using. Since VVs are very type specific, code is typically generated to handle the specifics of each type. Accessing

apache-drill-1.6.0: failed to connect to hdfs 1.0.4

2016-06-16 Thread Tao,Mingyuan
This is my hdfs namenode: NameNode '10.207.78.21:38234' Started: Mon Feb 02 19:16:43 CST 2015 Version: 1.0.4, r1393290 This is the config of my file system plugin `rpmp`: { "type": "file", "enabled": true, "connection": "hdfs://10.207.78.21:38234", "config": null, "workspaces": {

Re: Drill on Phoenix

2016-06-16 Thread Neeraja Rentachintala
Alex can yo briefly describe your use case for using Drill with Phoenix. On Thu, Jun 16, 2016 at 10:42 AM, James Taylor wrote: > Yes, we've created a new Phoenix storage plugin for Drill here[1], and > there's a good presentation put together by Jacques on here[2] that

Re: Drill on Phoenix

2016-06-16 Thread James Taylor
Yes, we've created a new Phoenix storage plugin for Drill here[1], and there's a good presentation put together by Jacques on here[2] that covers Drillix (that's our initiative name) plus Drill and Arrow. This is definitely a work in progress at the POC level, but IMHO is very promising. We need a

Re: Dynamic UDFs support

2016-06-16 Thread Ted Dunning
Since UDF's use source code transformation, using Jython would be difficult. On Thu, Jun 16, 2016 at 9:42 AM, Arina Yelchiyeva < arina.yelchiy...@gmail.com> wrote: > Hi Charles, > > not that I am aware of. Proposed solution doesn't invent anything new, just > adds possibility to add UDFs

Fwd: Hbase query using Drill

2016-06-16 Thread Krishnaprasad A S
Hi, I have a doubt on Hbase storage in Drill. My HBase table 'X' has a column family 'CF' and a column name which is an integer. we store the data in hbase with column name as bytes array, using the following code to create the byte[] (as normal in Hbase). byte[] column =

Re: Dynamic UDFs support

2016-06-16 Thread Arina Yelchiyeva
Hi Charles, not that I am aware of. Proposed solution doesn't invent anything new, just adds possibility to add UDFs without drillbit restart. But contributions are welcomed. On Thu, Jun 16, 2016 at 4:52 PM Charles Givre wrote: > Arina, > Has there been any discussion about

Drill on Phoenix

2016-06-16 Thread Alex Kamil
Can Drill be integrated with Apache Phoenix? Thanks Alex

Re: Dynamic UDFs support

2016-06-16 Thread Charles Givre
Arina, Has there been any discussion about making it possible via Jython or something for users to write simple UDFs in Python? My ideal would be to have this capability integrated in the web GUI such that a user could write their UDF (in Python) right there, submit it and it would be