Hi,
I have a doubt on Hbase storage in Drill.
My HBase table 'X' has a column family 'CF' and a column name which is an
integer.
we store the data in hbase with column name as bytes array, using the
following code to create the byte[] (as normal in Hbase).
byte[] column =
Can you process this file when it is stored on a local file system?
On Thu, Jun 16, 2016 at 11:29 AM, Tanmay Solanki <
tsolank...@yahoo.in.invalid> wrote:
> Hello,
> I am currently running Apache Drill on a 20 node cluster and was running
> into some errors that I was wondering if you would
Yes. A UDF could be built that constructs row objects and sends them to an
external function. There are some issues since you need to somehow know
what columns and types are being passed.
A more promising approach I think is dependent on Arrow succeeding. At that
point, you can define a batch UDF
Two quick thoughts:
- (user) In the design document I didn't see any discussion of
ownership/conflicts or unloading. Would be helpful to see the thinking there
- (dev) There is a row oriented facade via the
FieldReader/FieldWriter/ComplexWriter classes. That would be a good place
to start when
Hello,
I am currently running Apache Drill on a 20 node cluster and was running into
some errors that I was wondering if you would be able to help me with this.
I am attempting to run the following query to create a parquet table in a new
S3 bucket from another table that is in a tsv format:
Hello,
I am currently running Apache Drill on a 20 node cluster and was running into
some errors that I was wondering if you would be able to help me with this.
I am attempting to run the following query to create a parquet table in a new
S3 bucket from another table that is in a tsv format:
Qiang/Aditya can you create a JIRA for this and mark it for 1.7. thanks.
On Thu, Jun 16, 2016 at 11:25 AM, Aditya wrote:
> Thanks for reporting, I'm looking into it and will post a patch soon.
>
> On Wed, Jun 15, 2016 at 7:27 PM, qiang li wrote:
Hello,
I am currently running Apache Drill on a 20 node cluster and was running into
some errors that I was wondering if you would be able to help me with this.
I am attempting to run the following query to create a parquet table in a new
S3 bucket from another table that is in a tsv format:
Honestly, I don't see it as a priority issue. I think some of the ideas
around community java UDFs could be a better approach. I'd hate to take
away from other work to hack in something like this.
On Thu, Jun 16, 2016 at 1:19 PM, Paul Rogers wrote:
> Ted refers to source
The error seems to be thrown by hdfs, when Drill tries to get list of
file status. This might be caused by incompatibility of hdfs version.
Per Drill's doc [1] , HDFS 2.3 + API is required. Did you say you are
running on hdfs 1.0.4?
"
Hadoop: All Hadoop distributions (HDFS API 2.3+),
"
Ted refers to source code transformation. Drill gains its speed from value
vectors. However, VVs are a far cry from the row-based interface that most
mere mortals are accustomed to using. Since VVs are very type specific,
code is typically generated to handle the specifics of each type. Accessing
This is my hdfs namenode:
NameNode '10.207.78.21:38234'
Started: Mon Feb 02 19:16:43 CST 2015
Version: 1.0.4, r1393290
This is the config of my file system plugin `rpmp`:
{
"type": "file",
"enabled": true,
"connection": "hdfs://10.207.78.21:38234",
"config": null,
"workspaces": {
Alex
can yo briefly describe your use case for using Drill with Phoenix.
On Thu, Jun 16, 2016 at 10:42 AM, James Taylor
wrote:
> Yes, we've created a new Phoenix storage plugin for Drill here[1], and
> there's a good presentation put together by Jacques on here[2] that
Yes, we've created a new Phoenix storage plugin for Drill here[1], and
there's a good presentation put together by Jacques on here[2] that covers
Drillix (that's our initiative name) plus Drill and Arrow. This is
definitely a work in progress at the POC level, but IMHO is very promising.
We need a
Since UDF's use source code transformation, using Jython would be difficult.
On Thu, Jun 16, 2016 at 9:42 AM, Arina Yelchiyeva <
arina.yelchiy...@gmail.com> wrote:
> Hi Charles,
>
> not that I am aware of. Proposed solution doesn't invent anything new, just
> adds possibility to add UDFs
Hi,
I have a doubt on Hbase storage in Drill.
My HBase table 'X' has a column family 'CF' and a column name which is an
integer.
we store the data in hbase with column name as bytes array, using the
following code to create the byte[] (as normal in Hbase).
byte[] column =
Hi Charles,
not that I am aware of. Proposed solution doesn't invent anything new, just
adds possibility to add UDFs without drillbit restart. But contributions
are welcomed.
On Thu, Jun 16, 2016 at 4:52 PM Charles Givre wrote:
> Arina,
> Has there been any discussion about
Can Drill be integrated with Apache Phoenix?
Thanks
Alex
Arina,
Has there been any discussion about making it possible via Jython or something
for users to write simple UDFs in Python?
My ideal would be to have this capability integrated in the web GUI such that a
user could write their UDF (in Python) right there, submit it and it would be
19 matches
Mail list logo