Jacques, not sure if you caught this, in the stacktrace it mentions broadcast sender. Did the plan for your test query include a broadcast join?
* (com.fasterxml.jackson.databind.JsonMappingException) Already had POJO for id (java.lang.Integer) [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8] (through reference chain: org.apache.drill.exec.physical.config.BroadcastSender["destinations"])* On Thu, Jan 14, 2016 at 2:02 PM, Jacques Nadeau <jacq...@dremio.com> wrote: > Hey Rohit, > > I'm having trouble reproducing this in my environment (pointing at derby + > hdfs instead of redshift/postgres). Can you turn on debug logging and then > run this query? You can enable the debug logging we are interested in by > adding the following item to logback.xml: > > <logger > name="org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler" > additivity="false"> > <level value="debug" /> > <appender-ref ref="FILE" /> > </logger> > > This is the query that I have succesfully completing. Please confirm it is > similar to your query: > SELECT count(*) > FROM dfs.`/data/tpch01/line/` a > RIGHT JOIN derby.DRILL_DERBY_TEST.PERSON b > ON a.cvalue = b.person_id > > > > -- > Jacques Nadeau > CTO and Co-Founder, Dremio > > On Wed, Jan 13, 2016 at 7:36 PM, <rohitkulkarn...@gmail.com> wrote: > > > Thanks a bunch Jacques! > > > > Sent from my iPhone > > > > > On 14-Jan-2016, at 12:48 AM, Jacques Nadeau <jacq...@dremio.com> > wrote: > > > > > > I think it is most likely probably trivial but unfortunately haven't > had > > > the time to look at it yet. It looks like, for some reason, we're > having > > a > > > failure when serializing the query to pass around between nodes. Let me > > try > > > to take a look today. > > > > > > -- > > > Jacques Nadeau > > > CTO and Co-Founder, Dremio > > > > > > On Wed, Jan 13, 2016 at 3:17 AM, Rohit Kulkarni < > > rohitkulkarn...@gmail.com> > > > wrote: > > > > > >> Is this trivial, or big? > > >> > > >> Thanks, > > >> Rohit > > >> > > >> On Thu, Jan 7, 2016 at 11:29 PM, Boris Chmiel < > > >> boris.chm...@yahoo.com.invalid> wrote: > > >> > > >>> Hi everyone,I do also have this error trying to join MSSQL data > source > > >>> with dfs parquet files. Here the stack : > > >>> org.apache.drill.common.exceptions.UserRemoteException: SYSTEM > > >>> ERROR:IllegalStateException: Already had POJO for id > > >> > > > (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8 > > >> ] > > >>> Fragment2:0 [Error Id: 8431453e-94cb-459a-bc6c-5b5508c7ff84 on > > >>> PC-PC:31010](com.fasterxml.jackson.databind.JsonMappingException) > > Already > > >>> had POJO for > > >> > > > id(java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8 > > >> ] > > >>> (throughreference chain: > > >> > > > org.apache.drill.exec.physical.config.BroadcastSender["destinations"])com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745 > > >>> Caused By (java.lang.IllegalStateException) Alreadyhad POJO for id > > >> > > > (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8 > > >> ] > > >> > > > com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745 > > >>> > > >>> > > >>> > > >>> Le Jeudi 7 janvier 2016 17h08, Rohit Kulkarni < > > >>> rohitkulkarn...@gmail.com> a écrit : > > >>> > > >>> > > >>> Hi Jacques, > > >>> > > >>> Here is the full stack trace as you asked - > > >>> > > >>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO for id > > >>> (java.lang.Integer) > > >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]* > > >>> > > >>> *Fragment 2:0* > > >>> > > >>> *[Error Id: 57494209-04e8-4580-860d-461cf50b41f8 on > > >>> ip-x-x-x-x.ec2.internal:31010]* > > >>> > > >>> * (com.fasterxml.jackson.databind.JsonMappingException) Already had > > POJO > > >>> for id (java.lang.Integer) > > >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8] > > >>> (through reference chain: > > >>> > org.apache.drill.exec.physical.config.BroadcastSender["destinations"])* > > >>> * > > >>> > com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210* > > >>> * > > >>> > com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154* > > >>> * > > >>> > > com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126* > > >>> * > > >> > > > com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113* > > >>> * > > >> > > > com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41* > > >>> * > com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269* > > >>> * com.fasterxml.jackson.databind.ObjectReader.readValue():896* > > >>> * > > >> > > > org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94* > > >>> * org.apache.drill.exec.work.fragment.FragmentExecutor.run():227* > > >>> * org.apache.drill.common.SelfCleaningRunnable.run():38* > > >>> * java.util.concurrent.ThreadPoolExecutor.runWorker():1145* > > >>> * java.util.concurrent.ThreadPoolExecutor$Worker.run():615* > > >>> * java.lang.Thread.run():745* > > >>> * Caused By (java.lang.IllegalStateException) Already had POJO for > id > > >>> (java.lang.Integer) > > >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]* > > >>> * > > >> com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20* > > >>> * > > >>> > > com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154* > > >>> * > > >>> > > com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126* > > >>> * > > >> > > > com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113* > > >>> * > > >> > > > com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132* > > >>> * > > >> > > > com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41* > > >>> * > com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269* > > >>> * com.fasterxml.jackson.databind.ObjectReader.readValue():896* > > >>> * > > >> > > > org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94* > > >>> * org.apache.drill.exec.work.fragment.FragmentExecutor.run():227* > > >>> * org.apache.drill.common.SelfCleaningRunnable.run():38* > > >>> * java.util.concurrent.ThreadPoolExecutor.runWorker():1145* > > >>> * java.util.concurrent.ThreadPoolExecutor$Worker.run():615* > > >>> * java.lang.Thread.run():745 (state=,code=0)* > > >>> > > >>> > > >>>> On Wed, Jan 6, 2016 at 9:44 PM, Jacques Nadeau <jacq...@dremio.com> > > >>> wrote: > > >>> > > >>>> Can you turn on verbose errors and post the full stack trace of the > > >>> error? > > >>>> You can enable verbose errors per the instructions here: > > >>>> > > >>>> > https://drill.apache.org/docs/troubleshooting/#enable-verbose-errors > > >>>> > > >>>> -- > > >>>> Jacques Nadeau > > >>>> CTO and Co-Founder, Dremio > > >>>> > > >>>>> On Wed, Jan 6, 2016 at 6:10 AM, <rohitkulkarn...@gmail.com> wrote: > > >>>>> > > >>>>> Any thoughts on this? I tried so many variants of this query but > same > > >>>>> error! > > >>>>> > > >>>>> Thanks, > > >>>>> Rohit > > >>>>> > > >>>>>> On 06-Jan-2016, at 12:26 AM, Rohit Kulkarni < > > >>> rohitkulkarn...@gmail.com > > >>>>> > > >>>>> wrote: > > >>>>>> > > >>>>>> Thanks a bunch for replying! I quickly ran this - the TAGS_US data > > >> in > > >>>>> HDFS is in parquet format. > > >>>>>> > > >>>>>> select distinct typeof(cvalue) > > >>>>>> > > >>>>>> from hdfs.drill.TAGS_US; > > >>>>>> > > >>>>>> +----------+ > > >>>>>> > > >>>>>> | EXPR$0 | > > >>>>>> > > >>>>>> +----------+ > > >>>>>> > > >>>>>> | VARCHAR | > > >>>>>> > > >>>>>> +----------+ > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> Same with the table in Redshift. I changed my query to > specifically > > >>>> cast > > >>>>> the columns to VARCHAR again - > > >>>>>> > > >>>>>> select count(*) > > >>>>>> > > >>>>>> from redshift.reports.public.us_tags as a > > >>>>>> > > >>>>>> join hdfs.drill.TAGS_US as b > > >>>>>> > > >>>>>> on cast(b.cvalue as varchar) = cast(a.tag_value as varchar) ; > > >>>>>> > > >>>>>> > > >>>>>> I see the same error again. > > >>>>>> > > >>>>>> Here is the explain plan for the query - > > >>>>>> > > >>>>>> select count(*) > > >>>>>> from hdfs.drill.TAGS_US as a > > >>>>>> join redshift.reports.public.us_tags as b > > >>>>>> on a.cvalue = b.tag_value; > > >>>>>> > > >>>>>> Error: SYSTEM ERROR: IllegalStateException: Already had POJO for > id > > >>>>> (java.lang.Integer) > > >>>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8 > ] > > >>>>>> > > >>>>>> > > >>>>>> +------+------+ > > >>>>>> | text | json | > > >>>>>> +------+------+ > > >>>>>> | 00-00 Screen > > >>>>>> 00-01 StreamAgg(group=[{}], EXPR$0=[$SUM0($0)]) > > >>>>>> 00-02 UnionExchange > > >>>>>> 01-01 StreamAgg(group=[{}], EXPR$0=[COUNT()]) > > >>>>>> 01-02 Project($f0=[0]) > > >>>>>> 01-03 HashJoin(condition=[=($0, $1)], > > >> joinType=[inner]) > > >>>>>> 01-05 Scan(groupscan=[ParquetGroupScan > > >>>>> [entries=[ReadEntryWithPath [path=hdfs:// > > >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US]], > > >>>>> selectionRoot=hdfs:// > > >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US, > > >>> numFiles=1, > > >>>>> usedMetadataFile=false, columns=[`cvalue`]]]) > > >>>>>> 01-04 BroadcastExchange > > >>>>>> 02-01 Project(tag_value=[$2]) > > >>>>>> 02-02 Jdbc(sql=[SELECT * > > >>>>>> FROM "reports"."public"."us_tags"]) > > >>>>>> | { > > >>>>>> "head" : { > > >>>>>> "version" : 1, > > >>>>>> "generator" : { > > >>>>>> "type" : "ExplainHandler", > > >>>>>> "info" : "" > > >>>>>> }, > > >>>>>> "type" : "APACHE_DRILL_PHYSICAL", > > >>>>>> "options" : [ ], > > >>>>>> "queue" : 0, > > >>>>>> "resultMode" : "EXEC" > > >>>>>> }, > > >>>>>> "graph" : [ { > > >>>>>> "pop" : "jdbc-scan", > > >>>>>> "@id" : 0, > > >>>>>> "sql" : "SELECT *\nFROM \"reports\".\"public\".\"us_tags\"", > > >>>>>> "config" : { > > >>>>>> "type" : "jdbc", > > >>>>>> "driver" : "com.amazon.redshift.jdbc4.Driver", > > >>>>>> "url" : "", > > >>>>>> "username" : "", > > >>>>>> "password" : "", > > >>>>>> "enabled" : true > > >>>>>> }, > > >>>>>> "userName" : "", > > >>>>>> "cost" : 0.0 > > >>>>>> }, { > > >>>>>> "pop" : "project", > > >>>>>> "@id" : 131073, > > >>>>>> "exprs" : [ { > > >>>>>> "ref" : "`tag_value`", > > >>>>>> "expr" : "`tag_value`" > > >>>>>> } ], > > >>>>>> "child" : 0, > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 100.0 > > >>>>>> }, { > > >>>>>> "pop" : "broadcast-exchange", > > >>>>>> "@id" : 65540, > > >>>>>> "child" : 131073, > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 100.0 > > >>>>>> }, { > > >>>>>> "pop" : "parquet-scan", > > >>>>>> "@id" : 65541, > > >>>>>> "userName" : "XXXX", > > >>>>>> "entries" : [ { > > >>>>>> "path" : "hdfs:// > > >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US" > > >>>>>> } ], > > >>>>>> "storage" : { > > >>>>>> "type" : "file", > > >>>>>> "enabled" : true, > > >>>>>> "connection" : "hdfs:// > > >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/", > > >>>>>> "workspaces" : { > > >>>>>> "root" : { > > >>>>>> "location" : "/", > > >>>>>> "writable" : true, > > >>>>>> "defaultInputFormat" : null > > >>>>>> }, > > >>>>>> "tmp" : { > > >>>>>> "location" : "/tmp", > > >>>>>> "writable" : true, > > >>>>>> "defaultInputFormat" : null > > >>>>>> }, > > >>>>>> "drill" : { > > >>>>>> "location" : "/drill", > > >>>>>> "writable" : true, > > >>>>>> "defaultInputFormat" : "tsv" > > >>>>>> }, > > >>>>>> "drill2" : { > > >>>>>> "location" : "/drill", > > >>>>>> "writable" : true, > > >>>>>> "defaultInputFormat" : "csv" > > >>>>>> } > > >>>>>> }, > > >>>>>> "formats" : { > > >>>>>> "psv" : { > > >>>>>> "type" : "text", > > >>>>>> "extensions" : [ "tbl" ], > > >>>>>> "delimiter" : "|" > > >>>>>> }, > > >>>>>> "csv" : { > > >>>>>> "type" : "text", > > >>>>>> "extensions" : [ "csv" ], > > >>>>>> "delimiter" : "," > > >>>>>> }, > > >>>>>> "tsv" : { > > >>>>>> "type" : "text", > > >>>>>> "extensions" : [ "tsv" ], > > >>>>>> "delimiter" : "\t" > > >>>>>> }, > > >>>>>> "parquet" : { > > >>>>>> "type" : "parquet" > > >>>>>> }, > > >>>>>> "json" : { > > >>>>>> "type" : "json" > > >>>>>> }, > > >>>>>> "avro" : { > > >>>>>> "type" : "avro" > > >>>>>> }, > > >>>>>> "sequencefile" : { > > >>>>>> "type" : "sequencefile", > > >>>>>> "extensions" : [ "seq" ] > > >>>>>> }, > > >>>>>> "csvh" : { > > >>>>>> "type" : "text", > > >>>>>> "extensions" : [ "csvh" ], > > >>>>>> "extractHeader" : true, > > >>>>>> "delimiter" : "," > > >>>>>> } > > >>>>>> } > > >>>>>> }, > > >>>>>> "format" : { > > >>>>>> "type" : "parquet" > > >>>>>> }, > > >>>>>> "columns" : [ "`cvalue`" ], > > >>>>>> "selectionRoot" : "hdfs:// > > >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US", > > >>>>>> "fileSet" : [ "/drill/TAGS_US/0_0_1.parquet", > > >>>>> "/drill/TAGS_US/0_0_0.parquet" ], > > >>>>>> "cost" : 4.1667342E7 > > >>>>>> }, { > > >>>>>> "pop" : "hash-join", > > >>>>>> "@id" : 65539, > > >>>>>> "left" : 65541, > > >>>>>> "right" : 65540, > > >>>>>> "conditions" : [ { > > >>>>>> "relationship" : "EQUALS", > > >>>>>> "left" : "`cvalue`", > > >>>>>> "right" : "`tag_value`" > > >>>>>> } ], > > >>>>>> "joinType" : "INNER", > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 4.1667342E7 > > >>>>>> }, { > > >>>>>> "pop" : "project", > > >>>>>> "@id" : 65538, > > >>>>>> "exprs" : [ { > > >>>>>> "ref" : "`$f0`", > > >>>>>> "expr" : "0" > > >>>>>> } ], > > >>>>>> "child" : 65539, > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 4.1667342E7 > > >>>>>> }, { > > >>>>>> "pop" : "streaming-aggregate", > > >>>>>> "@id" : 65537, > > >>>>>> "child" : 65538, > > >>>>>> "keys" : [ ], > > >>>>>> "exprs" : [ { > > >>>>>> "ref" : "`EXPR$0`", > > >>>>>> "expr" : "count(1) " > > >>>>>> } ], > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 1.0 > > >>>>>> }, { > > >>>>>> "pop" : "union-exchange", > > >>>>>> "@id" : 2, > > >>>>>> "child" : 65537, > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 1.0 > > >>>>>> }, { > > >>>>>> "pop" : "streaming-aggregate", > > >>>>>> "@id" : 1, > > >>>>>> "child" : 2, > > >>>>>> "keys" : [ ], > > >>>>>> "exprs" : [ { > > >>>>>> "ref" : "`EXPR$0`", > > >>>>>> "expr" : "$sum0(`EXPR$0`) " > > >>>>>> } ], > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 1.0 > > >>>>>> }, { > > >>>>>> "pop" : "screen", > > >>>>>> "@id" : 0, > > >>>>>> "child" : 1, > > >>>>>> "initialAllocation" : 1000000, > > >>>>>> "maxAllocation" : 10000000000, > > >>>>>> "cost" : 1.0 > > >>>>>> } ] > > >>>>>> } | > > >>>>>> +------+------+ > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>>> On Mon, Jan 4, 2016 at 9:42 PM, Andries Engelbrecht < > > >>>>> aengelbre...@maprtech.com> wrote: > > >>>>>>> Perhaps check the data type of all the fields being used for the > > >>> join. > > >>>>>>> > > >>>>>>> Select cvalue, TYPEOF(cvalue) from hdfs...... limit 10 > > >>>>>>> > > >>>>>>> and similar for tag_value on redshift. > > >>>>>>> > > >>>>>>> You can then do a predicate to find records where the data type > > >> may > > >>> be > > >>>>> different. > > >>>>>>> where typeof(<field>) not like '<data type of field>' > > >>>>>>> > > >>>>>>> I believe there was a nice write up on they topic, but can't find > > >> it > > >>>>> now. > > >>>>>>> > > >>>>>>> > > >>>>>>> --Andries > > >>>>>>> > > >>>>>>> > > >>>>>>> > > >>>>>>> > > >>>>>>>> On Jan 3, 2016, at 8:45 PM, Rohit Kulkarni < > > >>>> rohitkulkarn...@gmail.com> > > >>>>> wrote: > > >>>>>>>> > > >>>>>>>> Hello all, > > >>>>>>>> > > >>>>>>>> I am sure if not all of you, but some of you must have seen this > > >>>>> error some > > >>>>>>>> time - > > >>>>>>>> > > >>>>>>>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO > > >> for > > >>> id > > >>>>>>>> (java.lang.Integer) > > >> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8 > > >>>> ]* > > >>>>>>>> > > >>>>>>>> I am trying to do a join between Redshift (JDBC) and HDFS like > > >>>> this > > >>>>> - > > >>>>>>>> > > >>>>>>>> > > >>>>>>>> > > >>>>>>>> > > >>>>>>>> *select count(*)from hdfs.drill.TAGS_US as aright join > > >>>>>>>> redshift.reports.public.us_tags as bon a.cvalue = b.tag_value;* > > >>>>>>>> > > >>>>>>>> > > >>>>>>>> I don't see anything wrong in the query. The two individual > > >> tables > > >>>>> return > > >>>>>>>> proper data when fired a query separately. Is something missing > > >> or > > >>>> am > > >>>>> I > > >>>>>>>> doing something wrong? > > >>>>>>>> > > >>>>>>>> Would very much appreciate your help! Thanks!! > > >>>>>>>> > > >>>>>>>> -- > > >>>>>>>> Warm Regards, > > >>>>>>>> Rohit Kulkarni > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> -- > > >>>>>> Warm Regards, > > >>>>>> Rohit Kulkarni > > >>>>>> Mo.: +91 89394 63593 > > >>> > > >>> > > >>> > > >>> -- > > >>> Warm Regards, > > >>> Rohit Kulkarni > > >>> Mo.: +91 89394 63593 > > >> > > >> > > >> > > >> -- > > >> Warm Regards, > > >> Rohit Kulkarni > > >> Mo.: +91 89394 63593 > > >> > > >