Hey Rohit,

I'm having trouble reproducing this in my environment (pointing at derby +
hdfs instead of redshift/postgres). Can you turn on debug logging and then
run this query? You can enable the debug logging we are interested in by
adding the following item to logback.xml:

  <logger
name="org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler"
additivity="false">
    <level value="debug" />
    <appender-ref ref="FILE" />
  </logger>

This is the query that I have succesfully completing. Please confirm it is
similar to your query:
SELECT count(*)
FROM dfs.`/data/tpch01/line/` a
RIGHT JOIN derby.DRILL_DERBY_TEST.PERSON b
ON a.cvalue = b.person_id



--
Jacques Nadeau
CTO and Co-Founder, Dremio

On Wed, Jan 13, 2016 at 7:36 PM, <[email protected]> wrote:

> Thanks a bunch Jacques!
>
> Sent from my iPhone
>
> > On 14-Jan-2016, at 12:48 AM, Jacques Nadeau <[email protected]> wrote:
> >
> > I think it is most likely probably trivial but unfortunately haven't had
> > the time to look at it yet. It looks like, for some reason, we're having
> a
> > failure when serializing the query to pass around between nodes. Let me
> try
> > to take a look today.
> >
> > --
> > Jacques Nadeau
> > CTO and Co-Founder, Dremio
> >
> > On Wed, Jan 13, 2016 at 3:17 AM, Rohit Kulkarni <
> [email protected]>
> > wrote:
> >
> >> Is this trivial, or big?
> >>
> >> Thanks,
> >> Rohit
> >>
> >> On Thu, Jan 7, 2016 at 11:29 PM, Boris Chmiel <
> >> [email protected]> wrote:
> >>
> >>> Hi everyone,I do also have this error trying to join MSSQL data source
> >>> with dfs parquet files. Here the stack :
> >>> org.apache.drill.common.exceptions.UserRemoteException: SYSTEM
> >>> ERROR:IllegalStateException: Already had POJO for id
> >>
> (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
> >> ]
> >>> Fragment2:0 [Error Id: 8431453e-94cb-459a-bc6c-5b5508c7ff84 on
> >>> PC-PC:31010](com.fasterxml.jackson.databind.JsonMappingException)
> Already
> >>> had POJO for
> >>
> id(java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
> >> ]
> >>> (throughreference chain:
> >>
> org.apache.drill.exec.physical.config.BroadcastSender["destinations"])com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745
> >>> Caused By (java.lang.IllegalStateException) Alreadyhad POJO for id
> >>
> (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
> >> ]
> >>
> com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745
> >>>
> >>>
> >>>
> >>>    Le Jeudi 7 janvier 2016 17h08, Rohit Kulkarni <
> >>> [email protected]> a écrit :
> >>>
> >>>
> >>> Hi Jacques,
> >>>
> >>> Here is the full stack trace as you asked -
> >>>
> >>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO for id
> >>> (java.lang.Integer)
> >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]*
> >>>
> >>> *Fragment 2:0*
> >>>
> >>> *[Error Id: 57494209-04e8-4580-860d-461cf50b41f8 on
> >>> ip-x-x-x-x.ec2.internal:31010]*
> >>>
> >>> *  (com.fasterxml.jackson.databind.JsonMappingException) Already had
> POJO
> >>> for id (java.lang.Integer)
> >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]
> >>> (through reference chain:
> >>> org.apache.drill.exec.physical.config.BroadcastSender["destinations"])*
> >>> *
> >>> com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210*
> >>> *
> >>> com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154*
> >>> *
> >>>
> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126*
> >>> *
> >>
> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113*
> >>> *
> >>
> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41*
> >>> *    com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269*
> >>> *    com.fasterxml.jackson.databind.ObjectReader.readValue():896*
> >>> *
> >>
> org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94*
> >>> *    org.apache.drill.exec.work.fragment.FragmentExecutor.run():227*
> >>> *    org.apache.drill.common.SelfCleaningRunnable.run():38*
> >>> *    java.util.concurrent.ThreadPoolExecutor.runWorker():1145*
> >>> *    java.util.concurrent.ThreadPoolExecutor$Worker.run():615*
> >>> *    java.lang.Thread.run():745*
> >>> *  Caused By (java.lang.IllegalStateException) Already had POJO for id
> >>> (java.lang.Integer)
> >>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]*
> >>> *
> >> com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20*
> >>> *
> >>>
> com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154*
> >>> *
> >>>
> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126*
> >>> *
> >>
> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113*
> >>> *
> >>
> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132*
> >>> *
> >>
> com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41*
> >>> *    com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269*
> >>> *    com.fasterxml.jackson.databind.ObjectReader.readValue():896*
> >>> *
> >>
> org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94*
> >>> *    org.apache.drill.exec.work.fragment.FragmentExecutor.run():227*
> >>> *    org.apache.drill.common.SelfCleaningRunnable.run():38*
> >>> *    java.util.concurrent.ThreadPoolExecutor.runWorker():1145*
> >>> *    java.util.concurrent.ThreadPoolExecutor$Worker.run():615*
> >>> *    java.lang.Thread.run():745 (state=,code=0)*
> >>>
> >>>
> >>>> On Wed, Jan 6, 2016 at 9:44 PM, Jacques Nadeau <[email protected]>
> >>> wrote:
> >>>
> >>>> Can you turn on verbose errors and post the full stack trace of the
> >>> error?
> >>>> You can enable verbose errors per the instructions here:
> >>>>
> >>>> https://drill.apache.org/docs/troubleshooting/#enable-verbose-errors
> >>>>
> >>>> --
> >>>> Jacques Nadeau
> >>>> CTO and Co-Founder, Dremio
> >>>>
> >>>>> On Wed, Jan 6, 2016 at 6:10 AM, <[email protected]> wrote:
> >>>>>
> >>>>> Any thoughts on this? I tried so many variants of this query but same
> >>>>> error!
> >>>>>
> >>>>> Thanks,
> >>>>> Rohit
> >>>>>
> >>>>>> On 06-Jan-2016, at 12:26 AM, Rohit Kulkarni <
> >>> [email protected]
> >>>>>
> >>>>> wrote:
> >>>>>>
> >>>>>> Thanks a bunch for replying! I quickly ran this - the TAGS_US data
> >> in
> >>>>> HDFS is in parquet format.
> >>>>>>
> >>>>>> select distinct typeof(cvalue)
> >>>>>>
> >>>>>> from hdfs.drill.TAGS_US;
> >>>>>>
> >>>>>> +----------+
> >>>>>>
> >>>>>> |  EXPR$0  |
> >>>>>>
> >>>>>> +----------+
> >>>>>>
> >>>>>> | VARCHAR  |
> >>>>>>
> >>>>>> +----------+
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> Same with the table in Redshift. I changed my query to specifically
> >>>> cast
> >>>>> the columns to VARCHAR again -
> >>>>>>
> >>>>>> select count(*)
> >>>>>>
> >>>>>> from redshift.reports.public.us_tags as a
> >>>>>>
> >>>>>> join hdfs.drill.TAGS_US as b
> >>>>>>
> >>>>>> on cast(b.cvalue as varchar) = cast(a.tag_value as varchar) ;
> >>>>>>
> >>>>>>
> >>>>>> I see the same error again.
> >>>>>>
> >>>>>> Here is the explain plan for the query -
> >>>>>>
> >>>>>> select count(*)
> >>>>>> from hdfs.drill.TAGS_US as a
> >>>>>> join redshift.reports.public.us_tags as b
> >>>>>> on a.cvalue = b.tag_value;
> >>>>>>
> >>>>>> Error: SYSTEM ERROR: IllegalStateException: Already had POJO for id
> >>>>> (java.lang.Integer)
> >>>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]
> >>>>>>
> >>>>>>
> >>>>>> +------+------+
> >>>>>> | text | json |
> >>>>>> +------+------+
> >>>>>> | 00-00    Screen
> >>>>>> 00-01      StreamAgg(group=[{}], EXPR$0=[$SUM0($0)])
> >>>>>> 00-02        UnionExchange
> >>>>>> 01-01          StreamAgg(group=[{}], EXPR$0=[COUNT()])
> >>>>>> 01-02            Project($f0=[0])
> >>>>>> 01-03              HashJoin(condition=[=($0, $1)],
> >> joinType=[inner])
> >>>>>> 01-05                Scan(groupscan=[ParquetGroupScan
> >>>>> [entries=[ReadEntryWithPath [path=hdfs://
> >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US]],
> >>>>> selectionRoot=hdfs://
> >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US,
> >>> numFiles=1,
> >>>>> usedMetadataFile=false, columns=[`cvalue`]]])
> >>>>>> 01-04                BroadcastExchange
> >>>>>> 02-01                  Project(tag_value=[$2])
> >>>>>> 02-02                    Jdbc(sql=[SELECT *
> >>>>>> FROM "reports"."public"."us_tags"])
> >>>>>> | {
> >>>>>> "head" : {
> >>>>>>   "version" : 1,
> >>>>>>   "generator" : {
> >>>>>>     "type" : "ExplainHandler",
> >>>>>>     "info" : ""
> >>>>>>   },
> >>>>>>   "type" : "APACHE_DRILL_PHYSICAL",
> >>>>>>   "options" : [ ],
> >>>>>>   "queue" : 0,
> >>>>>>   "resultMode" : "EXEC"
> >>>>>> },
> >>>>>> "graph" : [ {
> >>>>>>   "pop" : "jdbc-scan",
> >>>>>>   "@id" : 0,
> >>>>>>   "sql" : "SELECT *\nFROM \"reports\".\"public\".\"us_tags\"",
> >>>>>>   "config" : {
> >>>>>>     "type" : "jdbc",
> >>>>>>     "driver" : "com.amazon.redshift.jdbc4.Driver",
> >>>>>>     "url" : "",
> >>>>>>     "username" : "",
> >>>>>>     "password" : "",
> >>>>>>     "enabled" : true
> >>>>>>   },
> >>>>>>   "userName" : "",
> >>>>>>   "cost" : 0.0
> >>>>>> }, {
> >>>>>>   "pop" : "project",
> >>>>>>   "@id" : 131073,
> >>>>>>   "exprs" : [ {
> >>>>>>     "ref" : "`tag_value`",
> >>>>>>     "expr" : "`tag_value`"
> >>>>>>   } ],
> >>>>>>   "child" : 0,
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 100.0
> >>>>>> }, {
> >>>>>>   "pop" : "broadcast-exchange",
> >>>>>>   "@id" : 65540,
> >>>>>>   "child" : 131073,
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 100.0
> >>>>>> }, {
> >>>>>>   "pop" : "parquet-scan",
> >>>>>>   "@id" : 65541,
> >>>>>>   "userName" : "XXXX",
> >>>>>>   "entries" : [ {
> >>>>>>     "path" : "hdfs://
> >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US"
> >>>>>>   } ],
> >>>>>>   "storage" : {
> >>>>>>     "type" : "file",
> >>>>>>     "enabled" : true,
> >>>>>>     "connection" : "hdfs://
> >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/",
> >>>>>>     "workspaces" : {
> >>>>>>       "root" : {
> >>>>>>         "location" : "/",
> >>>>>>         "writable" : true,
> >>>>>>         "defaultInputFormat" : null
> >>>>>>       },
> >>>>>>       "tmp" : {
> >>>>>>         "location" : "/tmp",
> >>>>>>         "writable" : true,
> >>>>>>         "defaultInputFormat" : null
> >>>>>>       },
> >>>>>>       "drill" : {
> >>>>>>         "location" : "/drill",
> >>>>>>         "writable" : true,
> >>>>>>         "defaultInputFormat" : "tsv"
> >>>>>>       },
> >>>>>>       "drill2" : {
> >>>>>>         "location" : "/drill",
> >>>>>>         "writable" : true,
> >>>>>>         "defaultInputFormat" : "csv"
> >>>>>>       }
> >>>>>>     },
> >>>>>>     "formats" : {
> >>>>>>       "psv" : {
> >>>>>>         "type" : "text",
> >>>>>>         "extensions" : [ "tbl" ],
> >>>>>>         "delimiter" : "|"
> >>>>>>       },
> >>>>>>       "csv" : {
> >>>>>>         "type" : "text",
> >>>>>>         "extensions" : [ "csv" ],
> >>>>>>         "delimiter" : ","
> >>>>>>       },
> >>>>>>       "tsv" : {
> >>>>>>         "type" : "text",
> >>>>>>         "extensions" : [ "tsv" ],
> >>>>>>         "delimiter" : "\t"
> >>>>>>       },
> >>>>>>       "parquet" : {
> >>>>>>         "type" : "parquet"
> >>>>>>       },
> >>>>>>       "json" : {
> >>>>>>         "type" : "json"
> >>>>>>       },
> >>>>>>       "avro" : {
> >>>>>>         "type" : "avro"
> >>>>>>       },
> >>>>>>       "sequencefile" : {
> >>>>>>         "type" : "sequencefile",
> >>>>>>         "extensions" : [ "seq" ]
> >>>>>>       },
> >>>>>>       "csvh" : {
> >>>>>>         "type" : "text",
> >>>>>>         "extensions" : [ "csvh" ],
> >>>>>>         "extractHeader" : true,
> >>>>>>         "delimiter" : ","
> >>>>>>       }
> >>>>>>     }
> >>>>>>   },
> >>>>>>   "format" : {
> >>>>>>     "type" : "parquet"
> >>>>>>   },
> >>>>>>   "columns" : [ "`cvalue`" ],
> >>>>>>   "selectionRoot" : "hdfs://
> >>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US",
> >>>>>>   "fileSet" : [ "/drill/TAGS_US/0_0_1.parquet",
> >>>>> "/drill/TAGS_US/0_0_0.parquet" ],
> >>>>>>   "cost" : 4.1667342E7
> >>>>>> }, {
> >>>>>>   "pop" : "hash-join",
> >>>>>>   "@id" : 65539,
> >>>>>>   "left" : 65541,
> >>>>>>   "right" : 65540,
> >>>>>>   "conditions" : [ {
> >>>>>>     "relationship" : "EQUALS",
> >>>>>>     "left" : "`cvalue`",
> >>>>>>     "right" : "`tag_value`"
> >>>>>>   } ],
> >>>>>>   "joinType" : "INNER",
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 4.1667342E7
> >>>>>> }, {
> >>>>>>   "pop" : "project",
> >>>>>>   "@id" : 65538,
> >>>>>>   "exprs" : [ {
> >>>>>>     "ref" : "`$f0`",
> >>>>>>     "expr" : "0"
> >>>>>>   } ],
> >>>>>>   "child" : 65539,
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 4.1667342E7
> >>>>>> }, {
> >>>>>>   "pop" : "streaming-aggregate",
> >>>>>>   "@id" : 65537,
> >>>>>>   "child" : 65538,
> >>>>>>   "keys" : [ ],
> >>>>>>   "exprs" : [ {
> >>>>>>     "ref" : "`EXPR$0`",
> >>>>>>     "expr" : "count(1) "
> >>>>>>   } ],
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 1.0
> >>>>>> }, {
> >>>>>>   "pop" : "union-exchange",
> >>>>>>   "@id" : 2,
> >>>>>>   "child" : 65537,
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 1.0
> >>>>>> }, {
> >>>>>>   "pop" : "streaming-aggregate",
> >>>>>>   "@id" : 1,
> >>>>>>   "child" : 2,
> >>>>>>   "keys" : [ ],
> >>>>>>   "exprs" : [ {
> >>>>>>     "ref" : "`EXPR$0`",
> >>>>>>     "expr" : "$sum0(`EXPR$0`) "
> >>>>>>   } ],
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 1.0
> >>>>>> }, {
> >>>>>>   "pop" : "screen",
> >>>>>>   "@id" : 0,
> >>>>>>   "child" : 1,
> >>>>>>   "initialAllocation" : 1000000,
> >>>>>>   "maxAllocation" : 10000000000,
> >>>>>>   "cost" : 1.0
> >>>>>> } ]
> >>>>>> } |
> >>>>>> +------+------+
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>> On Mon, Jan 4, 2016 at 9:42 PM, Andries Engelbrecht <
> >>>>> [email protected]> wrote:
> >>>>>>> Perhaps check the data type of all the fields being used for the
> >>> join.
> >>>>>>>
> >>>>>>> Select cvalue, TYPEOF(cvalue) from hdfs...... limit 10
> >>>>>>>
> >>>>>>> and similar for  tag_value on redshift.
> >>>>>>>
> >>>>>>> You can then do a predicate to find records where the data type
> >> may
> >>> be
> >>>>> different.
> >>>>>>> where typeof(<field>) not like '<data type of field>'
> >>>>>>>
> >>>>>>> I believe there was a nice write up on they topic, but can't find
> >> it
> >>>>> now.
> >>>>>>>
> >>>>>>>
> >>>>>>> --Andries
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>> On Jan 3, 2016, at 8:45 PM, Rohit Kulkarni <
> >>>> [email protected]>
> >>>>> wrote:
> >>>>>>>>
> >>>>>>>> Hello all,
> >>>>>>>>
> >>>>>>>> I am sure if not all of you, but some of you must have seen this
> >>>>> error some
> >>>>>>>> time -
> >>>>>>>>
> >>>>>>>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO
> >> for
> >>> id
> >>>>>>>> (java.lang.Integer)
> >> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
> >>>> ]*
> >>>>>>>>
> >>>>>>>> ​I am trying to do a join between Redshift (JDBC​) and HDFS like
> >>>> this
> >>>>> -
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> *select count(*)from hdfs.drill.TAGS_US as aright join
> >>>>>>>> redshift.reports.public.us_tags as bon a.cvalue = b.tag_value;*
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> I don't see anything wrong in the query. The two individual
> >> tables
> >>>>> return
> >>>>>>>> proper data when fired a query separately. Is something missing
> >> or
> >>>> am
> >>>>> I
> >>>>>>>> doing something wrong?
> >>>>>>>>
> >>>>>>>> Would very much appreciate your help! Thanks!!
> >>>>>>>>
> >>>>>>>> --
> >>>>>>>> Warm Regards,
> >>>>>>>> Rohit Kulkarni
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> --
> >>>>>> Warm Regards,
> >>>>>> Rohit Kulkarni
> >>>>>> Mo.: +91 89394 63593
> >>>
> >>>
> >>>
> >>> --
> >>> Warm Regards,
> >>> Rohit Kulkarni
> >>> Mo.: +91 89394 63593
> >>
> >>
> >>
> >> --
> >> Warm Regards,
> >> Rohit Kulkarni
> >> Mo.: +91 89394 63593
> >>
>

Reply via email to