Thanks a bunch Jacques!

Sent from my iPhone

> On 14-Jan-2016, at 12:48 AM, Jacques Nadeau <[email protected]> wrote:
> 
> I think it is most likely probably trivial but unfortunately haven't had
> the time to look at it yet. It looks like, for some reason, we're having a
> failure when serializing the query to pass around between nodes. Let me try
> to take a look today.
> 
> --
> Jacques Nadeau
> CTO and Co-Founder, Dremio
> 
> On Wed, Jan 13, 2016 at 3:17 AM, Rohit Kulkarni <[email protected]>
> wrote:
> 
>> Is this trivial, or big?
>> 
>> Thanks,
>> Rohit
>> 
>> On Thu, Jan 7, 2016 at 11:29 PM, Boris Chmiel <
>> [email protected]> wrote:
>> 
>>> Hi everyone,I do also have this error trying to join MSSQL data source
>>> with dfs parquet files. Here the stack :
>>> org.apache.drill.common.exceptions.UserRemoteException: SYSTEM
>>> ERROR:IllegalStateException: Already had POJO for id
>> (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
>> ]
>>> Fragment2:0 [Error Id: 8431453e-94cb-459a-bc6c-5b5508c7ff84 on
>>> PC-PC:31010](com.fasterxml.jackson.databind.JsonMappingException) Already
>>> had POJO for
>> id(java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
>> ]
>>> (throughreference chain:
>> org.apache.drill.exec.physical.config.BroadcastSender["destinations"])com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745
>>> Caused By (java.lang.IllegalStateException) Alreadyhad POJO for id
>> (java.lang.Integer)[com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
>> ]
>> com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269com.fasterxml.jackson.databind.ObjectReader.readValue():896org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94org.apache.drill.exec.work.fragment.FragmentExecutor.run():227org.apache.drill.common.SelfCleaningRunnable.run():38java.util.concurrent.ThreadPoolExecutor.runWorker():1145java.util.concurrent.ThreadPoolExecutor$Worker.run():615java.lang.Thread.run():745
>>> 
>>> 
>>> 
>>>    Le Jeudi 7 janvier 2016 17h08, Rohit Kulkarni <
>>> [email protected]> a écrit :
>>> 
>>> 
>>> Hi Jacques,
>>> 
>>> Here is the full stack trace as you asked -
>>> 
>>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO for id
>>> (java.lang.Integer)
>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]*
>>> 
>>> *Fragment 2:0*
>>> 
>>> *[Error Id: 57494209-04e8-4580-860d-461cf50b41f8 on
>>> ip-x-x-x-x.ec2.internal:31010]*
>>> 
>>> *  (com.fasterxml.jackson.databind.JsonMappingException) Already had POJO
>>> for id (java.lang.Integer)
>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]
>>> (through reference chain:
>>> org.apache.drill.exec.physical.config.BroadcastSender["destinations"])*
>>> *
>>> com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():210*
>>> *
>>> com.fasterxml.jackson.databind.JsonMappingException.wrapWithPath():177*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.wrapAndThrow():1420*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():351*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154*
>>> *
>>> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126*
>>> *
>> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113*
>>> *
>> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84*
>>> *
>> com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132*
>>> *
>> com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41*
>>> *    com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269*
>>> *    com.fasterxml.jackson.databind.ObjectReader.readValue():896*
>>> *
>> org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94*
>>> *    org.apache.drill.exec.work.fragment.FragmentExecutor.run():227*
>>> *    org.apache.drill.common.SelfCleaningRunnable.run():38*
>>> *    java.util.concurrent.ThreadPoolExecutor.runWorker():1145*
>>> *    java.util.concurrent.ThreadPoolExecutor$Worker.run():615*
>>> *    java.lang.Thread.run():745*
>>> *  Caused By (java.lang.IllegalStateException) Already had POJO for id
>>> (java.lang.Integer)
>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]*
>>> *
>> com.fasterxml.jackson.annotation.SimpleObjectIdResolver.bindItem():20*
>>> *
>>> com.fasterxml.jackson.databind.deser.impl.ReadableObjectId.bindItem():66*
>>> *
>> com.fasterxml.jackson.databind.deser.impl.PropertyValueBuffer.handleIdValue():117*
>>> *
>> com.fasterxml.jackson.databind.deser.impl.PropertyBasedCreator.build():169*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased():349*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault():1056*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject():264*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeWithObjectId():1028*
>>> *
>> com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther():154*
>>> *
>>> com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize():126*
>>> *
>> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId():113*
>>> *
>> com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject():84*
>>> *
>> com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType():132*
>>> *
>> com.fasterxml.jackson.databind.deser.impl.TypeWrappedDeserializer.deserialize():41*
>>> *    com.fasterxml.jackson.databind.ObjectReader._bindAndClose():1269*
>>> *    com.fasterxml.jackson.databind.ObjectReader.readValue():896*
>>> *
>> org.apache.drill.exec.planner.PhysicalPlanReader.readFragmentOperator():94*
>>> *    org.apache.drill.exec.work.fragment.FragmentExecutor.run():227*
>>> *    org.apache.drill.common.SelfCleaningRunnable.run():38*
>>> *    java.util.concurrent.ThreadPoolExecutor.runWorker():1145*
>>> *    java.util.concurrent.ThreadPoolExecutor$Worker.run():615*
>>> *    java.lang.Thread.run():745 (state=,code=0)*
>>> 
>>> 
>>>> On Wed, Jan 6, 2016 at 9:44 PM, Jacques Nadeau <[email protected]>
>>> wrote:
>>> 
>>>> Can you turn on verbose errors and post the full stack trace of the
>>> error?
>>>> You can enable verbose errors per the instructions here:
>>>> 
>>>> https://drill.apache.org/docs/troubleshooting/#enable-verbose-errors
>>>> 
>>>> --
>>>> Jacques Nadeau
>>>> CTO and Co-Founder, Dremio
>>>> 
>>>>> On Wed, Jan 6, 2016 at 6:10 AM, <[email protected]> wrote:
>>>>> 
>>>>> Any thoughts on this? I tried so many variants of this query but same
>>>>> error!
>>>>> 
>>>>> Thanks,
>>>>> Rohit
>>>>> 
>>>>>> On 06-Jan-2016, at 12:26 AM, Rohit Kulkarni <
>>> [email protected]
>>>>> 
>>>>> wrote:
>>>>>> 
>>>>>> Thanks a bunch for replying! I quickly ran this - the TAGS_US data
>> in
>>>>> HDFS is in parquet format.
>>>>>> 
>>>>>> select distinct typeof(cvalue)
>>>>>> 
>>>>>> from hdfs.drill.TAGS_US;
>>>>>> 
>>>>>> +----------+
>>>>>> 
>>>>>> |  EXPR$0  |
>>>>>> 
>>>>>> +----------+
>>>>>> 
>>>>>> | VARCHAR  |
>>>>>> 
>>>>>> +----------+
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> Same with the table in Redshift. I changed my query to specifically
>>>> cast
>>>>> the columns to VARCHAR again -
>>>>>> 
>>>>>> select count(*)
>>>>>> 
>>>>>> from redshift.reports.public.us_tags as a
>>>>>> 
>>>>>> join hdfs.drill.TAGS_US as b
>>>>>> 
>>>>>> on cast(b.cvalue as varchar) = cast(a.tag_value as varchar) ;
>>>>>> 
>>>>>> 
>>>>>> I see the same error again.
>>>>>> 
>>>>>> Here is the explain plan for the query -
>>>>>> 
>>>>>> select count(*)
>>>>>> from hdfs.drill.TAGS_US as a
>>>>>> join redshift.reports.public.us_tags as b
>>>>>> on a.cvalue = b.tag_value;
>>>>>> 
>>>>>> Error: SYSTEM ERROR: IllegalStateException: Already had POJO for id
>>>>> (java.lang.Integer)
>>>>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8]
>>>>>> 
>>>>>> 
>>>>>> +------+------+
>>>>>> | text | json |
>>>>>> +------+------+
>>>>>> | 00-00    Screen
>>>>>> 00-01      StreamAgg(group=[{}], EXPR$0=[$SUM0($0)])
>>>>>> 00-02        UnionExchange
>>>>>> 01-01          StreamAgg(group=[{}], EXPR$0=[COUNT()])
>>>>>> 01-02            Project($f0=[0])
>>>>>> 01-03              HashJoin(condition=[=($0, $1)],
>> joinType=[inner])
>>>>>> 01-05                Scan(groupscan=[ParquetGroupScan
>>>>> [entries=[ReadEntryWithPath [path=hdfs://
>>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US]],
>>>>> selectionRoot=hdfs://
>>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US,
>>> numFiles=1,
>>>>> usedMetadataFile=false, columns=[`cvalue`]]])
>>>>>> 01-04                BroadcastExchange
>>>>>> 02-01                  Project(tag_value=[$2])
>>>>>> 02-02                    Jdbc(sql=[SELECT *
>>>>>> FROM "reports"."public"."us_tags"])
>>>>>> | {
>>>>>> "head" : {
>>>>>>   "version" : 1,
>>>>>>   "generator" : {
>>>>>>     "type" : "ExplainHandler",
>>>>>>     "info" : ""
>>>>>>   },
>>>>>>   "type" : "APACHE_DRILL_PHYSICAL",
>>>>>>   "options" : [ ],
>>>>>>   "queue" : 0,
>>>>>>   "resultMode" : "EXEC"
>>>>>> },
>>>>>> "graph" : [ {
>>>>>>   "pop" : "jdbc-scan",
>>>>>>   "@id" : 0,
>>>>>>   "sql" : "SELECT *\nFROM \"reports\".\"public\".\"us_tags\"",
>>>>>>   "config" : {
>>>>>>     "type" : "jdbc",
>>>>>>     "driver" : "com.amazon.redshift.jdbc4.Driver",
>>>>>>     "url" : "",
>>>>>>     "username" : "",
>>>>>>     "password" : "",
>>>>>>     "enabled" : true
>>>>>>   },
>>>>>>   "userName" : "",
>>>>>>   "cost" : 0.0
>>>>>> }, {
>>>>>>   "pop" : "project",
>>>>>>   "@id" : 131073,
>>>>>>   "exprs" : [ {
>>>>>>     "ref" : "`tag_value`",
>>>>>>     "expr" : "`tag_value`"
>>>>>>   } ],
>>>>>>   "child" : 0,
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 100.0
>>>>>> }, {
>>>>>>   "pop" : "broadcast-exchange",
>>>>>>   "@id" : 65540,
>>>>>>   "child" : 131073,
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 100.0
>>>>>> }, {
>>>>>>   "pop" : "parquet-scan",
>>>>>>   "@id" : 65541,
>>>>>>   "userName" : "XXXX",
>>>>>>   "entries" : [ {
>>>>>>     "path" : "hdfs://
>>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US"
>>>>>>   } ],
>>>>>>   "storage" : {
>>>>>>     "type" : "file",
>>>>>>     "enabled" : true,
>>>>>>     "connection" : "hdfs://
>>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/",
>>>>>>     "workspaces" : {
>>>>>>       "root" : {
>>>>>>         "location" : "/",
>>>>>>         "writable" : true,
>>>>>>         "defaultInputFormat" : null
>>>>>>       },
>>>>>>       "tmp" : {
>>>>>>         "location" : "/tmp",
>>>>>>         "writable" : true,
>>>>>>         "defaultInputFormat" : null
>>>>>>       },
>>>>>>       "drill" : {
>>>>>>         "location" : "/drill",
>>>>>>         "writable" : true,
>>>>>>         "defaultInputFormat" : "tsv"
>>>>>>       },
>>>>>>       "drill2" : {
>>>>>>         "location" : "/drill",
>>>>>>         "writable" : true,
>>>>>>         "defaultInputFormat" : "csv"
>>>>>>       }
>>>>>>     },
>>>>>>     "formats" : {
>>>>>>       "psv" : {
>>>>>>         "type" : "text",
>>>>>>         "extensions" : [ "tbl" ],
>>>>>>         "delimiter" : "|"
>>>>>>       },
>>>>>>       "csv" : {
>>>>>>         "type" : "text",
>>>>>>         "extensions" : [ "csv" ],
>>>>>>         "delimiter" : ","
>>>>>>       },
>>>>>>       "tsv" : {
>>>>>>         "type" : "text",
>>>>>>         "extensions" : [ "tsv" ],
>>>>>>         "delimiter" : "\t"
>>>>>>       },
>>>>>>       "parquet" : {
>>>>>>         "type" : "parquet"
>>>>>>       },
>>>>>>       "json" : {
>>>>>>         "type" : "json"
>>>>>>       },
>>>>>>       "avro" : {
>>>>>>         "type" : "avro"
>>>>>>       },
>>>>>>       "sequencefile" : {
>>>>>>         "type" : "sequencefile",
>>>>>>         "extensions" : [ "seq" ]
>>>>>>       },
>>>>>>       "csvh" : {
>>>>>>         "type" : "text",
>>>>>>         "extensions" : [ "csvh" ],
>>>>>>         "extractHeader" : true,
>>>>>>         "delimiter" : ","
>>>>>>       }
>>>>>>     }
>>>>>>   },
>>>>>>   "format" : {
>>>>>>     "type" : "parquet"
>>>>>>   },
>>>>>>   "columns" : [ "`cvalue`" ],
>>>>>>   "selectionRoot" : "hdfs://
>>>>> ec2-XX-XX-XX-XX.compute-1.amazonaws.com:8020/drill/TAGS_US",
>>>>>>   "fileSet" : [ "/drill/TAGS_US/0_0_1.parquet",
>>>>> "/drill/TAGS_US/0_0_0.parquet" ],
>>>>>>   "cost" : 4.1667342E7
>>>>>> }, {
>>>>>>   "pop" : "hash-join",
>>>>>>   "@id" : 65539,
>>>>>>   "left" : 65541,
>>>>>>   "right" : 65540,
>>>>>>   "conditions" : [ {
>>>>>>     "relationship" : "EQUALS",
>>>>>>     "left" : "`cvalue`",
>>>>>>     "right" : "`tag_value`"
>>>>>>   } ],
>>>>>>   "joinType" : "INNER",
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 4.1667342E7
>>>>>> }, {
>>>>>>   "pop" : "project",
>>>>>>   "@id" : 65538,
>>>>>>   "exprs" : [ {
>>>>>>     "ref" : "`$f0`",
>>>>>>     "expr" : "0"
>>>>>>   } ],
>>>>>>   "child" : 65539,
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 4.1667342E7
>>>>>> }, {
>>>>>>   "pop" : "streaming-aggregate",
>>>>>>   "@id" : 65537,
>>>>>>   "child" : 65538,
>>>>>>   "keys" : [ ],
>>>>>>   "exprs" : [ {
>>>>>>     "ref" : "`EXPR$0`",
>>>>>>     "expr" : "count(1) "
>>>>>>   } ],
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 1.0
>>>>>> }, {
>>>>>>   "pop" : "union-exchange",
>>>>>>   "@id" : 2,
>>>>>>   "child" : 65537,
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 1.0
>>>>>> }, {
>>>>>>   "pop" : "streaming-aggregate",
>>>>>>   "@id" : 1,
>>>>>>   "child" : 2,
>>>>>>   "keys" : [ ],
>>>>>>   "exprs" : [ {
>>>>>>     "ref" : "`EXPR$0`",
>>>>>>     "expr" : "$sum0(`EXPR$0`) "
>>>>>>   } ],
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 1.0
>>>>>> }, {
>>>>>>   "pop" : "screen",
>>>>>>   "@id" : 0,
>>>>>>   "child" : 1,
>>>>>>   "initialAllocation" : 1000000,
>>>>>>   "maxAllocation" : 10000000000,
>>>>>>   "cost" : 1.0
>>>>>> } ]
>>>>>> } |
>>>>>> +------+------+
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> On Mon, Jan 4, 2016 at 9:42 PM, Andries Engelbrecht <
>>>>> [email protected]> wrote:
>>>>>>> Perhaps check the data type of all the fields being used for the
>>> join.
>>>>>>> 
>>>>>>> Select cvalue, TYPEOF(cvalue) from hdfs...... limit 10
>>>>>>> 
>>>>>>> and similar for  tag_value on redshift.
>>>>>>> 
>>>>>>> You can then do a predicate to find records where the data type
>> may
>>> be
>>>>> different.
>>>>>>> where typeof(<field>) not like '<data type of field>'
>>>>>>> 
>>>>>>> I believe there was a nice write up on they topic, but can't find
>> it
>>>>> now.
>>>>>>> 
>>>>>>> 
>>>>>>> --Andries
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>>> On Jan 3, 2016, at 8:45 PM, Rohit Kulkarni <
>>>> [email protected]>
>>>>> wrote:
>>>>>>>> 
>>>>>>>> Hello all,
>>>>>>>> 
>>>>>>>> I am sure if not all of you, but some of you must have seen this
>>>>> error some
>>>>>>>> time -
>>>>>>>> 
>>>>>>>> *Error: SYSTEM ERROR: IllegalStateException: Already had POJO
>> for
>>> id
>>>>>>>> (java.lang.Integer)
>> [com.fasterxml.jackson.annotation.ObjectIdGenerator$IdKey@3372bbe8
>>>> ]*
>>>>>>>> 
>>>>>>>> ​I am trying to do a join between Redshift (JDBC​) and HDFS like
>>>> this
>>>>> -
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> *select count(*)from hdfs.drill.TAGS_US as aright join
>>>>>>>> redshift.reports.public.us_tags as bon a.cvalue = b.tag_value;*
>>>>>>>> 
>>>>>>>> 
>>>>>>>> I don't see anything wrong in the query. The two individual
>> tables
>>>>> return
>>>>>>>> proper data when fired a query separately. Is something missing
>> or
>>>> am
>>>>> I
>>>>>>>> doing something wrong?
>>>>>>>> 
>>>>>>>> Would very much appreciate your help! Thanks!!
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Warm Regards,
>>>>>>>> Rohit Kulkarni
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> --
>>>>>> Warm Regards,
>>>>>> Rohit Kulkarni
>>>>>> Mo.: +91 89394 63593
>>> 
>>> 
>>> 
>>> --
>>> Warm Regards,
>>> Rohit Kulkarni
>>> Mo.: +91 89394 63593
>> 
>> 
>> 
>> --
>> Warm Regards,
>> Rohit Kulkarni
>> Mo.: +91 89394 63593
>> 

Reply via email to