Hi Vipul,

I am afraid I cannot help you on that.

Thanks,
Marco

Il giorno lun 16 set 2019 alle ore 10:44 Vipul Rajan <vipul.s.p...@gmail.com>
ha scritto:

> Hi Marco,
>
> That does help. Thanks, for taking the time. I am confused as to how that
> Expression is created. There are methods like eval, nullSafeEval,
> doGenCode. Aren't there any architectural docs that could help with what is
> exactly happening? Reverse engineering seems a bit daunting.
>
> Regards
>
> On Mon, Sep 16, 2019 at 1:36 PM Marco Gaido <marcogaid...@gmail.com>
> wrote:
>
>> Hi Vipul,
>>
>> a function is never turned in a logical plan. A function is turned into
>> an Expression. And an Expression can be part of many Logical or Physical
>> Plans.
>> Hope this helps.
>>
>> Thanks,
>> Marco
>>
>> Il giorno lun 16 set 2019 alle ore 08:27 Vipul Rajan <
>> vipul.s.p...@gmail.com> ha scritto:
>>
>>> I am trying to create a function that reads data from Kafka,
>>> communicates with confluent schema registry and decodes avro data with
>>> evolving schemas. I am trying to not create hack-ish patches and to write
>>> proper code that I could maybe even create pull requests for. looking at
>>> the code I have been able to figure out a few things regarding how
>>> expressions are generated and how they help to accomplish what a function
>>> does, but there is still a ton I just cannot wrap my head around.
>>>
>>> I am unable to find any documentation that gets into such nitty gritties
>>> of Spark. *I am writing in hopes to find some help. Do you have any
>>> documentation that explains how a function
>>> (org.apache.spark.sql.function._) is turned into a logical plan?*
>>>
>>

Reply via email to