Almaz Murzabekov created SPARK-37539:
----------------------------------------

             Summary: Create Spark interface for creating UDFs
                 Key: SPARK-37539
                 URL: https://issues.apache.org/jira/browse/SPARK-37539
             Project: Spark
          Issue Type: Improvement
          Components: Spark Core
    Affects Versions: 3.2.0
            Reporter: Almaz Murzabekov


Currently, based on [spark 
documentation|https://spark.apache.org/docs/3.2.0/sql-ref-syntax-ddl-create-function.html],
 if you need to create a custom UDF with specific logic implemented in (let's 
say) Scala, you have to extend your class from:
 # UDF or UDAF from org.apache.hadoop.hive.ql.exec package
 # AbstractGenericUDAFResolver, GenericUDF, or GenericUDTF from 
org.apache.hadoop.hive.ql.exec package
 # UserDefinedAggregateFunction in org.apache.spark.sql.expressions (which is 
deprecated)

The third option is deprecated, and (I'm not sure) this option is used only for 
agg user functions. So, for plain UDF we have to follow these first two 
options. That means we MUST add a dependency for the hive-exec. Maybe, we can 
bring spark native interface for creating UDFs.

WDYT guys? 
Thanks in advance



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to