[ 
https://issues.apache.org/jira/browse/SPARK-34583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17296751#comment-17296751
 ] 

Hyukjin Kwon commented on SPARK-34583:
--------------------------------------

cc [~Ngone51] [~cloud_fan] FYI

> typed udf fails when it refers to type member in abstract class
> ---------------------------------------------------------------
>
>                 Key: SPARK-34583
>                 URL: https://issues.apache.org/jira/browse/SPARK-34583
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 3.0.1
>            Reporter: kondziolka9ld
>            Priority: Minor
>
>  Please consider a following scenario: 
> {code:java}
> scala> abstract class SomeAbstractClass {
>      |   type SomeTypeMember
>      | }
> defined class SomeAbstractClassscala> class SomeSpecificClass extends 
> SomeAbstractClass {
>      |   override type SomeTypeMember = Int
>      | }
> defined class SomeSpecificClassscala> def someFunction(someInstance: 
> SomeAbstractClass): Any = {
>      |   udf { _: someInstance.SomeTypeMember => 42 }
>      | }
> someFunction: (someInstance: SomeAbstractClass)Any
> scala> someFunction(new SomeSpecificClass)
> java.lang.NoClassDefFoundError: no Java class corresponding to 
> someInstance.SomeTypeMember found
>   at 
> scala.reflect.runtime.JavaMirrors$JavaMirror.typeToJavaClass(JavaMirrors.scala:1354)
>   at 
> scala.reflect.runtime.JavaMirrors$JavaMirror.runtimeClass(JavaMirrors.scala:227)
>   at 
> scala.reflect.runtime.JavaMirrors$JavaMirror.runtimeClass(JavaMirrors.scala:68)
>   at 
> org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$.apply(ExpressionEncoder.scala:56)
>   at org.apache.spark.sql.functions$.$anonfun$udf$1(functions.scala:4509)
>   at scala.util.Try$.apply(Try.scala:213)
>   at org.apache.spark.sql.functions$.udf(functions.scala:4509)
>   at someFunction(<console>:25)
>   ... 47 elided
> {code}
> On *spark-2.4.7* it works. I guess that it is related to:
> {code:java}
> In Spark 3.0, using org.apache.spark.sql.functions.udf(AnyRef, DataType) is 
> not allowed by default. Remove the return type parameter to automatically 
> switch to typed Scala udf is recommended, or set 
> spark.sql.legacy.allowUntypedScalaUDF to true to keep using it. In Spark 
> version 2.4 and below, if org.apache.spark.sql.functions.udf(AnyRef, 
> DataType) gets a Scala closure with primitive-type argument, the returned UDF 
> returns null if the input values is null. However, in Spark 3.0, the UDF 
> returns the default value of the Java type if the input value is null. For 
> example, val f = udf((x: Int) => x, IntegerType), f($"x") returns null in 
> Spark 2.4 and below if column x is null, and return 0 in Spark 3.0. This 
> behavior change is introduced because Spark 3.0 is built with Scala 2.12 by 
> default.
> {code}
> [https://spark.apache.org/docs/latest/sql-migration-guide.html#udfs-and-built-in-functions]
> Does spark try to do some type inferation? When it refers to 
> `SomeAbstractClass.SomeTypeMember` it really does not exist.
> Some workaround could be runtime type casting, something like: 
> {code:java}
> udf { param: Any => {
>      ...
>      param.asInstanceOf[someInstance.SomeTypeMember]
>      ...  
> } {code}
> ----
>  I classified it as bug since on previous versions of spark it worked. 
> However, I believe that it can work as designed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to