Github user dilipbiswal commented on a diff in the pull request:

    https://github.com/apache/spark/pull/10001#discussion_r46011485
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/sources/interfaces.scala ---
    @@ -606,9 +606,17 @@ abstract class HadoopFsRelation private[sql](
             // we need to cast into the data type that user specified.
             def castPartitionValuesToUserSchema(row: InternalRow) = {
               InternalRow((0 until row.numFields).map { i =>
    -            Cast(
    -              Literal.create(row.getString(i), StringType),
    -              userProvidedSchema.fields(i).dataType).eval()
    +            row.isNullAt(i) match {
    --- End diff --
    
    @davies Thanks for reviewing the change. In cases when we know in advance 
that schema does not allow nulls , sometimes we can skip this null check. By 
moving it to internalRow , would we loose the opportunity to optimize ? Pl. let 
me know.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to