Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22237#discussion_r224069683
  
    --- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/FailureSafeParser.scala
 ---
    @@ -15,50 +15,57 @@
      * limitations under the License.
      */
     
    -package org.apache.spark.sql.execution.datasources
    +package org.apache.spark.sql.catalyst.util
     
     import org.apache.spark.SparkException
     import org.apache.spark.sql.catalyst.InternalRow
     import org.apache.spark.sql.catalyst.expressions.GenericInternalRow
    -import org.apache.spark.sql.catalyst.util._
    -import org.apache.spark.sql.internal.SQLConf
    -import org.apache.spark.sql.types.StructType
    +import org.apache.spark.sql.types.{DataType, StructType}
     import org.apache.spark.unsafe.types.UTF8String
     
     class FailureSafeParser[IN](
    --- End diff --
    
    It's not about the real business logic, but an idea to make the code 
simpler.
    ```
    val parserSchema = nullableSchema match {
      case s: StructType => s
      case other => new StructType("value", other)
    }
    new FailureSafeParser[UTF8String](
          input => rawParser.parse(input, createParser, identity[UTF8String]),
          mode,
          parserSchema,
          parsedOptions.columnNameOfCorruptRecord,
          parsedOptions.multiLine)
    ```
    
    Then we don't need to change `FailureSafeParser`


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to