simplylizz commented on a change in pull request #25117: [SPARK-28454][PYTHON] Validate LongType in `createDataFrame(verifySchema=True)` URL: https://github.com/apache/spark/pull/25117#discussion_r311466185
########## File path: docs/sql-migration-guide-upgrade.md ########## @@ -149,6 +149,8 @@ license: | - Since Spark 3.0, if files or subdirectories disappear during recursive directory listing (i.e. they appear in an intermediate listing but then cannot be read or listed during later phases of the recursive directory listing, due to either concurrent file deletions or object store consistency issues) then the listing will fail with an exception unless `spark.sql.files.ignoreMissingFiles` is `true` (default `false`). In previous versions, these missing files or subdirectories would be ignored. Note that this change of behavior only applies during initial table file listing (or during `REFRESH TABLE`), not during query execution: the net change is that `spark.sql.files.ignoreMissingFiles` is now obeyed during table file listing / query planning, not only at query execution time. + - Since Spark 3.0, missing validation of `LongType` value in pyspark's `createDataframe` added. Previously `LongType` was not verified and resulted in `None` in case of wrong value. To have `None` instead of any validation error set `verifySchema=False`. Review comment: Fixed: 1f02d0c ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
