zero323 edited a comment on pull request #35296: URL: https://github.com/apache/spark/pull/35296#issuecomment-1019586587
Thank you for your proposal @bjornjorgensen. First of all, some formalities: - Could you [enable GitHub actions](https://github.com/apache/spark/pull/35296/checks?check_run_id=4914852660) in your fork, - It would be better to have JIRA ticket that specifically targets documentation update, as the one you linked doesn't even use `pyspark.pandas`. Regarding the change ‒ as-is, this doesn't really describe the actual behavior: - If `path` is not provided, `to_json` delegates transformation to pandas, so fields with NULL are preserved. - If `path` is provided, standard writer is used, so the same rules apply (in particular `ignoreNullFields` keyword option or `spark.sql.jsonGenerator.ignoreNullFields` config are respected). - In both cases, behavior doesn't depend on the number of null values in the column. Finally, you have a typo ‒ "The column **well** be deleted" -> "The column will be deleted." -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
