Hi Tim,

I think you can try setting the option *spark.sql.files.ignoreCorruptFiles *as
*true*. With the option enabled, the Spark jobs will continue to run
when encountering corrupted files and the contents that have been read will
still be returned.
The CSV/JSON data source supports the Permissive modes in reading files
because it is possible that users still want partial row results.
When reading corrupted Avro files, I think skipping the rest of files is
enough if users want to ignore them.
For processing data with function `from_avro`, I have created a PR to
support  *PERMISSIVE*/*FAILFAST* mode:
https://github.com/apache/spark/pull/22814

Gengliang


On Fri, Mar 8, 2019 at 6:25 AM tim <t...@gh.st> wrote:

> /facepalm
>
> Here we go: https://issues.apache.org/jira/browse/SPARK-27093
>
> Tim
>
>
>
> --
> Sent from: http://apache-spark-developers-list.1001551.n3.nabble.com/
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to