Hello, I'm developing a (java) DataSourceV2 to read a columnar fileformat popular in a number of physical sciences (https://root.cern.ch/). (I also understand that the API isn't fixed and subject to change).
My question is -- what is the expected way to transmit exceptions from the DataSource up to Spark? The DSV2 interface (unless I'm misreading it) doesn't specify any caught exceptions that can be thrown in the DS, so should I instead catch/rethrow any exceptions as uncaught exceptions? If so, is there a recommended hierarchy to throw from? thanks! Andrew --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org