I agree with the CatalystConf idea.---- On Fri, 14 Dec 2018
18:40:26 +0800 Wenchen Fan<cloud0...@gmail.com> wrote ----IIRC, the reason we
did it is: `SQLConf` was in SQLÂ core module. So we need to create methods in
`CatalystConf`, and `SQLConf` implements `CatalystConf`.Now the problem has
gone: we moved `SQLConf` to catalyst module. I think we can remove these
methods.On Fri, Dec 14, 2018 at 3:45 PM Reynold Xin <r...@databricks.com>
wrote:In SQLConf, for each config option, we declare them in two places:First
in the SQLConf object, e.g.:val CSV_PARSER_COLUMN_PRUNING =
buildConf("spark.sql.csv.parser.columnPruning.enabled") .internal() .doc("If
it is set to true, column names of the requested schema are passed to CSV
parser. " + "Other column values can be ignored during parsing even if they
are malformed.") .booleanConf .createWithDefault(true)Second in SQLConf
class:def csvColumnPruning: Boolean =
getConf(SQLConf.CSV_PARSER_COLUMN_PRUNING)As the person that introduced both,
I'm now thinking we should remove almost all of the latter, unless it is used
more than 5 times. The vast majority of config options are read only in one
place, so the functions are pretty redundant ...