I know some implementations of model save/load in MLlib use an
explicit version 1.0, 2.0, 3.0 mechanism. I've also seen that some
just decide based on the version of Spark that wrote the model.

Is one or the other preferred?

See https://github.com/apache/spark/pull/23549#discussion_r248318392
for example. In cases like this, is it simpler still to just select
all the values written in the model and decide what to do based on the
presence or absence of columns? That seems a little more robust. It
wouldn't be so much an option if the contents or meaning of the
columns had changed.

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to