Github user BryanCutler commented on the issue:

    https://github.com/apache/spark/pull/20124
  
    This basically works by splitting the array of ParamMaps into two.  One 
that has params that can be optimized by the estimator, and one that can be 
parallelized over.  These are then grouped together so that the estimator can 
fit a sequence of Models.  This allows us to reuse the previous API for fitting 
multiple Models and still keep the parallelization logic pretty 
straightforward.  Model specific optimization support is just how it was before 
there was any parallelism introduced too.  I can explain in further detail or 
make a design document if needed.
    
    cc @MLnick @WeichenXu123 @jkbradley 


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to