Github user leifker commented on the issue:
https://github.com/apache/spark/pull/17306
Closing this PR, will get back to this eventually, but dealing with some
other priorities at the moment.
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user leifker commented on the issue:
https://github.com/apache/spark/pull/17306
Interesting, let me think about this a bit. I think that there is probably
a better api around this approach for sure.
---
If your project is set up for it, you can reply to this email and have you
Github user BryanCutler commented on the issue:
https://github.com/apache/spark/pull/17306
Yeah, that's true in this case. You could just build the grids separately
and combine them like this
```scala
val pipeline1_grid = new ParamGridBuilder()
.baseOn(pipel
Github user leifker commented on the issue:
https://github.com/apache/spark/pull/17306
Sorry of the delayed response @BryanCutler, that's pretty neat, however
this will perform unneeded work as it will execute nonsensical combinations of
parameters. For example, if pipeline2 is execut
Github user BryanCutler commented on the issue:
https://github.com/apache/spark/pull/17306
Thanks @leifker for the PR, this is a good idea. I think though it can
already be accomplished with the current param grid builder. Since the stages
of a pipeline are actually a param, you can
Github user MLnick commented on the issue:
https://github.com/apache/spark/pull/17306
I commented on the linked JIRA also.
In principle I think this can be a useful enhancement and yes the better
efficiency on the caching side is a good benefit. I'd actually been thinking
abo