Hello All, I plan to run 2 clustering algorithms on a shared data (Algo A starts first and modifies data and then Algo B starts with the modified data part. Thereafter they run in parallel) in Spark MLLib. Is this possible to share data between two algorithms in a single pipeline?
Regards, Somabha Bhattacharjya