Hey all,

I'm evaluating using Spark Streaming with Kafka direct streaming, and I
have a couple of questions:

1.  Would it be possible to add / remove worker nodes without stopping and
restarting the spark streaming driver?

2.  I understand that we can enable checkpointing to recover from node
failures, and that it doesn't work across code changes.  What about in the
event that worker nodes failed due to load -> we added more worker nodes ->
restart Spark Streaming?  Would this incur data loss as well?


Best,
Augustus

-- 
[image: Branch Metrics mobile deep linking] <http://branch.io/>* Augustus
Hong*
 Data Analytics | Branch Metrics
 m 650-391-3369 | e augus...@branch.io

Reply via email to