Hey all, I'm evaluating using Spark Streaming with Kafka direct streaming, and I have a couple of questions:
1. Would it be possible to add / remove worker nodes without stopping and restarting the spark streaming driver? 2. I understand that we can enable checkpointing to recover from node failures, and that it doesn't work across code changes. What about in the event that worker nodes failed due to load -> we added more worker nodes -> restart Spark Streaming? Would this incur data loss as well? Best, Augustus -- [image: Branch Metrics mobile deep linking] <http://branch.io/>* Augustus Hong* Data Analytics | Branch Metrics m 650-391-3369 | e augus...@branch.io