Looks like currently solution to update spark-stream jars/configurations is
to

1) save current Kafka offsets somewhere (say zookeeper)
2) shutdown the cluster and restart
3) connect to Kafka with previously saved offset

Assuming we're reading from Kafka which provides nice persistence and
read-from-offset feature.

Is there a better solution? Is it possible to reload jars dynamically?
Resize cluster size dynamically (SPARK-3174)?

A reasonable ask is to have a strategy to update or restart executors
gracefully. Something like https://issues.apache.org/jira/browse/STORM-167

Any suggestion?

-- 
Jianshi Huang

LinkedIn: jianshi
Twitter: @jshuang
Github & Blog: http://huangjs.github.com/

Reply via email to