Unfortunately this opens a lot more questions and problems than it solves. What if you take something off the classpath, for example? change a class?
On Wed, Mar 9, 2022 at 8:22 AM Rafał Wojdyła <ravwojd...@gmail.com> wrote: > Thanks Sean, > To be clear, if you prefer to change the label on this issue from bug to > sth else, feel free to do so, no strong opinions on my end. What happens to > the classpath, whether spark uses some classloader magic, is probably an > implementation detail. That said, it's definitely not intuitive that you > can change the configuration and get the context (with the updated config) > without any warnings/errors. Also what would you recommend as a workaround > or solution to this problem? Any comments about the workaround in the > issue? Keep in mind that I can't restart the long running orchestration > process (python process if that matters). > Cheers - Rafal > > On Wed, 9 Mar 2022 at 13:15, Sean Owen <sro...@gmail.com> wrote: > >> That isn't a bug - you can't change the classpath once the JVM is >> executing. >> >> On Wed, Mar 9, 2022 at 7:11 AM Rafał Wojdyła <ravwojd...@gmail.com> >> wrote: >> >>> Hi, >>> My use case is that, I have a long running process (orchestrator) with >>> multiple tasks, some tasks might require extra spark dependencies. It seems >>> once the spark context is started it's not possible to update >>> `spark.jars.packages`? I have reported an issue at >>> https://issues.apache.org/jira/browse/SPARK-38438, together with a >>> workaround ("hard reset of the cluster"). I wonder if anyone has a solution >>> for this? >>> Cheers - Rafal >>> >>