Yes more specifically, you can't ask for executors once the app starts, in SparkConf like that. You set this when you launch it against a Spark cluster in spark-submit or otherwise.
On Tue, Mar 21, 2023 at 4:23 AM Mich Talebzadeh <[email protected]> wrote: > Hi Emmanouil, > > This means that your job is running on the driver as a single JVM, hence > active(1) > >
