hi all; --driver-java-options not support multiple JVM configuration.
the submot as following: Cores=16 sparkdriverextraJavaOptions="-XX:newsize=2096m -XX:MaxPermSize=512m -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseP arNewGC -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=80 -XX:GCTimeLimit=5 -XX:GCHeapFreeLimit=95" main1=com.suning.spark.streaming.ppsc.RecommendBasedShoppingCart spark-submit --deploy-mode cluster \ --total-executor-cores $Cores \ --executor-memory 8g \ --driver-memory 16g \ --conf spark.driver.cores=4 \ --driver-java-options $sparkdriverextraJavaOptions \ --class $main1 \ hdfs:///user/bdapp/$appjars error : Error: Unrecognized option '-XX:MaxPermSize=512m ; when change to : sparkdriverextraJavaOptions="-XX:newsize=2096m,-XX:MaxPermSize=512m,-XX:+PrintGCDetails,-XX:+PrintGCTimeStamps,-XX:+UseParNewGC,-XX:+UseConcMarkSweepGC,-XX:CMSInitiatingOccupancyFraction=80,-XX:GCTimeLimit=5,-XX:GCHeapFreeLimit=95" the driver errors is : Unrecognized VM option 'newsize=2096m,-XX:MaxPermSize=512m,-XX:+PrintGCDetails,-XX:+PrintGCTimeStamps,-XX:+UseParNewGC,-XX:+UseConcMarkSweepGC,-XX:CMSInitiatingOccupancyFraction=80,-XX:GCTimeLimit=5,-XX:GCHeapFreeLimit=95