[ https://issues.apache.org/jira/browse/SPARK-11555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Thomas Graves updated SPARK-11555: ---------------------------------- Description: Using the old spark-class and --num-workers interface, --num-workers parameter is ignored and always uses default number of executors (2). bin/spark-class org.apache.spark.deploy.yarn.Client --jar lib/spark-examples-1.5.2.0-hadoop2.6.0.16.1506060127.jar --class org.apache.spark.examples.SparkPi --num-workers 4 --worker-memory 2g --master-memory 1g --worker-cores 1 --queue default was: Calling into yarn with 0 number or workers shouldn't be allowed unless dynamic allocation is enabled. I have a test that does backwards compatibility testing using the old spark-class and --num-workers and passing in 0 is now broken. It allows 0 and shouldn't: bin/spark-class org.apache.spark.deploy.yarn.Client --jar lib/spark-examples-1.5.2.0-hadoop2.6.0.16.1506060127.jar --class org.apache.spark.examples.SparkPi --num-workers 0 --worker-memory 2g --master-memory 1g --worker-cores 1 --queue default > spark on yarn spark-class --num-workers doesn't work > ---------------------------------------------------- > > Key: SPARK-11555 > URL: https://issues.apache.org/jira/browse/SPARK-11555 > Project: Spark > Issue Type: Bug > Components: YARN > Affects Versions: 1.5.2 > Reporter: Thomas Graves > > Using the old spark-class and --num-workers interface, --num-workers > parameter is ignored and always uses default number of executors (2). > bin/spark-class org.apache.spark.deploy.yarn.Client --jar > lib/spark-examples-1.5.2.0-hadoop2.6.0.16.1506060127.jar --class > org.apache.spark.examples.SparkPi --num-workers 4 --worker-memory 2g > --master-memory 1g --worker-cores 1 --queue default -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org