Re: SparkSession via HS2 - Error -spark.yarn.jars not read
STS will refer spark-thrift-sparkconf.conf, Can you check if the spark.yarn.jars exists in this file? On Wed, Jul 5, 2017 at 2:01 PM, Sudha KS wrote: > The property “spark.yarn.jars” available via /usr/hdp/current/spark2- > client/conf/spark-default.conf > > > > spark.yarn.jars hdfs://ambari03.fuzzyl.com:8020/hdp/apps/2.6.1.0-129/ > spark2 > > > > > > Is there any other way to set/read/pass this property “spark.yarn.jars” ? > > > > *From:* Sudha KS [mailto:sudha...@fuzzylogix.com] > *Sent:* Wednesday, July 5, 2017 1:51 PM > *To:* user@spark.apache.org > *Subject:* SparkSession via HS2 - Error -spark.yarn.jars not read > > > > Why does “spark.yarn.jars” property not read, in this HDP 2.6 , Spark2.1.1 > cluster: > > 0: jdbc:hive2://localhost:1/db> set spark.yarn.jars; > > +--- > ---+--+ > > | set > | > > +--- > ---+--+ > > | spark.yarn.jars=hdfs://ambari03.fuzzyl.com:8020/hdp/apps/2. > 6.1.0-129/spark2 | > > +--- > ---+--+ > > 1 row selected (0.101 seconds) > > 0: jdbc:hive2://localhost:1/db> > > > > > > > > Error during launch of a SparkSession via HS2: > > Caused by: java.lang.IllegalStateException: Library directory > '/hadoop/yarn/local/usercache/hive/appcache/application_ > 1499235958765_0042/container_e04_1499235958765_0042_01_ > 05/assembly/target/scala-2.11/jars' does not exist; make sure Spark > is built. > > at org.apache.spark.launcher.CommandBuilderUtils.checkState( > CommandBuilderUtils.java:260) > > at org.apache.spark.launcher.CommandBuilderUtils.findJarsDir( > CommandBuilderUtils.java:380) > > at org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir( > YarnCommandBuilderUtils.scala:38) > > at org.apache.spark.deploy.yarn.Client.prepareLocalResources( > Client.scala:570) > > at org.apache.spark.deploy.yarn.Client. > createContainerLaunchContext(Client.scala:895) > > at org.apache.spark.deploy.yarn.Client.submitApplication( > Client.scala:171) > > at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend. > start(YarnClientSchedulerBackend.scala:56) > > at org.apache.spark.scheduler.TaskSchedulerImpl.start( > TaskSchedulerImpl.scala:156) > > at org.apache.spark.SparkContext.(SparkContext.scala:509) > > at org.apache.spark.SparkContext$.getOrCreate(SparkContext. > scala:2320) > > at org.apache.spark.sql.SparkSession$Builder$$anonfun$ > 6.apply(SparkSession.scala:868) > > at org.apache.spark.sql.SparkSession$Builder$$anonfun$ > 6.apply(SparkSession.scala:860) > > at scala.Option.getOrElse(Option.scala:121) > > at org.apache.spark.sql.SparkSession$Builder. > getOrCreate(SparkSession.scala:860) > > at SparkHiveUDTF.sparkJob(SparkHiveUDTF.java:97) > > at SparkHiveUDTF.process(SparkHiveUDTF.java:78) > > at org.apache.hadoop.hive.ql.exec.UDTFOperator.process( > UDTFOperator.java:109) > > at org.apache.hadoop.hive.ql.exec.Operator.forward( > Operator.java:841) > > at org.apache.hadoop.hive.ql.exec.SelectOperator.process( > SelectOperator.java:88) > > at org.apache.hadoop.hive.ql.exec.Operator.forward( > Operator.java:841) > > at org.apache.hadoop.hive.ql.exec.TableScanOperator. > process(TableScanOperator.java:133) > > at org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx. > forward(MapOperator.java:170) > > at org.apache.hadoop.hive.ql.exec.MapOperator.process( > MapOperator.java:555) > > ... 18 more > > > > > > > > > -- * Regards* * Sandeep Nemuri*
RE: SparkSession via HS2 - Error -spark.yarn.jars not read
The property "spark.yarn.jars" available via /usr/hdp/current/spark2-client/conf/spark-default.conf spark.yarn.jars hdfs://ambari03.fuzzyl.com:8020/hdp/apps/2.6.1.0-129/spark2 Is there any other way to set/read/pass this property "spark.yarn.jars" ? From: Sudha KS [mailto:sudha...@fuzzylogix.com] Sent: Wednesday, July 5, 2017 1:51 PM To: user@spark.apache.org Subject: SparkSession via HS2 - Error -spark.yarn.jars not read Why does "spark.yarn.jars" property not read, in this HDP 2.6 , Spark2.1.1 cluster: 0: jdbc:hive2://localhost:1/db> set spark.yarn.jars; +--+--+ | set | +--+--+ | spark.yarn.jars=hdfs://ambari03.fuzzyl.com:8020/hdp/apps/2.6.1.0-129/spark2 | +--+--+ 1 row selected (0.101 seconds) 0: jdbc:hive2://localhost:1/db> Error during launch of a SparkSession via HS2: Caused by: java.lang.IllegalStateException: Library directory '/hadoop/yarn/local/usercache/hive/appcache/application_1499235958765_0042/container_e04_1499235958765_0042_01_05/assembly/target/scala-2.11/jars' does not exist; make sure Spark is built. at org.apache.spark.launcher.CommandBuilderUtils.checkState(CommandBuilderUtils.java:260) at org.apache.spark.launcher.CommandBuilderUtils.findJarsDir(CommandBuilderUtils.java:380) at org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir(YarnCommandBuilderUtils.scala:38) at org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:570) at org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:895) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:171) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:56) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:156) at org.apache.spark.SparkContext.(SparkContext.scala:509) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2320) at org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:868) at org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:860) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:860) at SparkHiveUDTF.sparkJob(SparkHiveUDTF.java:97) at SparkHiveUDTF.process(SparkHiveUDTF.java:78) at org.apache.hadoop.hive.ql.exec.UDTFOperator.process(UDTFOperator.java:109) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:841) at org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:841) at org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:133) at org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:170) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:555) ... 18 more
SparkSession via HS2 - Error -spark.yarn.jars not read
Why does "spark.yarn.jars" property not read, in this HDP 2.6 , Spark2.1.1 cluster: 0: jdbc:hive2://localhost:1/db> set spark.yarn.jars; +--+--+ | set | +--+--+ | spark.yarn.jars=hdfs://ambari03.fuzzyl.com:8020/hdp/apps/2.6.1.0-129/spark2 | +--+--+ 1 row selected (0.101 seconds) 0: jdbc:hive2://localhost:1/db> Error during launch of a SparkSession via HS2: Caused by: java.lang.IllegalStateException: Library directory '/hadoop/yarn/local/usercache/hive/appcache/application_1499235958765_0042/container_e04_1499235958765_0042_01_05/assembly/target/scala-2.11/jars' does not exist; make sure Spark is built. at org.apache.spark.launcher.CommandBuilderUtils.checkState(CommandBuilderUtils.java:260) at org.apache.spark.launcher.CommandBuilderUtils.findJarsDir(CommandBuilderUtils.java:380) at org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir(YarnCommandBuilderUtils.scala:38) at org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:570) at org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:895) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:171) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:56) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:156) at org.apache.spark.SparkContext.(SparkContext.scala:509) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2320) at org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:868) at org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:860) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:860) at SparkHiveUDTF.sparkJob(SparkHiveUDTF.java:97) at SparkHiveUDTF.process(SparkHiveUDTF.java:78) at org.apache.hadoop.hive.ql.exec.UDTFOperator.process(UDTFOperator.java:109) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:841) at org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:841) at org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:133) at org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:170) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:555) ... 18 more