Kent Yao created SPARK-22463:
--------------------------------

             Summary: Missing hadoop/hive/hbase/etc configuration files in 
SPARK_CONF_DIR to distributed archive
                 Key: SPARK-22463
                 URL: https://issues.apache.org/jira/browse/SPARK-22463
             Project: Spark
          Issue Type: Bug
          Components: YARN
    Affects Versions: 2.2.0, 2.1.2
            Reporter: Kent Yao


When I ran self contained sql apps, such as
{code:java}
import org.apache.spark.sql.SparkSession

object ShowHiveTables {
  def main(args: Array[String]): Unit = {
    val spark = SparkSession
      .builder()
      .appName("Show Hive Tables")
      .enableHiveSupport()
      .getOrCreate()
    spark.sql("show tables").show()
    spark.stop()
  }
}
{code}
with **yarn cluster** mode and `hive-site.xml` correctly within 
`$SPARK_HOME/conf`,they failed to connect the right hive metestore for not 
seeing hive-site.xml in AM/Driver's classpath.

Although submitting them with `--files/--jars local/path/to/hive-site.xml` or 
puting it to `$HADOOP_CONF_DIR/YARN_CONF_DIR` can make these apps works well in 
cluster mode as client mode, according to the official doc, see @ 
http://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables
> Configuration of Hive is done by placing your hive-site.xml, core-site.xml 
> (for security configuration), and hdfs-site.xml (for HDFS configuration) file 
> in conf/.

We may respect these configuration files too or modify the doc for hive-tables 
in cluster mode.





--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to