[ https://issues.apache.org/jira/browse/SPARK-16224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Davies Liu resolved SPARK-16224. -------------------------------- Resolution: Fixed Fix Version/s: 2.0.1 Issue resolved by pull request 13931 [https://github.com/apache/spark/pull/13931] > Hive context created by HiveContext can't access Hive databases when used in > a script launched be spark-submit > -------------------------------------------------------------------------------------------------------------- > > Key: SPARK-16224 > URL: https://issues.apache.org/jira/browse/SPARK-16224 > Project: Spark > Issue Type: Bug > Components: PySpark > Affects Versions: 2.0.0 > Environment: branch-2.0 > Reporter: Piotr Milanowski > Assignee: Yin Huai > Priority: Blocker > Fix For: 2.0.1 > > > Hi, > This is a continuation of a resolved bug > [SPARK-15345|https://issues.apache.org/jira/browse/SPARK-15345] > I can access databases when using new methodology, i.e: > {code} > from pyspark.sql import SparkSession > from pyspark import SparkConf > if __name__ == "__main__": > conf = SparkConf() > hc = > SparkSession.builder.config(conf=conf).enableHiveSupport().getOrCreate() > print(hc.sql("show databases").collect()) > {code} > This shows all database in hive. > However, using HiveContext, i.e.: > {code} > from pyspark.sql import HiveContext > from pyspark import SparkContext, SparkConf > if __name__ == "__main__": > conf = SparkConf() > sc = SparkContext(conf=conf) > hive_context = HiveContext(sc) > print(hive_context.sql("show databases").collect()) > # The result is > #[Row(result='default')] > {code} > prints only default database. > I have {{hive-site.xml}} file configured. > Those snippets are for scripts launched with {{spark-submit}} command. With > pyspark those code fragments work fine, displaying all the databases. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org