[ 
https://issues.apache.org/jira/browse/SPARK-17330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Owen updated SPARK-17330:
------------------------------
      Assignee: tone
    Issue Type: Improvement  (was: Bug)

> Clean up spark-warehouse in UT
> ------------------------------
>
>                 Key: SPARK-17330
>                 URL: https://issues.apache.org/jira/browse/SPARK-17330
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL, Tests
>    Affects Versions: 2.0.0
>            Reporter: tone
>            Assignee: tone
>            Priority: Minor
>             Fix For: 2.1.0
>
>
> When run Spark UT based on the latest version of master branch, the UT case 
> (SPARK-8368) can be passed at the first time, but always fail if run it 
> again. The error log is as below:
> [info]   2016-08-31 09:35:51.967 - stderr> 16/08/31 09:35:51 ERROR 
> RetryingHMSHandler: AlreadyExistsException(message:Database default already 
> exists)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_database(HiveMetaStore.java:891)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> java.lang.reflect.Method.invoke(Method.java:498)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> com.sun.proxy.$Proxy18.create_database(Unknown Source)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:644)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> java.lang.reflect.Method.invoke(Method.java:498)
> [info]   2016-08-31 09:35:51.967 - stderr>      at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:156)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> com.sun.proxy.$Proxy19.createDatabase(Unknown Source)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.hadoop.hive.ql.metadata.Hive.createDatabase(Hive.java:306)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply$mcV$sp(HiveClientImpl.scala:310)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply(HiveClientImpl.scala:310)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply(HiveClientImpl.scala:310)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:281)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:228)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:227)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:270)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.client.HiveClientImpl.createDatabase(HiveClientImpl.scala:309)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply$mcV$sp(HiveExternalCatalog.scala:120)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply(HiveExternalCatalog.scala:120)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply(HiveExternalCatalog.scala:120)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:87)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveExternalCatalog.createDatabase(HiveExternalCatalog.scala:119)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.catalyst.catalog.SessionCatalog.createDatabase(SessionCatalog.scala:147)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.catalyst.catalog.SessionCatalog.<init>(SessionCatalog.scala:89)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionCatalog.<init>(HiveSessionCatalog.scala:49)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:46)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:45)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:59)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:59)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:58)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:61)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.SparkSession.createDataFrame(SparkSession.scala:261)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.SQLContext.createDataFrame(SQLContext.scala:290)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.SparkSubmitClassLoaderTest$.main(HiveSparkSubmitSuite.scala:596)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.sql.hive.SparkSubmitClassLoaderTest.main(HiveSparkSubmitSuite.scala)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> java.lang.reflect.Method.invoke(Method.java:498)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:729)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124)
> [info]   2016-08-31 09:35:51.968 - stderr>      at 
> org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> From the log, the database file has been existed when run the UT case. Need 
> to fix the issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to