Strange: the JIRAs below were marked Fixed in 1.5.0 On Fri, Mar 25, 2016 at 3:43 PM, Mich Talebzadeh <mich.talebza...@gmail.com> wrote:
> Is this 1.6 Ted? > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > > On 25 March 2016 at 22:40, Ted Yu <yuzhih...@gmail.com> wrote: > >> Looks like database support was fixed by: >> >> [SPARK-7943] [SPARK-8105] [SPARK-8435] [SPARK-8714] [SPARK-8561] Fixes >> multi-database support >> >> On Fri, Mar 25, 2016 at 3:35 PM, Ashok Kumar <ashok34...@yahoo.com> >> wrote: >> >>> 1.5.2 Ted. >>> >>> Those two lines I don't know where they come. It finds and gets the >>> table info OK >>> >>> HTH >>> >>> >>> On Friday, 25 March 2016, 22:32, Ted Yu <yuzhih...@gmail.com> wrote: >>> >>> >>> Which release of Spark do you use, Mich ? >>> >>> In master branch, the message is more accurate >>> (sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/NoSuchItemException.scala): >>> >>> override def getMessage: String = s"Table $table not found in database >>> $db" >>> >>> >>> On Fri, Mar 25, 2016 at 3:21 PM, Mich Talebzadeh < >>> mich.talebza...@gmail.com> wrote: >>> >>> You can use DESCRIBE FORMATTED <DATABASE>.<TABLE_NAME> to get that info. >>> >>> This is based on the same command in Hive however, it throws two >>> erroneous error lines as shown below (don't see them in Hive DESCRIBE ...) >>> >>> Example >>> >>> scala> sql("describe formatted test.t14").collect.foreach(println) >>> 16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not >>> found >>> 16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not >>> found >>> [# col_name data_type comment ] >>> [ ] >>> [invoicenumber int ] >>> [paymentdate date ] >>> [net decimal(20,2) ] >>> [vat decimal(20,2) ] >>> [total decimal(20,2) ] >>> [ ] >>> [# Detailed Table Information ] >>> [Database: test ] >>> [Owner: hduser ] >>> [ >>> *CreateTime: Fri Mar 25 22:13:44 GMT 2016 ]*[LastAccessTime: >>> UNKNOWN ] >>> [Protect Mode: None ] >>> [Retention: 0 ] >>> [Location: >>> hdfs://rhes564:9000/user/hive/warehouse/test.db/t14 ] >>> [Table Type: MANAGED_TABLE ] >>> [Table Parameters: ] >>> [ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}] >>> [ comment from csv file from excel sheet] >>> [ numFiles 2 ] >>> [ orc.compress ZLIB ] >>> [ totalSize 1090 ] >>> [ transient_lastDdlTime 1458944025 ] >>> [ ] >>> [# Storage Information ] >>> [SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde >>> ] >>> [InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat >>> ] >>> [OutputFormat: >>> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat ] >>> [Compressed: No ] >>> [Num Buckets: -1 ] >>> [Bucket Columns: [] ] >>> [Sort Columns: [] ] >>> [Storage Desc Params: ] >>> [ serialization.format 1 ] >>> >>> HTH >>> >>> Dr Mich Talebzadeh >>> >>> LinkedIn * >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>> >>> http://talebzadehmich.wordpress.com >>> >>> >>> On 25 March 2016 at 22:12, Ashok Kumar <ashok34...@yahoo.com.invalid> >>> wrote: >>> >>> Experts, >>> >>> I would like to know when a table was created in Hive database using >>> Spark shell? >>> >>> Thanks >>> >>> >>> >>> >>> >>> >> >