Ah, I see,
thanks, Yin
--
Nan Zhu
On Monday, July 21, 2014 at 5:00 PM, Yin Huai wrote:
Hi Nan,
It is basically a log entry because your table does not exist. It is not a
real exception.
Thanks,
Yin
On Mon, Jul 21, 2014 at 7:10 AM, Nan Zhu zhunanmcg...@gmail.com
(mailto:zhunanmcg...@gmail.com) wrote:
a related JIRA: https://issues.apache.org/jira/browse/SPARK-2605
--
Nan Zhu
On Monday, July 21, 2014 at 10:10 AM, Nan Zhu wrote:
Hi, all
When I try hiveContext.hql(drop table if exists abc) where abc is a
non-exist table
I still received an exception about non-exist table though if exists is
there
the same statement runs well in hive shell
Some feedback from Hive community is here:
https://issues.apache.org/jira/browse/HIVE-7458
“Your are doing hiveContext.hql(DROP TABLE IF EXISTS hivetesting) in
Scala schell of the Spark project.
What this shell is doing ? Query to remote metastore on non existing
table (see on your provided stack).
The remote metastore throws
NoSuchObjectException(message:default.hivetesting table not found)because
Spark code call
HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:854) on
non-existing table. It's the right behavior.
You should check on Spark code why a query is done on non existing table.
I think Spark does not handle well the IF EXISTS part of this query.
Maybe you could fill a ticket on Spark JIRA.
BUT, it's not a bug in HIVE IMHO.”
My question is the DDL is executed by Hive itself, doesn’t it?
Best,
--
Nan Zhu