Hey Yin, Thanks for answer. I thought that this could be problem but i can not create HiveContext because i can not import org.apache.spark.sql.hive.HiveContext. It does not see this package.
I read that i should build spark with -PHive but i’m running on Amazon EMR 1.4.1 and on spark-shell i can import hive package but can not do the same on spark-submit. Do you have any idea why? Because if it’s related to build with -PHive, how can i import it in spark-shell? > On 19 Aug 2015, at 18:59, Yin Huai <yh...@databricks.com> wrote: > > Can you try to use HiveContext instead of SQLContext? Your query is trying to > create a table and persist the metadata of the table in metastore, which is > only supported by HiveContext. > > On Wed, Aug 19, 2015 at 8:44 AM, Yusuf Can Gürkan <yu...@useinsider.com > <mailto:yu...@useinsider.com>> wrote: > Hello, > > I’m trying to create a table with sqlContext.sql method as below: > > val sc = new SparkContext() > val sqlContext = new SQLContext(sc) > > import sqlContext.implicits._ > > sqlContext.sql(s""" > create table if not exists landing ( > date string, > referrer string > ) > partitioned by (partnerid string,dt string) > row format delimited fields terminated by '\t' lines terminated by '\n' > STORED AS TEXTFILE LOCATION 's3n://...' <> > "”") > > > It gives error on spark-submit: > > Exception in thread "main" java.lang.RuntimeException: [2.1] failure: > ``with'' expected but identifier create found > > create external table if not exists landing ( > > ^ > at scala.sys.package$.error(package.scala:27) > at > org.apache.spark.sql.catalyst.AbstractSparkSQLParser.parse(AbstractSparkSQLParser.scala:36) > at > org.apache.spark.sql.catalyst.DefaultParserDialect.parse(ParserDialect.scala:67) > > > > What can be the reason?? >