Hi, I met the same problem before, and according to Matei Zaharia: /The issue is that you're using SQLContext instead of HiveContext. SQLContext implements a smaller subset of the SQL language and so you're getting a SQL parse error because it doesn't support the syntax you have. Look at how you'd write this in HiveQL, and then try doing that with HiveContext./
In fact, there are more problems than that. The sparkSQL will conserve (15+5=20) columns in the final table, if I remember well. Therefore, when you are doing join on two tables which have the same columns will cause doublecolumn error. Cheers Gen Hao Ren wrote > Update: > > This syntax is mainly for avoiding retyping column names. > > Let's take the example in my previous post, where * > a * > is a table of 15 columns, * > b * > has 5 columns, after a join, I have a table of (15 + 5 - 1(key in b)) = > 19 columns and register the table in sqlContext. > > I don't want to actually retype all the 19 columns' name when querying > with select. This feature exists in hive. > But in SparkSql, it gives an exception. > > Any ideas ? Thx > > Hao -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/SparkSQL-select-syntax-tp16299p16367.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org