Are you using SQLContext? Right now, the parser in the SQLContext is quite limited on the data type keywords that it handles (see here <https://github.com/apache/spark/blob/master/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala#L391>) and unfortunately "bigint" is not handled in it right now. We will add other data types in there (https://issues.apache.org/jira/browse/SPARK-6146 is used to track it). Can you try HiveContext for now?
On Fri, Mar 13, 2015 at 4:48 AM, Masf <masfwo...@gmail.com> wrote: > Hi. > > I have a query in Spark SQL and I can not covert a value to BIGINT: > CAST(column AS BIGINT) or > CAST(0 AS BIGINT) > > The output is: > Exception in thread "main" java.lang.RuntimeException: [34.62] failure: > ``DECIMAL'' expected but identifier BIGINT found > > Thanks!! > Regards. > Miguel Ángel >