Thanks -- it does appear that I misdiagnosed a bit: case works generally
but it doesn't seem to like the bit operation, which does not seem to work
(type of bit_field in Hive is bigint):

Error: java.lang.RuntimeException:
Unsupported language features in query: select (case when bit_field &
1=1 then r_end - r_start else NULL end) from mytable where
pkey='0178-2014-07' LIMIT 2
TOK_QUERY
  TOK_FROM
    TOK_TABREF
      TOK_TABNAME
       mytable
  TOK_INSERT
    TOK_DESTINATION
      TOK_DIR
        TOK_TMP_FILE
    TOK_SELECT
      TOK_SELEXPR
        TOK_FUNCTION
          when
          =
            &
              TOK_TABLE_OR_COL
                bit_field
              1
            1
          -
            TOK_TABLE_OR_COL
              r_end
            TOK_TABLE_OR_COL
              r_start
          TOK_NULL
    TOK_WHERE
      =
        TOK_TABLE_OR_COL
          pkey
        '0178-2014-07'
    TOK_LIMIT
      2


SQLState:  null
ErrorCode: 0

​

similarly, concat seems to work but I get a failure in this query (due to
LPAD I believe) :

select customer_id from mytable where
pkey=concat_ws('-',LPAD('077',4,'0'),'2014-07') LIMIT 2

(there is something going on with the fact that the function is in the
where clause....the following work fine:

select concat_ws('-', LPAD(cast(112717 % 1024 AS STRING),4,'0'),'2014-07')
from mytable where pkey='0077-2014-07' LIMIT 2 select customer_id from
mytable where pkey=concat_ws('-','0077','2014-07') LIMIT 2
​
)

14/10/03 14:51:35 ERROR server.SparkSQLOperationManager: Error executing query:
org.apache.spark.SparkException: Task not serializable
        at 
org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:166)
        at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:158)
        at org.apache.spark.SparkContext.clean(SparkContext.scala:1242)
        at org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:597)
        at 
org.apache.spark.sql.execution.Limit.execute(basicOperators.scala:146)
        at 
org.apache.spark.sql.hive.HiveContext$QueryExecution.toRdd$lzycompute(HiveContext.scala:360)
        at 
org.apache.spark.sql.hive.HiveContext$QueryExecution.toRdd(HiveContext.scala:360)
        at 
org.apache.spark.sql.hive.thriftserver.server.SparkSQLOperationManager$$anon$1.run(SparkSQLOperationManager.scala:185)
        at 
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:193)
        at 
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:175)
        at 
org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:150)
        at 
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:207)
        at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
        at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
        at 
org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:58)
        at 
org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:55)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
        at 
org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526)
        at 
org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:55)
        at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:744)
Caused by: java.io.NotSerializableException: java.lang.reflect.Constructor
        at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1183)
        at 
java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
        at 
java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
        at 
java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
        at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
        at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1377)
        at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1173)
        at 
java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
        at 
java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
        at 
java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
        at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
        at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347)
        at scala.collection.immutable.$colon$colon.writeObject(List.scala:379)
        at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

​


Let me know if any of these warrant a JIRA

thanks




On Thu, Oct 2, 2014 at 2:00 PM, Michael Armbrust <mich...@databricks.com>
wrote:

> What are the errors you are seeing.  All of those functions should work.
>
> On Thu, Oct 2, 2014 at 6:56 AM, Yana Kadiyska <yana.kadiy...@gmail.com>
> wrote:
>
>> Hi, in an effort to migrate off of Shark I recently tried the Thrift JDBC
>> server that comes with Spark 1.1.0.
>>
>> However I observed that conditional functions do not work (I tried 'case'
>> and 'coalesce')
>>
>> some string functions like 'concat' also did not work.
>>
>> Is there a list of what's missing or a roadmap of when it will be added?
>> (I know percentiles are pending, for example but do not see JIRAs for the
>> others in this email).
>>
>
>

Reply via email to