I have created

https://issues.apache.org/jira/browse/SPARK-3814
https://issues.apache.org/jira/browse/SPARK-3815

Will probably try my hand at 3814, seems like a good place to get started...

On Fri, Oct 3, 2014 at 3:06 PM, Michael Armbrust <mich...@databricks.com>
wrote:

> Thanks for digging in!  These both look like they should have JIRAs.
>
> On Fri, Oct 3, 2014 at 8:14 AM, Yana Kadiyska <yana.kadiy...@gmail.com>
> wrote:
>
>> Thanks -- it does appear that I misdiagnosed a bit: case works generally
>> but it doesn't seem to like the bit operation, which does not seem to work
>> (type of bit_field in Hive is bigint):
>>
>> Error: java.lang.RuntimeException:
>> Unsupported language features in query: select (case when bit_field & 1=1 
>> then r_end - r_start else NULL end) from mytable where pkey='0178-2014-07' 
>> LIMIT 2
>> TOK_QUERY
>>   TOK_FROM
>>     TOK_TABREF
>>       TOK_TABNAME
>>        mytable
>>   TOK_INSERT
>>     TOK_DESTINATION
>>       TOK_DIR
>>         TOK_TMP_FILE
>>     TOK_SELECT
>>       TOK_SELEXPR
>>         TOK_FUNCTION
>>           when
>>           =
>>             &
>>               TOK_TABLE_OR_COL
>>                 bit_field
>>               1
>>             1
>>           -
>>             TOK_TABLE_OR_COL
>>               r_end
>>             TOK_TABLE_OR_COL
>>               r_start
>>           TOK_NULL
>>     TOK_WHERE
>>       =
>>         TOK_TABLE_OR_COL
>>           pkey
>>         '0178-2014-07'
>>     TOK_LIMIT
>>       2
>>
>>
>> SQLState:  null
>> ErrorCode: 0
>>
>> ​
>>
>> similarly, concat seems to work but I get a failure in this query (due to
>> LPAD I believe) :
>>
>> select customer_id from mytable where
>> pkey=concat_ws('-',LPAD('077',4,'0'),'2014-07') LIMIT 2
>>
>> (there is something going on with the fact that the function is in the
>> where clause....the following work fine:
>>
>> select concat_ws('-', LPAD(cast(112717 % 1024 AS
>> STRING),4,'0'),'2014-07') from mytable where pkey='0077-2014-07' LIMIT 2
>> select customer_id from mytable where pkey=concat_ws('-','0077','2014-07')
>> LIMIT 2
>> ​
>> )
>>
>> 14/10/03 14:51:35 ERROR server.SparkSQLOperationManager: Error executing 
>> query:
>> org.apache.spark.SparkException: Task not serializable
>>         at 
>> org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:166)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:158)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:1242)
>>         at org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:597)
>>         at 
>> org.apache.spark.sql.execution.Limit.execute(basicOperators.scala:146)
>>         at 
>> org.apache.spark.sql.hive.HiveContext$QueryExecution.toRdd$lzycompute(HiveContext.scala:360)
>>         at 
>> org.apache.spark.sql.hive.HiveContext$QueryExecution.toRdd(HiveContext.scala:360)
>>         at 
>> org.apache.spark.sql.hive.thriftserver.server.SparkSQLOperationManager$$anon$1.run(SparkSQLOperationManager.scala:185)
>>         at 
>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:193)
>>         at 
>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:175)
>>         at 
>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:150)
>>         at 
>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:207)
>>         at 
>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
>>         at 
>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
>>         at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>>         at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>>         at 
>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:58)
>>         at 
>> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:55)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:415)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>>         at 
>> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:526)
>>         at 
>> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:55)
>>         at 
>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>>         at 
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>         at 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:744)
>> Caused by: java.io.NotSerializableException: java.lang.reflect.Constructor
>>         at 
>> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1183)
>>         at 
>> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
>>         at 
>> java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
>>         at 
>> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
>>         at 
>> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
>>         at 
>> java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1377)
>>         at 
>> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1173)
>>         at 
>> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
>>         at 
>> java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
>>         at 
>> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
>>         at 
>> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
>>         at 
>> java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347)
>>         at 
>> scala.collection.immutable.$colon$colon.writeObject(List.scala:379)
>>         at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
>>         at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>
>> ​
>>
>>
>> Let me know if any of these warrant a JIRA
>>
>> thanks
>>
>>
>>
>>
>> On Thu, Oct 2, 2014 at 2:00 PM, Michael Armbrust <mich...@databricks.com>
>> wrote:
>>
>>> What are the errors you are seeing.  All of those functions should work.
>>>
>>> On Thu, Oct 2, 2014 at 6:56 AM, Yana Kadiyska <yana.kadiy...@gmail.com>
>>> wrote:
>>>
>>>> Hi, in an effort to migrate off of Shark I recently tried the Thrift
>>>> JDBC server that comes with Spark 1.1.0.
>>>>
>>>> However I observed that conditional functions do not work (I tried
>>>> 'case' and 'coalesce')
>>>>
>>>> some string functions like 'concat' also did not work.
>>>>
>>>> Is there a list of what's missing or a roadmap of when it will be
>>>> added? (I know percentiles are pending, for example but do not see JIRAs
>>>> for the others in this email).
>>>>
>>>
>>>
>>
>

Reply via email to