Re: Discussion about getting excution duration about a query when using sparkshell+carbondata

2017-02-08 Thread Ravindra Pesala
Hi Libis, spark-sql CLI is not supported by carbondata. Why don't you use carbon thrift server and beeline, it is also same as spark-sql CLI and it gives execution time for each query. Start carbondata thrift server script. bin/spark-submit --class org.apache.carbondata.spark.thriftserver.CarbonT

Re: Discussion about getting excution duration about a query when using sparkshell+carbondata

2017-02-08 Thread 范范欣欣
Hi Now i can use carbondata 1.0.0 with spark-shell(spark 2.1) as: ./bin/spark-shell --jars but it's inconvenient to get the query time , so i try to use ./bin/spark-sql --jars ,but i found some errors when create table : spark-sql> create table if not exists test_table(id string, name string,

Re: Discussion about getting excution duration about a query when using sparkshell+carbondata

2017-02-06 Thread Liang Chen
Hi I used the below method in spark shell for DEMO, for your reference: import org.apache.spark.sql.catalyst.util._ benchmark { carbondf.filter($"name" === "Allen" and $"gender" === "Male" and $"province" === "NB" and $"singler" === "false").count } Regards Liang 2017-02-06 22:07 GMT-05:00 Y

Discussion about getting excution duration about a query when using sparkshell+carbondata

2017-02-06 Thread Yinwei Li
Hi all, When we are using sparkshell + carbondata to send a query, how can we get the excution duration? Some topics are thrown as follows: 1. One query can produce one or more jobs, and some of the jobs may have DAG dependence, thus we can't get the excution duration by sum up all the jo