Hi Libis,
spark-sql CLI is not supported by carbondata.
Why don't you use carbon thrift server and beeline, it is also same as
spark-sql CLI and it gives execution time for each query.
Start carbondata thrift server script.
bin/spark-submit --class
org.apache.carbondata.spark.thriftserver.CarbonT
Hi
Now i can use carbondata 1.0.0 with spark-shell(spark 2.1) as:
./bin/spark-shell --jars
but it's inconvenient to get the query time , so i try to use
./bin/spark-sql --jars ,but i found some
errors when create table :
spark-sql> create table if not exists test_table(id string, name string,
Hi
I used the below method in spark shell for DEMO, for your reference:
import org.apache.spark.sql.catalyst.util._
benchmark { carbondf.filter($"name" === "Allen" and $"gender" === "Male"
and $"province" === "NB" and $"singler" === "false").count }
Regards
Liang
2017-02-06 22:07 GMT-05:00 Y
Hi all,
When we are using sparkshell + carbondata to send a query, how can we get the
excution duration? Some topics are thrown as follows:
1. One query can produce one or more jobs, and some of the jobs may have DAG
dependence, thus we can't get the excution duration by sum up all the jo