Hi all
now, I'm trying the SparkSQL with hivecontext.
when I execute the hql like the following.
---
val ctx = new org.apache.spark.sql.hive.HiveContext(sc)
import ctx._
val queries = ctx.hql("select keyword from queries where dt =
'2015-02-01' limit 10000000")
---
It seem that the number of the partitions ot the queries is set by 1.
Is this the specifications for schemaRDD, SparkSQL, HiveContext ?
Are there any means to set the number of partitions arbitrary value
except for explicit repartition
Masaki Rikitoku
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]