Spark Converting dataframe to Rdd reduces partitions

2017-01-02 Thread manish jaiswal
Hi, I am getting issue while converting dataframe to Rdd, it reduces partitions. In our code, Dataframe was created as : DataFrame DF = hiveContext.sql("select * from table_instance"); When I convert my dataframe to rdd and try to get its number of partitions as RDD newRDD = Df.rdd();

Re: SPARK SQL READING FROM HIVE

2016-08-08 Thread manish jaiswal
correct its creating delta file in hdfs.but after compaction it merge all data and create extra directory where all bucketed data present.( i am able to read data from hive but not from sparksql).

Re: SPARK SQL READING FROM HIVE

2016-08-08 Thread manish jaiswal
i am using spark 1.6.0 and hive 1.2.1. reading from hive transactional table is not supported yet by sparl sql? On Tue, Aug 9, 2016 at 12:18 AM, manish jaiswal <manishsr...@gmail.com> wrote: > Hi, > > I am not able to read data from hive transactional table using sparksql. >

SPARK SQL READING FROM HIVE

2016-08-08 Thread manish jaiswal
Hi, I am not able to read data from hive transactional table using sparksql. (i don't want read via hive jdbc) Please help.

Spark Job trigger in production

2016-07-18 Thread manish jaiswal
Hi, What is the best approach to trigger spark job in production cluster?

HiveContext

2016-07-01 Thread manish jaiswal
Hi, Using sparkHiveContext when we read all rows where age was between 0 and 100, even though we requested rows where age was less than 15. Such full table scanning is an expensive operation. ORC avoids this type of overhead by using predicate push-down with three levels of built-in indexes

HiveContext

2016-06-30 Thread manish jaiswal
-- Forwarded message -- From: "manish jaiswal" <manishsr...@gmail.com> Date: Jun 30, 2016 17:35 Subject: HiveContext To: <user@spark.apache.org>, <user-subscr...@spark.apache.org>, < user-h...@spark.apache.org> Cc: Hi, I am new to Spark.I foun