Re: How to increase Spark partitions for the DataFrame?

2015-10-08 Thread Umesh Kacha
ntext.read().format("orc").load("/hdfs/path/to/orc/files/"); > > df.select().groupby(..) > > > > > > > > > > -- > > View this message in context: > http:/

Re: How to increase Spark partitions for the DataFrame?

2015-10-08 Thread Lan Jiang
files through shuffle in 12 > > partitions. Please guide. > > > > DataFrame df = > > hiveContext.read().format("orc").load("/hdfs/path/to/orc/files/"); > > df.select().groupby(..) > > > > > > > > > > -- > > Vi

Re: How to increase Spark partitions for the DataFrame?

2015-10-08 Thread Umesh Kacha
>> > >> > DataFrame df = >> > hiveContext.read().format("orc").load("/hdfs/path/to/orc/files/"); >> > df.select().groupby(..) >> > >> > >> > >> > >> > -- >> > View this message in context: >&

Re: How to increase Spark partitions for the DataFrame?

2015-10-08 Thread Ted Yu
t; does not >>> > hang for long time because of reading 10 GB files through shuffle in 12 >>> > partitions. Please guide. >>> > >>> > DataFrame df = >>> > hiveContext.read().format("orc").load("/hdfs/path/to/orc/fi

How to increase Spark partitions for the DataFrame?

2015-10-08 Thread unk1102
upby(..) -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-increase-Spark-partitions-for-the-DataFrame-tp24980.html Sent from the Apache Spark User List mailing list archive at Nabble.com. ---

Re: How to increase Spark partitions for the DataFrame?

2015-10-08 Thread Lan Jiang
d("/hdfs/path/to/orc/files/"); > df.select().groupby(..) > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/How-to-increase-Spark-partitions-for-the-DataFrame-