Re: java vs scala for Apache Spark - is there a performance difference ?

2018-10-29 Thread Gourav Sengupta
I genuinely do not think that Scala for Spark needs us to be super in Scala. There is infact a tutorial called as "Just enough Scala for Spark" which even with my IQ does not take more than 40 mins to go through. Also the sytax of Scala is almost always similar to that of Python. Data processing

Re: java vs scala for Apache Spark - is there a performance difference ?

2018-10-29 Thread kant kodali
Most people when they compare two different programming languages 99% of the time it all seems to boil down to syntax sugar. Performance I doubt Scala is ever faster than Java given that Scala likes Heap more than Java. I had also written some pointless micro-benchmarking code like (Random String

Re: java vs scala for Apache Spark - is there a performance difference ?

2018-10-29 Thread Jean Georges Perrin
did not see anything, but curious if you find something. I think one of the big benefit of using Java, for data engineering in the context of Spark, is that you do not have to train a lot of your team to Scala. Now if you want to do data science, Java is probably not the best tool yet... >

Re: dremel paper example schema

2018-10-29 Thread Debasish Das
Open source impl of dremel is parquet ! On Mon, Oct 29, 2018, 8:42 AM Gourav Sengupta wrote: > Hi, > > why not just use dremel? > > Regards, > Gourav Sengupta > > On Mon, Oct 29, 2018 at 1:35 PM lchorbadjiev < > lubomir.chorbadj...@gmail.com> wrote: > >> Hi, >> >> I'm trying to reproduce the

Re: dremel paper example schema

2018-10-29 Thread Gourav Sengupta
Hi, why not just use dremel? Regards, Gourav Sengupta On Mon, Oct 29, 2018 at 1:35 PM lchorbadjiev wrote: > Hi, > > I'm trying to reproduce the example from dremel paper > (https://research.google.com/pubs/archive/36632.pdf) in Apache Spark using > pyspark and I wonder if it is possible at

dremel paper example schema

2018-10-29 Thread lchorbadjiev
Hi, I'm trying to reproduce the example from dremel paper (https://research.google.com/pubs/archive/36632.pdf) in Apache Spark using pyspark and I wonder if it is possible at all? Trying to follow the paper example as close as possible I created this document type: from pyspark.sql.types import

Re: Processing Flexibility Between RDD and Dataframe API

2018-10-29 Thread Gourav Sengupta
Hi, I would recommend reading the book by Matei Zaharia. One of the main differentiating factors between Spark 1.x and subsequent releases has been optimization and hence dataframes, and in no way RDD is going away because dataframes are built on RDD's. The use of RDD's are allowed and is

Re: Processing Flexibility Between RDD and Dataframe API

2018-10-29 Thread Jungtaek Lim
Just 2 cents on just one of contributors: while SQL semantic can express various use cases data scientists encounter, I also agree someone who are end users who are more familiar with code instead of SQL can feel it is not flexible. But counterless efforts have been incorporated into Spark SQL