RE: Support of other languages?

2015-09-22 Thread Sun, Rui
Palamuttam [mailto:rahulpala...@gmail.com] Sent: Thursday, September 17, 2015 3:09 PM To: Sun, Rui Cc: user@spark.apache.org Subject: Re: Support of other languages? Hi, Thank you for both responses. Sun you pointed out the exact issue I was referring to, which is copying,serializing, deserializing

Re: Support of other languages?

2015-09-17 Thread Rahul Palamuttam
Hi, Thank you for both responses. Sun you pointed out the exact issue I was referring to, which is copying,serializing, deserializing, the byte-array between the JVM heap and the worker memory. It also doesn't make sense why the byte-array should be kept on-heap, since the data of the parent

RE: Support of other languages?

2015-09-09 Thread Sun, Rui
Hi, Rahul, To support a new language other than Java/Scala in spark, it is different between RDD API and DataFrame API. For RDD API: RDD is a distributed collection of the language-specific data types whose representation is unknown to JVM. Also transformation functions for RDD are written

Re: Support of other languages?

2015-09-08 Thread Nagaraj Chandrashekar
Hi Rahul, I may not have the answer for what you are looking for but my thoughts are given below. I have worked with HP Vertica and R VIA UDF¹s (User Defined Functions). I don¹t have any experience with Spark R till now. I would expect it might follow the similar route. UDF functions