Re: Help Required on Spark - Convert DataFrame to List with out using collect

2017-12-22 Thread Naresh Dulam
Hi Sunitha, Make the class which is having the common function your calling as serializable. Thank you, Naresh On Wed, Dec 20, 2017 at 9:58 PM Sunitha Chennareddy < chennareddysuni...@gmail.com> wrote: > Hi, > > Thank You All.. > > Here is my requirement, I have a dataframe which contains

Storage at node or executor level

2017-12-22 Thread Jean Georges Perrin
Hi all, This is more of a general architecture question, I have my idea, but wanted to confirm/infirm... When your executor is accessing data, where is it stored: at the executor level or at the worker level? jg - To

Passing an array of more than 22 elements in a UDF

2017-12-22 Thread Aakash Basu
Hi, I am using Spark 2.2 using Java, can anyone please suggest me how to take more than 22 parameters in an UDF? I mean, if I want to pass all the parameters as an array of integers? Thanks, Aakash.

Re: Passing an array of more than 22 elements in a UDF

2017-12-22 Thread ayan guha
Hi I think you are in correct track. You can stuff all your param in a suitable data structure like array or dict and pass this structure as a single param in your udf. On Fri, 22 Dec 2017 at 2:55 pm, Aakash Basu wrote: > Hi, > > I am using Spark 2.2 using Java, can

unsubscribe

2017-12-22 Thread 施朝浩
unsubscribe - To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Re: [E] How to do stop streaming before the application got killed

2017-12-22 Thread Rastogi, Pankaj
You can add a shutdown hook to your JVM and request spark streaming context to stop gracefully. /** * Shutdown hook to shutdown JVM gracefully * @param ssCtx */ def addShutdownHook(ssCtx: StreamingContext) = { Runtime.getRuntime.addShutdownHook( new Thread() { override

How to do stop streaming before the application got killed

2017-12-22 Thread Toy
I'm trying to write a deployment job for Spark application. Basically the job will send yarn application --kill app_id to the cluster but after the application received the signal it dies without finishing whatever is processing or stopping the stream. I'm using Spark Streaming. What's the best