Re: How more than one spark job can write to same partition in the parquet file

2019-12-11 Thread ayan guha
We partitioned data logically for 2 different jobs...in our use case based on geography... On Thu, 12 Dec 2019 at 3:39 pm, Chetan Khatri wrote: > Thanks, If you can share alternative change in design. I would love to > hear from you. > > On Wed, Dec 11, 2019 at 9:34 PM ayan guha wrote: > >> No

Re: How more than one spark job can write to same partition in the parquet file

2019-12-11 Thread Chetan Khatri
Thanks, If you can share alternative change in design. I would love to hear from you. On Wed, Dec 11, 2019 at 9:34 PM ayan guha wrote: > No we faced problem with that setup. > > On Thu, 12 Dec 2019 at 11:14 am, Chetan Khatri < > chetan.opensou...@gmail.com> wrote: > >> Hi Spark Users, >> would

subscribe

2019-12-11 Thread Genieliu
| | Genieliu | | feixiang...@163.com China | 签名由网易邮箱大师定制

Re: How more than one spark job can write to same partition in the parquet file

2019-12-11 Thread ayan guha
No we faced problem with that setup. On Thu, 12 Dec 2019 at 11:14 am, Chetan Khatri wrote: > Hi Spark Users, > would that be possible to write to same partition to the parquet file > through concurrent two spark jobs with different spark session. > > thanks > -- Best Regards, Ayan Guha

Re: spark-shell, how it works internally

2019-12-11 Thread mykidong
I have found a source how to compile spark codes and dynamically load them into distributed executors in spark repl: https://ardoris.wordpress.com/2014/03/30/how-spark-does-class-loading/ If you run spark repl, you can find the spark configuration like this :

How more than one spark job can write to same partition in the parquet file

2019-12-11 Thread Chetan Khatri
Hi Spark Users, would that be possible to write to same partition to the parquet file through concurrent two spark jobs with different spark session. thanks

Unsubscribe

2019-12-11 Thread Davide Mandrini

Unsubscribe

2019-12-11 Thread Ryan Victory

Spark streaming when a node or nodes go down

2019-12-11 Thread Mich Talebzadeh
Hi, I know this is a basic question but someone enquired about it and I just wanted to fill my knowledge gap so to speak. Within the context of Spark streaming, the RDD is created from the incoming topic and RDD is partitioned and each node of Spark is operating on a partition at that time. OK