Re: How to track batch jobs in spark ?

2018-12-05 Thread Mark Hamstra
That will kill an entire Spark application, not a batch Job. On Wed, Dec 5, 2018 at 3:07 PM Priya Matpadi wrote: > if you are deploying your spark application on YARN cluster, > 1. ssh into master node > 2. List the currently running application and retreive the application_id > yarn

Re: How to track batch jobs in spark ?

2018-12-05 Thread Priya Matpadi
if you are deploying your spark application on YARN cluster, 1. ssh into master node 2. List the currently running application and retreive the application_id yarn application --list 3. Kill the application using application_id of the form application_x_ from output of list command

How to track batch jobs in spark ?

2018-12-05 Thread kant kodali
Hi All, How to track batch jobs in spark? For example, is there some id or token i can get after I spawn a batch job and use it to track the progress or to kill the batch job itself? For Streaming, we have StreamingQuery.id() Thanks!

Re: how to change temp directory when spark write data ?

2018-12-05 Thread JF Chen
Directory to use for "scratch" space in Spark, including map output files and RDDs that get stored on disk. It seems a directory on my local disk. But currently the temp directory is on hdfs, under the path of df.write.parquet(path) Regard, Junfeng Chen On Wed, Dec 5, 2018 at 6:28 PM Sandip

Re: OData compliant API for Spark

2018-12-05 Thread Jean Georges Perrin
I was involved in a project like that and we decided to deploy the data in https://ckan.org/. We used Spark for the data pipeline and transformation. Hih. jg > On Dec 4, 2018, at 21:14, Affan Syed wrote: > > All, > > We have been thinking about exposing our platform for analytics an OData

Re: how to change temp directory when spark write data ?

2018-12-05 Thread Sandip Mehta
tryspark.local.dir property. On Wed, Dec 5, 2018 at 1:42 PM JF Chen wrote: > I have two spark apps writing data to one directory. I notice they share > one temp directory, and the spark fist finish writing will clear the temp > directory and the slower one may throw "No lease on *** File does

how to change temp directory when spark write data ?

2018-12-05 Thread JF Chen
I have two spark apps writing data to one directory. I notice they share one temp directory, and the spark fist finish writing will clear the temp directory and the slower one may throw "No lease on *** File does not exist" error So how to specify the temp directory? Thanks! Regard, Junfeng Chen