Re: Spark job stuck at s3a-file-system metrics system started

2020-05-01 Thread Gourav Sengupta
Hi, I think that we should stop using S3a, and use S3. Please try refer about EMRFS and how it provides fantastic advantages :) Regards, Gourav Sengupta On Thu, Apr 30, 2020 at 12:54 AM Aniruddha P Tekade wrote: > Hello, > > I am trying to run a spark job that is trying to write the data

Re: Spark job stuck at s3a-file-system metrics system started

2020-04-30 Thread Abhisheks
Hi there, Read your question and I do believe you are on right path. But what could be worth checking is - are you able to connect to s3 bucket from your worker nodes. I did read that you are able to do it from your machine but since write happens at the the worker end, it might be worth

Spark job stuck at s3a-file-system metrics system started

2020-04-29 Thread Aniruddha P Tekade
Hello, I am trying to run a spark job that is trying to write the data into a custom s3 endpoint bucket. But I am stuck at this line of output and job is not moving forward at all - 20/04/29 16:03:59 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of