Hi,

Have a look here -
https://repost.aws/knowledge-center/spark-driver-logs-emr-cluster.
Usually, you have application logs out-of-the-box in the driver stdout.
It looks like
"s3://aws-logs-111111111111-us-east-1/elasticmapreduce/j-35PUYZBQVIJNM/containers/application_1572839353552_0008/container_1572839353552_0008_01_000001/stdout.gz".

Grisha

On Sat, Dec 30, 2023 at 2:49 PM PRASHANT L <prashant...@gmail.com> wrote:

> Hi Grisha
> This is Great :) It worked thanks alot
>
> I have this requirement , I will be running my spark application on EMR
> and build a custom logging to create logs on S3. Any idea what should I do?
> or In general if i create a custom log (with my Application name ), where
> will logs be generated when run in cluster mode (since in cluster mode jobs
> are executed all over different machine)
>
> On Sat, Dec 30, 2023 at 1:56 PM Grisha Weintraub <
> grisha.weintr...@gmail.com> wrote:
>
>> In Java, it expects an array of Columns, so you can simply cast your list
>> to an array:
>>
>> array_df.select(fields.toArray(new Column[0]))
>>
>>
>>
>>
>> On Fri, Dec 29, 2023 at 10:58 PM PRASHANT L <prashant...@gmail.com>
>> wrote:
>>
>>>
>>> Team
>>> I am using Java and want to select columns from Dataframe , columns are
>>> stored in List<Column>
>>> equivalent of below scala code
>>> *  array_df=array_df.select(fields: _*)*
>>>
>>>
>>> When I try array_df=array_df.select(fields) , I get error saying Cast to
>>> Column
>>>
>>> I am using Spark 3.4
>>>
>>

Reply via email to