Hi Till,

I understand the errors which appears in my logs are not stopping me from
running the job. I am running flink session cluster in ECS and also
configured graylog to get the container logs. So getting the docker logs is
also not an issue.
But is there a way to suppress this error or any work around ?

Thanks
Sidhant Gupta

On Wed, Oct 7, 2020, 9:15 PM Till Rohrmann <trohrm...@apache.org> wrote:

> Hi Sidhant,
>
> when using Flink's Docker image, then the cluster won't create the out
> files. Instead the components will directly write to STDOUT which is
> captured by Kubernetes and can be viewed using `kubectl logs POD_NAME`. The
> error which appears in your logs is not a problem. It is simply the REST
> handler which tries to serve the out files.
>
> Cheers,
> Till
>
> On Wed, Oct 7, 2020 at 5:11 PM 大森林 <appleyu...@foxmail.com> wrote:
>
>> what's your running mode?
>> if your flink cluster is on yarn mode,then the output you need has no
>> relation to $FLINK_HOME/logs/*.out
>>
>>
>> ------------------ 原始邮件 ------------------
>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>> *发送时间:* 2020年10月7日(星期三) 晚上11:33
>> *收件人:* "大森林"<appleyu...@foxmail.com>;"user"<user@flink.apache.org>;
>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>>
>> Hi,
>>
>> I'm running flink cluster in ecs. There is a pipeline which creates the
>> job manager and then the task manager using the docker image.
>>
>> Not sure if we would want to restart the cluster in production.
>>
>> Is there any way we can make sure the .out files will be created without
>> restart ?
>>
>> I am able to see the logs in the logs tab but not the stdout logs in the
>> web ui and getting the below mentioned error after running the job.
>>
>> Thanks
>> Sidhant Gupta
>>
>>
>> On Wed, Oct 7, 2020, 8:00 PM 大森林 <appleyu...@foxmail.com> wrote:
>>
>>> it's easy,
>>> just restart your flink cluster(standalone mode)
>>>
>>> if you run flink in yarn mode,then the result will display on
>>> $HADOOP/logs/*.out files
>>>
>>> ------------------ 原始邮件 ------------------
>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>>> *发送时间:* 2020年10月7日(星期三) 晚上9:52
>>> *收件人:* "大森林"<appleyu...@foxmail.com>;
>>> *抄送:* "user"<user@flink.apache.org>;
>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>>>
>>> ++ user
>>>
>>> On Wed, Oct 7, 2020, 6:47 PM sidhant gupta <sidhan...@gmail.com> wrote:
>>>
>>>> Hi
>>>>
>>>> I checked in the $FLINK_HOME/logs. The .out file was not there. Can
>>>> you suggest what should be the action item ?
>>>>
>>>> Thanks
>>>> Sidhant Gupta
>>>>
>>>>
>>>> On Wed, Oct 7, 2020, 7:17 AM 大森林 <appleyu...@foxmail.com> wrote:
>>>>
>>>>>
>>>>> check if the .out file is in $FLINK_HOME/logs  please.
>>>>>
>>>>> ------------------ 原始邮件 ------------------
>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>>>>> *发送时间:* 2020年10月7日(星期三) 凌晨1:52
>>>>> *收件人:* "大森林"<appleyu...@foxmail.com>;
>>>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>>>>>
>>>>> Hi,
>>>>>
>>>>> I am just running the docker container as it is by adding just the
>>>>> conf/flink.yaml .
>>>>> I am not sure if the .out file got deleted. Do we need to expose some
>>>>> ports ?
>>>>>
>>>>> Thanks
>>>>> Sidhant Gupta
>>>>>
>>>>>
>>>>>
>>>>> On Tue, Oct 6, 2020, 8:51 PM 大森林 <appleyu...@foxmail.com> wrote:
>>>>>
>>>>>>
>>>>>> Hi,I guess you may deleted .out file in $FLINK_HOME/logs.
>>>>>> you can just use your default log settings.
>>>>>> ------------------ 原始邮件 ------------------
>>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>>>>>> *发送时间:* 2020年10月6日(星期二) 晚上10:59
>>>>>> *收件人:* "user"<user@flink.apache.org>;
>>>>>> *主题:* The file STDOUT does not exist on the TaskExecutor
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I am running dockerized flink:1.11.0-scala_2.11 container in ecs. I
>>>>>> am getting the following error after the job runs:
>>>>>>
>>>>>> ERROR org.apache.flink.runtime.rest.handler.taskmanager.
>>>>>> TaskManagerStdoutFileHandler [] - Unhandled exception.
>>>>>> org.apache.flink.util.FlinkException: The file STDOUT does not exist
>>>>>> on the TaskExecutor.
>>>>>>     at org.apache.flink.runtime.taskexecutor.TaskExecutor
>>>>>> .lambda$requestFileUploadByFilePath$25(TaskExecutor.java:1742)
>>>>>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>>>>>     at java.util.concurrent.CompletableFuture$AsyncSupply.run(
>>>>>> CompletableFuture.java:1604) ~[?:1.8.0_262]
>>>>>>     at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>> ThreadPoolExecutor.java:1149) ~[?:1.8.0_262]
>>>>>>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>>> ThreadPoolExecutor.java:624) ~[?:1.8.0_262]
>>>>>>     at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_262]
>>>>>>
>>>>>> I guess "file" needs to be added in log4j.properties in the docker
>>>>>> container e.g. log4j.rootLogger=INFO, file
>>>>>> Are there any other properties which needs to be configured in any of
>>>>>> the other property files or any jar needs to be added in the */opt/flink
>>>>>> *path ?
>>>>>> Thanks
>>>>>> Sidhant Gupta
>>>>>>
>>>>>>

Reply via email to