Hi Till, I understand the errors which appears in my logs are not stopping me from running the job. I am running flink session cluster in ECS and also configured graylog to get the container logs. So getting the docker logs is also not an issue. But is there a way to suppress this error or any work around ?
Thanks Sidhant Gupta On Wed, Oct 7, 2020, 9:15 PM Till Rohrmann <trohrm...@apache.org> wrote: > Hi Sidhant, > > when using Flink's Docker image, then the cluster won't create the out > files. Instead the components will directly write to STDOUT which is > captured by Kubernetes and can be viewed using `kubectl logs POD_NAME`. The > error which appears in your logs is not a problem. It is simply the REST > handler which tries to serve the out files. > > Cheers, > Till > > On Wed, Oct 7, 2020 at 5:11 PM 大森林 <appleyu...@foxmail.com> wrote: > >> what's your running mode? >> if your flink cluster is on yarn mode,then the output you need has no >> relation to $FLINK_HOME/logs/*.out >> >> >> ------------------ 原始邮件 ------------------ >> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >> *发送时间:* 2020年10月7日(星期三) 晚上11:33 >> *收件人:* "大森林"<appleyu...@foxmail.com>;"user"<user@flink.apache.org>; >> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >> >> Hi, >> >> I'm running flink cluster in ecs. There is a pipeline which creates the >> job manager and then the task manager using the docker image. >> >> Not sure if we would want to restart the cluster in production. >> >> Is there any way we can make sure the .out files will be created without >> restart ? >> >> I am able to see the logs in the logs tab but not the stdout logs in the >> web ui and getting the below mentioned error after running the job. >> >> Thanks >> Sidhant Gupta >> >> >> On Wed, Oct 7, 2020, 8:00 PM 大森林 <appleyu...@foxmail.com> wrote: >> >>> it's easy, >>> just restart your flink cluster(standalone mode) >>> >>> if you run flink in yarn mode,then the result will display on >>> $HADOOP/logs/*.out files >>> >>> ------------------ 原始邮件 ------------------ >>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>> *发送时间:* 2020年10月7日(星期三) 晚上9:52 >>> *收件人:* "大森林"<appleyu...@foxmail.com>; >>> *抄送:* "user"<user@flink.apache.org>; >>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >>> >>> ++ user >>> >>> On Wed, Oct 7, 2020, 6:47 PM sidhant gupta <sidhan...@gmail.com> wrote: >>> >>>> Hi >>>> >>>> I checked in the $FLINK_HOME/logs. The .out file was not there. Can >>>> you suggest what should be the action item ? >>>> >>>> Thanks >>>> Sidhant Gupta >>>> >>>> >>>> On Wed, Oct 7, 2020, 7:17 AM 大森林 <appleyu...@foxmail.com> wrote: >>>> >>>>> >>>>> check if the .out file is in $FLINK_HOME/logs please. >>>>> >>>>> ------------------ 原始邮件 ------------------ >>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>>>> *发送时间:* 2020年10月7日(星期三) 凌晨1:52 >>>>> *收件人:* "大森林"<appleyu...@foxmail.com>; >>>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >>>>> >>>>> Hi, >>>>> >>>>> I am just running the docker container as it is by adding just the >>>>> conf/flink.yaml . >>>>> I am not sure if the .out file got deleted. Do we need to expose some >>>>> ports ? >>>>> >>>>> Thanks >>>>> Sidhant Gupta >>>>> >>>>> >>>>> >>>>> On Tue, Oct 6, 2020, 8:51 PM 大森林 <appleyu...@foxmail.com> wrote: >>>>> >>>>>> >>>>>> Hi,I guess you may deleted .out file in $FLINK_HOME/logs. >>>>>> you can just use your default log settings. >>>>>> ------------------ 原始邮件 ------------------ >>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>>>>> *发送时间:* 2020年10月6日(星期二) 晚上10:59 >>>>>> *收件人:* "user"<user@flink.apache.org>; >>>>>> *主题:* The file STDOUT does not exist on the TaskExecutor >>>>>> >>>>>> Hi, >>>>>> >>>>>> I am running dockerized flink:1.11.0-scala_2.11 container in ecs. I >>>>>> am getting the following error after the job runs: >>>>>> >>>>>> ERROR org.apache.flink.runtime.rest.handler.taskmanager. >>>>>> TaskManagerStdoutFileHandler [] - Unhandled exception. >>>>>> org.apache.flink.util.FlinkException: The file STDOUT does not exist >>>>>> on the TaskExecutor. >>>>>> at org.apache.flink.runtime.taskexecutor.TaskExecutor >>>>>> .lambda$requestFileUploadByFilePath$25(TaskExecutor.java:1742) >>>>>> ~[flink-dist_2.11-1.11.0.jar:1.11.0] >>>>>> at java.util.concurrent.CompletableFuture$AsyncSupply.run( >>>>>> CompletableFuture.java:1604) ~[?:1.8.0_262] >>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>> ThreadPoolExecutor.java:1149) ~[?:1.8.0_262] >>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>> ThreadPoolExecutor.java:624) ~[?:1.8.0_262] >>>>>> at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_262] >>>>>> >>>>>> I guess "file" needs to be added in log4j.properties in the docker >>>>>> container e.g. log4j.rootLogger=INFO, file >>>>>> Are there any other properties which needs to be configured in any of >>>>>> the other property files or any jar needs to be added in the */opt/flink >>>>>> *path ? >>>>>> Thanks >>>>>> Sidhant Gupta >>>>>> >>>>>>