Hi,

I would suggest the same thing as Vino did: it might be possible to use stdout 
somehow, but it’s a better idea to coordinate in some other way. Produce some 
(side?) output with a control message from one job once it finishes, that will 
control the second job.

Piotrek

> On 25 Nov 2019, at 09:38, vino yang <yanghua1...@gmail.com> wrote:
> 
> Hi Komal,
> 
> > Thank you! That's exactly what's happening. Is there any way to force it 
> > write to a specific .out of a TaskManager?
> 
> No, I am curious why the two jobs depend on stdout? Can we introduce another 
> coordinator other than stdout? IMO, this mechanism is not always available.
> 
> Best,
> Vino
> 
> Komal Mariam <komal.mar...@gmail.com <mailto:komal.mar...@gmail.com>> 
> 于2019年11月25日周一 上午10:46写道:
> Hi Theo,
> 
> I want to interrupt/cancel my current job as it has produced the desired 
> results even though it runs infinitely,  and the next one requires full 
> resources. 
> 
> Due to some technical issue we cannot access the web UI so just working with 
> the CLI, for now. 
> 
> I found a less crude way by running the command ./bin/flink cancel <job id>  
> specified by the commands listed here: 
> https://ci.apache.org/projects/flink/flink-docs-stable/ops/cli.html 
> <https://ci.apache.org/projects/flink/flink-docs-stable/ops/cli.html>
> 
> ------------
> 
> Hello Vino,
> 
> Thank you! That's exactly what's happening. Is there any way to force it 
> write to a specific .out of a TaskManager?
> 
> 
> Best Regards,
> Komal
> 
> 
> 
> On Mon, 25 Nov 2019 at 11:10, vino yang <yanghua1...@gmail.com 
> <mailto:yanghua1...@gmail.com>> wrote:
> Hi Komal,
> 
> Since you use the Flink standalone deployment mode, the tasks of the jobs 
> which print information to the STDOUT may randomly deploy in any task manager 
> of the cluster. Did you check other Task Managers out file?
> 
> Best,
> Vino
> 
> Komal Mariam <komal.mar...@gmail.com <mailto:komal.mar...@gmail.com>> 
> 于2019年11月22日周五 下午6:59写道:
> Dear all,
> 
> Thank you for your help regarding my previous queries. Unfortunately, I'm 
> stuck with another one and will really appreciate your input. 
> 
> I can't seem to produce any outputs in "flink-taskexecutor-0.out" from my 
> second job after submitting the first one in my 3-node-flink standalone 
> cluster.
> 
> Say I want to test out two jobs sequentially. (I do not want to run them 
> concurrently/in parallel).
> 
> After submitting "job1.jar " via command line, I press "Ctrl + C" to exit 
> from it (as it runs infinitely). After that I 
> try to submit a second jar file having the same properties (group-id, topic, 
> etc) with the only difference being the query written in main function.
> 
> The first job produces relevant outputs in "flink-taskexecutor-0.out" but the 
> second one doesn't.
> 
> The only way I can see the output produced is if I restart the cluster after 
> job1 and then submit job2 as it produces another .out file.
> 
> But I want to submit 2 jobs sequentially and see their outputs without having 
> to restart my cluster. Is there any way to do this?
> 
> Additional info:
> For both jobs I'm using DataStream API and I have set: 
>  StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
> 
> Best Regards,
> Komal

Reply via email to