Hi Komal,

> Thank you! That's exactly what's happening. Is there any way to force it
write to a specific .out of a TaskManager?

No, I am curious why the two jobs depend on stdout? Can we introduce
another coordinator other than stdout? IMO, this mechanism is not always
available.

Best,
Vino

Komal Mariam <komal.mar...@gmail.com> 于2019年11月25日周一 上午10:46写道:

> Hi Theo,
>
> I want to interrupt/cancel my current job as it has produced the desired
> results even though it runs infinitely,  and the next one requires full
> resources.
>
> Due to some technical issue we cannot access the web UI so just working
> with the CLI, for now.
>
> I found a less crude way by running the command ./bin/flink cancel <job
> id>  specified by the commands listed here:
> https://ci.apache.org/projects/flink/flink-docs-stable/ops/cli.html
>
> ------------
>
> Hello Vino,
>
> Thank you! That's exactly what's happening. Is there any way to force it
> write to a specific .out of a TaskManager?
>
>
> Best Regards,
> Komal
>
>
>
> On Mon, 25 Nov 2019 at 11:10, vino yang <yanghua1...@gmail.com> wrote:
>
>> Hi Komal,
>>
>> Since you use the Flink standalone deployment mode, the tasks of the jobs
>> which print information to the STDOUT may randomly deploy in any task
>> manager of the cluster. Did you check other Task Managers out file?
>>
>> Best,
>> Vino
>>
>> Komal Mariam <komal.mar...@gmail.com> 于2019年11月22日周五 下午6:59写道:
>>
>>> Dear all,
>>>
>>> Thank you for your help regarding my previous queries. Unfortunately,
>>> I'm stuck with another one and will really appreciate your input.
>>>
>>> I can't seem to produce any outputs in "flink-taskexecutor-0.out" from
>>> my second job after submitting the first one in my 3-node-flink standalone
>>> cluster.
>>>
>>> Say I want to test out two jobs sequentially. (I do not want to run them
>>> concurrently/in parallel).
>>>
>>> After submitting "job1.jar " via command line, I press "Ctrl + C" to
>>> exit from it (as it runs infinitely). After that I
>>> try to submit a second jar file having the same properties (group-id,
>>> topic, etc) with the only difference being the query written in main
>>> function.
>>>
>>> The first job produces relevant outputs in "flink-taskexecutor-0.out"
>>> but the second one doesn't.
>>>
>>> The only way I can see the output produced is if I restart the cluster
>>> after job1 and then submit job2 as it produces another .out file.
>>>
>>> But I want to submit 2 jobs sequentially and see their outputs without
>>> having to restart my cluster. Is there any way to do this?
>>>
>>> Additional info:
>>> For both jobs I'm using DataStream API and I have set:
>>>  StreamExecutionEnvironment env =
>>> StreamExecutionEnvironment.getExecutionEnvironment();
>>>
>>> Best Regards,
>>> Komal
>>>
>>

Reply via email to