Hi Diego,

If you want the data of all streams to be written to the same files, you
can also union the streams before sending them to the sink.

Best, Fabian

2016-11-29 15:50 GMT+01:00 Kostas Kloudas <k.klou...@data-artisans.com>:

> Hi Diego,
>
> You cannot prefix each stream with a different
> string so that the paths do not collide?
>
> If I understand your use-case correctly, this might work.
>
> Cheers,
> Kostas
>
> On Nov 29, 2016, at 10:04 AM, Diego Fustes Villadóniga <dfus...@oesia.com>
> wrote:
>
> Hi Kostas,
>
> Thanks for your reply.
>
> The problem is at the initialization of the job.  The reason was that I
> was using the same HDFS path as sink for 3 different streams, which is
> something that I would like. I can fix it by using different paths
> for each stream.
>
> Maybe there is a way to achieve this in a different manner by joining the
> streams somehow before sinking… maybe through Kafka?
>
> Kind Regards,
>
> Diego
>
>
>
> *De:* Kostas Kloudas [mailto:k.klou...@data-artisans.com
> <k.klou...@data-artisans.com>]
> *Enviado el:* lunes, 28 de noviembre de 2016 19:13
> *Para:* user@flink.apache.org
> *Asunto:* Re: Problems with RollingSink
>
> Hi Diego,
>
> The message shows that two tasks are trying to touch concurrently the same
> file.
>
> This message is thrown upon recovery after a failure, or at the
> initialization of the job?
> Could you please check the logs for other exceptions before this?
>
> Can this be related to this issue?
> https://www.mail-archive.com/issues@flink.apache.org/msg73871.html
>
> Thanks,
> Kostas
>
>
> On Nov 28, 2016, at 5:37 PM, Diego Fustes Villadóniga <dfus...@oesia.com>
> wrote:
>
> Hi colleagues,
>
> I am experiencing problems when trying to write events from a stream to
> HDFS. I get the following exception:
>
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException):
> failed to create file 
> /user/biguardian/events/2016-11-28--15/flinkpart-0-0.text
> for DFSClient_NONMAPREDUCE_1634980080_43 for client 172.21.40.75 because
> current leaseholder is trying to recreate file.
>
> My Flink version is 1.1.3 and I am running it directly from a JAR (not in
> YARN) with java -jar.
>
> Do you know the reason of this error?
>
> Kind regards,
>
> Diego
>
>
>

Reply via email to