[ANNOUNCE] Beam 2.23.0 Released

2020-07-29 Thread Valentyn Tymofieiev
The Apache Beam team is pleased to announce the release of version 2.23.0. Apache Beam is an open source unified programming model to define and execute data processing pipelines, including ETL, batch and stream (continuous) processing. See: https://beam.apache.org You can download the release

Re: Program and registration for Beam Digital Summit

2020-07-29 Thread Maximilian Michels
Thanks Pedro! Great to see the program! This is going to be an exciting event. Forwarding to the dev mailing list, in case people didn't see this here. -Max On 29.07.20 20:25, Pedro Galvan wrote: Hello! Just a quick message to let everybody know that we have published the program for the

Program and registration for Beam Digital Summit

2020-07-29 Thread Pedro Galvan
Hello! Just a quick message to let everybody know that we have published the program for the Beam Digital Summit. It is available at https://2020.beamsummit.org/program With more than 30 talks and workshops covering all the scope from introductory sessions to advanced scenarios and use

Re: Exceptions: Attempt to deliver a timer to a DoFn, but timers are not supported in Dataflow.

2020-07-29 Thread Mohil Khare
Hi Kenneth, I am on beam java sdk 2.19 With enableStreamingEngine set to true and using default machine type (n1-standard-2). Thanks and regards Mohil On Wed, Jul 29, 2020 at 10:36 AM Kenneth Knowles wrote: > Hi Mohil, > > It helps also to tell us what version of Beam you are using and some

Re: Exceptions: Attempt to deliver a timer to a DoFn, but timers are not supported in Dataflow.

2020-07-29 Thread Kenneth Knowles
Hi Mohil, It helps also to tell us what version of Beam you are using and some more details. This looks related to https://issues.apache.org/jira/browse/BEAM-6855 which claims to be resolved in 2.17.0 Kenn On Mon, Jul 27, 2020 at 11:47 PM Mohil Khare wrote: > Hello all, > > I think I found

Re: KafkaUnboundedReader

2020-07-29 Thread Maximilian Michels
Hi Dinh, The check only de-duplicates in case the consumer processes the same offset multiple times. It ensures the offset is always increasing. If this has been fixed in Kafka, which the comment assumes, the condition will never be true. Which Kafka version are you using? -Max On

KafkaUnboundedReader

2020-07-29 Thread wang Wu
Hi, I am curious about this comment: if (offset < expected) { // -- (a) // this can happen when compression is enabled in Kafka (seems to be fixed in 0.10) // should we check if the offset is way off from consumedOffset (say > 1M)? LOG.warn( "{}: