Here are some figures (small warn is I did priviledge beam a lot in this benchmark, a bit more than it should in a real impl, I'll say more about it after):
I copied the code at: https://gist.github.com/rmannibucau/fd98fb6a10f9557613fb145c8e7e2de1 And results at: https://gist.github.com/rmannibucau/3d3d0d61e85d45f2959900f208e71c5e Summary is: Benchmark Mode Cnt Score Error Units Comparison.beam thrpt 5 217003888,048 ± 4951801,578 ops/s Comparison.beamCompletionStage thrpt 5 55437442,544 ± 1750845,680 ops/s Comparison.beamFastCompletionStage thrpt 5 128422225,642 ± 3215651,832 ops/s Comparison.defaultCompletionStage thrpt 5 57284761,644 ± 1022305,051 ops/s Comparison.fastCompletionStage thrpt 5 202739102,801 ± 5384078,170 ops/s Comparison.passthrough thrpt 5 391503537,773 ± 6147672,843 ops/s Comments: 1. completionstage then*(Function|Consumer) overhead is visible and creating more objects than beam it is a bit slower 2. it is trivial to make completionstage faster than the default in all the "precomputed" cases beam can know about (we mentionned it earlier) and it falls in the "fastCompletionStage" benchmarks which are x2 adapted to beam case (and allows to dropp one instantiation and compare and swap logic) 3. the diff between the fast completionstage and beam is mainly about the creation of one more wrapper + lambda handling 4. however figures stays high (milllions ops/s!). if you add any IO or real computation into the mix you will converge to a sensitively close number 5. don't forget that being based on CompletionStage we can become reactive/async and scale more appropriately if runners embrace that and at the end be faster anyway (the old NIO vs BIO topic ;)) To summarize: I don't think these figures means that completionstage is a bad fit for beam but actually the opposite even if the raw difference is significative. It still shows that the completionstage based throughoutput is high enough to fit real world batches IMHO. Just to illustrate that point, if I just do a System.getProperty("user.name").length() (my username being rmannibucau): @Benchmark public void stringLength(final Blackhole blackhole) { blackhole.consume(System.getProperty("user.name").length()); } the execution is slower than completionfuture in beam case: Comparison.stringLength thrpt 5 42428432,508 ± 4555866,693 ops/s Romain Manni-Bucau @rmannibucau <https://twitter.com/rmannibucau> | Blog <https://rmannibucau.metawerx.net/> | Old Blog <http://rmannibucau.wordpress.com/> | Github <https://github.com/rmannibucau> | LinkedIn <https://www.linkedin.com/in/rmannibucau> | Book <https://www.packtpub.com/application-development/java-ee-8-high-performance> 2018-03-12 18:58 GMT+01:00 Romain Manni-Bucau <rmannibu...@gmail.com>: > > No more but can try to gather some figures and compare it to beam dofn overhead which should be at the same level or a bit more here since it is never unwrapped whereas completionfuture is a raw code chain without beam in the middle. > > Le 12 mars 2018 18:18, "Lukasz Cwik" <lc...@google.com> a écrit : >> >> Do you have data that supports this? >> >> Note that in reality for something like passing an element between DoFns, the constant in o(1) actually matters. Decreasing SDK harness overhead is a good thing though. >> >> On Mon, Mar 12, 2018 at 10:14 AM, Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>> >>> By itself just the overhead of instantiating a wrapper (so nothing with the recent JVM GC improvement done for the stream/optional usages). After if you use the chaining you have a light overhead but still o(1) you can desire to skip when doing sync code but which will enable you to run way faster doing IO/async code by optimizing the CPU usage properly when you tune your slaves/workers. So tempted to summarize it as "has an overhead allowing to not run slower". It doesn't prevent beam to still expose a synchronous API collapse at evaluation time in a single fn which will give you the best of both worlds. >>> >>> >>> Romain Manni-Bucau >>> @rmannibucau | Blog | Old Blog | Github | LinkedIn | Book >>> >>> 2018-03-12 18:08 GMT+01:00 Lukasz Cwik <lc...@google.com>: >>>> >>>> It is expected that SDKs will have all their cores fully utilized by processing bundles in parallel and not by performing intrabundle parallelization. This allows for DoFns to be chained together via regular method calls because the overhead to pass a single element through all the DoFn's should be as minimal as possible >>>> >>>> What is the overhead of using a completion stage vs using a regular method call? >>>> >>>> On Sun, Mar 11, 2018 at 10:18 PM, Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>> >>>>> >>>>> >>>>> Le 12 mars 2018 00:19, "Reuven Lax" <re...@google.com> a écrit : >>>>> >>>>> >>>>> >>>>> >>>>> On Sun, Mar 11, 2018 at 7:40 AM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>> >>>>>> makes me think beam should maybe do 2 internals changes before moving forward on (s)df API changes: >>>>>> >>>>>> 1. define a beam singleton per JVM (classloader hierarchy actually but you get the idea I think) which can be used to store things locally for reuse - see 2 for an example or metrics pusher work Etienne does could benefit from it too >>>>> >>>>> >>>>> I think we do need something like this, but it needs to be a bit more than a singleton per JVM. For one thing it needs to be at least per pipeline within a JVM. You might run multiple tests in a single JVM, and it should also be possible to run those tests in parallel without the static state interfering with each other. I also think the state needs to be addressable per step (i.e. a ParDo can look up its static state without caring about static state belonging to another ParDo). >>>>> >>>>> >>>>> Agree but you can register in a singleton the pipeline (as "ref" not as instance) and therefore hit the same need. +1 to have scopes (singleton, pipeline, thread) but it still requires a single singleton to handle serialization ;). >>>>> >>>>> >>>>> >>>>> >>>>>> 2. define a SPI to load (s)dofn parameter provider instead of having an ArgProvider which provides everything which is supported. This way you can use any kind of parameter and the parameterproviders can use 1. to handle their own state. First impl of the parameterprovider SPI would be a) state b) timer c) reactive handlers and potentially user parameter providers (service like which can be singleton in the scope of a "JVM" thanks to 1). >>>>>> >>>>>> >>>>>> Romain Manni-Bucau >>>>>> @rmannibucau | Blog | Old Blog | Github | LinkedIn | Book >>>>>> >>>>>> 2018-03-11 15:32 GMT+01:00 Reuven Lax <re...@google.com>: >>>>>>> >>>>>>> Yep. Introduce OutputEmitter, and Process context no longer has much use. >>>>>>> >>>>>>> On Sun, Mar 11, 2018, 11:19 AM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>> >>>>>>>> Which is still a key feature for sdf but agree it can be dropped for an outputemitter pattern and the dofn moved to a plain parameters injection based pattern. Both (which completionstage) stays compatible :). >>>>>>>> >>>>>>>> Le 11 mars 2018 13:12, "Reuven Lax" <re...@google.com> a écrit : >>>>>>>>> >>>>>>>>> I think process context should go away completely. At that point it has little use except for a way to send output downstream. >>>>>>>>> >>>>>>>>> On Sun, Mar 11, 2018, 6:07 AM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>> Hmm, thinking out loud but completionstage should/could be extended to replace processcontext since it represents element and output at the same time no? >>>>>>>>>> >>>>>>>>>> Le 11 mars 2018 00:57, "Kenneth Knowles" <k...@google.com> a écrit : >>>>>>>>>>> >>>>>>>>>>> Yea, I think it could. But it is probably more readable to not overload the term, plus certainly a bit simpler in implementation. So perhaps @AsyncElement to make it very clear. >>>>>>>>>>> >>>>>>>>>>> Kenn >>>>>>>>>>> >>>>>>>>>>> On Sat, Mar 10, 2018 at 1:32 PM Reuven Lax <re...@google.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Ken, can NewDoFn distinguish at generation time the difference between: >>>>>>>>>>>> >>>>>>>>>>>> public void process(@Element CompletionStage<InputT> element, ...) { >>>>>>>>>>>> >>>>>>>>>>>> and >>>>>>>>>>>> >>>>>>>>>>>> public void process(@Element Input element, ...) { >>>>>>>>>>>> >>>>>>>>>>>> If not, then we would probably need separate annotations.... >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Sat, Mar 10, 2018 at 11:09 AM Kenneth Knowles < k...@google.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Nice! I agree that providing a CompletionStage for chaining is much better than an ExecutorService, and very clear. >>>>>>>>>>>>> >>>>>>>>>>>>> It is very feasible to add support that looks like >>>>>>>>>>>>> >>>>>>>>>>>>> new DoFn<InputT, OutputT>() { >>>>>>>>>>>>> @ProcessElement >>>>>>>>>>>>> public void process(@Element CompletionStage<InputT> element, ...) { >>>>>>>>>>>>> element.thenApply(...) >>>>>>>>>>>>> } >>>>>>>>>>>>> } >>>>>>>>>>>>> >>>>>>>>>>>>> If we had this available, I think users could even experiment with this often as it might help even where it isn't obvious. >>>>>>>>>>>>> >>>>>>>>>>>>> My main hesitation is that big part of Beam is giving a basic/imperative style of programming a DoFn that executes in a very smart functional/parallel way. Full future-oriented programming is not explored much outside of Javascript (and maybe Haskell) and requires greater discipline in programming in a functional manner - if you are mutating stuff in your callback you are going to have bugs, and then when you add concurrency control you are going to have bad performance and deadlocks. So I definitely wouldn't make it the default or want to spend all our support effort on teaching advanced programming technique. >>>>>>>>>>>>> >>>>>>>>>>>>> Kenn >>>>>>>>>>>>> >>>>>>>>>>>>> On Sat, Mar 10, 2018 at 9:31 AM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2018-03-10 17:30 GMT+01:00 Reuven Lax <re...@google.com>: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Have you considered drafting in detail what you think this API might look like? >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Yes, but it is after the "enhancements" - for my use cases - and "bugs" list so didn't started to work on it much. >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> If it's a radically different API, it might be more appropriate as an alternative parallel Beam API rather than a replacement for the current API (there is also one such fluent API in the works). >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> What I plan is to draft it on top of beam (so the "useless" case I spoke about before) and then propose to impl it ~natively and move it as main API for another major. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Sat, Mar 10, 2018 at 7:23 AM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2018-03-10 16:19 GMT+01:00 Reuven Lax <re...@google.com>: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> This is another version (maybe a better, Java 8 idiomatic one?) of what Kenn suggested. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Note that with NewDoFn this need not be incompatible (so might not require waiting till Beam 3.0). We can recognize new parameters to processElement and populate add needed. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> This is right however in my head it was a single way movemenent to enforce the design to be reactive and not fake a reactive API with a sync and not reactive impl which is what would be done today with both support I fear. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Sat, Mar 10, 2018, 12:13 PM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Yes, for the dofn for instance, instead of having processcontext.element()=<T> you get a CompletionStage<T> and output gets it as well. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> This way you register an execution chain. Mixed with streams you get a big data java 8/9/10 API which enabkes any connectivity in a wel performing manner ;). >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Le 10 mars 2018 13:56, "Reuven Lax" <re...@google.com> a écrit : >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> So you mean the user should have a way of registering asynchronous activity with a callback (the callback must be registered with Beam, because Beam needs to know not to mark the element as done until all associated callbacks have completed). I think that's basically what Kenn was suggesting, unless I'm missing something. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Fri, Mar 9, 2018 at 11:07 PM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Yes, callback based. Beam today is synchronous and until bundles+combines are reactive friendly, beam will be synchronous whatever other parts do. Becoming reactive will enable to manage the threading issues properly and to have better scalability on the overall execution when remote IO are involved. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> However it requires to break source, sdf design to use completionstage - or equivalent - to chain the processing properly and in an unified fashion. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Le 9 mars 2018 23:48, "Reuven Lax" <re...@google.com> a écrit : >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> If you're talking about reactive programming, at a certain level beam is already reactive. Are you referring to a specific way of writing the code? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On Fri, Mar 9, 2018 at 1:59 PM Reuven Lax < re...@google.com> wrote: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> What do you mean by reactive? >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> On Fri, Mar 9, 2018, 6:58 PM Romain Manni-Bucau < rmannibu...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> @Kenn: why not preferring to make beam reactive? Would alow to scale way more without having to hardly synchronize multithreading. Elegant and efficient :). Beam 3? >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Le 9 mars 2018 22:49, "Kenneth Knowles" < k...@google.com> a écrit : >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> I will start with the "exciting futuristic" answer, which is that we envision the new DoFn to be able to provide an automatic ExecutorService parameters that you can use as you wish. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> new DoFn<>() { >>>>>>>>>>>>>>>>>>>>>>> @ProcessElement >>>>>>>>>>>>>>>>>>>>>>> public void process(ProcessContext ctx, ExecutorService executorService) { >>>>>>>>>>>>>>>>>>>>>>> ... launch some futures, put them in instance vars ... >>>>>>>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> @FinishBundle >>>>>>>>>>>>>>>>>>>>>>> public void finish(...) { >>>>>>>>>>>>>>>>>>>>>>> ... block on futures, output results if appropriate ... >>>>>>>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> This way, the Java SDK harness can use its overarching knowledge of what is going on in a computation to, for example, share a thread pool between different bits. This was one reason to delete IntraBundleParallelization - it didn't allow the runner and user code to properly manage how many things were going on concurrently. And mostly the runner should own parallelizing to max out cores and what user code needs is asynchrony hooks that can interact with that. However, this feature is not thoroughly considered. TBD how much the harness itself manages blocking on outstanding requests versus it being your responsibility in FinishBundle, etc. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> I haven't explored rolling your own here, if you are willing to do the knob tuning to get the threading acceptable for your particular use case. Perhaps someone else can weigh in. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Kenn >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> On Fri, Mar 9, 2018 at 1:38 PM Josh Ferge < josh.fe...@bounceexchange.com> wrote: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Hello all: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Our team has a pipeline that make external network calls. These pipelines are currently super slow, and the hypothesis is that they are slow because we are not threading for our network calls. The github issue below provides some discussion around this: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/beam/pull/957 >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> In beam 1.0, there was IntraBundleParallelization, which helped with this. However, this was removed because it didn't comply with a few BEAM paradigms. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Questions going forward: >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> What is advised for jobs that make blocking network calls? It seems bundling the elements into groups of size X prior to passing to the DoFn, and managing the threading within the function might work. thoughts? >>>>>>>>>>>>>>>>>>>>>>>> Are these types of jobs even suitable for beam? >>>>>>>>>>>>>>>>>>>>>>>> Are there any plans to develop features that help with this? >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>> >>>>> >>>> >>> >>