Creating composite DoFns is tricky today due to how they are implemented
(via annotated methods). However providing such a method to compose DoFns
would be very useful IMO.

On Fri, Sep 15, 2023 at 9:33 AM Joey Tran <joey.t...@schrodinger.com> wrote:

> Yeah for (1) the concern would be adding a shuffle/fusion break and (2)
> sounds like the likely solution, was just hoping there'd be one that could
> wrap at the PTransform level but I realize now the PTransform abstraction
> is too general as you mentioned to do something like that.
>
> (2) will be likely what we do, though now I'm wondering if it might be
> possible to create a ParDo wrapper that can take a ParDo, extract it's
> dofn, wrap it, and return a new ParDo
>
> On Fri, Sep 15, 2023, 11:53 AM Robert Bradshaw via user <
> user@beam.apache.org> wrote:
>
>> +1 to looking at composite transforms. You could even have a composite
>> transform that takes another transform as one of its construction arguments
>> and whose expand method does pre- and post-processing to the inputs/outputs
>> before/after applying the transform in question. (You could even implement
>> this as a Python decorator if you really wanted, either decorating the
>> expand method itself or the full class...)
>>
>> One of the difficulties is that for a general transform there isn't
>> necessarily a 1:N relationship between outputs and inputs as one has for a
>> DoFn (especially if there is any aggregation involved). There are, however,
>> two partial solutions that might help.
>>
>> (1) You can do a CombineGlobally with a CombineFn (Like Sample) that
>> returns at most N elements. You could do this with a CombinePerKey if you
>> can come up with a reasonable key (e.g. the id of your input elements) that
>> the limit should be a applied to. Note that this may cause a lot of data to
>> be shuffled (though due to combiner lifting, no more than N per bundle).
>>
>> (2) You could have a DoFn that limits to N per bundle by initializing a
>> counter in its start_bundle and passing elements through until the counter
>> reaches a threshold. (Again, one could do this per id if one is available.)
>> It wouldn't stop production of the elements, but if things get fused it
>> would still likely be fairly cheap.
>>
>> Both of these could be prepended to the problematic consuming PTransform
>> as well.
>>
>> - Robert
>>
>>
>>
>> On Fri, Sep 15, 2023 at 8:13 AM Joey Tran <joey.t...@schrodinger.com>
>> wrote:
>>
>>> I'm aware of composite transforms and of the distributed nature of
>>> PTransforms. I'm not suggesting limiting the entire set and my example was
>>> more illustrative than the actual use case.
>>>
>>> My actual use case is basically: I have multiple PTransforms, and let's
>>> say most of them average ~100 generated outputs for a single input. Most of
>>> these PTransforms will occasionally run into an input though that might
>>> output maybe 1M outputs. This can cause issues if for example there are
>>> transforms that follow it that require a lot of compute per input.
>>>
>>> The simplest way to deal with this is to modify the `DoFn`s in our
>>> Ptransforms and add a limiter in the logic (e.g. `if num_outputs_generated
>>> >= OUTPUTS_PER_INPUT_LIMIT: return`). We could duplicate this logic across
>>> our transforms, but it'd be much cleaner if we could lift up this limiting
>>> logic out of the application logic and have some generic wrapper that
>>> extends our transforms.
>>>
>>> Thanks for the discussion!
>>>
>>> On Fri, Sep 15, 2023 at 10:29 AM Alexey Romanenko <
>>> aromanenko....@gmail.com> wrote:
>>>
>>>> I don’t think it’s possible to extend in a way that you are asking
>>>> (like, Java classes “*extend*"). Though, you can create your own
>>>> composite PTransform that will incorporate one or several others inside
>>>> *“expand()”* method. Actually, most of the Beam native PTransforms are
>>>> composite transforms. Please, take a look on doc and examples [1]
>>>>
>>>> Regarding your example, please, be aware that all PTransforms are
>>>> supposed to be executed in distributed environment and the order of records
>>>> is not guaranteed. So, limiting the whole output by fixed number of records
>>>> can be challenging - you’d need to make sure that it will be processed on
>>>> only one worker, that means that you’d need to shuffle all your records by
>>>> the same key and probably sort the records in way that you need.
>>>>
>>>> Did you consider to use “*org.apache.beam.sdk.transforms.Top*” for
>>>> that? [2]
>>>>
>>>> If it doesn’t work for you, could you provide more details of your use
>>>> case? Then we probably can propose the more suitable solutions for that.
>>>>
>>>> [1]
>>>> https://beam.apache.org/documentation/programming-guide/#composite-transforms
>>>> [2]
>>>> https://beam.apache.org/releases/javadoc/2.50.0/org/apache/beam/sdk/transforms/Top.html
>>>>
>>>> —
>>>> Alexey
>>>>
>>>> On 15 Sep 2023, at 14:22, Joey Tran <joey.t...@schrodinger.com> wrote:
>>>>
>>>> Is there a way to extend already defined PTransforms? My question is
>>>> probably better illustrated with an example. Let's say I have a PTransform
>>>> that generates a very variable number of outputs. I'd like to "wrap" that
>>>> PTransform such that if it ever creates more than say 1,000 outputs, then I
>>>> just take the first 1,000 outputs without generating the rest of the
>>>> outputs.
>>>>
>>>> It'd be trivial if I have access to the DoFn, but what if the
>>>> PTransform in question doesn't expose the `DoFn`?
>>>>
>>>>
>>>>

Reply via email to