Thanks Cham for this.
I actually came across this yesterday and am now looking into it. Not sure
if it will work but trying it out.

Regards,
Pravin

On Tue, Nov 7, 2023 at 12:36 PM Chamikara Jayalath via user <
user@beam.apache.org> wrote:

> Not sure what runner you are using but some runners support updating or
> redeploying your pipeline with the changes which might be easier than doing
> it manually using your workflow. For example, Dataflow streaming can deploy
> a replacement job for you with the updated config for PubSub:
> https://cloud.google.com/dataflow/docs/guides/updating-a-pipeline#Launching
>
>
> Thanks,
> Cham
>
> On Tue, Nov 7, 2023 at 9:25 AM Pravin DSouza <pravin...@gmail.com> wrote:
>
>> Hi Wisniowski,
>>
>> Thank you for the inputs.
>> It is not possible for us to identify all the topics in advance as new
>> requirements in future might warrant new topics or may be able to use one
>> of existing topics.
>> As such, we have decided to proceed with the redeployment option for now
>> if the input topic list changes.
>>
>> Regards,
>> Pravin
>>
>> On Sun, Nov 5, 2023 at 3:10 AM Piotr Wiśniowski <
>> contact.wisniowskipi...@gmail.com> wrote:
>>
>>> Hi,
>>> Another idea. Less cost effective, but easier to implement and maintain:
>>> 1. Read from all possible topics.
>>> 2. 'Map' to add constant topic name to every topic. (Or some other
>>> variations of this step)
>>> 3. 'Filter' with side input to just filter out all topics that should be
>>> silenced.
>>> 4. 'Flatten'
>>> But this requires that new topics are not created during run of the
>>> pipeline and You know them. If not then suggestion with custom 'ParDo' is
>>> only option, but I would also suggest rethinking Your infrastructure setup.
>>>
>>> Best
>>> Wiśniowski Piotr
>>>
>>> śr., 1 lis 2023, 19:06 użytkownik Chamikara Jayalath via user <
>>> user@beam.apache.org> napisał:
>>>
>>>> Currently only some Beam sources are able to consume a configuration
>>>> (set of topics here) that is dynamically generated and I don't think
>>>> PubSubIO is one of them. So probably you'll have to implement a custom DoFn
>>>> that reads from Cloud Pub/Sub to support this. Also, probably you'll have
>>>> to constrain reading with size/time since Cloud Pub/Sub is an unbounded
>>>> input source.
>>>>
>>>> Thanks,
>>>> Cham
>>>>
>>>>
>>>>
>>>> On Tue, Oct 31, 2023 at 2:22 PM Pravin DSouza <pravin...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I have a use case where I want to listen to multiple Pub/Sub input
>>>>> topics
>>>>> and route messages from all multiple input topics to a single
>>>>> destination
>>>>> output topic.
>>>>> The number of input topics can change any time and are stored in a
>>>>> table.
>>>>> For example when I deploy for the first time, I might be reading from 3
>>>>> topics, but after 2 days the table is updated and now I want to read
>>>>> from 4
>>>>> topics.I don't want to redeploy the job because of a change in input
>>>>> topics
>>>>> list.
>>>>> I know how to use SideInputs for refresh, but I am unable to use
>>>>> SideInputs
>>>>> to read the input topic details and pass to PubSubIO.
>>>>> Can you please suggest which of the following ways this can be
>>>>> achieved if
>>>>> new input topics are added:
>>>>> 1. Using SideInput
>>>>> 2. Any other approach
>>>>> 3. Redeploy the job (this is the last course I want to rely on)
>>>>>
>>>>> Please suggest.
>>>>>
>>>>> Thank you,
>>>>> Pravin
>>>>>
>>>>

Reply via email to