You could create a batch pipeline that reads GCS and writes to BigQuery. And you can use this template https://cloud.google.com/dataflow/docs/guides/templates/provided/cloud-storage-to-bigquery .
On Sat, May 6, 2023 at 1:10 AM Utkarsh Parekh <utkarsh.s.par...@gmail.com> wrote: > Hi, > > I'm writing a simple streaming beam application. The application job is > doing following tasks: > > 1. Reads data from GCS bucket (project 1) and loads into Kafka topic > 2. Reads data from Kafka topic and loads into BigQUery (project 3) > > Composer running in Project 1 > Data Flow running in project 2 > > I'm using BeamRunPythonPipelineOperator and DataflowConfiguration > configuration. > > Is this the right setup? What would be the gcp_conn_id? Any suggestions? > > Utkarsh >