https://github.com/apache/beam/issues/26515 tracks this issue. The fix was
merged. Thanks a lot for reporting this issue, Binh!
On Mon, Apr 17, 2023 at 12:58 PM Binh Nguyen Van wrote:
> Hi,
>
> I tested with streaming insert and file load, and they all worked as
> expected. But looks like storag
Hi,
I tested with streaming insert and file load, and they all worked as
expected. But looks like storage API is the new way to go so want to test
it too. I am using Apache Beam v2.46.0 and running it with Google Dataflow.
Thanks
-Binh
On Mon, Apr 17, 2023 at 9:53 AM Reuven Lax via user
wrote:
What version of Beam are you using? There are no known data-loss bugs in
the connector, however if there has been a regression we would like to
address it with high priority.
On Mon, Apr 17, 2023 at 12:47 AM Binh Nguyen Van wrote:
> Hi,
>
> I have a job that uses BigQuery IO Connector to write t
Does FILE_LOADS (
https://beam.apache.org/releases/javadoc/2.46.0/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIO.Write.Method.html#FILE_LOADS)
work for your case?
For STORAGE_WRITE_API, it has been actively improved. If the latest SDK
still has this issue, I highly recommend you to create a Google
Hi,
I have a job that uses BigQuery IO Connector to write to a BigQuery table.
When I test it with a small number of records (100) it works as expected
but when I tested it with a larger number of records (1), I don’t see
all of the records written to the output table but only part of it. It
c