chunyang commented on a change in pull request #10979: [BEAM-8841] Support writing data to BigQuery via Avro in Python SDK URL: https://github.com/apache/beam/pull/10979#discussion_r385442046
########## File path: sdks/python/apache_beam/io/gcp/bigquery.py ########## @@ -1361,87 +1369,18 @@ def __init__( self.triggering_frequency = triggering_frequency self.insert_retry_strategy = insert_retry_strategy self._validate = validate + self._temp_file_format = temp_file_format or bigquery_tools.FileFormat.JSON Review comment: AFAICT using Avro has no disadvantages compared to JSON for loading data into BigQuery, but would requiring a schema constitute a breaking API change for semantic versioning purposes? Personally I'm for using Avro as default. I guess when users update Beam, they'll specify a `temp_file_format` explicitly to get the old behavior. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services