Regarding P2s and P3s getting resolved as well: pretty much every healthy project has a backlog that grows without bound. So we do need a place to put that backlog. I think P3 is where things tend to end up, because P2s that do not receive a comment are automatically downgraded to P3. These may still be resolved, but there isn't any hope that _all_ of them get resolved.
Kenn On Fri, Jun 24, 2022 at 10:32 AM Alexey Romanenko <aromanenko....@gmail.com> wrote: > Thanks, Danny! > > On 24 Jun 2022, at 19:23, Danny McCormick <dannymccorm...@google.com> > wrote: > > Sure, I put up a fix - https://github.com/apache/beam/pull/22048 > > On Fri, Jun 24, 2022 at 1:20 PM Alexey Romanenko <aromanenko....@gmail.com> > wrote: > >> >> >> > 2. The links in this report start with api.github.* and don’t take us >> directly to the issues. >> >> > Yeah Danny pointed that out as well. I'm assuming he knows how to fix >> it? >> >> This is already fixed - Pablo actually beat me to it! >> <https://github.com/apache/beam/pull/22033> >> >> >> It adds also a colon after URL and some mail clients consider it as a >> part of URL which leads to a broken link. >> Should we just remove a colon there or add a space between? >> >> — >> Alexey >> >> >> Thanks, >> Danny >> >> On Thu, Jun 23, 2022 at 8:30 PM Brian Hulette <bhule...@google.com> >> wrote: >> >>> +1 for that proposal! >>> >>> > 1. P2 and P3 issues should be noticed and resolved as well. Shall we >>> have a longer time window for the rest of not triaged or stagnate issues >>> and include them? >>> >>> I worry these lists would get _very_ long and wouldn't be actionable. >>> But maybe it's worth reporting something like "There are 376 P2's with no >>> update in the last 6 months" with a link to a query? >>> >>> > 2. The links in this report start with api.github.* and don’t take us >>> directly to the issues. >>> >>> Yeah Danny pointed that out as well. I'm assuming he knows how to fix it? >>> >>> On Thu, Jun 23, 2022 at 2:37 PM Pablo Estrada <pabl...@google.com> >>> wrote: >>> >>>> Thanks. I like the proposal, and I've found the emails useful. >>>> Best >>>> -P. >>>> >>>> On Thu, Jun 23, 2022 at 2:33 PM Manu Zhang <owenzhang1...@gmail.com> >>>> wrote: >>>> >>>>> Sounds good! It’s like our internal reports of JIRA tickets exceeding >>>>> SLA time and having no response from engineers. We either resolve them or >>>>> downgrade the priority to extend time window. >>>>> >>>>> Besides, >>>>> 1. P2 and P3 issues should be noticed and resolved as well. Shall we >>>>> have a longer time window for the rest of not triaged or stagnate issues >>>>> and include them? >>>>> 2. The links in this report start with api.github.* and don’t take us >>>>> directly to the issues. >>>>> >>>>> >>>>> Danny McCormick <dannymccorm...@google.com>于2022年6月24日 周五04:48写道: >>>>> >>>>>> That generally sounds right to me - I also would vote that we >>>>>> consolidate to 1 email and stop distinguishing between flaky P1s and >>>>>> normal >>>>>> P1s. >>>>>> >>>>>> So the single daily report would be: >>>>>> >>>>>> - Unassigned P0s >>>>>> - P0s with no update in the last 36 hours >>>>>> - Unassigned P1s >>>>>> - P1s with no update in the last 7 days >>>>>> >>>>>> I think that will generate a pretty good list of issues that require >>>>>> some kind of action. >>>>>> >>>>>> On Thu, Jun 23, 2022 at 4:43 PM Kenneth Knowles <k...@apache.org> >>>>>> wrote: >>>>>> >>>>>>> Sounds good to me. Perhaps P0s > 36 hours ago (presumably they are >>>>>>> more like ~hours for true outages of CI/website/etc) and P1s > 7 days? >>>>>>> >>>>>>> On Thu, Jun 23, 2022 at 1:27 PM Brian Hulette <bhule...@google.com> >>>>>>> wrote: >>>>>>> >>>>>>>> I think that Danny's alternate proposal (a daily email that show >>>>>>>> only issues last updated >7 days ago, and those with no assignee) fits >>>>>>>> well >>>>>>>> with the two goals you describe, if we include "triage needed" issues >>>>>>>> in >>>>>>>> the latter category. Maybe we also explicitly separate these two >>>>>>>> concerns >>>>>>>> in the report? >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Jun 23, 2022 at 1:14 PM Kenneth Knowles <k...@apache.org> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Forking thread because lots of people may just ignore this topic, >>>>>>>>> per the discussion :-) >>>>>>>>> >>>>>>>>> (sometimes gmail doesn't fork thread properly, but here's >>>>>>>>> hoping...) >>>>>>>>> >>>>>>>>> I'll add some other outcomes of these emails: >>>>>>>>> >>>>>>>>> - people file P0s that are not outages and P1s that are not data >>>>>>>>> loss and I downgrade them >>>>>>>>> - I randomly open up a few flaky test bugs and see if I can fix >>>>>>>>> them really quick >>>>>>>>> - people file legit P0s and P1s and I subscribe and follow them >>>>>>>>> >>>>>>>>> Of these, only the last one seems important (not just that *I* >>>>>>>>> follow them, but that new P0s and P1s get immediate attention from >>>>>>>>> many >>>>>>>>> eyes) >>>>>>>>> >>>>>>>>> So maybe one take on the goal is to: >>>>>>>>> >>>>>>>>> - have new P0s and P1s evaluated quickly: P0s are an outage or >>>>>>>>> outage-like occurrence that needs immediate remedy, and P1s need to be >>>>>>>>> evaluated for release blocking, etc. >>>>>>>>> - make sure P0s and P1s get attention appropriate to their >>>>>>>>> priority >>>>>>>>> >>>>>>>>> It can also be helpful to just state the failure modes which would >>>>>>>>> happen by default if we don't have a good process or automation: >>>>>>>>> >>>>>>>>> - Real P0 gets filed and not noticed or fixed in a timely manner, >>>>>>>>> blocking users and/or community in real time >>>>>>>>> - Real P1 gets filed and not noticed, so release goes out with >>>>>>>>> known data loss bug or other total loss of functionality >>>>>>>>> - Non-real P0s and P1s accumulate, throwing off our data and >>>>>>>>> making it hard to find the real problems >>>>>>>>> - Flakes are never fixed >>>>>>>>> >>>>>>>>> WDYT? >>>>>>>>> >>>>>>>>> If we have P0s and P1s in the "awaiting triage" state, those are >>>>>>>>> the ones we need to notice. Then for a P0 or P1 outside of that >>>>>>>>> state, we >>>>>>>>> just need some way of making sure it doesn't stagnate. Or if it does >>>>>>>>> stagnate, that empirically demonstrates it isn't really P1 (just like >>>>>>>>> our >>>>>>>>> P2 to P3 downgrade automation). If everything is P1, nothing is, as >>>>>>>>> they >>>>>>>>> say. >>>>>>>>> >>>>>>>>> Kenn >>>>>>>>> >>>>>>>>> On Thu, Jun 23, 2022 at 10:01 AM Danny McCormick < >>>>>>>>> dannymccorm...@google.com> wrote: >>>>>>>>> >>>>>>>>>> > Maybe it would be helpful to sort these by last update time >>>>>>>>>> (and potentially include that information in the email). Then we can >>>>>>>>>> at >>>>>>>>>> least prioritize them instead of looking at a big wall of issues. >>>>>>>>>> >>>>>>>>>> I agree that this is a good idea (and pretty trivial to do). I'll >>>>>>>>>> update the automation to do that once we get consensus on an >>>>>>>>>> approach. >>>>>>>>>> >>>>>>>>>> > I think the motivation for daily emails is that per the >>>>>>>>>> priorities guide [1] P1 issues should be getting "continuous status >>>>>>>>>> updates". If these issues aren't actually that important, I think >>>>>>>>>> the noise >>>>>>>>>> is good as it should motivate us to prioritize them correctly. In >>>>>>>>>> practice >>>>>>>>>> that hasn't been happening though... >>>>>>>>>> >>>>>>>>>> I guess the questions here are: >>>>>>>>>> >>>>>>>>>> 1) What is the goal of this email? >>>>>>>>>> 2) Is it effective at accomplishing that goal. >>>>>>>>>> >>>>>>>>>> I think you're saying that the goal (or a goal) is to highlight >>>>>>>>>> issues that aren't getting the attention they need; if that's our >>>>>>>>>> goal, >>>>>>>>>> then I don't think this is a particularly effective mechanism for it >>>>>>>>>> because (a) its very unclear which issues fall into that category >>>>>>>>>> and (b) >>>>>>>>>> there are too many to manually go through on a daily basis. From the >>>>>>>>>> email >>>>>>>>>> alone, it's not clear to me that any of the issues above "shouldn't" >>>>>>>>>> be P1s >>>>>>>>>> (though I'd guess you're right that some/many of them don't belong >>>>>>>>>> since >>>>>>>>>> most were created before the Jira -> GH migration based on the >>>>>>>>>> titles). I'd >>>>>>>>>> also argue that a daily email just desensitizes us to them since >>>>>>>>>> there almost always will be *some *valid P1s that don't need >>>>>>>>>> extra attention. >>>>>>>>>> >>>>>>>>>> I do still think this could have value as a weekly email, with >>>>>>>>>> the goal being "it's probably a good idea for someone to take a look >>>>>>>>>> at >>>>>>>>>> each of these". Another option would be to only include issues with >>>>>>>>>> no >>>>>>>>>> action in the last 7 days and/or no assignees and keep it daily. >>>>>>>>>> >>>>>>>>>> A couple side notes: >>>>>>>>>> - No matter what we do, if we keep the current automation in any >>>>>>>>>> form we should fix the url from >>>>>>>>>> https://api.github.com/repos/apache/beam/issues/# to >>>>>>>>>> https://github.com/apache/beam/issues/# - the current links are >>>>>>>>>> very annoying. >>>>>>>>>> - After I send this, I will do a pass of the current P1s since it >>>>>>>>>> does indeed seem like too many are P1s and many should actually be >>>>>>>>>> P2s (or >>>>>>>>>> lower). >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> Danny >>>>>>>>>> >>>>>>>>>> On Thu, Jun 23, 2022 at 12:21 PM Brian Hulette < >>>>>>>>>> bhule...@google.com> wrote: >>>>>>>>>> >>>>>>>>>>> I think the motivation for daily emails is that per the >>>>>>>>>>> priorities guide [1] P1 issues should be getting "continuous status >>>>>>>>>>> updates". If these issues aren't actually that important, I think >>>>>>>>>>> the noise >>>>>>>>>>> is good as it should motivate us to prioritize them correctly. In >>>>>>>>>>> practice >>>>>>>>>>> that hasn't been happening though... >>>>>>>>>>> >>>>>>>>>>> Maybe it would be helpful to sort these by last update time (and >>>>>>>>>>> potentially include that information in the email). Then we can at >>>>>>>>>>> least >>>>>>>>>>> prioritize them instead of looking at a big wall of issues. >>>>>>>>>>> >>>>>>>>>>> Brian >>>>>>>>>>> >>>>>>>>>>> [1] https://beam.apache.org/contribute/issue-priorities/ >>>>>>>>>>> >>>>>>>>>>> On Thu, Jun 23, 2022 at 6:07 AM Danny McCormick < >>>>>>>>>>> dannymccorm...@google.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> I think a weekly summary seems like a good idea for the P1 >>>>>>>>>>>> issues and flaky tests, though daily still seems appropriate for >>>>>>>>>>>> P0 issues. >>>>>>>>>>>> I put up https://github.com/apache/beam/pull/22017 to just >>>>>>>>>>>> send the P1/flaky test reports on Wednesdays, if anyone objects >>>>>>>>>>>> please let >>>>>>>>>>>> me know - I'll wait on merging til tomorrow to leave time for >>>>>>>>>>>> feedback (and >>>>>>>>>>>> it's always reversible 🙂). >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> Danny >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Jun 22, 2022 at 7:05 PM Manu Zhang < >>>>>>>>>>>> owenzhang1...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi all, >>>>>>>>>>>>> >>>>>>>>>>>>> what is this daily summary intended for? Not all issues look >>>>>>>>>>>>> like P1. And will a weekly summary be less noise? >>>>>>>>>>>>> >>>>>>>>>>>>> <beamacti...@gmail.com>于2022年6月22日 周三23:45写道: >>>>>>>>>>>>> >>>>>>>>>>>>>> This is your daily summary of Beam's current P1 issues, not >>>>>>>>>>>>>> including flaky tests. >>>>>>>>>>>>>> >>>>>>>>>>>>>> See >>>>>>>>>>>>>> https://beam.apache.org/contribute/issue-priorities/#p1-critical >>>>>>>>>>>>>> for the meaning and expectations around P1 issues. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21978: >>>>>>>>>>>>>> [Playground] Implement Share Any Code feature on the frontend >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21946: >>>>>>>>>>>>>> [Bug]: No way to read or write to file when running Beam in Flink >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21935: >>>>>>>>>>>>>> [Bug]: Reject illformed GBK Coders >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21897: >>>>>>>>>>>>>> [Feature Request]: Flink runner savepoint backward compatibility >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21893: >>>>>>>>>>>>>> [Bug]: BigQuery Storage Write API implementation does not >>>>>>>>>>>>>> support table >>>>>>>>>>>>>> partitioning >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21794: >>>>>>>>>>>>>> Dataflow runner creates a new timer whenever the output >>>>>>>>>>>>>> timestamp is change >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21763: >>>>>>>>>>>>>> [Playground Task]: Migrate from Google Analytics to Matomo Cloud >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21715: Data >>>>>>>>>>>>>> missing when using CassandraIO.Read >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21713: 404s >>>>>>>>>>>>>> in BigQueryIO don't get output to Failed Inserts PCollection >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21711: >>>>>>>>>>>>>> Python Streaming job failing to drain with BigQueryIO write >>>>>>>>>>>>>> errors >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21703: >>>>>>>>>>>>>> pubsublite.ReadWriteIT failing in >>>>>>>>>>>>>> beam_PostCommit_Java_DataflowV1 and V2 >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21702: >>>>>>>>>>>>>> SpannerWriteIT failing in beam PostCommit Java V1 >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21700: >>>>>>>>>>>>>> --dataflowServiceOptions=use_runner_v2 is broken >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21695: >>>>>>>>>>>>>> DataflowPipelineResult does not raise exception for unsuccessful >>>>>>>>>>>>>> states. >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21694: >>>>>>>>>>>>>> BigQuery Storage API insert with writeResult retry and write to >>>>>>>>>>>>>> error table >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21479: >>>>>>>>>>>>>> Install Python wheel and dependencies to local venv in SDK >>>>>>>>>>>>>> harness >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21478: >>>>>>>>>>>>>> KafkaIO.read.withDynamicRead() doesn't pick up new >>>>>>>>>>>>>> TopicPartitions >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21477: Add >>>>>>>>>>>>>> integration testing for BQ Storage API write modes >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21476: >>>>>>>>>>>>>> WriteToBigQuery Dynamic table destinations returns wrong tableId >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21475: Beam >>>>>>>>>>>>>> x-lang Dataflow tests failing due to _InactiveRpcError >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21473: >>>>>>>>>>>>>> PVR_Spark2_Streaming perma-red >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21466: >>>>>>>>>>>>>> Simplify version override for Dev versions of the Go SDK. >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21465: Kafka >>>>>>>>>>>>>> commit offset drop data on failure for runners that have >>>>>>>>>>>>>> non-checkpointing >>>>>>>>>>>>>> shuffle >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21269: >>>>>>>>>>>>>> Delete orphaned files >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21268: Race >>>>>>>>>>>>>> between member variable being accessed due to leaking >>>>>>>>>>>>>> uninitialized state >>>>>>>>>>>>>> via OutboundObserverFactory >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21267: >>>>>>>>>>>>>> WriteToBigQuery submits a duplicate BQ load job if a 503 error >>>>>>>>>>>>>> code is >>>>>>>>>>>>>> returned from googleapi >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21265: >>>>>>>>>>>>>> apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest.test_run_packable_combine_globally >>>>>>>>>>>>>> 'apache_beam.coders.coder_impl._AbstractIterable' object is not >>>>>>>>>>>>>> reversible >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21263: >>>>>>>>>>>>>> (Broken Pipe induced) Bricked Dataflow Pipeline >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21262: >>>>>>>>>>>>>> Python AfterAny, AfterAll do not follow spec >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21260: >>>>>>>>>>>>>> Python DirectRunner does not emit data at GC time >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21259: >>>>>>>>>>>>>> Consumer group with random prefix >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21258: >>>>>>>>>>>>>> Dataflow error in CombinePerKey operation >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21257: >>>>>>>>>>>>>> Either Create or DirectRunner fails to produce all elements to >>>>>>>>>>>>>> the >>>>>>>>>>>>>> following transform >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21123: >>>>>>>>>>>>>> Multiple jobs running on Flink session cluster reuse the >>>>>>>>>>>>>> persistent Python >>>>>>>>>>>>>> environment. >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21119: >>>>>>>>>>>>>> Migrate to the next version of Python `requests` when released >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21117: "Java >>>>>>>>>>>>>> IO IT Tests" - missing data in grafana >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21115: >>>>>>>>>>>>>> JdbcIO date conversion is sensitive to OS >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21112: >>>>>>>>>>>>>> Dataflow SocketException (SSLException) error while trying to >>>>>>>>>>>>>> send message >>>>>>>>>>>>>> from Cloud Pub/Sub to BigQuery >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21111: Java >>>>>>>>>>>>>> creates an incorrect pipeline proto when core-construction-java >>>>>>>>>>>>>> jar is not >>>>>>>>>>>>>> in the CLASSPATH >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21110: >>>>>>>>>>>>>> codecov/patch has poor behavior >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21109: SDF >>>>>>>>>>>>>> BoundedSource seems to execute significantly slower than 'normal' >>>>>>>>>>>>>> BoundedSource >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21108: >>>>>>>>>>>>>> java.io.InvalidClassException With Flink Kafka >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20979: >>>>>>>>>>>>>> Portable runners should be able to issue checkpoints to >>>>>>>>>>>>>> Splittable DoFn >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20978: >>>>>>>>>>>>>> PubsubIO.readAvroGenericRecord creates SchemaCoder that fails to >>>>>>>>>>>>>> decode >>>>>>>>>>>>>> some Avro logical types >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20973: >>>>>>>>>>>>>> Python Beam SDK Harness hangs when installing pip packages >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20818: >>>>>>>>>>>>>> XmlIO.Read does not handle XML encoding per spec >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20814: JmsIO >>>>>>>>>>>>>> is not acknowledging messages correctly >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20813: No >>>>>>>>>>>>>> trigger early repeatedly for session windows >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20812: >>>>>>>>>>>>>> Cross-language consistency (RequiresStableInputs) is quietly >>>>>>>>>>>>>> broken (at >>>>>>>>>>>>>> least on portable flink runner) >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20692: Timer >>>>>>>>>>>>>> with dataflow runner can be set multiple times (dataflow runner) >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20691: Beam >>>>>>>>>>>>>> metrics should be displayed in Flink UI "Metrics" tab >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20689: Kafka >>>>>>>>>>>>>> commitOffsetsInFinalize OOM on Flink >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20532: >>>>>>>>>>>>>> Support for coder argument in WriteToBigQuery >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20531: >>>>>>>>>>>>>> FileBasedSink: allow setting temp directory provider per dynamic >>>>>>>>>>>>>> destination >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20530: Make >>>>>>>>>>>>>> non-portable Splittable DoFn the only option when executing Java >>>>>>>>>>>>>> "Read" >>>>>>>>>>>>>> transforms >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20529: >>>>>>>>>>>>>> SpannerIO tests don't actually assert anything. >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20528: >>>>>>>>>>>>>> python CombineGlobally().with_fanout() cause duplicate combine >>>>>>>>>>>>>> results for >>>>>>>>>>>>>> sliding windows >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20333: >>>>>>>>>>>>>> beam_PerformanceTests_Kafka_IO failing due to " provided port is >>>>>>>>>>>>>> already >>>>>>>>>>>>>> allocated" >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20332: >>>>>>>>>>>>>> FileIO writeDynamic with AvroIO.sink not writing all data >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20330: >>>>>>>>>>>>>> Remove insecure ssl options from MongoDBIO >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20109: >>>>>>>>>>>>>> SortValues should fail if SecondaryKey coder is not deterministic >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20108: >>>>>>>>>>>>>> Python direct runner doesn't emit empty pane when it should >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20009: >>>>>>>>>>>>>> Environment-sensitive provisioning for Dataflow >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19971: [SQL] >>>>>>>>>>>>>> Some Hive tests throw NullPointerException, but get marked as >>>>>>>>>>>>>> passing >>>>>>>>>>>>>> (Direct Runner) >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19817: >>>>>>>>>>>>>> datetime and decimal should be logical types >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19815: Add >>>>>>>>>>>>>> support for remaining data types in python RowCoder >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19813: >>>>>>>>>>>>>> PubsubIO returns empty message bodies for all messages read >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19556: User >>>>>>>>>>>>>> reports protobuf ClassChangeError running against 2.6.0 or above >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19369: >>>>>>>>>>>>>> KafkaIO doesn't commit offsets while being used as bounded source >>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/17950: >>>>>>>>>>>>>> [Bug]: Java Precommit permared >>>>>>>>>>>>>> >>>>>>>>>>>>> >> >