Regarding P2s and P3s getting resolved as well: pretty much every healthy
project has a backlog that grows without bound. So we do need a place to
put that backlog. I think P3 is where things tend to end up, because P2s
that do not receive a comment are automatically downgraded to P3. These may
still be resolved, but there isn't any hope that _all_ of them get resolved.

Kenn

On Fri, Jun 24, 2022 at 10:32 AM Alexey Romanenko <aromanenko....@gmail.com>
wrote:

> Thanks, Danny!
>
> On 24 Jun 2022, at 19:23, Danny McCormick <dannymccorm...@google.com>
> wrote:
>
> Sure, I put up a fix - https://github.com/apache/beam/pull/22048
>
> On Fri, Jun 24, 2022 at 1:20 PM Alexey Romanenko <aromanenko....@gmail.com>
> wrote:
>
>>
>>
>> > 2. The links in this report start with api.github.* and don’t take us
>> directly to the issues.
>>
>> > Yeah Danny pointed that out as well. I'm assuming he knows how to fix
>> it?
>>
>> This is already fixed - Pablo actually beat me to it!
>> <https://github.com/apache/beam/pull/22033>
>>
>>
>> It adds also a colon after URL and some mail clients consider it as a
>> part of URL which leads to a broken link.
>> Should we just remove a colon there or add a space between?
>>
>> —
>> Alexey
>>
>>
>> Thanks,
>> Danny
>>
>> On Thu, Jun 23, 2022 at 8:30 PM Brian Hulette <bhule...@google.com>
>> wrote:
>>
>>> +1 for that proposal!
>>>
>>> > 1. P2 and P3 issues should be noticed and resolved as well. Shall we
>>> have a longer time window for the rest of not triaged or stagnate issues
>>> and include them?
>>>
>>> I worry these lists would get _very_ long and wouldn't be actionable.
>>> But maybe it's worth reporting something like "There are 376 P2's with no
>>> update in the last 6 months" with a link to a query?
>>>
>>> > 2. The links in this report start with api.github.* and don’t take us
>>> directly to the issues.
>>>
>>> Yeah Danny pointed that out as well. I'm assuming he knows how to fix it?
>>>
>>> On Thu, Jun 23, 2022 at 2:37 PM Pablo Estrada <pabl...@google.com>
>>> wrote:
>>>
>>>> Thanks. I like the proposal, and I've found the emails useful.
>>>> Best
>>>> -P.
>>>>
>>>> On Thu, Jun 23, 2022 at 2:33 PM Manu Zhang <owenzhang1...@gmail.com>
>>>> wrote:
>>>>
>>>>> Sounds good! It’s like our internal reports of JIRA tickets exceeding
>>>>> SLA time and having no response from engineers.  We either resolve them or
>>>>> downgrade the priority to extend time window.
>>>>>
>>>>> Besides,
>>>>> 1. P2 and P3 issues should be noticed and resolved as well. Shall we
>>>>> have a longer time window for the rest of not triaged or stagnate issues
>>>>> and include them?
>>>>> 2. The links in this report start with api.github.* and don’t take us
>>>>> directly to the issues.
>>>>>
>>>>>
>>>>> Danny McCormick <dannymccorm...@google.com>于2022年6月24日 周五04:48写道:
>>>>>
>>>>>> That generally sounds right to me - I also would vote that we
>>>>>> consolidate to 1 email and stop distinguishing between flaky P1s and 
>>>>>> normal
>>>>>> P1s.
>>>>>>
>>>>>> So the single daily report would be:
>>>>>>
>>>>>> - Unassigned P0s
>>>>>> - P0s with no update in the last 36 hours
>>>>>> - Unassigned P1s
>>>>>> - P1s with no update in the last 7 days
>>>>>>
>>>>>> I think that will generate a pretty good list of issues that require
>>>>>> some kind of action.
>>>>>>
>>>>>> On Thu, Jun 23, 2022 at 4:43 PM Kenneth Knowles <k...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Sounds good to me. Perhaps P0s > 36 hours ago (presumably they are
>>>>>>> more like ~hours for true outages of CI/website/etc) and P1s > 7 days?
>>>>>>>
>>>>>>> On Thu, Jun 23, 2022 at 1:27 PM Brian Hulette <bhule...@google.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I think that Danny's alternate proposal (a daily email that show
>>>>>>>> only issues last updated >7 days ago, and those with no assignee) fits 
>>>>>>>> well
>>>>>>>> with the two goals you describe, if we include "triage needed" issues 
>>>>>>>> in
>>>>>>>> the latter category. Maybe we also explicitly separate these two 
>>>>>>>> concerns
>>>>>>>> in the report?
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, Jun 23, 2022 at 1:14 PM Kenneth Knowles <k...@apache.org>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Forking thread because lots of people may just ignore this topic,
>>>>>>>>> per the discussion :-)
>>>>>>>>>
>>>>>>>>> (sometimes gmail doesn't fork thread properly, but here's
>>>>>>>>> hoping...)
>>>>>>>>>
>>>>>>>>> I'll add some other outcomes of these emails:
>>>>>>>>>
>>>>>>>>>  - people file P0s that are not outages and P1s that are not data
>>>>>>>>> loss and I downgrade them
>>>>>>>>>  - I randomly open up a few flaky test bugs and see if I can fix
>>>>>>>>> them really quick
>>>>>>>>>  - people file legit P0s and P1s and I subscribe and follow them
>>>>>>>>>
>>>>>>>>> Of these, only the last one seems important (not just that *I*
>>>>>>>>> follow them, but that new P0s and P1s get immediate attention from 
>>>>>>>>> many
>>>>>>>>> eyes)
>>>>>>>>>
>>>>>>>>> So maybe one take on the goal is to:
>>>>>>>>>
>>>>>>>>>  - have new P0s and P1s evaluated quickly: P0s are an outage or
>>>>>>>>> outage-like occurrence that needs immediate remedy, and P1s need to be
>>>>>>>>> evaluated for release blocking, etc.
>>>>>>>>>  - make sure P0s and P1s get attention appropriate to their
>>>>>>>>> priority
>>>>>>>>>
>>>>>>>>> It can also be helpful to just state the failure modes which would
>>>>>>>>> happen by default if we don't have a good process or automation:
>>>>>>>>>
>>>>>>>>>  - Real P0 gets filed and not noticed or fixed in a timely manner,
>>>>>>>>> blocking users and/or community in real time
>>>>>>>>>  - Real P1 gets filed and not noticed, so release goes out with
>>>>>>>>> known data loss bug or other total loss of functionality
>>>>>>>>>  - Non-real P0s and P1s accumulate, throwing off our data and
>>>>>>>>> making it hard to find the real problems
>>>>>>>>>  - Flakes are never fixed
>>>>>>>>>
>>>>>>>>> WDYT?
>>>>>>>>>
>>>>>>>>> If we have P0s and P1s in the "awaiting triage" state, those are
>>>>>>>>> the ones we need to notice. Then for a P0 or P1 outside of that 
>>>>>>>>> state, we
>>>>>>>>> just need some way of making sure it doesn't stagnate. Or if it does
>>>>>>>>> stagnate, that empirically demonstrates it isn't really P1 (just like 
>>>>>>>>> our
>>>>>>>>> P2 to P3 downgrade automation). If everything is P1, nothing is, as 
>>>>>>>>> they
>>>>>>>>> say.
>>>>>>>>>
>>>>>>>>> Kenn
>>>>>>>>>
>>>>>>>>> On Thu, Jun 23, 2022 at 10:01 AM Danny McCormick <
>>>>>>>>> dannymccorm...@google.com> wrote:
>>>>>>>>>
>>>>>>>>>> > Maybe it would be helpful to sort these by last update time
>>>>>>>>>> (and potentially include that information in the email). Then we can 
>>>>>>>>>> at
>>>>>>>>>> least prioritize them instead of looking at a big wall of issues.
>>>>>>>>>>
>>>>>>>>>> I agree that this is a good idea (and pretty trivial to do). I'll
>>>>>>>>>> update the automation to do that once we get consensus on an 
>>>>>>>>>> approach.
>>>>>>>>>>
>>>>>>>>>> > I think the motivation for daily emails is that per the
>>>>>>>>>> priorities guide [1] P1 issues should be getting "continuous status
>>>>>>>>>> updates". If these issues aren't actually that important, I think 
>>>>>>>>>> the noise
>>>>>>>>>> is good as it should motivate us to prioritize them correctly. In 
>>>>>>>>>> practice
>>>>>>>>>> that hasn't been happening though...
>>>>>>>>>>
>>>>>>>>>> I guess the questions here are:
>>>>>>>>>>
>>>>>>>>>> 1) What is the goal of this email?
>>>>>>>>>> 2) Is it effective at accomplishing that goal.
>>>>>>>>>>
>>>>>>>>>> I think you're saying that the goal (or a goal) is to highlight
>>>>>>>>>> issues that aren't getting the attention they need; if that's our 
>>>>>>>>>> goal,
>>>>>>>>>> then I don't think this is a particularly effective mechanism for it
>>>>>>>>>> because (a) its very unclear which issues fall into that category 
>>>>>>>>>> and (b)
>>>>>>>>>> there are too many to manually go through on a daily basis. From the 
>>>>>>>>>> email
>>>>>>>>>> alone, it's not clear to me that any of the issues above "shouldn't" 
>>>>>>>>>> be P1s
>>>>>>>>>> (though I'd guess you're right that some/many of them don't belong 
>>>>>>>>>> since
>>>>>>>>>> most were created before the Jira -> GH migration based on the 
>>>>>>>>>> titles). I'd
>>>>>>>>>> also argue that a daily email just desensitizes us to them since
>>>>>>>>>> there almost always will be *some *valid P1s that don't need
>>>>>>>>>> extra attention.
>>>>>>>>>>
>>>>>>>>>> I do still think this could have value as a weekly email, with
>>>>>>>>>> the goal being "it's probably a good idea for someone to take a look 
>>>>>>>>>> at
>>>>>>>>>> each of these". Another option would be to only include issues with 
>>>>>>>>>> no
>>>>>>>>>> action in the last 7 days and/or no assignees and keep it daily.
>>>>>>>>>>
>>>>>>>>>> A couple side notes:
>>>>>>>>>> - No matter what we do, if we keep the current automation in any
>>>>>>>>>> form we should fix the url from
>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/# to
>>>>>>>>>> https://github.com/apache/beam/issues/# - the current links are
>>>>>>>>>> very annoying.
>>>>>>>>>> - After I send this, I will do a pass of the current P1s since it
>>>>>>>>>> does indeed seem like too many are P1s and many should actually be 
>>>>>>>>>> P2s (or
>>>>>>>>>> lower).
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>> Danny
>>>>>>>>>>
>>>>>>>>>> On Thu, Jun 23, 2022 at 12:21 PM Brian Hulette <
>>>>>>>>>> bhule...@google.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> I think the motivation for daily emails is that per the
>>>>>>>>>>> priorities guide [1] P1 issues should be getting "continuous status
>>>>>>>>>>> updates". If these issues aren't actually that important, I think 
>>>>>>>>>>> the noise
>>>>>>>>>>> is good as it should motivate us to prioritize them correctly. In 
>>>>>>>>>>> practice
>>>>>>>>>>> that hasn't been happening though...
>>>>>>>>>>>
>>>>>>>>>>> Maybe it would be helpful to sort these by last update time (and
>>>>>>>>>>> potentially include that information in the email). Then we can at 
>>>>>>>>>>> least
>>>>>>>>>>> prioritize them instead of looking at a big wall of issues.
>>>>>>>>>>>
>>>>>>>>>>> Brian
>>>>>>>>>>>
>>>>>>>>>>> [1] https://beam.apache.org/contribute/issue-priorities/
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Jun 23, 2022 at 6:07 AM Danny McCormick <
>>>>>>>>>>> dannymccorm...@google.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I think a weekly summary seems like a good idea for the P1
>>>>>>>>>>>> issues and flaky tests, though daily still seems appropriate for 
>>>>>>>>>>>> P0 issues.
>>>>>>>>>>>> I put up https://github.com/apache/beam/pull/22017 to just
>>>>>>>>>>>> send the P1/flaky test reports on Wednesdays, if anyone objects 
>>>>>>>>>>>> please let
>>>>>>>>>>>> me know - I'll wait on merging til tomorrow to leave time for 
>>>>>>>>>>>> feedback (and
>>>>>>>>>>>> it's always reversible 🙂).
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Danny
>>>>>>>>>>>>
>>>>>>>>>>>> On Wed, Jun 22, 2022 at 7:05 PM Manu Zhang <
>>>>>>>>>>>> owenzhang1...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi all,
>>>>>>>>>>>>>
>>>>>>>>>>>>> what is this daily summary intended for? Not all issues look
>>>>>>>>>>>>> like P1. And will a weekly summary be less noise?
>>>>>>>>>>>>>
>>>>>>>>>>>>> <beamacti...@gmail.com>于2022年6月22日 周三23:45写道:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> This is your daily summary of Beam's current P1 issues, not
>>>>>>>>>>>>>> including flaky tests.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>     See
>>>>>>>>>>>>>> https://beam.apache.org/contribute/issue-priorities/#p1-critical
>>>>>>>>>>>>>> for the meaning and expectations around P1 issues.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21978:
>>>>>>>>>>>>>> [Playground] Implement Share Any Code feature on the frontend
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21946:
>>>>>>>>>>>>>> [Bug]: No way to read or write to file when running Beam in Flink
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21935:
>>>>>>>>>>>>>> [Bug]: Reject illformed GBK Coders
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21897:
>>>>>>>>>>>>>> [Feature Request]: Flink runner savepoint backward compatibility
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21893:
>>>>>>>>>>>>>> [Bug]: BigQuery Storage Write API implementation does not 
>>>>>>>>>>>>>> support table
>>>>>>>>>>>>>> partitioning
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21794:
>>>>>>>>>>>>>> Dataflow runner creates a new timer whenever the output 
>>>>>>>>>>>>>> timestamp is change
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21763:
>>>>>>>>>>>>>> [Playground Task]: Migrate from Google Analytics to Matomo Cloud
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21715: Data
>>>>>>>>>>>>>> missing when using CassandraIO.Read
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21713: 404s
>>>>>>>>>>>>>> in BigQueryIO don't get output to Failed Inserts PCollection
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21711:
>>>>>>>>>>>>>> Python Streaming job failing to drain with BigQueryIO write 
>>>>>>>>>>>>>> errors
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21703:
>>>>>>>>>>>>>> pubsublite.ReadWriteIT failing in 
>>>>>>>>>>>>>> beam_PostCommit_Java_DataflowV1 and V2
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21702:
>>>>>>>>>>>>>> SpannerWriteIT failing in beam PostCommit Java V1
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21700:
>>>>>>>>>>>>>> --dataflowServiceOptions=use_runner_v2 is broken
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21695:
>>>>>>>>>>>>>> DataflowPipelineResult does not raise exception for unsuccessful 
>>>>>>>>>>>>>> states.
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21694:
>>>>>>>>>>>>>> BigQuery Storage API insert with writeResult retry and write to 
>>>>>>>>>>>>>> error table
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21479:
>>>>>>>>>>>>>> Install Python wheel and dependencies to local venv in SDK 
>>>>>>>>>>>>>> harness
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21478:
>>>>>>>>>>>>>> KafkaIO.read.withDynamicRead() doesn't pick up new 
>>>>>>>>>>>>>> TopicPartitions
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21477: Add
>>>>>>>>>>>>>> integration testing for BQ Storage API  write modes
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21476:
>>>>>>>>>>>>>> WriteToBigQuery Dynamic table destinations returns wrong tableId
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21475: Beam
>>>>>>>>>>>>>> x-lang Dataflow tests failing due to _InactiveRpcError
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21473:
>>>>>>>>>>>>>> PVR_Spark2_Streaming perma-red
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21466:
>>>>>>>>>>>>>> Simplify version override for Dev versions of the Go SDK.
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21465: Kafka
>>>>>>>>>>>>>> commit offset drop data on failure for runners that have 
>>>>>>>>>>>>>> non-checkpointing
>>>>>>>>>>>>>> shuffle
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21269:
>>>>>>>>>>>>>> Delete orphaned files
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21268: Race
>>>>>>>>>>>>>> between member variable being accessed due to leaking 
>>>>>>>>>>>>>> uninitialized state
>>>>>>>>>>>>>> via OutboundObserverFactory
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21267:
>>>>>>>>>>>>>> WriteToBigQuery submits a duplicate BQ load job if a 503 error 
>>>>>>>>>>>>>> code is
>>>>>>>>>>>>>> returned from googleapi
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21265:
>>>>>>>>>>>>>> apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest.test_run_packable_combine_globally
>>>>>>>>>>>>>> 'apache_beam.coders.coder_impl._AbstractIterable' object is not 
>>>>>>>>>>>>>> reversible
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21263:
>>>>>>>>>>>>>> (Broken Pipe induced) Bricked Dataflow Pipeline
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21262:
>>>>>>>>>>>>>> Python AfterAny, AfterAll do not follow spec
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21260:
>>>>>>>>>>>>>> Python DirectRunner does not emit data at GC time
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21259:
>>>>>>>>>>>>>> Consumer group with random prefix
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21258:
>>>>>>>>>>>>>> Dataflow error in CombinePerKey operation
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21257:
>>>>>>>>>>>>>> Either Create or DirectRunner fails to produce all elements to 
>>>>>>>>>>>>>> the
>>>>>>>>>>>>>> following transform
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21123:
>>>>>>>>>>>>>> Multiple jobs running on Flink session cluster reuse the 
>>>>>>>>>>>>>> persistent Python
>>>>>>>>>>>>>> environment.
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21119:
>>>>>>>>>>>>>> Migrate to the next version of Python `requests` when released
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21117: "Java
>>>>>>>>>>>>>> IO IT Tests" - missing data in grafana
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21115:
>>>>>>>>>>>>>> JdbcIO date conversion is sensitive to OS
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21112:
>>>>>>>>>>>>>> Dataflow SocketException (SSLException) error while trying to 
>>>>>>>>>>>>>> send message
>>>>>>>>>>>>>> from Cloud Pub/Sub to BigQuery
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21111: Java
>>>>>>>>>>>>>> creates an incorrect pipeline proto when core-construction-java 
>>>>>>>>>>>>>> jar is not
>>>>>>>>>>>>>> in the CLASSPATH
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21110:
>>>>>>>>>>>>>> codecov/patch has poor behavior
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21109: SDF
>>>>>>>>>>>>>> BoundedSource seems to execute significantly slower than 'normal'
>>>>>>>>>>>>>> BoundedSource
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/21108:
>>>>>>>>>>>>>> java.io.InvalidClassException With Flink Kafka
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20979:
>>>>>>>>>>>>>> Portable runners should be able to issue checkpoints to 
>>>>>>>>>>>>>> Splittable DoFn
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20978:
>>>>>>>>>>>>>> PubsubIO.readAvroGenericRecord creates SchemaCoder that fails to 
>>>>>>>>>>>>>> decode
>>>>>>>>>>>>>> some Avro logical types
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20973:
>>>>>>>>>>>>>> Python Beam SDK Harness hangs when installing pip packages
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20818:
>>>>>>>>>>>>>> XmlIO.Read does not handle XML encoding per spec
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20814: JmsIO
>>>>>>>>>>>>>> is not acknowledging messages correctly
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20813: No
>>>>>>>>>>>>>> trigger early repeatedly for session windows
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20812:
>>>>>>>>>>>>>> Cross-language consistency (RequiresStableInputs) is quietly 
>>>>>>>>>>>>>> broken (at
>>>>>>>>>>>>>> least on portable flink runner)
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20692: Timer
>>>>>>>>>>>>>> with dataflow runner can be set multiple times (dataflow runner)
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20691: Beam
>>>>>>>>>>>>>> metrics should be displayed in Flink UI "Metrics" tab
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20689: Kafka
>>>>>>>>>>>>>> commitOffsetsInFinalize OOM on Flink
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20532:
>>>>>>>>>>>>>> Support for coder argument in WriteToBigQuery
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20531:
>>>>>>>>>>>>>> FileBasedSink: allow setting temp directory provider per dynamic 
>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20530: Make
>>>>>>>>>>>>>> non-portable Splittable DoFn the only option when executing Java 
>>>>>>>>>>>>>> "Read"
>>>>>>>>>>>>>> transforms
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20529:
>>>>>>>>>>>>>> SpannerIO tests don't actually assert anything.
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20528:
>>>>>>>>>>>>>> python CombineGlobally().with_fanout() cause duplicate combine 
>>>>>>>>>>>>>> results for
>>>>>>>>>>>>>> sliding windows
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20333:
>>>>>>>>>>>>>> beam_PerformanceTests_Kafka_IO failing due to " provided port is 
>>>>>>>>>>>>>> already
>>>>>>>>>>>>>> allocated"
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20332:
>>>>>>>>>>>>>> FileIO writeDynamic with AvroIO.sink not writing all data
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20330:
>>>>>>>>>>>>>> Remove insecure ssl options from MongoDBIO
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20109:
>>>>>>>>>>>>>> SortValues should fail if SecondaryKey coder is not deterministic
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20108:
>>>>>>>>>>>>>> Python direct runner doesn't emit empty pane when it should
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/20009:
>>>>>>>>>>>>>> Environment-sensitive provisioning for Dataflow
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19971: [SQL]
>>>>>>>>>>>>>> Some Hive tests throw NullPointerException, but get marked as 
>>>>>>>>>>>>>> passing
>>>>>>>>>>>>>> (Direct Runner)
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19817:
>>>>>>>>>>>>>> datetime and decimal should be logical types
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19815: Add
>>>>>>>>>>>>>> support for remaining data types in python RowCoder
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19813:
>>>>>>>>>>>>>> PubsubIO returns empty message bodies for all messages read
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19556: User
>>>>>>>>>>>>>> reports protobuf ClassChangeError running against 2.6.0 or above
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/19369:
>>>>>>>>>>>>>> KafkaIO doesn't commit offsets while being used as bounded source
>>>>>>>>>>>>>> https://api.github.com/repos/apache/beam/issues/17950:
>>>>>>>>>>>>>> [Bug]: Java Precommit permared
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>
>

Reply via email to