Thanks, Danny!

> On 24 Jun 2022, at 19:23, Danny McCormick <dannymccorm...@google.com> wrote:
> 
> Sure, I put up a fix - https://github.com/apache/beam/pull/22048 
> <https://github.com/apache/beam/pull/22048>
> On Fri, Jun 24, 2022 at 1:20 PM Alexey Romanenko <aromanenko....@gmail.com 
> <mailto:aromanenko....@gmail.com>> wrote:
> 
> 
>> > 2. The links in this report start with api.github.* and don’t take us 
>> > directly to the issues.
>> 
>> > Yeah Danny pointed that out as well. I'm assuming he knows how to fix it?
>> 
>> This is already fixed - Pablo actually beat me to it! 
>> <https://github.com/apache/beam/pull/22033>
> It adds also a colon after URL and some mail clients consider it as a part of 
> URL which leads to a broken link.
> Should we just remove a colon there or add a space between?
> 
> —
> Alexey
> 
>> 
>> Thanks,
>> Danny
>> 
>> On Thu, Jun 23, 2022 at 8:30 PM Brian Hulette <bhule...@google.com 
>> <mailto:bhule...@google.com>> wrote:
>> +1 for that proposal!
>> 
>> > 1. P2 and P3 issues should be noticed and resolved as well. Shall we have 
>> > a longer time window for the rest of not triaged or stagnate issues and 
>> > include them?
>> 
>> I worry these lists would get _very_ long and wouldn't be actionable. But 
>> maybe it's worth reporting something like "There are 376 P2's with no update 
>> in the last 6 months" with a link to a query?
>> 
>> > 2. The links in this report start with api.github.* and don’t take us 
>> > directly to the issues.
>> 
>> Yeah Danny pointed that out as well. I'm assuming he knows how to fix it?
>> 
>> On Thu, Jun 23, 2022 at 2:37 PM Pablo Estrada <pabl...@google.com 
>> <mailto:pabl...@google.com>> wrote:
>> Thanks. I like the proposal, and I've found the emails useful.
>> Best
>> -P.
>> 
>> On Thu, Jun 23, 2022 at 2:33 PM Manu Zhang <owenzhang1...@gmail.com 
>> <mailto:owenzhang1...@gmail.com>> wrote:
>> Sounds good! It’s like our internal reports of JIRA tickets exceeding SLA 
>> time and having no response from engineers.  We either resolve them or 
>> downgrade the priority to extend time window.
>> 
>> Besides,
>> 1. P2 and P3 issues should be noticed and resolved as well. Shall we have a 
>> longer time window for the rest of not triaged or stagnate issues and 
>> include them?
>> 2. The links in this report start with api.github.* and don’t take us 
>> directly to the issues.
>> 
>> 
>> Danny McCormick <dannymccorm...@google.com 
>> <mailto:dannymccorm...@google.com>>于2022年6月24日 周五04:48写道:
>> That generally sounds right to me - I also would vote that we consolidate to 
>> 1 email and stop distinguishing between flaky P1s and normal P1s.
>> 
>> So the single daily report would be:
>> 
>> - Unassigned P0s
>> - P0s with no update in the last 36 hours
>> - Unassigned P1s
>> - P1s with no update in the last 7 days
>> 
>> I think that will generate a pretty good list of issues that require some 
>> kind of action.
>> 
>> On Thu, Jun 23, 2022 at 4:43 PM Kenneth Knowles <k...@apache.org 
>> <mailto:k...@apache.org>> wrote:
>> Sounds good to me. Perhaps P0s > 36 hours ago (presumably they are more like 
>> ~hours for true outages of CI/website/etc) and P1s > 7 days?
>> 
>> On Thu, Jun 23, 2022 at 1:27 PM Brian Hulette <bhule...@google.com 
>> <mailto:bhule...@google.com>> wrote:
>> I think that Danny's alternate proposal (a daily email that show only issues 
>> last updated >7 days ago, and those with no assignee) fits well with the two 
>> goals you describe, if we include "triage needed" issues in the latter 
>> category. Maybe we also explicitly separate these two concerns in the report?
>> 
>> 
>> On Thu, Jun 23, 2022 at 1:14 PM Kenneth Knowles <k...@apache.org 
>> <mailto:k...@apache.org>> wrote:
>> Forking thread because lots of people may just ignore this topic, per the 
>> discussion :-)
>> 
>> (sometimes gmail doesn't fork thread properly, but here's hoping...)
>> 
>> I'll add some other outcomes of these emails:
>> 
>>  - people file P0s that are not outages and P1s that are not data loss and I 
>> downgrade them
>>  - I randomly open up a few flaky test bugs and see if I can fix them really 
>> quick
>>  - people file legit P0s and P1s and I subscribe and follow them
>> 
>> Of these, only the last one seems important (not just that *I* follow them, 
>> but that new P0s and P1s get immediate attention from many eyes)
>> 
>> So maybe one take on the goal is to:
>> 
>>  - have new P0s and P1s evaluated quickly: P0s are an outage or outage-like 
>> occurrence that needs immediate remedy, and P1s need to be evaluated for 
>> release blocking, etc.
>>  - make sure P0s and P1s get attention appropriate to their priority
>> 
>> It can also be helpful to just state the failure modes which would happen by 
>> default if we don't have a good process or automation:
>> 
>>  - Real P0 gets filed and not noticed or fixed in a timely manner, blocking 
>> users and/or community in real time
>>  - Real P1 gets filed and not noticed, so release goes out with known data 
>> loss bug or other total loss of functionality
>>  - Non-real P0s and P1s accumulate, throwing off our data and making it hard 
>> to find the real problems
>>  - Flakes are never fixed
>> 
>> WDYT?
>> 
>> If we have P0s and P1s in the "awaiting triage" state, those are the ones we 
>> need to notice. Then for a P0 or P1 outside of that state, we just need some 
>> way of making sure it doesn't stagnate. Or if it does stagnate, that 
>> empirically demonstrates it isn't really P1 (just like our P2 to P3 
>> downgrade automation). If everything is P1, nothing is, as they say.
>> 
>> Kenn
>> 
>> On Thu, Jun 23, 2022 at 10:01 AM Danny McCormick <dannymccorm...@google.com 
>> <mailto:dannymccorm...@google.com>> wrote:
>> > Maybe it would be helpful to sort these by last update time (and 
>> > potentially include that information in the email). Then we can at least 
>> > prioritize them instead of looking at a big wall of issues.
>> 
>> I agree that this is a good idea (and pretty trivial to do). I'll update the 
>> automation to do that once we get consensus on an approach.
>> 
>> > I think the motivation for daily emails is that per the priorities guide 
>> > [1] P1 issues should be getting "continuous status updates". If these 
>> > issues aren't actually that important, I think the noise is good as it 
>> > should motivate us to prioritize them correctly. In practice that hasn't 
>> > been happening though...
>> 
>> I guess the questions here are:
>> 
>> 1) What is the goal of this email?
>> 2) Is it effective at accomplishing that goal.
>> 
>> I think you're saying that the goal (or a goal) is to highlight issues that 
>> aren't getting the attention they need; if that's our goal, then I don't 
>> think this is a particularly effective mechanism for it because (a) its very 
>> unclear which issues fall into that category and (b) there are too many to 
>> manually go through on a daily basis. From the email alone, it's not clear 
>> to me that any of the issues above "shouldn't" be P1s (though I'd guess 
>> you're right that some/many of them don't belong since most were created 
>> before the Jira -> GH migration based on the titles). I'd also argue that a 
>> daily email just desensitizes us to them since there almost always will be 
>> some valid P1s that don't need extra attention.
>> 
>> I do still think this could have value as a weekly email, with the goal 
>> being "it's probably a good idea for someone to take a look at each of 
>> these". Another option would be to only include issues with no action in the 
>> last 7 days and/or no assignees and keep it daily.
>> 
>> A couple side notes:
>> - No matter what we do, if we keep the current automation in any form we 
>> should fix the url from https://api.github.com/repos/apache/beam/issues/# 
>> <https://api.github.com/repos/apache/beam/issues/#> to 
>> https://github.com/apache/beam/issues/# 
>> <https://github.com/apache/beam/issues/#> - the current links are very 
>> annoying.
>> - After I send this, I will do a pass of the current P1s since it does 
>> indeed seem like too many are P1s and many should actually be P2s (or lower).
>> 
>> Thanks,
>> Danny
>> 
>> On Thu, Jun 23, 2022 at 12:21 PM Brian Hulette <bhule...@google.com 
>> <mailto:bhule...@google.com>> wrote:
>> I think the motivation for daily emails is that per the priorities guide [1] 
>> P1 issues should be getting "continuous status updates". If these issues 
>> aren't actually that important, I think the noise is good as it should 
>> motivate us to prioritize them correctly. In practice that hasn't been 
>> happening though...
>> 
>> Maybe it would be helpful to sort these by last update time (and potentially 
>> include that information in the email). Then we can at least prioritize them 
>> instead of looking at a big wall of issues.
>> 
>> Brian
>> 
>> [1] https://beam.apache.org/contribute/issue-priorities/ 
>> <https://beam.apache.org/contribute/issue-priorities/>
>> On Thu, Jun 23, 2022 at 6:07 AM Danny McCormick <dannymccorm...@google.com 
>> <mailto:dannymccorm...@google.com>> wrote:
>> I think a weekly summary seems like a good idea for the P1 issues and flaky 
>> tests, though daily still seems appropriate for P0 issues. I put up 
>> https://github.com/apache/beam/pull/22017 
>> <https://github.com/apache/beam/pull/22017> to just send the P1/flaky test 
>> reports on Wednesdays, if anyone objects please let me know - I'll wait on 
>> merging til tomorrow to leave time for feedback (and it's always reversible 
>> 🙂).
>> 
>> Thanks,
>> Danny
>> 
>> On Wed, Jun 22, 2022 at 7:05 PM Manu Zhang <owenzhang1...@gmail.com 
>> <mailto:owenzhang1...@gmail.com>> wrote:
>> Hi all,
>> 
>> what is this daily summary intended for? Not all issues look like P1. And 
>> will a weekly summary be less noise?
>> 
>> <beamacti...@gmail.com <mailto:beamacti...@gmail.com>>于2022年6月22日 周三23:45写道:
>> This is your daily summary of Beam's current P1 issues, not including flaky 
>> tests.
>> 
>>     See https://beam.apache.org/contribute/issue-priorities/#p1-critical 
>> <https://beam.apache.org/contribute/issue-priorities/#p1-critical> for the 
>> meaning and expectations around P1 issues.
>> 
>> 
>> 
>> https://api.github.com/repos/apache/beam/issues/21978 
>> <https://api.github.com/repos/apache/beam/issues/21978>: [Playground] 
>> Implement Share Any Code feature on the frontend
>> https://api.github.com/repos/apache/beam/issues/21946 
>> <https://api.github.com/repos/apache/beam/issues/21946>: [Bug]: No way to 
>> read or write to file when running Beam in Flink
>> https://api.github.com/repos/apache/beam/issues/21935 
>> <https://api.github.com/repos/apache/beam/issues/21935>: [Bug]: Reject 
>> illformed GBK Coders
>> https://api.github.com/repos/apache/beam/issues/21897 
>> <https://api.github.com/repos/apache/beam/issues/21897>: [Feature Request]: 
>> Flink runner savepoint backward compatibility 
>> https://api.github.com/repos/apache/beam/issues/21893 
>> <https://api.github.com/repos/apache/beam/issues/21893>: [Bug]: BigQuery 
>> Storage Write API implementation does not support table partitioning
>> https://api.github.com/repos/apache/beam/issues/21794 
>> <https://api.github.com/repos/apache/beam/issues/21794>: Dataflow runner 
>> creates a new timer whenever the output timestamp is change
>> https://api.github.com/repos/apache/beam/issues/21763 
>> <https://api.github.com/repos/apache/beam/issues/21763>: [Playground Task]: 
>> Migrate from Google Analytics to Matomo Cloud
>> https://api.github.com/repos/apache/beam/issues/21715 
>> <https://api.github.com/repos/apache/beam/issues/21715>: Data missing when 
>> using CassandraIO.Read
>> https://api.github.com/repos/apache/beam/issues/21713 
>> <https://api.github.com/repos/apache/beam/issues/21713>: 404s in BigQueryIO 
>> don't get output to Failed Inserts PCollection
>> https://api.github.com/repos/apache/beam/issues/21711 
>> <https://api.github.com/repos/apache/beam/issues/21711>: Python Streaming 
>> job failing to drain with BigQueryIO write errors
>> https://api.github.com/repos/apache/beam/issues/21703 
>> <https://api.github.com/repos/apache/beam/issues/21703>: 
>> pubsublite.ReadWriteIT failing in beam_PostCommit_Java_DataflowV1 and V2
>> https://api.github.com/repos/apache/beam/issues/21702 
>> <https://api.github.com/repos/apache/beam/issues/21702>: SpannerWriteIT 
>> failing in beam PostCommit Java V1
>> https://api.github.com/repos/apache/beam/issues/21700 
>> <https://api.github.com/repos/apache/beam/issues/21700>: 
>> --dataflowServiceOptions=use_runner_v2 is broken
>> https://api.github.com/repos/apache/beam/issues/21695 
>> <https://api.github.com/repos/apache/beam/issues/21695>: 
>> DataflowPipelineResult does not raise exception for unsuccessful states.
>> https://api.github.com/repos/apache/beam/issues/21694 
>> <https://api.github.com/repos/apache/beam/issues/21694>: BigQuery Storage 
>> API insert with writeResult retry and write to error table
>> https://api.github.com/repos/apache/beam/issues/21479 
>> <https://api.github.com/repos/apache/beam/issues/21479>: Install Python 
>> wheel and dependencies to local venv in SDK harness
>> https://api.github.com/repos/apache/beam/issues/21478 
>> <https://api.github.com/repos/apache/beam/issues/21478>: 
>> KafkaIO.read.withDynamicRead() doesn't pick up new TopicPartitions
>> https://api.github.com/repos/apache/beam/issues/21477 
>> <https://api.github.com/repos/apache/beam/issues/21477>: Add integration 
>> testing for BQ Storage API  write modes
>> https://api.github.com/repos/apache/beam/issues/21476 
>> <https://api.github.com/repos/apache/beam/issues/21476>: WriteToBigQuery 
>> Dynamic table destinations returns wrong tableId
>> https://api.github.com/repos/apache/beam/issues/21475 
>> <https://api.github.com/repos/apache/beam/issues/21475>: Beam x-lang 
>> Dataflow tests failing due to _InactiveRpcError
>> https://api.github.com/repos/apache/beam/issues/21473 
>> <https://api.github.com/repos/apache/beam/issues/21473>: 
>> PVR_Spark2_Streaming perma-red
>> https://api.github.com/repos/apache/beam/issues/21466 
>> <https://api.github.com/repos/apache/beam/issues/21466>: Simplify version 
>> override for Dev versions of the Go SDK.
>> https://api.github.com/repos/apache/beam/issues/21465 
>> <https://api.github.com/repos/apache/beam/issues/21465>: Kafka commit offset 
>> drop data on failure for runners that have non-checkpointing shuffle
>> https://api.github.com/repos/apache/beam/issues/21269 
>> <https://api.github.com/repos/apache/beam/issues/21269>: Delete orphaned 
>> files
>> https://api.github.com/repos/apache/beam/issues/21268 
>> <https://api.github.com/repos/apache/beam/issues/21268>: Race between member 
>> variable being accessed due to leaking uninitialized state via 
>> OutboundObserverFactory
>> https://api.github.com/repos/apache/beam/issues/21267 
>> <https://api.github.com/repos/apache/beam/issues/21267>: WriteToBigQuery 
>> submits a duplicate BQ load job if a 503 error code is returned from 
>> googleapi
>> https://api.github.com/repos/apache/beam/issues/21265 
>> <https://api.github.com/repos/apache/beam/issues/21265>: 
>> apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest.test_run_packable_combine_globally
>>  'apache_beam.coders.coder_impl._AbstractIterable' object is not reversible
>> https://api.github.com/repos/apache/beam/issues/21263 
>> <https://api.github.com/repos/apache/beam/issues/21263>: (Broken Pipe 
>> induced) Bricked Dataflow Pipeline 
>> https://api.github.com/repos/apache/beam/issues/21262 
>> <https://api.github.com/repos/apache/beam/issues/21262>: Python AfterAny, 
>> AfterAll do not follow spec
>> https://api.github.com/repos/apache/beam/issues/21260 
>> <https://api.github.com/repos/apache/beam/issues/21260>: Python DirectRunner 
>> does not emit data at GC time
>> https://api.github.com/repos/apache/beam/issues/21259 
>> <https://api.github.com/repos/apache/beam/issues/21259>: Consumer group with 
>> random prefix
>> https://api.github.com/repos/apache/beam/issues/21258 
>> <https://api.github.com/repos/apache/beam/issues/21258>: Dataflow error in 
>> CombinePerKey operation
>> https://api.github.com/repos/apache/beam/issues/21257 
>> <https://api.github.com/repos/apache/beam/issues/21257>: Either Create or 
>> DirectRunner fails to produce all elements to the following transform
>> https://api.github.com/repos/apache/beam/issues/21123 
>> <https://api.github.com/repos/apache/beam/issues/21123>: Multiple jobs 
>> running on Flink session cluster reuse the persistent Python environment.
>> https://api.github.com/repos/apache/beam/issues/21119 
>> <https://api.github.com/repos/apache/beam/issues/21119>: Migrate to the next 
>> version of Python `requests` when released
>> https://api.github.com/repos/apache/beam/issues/21117 
>> <https://api.github.com/repos/apache/beam/issues/21117>: "Java IO IT Tests" 
>> - missing data in grafana
>> https://api.github.com/repos/apache/beam/issues/21115 
>> <https://api.github.com/repos/apache/beam/issues/21115>: JdbcIO date 
>> conversion is sensitive to OS
>> https://api.github.com/repos/apache/beam/issues/21112 
>> <https://api.github.com/repos/apache/beam/issues/21112>: Dataflow 
>> SocketException (SSLException) error while trying to send message from Cloud 
>> Pub/Sub to BigQuery
>> https://api.github.com/repos/apache/beam/issues/21111 
>> <https://api.github.com/repos/apache/beam/issues/21111>: Java creates an 
>> incorrect pipeline proto when core-construction-java jar is not in the 
>> CLASSPATH
>> https://api.github.com/repos/apache/beam/issues/21110 
>> <https://api.github.com/repos/apache/beam/issues/21110>: codecov/patch has 
>> poor behavior
>> https://api.github.com/repos/apache/beam/issues/21109 
>> <https://api.github.com/repos/apache/beam/issues/21109>: SDF BoundedSource 
>> seems to execute significantly slower than 'normal' BoundedSource
>> https://api.github.com/repos/apache/beam/issues/21108 
>> <https://api.github.com/repos/apache/beam/issues/21108>: 
>> java.io.InvalidClassException With Flink Kafka
>> https://api.github.com/repos/apache/beam/issues/20979 
>> <https://api.github.com/repos/apache/beam/issues/20979>: Portable runners 
>> should be able to issue checkpoints to Splittable DoFn
>> https://api.github.com/repos/apache/beam/issues/20978 
>> <https://api.github.com/repos/apache/beam/issues/20978>: 
>> PubsubIO.readAvroGenericRecord creates SchemaCoder that fails to decode some 
>> Avro logical types
>> https://api.github.com/repos/apache/beam/issues/20973 
>> <https://api.github.com/repos/apache/beam/issues/20973>: Python Beam SDK 
>> Harness hangs when installing pip packages
>> https://api.github.com/repos/apache/beam/issues/20818 
>> <https://api.github.com/repos/apache/beam/issues/20818>: XmlIO.Read does not 
>> handle XML encoding per spec
>> https://api.github.com/repos/apache/beam/issues/20814 
>> <https://api.github.com/repos/apache/beam/issues/20814>: JmsIO is not 
>> acknowledging messages correctly
>> https://api.github.com/repos/apache/beam/issues/20813 
>> <https://api.github.com/repos/apache/beam/issues/20813>: No trigger early 
>> repeatedly for session windows
>> https://api.github.com/repos/apache/beam/issues/20812 
>> <https://api.github.com/repos/apache/beam/issues/20812>: Cross-language 
>> consistency (RequiresStableInputs) is quietly broken (at least on portable 
>> flink runner)
>> https://api.github.com/repos/apache/beam/issues/20692 
>> <https://api.github.com/repos/apache/beam/issues/20692>: Timer with dataflow 
>> runner can be set multiple times (dataflow runner)
>> https://api.github.com/repos/apache/beam/issues/20691 
>> <https://api.github.com/repos/apache/beam/issues/20691>: Beam metrics should 
>> be displayed in Flink UI "Metrics" tab
>> https://api.github.com/repos/apache/beam/issues/20689 
>> <https://api.github.com/repos/apache/beam/issues/20689>: Kafka 
>> commitOffsetsInFinalize OOM on Flink
>> https://api.github.com/repos/apache/beam/issues/20532 
>> <https://api.github.com/repos/apache/beam/issues/20532>: Support for coder 
>> argument in WriteToBigQuery
>> https://api.github.com/repos/apache/beam/issues/20531 
>> <https://api.github.com/repos/apache/beam/issues/20531>: FileBasedSink: 
>> allow setting temp directory provider per dynamic destination
>> https://api.github.com/repos/apache/beam/issues/20530 
>> <https://api.github.com/repos/apache/beam/issues/20530>: Make non-portable 
>> Splittable DoFn the only option when executing Java "Read" transforms
>> https://api.github.com/repos/apache/beam/issues/20529 
>> <https://api.github.com/repos/apache/beam/issues/20529>: SpannerIO tests 
>> don't actually assert anything.
>> https://api.github.com/repos/apache/beam/issues/20528 
>> <https://api.github.com/repos/apache/beam/issues/20528>: python 
>> CombineGlobally().with_fanout() cause duplicate combine results for sliding 
>> windows
>> https://api.github.com/repos/apache/beam/issues/20333 
>> <https://api.github.com/repos/apache/beam/issues/20333>: 
>> beam_PerformanceTests_Kafka_IO failing due to " provided port is already 
>> allocated"
>> https://api.github.com/repos/apache/beam/issues/20332 
>> <https://api.github.com/repos/apache/beam/issues/20332>: FileIO writeDynamic 
>> with AvroIO.sink not writing all data
>> https://api.github.com/repos/apache/beam/issues/20330 
>> <https://api.github.com/repos/apache/beam/issues/20330>: Remove insecure ssl 
>> options from MongoDBIO
>> https://api.github.com/repos/apache/beam/issues/20109 
>> <https://api.github.com/repos/apache/beam/issues/20109>: SortValues should 
>> fail if SecondaryKey coder is not deterministic
>> https://api.github.com/repos/apache/beam/issues/20108 
>> <https://api.github.com/repos/apache/beam/issues/20108>: Python direct 
>> runner doesn't emit empty pane when it should
>> https://api.github.com/repos/apache/beam/issues/20009 
>> <https://api.github.com/repos/apache/beam/issues/20009>: 
>> Environment-sensitive provisioning for Dataflow
>> https://api.github.com/repos/apache/beam/issues/19971 
>> <https://api.github.com/repos/apache/beam/issues/19971>: [SQL] Some Hive 
>> tests throw NullPointerException, but get marked as passing (Direct Runner)
>> https://api.github.com/repos/apache/beam/issues/19817 
>> <https://api.github.com/repos/apache/beam/issues/19817>: datetime and 
>> decimal should be logical types
>> https://api.github.com/repos/apache/beam/issues/19815 
>> <https://api.github.com/repos/apache/beam/issues/19815>: Add support for 
>> remaining data types in python RowCoder 
>> https://api.github.com/repos/apache/beam/issues/19813 
>> <https://api.github.com/repos/apache/beam/issues/19813>: PubsubIO returns 
>> empty message bodies for all messages read
>> https://api.github.com/repos/apache/beam/issues/19556 
>> <https://api.github.com/repos/apache/beam/issues/19556>: User reports 
>> protobuf ClassChangeError running against 2.6.0 or above
>> https://api.github.com/repos/apache/beam/issues/19369 
>> <https://api.github.com/repos/apache/beam/issues/19369>: KafkaIO doesn't 
>> commit offsets while being used as bounded source
>> https://api.github.com/repos/apache/beam/issues/17950 
>> <https://api.github.com/repos/apache/beam/issues/17950>: [Bug]: Java 
>> Precommit permared
> 

Reply via email to