I wonder if we can add graph to community metrics showing ignored tests by
language/project/overall. That can be useful to see focus area.

On Tue, May 12, 2020 at 12:28 PM Jan Lukavský <je...@seznam.cz> wrote:

> +1, visualizing the number of ignored tests in a graph seems useful. Even
> better with some slices (e.g. per runner, module, ...).
> On 5/12/20 8:02 PM, Ahmet Altay wrote:
>
> +1 to generate a report instead of removing these tests. A report like
> this could help us with prioritization. It is easier to address issues when
> we can quantify how much of a problem it is.
>
> I am curious what we can do to incentivize reducing the number of
> flaky/ignored tests? A report itself might provide incentive, it is
> rewarding to see ignored tests numbers go down over time.
>
> On Mon, May 11, 2020 at 8:30 AM Luke Cwik <lc...@google.com> wrote:
>
>> Deleting ignored tests does lead us to losing the reason as to why the
>> test case was around so I would rather keep it around. I think it would be
>> more valuable to generate a report that goes on the website/wiki showing
>> stability of the modules (num tests, num passed, num skipped, num failed
>> (running averages over the past N runs)). We had discussed doing something
>> like this for ValidatesRunner so we could show which runner supports what
>> automatically.
>>
>> On Mon, May 11, 2020 at 12:53 AM Jan Lukavský <je...@seznam.cz> wrote:
>>
>>> I think that we do have Jira issues for ignored test, there should be no
>>> problem with that. The questionable point is that when test gets Ignored,
>>> people might consider the problem as "less painful" and postpone the
>>> correct solution until ... forever. I'd just like to discuss if people see
>>> this as an issue. If yes, should we do something about that, or if no,
>>> maybe we can create a rule that test marked as Ignored for long time might
>>> be deleted, because apparently is only a dead code.
>>> On 5/6/20 6:30 PM, Kenneth Knowles wrote:
>>>
>>> Good point.
>>>
>>> The raw numbers are available in the test run output. See
>>> https://builds.apache.org/view/A-D/view/Beam/view/PostCommit/job/beam_PreCommit_Java_Cron/2718/testReport/
>>>  for
>>> the "skipped" column.
>>> And you get the same on console or Gradle Scan:
>>> https://scans.gradle.com/s/ml3jv5xctkrmg/tests?collapse-all
>>> This would be good to review periodically for obvious trouble spots.
>>>
>>> But I think you mean something more detailed. Some report with columns:
>>> Test Suite, Test Method, Jira, Date Ignored, Most Recent Update
>>>
>>> I think we can get most of this from Jira, if we just make sure that
>>> each ignored test has a Jira and they are all labeled in a consistent way.
>>> That would be the quickest way to get some result, even though it is not
>>> perfectly automated and audited.
>>>
>>> Kenn
>>>
>>> On Tue, May 5, 2020 at 2:41 PM Jan Lukavský <je...@seznam.cz> wrote:
>>>
>>>> Hi,
>>>>
>>>> it seems we are accumulating test cases (see discussion in [1]) that
>>>> are
>>>> marked as @Ignored (mostly due to flakiness), which is generally
>>>> undesirable. Associated JIRAs seem to be open for a long time, and this
>>>> might generally cause that we loose code coverage. Would anyone have
>>>> idea on how to visualize these Ignored tests better? My first idea
>>>> would
>>>> be something similar to "Beam dependency check report", but that seems
>>>> to be not the best example (which is completely different issue :)).
>>>>
>>>> Jan
>>>>
>>>> [1] https://github.com/apache/beam/pull/11614
>>>>
>>>>

Reply via email to