Oh, it is the build artifacts, not the jars...

Most of our build artifacts are build logs, but maybe the problem is that
some of the logs are very large if the test hangs...

张铎(Duo Zhang) <palomino...@gmail.com> 于2019年6月11日周二 上午8:16写道:

> For flakey we just need the commit id in the console output then we can
> build the artifacts locally. +1 on removing artifacts caching.
>
> Josh Elser <els...@apache.org> 于2019年6月11日周二 上午7:50写道:
>
>> Sure, Misty. No arguments here.
>>
>> I think that might be a bigger untangling. Maybe Peter or Busbey know
>> better about how these could be de-coupled (e.g. I think flakies
>> actually look back at old artifacts), but I'm not sure off the top of my
>> head. I was just going for a quick fix to keep Infra from doing
>> something super-destructive.
>>
>> For context, I've dropped them a note in Slack to make sure what I'm
>> doing is having a positive effect.
>>
>> On 6/10/19 7:34 PM, Misty Linville wrote:
>> > Keeping artifacts and keeping build logs are two separate things. I
>> don’t
>> > see a need to keep any artifacts past the most recent green and most
>> recent
>> > red builds. Alternately if we need the artifacts let’s have Jenkins put
>> > them somewhere rather than keeping them there. You can get back to
>> whatever
>> > hash you need within git to reproduce a build problem.
>> >
>> > On Mon, Jun 10, 2019 at 2:26 PM Josh Elser <els...@apache.org> wrote:
>> >
>> >> https://issues.apache.org/jira/browse/HBASE-22563 for a quick bandaid
>> (I
>> >> hope).
>> >>
>> >> On 6/10/19 4:31 PM, Josh Elser wrote:
>> >>> Eyes on.
>> >>>
>> >>> Looking at master, we already have the linked configuration, set to
>> >>> retain 30 builds.
>> >>>
>> >>> We have some extra branches which we can lop off (branch-1.2,
>> >>> branch-2.0, maybe some feature branches too). A quick fix might be to
>> >>> just pull back that 30 to 10.
>> >>>
>> >>> Largely figuring out how this stuff works now, give me a shout in
>> Slack
>> >>> if anyone else has cycles.
>> >>>
>> >>> On 6/10/19 2:34 PM, Peter Somogyi wrote:
>> >>>> Hi,
>> >>>>
>> >>>> HBase jobs are using more than 400GB based on this list.
>> >>>> Could someone take a look at the job configurations today?
>> Otherwise, I
>> >>>> will look into it tomorrow morning.
>> >>>>
>> >>>> Thanks,
>> >>>> Peter
>> >>>>
>> >>>> ---------- Forwarded message ---------
>> >>>> From: Chris Lambertus <c...@apache.org>
>> >>>> Date: Mon, Jun 10, 2019 at 7:57 PM
>> >>>> Subject: ACTION REQUIRED: disk space on jenkins master nearly full
>> >>>> To: <bui...@apache.org>
>> >>>> Cc: <d...@mesos.apache.org>, <d...@pulsar.apache.org>
>> >>>>
>> >>>>
>> >>>> Hello,
>> >>>>
>> >>>> The jenkins master is nearly full.
>> >>>>
>> >>>> The workspaces listed below need significant size reduction within 24
>> >>>> hours
>> >>>> or Infra will need to perform some manual pruning of old builds to
>> >>>> keep the
>> >>>> jenkins system running. The Mesos “Packaging” job also needs to be
>> >>>> corrected to include the project name (mesos-packaging) please.
>> >>>>
>> >>>> It appears that the typical ‘Discard Old Builds’ checkbox in the job
>> >>>> configuration may not be working for multibranch pipeline jobs.
>> Please
>> >>>> refer to these articles for information on discarding builds in
>> >>>> multibranch
>> >>>> jobs:
>> >>>>
>> >>>>
>> >>
>> https://support.cloudbees.com/hc/en-us/articles/115000237071-How-do-I-set-discard-old-builds-for-a-Multi-Branch-Pipeline-Job-
>> >>>>
>> >>>> https://issues.jenkins-ci.org/browse/JENKINS-35642
>> >>>>
>> >>
>> https://issues.jenkins-ci.org/browse/JENKINS-34738?focusedCommentId=263489&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-263489
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> NB: I have not fully vetted the above information, I just notice that
>> >>>> many
>> >>>> of these jobs have ‘Discard old builds’ checked, but it is clearly
>> not
>> >>>> working.
>> >>>>
>> >>>>
>> >>>> If you are unable to reduce your disk usage beyond what is listed,
>> >> please
>> >>>> let me know what the reasons are and we’ll see if we can find a
>> >> solution.
>> >>>> If you believe you’ve configured your job properly and the space
>> usage
>> >> is
>> >>>> more than you expect, please comment here and we’ll take a look at
>> what
>> >>>> might be going on.
>> >>>>
>> >>>> I cut this list off arbitrarily at 40GB workspaces and larger. There
>> are
>> >>>> many which are between 20 and 30GB which also need to be addressed,
>> but
>> >>>> these are the current top contributors to the disk space situation.
>> >>>>
>> >>>>
>> >>>> 594G    Packaging
>> >>>> 425G    pulsar-website-build
>> >>>> 274G    pulsar-master
>> >>>> 195G    hadoop-multibranch
>> >>>> 173G    HBase Nightly
>> >>>> 138G    HBase-Flaky-Tests
>> >>>> 119G    netbeans-release
>> >>>> 108G    Any23-trunk
>> >>>> 101G    netbeans-linux-experiment
>> >>>> 96G     Jackrabbit-Oak-Windows
>> >>>> 94G     HBase-Find-Flaky-Tests
>> >>>> 88G     PreCommit-ZOOKEEPER-github-pr-build
>> >>>> 74G     netbeans-windows
>> >>>> 71G     stanbol-0.12
>> >>>> 68G     Sling
>> >>>> 63G     Atlas-master-NoTests
>> >>>> 48G     FlexJS Framework (maven)
>> >>>> 45G     HBase-PreCommit-GitHub-PR
>> >>>> 42G     pulsar-pull-request
>> >>>> 40G     Atlas-1.0-NoTests
>> >>>>
>> >>>>
>> >>>>
>> >>>> Thanks,
>> >>>> Chris
>> >>>> ASF Infra
>> >>>>
>> >>
>> >
>>
>

Reply via email to