For flakey we just need the commit id in the console output then we can
build the artifacts locally. +1 on removing artifacts caching.

Josh Elser <els...@apache.org> 于2019年6月11日周二 上午7:50写道:

> Sure, Misty. No arguments here.
>
> I think that might be a bigger untangling. Maybe Peter or Busbey know
> better about how these could be de-coupled (e.g. I think flakies
> actually look back at old artifacts), but I'm not sure off the top of my
> head. I was just going for a quick fix to keep Infra from doing
> something super-destructive.
>
> For context, I've dropped them a note in Slack to make sure what I'm
> doing is having a positive effect.
>
> On 6/10/19 7:34 PM, Misty Linville wrote:
> > Keeping artifacts and keeping build logs are two separate things. I don’t
> > see a need to keep any artifacts past the most recent green and most
> recent
> > red builds. Alternately if we need the artifacts let’s have Jenkins put
> > them somewhere rather than keeping them there. You can get back to
> whatever
> > hash you need within git to reproduce a build problem.
> >
> > On Mon, Jun 10, 2019 at 2:26 PM Josh Elser <els...@apache.org> wrote:
> >
> >> https://issues.apache.org/jira/browse/HBASE-22563 for a quick bandaid
> (I
> >> hope).
> >>
> >> On 6/10/19 4:31 PM, Josh Elser wrote:
> >>> Eyes on.
> >>>
> >>> Looking at master, we already have the linked configuration, set to
> >>> retain 30 builds.
> >>>
> >>> We have some extra branches which we can lop off (branch-1.2,
> >>> branch-2.0, maybe some feature branches too). A quick fix might be to
> >>> just pull back that 30 to 10.
> >>>
> >>> Largely figuring out how this stuff works now, give me a shout in Slack
> >>> if anyone else has cycles.
> >>>
> >>> On 6/10/19 2:34 PM, Peter Somogyi wrote:
> >>>> Hi,
> >>>>
> >>>> HBase jobs are using more than 400GB based on this list.
> >>>> Could someone take a look at the job configurations today? Otherwise,
> I
> >>>> will look into it tomorrow morning.
> >>>>
> >>>> Thanks,
> >>>> Peter
> >>>>
> >>>> ---------- Forwarded message ---------
> >>>> From: Chris Lambertus <c...@apache.org>
> >>>> Date: Mon, Jun 10, 2019 at 7:57 PM
> >>>> Subject: ACTION REQUIRED: disk space on jenkins master nearly full
> >>>> To: <bui...@apache.org>
> >>>> Cc: <d...@mesos.apache.org>, <d...@pulsar.apache.org>
> >>>>
> >>>>
> >>>> Hello,
> >>>>
> >>>> The jenkins master is nearly full.
> >>>>
> >>>> The workspaces listed below need significant size reduction within 24
> >>>> hours
> >>>> or Infra will need to perform some manual pruning of old builds to
> >>>> keep the
> >>>> jenkins system running. The Mesos “Packaging” job also needs to be
> >>>> corrected to include the project name (mesos-packaging) please.
> >>>>
> >>>> It appears that the typical ‘Discard Old Builds’ checkbox in the job
> >>>> configuration may not be working for multibranch pipeline jobs. Please
> >>>> refer to these articles for information on discarding builds in
> >>>> multibranch
> >>>> jobs:
> >>>>
> >>>>
> >>
> https://support.cloudbees.com/hc/en-us/articles/115000237071-How-do-I-set-discard-old-builds-for-a-Multi-Branch-Pipeline-Job-
> >>>>
> >>>> https://issues.jenkins-ci.org/browse/JENKINS-35642
> >>>>
> >>
> https://issues.jenkins-ci.org/browse/JENKINS-34738?focusedCommentId=263489&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-263489
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> NB: I have not fully vetted the above information, I just notice that
> >>>> many
> >>>> of these jobs have ‘Discard old builds’ checked, but it is clearly not
> >>>> working.
> >>>>
> >>>>
> >>>> If you are unable to reduce your disk usage beyond what is listed,
> >> please
> >>>> let me know what the reasons are and we’ll see if we can find a
> >> solution.
> >>>> If you believe you’ve configured your job properly and the space usage
> >> is
> >>>> more than you expect, please comment here and we’ll take a look at
> what
> >>>> might be going on.
> >>>>
> >>>> I cut this list off arbitrarily at 40GB workspaces and larger. There
> are
> >>>> many which are between 20 and 30GB which also need to be addressed,
> but
> >>>> these are the current top contributors to the disk space situation.
> >>>>
> >>>>
> >>>> 594G    Packaging
> >>>> 425G    pulsar-website-build
> >>>> 274G    pulsar-master
> >>>> 195G    hadoop-multibranch
> >>>> 173G    HBase Nightly
> >>>> 138G    HBase-Flaky-Tests
> >>>> 119G    netbeans-release
> >>>> 108G    Any23-trunk
> >>>> 101G    netbeans-linux-experiment
> >>>> 96G     Jackrabbit-Oak-Windows
> >>>> 94G     HBase-Find-Flaky-Tests
> >>>> 88G     PreCommit-ZOOKEEPER-github-pr-build
> >>>> 74G     netbeans-windows
> >>>> 71G     stanbol-0.12
> >>>> 68G     Sling
> >>>> 63G     Atlas-master-NoTests
> >>>> 48G     FlexJS Framework (maven)
> >>>> 45G     HBase-PreCommit-GitHub-PR
> >>>> 42G     pulsar-pull-request
> >>>> 40G     Atlas-1.0-NoTests
> >>>>
> >>>>
> >>>>
> >>>> Thanks,
> >>>> Chris
> >>>> ASF Infra
> >>>>
> >>
> >
>

Reply via email to