Cool. If it is /home/jenkins it should be just fine. Thanks for checking!

Kenn

On Tue, Jul 28, 2020 at 10:23 AM Damian Gadomski <
damian.gadom...@polidea.com> wrote:

> Sorry, mistake while copying, [1] should be:
> [1]
> https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L63
>
>
> On Tue, Jul 28, 2020 at 7:21 PM Damian Gadomski <
> damian.gadom...@polidea.com> wrote:
>
>> That's interesting. I didn't check that myself but all the Jenkins jobs
>> are configured to wipe the workspace just before the actual build happens
>> [1]
>> <https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6>.
>> Git SCM plugin is used for that and it enables the option called "Wipe out
>> repository and force clone". Docs state that it "deletes the contents of
>> the workspace before build and before checkout" [2]
>> <https://plugins.jenkins.io/git/>. Therefore I assume that removing
>> workspace just after the build won't change anything.
>>
>> The ./.gradle/caches/modules-2/files-2.1 dir is indeed present on the
>> worker machines but it's rather in /home/jenkins dir.
>>
>> damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh
>> 11G .
>> damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh
>> caches/modules-2/files-2.1
>> 2.3G caches/modules-2/files-2.1
>>
>> I can't find that directory structure inside workspaces.
>>
>> damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$
>> sudo find -name "files-2.1"
>> damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$
>>
>> [1]
>> https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6
>> [2] https://plugins.jenkins.io/git/
>>
>> On Tue, Jul 28, 2020 at 5:47 PM Kenneth Knowles <k...@apache.org> wrote:
>>
>>> Just checking - will this wipe out dependency cache? That will slow
>>> things down and significantly increase flakiness. If I recall correctly,
>>> the default Jenkins layout was:
>>>
>>>     /home/jenkins/jenkins-slave/workspace/$jobname
>>>     /home/jenkins/jenkins-slave/workspace/$jobname/.m2
>>>     /home/jenkins/jenkins-slave/workspace/$jobname/.git
>>>
>>> Where you can see that it did a `git clone` right into the root
>>> workspace directory, adjacent to .m2. This was not hygienic. One important
>>> thing was that `git clean` would wipe the maven cache with every build. So
>>> in https://github.com/apache/beam/pull/3976 we changed it to
>>>
>>>     /home/jenkins/jenkins-slave/workspace/$jobname
>>>     /home/jenkins/jenkins-slave/workspace/$jobname/.m2
>>>     /home/jenkins/jenkins-slave/workspace/$jobname/src/.git
>>>
>>> Now the .m2 directory survives and we do not constantly see flakes
>>> re-downloading deps that are immutable. This does, of course, use disk
>>> space.
>>>
>>> That was in the maven days. Gradle is the same except for $HOME/.m2 is
>>> replaced by $HOME/.gradle/caches/modules-2/files-2.1. Is Jenkins configured
>>> the same way so we will be wiping out the dependencies? If so, can you
>>> address this issue? Everything in that directory should be immutable and
>>> just a cache to avoid pointless re-download.
>>>
>>> Kenn
>>>
>>> On Tue, Jul 28, 2020 at 2:25 AM Damian Gadomski <
>>> damian.gadom...@polidea.com> wrote:
>>>
>>>> Agree with Udi, workspaces seem to be the third culprit, not yet
>>>> addressed in any way (until PR#12326
>>>> <https://github.com/apache/beam/pull/12326> is merged). I feel that
>>>> it'll solve the issue of filling up the disks for a long time ;)
>>>>
>>>> I'm also OK with moving /tmp cleanup to option B, and will happily
>>>> investigate on proper TMPDIR config.
>>>>
>>>>
>>>>
>>>> On Tue, Jul 28, 2020 at 3:07 AM Udi Meiri <eh...@google.com> wrote:
>>>>
>>>>> What about the workspaces, which can take up 175GB in some cases (see
>>>>> above)?
>>>>> I'm working on getting them cleaned up automatically:
>>>>> https://github.com/apache/beam/pull/12326
>>>>>
>>>>> My opinion is that we would get more mileage out of fixing the jobs
>>>>> that leave behind files in /tmp and images/containers in Docker.
>>>>> This would also help keep development machines clean.
>>>>>
>>>>>
>>>>> On Mon, Jul 27, 2020 at 5:31 PM Tyson Hamilton <tyso...@google.com>
>>>>> wrote:
>>>>>
>>>>>> Here is a summery of how I understand things,
>>>>>>
>>>>>>   - /tmp and /var/lib/docker are the culprit for filling up disks
>>>>>>   - inventory Jenkins job runs every 12 hours and runs a docker prune
>>>>>> to clean up images older than 24hr
>>>>>>   - crontab on each machine cleans up /tmp files older than three
>>>>>> days weekly
>>>>>>
>>>>>> This doesn't seem to be working since we're still running out of disk
>>>>>> periodically and requiring manual intervention. Knobs and options we have
>>>>>> available:
>>>>>>
>>>>>>   1. increase frequency of deleting files
>>>>>>   2. decrease the number of days required to delete a file (e.g.
>>>>>> older than 2 days)
>>>>>>
>>>>>> The execution methods we have available are:
>>>>>>
>>>>>>   A. cron
>>>>>>     - pro: runs even if a job gets stuck in Jenkins due to full disk
>>>>>>     - con: config baked into VM which is tough to update, not
>>>>>> discoverable or documented well
>>>>>>   B. inventory job
>>>>>>     - pro: easy to update, runs every 12h already
>>>>>>     - con: could get stuck if Jenkins agent runs out of disk or is
>>>>>> otherwise stuck, tied to all other inventory job frequency
>>>>>>   C. configure startup scripts for the VMs that set up the cron job
>>>>>> anytime the VM is restarted
>>>>>>     - pro: similar to A. and easy to update
>>>>>>     - con: similar to A.
>>>>>>
>>>>>> Between the three I prefer B. because it is consistent with other
>>>>>> inventory jobs. If it ends up that stuck jobs prohibit scheduling of the
>>>>>> inventory job often we could further investigate C to avoid having to
>>>>>> rebuild the VM images repeatedly.
>>>>>>
>>>>>> Any objections or comments? If not, we'll go forward with B. and
>>>>>> reduce the date check from 3 days to 2 days.
>>>>>>
>>>>>>
>>>>>> On 2020/07/24 20:13:29, Ahmet Altay <al...@google.com> wrote:
>>>>>> > Tests may not be doing docker cleanup. Inventory job runs a docker
>>>>>> prune
>>>>>> > every 12 hours for images older than 24 hrs [1]. Randomly looking
>>>>>> at one of
>>>>>> > the recent runs [2], it cleaned up a long list of containers
>>>>>> consuming
>>>>>> > 30+GB space. That should be just 12 hours worth of containers.
>>>>>> >
>>>>>> > [1]
>>>>>> >
>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L69
>>>>>> > [2]
>>>>>> >
>>>>>> https://ci-beam.apache.org/job/beam_Inventory_apache-beam-jenkins-14/501/console
>>>>>> >
>>>>>> > On Fri, Jul 24, 2020 at 1:07 PM Tyson Hamilton <tyso...@google.com>
>>>>>> wrote:
>>>>>> >
>>>>>> > > Yes, these are on the same volume in the /var/lib/docker
>>>>>> directory. I'm
>>>>>> > > unsure if they clean up leftover images.
>>>>>> > >
>>>>>> > > On Fri, Jul 24, 2020 at 12:52 PM Udi Meiri <eh...@google.com>
>>>>>> wrote:
>>>>>> > >
>>>>>> > >> I forgot Docker images:
>>>>>> > >>
>>>>>> > >> ehudm@apache-ci-beam-jenkins-3:~$ sudo docker system df
>>>>>> > >> TYPE                TOTAL               ACTIVE              SIZE
>>>>>> > >>        RECLAIMABLE
>>>>>> > >> Images              88                  9
>>>>>>  125.4GB
>>>>>> > >>       124.2GB (99%)
>>>>>> > >> Containers          40                  4
>>>>>>  7.927GB
>>>>>> > >>       7.871GB (99%)
>>>>>> > >> Local Volumes       47                  0
>>>>>>  3.165GB
>>>>>> > >>       3.165GB (100%)
>>>>>> > >> Build Cache         0                   0                   0B
>>>>>> > >>        0B
>>>>>> > >>
>>>>>> > >> There are about 90 images on that machine, with all but 1 less
>>>>>> than 48
>>>>>> > >> hours old.
>>>>>> > >> I think the docker test jobs need to try harder at cleaning up
>>>>>> their
>>>>>> > >> leftover images. (assuming they're already doing it?)
>>>>>> > >>
>>>>>> > >> On Fri, Jul 24, 2020 at 12:31 PM Udi Meiri <eh...@google.com>
>>>>>> wrote:
>>>>>> > >>
>>>>>> > >>> The additional slots (@3 directories) take up even more space
>>>>>> now than
>>>>>> > >>> before.
>>>>>> > >>>
>>>>>> > >>> I'm testing out https://github.com/apache/beam/pull/12326
>>>>>> which could
>>>>>> > >>> help by cleaning up workspaces after a run (just started a seed
>>>>>> job).
>>>>>> > >>>
>>>>>> > >>> On Fri, Jul 24, 2020 at 12:13 PM Tyson Hamilton <
>>>>>> tyso...@google.com>
>>>>>> > >>> wrote:
>>>>>> > >>>
>>>>>> > >>>> 664M    beam_PreCommit_JavaPortabilityApi_Commit
>>>>>> > >>>> 656M    beam_PreCommit_JavaPortabilityApi_Commit@2
>>>>>> > >>>> 611M    beam_PreCommit_JavaPortabilityApi_Cron
>>>>>> > >>>> 616M    beam_PreCommit_JavaPortabilityApiJava11_Commit
>>>>>> > >>>> 598M    beam_PreCommit_JavaPortabilityApiJava11_Commit@2
>>>>>> > >>>> 662M    beam_PreCommit_JavaPortabilityApiJava11_Cron
>>>>>> > >>>> 2.9G    beam_PreCommit_Portable_Python_Commit
>>>>>> > >>>> 2.9G    beam_PreCommit_Portable_Python_Commit@2
>>>>>> > >>>> 1.7G    beam_PreCommit_Portable_Python_Commit@3
>>>>>> > >>>> 3.4G    beam_PreCommit_Portable_Python_Cron
>>>>>> > >>>> 1.9G    beam_PreCommit_Python2_PVR_Flink_Commit
>>>>>> > >>>> 1.4G    beam_PreCommit_Python2_PVR_Flink_Cron
>>>>>> > >>>> 1.3G    beam_PreCommit_Python2_PVR_Flink_Phrase
>>>>>> > >>>> 6.2G    beam_PreCommit_Python_Commit
>>>>>> > >>>> 7.5G    beam_PreCommit_Python_Commit@2
>>>>>> > >>>> 7.5G    beam_PreCommit_Python_Cron
>>>>>> > >>>> 1012M   beam_PreCommit_PythonDocker_Commit
>>>>>> > >>>> 1011M   beam_PreCommit_PythonDocker_Commit@2
>>>>>> > >>>> 1011M   beam_PreCommit_PythonDocker_Commit@3
>>>>>> > >>>> 1002M   beam_PreCommit_PythonDocker_Cron
>>>>>> > >>>> 877M    beam_PreCommit_PythonFormatter_Commit
>>>>>> > >>>> 988M    beam_PreCommit_PythonFormatter_Cron
>>>>>> > >>>> 986M    beam_PreCommit_PythonFormatter_Phrase
>>>>>> > >>>> 1.7G    beam_PreCommit_PythonLint_Commit
>>>>>> > >>>> 2.1G    beam_PreCommit_PythonLint_Cron
>>>>>> > >>>> 7.5G    beam_PreCommit_Python_Phrase
>>>>>> > >>>> 346M    beam_PreCommit_RAT_Commit
>>>>>> > >>>> 341M    beam_PreCommit_RAT_Cron
>>>>>> > >>>> 338M    beam_PreCommit_Spotless_Commit
>>>>>> > >>>> 339M    beam_PreCommit_Spotless_Cron
>>>>>> > >>>> 5.5G    beam_PreCommit_SQL_Commit
>>>>>> > >>>> 5.5G    beam_PreCommit_SQL_Cron
>>>>>> > >>>> 5.5G    beam_PreCommit_SQL_Java11_Commit
>>>>>> > >>>> 750M    beam_PreCommit_Website_Commit
>>>>>> > >>>> 750M    beam_PreCommit_Website_Commit@2
>>>>>> > >>>> 750M    beam_PreCommit_Website_Cron
>>>>>> > >>>> 764M    beam_PreCommit_Website_Stage_GCS_Commit
>>>>>> > >>>> 771M    beam_PreCommit_Website_Stage_GCS_Cron
>>>>>> > >>>> 336M    beam_Prober_CommunityMetrics
>>>>>> > >>>> 693M    beam_python_mongoio_load_test
>>>>>> > >>>> 339M    beam_SeedJob
>>>>>> > >>>> 333M    beam_SeedJob_Standalone
>>>>>> > >>>> 334M    beam_sonarqube_report
>>>>>> > >>>> 556M    beam_SQLBigQueryIO_Batch_Performance_Test_Java
>>>>>> > >>>> 175G    total
>>>>>> > >>>>
>>>>>> > >>>> On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <
>>>>>> tyso...@google.com>
>>>>>> > >>>> wrote:
>>>>>> > >>>>
>>>>>> > >>>>> Ya looks like something in the workspaces is taking up room:
>>>>>> > >>>>>
>>>>>> > >>>>> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc .
>>>>>> > >>>>> 191G    .
>>>>>> > >>>>> 191G    total
>>>>>> > >>>>>
>>>>>> > >>>>>
>>>>>> > >>>>> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <
>>>>>> tyso...@google.com>
>>>>>> > >>>>> wrote:
>>>>>> > >>>>>
>>>>>> > >>>>>> Node 8 is also full. The partition that /tmp is on is here:
>>>>>> > >>>>>>
>>>>>> > >>>>>> Filesystem      Size  Used Avail Use% Mounted on
>>>>>> > >>>>>> /dev/sda1       485G  482G  2.9G 100% /
>>>>>> > >>>>>>
>>>>>> > >>>>>> however after cleaning up tmp with the crontab command,
>>>>>> there is only
>>>>>> > >>>>>> 8G usage yet it still remains 100% full:
>>>>>> > >>>>>>
>>>>>> > >>>>>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp
>>>>>> > >>>>>> 8.0G    /tmp
>>>>>> > >>>>>> 8.0G    total
>>>>>> > >>>>>>
>>>>>> > >>>>>> The workspaces are in the
>>>>>> /home/jenkins/jenkins-slave/workspace
>>>>>> > >>>>>> directory. When I run a du on that, it takes really long.
>>>>>> I'll let it keep
>>>>>> > >>>>>> running for a while to see if it ever returns a result but
>>>>>> so far this
>>>>>> > >>>>>> seems suspect.
>>>>>> > >>>>>>
>>>>>> > >>>>>>
>>>>>> > >>>>>>
>>>>>> > >>>>>>
>>>>>> > >>>>>>
>>>>>> > >>>>>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <
>>>>>> tyso...@google.com>
>>>>>> > >>>>>> wrote:
>>>>>> > >>>>>>
>>>>>> > >>>>>>> Everything I've been looking at is in the /tmp dir. Where
>>>>>> are the
>>>>>> > >>>>>>> workspaces, or what are the named?
>>>>>> > >>>>>>>
>>>>>> > >>>>>>>
>>>>>> > >>>>>>>
>>>>>> > >>>>>>>
>>>>>> > >>>>>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <
>>>>>> eh...@google.com> wrote:
>>>>>> > >>>>>>>
>>>>>> > >>>>>>>> I'm curious to what you find. Was it /tmp or the
>>>>>> workspaces using
>>>>>> > >>>>>>>> up the space?
>>>>>> > >>>>>>>>
>>>>>> > >>>>>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <
>>>>>> tyso...@google.com>
>>>>>> > >>>>>>>> wrote:
>>>>>> > >>>>>>>>
>>>>>> > >>>>>>>>> Bleck. I just realized that it is 'offline' so that won't
>>>>>> work.
>>>>>> > >>>>>>>>> I'll clean up manually on the machine using the cron
>>>>>> command.
>>>>>> > >>>>>>>>>
>>>>>> > >>>>>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <
>>>>>> > >>>>>>>>> tyso...@google.com> wrote:
>>>>>> > >>>>>>>>>
>>>>>> > >>>>>>>>>> Something isn't working with the current set up because
>>>>>> node 15
>>>>>> > >>>>>>>>>> appears to be out of space and is currently 'offline'
>>>>>> according to Jenkins.
>>>>>> > >>>>>>>>>> Can someone run the cleanup job? The machine is full,
>>>>>> > >>>>>>>>>>
>>>>>> > >>>>>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h
>>>>>> > >>>>>>>>>> Filesystem      Size  Used Avail Use% Mounted on
>>>>>> > >>>>>>>>>> udev             52G     0   52G   0% /dev
>>>>>> > >>>>>>>>>> tmpfs            11G  265M   10G   3% /run
>>>>>> > >>>>>>>>>> */dev/sda1       485G  484G  880M 100% /*
>>>>>> > >>>>>>>>>> tmpfs            52G     0   52G   0% /dev/shm
>>>>>> > >>>>>>>>>> tmpfs           5.0M     0  5.0M   0% /run/lock
>>>>>> > >>>>>>>>>> tmpfs            52G     0   52G   0% /sys/fs/cgroup
>>>>>> > >>>>>>>>>> tmpfs            11G     0   11G   0% /run/user/1017
>>>>>> > >>>>>>>>>> tmpfs            11G     0   11G   0% /run/user/1037
>>>>>> > >>>>>>>>>>
>>>>>> > >>>>>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . |
>>>>>> sort -rhk
>>>>>> > >>>>>>>>>> 1,1 | head -n 20
>>>>>> > >>>>>>>>>> 20G     2020-07-24 17:52        .
>>>>>> > >>>>>>>>>> 580M    2020-07-22 17:31
>>>>>> ./junit1031982597110125586
>>>>>> > >>>>>>>>>> 517M    2020-07-22 17:31
>>>>>> > >>>>>>>>>>
>>>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof
>>>>>> > >>>>>>>>>> 517M    2020-07-22 17:31
>>>>>> > >>>>>>>>>>  ./junit1031982597110125586/junit8739924829337821410
>>>>>> > >>>>>>>>>> 263M    2020-07-22 12:23        ./pip-install-2GUhO_
>>>>>> > >>>>>>>>>> 263M    2020-07-20 09:30        ./pip-install-sxgwqr
>>>>>> > >>>>>>>>>> 263M    2020-07-17 13:56        ./pip-install-bWSKIV
>>>>>> > >>>>>>>>>> 242M    2020-07-21 20:25
>>>>>> ./beam-pipeline-tempmByU6T
>>>>>> > >>>>>>>>>> 242M    2020-07-21 20:21
>>>>>> ./beam-pipeline-tempV85xeK
>>>>>> > >>>>>>>>>> 242M    2020-07-21 20:15
>>>>>> ./beam-pipeline-temp7dJROJ
>>>>>> > >>>>>>>>>> 236M    2020-07-21 20:25
>>>>>> > >>>>>>>>>>  ./beam-pipeline-tempmByU6T/tmpOWj3Yr
>>>>>> > >>>>>>>>>> 236M    2020-07-21 20:21
>>>>>> > >>>>>>>>>>  ./beam-pipeline-tempV85xeK/tmppbQHB3
>>>>>> > >>>>>>>>>> 236M    2020-07-21 20:15
>>>>>> > >>>>>>>>>>  ./beam-pipeline-temp7dJROJ/tmpgOXPKW
>>>>>> > >>>>>>>>>> 111M    2020-07-23 00:57        ./pip-install-1JnyNE
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:17
>>>>>> ./beam-artifact1374651823280819755
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:16
>>>>>> ./beam-artifact5050755582921936972
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:16
>>>>>> ./beam-artifact1834064452502646289
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:15
>>>>>> ./beam-artifact682561790267074916
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:15
>>>>>> ./beam-artifact4691304965824489394
>>>>>> > >>>>>>>>>> 105M    2020-07-23 00:14
>>>>>> ./beam-artifact4050383819822604421
>>>>>> > >>>>>>>>>>
>>>>>> > >>>>>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw <
>>>>>> > >>>>>>>>>> rober...@google.com> wrote:
>>>>>> > >>>>>>>>>>
>>>>>> > >>>>>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton <
>>>>>> > >>>>>>>>>>> tyso...@google.com> wrote:
>>>>>> > >>>>>>>>>>>
>>>>>> > >>>>>>>>>>>> Ah I see, thanks Kenn. I found some advice from the
>>>>>> Apache
>>>>>> > >>>>>>>>>>>> infra wiki that also suggests using a tmpdir inside
>>>>>> the workspace [1]:
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>> Procedures Projects can take to clean up disk space
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>> Projects can help themselves and Infra by taking some
>>>>>> basic
>>>>>> > >>>>>>>>>>>> steps to help clean up their jobs after themselves on
>>>>>> the build nodes.
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>    1. Use a ./tmp dir in your jobs workspace. That way
>>>>>> it gets
>>>>>> > >>>>>>>>>>>>    cleaned up when job workspaces expire.
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>> Tests should be (able to be) written to use the standard
>>>>>> > >>>>>>>>>>> temporary file mechanisms, and the environment set up
>>>>>> on Jenkins such that
>>>>>> > >>>>>>>>>>> that falls into the respective workspaces. Ideally this
>>>>>> should be as simple
>>>>>> > >>>>>>>>>>> as setting the TMPDIR (or similar) environment variable
>>>>>> (and making sure it
>>>>>> > >>>>>>>>>>> exists/is writable).
>>>>>> > >>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>    1. Configure your jobs to wipe workspaces on start
>>>>>> or
>>>>>> > >>>>>>>>>>>>    finish.
>>>>>> > >>>>>>>>>>>>    2. Configure your jobs to only keep 5 or 10
>>>>>> previous builds.
>>>>>> > >>>>>>>>>>>>    3. Configure your jobs to only keep 5 or 10 previous
>>>>>> > >>>>>>>>>>>>    artifacts.
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>> [1]:
>>>>>> > >>>>>>>>>>>>
>>>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <
>>>>>> > >>>>>>>>>>>> k...@apache.org> wrote:
>>>>>> > >>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>> Those file listings look like the result of using
>>>>>> standard
>>>>>> > >>>>>>>>>>>>> temp file APIs but with TMPDIR set to /tmp.
>>>>>> > >>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton <
>>>>>> > >>>>>>>>>>>>> tyso...@google.com> wrote:
>>>>>> > >>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique
>>>>>> > >>>>>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look
>>>>>> into two examples:
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time .
>>>>>> | sort
>>>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20
>>>>>> > >>>>>>>>>>>>>> 1.6G    2020-07-21 02:25        .
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:48
>>>>>> ./beam-pipeline-temp3ybuY4
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:46
>>>>>> ./beam-pipeline-tempuxjiPT
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:44
>>>>>> ./beam-pipeline-tempVpg1ME
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:42
>>>>>> ./beam-pipeline-tempJ4EpyB
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:39
>>>>>> ./beam-pipeline-tempepea7Q
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:35
>>>>>> ./beam-pipeline-temp79qot2
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:48
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmpy_Ytzz
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:46
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpN5_UfJ
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:44
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpxSm8pX
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:42
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpMZJU76
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:39
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmpWy1vWX
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:35
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmpvN7vWA
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:48
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmprlh_di
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:46
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpLmVWfe
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:44
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpvrxbY7
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:42
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:39
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmptYF1v1
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:35
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmplfV0Rg
>>>>>> > >>>>>>>>>>>>>> 2.7M    2020-07-17 20:10
>>>>>> ./pip-install-q9l227ef
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time
>>>>>> . | sort
>>>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20
>>>>>> > >>>>>>>>>>>>>> 817M    2020-07-21 02:26        .
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:14
>>>>>> ./beam-pipeline-tempUTXqlM
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:11
>>>>>> ./beam-pipeline-tempx3Yno3
>>>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:05
>>>>>> ./beam-pipeline-tempyCrMYq
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:14
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpstXoL0
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:11
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpnnVn65
>>>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:05
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpRF0iNs
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:14
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpbJjUAQ
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:11
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpsmmzqe
>>>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:05
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmp5b3ZvY
>>>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:14
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpoj3orz
>>>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:11
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmptng9sZ
>>>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:05
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpWp6njc
>>>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:14
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmphgdj35
>>>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:11
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmp8ySXpm
>>>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:05
>>>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpNVEJ4e
>>>>>> > >>>>>>>>>>>>>> 992K    2020-07-12 12:00
>>>>>> ./junit642086915811430564
>>>>>> > >>>>>>>>>>>>>> 988K    2020-07-12 12:00
>>>>>> ./junit642086915811430564/beam
>>>>>> > >>>>>>>>>>>>>> 984K    2020-07-12 12:00
>>>>>> > >>>>>>>>>>>>>>  ./junit642086915811430564/beam/nodes
>>>>>> > >>>>>>>>>>>>>> 980K    2020-07-12 12:00
>>>>>> > >>>>>>>>>>>>>>  ./junit642086915811430564/beam/nodes/0
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <
>>>>>> eh...@google.com>
>>>>>> > >>>>>>>>>>>>>> wrote:
>>>>>> > >>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>> You're right, job workspaces should be hermetic.
>>>>>> > >>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles <
>>>>>> > >>>>>>>>>>>>>>> k...@apache.org> wrote:
>>>>>> > >>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>> I'm probably late to this discussion and missing
>>>>>> something,
>>>>>> > >>>>>>>>>>>>>>>> but why are we writing to /tmp at all? I would
>>>>>> expect TMPDIR to point
>>>>>> > >>>>>>>>>>>>>>>> somewhere inside the job directory that will be
>>>>>> wiped by Jenkins, and I
>>>>>> > >>>>>>>>>>>>>>>> would expect code to always create temp files via
>>>>>> APIs that respect this.
>>>>>> > >>>>>>>>>>>>>>>> Is Jenkins not cleaning up? Do we not have the
>>>>>> ability to set this up? Do
>>>>>> > >>>>>>>>>>>>>>>> we have bugs in our code (that we could probably
>>>>>> find by setting TMPDIR to
>>>>>> > >>>>>>>>>>>>>>>> somewhere not-/tmp and running the tests without
>>>>>> write permission to /tmp,
>>>>>> > >>>>>>>>>>>>>>>> etc)
>>>>>> > >>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>> Kenn
>>>>>> > >>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay <
>>>>>> > >>>>>>>>>>>>>>>> al...@google.com> wrote:
>>>>>> > >>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri
>>>>>> > >>>>>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue
>>>>>> previously (
>>>>>> > >>>>>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865)
>>>>>> for
>>>>>> > >>>>>>>>>>>>>>>>> cleaning up workspace directory after successful
>>>>>> jobs. Alternatively, we
>>>>>> > >>>>>>>>>>>>>>>>> can consider periodically cleaning up the /src
>>>>>> directories.
>>>>>> > >>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>> I would suggest moving the cron task from
>>>>>> internal cron
>>>>>> > >>>>>>>>>>>>>>>>> scripts to the inventory job (
>>>>>> > >>>>>>>>>>>>>>>>>
>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51
>>>>>> ).
>>>>>> > >>>>>>>>>>>>>>>>> That way, we can see all the cron jobs as part of
>>>>>> the source tree, adjust
>>>>>> > >>>>>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not
>>>>>> know how internal cron
>>>>>> > >>>>>>>>>>>>>>>>> scripts are created, maintained, and how would
>>>>>> they be recreated for new
>>>>>> > >>>>>>>>>>>>>>>>> worker instances.
>>>>>> > >>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com>
>>>>>> > >>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski <
>>>>>> > >>>>>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> Hey,
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> I've recently created a solution for the growing
>>>>>> /tmp
>>>>>> > >>>>>>>>>>>>>>>>>> directory. Part of it is the job mentioned by
>>>>>> Tyson:
>>>>>> > >>>>>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally
>>>>>> not
>>>>>> > >>>>>>>>>>>>>>>>>> triggered by cron and should be a last resort
>>>>>> solution for some strange
>>>>>> > >>>>>>>>>>>>>>>>>> cases.
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> Along with that job, I've also updated every
>>>>>> worker with
>>>>>> > >>>>>>>>>>>>>>>>>> an internal cron script. It's being executed
>>>>>> once a week and deletes all
>>>>>> > >>>>>>>>>>>>>>>>>> the files (and only files) that were not
>>>>>> accessed for at least three days.
>>>>>> > >>>>>>>>>>>>>>>>>> That's designed to be as safe as possible for
>>>>>> the running jobs on the
>>>>>> > >>>>>>>>>>>>>>>>>> worker (not to delete the files that are still
>>>>>> in use), and also to be
>>>>>> > >>>>>>>>>>>>>>>>>> insensitive to the current workload on the
>>>>>> machine. The cleanup will always
>>>>>> > >>>>>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are
>>>>>> blocking the machine.
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> I also think that currently the "No space left"
>>>>>> errors
>>>>>> > >>>>>>>>>>>>>>>>>> may be a consequence of growing workspace
>>>>>> directory rather than /tmp. I
>>>>>> > >>>>>>>>>>>>>>>>>> didn't do any detailed analysis but e.g.
>>>>>> currently, on
>>>>>> > >>>>>>>>>>>>>>>>>> apache-beam-jenkins-7 the workspace directory
>>>>>> size is 158 GB while /tmp is
>>>>>> > >>>>>>>>>>>>>>>>>> only 16 GB. We should either guarantee the disk
>>>>>> size to hold workspaces for
>>>>>> > >>>>>>>>>>>>>>>>>> all jobs (because eventually, every worker will
>>>>>> execute each job) or clear
>>>>>> > >>>>>>>>>>>>>>>>>> also the workspaces in some way.
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> Regards,
>>>>>> > >>>>>>>>>>>>>>>>>> Damian
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian
>>>>>> Michels <
>>>>>> > >>>>>>>>>>>>>>>>>> m...@apache.org> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't
>>>>>> lead to
>>>>>> > >>>>>>>>>>>>>>>>>>> test failures
>>>>>> > >>>>>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe
>>>>>> there is
>>>>>> > >>>>>>>>>>>>>>>>>>> the notion of
>>>>>> > >>>>>>>>>>>>>>>>>>> running exclusively while no other tasks are
>>>>>> running?
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> -Max
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in
>>>>>> Jenkins:
>>>>>> > >>>>>>>>>>>>>>>>>>> beam_Clean_tmp_directory
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm
>>>>>> seeing some
>>>>>> > >>>>>>>>>>>>>>>>>>> out of disk related errors in precommit tests
>>>>>> currently, perhaps we should
>>>>>> > >>>>>>>>>>>>>>>>>>> schedule this job with cron?
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <
>>>>>> > >>>>>>>>>>>>>>>>>>> heej...@google.com> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>> >> Still seeing no space left on device errors
>>>>>> on
>>>>>> > >>>>>>>>>>>>>>>>>>> jenkins-7 (for example:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/
>>>>>> > >>>>>>>>>>>>>>>>>>> )
>>>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>>>> > >>>>>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold <
>>>>>> > >>>>>>>>>>>>>>>>>>> amyrv...@google.com> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned
>>>>>> by jenkins
>>>>>> > >>>>>>>>>>>>>>>>>>> older than 3 days.
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution.
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except
>>>>>> > >>>>>>>>>>>>>>>>>>> jenkins-12, which has not
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13
>>>>>> days. Not
>>>>>> > >>>>>>>>>>>>>>>>>>> scheduling:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> Recent passing builds:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D
>>>>>> > >>>>>>>>>>>>>>>>>>> >
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay
>>>>>> <
>>>>>> > >>>>>>>>>>>>>>>>>>> al...@google.com> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is
>>>>>> doing a one
>>>>>> > >>>>>>>>>>>>>>>>>>> time cleanup. I agree
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>> that we need to have a solution to
>>>>>> automate this
>>>>>> > >>>>>>>>>>>>>>>>>>> task or address the root
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>> cause of the buildup.
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał
>>>>>> Walenia <
>>>>>> > >>>>>>>>>>>>>>>>>>> michal.wale...@polidea.com>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>> wrote:
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Hi there,
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins
>>>>>> workers
>>>>>> > >>>>>>>>>>>>>>>>>>> again. Nodes 1 and 7
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on
>>>>>> device".
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in
>>>>>> these cases
>>>>>> > >>>>>>>>>>>>>>>>>>> (someone with access
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> permissions to the workers).
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are
>>>>>> becoming more
>>>>>> > >>>>>>>>>>>>>>>>>>> and more frequent
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can
>>>>>> this be
>>>>>> > >>>>>>>>>>>>>>>>>>> remedied. Can a cleanup
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow?
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Regards
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michal
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> --
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michał Walenia
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> |
>>>>>> Software
>>>>>> > >>>>>>>>>>>>>>>>>>> Engineer
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002
>>>>>> <+48%20791%20432%20002> <+48%20791%20432%20002> <
>>>>>> > >>>>>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>
>>>>>> <+48%20791%20432%20002>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Unique Tech
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Check out our projects! <
>>>>>> > >>>>>>>>>>>>>>>>>>> https://www.polidea.com/our-work>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>>>> > >>>>>>>>>>>>>>>>>>>
>>>>>> > >>>>>>>>>>>>>>>>>>
>>>>>> >
>>>>>>
>>>>>

Reply via email to