That's interesting. I didn't check that myself but all the Jenkins jobs are
configured to wipe the workspace just before the actual build happens [1]
<https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6>.
Git SCM plugin is used for that and it enables the option called "Wipe out
repository and force clone". Docs state that it "deletes the contents of
the workspace before build and before checkout" [2]
<https://plugins.jenkins.io/git/>. Therefore I assume that removing
workspace just after the build won't change anything.

The ./.gradle/caches/modules-2/files-2.1 dir is indeed present on the
worker machines but it's rather in /home/jenkins dir.

damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh
11G .
damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh
caches/modules-2/files-2.1
2.3G caches/modules-2/files-2.1

I can't find that directory structure inside workspaces.

damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$
sudo find -name "files-2.1"
damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$

[1]
https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6
[2] https://plugins.jenkins.io/git/

On Tue, Jul 28, 2020 at 5:47 PM Kenneth Knowles <k...@apache.org> wrote:

> Just checking - will this wipe out dependency cache? That will slow things
> down and significantly increase flakiness. If I recall correctly, the
> default Jenkins layout was:
>
>     /home/jenkins/jenkins-slave/workspace/$jobname
>     /home/jenkins/jenkins-slave/workspace/$jobname/.m2
>     /home/jenkins/jenkins-slave/workspace/$jobname/.git
>
> Where you can see that it did a `git clone` right into the root workspace
> directory, adjacent to .m2. This was not hygienic. One important thing was
> that `git clean` would wipe the maven cache with every build. So in
> https://github.com/apache/beam/pull/3976 we changed it to
>
>     /home/jenkins/jenkins-slave/workspace/$jobname
>     /home/jenkins/jenkins-slave/workspace/$jobname/.m2
>     /home/jenkins/jenkins-slave/workspace/$jobname/src/.git
>
> Now the .m2 directory survives and we do not constantly see flakes
> re-downloading deps that are immutable. This does, of course, use disk
> space.
>
> That was in the maven days. Gradle is the same except for $HOME/.m2 is
> replaced by $HOME/.gradle/caches/modules-2/files-2.1. Is Jenkins configured
> the same way so we will be wiping out the dependencies? If so, can you
> address this issue? Everything in that directory should be immutable and
> just a cache to avoid pointless re-download.
>
> Kenn
>
> On Tue, Jul 28, 2020 at 2:25 AM Damian Gadomski <
> damian.gadom...@polidea.com> wrote:
>
>> Agree with Udi, workspaces seem to be the third culprit, not yet
>> addressed in any way (until PR#12326
>> <https://github.com/apache/beam/pull/12326> is merged). I feel that
>> it'll solve the issue of filling up the disks for a long time ;)
>>
>> I'm also OK with moving /tmp cleanup to option B, and will happily
>> investigate on proper TMPDIR config.
>>
>>
>>
>> On Tue, Jul 28, 2020 at 3:07 AM Udi Meiri <eh...@google.com> wrote:
>>
>>> What about the workspaces, which can take up 175GB in some cases (see
>>> above)?
>>> I'm working on getting them cleaned up automatically:
>>> https://github.com/apache/beam/pull/12326
>>>
>>> My opinion is that we would get more mileage out of fixing the jobs that
>>> leave behind files in /tmp and images/containers in Docker.
>>> This would also help keep development machines clean.
>>>
>>>
>>> On Mon, Jul 27, 2020 at 5:31 PM Tyson Hamilton <tyso...@google.com>
>>> wrote:
>>>
>>>> Here is a summery of how I understand things,
>>>>
>>>>   - /tmp and /var/lib/docker are the culprit for filling up disks
>>>>   - inventory Jenkins job runs every 12 hours and runs a docker prune
>>>> to clean up images older than 24hr
>>>>   - crontab on each machine cleans up /tmp files older than three days
>>>> weekly
>>>>
>>>> This doesn't seem to be working since we're still running out of disk
>>>> periodically and requiring manual intervention. Knobs and options we have
>>>> available:
>>>>
>>>>   1. increase frequency of deleting files
>>>>   2. decrease the number of days required to delete a file (e.g. older
>>>> than 2 days)
>>>>
>>>> The execution methods we have available are:
>>>>
>>>>   A. cron
>>>>     - pro: runs even if a job gets stuck in Jenkins due to full disk
>>>>     - con: config baked into VM which is tough to update, not
>>>> discoverable or documented well
>>>>   B. inventory job
>>>>     - pro: easy to update, runs every 12h already
>>>>     - con: could get stuck if Jenkins agent runs out of disk or is
>>>> otherwise stuck, tied to all other inventory job frequency
>>>>   C. configure startup scripts for the VMs that set up the cron job
>>>> anytime the VM is restarted
>>>>     - pro: similar to A. and easy to update
>>>>     - con: similar to A.
>>>>
>>>> Between the three I prefer B. because it is consistent with other
>>>> inventory jobs. If it ends up that stuck jobs prohibit scheduling of the
>>>> inventory job often we could further investigate C to avoid having to
>>>> rebuild the VM images repeatedly.
>>>>
>>>> Any objections or comments? If not, we'll go forward with B. and reduce
>>>> the date check from 3 days to 2 days.
>>>>
>>>>
>>>> On 2020/07/24 20:13:29, Ahmet Altay <al...@google.com> wrote:
>>>> > Tests may not be doing docker cleanup. Inventory job runs a docker
>>>> prune
>>>> > every 12 hours for images older than 24 hrs [1]. Randomly looking at
>>>> one of
>>>> > the recent runs [2], it cleaned up a long list of containers consuming
>>>> > 30+GB space. That should be just 12 hours worth of containers.
>>>> >
>>>> > [1]
>>>> >
>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L69
>>>> > [2]
>>>> >
>>>> https://ci-beam.apache.org/job/beam_Inventory_apache-beam-jenkins-14/501/console
>>>> >
>>>> > On Fri, Jul 24, 2020 at 1:07 PM Tyson Hamilton <tyso...@google.com>
>>>> wrote:
>>>> >
>>>> > > Yes, these are on the same volume in the /var/lib/docker directory.
>>>> I'm
>>>> > > unsure if they clean up leftover images.
>>>> > >
>>>> > > On Fri, Jul 24, 2020 at 12:52 PM Udi Meiri <eh...@google.com>
>>>> wrote:
>>>> > >
>>>> > >> I forgot Docker images:
>>>> > >>
>>>> > >> ehudm@apache-ci-beam-jenkins-3:~$ sudo docker system df
>>>> > >> TYPE                TOTAL               ACTIVE              SIZE
>>>> > >>        RECLAIMABLE
>>>> > >> Images              88                  9                   125.4GB
>>>> > >>       124.2GB (99%)
>>>> > >> Containers          40                  4                   7.927GB
>>>> > >>       7.871GB (99%)
>>>> > >> Local Volumes       47                  0                   3.165GB
>>>> > >>       3.165GB (100%)
>>>> > >> Build Cache         0                   0                   0B
>>>> > >>        0B
>>>> > >>
>>>> > >> There are about 90 images on that machine, with all but 1 less
>>>> than 48
>>>> > >> hours old.
>>>> > >> I think the docker test jobs need to try harder at cleaning up
>>>> their
>>>> > >> leftover images. (assuming they're already doing it?)
>>>> > >>
>>>> > >> On Fri, Jul 24, 2020 at 12:31 PM Udi Meiri <eh...@google.com>
>>>> wrote:
>>>> > >>
>>>> > >>> The additional slots (@3 directories) take up even more space now
>>>> than
>>>> > >>> before.
>>>> > >>>
>>>> > >>> I'm testing out https://github.com/apache/beam/pull/12326 which
>>>> could
>>>> > >>> help by cleaning up workspaces after a run (just started a seed
>>>> job).
>>>> > >>>
>>>> > >>> On Fri, Jul 24, 2020 at 12:13 PM Tyson Hamilton <
>>>> tyso...@google.com>
>>>> > >>> wrote:
>>>> > >>>
>>>> > >>>> 664M    beam_PreCommit_JavaPortabilityApi_Commit
>>>> > >>>> 656M    beam_PreCommit_JavaPortabilityApi_Commit@2
>>>> > >>>> 611M    beam_PreCommit_JavaPortabilityApi_Cron
>>>> > >>>> 616M    beam_PreCommit_JavaPortabilityApiJava11_Commit
>>>> > >>>> 598M    beam_PreCommit_JavaPortabilityApiJava11_Commit@2
>>>> > >>>> 662M    beam_PreCommit_JavaPortabilityApiJava11_Cron
>>>> > >>>> 2.9G    beam_PreCommit_Portable_Python_Commit
>>>> > >>>> 2.9G    beam_PreCommit_Portable_Python_Commit@2
>>>> > >>>> 1.7G    beam_PreCommit_Portable_Python_Commit@3
>>>> > >>>> 3.4G    beam_PreCommit_Portable_Python_Cron
>>>> > >>>> 1.9G    beam_PreCommit_Python2_PVR_Flink_Commit
>>>> > >>>> 1.4G    beam_PreCommit_Python2_PVR_Flink_Cron
>>>> > >>>> 1.3G    beam_PreCommit_Python2_PVR_Flink_Phrase
>>>> > >>>> 6.2G    beam_PreCommit_Python_Commit
>>>> > >>>> 7.5G    beam_PreCommit_Python_Commit@2
>>>> > >>>> 7.5G    beam_PreCommit_Python_Cron
>>>> > >>>> 1012M   beam_PreCommit_PythonDocker_Commit
>>>> > >>>> 1011M   beam_PreCommit_PythonDocker_Commit@2
>>>> > >>>> 1011M   beam_PreCommit_PythonDocker_Commit@3
>>>> > >>>> 1002M   beam_PreCommit_PythonDocker_Cron
>>>> > >>>> 877M    beam_PreCommit_PythonFormatter_Commit
>>>> > >>>> 988M    beam_PreCommit_PythonFormatter_Cron
>>>> > >>>> 986M    beam_PreCommit_PythonFormatter_Phrase
>>>> > >>>> 1.7G    beam_PreCommit_PythonLint_Commit
>>>> > >>>> 2.1G    beam_PreCommit_PythonLint_Cron
>>>> > >>>> 7.5G    beam_PreCommit_Python_Phrase
>>>> > >>>> 346M    beam_PreCommit_RAT_Commit
>>>> > >>>> 341M    beam_PreCommit_RAT_Cron
>>>> > >>>> 338M    beam_PreCommit_Spotless_Commit
>>>> > >>>> 339M    beam_PreCommit_Spotless_Cron
>>>> > >>>> 5.5G    beam_PreCommit_SQL_Commit
>>>> > >>>> 5.5G    beam_PreCommit_SQL_Cron
>>>> > >>>> 5.5G    beam_PreCommit_SQL_Java11_Commit
>>>> > >>>> 750M    beam_PreCommit_Website_Commit
>>>> > >>>> 750M    beam_PreCommit_Website_Commit@2
>>>> > >>>> 750M    beam_PreCommit_Website_Cron
>>>> > >>>> 764M    beam_PreCommit_Website_Stage_GCS_Commit
>>>> > >>>> 771M    beam_PreCommit_Website_Stage_GCS_Cron
>>>> > >>>> 336M    beam_Prober_CommunityMetrics
>>>> > >>>> 693M    beam_python_mongoio_load_test
>>>> > >>>> 339M    beam_SeedJob
>>>> > >>>> 333M    beam_SeedJob_Standalone
>>>> > >>>> 334M    beam_sonarqube_report
>>>> > >>>> 556M    beam_SQLBigQueryIO_Batch_Performance_Test_Java
>>>> > >>>> 175G    total
>>>> > >>>>
>>>> > >>>> On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <
>>>> tyso...@google.com>
>>>> > >>>> wrote:
>>>> > >>>>
>>>> > >>>>> Ya looks like something in the workspaces is taking up room:
>>>> > >>>>>
>>>> > >>>>> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc .
>>>> > >>>>> 191G    .
>>>> > >>>>> 191G    total
>>>> > >>>>>
>>>> > >>>>>
>>>> > >>>>> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <
>>>> tyso...@google.com>
>>>> > >>>>> wrote:
>>>> > >>>>>
>>>> > >>>>>> Node 8 is also full. The partition that /tmp is on is here:
>>>> > >>>>>>
>>>> > >>>>>> Filesystem      Size  Used Avail Use% Mounted on
>>>> > >>>>>> /dev/sda1       485G  482G  2.9G 100% /
>>>> > >>>>>>
>>>> > >>>>>> however after cleaning up tmp with the crontab command, there
>>>> is only
>>>> > >>>>>> 8G usage yet it still remains 100% full:
>>>> > >>>>>>
>>>> > >>>>>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp
>>>> > >>>>>> 8.0G    /tmp
>>>> > >>>>>> 8.0G    total
>>>> > >>>>>>
>>>> > >>>>>> The workspaces are in the /home/jenkins/jenkins-slave/workspace
>>>> > >>>>>> directory. When I run a du on that, it takes really long. I'll
>>>> let it keep
>>>> > >>>>>> running for a while to see if it ever returns a result but so
>>>> far this
>>>> > >>>>>> seems suspect.
>>>> > >>>>>>
>>>> > >>>>>>
>>>> > >>>>>>
>>>> > >>>>>>
>>>> > >>>>>>
>>>> > >>>>>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <
>>>> tyso...@google.com>
>>>> > >>>>>> wrote:
>>>> > >>>>>>
>>>> > >>>>>>> Everything I've been looking at is in the /tmp dir. Where are
>>>> the
>>>> > >>>>>>> workspaces, or what are the named?
>>>> > >>>>>>>
>>>> > >>>>>>>
>>>> > >>>>>>>
>>>> > >>>>>>>
>>>> > >>>>>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com>
>>>> wrote:
>>>> > >>>>>>>
>>>> > >>>>>>>> I'm curious to what you find. Was it /tmp or the workspaces
>>>> using
>>>> > >>>>>>>> up the space?
>>>> > >>>>>>>>
>>>> > >>>>>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <
>>>> tyso...@google.com>
>>>> > >>>>>>>> wrote:
>>>> > >>>>>>>>
>>>> > >>>>>>>>> Bleck. I just realized that it is 'offline' so that won't
>>>> work.
>>>> > >>>>>>>>> I'll clean up manually on the machine using the cron
>>>> command.
>>>> > >>>>>>>>>
>>>> > >>>>>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <
>>>> > >>>>>>>>> tyso...@google.com> wrote:
>>>> > >>>>>>>>>
>>>> > >>>>>>>>>> Something isn't working with the current set up because
>>>> node 15
>>>> > >>>>>>>>>> appears to be out of space and is currently 'offline'
>>>> according to Jenkins.
>>>> > >>>>>>>>>> Can someone run the cleanup job? The machine is full,
>>>> > >>>>>>>>>>
>>>> > >>>>>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h
>>>> > >>>>>>>>>> Filesystem      Size  Used Avail Use% Mounted on
>>>> > >>>>>>>>>> udev             52G     0   52G   0% /dev
>>>> > >>>>>>>>>> tmpfs            11G  265M   10G   3% /run
>>>> > >>>>>>>>>> */dev/sda1       485G  484G  880M 100% /*
>>>> > >>>>>>>>>> tmpfs            52G     0   52G   0% /dev/shm
>>>> > >>>>>>>>>> tmpfs           5.0M     0  5.0M   0% /run/lock
>>>> > >>>>>>>>>> tmpfs            52G     0   52G   0% /sys/fs/cgroup
>>>> > >>>>>>>>>> tmpfs            11G     0   11G   0% /run/user/1017
>>>> > >>>>>>>>>> tmpfs            11G     0   11G   0% /run/user/1037
>>>> > >>>>>>>>>>
>>>> > >>>>>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . |
>>>> sort -rhk
>>>> > >>>>>>>>>> 1,1 | head -n 20
>>>> > >>>>>>>>>> 20G     2020-07-24 17:52        .
>>>> > >>>>>>>>>> 580M    2020-07-22 17:31        ./junit1031982597110125586
>>>> > >>>>>>>>>> 517M    2020-07-22 17:31
>>>> > >>>>>>>>>>
>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof
>>>> > >>>>>>>>>> 517M    2020-07-22 17:31
>>>> > >>>>>>>>>>  ./junit1031982597110125586/junit8739924829337821410
>>>> > >>>>>>>>>> 263M    2020-07-22 12:23        ./pip-install-2GUhO_
>>>> > >>>>>>>>>> 263M    2020-07-20 09:30        ./pip-install-sxgwqr
>>>> > >>>>>>>>>> 263M    2020-07-17 13:56        ./pip-install-bWSKIV
>>>> > >>>>>>>>>> 242M    2020-07-21 20:25        ./beam-pipeline-tempmByU6T
>>>> > >>>>>>>>>> 242M    2020-07-21 20:21        ./beam-pipeline-tempV85xeK
>>>> > >>>>>>>>>> 242M    2020-07-21 20:15        ./beam-pipeline-temp7dJROJ
>>>> > >>>>>>>>>> 236M    2020-07-21 20:25
>>>> > >>>>>>>>>>  ./beam-pipeline-tempmByU6T/tmpOWj3Yr
>>>> > >>>>>>>>>> 236M    2020-07-21 20:21
>>>> > >>>>>>>>>>  ./beam-pipeline-tempV85xeK/tmppbQHB3
>>>> > >>>>>>>>>> 236M    2020-07-21 20:15
>>>> > >>>>>>>>>>  ./beam-pipeline-temp7dJROJ/tmpgOXPKW
>>>> > >>>>>>>>>> 111M    2020-07-23 00:57        ./pip-install-1JnyNE
>>>> > >>>>>>>>>> 105M    2020-07-23 00:17
>>>> ./beam-artifact1374651823280819755
>>>> > >>>>>>>>>> 105M    2020-07-23 00:16
>>>> ./beam-artifact5050755582921936972
>>>> > >>>>>>>>>> 105M    2020-07-23 00:16
>>>> ./beam-artifact1834064452502646289
>>>> > >>>>>>>>>> 105M    2020-07-23 00:15
>>>> ./beam-artifact682561790267074916
>>>> > >>>>>>>>>> 105M    2020-07-23 00:15
>>>> ./beam-artifact4691304965824489394
>>>> > >>>>>>>>>> 105M    2020-07-23 00:14
>>>> ./beam-artifact4050383819822604421
>>>> > >>>>>>>>>>
>>>> > >>>>>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw <
>>>> > >>>>>>>>>> rober...@google.com> wrote:
>>>> > >>>>>>>>>>
>>>> > >>>>>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton <
>>>> > >>>>>>>>>>> tyso...@google.com> wrote:
>>>> > >>>>>>>>>>>
>>>> > >>>>>>>>>>>> Ah I see, thanks Kenn. I found some advice from the
>>>> Apache
>>>> > >>>>>>>>>>>> infra wiki that also suggests using a tmpdir inside the
>>>> workspace [1]:
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>> Procedures Projects can take to clean up disk space
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>> Projects can help themselves and Infra by taking some
>>>> basic
>>>> > >>>>>>>>>>>> steps to help clean up their jobs after themselves on
>>>> the build nodes.
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>    1. Use a ./tmp dir in your jobs workspace. That way
>>>> it gets
>>>> > >>>>>>>>>>>>    cleaned up when job workspaces expire.
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>> Tests should be (able to be) written to use the standard
>>>> > >>>>>>>>>>> temporary file mechanisms, and the environment set up on
>>>> Jenkins such that
>>>> > >>>>>>>>>>> that falls into the respective workspaces. Ideally this
>>>> should be as simple
>>>> > >>>>>>>>>>> as setting the TMPDIR (or similar) environment variable
>>>> (and making sure it
>>>> > >>>>>>>>>>> exists/is writable).
>>>> > >>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>    1. Configure your jobs to wipe workspaces on start or
>>>> > >>>>>>>>>>>>    finish.
>>>> > >>>>>>>>>>>>    2. Configure your jobs to only keep 5 or 10 previous
>>>> builds.
>>>> > >>>>>>>>>>>>    3. Configure your jobs to only keep 5 or 10 previous
>>>> > >>>>>>>>>>>>    artifacts.
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>> [1]:
>>>> > >>>>>>>>>>>>
>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <
>>>> > >>>>>>>>>>>> k...@apache.org> wrote:
>>>> > >>>>>>>>>>>>
>>>> > >>>>>>>>>>>>> Those file listings look like the result of using
>>>> standard
>>>> > >>>>>>>>>>>>> temp file APIs but with TMPDIR set to /tmp.
>>>> > >>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton <
>>>> > >>>>>>>>>>>>> tyso...@google.com> wrote:
>>>> > >>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique
>>>> > >>>>>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look
>>>> into two examples:
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . |
>>>> sort
>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20
>>>> > >>>>>>>>>>>>>> 1.6G    2020-07-21 02:25        .
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:48
>>>> ./beam-pipeline-temp3ybuY4
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:46
>>>> ./beam-pipeline-tempuxjiPT
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:44
>>>> ./beam-pipeline-tempVpg1ME
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:42
>>>> ./beam-pipeline-tempJ4EpyB
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:39
>>>> ./beam-pipeline-tempepea7Q
>>>> > >>>>>>>>>>>>>> 242M    2020-07-17 18:35
>>>> ./beam-pipeline-temp79qot2
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:48
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmpy_Ytzz
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:46
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpN5_UfJ
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:44
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpxSm8pX
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:42
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpMZJU76
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:39
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmpWy1vWX
>>>> > >>>>>>>>>>>>>> 236M    2020-07-17 18:35
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmpvN7vWA
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:48
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmprlh_di
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:46
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpLmVWfe
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:44
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpvrxbY7
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:42
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:39
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmptYF1v1
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-17 18:35
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmplfV0Rg
>>>> > >>>>>>>>>>>>>> 2.7M    2020-07-17 20:10        ./pip-install-q9l227ef
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time .
>>>> | sort
>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20
>>>> > >>>>>>>>>>>>>> 817M    2020-07-21 02:26        .
>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:14
>>>> ./beam-pipeline-tempUTXqlM
>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:11
>>>> ./beam-pipeline-tempx3Yno3
>>>> > >>>>>>>>>>>>>> 242M    2020-07-19 12:05
>>>> ./beam-pipeline-tempyCrMYq
>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:14
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpstXoL0
>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:11
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpnnVn65
>>>> > >>>>>>>>>>>>>> 236M    2020-07-19 12:05
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpRF0iNs
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:14
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpbJjUAQ
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:11
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpsmmzqe
>>>> > >>>>>>>>>>>>>> 3.7M    2020-07-19 12:05
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmp5b3ZvY
>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:14
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpoj3orz
>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:11
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmptng9sZ
>>>> > >>>>>>>>>>>>>> 2.0M    2020-07-19 12:05
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpWp6njc
>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:14
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmphgdj35
>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:11
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmp8ySXpm
>>>> > >>>>>>>>>>>>>> 1.2M    2020-07-19 12:05
>>>> > >>>>>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpNVEJ4e
>>>> > >>>>>>>>>>>>>> 992K    2020-07-12 12:00
>>>> ./junit642086915811430564
>>>> > >>>>>>>>>>>>>> 988K    2020-07-12 12:00
>>>> ./junit642086915811430564/beam
>>>> > >>>>>>>>>>>>>> 984K    2020-07-12 12:00
>>>> > >>>>>>>>>>>>>>  ./junit642086915811430564/beam/nodes
>>>> > >>>>>>>>>>>>>> 980K    2020-07-12 12:00
>>>> > >>>>>>>>>>>>>>  ./junit642086915811430564/beam/nodes/0
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <
>>>> eh...@google.com>
>>>> > >>>>>>>>>>>>>> wrote:
>>>> > >>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>> You're right, job workspaces should be hermetic.
>>>> > >>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles <
>>>> > >>>>>>>>>>>>>>> k...@apache.org> wrote:
>>>> > >>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>> I'm probably late to this discussion and missing
>>>> something,
>>>> > >>>>>>>>>>>>>>>> but why are we writing to /tmp at all? I would
>>>> expect TMPDIR to point
>>>> > >>>>>>>>>>>>>>>> somewhere inside the job directory that will be
>>>> wiped by Jenkins, and I
>>>> > >>>>>>>>>>>>>>>> would expect code to always create temp files via
>>>> APIs that respect this.
>>>> > >>>>>>>>>>>>>>>> Is Jenkins not cleaning up? Do we not have the
>>>> ability to set this up? Do
>>>> > >>>>>>>>>>>>>>>> we have bugs in our code (that we could probably
>>>> find by setting TMPDIR to
>>>> > >>>>>>>>>>>>>>>> somewhere not-/tmp and running the tests without
>>>> write permission to /tmp,
>>>> > >>>>>>>>>>>>>>>> etc)
>>>> > >>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>> Kenn
>>>> > >>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay <
>>>> > >>>>>>>>>>>>>>>> al...@google.com> wrote:
>>>> > >>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri
>>>> > >>>>>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue
>>>> previously (
>>>> > >>>>>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865)
>>>> for
>>>> > >>>>>>>>>>>>>>>>> cleaning up workspace directory after successful
>>>> jobs. Alternatively, we
>>>> > >>>>>>>>>>>>>>>>> can consider periodically cleaning up the /src
>>>> directories.
>>>> > >>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>> I would suggest moving the cron task from internal
>>>> cron
>>>> > >>>>>>>>>>>>>>>>> scripts to the inventory job (
>>>> > >>>>>>>>>>>>>>>>>
>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51
>>>> ).
>>>> > >>>>>>>>>>>>>>>>> That way, we can see all the cron jobs as part of
>>>> the source tree, adjust
>>>> > >>>>>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not
>>>> know how internal cron
>>>> > >>>>>>>>>>>>>>>>> scripts are created, maintained, and how would they
>>>> be recreated for new
>>>> > >>>>>>>>>>>>>>>>> worker instances.
>>>> > >>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com>
>>>> > >>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski <
>>>> > >>>>>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote:
>>>> > >>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> Hey,
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> I've recently created a solution for the growing
>>>> /tmp
>>>> > >>>>>>>>>>>>>>>>>> directory. Part of it is the job mentioned by
>>>> Tyson:
>>>> > >>>>>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not
>>>> > >>>>>>>>>>>>>>>>>> triggered by cron and should be a last resort
>>>> solution for some strange
>>>> > >>>>>>>>>>>>>>>>>> cases.
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> Along with that job, I've also updated every
>>>> worker with
>>>> > >>>>>>>>>>>>>>>>>> an internal cron script. It's being executed once
>>>> a week and deletes all
>>>> > >>>>>>>>>>>>>>>>>> the files (and only files) that were not accessed
>>>> for at least three days.
>>>> > >>>>>>>>>>>>>>>>>> That's designed to be as safe as possible for the
>>>> running jobs on the
>>>> > >>>>>>>>>>>>>>>>>> worker (not to delete the files that are still in
>>>> use), and also to be
>>>> > >>>>>>>>>>>>>>>>>> insensitive to the current workload on the
>>>> machine. The cleanup will always
>>>> > >>>>>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are
>>>> blocking the machine.
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> I also think that currently the "No space left"
>>>> errors
>>>> > >>>>>>>>>>>>>>>>>> may be a consequence of growing workspace
>>>> directory rather than /tmp. I
>>>> > >>>>>>>>>>>>>>>>>> didn't do any detailed analysis but e.g.
>>>> currently, on
>>>> > >>>>>>>>>>>>>>>>>> apache-beam-jenkins-7 the workspace directory size
>>>> is 158 GB while /tmp is
>>>> > >>>>>>>>>>>>>>>>>> only 16 GB. We should either guarantee the disk
>>>> size to hold workspaces for
>>>> > >>>>>>>>>>>>>>>>>> all jobs (because eventually, every worker will
>>>> execute each job) or clear
>>>> > >>>>>>>>>>>>>>>>>> also the workspaces in some way.
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> Regards,
>>>> > >>>>>>>>>>>>>>>>>> Damian
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian
>>>> Michels <
>>>> > >>>>>>>>>>>>>>>>>> m...@apache.org> wrote:
>>>> > >>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't
>>>> lead to
>>>> > >>>>>>>>>>>>>>>>>>> test failures
>>>> > >>>>>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe
>>>> there is
>>>> > >>>>>>>>>>>>>>>>>>> the notion of
>>>> > >>>>>>>>>>>>>>>>>>> running exclusively while no other tasks are
>>>> running?
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>>> -Max
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote:
>>>> > >>>>>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in
>>>> Jenkins:
>>>> > >>>>>>>>>>>>>>>>>>> beam_Clean_tmp_directory
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm
>>>> seeing some
>>>> > >>>>>>>>>>>>>>>>>>> out of disk related errors in precommit tests
>>>> currently, perhaps we should
>>>> > >>>>>>>>>>>>>>>>>>> schedule this job with cron?
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <
>>>> > >>>>>>>>>>>>>>>>>>> heej...@google.com> wrote:
>>>> > >>>>>>>>>>>>>>>>>>> >> Still seeing no space left on device errors on
>>>> > >>>>>>>>>>>>>>>>>>> jenkins-7 (for example:
>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/
>>>> > >>>>>>>>>>>>>>>>>>> )
>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>> > >>>>>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold <
>>>> > >>>>>>>>>>>>>>>>>>> amyrv...@google.com> wrote:
>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>> > >>>>>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by
>>>> jenkins
>>>> > >>>>>>>>>>>>>>>>>>> older than 3 days.
>>>> > >>>>>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution.
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except
>>>> > >>>>>>>>>>>>>>>>>>> jenkins-12, which has not
>>>> > >>>>>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 days.
>>>> Not
>>>> > >>>>>>>>>>>>>>>>>>> scheduling:
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>> Recent passing builds:
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds
>>>> > >>>>>>>>>>>>>>>>>>> >>> <
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D
>>>> > >>>>>>>>>>>>>>>>>>> >
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay <
>>>> > >>>>>>>>>>>>>>>>>>> al...@google.com> wrote:
>>>> > >>>>>>>>>>>>>>>>>>> >>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is
>>>> doing a one
>>>> > >>>>>>>>>>>>>>>>>>> time cleanup. I agree
>>>> > >>>>>>>>>>>>>>>>>>> >>>> that we need to have a solution to automate
>>>> this
>>>> > >>>>>>>>>>>>>>>>>>> task or address the root
>>>> > >>>>>>>>>>>>>>>>>>> >>>> cause of the buildup.
>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał
>>>> Walenia <
>>>> > >>>>>>>>>>>>>>>>>>> michal.wale...@polidea.com>
>>>> > >>>>>>>>>>>>>>>>>>> >>>> wrote:
>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Hi there,
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins
>>>> workers
>>>> > >>>>>>>>>>>>>>>>>>> again. Nodes 1 and 7
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on
>>>> device".
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these
>>>> cases
>>>> > >>>>>>>>>>>>>>>>>>> (someone with access
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> permissions to the workers).
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are
>>>> becoming more
>>>> > >>>>>>>>>>>>>>>>>>> and more frequent
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can
>>>> this be
>>>> > >>>>>>>>>>>>>>>>>>> remedied. Can a cleanup
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow?
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Regards
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michal
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> --
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michał Walenia
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> |
>>>> Software
>>>> > >>>>>>>>>>>>>>>>>>> Engineer
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002>
>>>> <+48%20791%20432%20002> <
>>>> > >>>>>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>
>>>> <+48%20791%20432%20002>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Unique Tech
>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Check out our projects! <
>>>> > >>>>>>>>>>>>>>>>>>> https://www.polidea.com/our-work>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>>>
>>>> > >>>>>>>>>>>>>>>>>>> >>
>>>> > >>>>>>>>>>>>>>>>>>>
>>>> > >>>>>>>>>>>>>>>>>>
>>>> >
>>>>
>>>

Reply via email to