Sorry, mistake while copying, [1] should be: [1] https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L63
On Tue, Jul 28, 2020 at 7:21 PM Damian Gadomski <damian.gadom...@polidea.com> wrote: > That's interesting. I didn't check that myself but all the Jenkins jobs > are configured to wipe the workspace just before the actual build happens > [1] > <https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6>. > Git SCM plugin is used for that and it enables the option called "Wipe out > repository and force clone". Docs state that it "deletes the contents of > the workspace before build and before checkout" [2] > <https://plugins.jenkins.io/git/>. Therefore I assume that removing > workspace just after the build won't change anything. > > The ./.gradle/caches/modules-2/files-2.1 dir is indeed present on the > worker machines but it's rather in /home/jenkins dir. > > damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh > 11G . > damgad@apache-ci-beam-jenkins-13:/home/jenkins/.gradle$ sudo du -sh > caches/modules-2/files-2.1 > 2.3G caches/modules-2/files-2.1 > > I can't find that directory structure inside workspaces. > > damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$ > sudo find -name "files-2.1" > damgad@apache-ci-beam-jenkins-13:/home/jenkins/jenkins-slave/workspace$ > > [1] > https://github.com/apache/beam/blob/8aca8ccc7f1a14516ad769b63845ddd4dc163d92/.test-infra/jenkins/CommonJobProperties.groovy#L6 > [2] https://plugins.jenkins.io/git/ > > On Tue, Jul 28, 2020 at 5:47 PM Kenneth Knowles <k...@apache.org> wrote: > >> Just checking - will this wipe out dependency cache? That will slow >> things down and significantly increase flakiness. If I recall correctly, >> the default Jenkins layout was: >> >> /home/jenkins/jenkins-slave/workspace/$jobname >> /home/jenkins/jenkins-slave/workspace/$jobname/.m2 >> /home/jenkins/jenkins-slave/workspace/$jobname/.git >> >> Where you can see that it did a `git clone` right into the root workspace >> directory, adjacent to .m2. This was not hygienic. One important thing was >> that `git clean` would wipe the maven cache with every build. So in >> https://github.com/apache/beam/pull/3976 we changed it to >> >> /home/jenkins/jenkins-slave/workspace/$jobname >> /home/jenkins/jenkins-slave/workspace/$jobname/.m2 >> /home/jenkins/jenkins-slave/workspace/$jobname/src/.git >> >> Now the .m2 directory survives and we do not constantly see flakes >> re-downloading deps that are immutable. This does, of course, use disk >> space. >> >> That was in the maven days. Gradle is the same except for $HOME/.m2 is >> replaced by $HOME/.gradle/caches/modules-2/files-2.1. Is Jenkins configured >> the same way so we will be wiping out the dependencies? If so, can you >> address this issue? Everything in that directory should be immutable and >> just a cache to avoid pointless re-download. >> >> Kenn >> >> On Tue, Jul 28, 2020 at 2:25 AM Damian Gadomski < >> damian.gadom...@polidea.com> wrote: >> >>> Agree with Udi, workspaces seem to be the third culprit, not yet >>> addressed in any way (until PR#12326 >>> <https://github.com/apache/beam/pull/12326> is merged). I feel that >>> it'll solve the issue of filling up the disks for a long time ;) >>> >>> I'm also OK with moving /tmp cleanup to option B, and will happily >>> investigate on proper TMPDIR config. >>> >>> >>> >>> On Tue, Jul 28, 2020 at 3:07 AM Udi Meiri <eh...@google.com> wrote: >>> >>>> What about the workspaces, which can take up 175GB in some cases (see >>>> above)? >>>> I'm working on getting them cleaned up automatically: >>>> https://github.com/apache/beam/pull/12326 >>>> >>>> My opinion is that we would get more mileage out of fixing the jobs >>>> that leave behind files in /tmp and images/containers in Docker. >>>> This would also help keep development machines clean. >>>> >>>> >>>> On Mon, Jul 27, 2020 at 5:31 PM Tyson Hamilton <tyso...@google.com> >>>> wrote: >>>> >>>>> Here is a summery of how I understand things, >>>>> >>>>> - /tmp and /var/lib/docker are the culprit for filling up disks >>>>> - inventory Jenkins job runs every 12 hours and runs a docker prune >>>>> to clean up images older than 24hr >>>>> - crontab on each machine cleans up /tmp files older than three days >>>>> weekly >>>>> >>>>> This doesn't seem to be working since we're still running out of disk >>>>> periodically and requiring manual intervention. Knobs and options we have >>>>> available: >>>>> >>>>> 1. increase frequency of deleting files >>>>> 2. decrease the number of days required to delete a file (e.g. older >>>>> than 2 days) >>>>> >>>>> The execution methods we have available are: >>>>> >>>>> A. cron >>>>> - pro: runs even if a job gets stuck in Jenkins due to full disk >>>>> - con: config baked into VM which is tough to update, not >>>>> discoverable or documented well >>>>> B. inventory job >>>>> - pro: easy to update, runs every 12h already >>>>> - con: could get stuck if Jenkins agent runs out of disk or is >>>>> otherwise stuck, tied to all other inventory job frequency >>>>> C. configure startup scripts for the VMs that set up the cron job >>>>> anytime the VM is restarted >>>>> - pro: similar to A. and easy to update >>>>> - con: similar to A. >>>>> >>>>> Between the three I prefer B. because it is consistent with other >>>>> inventory jobs. If it ends up that stuck jobs prohibit scheduling of the >>>>> inventory job often we could further investigate C to avoid having to >>>>> rebuild the VM images repeatedly. >>>>> >>>>> Any objections or comments? If not, we'll go forward with B. and >>>>> reduce the date check from 3 days to 2 days. >>>>> >>>>> >>>>> On 2020/07/24 20:13:29, Ahmet Altay <al...@google.com> wrote: >>>>> > Tests may not be doing docker cleanup. Inventory job runs a docker >>>>> prune >>>>> > every 12 hours for images older than 24 hrs [1]. Randomly looking at >>>>> one of >>>>> > the recent runs [2], it cleaned up a long list of containers >>>>> consuming >>>>> > 30+GB space. That should be just 12 hours worth of containers. >>>>> > >>>>> > [1] >>>>> > >>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L69 >>>>> > [2] >>>>> > >>>>> https://ci-beam.apache.org/job/beam_Inventory_apache-beam-jenkins-14/501/console >>>>> > >>>>> > On Fri, Jul 24, 2020 at 1:07 PM Tyson Hamilton <tyso...@google.com> >>>>> wrote: >>>>> > >>>>> > > Yes, these are on the same volume in the /var/lib/docker >>>>> directory. I'm >>>>> > > unsure if they clean up leftover images. >>>>> > > >>>>> > > On Fri, Jul 24, 2020 at 12:52 PM Udi Meiri <eh...@google.com> >>>>> wrote: >>>>> > > >>>>> > >> I forgot Docker images: >>>>> > >> >>>>> > >> ehudm@apache-ci-beam-jenkins-3:~$ sudo docker system df >>>>> > >> TYPE TOTAL ACTIVE SIZE >>>>> > >> RECLAIMABLE >>>>> > >> Images 88 9 >>>>> 125.4GB >>>>> > >> 124.2GB (99%) >>>>> > >> Containers 40 4 >>>>> 7.927GB >>>>> > >> 7.871GB (99%) >>>>> > >> Local Volumes 47 0 >>>>> 3.165GB >>>>> > >> 3.165GB (100%) >>>>> > >> Build Cache 0 0 0B >>>>> > >> 0B >>>>> > >> >>>>> > >> There are about 90 images on that machine, with all but 1 less >>>>> than 48 >>>>> > >> hours old. >>>>> > >> I think the docker test jobs need to try harder at cleaning up >>>>> their >>>>> > >> leftover images. (assuming they're already doing it?) >>>>> > >> >>>>> > >> On Fri, Jul 24, 2020 at 12:31 PM Udi Meiri <eh...@google.com> >>>>> wrote: >>>>> > >> >>>>> > >>> The additional slots (@3 directories) take up even more space >>>>> now than >>>>> > >>> before. >>>>> > >>> >>>>> > >>> I'm testing out https://github.com/apache/beam/pull/12326 which >>>>> could >>>>> > >>> help by cleaning up workspaces after a run (just started a seed >>>>> job). >>>>> > >>> >>>>> > >>> On Fri, Jul 24, 2020 at 12:13 PM Tyson Hamilton < >>>>> tyso...@google.com> >>>>> > >>> wrote: >>>>> > >>> >>>>> > >>>> 664M beam_PreCommit_JavaPortabilityApi_Commit >>>>> > >>>> 656M beam_PreCommit_JavaPortabilityApi_Commit@2 >>>>> > >>>> 611M beam_PreCommit_JavaPortabilityApi_Cron >>>>> > >>>> 616M beam_PreCommit_JavaPortabilityApiJava11_Commit >>>>> > >>>> 598M beam_PreCommit_JavaPortabilityApiJava11_Commit@2 >>>>> > >>>> 662M beam_PreCommit_JavaPortabilityApiJava11_Cron >>>>> > >>>> 2.9G beam_PreCommit_Portable_Python_Commit >>>>> > >>>> 2.9G beam_PreCommit_Portable_Python_Commit@2 >>>>> > >>>> 1.7G beam_PreCommit_Portable_Python_Commit@3 >>>>> > >>>> 3.4G beam_PreCommit_Portable_Python_Cron >>>>> > >>>> 1.9G beam_PreCommit_Python2_PVR_Flink_Commit >>>>> > >>>> 1.4G beam_PreCommit_Python2_PVR_Flink_Cron >>>>> > >>>> 1.3G beam_PreCommit_Python2_PVR_Flink_Phrase >>>>> > >>>> 6.2G beam_PreCommit_Python_Commit >>>>> > >>>> 7.5G beam_PreCommit_Python_Commit@2 >>>>> > >>>> 7.5G beam_PreCommit_Python_Cron >>>>> > >>>> 1012M beam_PreCommit_PythonDocker_Commit >>>>> > >>>> 1011M beam_PreCommit_PythonDocker_Commit@2 >>>>> > >>>> 1011M beam_PreCommit_PythonDocker_Commit@3 >>>>> > >>>> 1002M beam_PreCommit_PythonDocker_Cron >>>>> > >>>> 877M beam_PreCommit_PythonFormatter_Commit >>>>> > >>>> 988M beam_PreCommit_PythonFormatter_Cron >>>>> > >>>> 986M beam_PreCommit_PythonFormatter_Phrase >>>>> > >>>> 1.7G beam_PreCommit_PythonLint_Commit >>>>> > >>>> 2.1G beam_PreCommit_PythonLint_Cron >>>>> > >>>> 7.5G beam_PreCommit_Python_Phrase >>>>> > >>>> 346M beam_PreCommit_RAT_Commit >>>>> > >>>> 341M beam_PreCommit_RAT_Cron >>>>> > >>>> 338M beam_PreCommit_Spotless_Commit >>>>> > >>>> 339M beam_PreCommit_Spotless_Cron >>>>> > >>>> 5.5G beam_PreCommit_SQL_Commit >>>>> > >>>> 5.5G beam_PreCommit_SQL_Cron >>>>> > >>>> 5.5G beam_PreCommit_SQL_Java11_Commit >>>>> > >>>> 750M beam_PreCommit_Website_Commit >>>>> > >>>> 750M beam_PreCommit_Website_Commit@2 >>>>> > >>>> 750M beam_PreCommit_Website_Cron >>>>> > >>>> 764M beam_PreCommit_Website_Stage_GCS_Commit >>>>> > >>>> 771M beam_PreCommit_Website_Stage_GCS_Cron >>>>> > >>>> 336M beam_Prober_CommunityMetrics >>>>> > >>>> 693M beam_python_mongoio_load_test >>>>> > >>>> 339M beam_SeedJob >>>>> > >>>> 333M beam_SeedJob_Standalone >>>>> > >>>> 334M beam_sonarqube_report >>>>> > >>>> 556M beam_SQLBigQueryIO_Batch_Performance_Test_Java >>>>> > >>>> 175G total >>>>> > >>>> >>>>> > >>>> On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton < >>>>> tyso...@google.com> >>>>> > >>>> wrote: >>>>> > >>>> >>>>> > >>>>> Ya looks like something in the workspaces is taking up room: >>>>> > >>>>> >>>>> > >>>>> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc . >>>>> > >>>>> 191G . >>>>> > >>>>> 191G total >>>>> > >>>>> >>>>> > >>>>> >>>>> > >>>>> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton < >>>>> tyso...@google.com> >>>>> > >>>>> wrote: >>>>> > >>>>> >>>>> > >>>>>> Node 8 is also full. The partition that /tmp is on is here: >>>>> > >>>>>> >>>>> > >>>>>> Filesystem Size Used Avail Use% Mounted on >>>>> > >>>>>> /dev/sda1 485G 482G 2.9G 100% / >>>>> > >>>>>> >>>>> > >>>>>> however after cleaning up tmp with the crontab command, there >>>>> is only >>>>> > >>>>>> 8G usage yet it still remains 100% full: >>>>> > >>>>>> >>>>> > >>>>>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp >>>>> > >>>>>> 8.0G /tmp >>>>> > >>>>>> 8.0G total >>>>> > >>>>>> >>>>> > >>>>>> The workspaces are in the >>>>> /home/jenkins/jenkins-slave/workspace >>>>> > >>>>>> directory. When I run a du on that, it takes really long. >>>>> I'll let it keep >>>>> > >>>>>> running for a while to see if it ever returns a result but so >>>>> far this >>>>> > >>>>>> seems suspect. >>>>> > >>>>>> >>>>> > >>>>>> >>>>> > >>>>>> >>>>> > >>>>>> >>>>> > >>>>>> >>>>> > >>>>>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton < >>>>> tyso...@google.com> >>>>> > >>>>>> wrote: >>>>> > >>>>>> >>>>> > >>>>>>> Everything I've been looking at is in the /tmp dir. Where >>>>> are the >>>>> > >>>>>>> workspaces, or what are the named? >>>>> > >>>>>>> >>>>> > >>>>>>> >>>>> > >>>>>>> >>>>> > >>>>>>> >>>>> > >>>>>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com> >>>>> wrote: >>>>> > >>>>>>> >>>>> > >>>>>>>> I'm curious to what you find. Was it /tmp or the workspaces >>>>> using >>>>> > >>>>>>>> up the space? >>>>> > >>>>>>>> >>>>> > >>>>>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton < >>>>> tyso...@google.com> >>>>> > >>>>>>>> wrote: >>>>> > >>>>>>>> >>>>> > >>>>>>>>> Bleck. I just realized that it is 'offline' so that won't >>>>> work. >>>>> > >>>>>>>>> I'll clean up manually on the machine using the cron >>>>> command. >>>>> > >>>>>>>>> >>>>> > >>>>>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton < >>>>> > >>>>>>>>> tyso...@google.com> wrote: >>>>> > >>>>>>>>> >>>>> > >>>>>>>>>> Something isn't working with the current set up because >>>>> node 15 >>>>> > >>>>>>>>>> appears to be out of space and is currently 'offline' >>>>> according to Jenkins. >>>>> > >>>>>>>>>> Can someone run the cleanup job? The machine is full, >>>>> > >>>>>>>>>> >>>>> > >>>>>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h >>>>> > >>>>>>>>>> Filesystem Size Used Avail Use% Mounted on >>>>> > >>>>>>>>>> udev 52G 0 52G 0% /dev >>>>> > >>>>>>>>>> tmpfs 11G 265M 10G 3% /run >>>>> > >>>>>>>>>> */dev/sda1 485G 484G 880M 100% /* >>>>> > >>>>>>>>>> tmpfs 52G 0 52G 0% /dev/shm >>>>> > >>>>>>>>>> tmpfs 5.0M 0 5.0M 0% /run/lock >>>>> > >>>>>>>>>> tmpfs 52G 0 52G 0% /sys/fs/cgroup >>>>> > >>>>>>>>>> tmpfs 11G 0 11G 0% /run/user/1017 >>>>> > >>>>>>>>>> tmpfs 11G 0 11G 0% /run/user/1037 >>>>> > >>>>>>>>>> >>>>> > >>>>>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . | >>>>> sort -rhk >>>>> > >>>>>>>>>> 1,1 | head -n 20 >>>>> > >>>>>>>>>> 20G 2020-07-24 17:52 . >>>>> > >>>>>>>>>> 580M 2020-07-22 17:31 ./junit1031982597110125586 >>>>> > >>>>>>>>>> 517M 2020-07-22 17:31 >>>>> > >>>>>>>>>> >>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof >>>>> > >>>>>>>>>> 517M 2020-07-22 17:31 >>>>> > >>>>>>>>>> ./junit1031982597110125586/junit8739924829337821410 >>>>> > >>>>>>>>>> 263M 2020-07-22 12:23 ./pip-install-2GUhO_ >>>>> > >>>>>>>>>> 263M 2020-07-20 09:30 ./pip-install-sxgwqr >>>>> > >>>>>>>>>> 263M 2020-07-17 13:56 ./pip-install-bWSKIV >>>>> > >>>>>>>>>> 242M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T >>>>> > >>>>>>>>>> 242M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK >>>>> > >>>>>>>>>> 242M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ >>>>> > >>>>>>>>>> 236M 2020-07-21 20:25 >>>>> > >>>>>>>>>> ./beam-pipeline-tempmByU6T/tmpOWj3Yr >>>>> > >>>>>>>>>> 236M 2020-07-21 20:21 >>>>> > >>>>>>>>>> ./beam-pipeline-tempV85xeK/tmppbQHB3 >>>>> > >>>>>>>>>> 236M 2020-07-21 20:15 >>>>> > >>>>>>>>>> ./beam-pipeline-temp7dJROJ/tmpgOXPKW >>>>> > >>>>>>>>>> 111M 2020-07-23 00:57 ./pip-install-1JnyNE >>>>> > >>>>>>>>>> 105M 2020-07-23 00:17 >>>>> ./beam-artifact1374651823280819755 >>>>> > >>>>>>>>>> 105M 2020-07-23 00:16 >>>>> ./beam-artifact5050755582921936972 >>>>> > >>>>>>>>>> 105M 2020-07-23 00:16 >>>>> ./beam-artifact1834064452502646289 >>>>> > >>>>>>>>>> 105M 2020-07-23 00:15 >>>>> ./beam-artifact682561790267074916 >>>>> > >>>>>>>>>> 105M 2020-07-23 00:15 >>>>> ./beam-artifact4691304965824489394 >>>>> > >>>>>>>>>> 105M 2020-07-23 00:14 >>>>> ./beam-artifact4050383819822604421 >>>>> > >>>>>>>>>> >>>>> > >>>>>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw < >>>>> > >>>>>>>>>> rober...@google.com> wrote: >>>>> > >>>>>>>>>> >>>>> > >>>>>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton < >>>>> > >>>>>>>>>>> tyso...@google.com> wrote: >>>>> > >>>>>>>>>>> >>>>> > >>>>>>>>>>>> Ah I see, thanks Kenn. I found some advice from the >>>>> Apache >>>>> > >>>>>>>>>>>> infra wiki that also suggests using a tmpdir inside the >>>>> workspace [1]: >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> Procedures Projects can take to clean up disk space >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> Projects can help themselves and Infra by taking some >>>>> basic >>>>> > >>>>>>>>>>>> steps to help clean up their jobs after themselves on >>>>> the build nodes. >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> 1. Use a ./tmp dir in your jobs workspace. That way >>>>> it gets >>>>> > >>>>>>>>>>>> cleaned up when job workspaces expire. >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>> Tests should be (able to be) written to use the standard >>>>> > >>>>>>>>>>> temporary file mechanisms, and the environment set up on >>>>> Jenkins such that >>>>> > >>>>>>>>>>> that falls into the respective workspaces. Ideally this >>>>> should be as simple >>>>> > >>>>>>>>>>> as setting the TMPDIR (or similar) environment variable >>>>> (and making sure it >>>>> > >>>>>>>>>>> exists/is writable). >>>>> > >>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> 1. Configure your jobs to wipe workspaces on start or >>>>> > >>>>>>>>>>>> finish. >>>>> > >>>>>>>>>>>> 2. Configure your jobs to only keep 5 or 10 previous >>>>> builds. >>>>> > >>>>>>>>>>>> 3. Configure your jobs to only keep 5 or 10 previous >>>>> > >>>>>>>>>>>> artifacts. >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> [1]: >>>>> > >>>>>>>>>>>> >>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles < >>>>> > >>>>>>>>>>>> k...@apache.org> wrote: >>>>> > >>>>>>>>>>>> >>>>> > >>>>>>>>>>>>> Those file listings look like the result of using >>>>> standard >>>>> > >>>>>>>>>>>>> temp file APIs but with TMPDIR set to /tmp. >>>>> > >>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton < >>>>> > >>>>>>>>>>>>> tyso...@google.com> wrote: >>>>> > >>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique >>>>> > >>>>>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look >>>>> into two examples: >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . >>>>> | sort >>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20 >>>>> > >>>>>>>>>>>>>> 1.6G 2020-07-21 02:25 . >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:48 >>>>> ./beam-pipeline-temp3ybuY4 >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:46 >>>>> ./beam-pipeline-tempuxjiPT >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:44 >>>>> ./beam-pipeline-tempVpg1ME >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:42 >>>>> ./beam-pipeline-tempJ4EpyB >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:39 >>>>> ./beam-pipeline-tempepea7Q >>>>> > >>>>>>>>>>>>>> 242M 2020-07-17 18:35 >>>>> ./beam-pipeline-temp79qot2 >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:48 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmpy_Ytzz >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:46 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpN5_UfJ >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:44 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpxSm8pX >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:42 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpMZJU76 >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:39 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmpWy1vWX >>>>> > >>>>>>>>>>>>>> 236M 2020-07-17 18:35 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-temp79qot2/tmpvN7vWA >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:48 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmprlh_di >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:46 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpLmVWfe >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:44 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpvrxbY7 >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:42 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:39 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmptYF1v1 >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-17 18:35 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-temp79qot2/tmplfV0Rg >>>>> > >>>>>>>>>>>>>> 2.7M 2020-07-17 20:10 ./pip-install-q9l227ef >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time . >>>>> | sort >>>>> > >>>>>>>>>>>>>> -rhk 1,1 | head -n 20 >>>>> > >>>>>>>>>>>>>> 817M 2020-07-21 02:26 . >>>>> > >>>>>>>>>>>>>> 242M 2020-07-19 12:14 >>>>> ./beam-pipeline-tempUTXqlM >>>>> > >>>>>>>>>>>>>> 242M 2020-07-19 12:11 >>>>> ./beam-pipeline-tempx3Yno3 >>>>> > >>>>>>>>>>>>>> 242M 2020-07-19 12:05 >>>>> ./beam-pipeline-tempyCrMYq >>>>> > >>>>>>>>>>>>>> 236M 2020-07-19 12:14 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpstXoL0 >>>>> > >>>>>>>>>>>>>> 236M 2020-07-19 12:11 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpnnVn65 >>>>> > >>>>>>>>>>>>>> 236M 2020-07-19 12:05 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpRF0iNs >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-19 12:14 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpbJjUAQ >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-19 12:11 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpsmmzqe >>>>> > >>>>>>>>>>>>>> 3.7M 2020-07-19 12:05 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmp5b3ZvY >>>>> > >>>>>>>>>>>>>> 2.0M 2020-07-19 12:14 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpoj3orz >>>>> > >>>>>>>>>>>>>> 2.0M 2020-07-19 12:11 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmptng9sZ >>>>> > >>>>>>>>>>>>>> 2.0M 2020-07-19 12:05 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpWp6njc >>>>> > >>>>>>>>>>>>>> 1.2M 2020-07-19 12:14 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmphgdj35 >>>>> > >>>>>>>>>>>>>> 1.2M 2020-07-19 12:11 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmp8ySXpm >>>>> > >>>>>>>>>>>>>> 1.2M 2020-07-19 12:05 >>>>> > >>>>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpNVEJ4e >>>>> > >>>>>>>>>>>>>> 992K 2020-07-12 12:00 >>>>> ./junit642086915811430564 >>>>> > >>>>>>>>>>>>>> 988K 2020-07-12 12:00 >>>>> ./junit642086915811430564/beam >>>>> > >>>>>>>>>>>>>> 984K 2020-07-12 12:00 >>>>> > >>>>>>>>>>>>>> ./junit642086915811430564/beam/nodes >>>>> > >>>>>>>>>>>>>> 980K 2020-07-12 12:00 >>>>> > >>>>>>>>>>>>>> ./junit642086915811430564/beam/nodes/0 >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri < >>>>> eh...@google.com> >>>>> > >>>>>>>>>>>>>> wrote: >>>>> > >>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>> You're right, job workspaces should be hermetic. >>>>> > >>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles < >>>>> > >>>>>>>>>>>>>>> k...@apache.org> wrote: >>>>> > >>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>> I'm probably late to this discussion and missing >>>>> something, >>>>> > >>>>>>>>>>>>>>>> but why are we writing to /tmp at all? I would >>>>> expect TMPDIR to point >>>>> > >>>>>>>>>>>>>>>> somewhere inside the job directory that will be >>>>> wiped by Jenkins, and I >>>>> > >>>>>>>>>>>>>>>> would expect code to always create temp files via >>>>> APIs that respect this. >>>>> > >>>>>>>>>>>>>>>> Is Jenkins not cleaning up? Do we not have the >>>>> ability to set this up? Do >>>>> > >>>>>>>>>>>>>>>> we have bugs in our code (that we could probably >>>>> find by setting TMPDIR to >>>>> > >>>>>>>>>>>>>>>> somewhere not-/tmp and running the tests without >>>>> write permission to /tmp, >>>>> > >>>>>>>>>>>>>>>> etc) >>>>> > >>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>> Kenn >>>>> > >>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay < >>>>> > >>>>>>>>>>>>>>>> al...@google.com> wrote: >>>>> > >>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri >>>>> > >>>>>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue >>>>> previously ( >>>>> > >>>>>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865) >>>>> for >>>>> > >>>>>>>>>>>>>>>>> cleaning up workspace directory after successful >>>>> jobs. Alternatively, we >>>>> > >>>>>>>>>>>>>>>>> can consider periodically cleaning up the /src >>>>> directories. >>>>> > >>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>> I would suggest moving the cron task from internal >>>>> cron >>>>> > >>>>>>>>>>>>>>>>> scripts to the inventory job ( >>>>> > >>>>>>>>>>>>>>>>> >>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51 >>>>> ). >>>>> > >>>>>>>>>>>>>>>>> That way, we can see all the cron jobs as part of >>>>> the source tree, adjust >>>>> > >>>>>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not >>>>> know how internal cron >>>>> > >>>>>>>>>>>>>>>>> scripts are created, maintained, and how would >>>>> they be recreated for new >>>>> > >>>>>>>>>>>>>>>>> worker instances. >>>>> > >>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com> >>>>> > >>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski < >>>>> > >>>>>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote: >>>>> > >>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> Hey, >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> I've recently created a solution for the growing >>>>> /tmp >>>>> > >>>>>>>>>>>>>>>>>> directory. Part of it is the job mentioned by >>>>> Tyson: >>>>> > >>>>>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not >>>>> > >>>>>>>>>>>>>>>>>> triggered by cron and should be a last resort >>>>> solution for some strange >>>>> > >>>>>>>>>>>>>>>>>> cases. >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> Along with that job, I've also updated every >>>>> worker with >>>>> > >>>>>>>>>>>>>>>>>> an internal cron script. It's being executed once >>>>> a week and deletes all >>>>> > >>>>>>>>>>>>>>>>>> the files (and only files) that were not accessed >>>>> for at least three days. >>>>> > >>>>>>>>>>>>>>>>>> That's designed to be as safe as possible for the >>>>> running jobs on the >>>>> > >>>>>>>>>>>>>>>>>> worker (not to delete the files that are still in >>>>> use), and also to be >>>>> > >>>>>>>>>>>>>>>>>> insensitive to the current workload on the >>>>> machine. The cleanup will always >>>>> > >>>>>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are >>>>> blocking the machine. >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> I also think that currently the "No space left" >>>>> errors >>>>> > >>>>>>>>>>>>>>>>>> may be a consequence of growing workspace >>>>> directory rather than /tmp. I >>>>> > >>>>>>>>>>>>>>>>>> didn't do any detailed analysis but e.g. >>>>> currently, on >>>>> > >>>>>>>>>>>>>>>>>> apache-beam-jenkins-7 the workspace directory >>>>> size is 158 GB while /tmp is >>>>> > >>>>>>>>>>>>>>>>>> only 16 GB. We should either guarantee the disk >>>>> size to hold workspaces for >>>>> > >>>>>>>>>>>>>>>>>> all jobs (because eventually, every worker will >>>>> execute each job) or clear >>>>> > >>>>>>>>>>>>>>>>>> also the workspaces in some way. >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> Regards, >>>>> > >>>>>>>>>>>>>>>>>> Damian >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian >>>>> Michels < >>>>> > >>>>>>>>>>>>>>>>>> m...@apache.org> wrote: >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't >>>>> lead to >>>>> > >>>>>>>>>>>>>>>>>>> test failures >>>>> > >>>>>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe >>>>> there is >>>>> > >>>>>>>>>>>>>>>>>>> the notion of >>>>> > >>>>>>>>>>>>>>>>>>> running exclusively while no other tasks are >>>>> running? >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>>> -Max >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote: >>>>> > >>>>>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in >>>>> Jenkins: >>>>> > >>>>>>>>>>>>>>>>>>> beam_Clean_tmp_directory >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm >>>>> seeing some >>>>> > >>>>>>>>>>>>>>>>>>> out of disk related errors in precommit tests >>>>> currently, perhaps we should >>>>> > >>>>>>>>>>>>>>>>>>> schedule this job with cron? >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee < >>>>> > >>>>>>>>>>>>>>>>>>> heej...@google.com> wrote: >>>>> > >>>>>>>>>>>>>>>>>>> >> Still seeing no space left on device errors on >>>>> > >>>>>>>>>>>>>>>>>>> jenkins-7 (for example: >>>>> > >>>>>>>>>>>>>>>>>>> >> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/ >>>>> > >>>>>>>>>>>>>>>>>>> ) >>>>> > >>>>>>>>>>>>>>>>>>> >> >>>>> > >>>>>>>>>>>>>>>>>>> >> >>>>> > >>>>>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold < >>>>> > >>>>>>>>>>>>>>>>>>> amyrv...@google.com> wrote: >>>>> > >>>>>>>>>>>>>>>>>>> >> >>>>> > >>>>>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by >>>>> jenkins >>>>> > >>>>>>>>>>>>>>>>>>> older than 3 days. >>>>> > >>>>>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution. >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except >>>>> > >>>>>>>>>>>>>>>>>>> jenkins-12, which has not >>>>> > >>>>>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 >>>>> days. Not >>>>> > >>>>>>>>>>>>>>>>>>> scheduling: >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> Recent passing builds: >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds >>>>> > >>>>>>>>>>>>>>>>>>> >>> < >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D >>>>> > >>>>>>>>>>>>>>>>>>> > >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay < >>>>> > >>>>>>>>>>>>>>>>>>> al...@google.com> wrote: >>>>> > >>>>>>>>>>>>>>>>>>> >>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is >>>>> doing a one >>>>> > >>>>>>>>>>>>>>>>>>> time cleanup. I agree >>>>> > >>>>>>>>>>>>>>>>>>> >>>> that we need to have a solution to automate >>>>> this >>>>> > >>>>>>>>>>>>>>>>>>> task or address the root >>>>> > >>>>>>>>>>>>>>>>>>> >>>> cause of the buildup. >>>>> > >>>>>>>>>>>>>>>>>>> >>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał >>>>> Walenia < >>>>> > >>>>>>>>>>>>>>>>>>> michal.wale...@polidea.com> >>>>> > >>>>>>>>>>>>>>>>>>> >>>> wrote: >>>>> > >>>>>>>>>>>>>>>>>>> >>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Hi there, >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins >>>>> workers >>>>> > >>>>>>>>>>>>>>>>>>> again. Nodes 1 and 7 >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on >>>>> device". >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these >>>>> cases >>>>> > >>>>>>>>>>>>>>>>>>> (someone with access >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> permissions to the workers). >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are >>>>> becoming more >>>>> > >>>>>>>>>>>>>>>>>>> and more frequent >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can >>>>> this be >>>>> > >>>>>>>>>>>>>>>>>>> remedied. Can a cleanup >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow? >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Regards >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michal >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> -- >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Michał Walenia >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> | >>>>> Software >>>>> > >>>>>>>>>>>>>>>>>>> Engineer >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002> >>>>> <+48%20791%20432%20002> < >>>>> > >>>>>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002> >>>>> <+48%20791%20432%20002>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Unique Tech >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> Check out our projects! < >>>>> > >>>>>>>>>>>>>>>>>>> https://www.polidea.com/our-work> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> >>>>> > >>>>>>>>>>>>>>>>>>> >>>> >>>>> > >>>>>>>>>>>>>>>>>>> >> >>>>> > >>>>>>>>>>>>>>>>>>> >>>>> > >>>>>>>>>>>>>>>>>> >>>>> > >>>>> >>>>