664M beam_PreCommit_JavaPortabilityApi_Commit 656M beam_PreCommit_JavaPortabilityApi_Commit@2 611M beam_PreCommit_JavaPortabilityApi_Cron 616M beam_PreCommit_JavaPortabilityApiJava11_Commit 598M beam_PreCommit_JavaPortabilityApiJava11_Commit@2 662M beam_PreCommit_JavaPortabilityApiJava11_Cron 2.9G beam_PreCommit_Portable_Python_Commit 2.9G beam_PreCommit_Portable_Python_Commit@2 1.7G beam_PreCommit_Portable_Python_Commit@3 3.4G beam_PreCommit_Portable_Python_Cron 1.9G beam_PreCommit_Python2_PVR_Flink_Commit 1.4G beam_PreCommit_Python2_PVR_Flink_Cron 1.3G beam_PreCommit_Python2_PVR_Flink_Phrase 6.2G beam_PreCommit_Python_Commit 7.5G beam_PreCommit_Python_Commit@2 7.5G beam_PreCommit_Python_Cron 1012M beam_PreCommit_PythonDocker_Commit 1011M beam_PreCommit_PythonDocker_Commit@2 1011M beam_PreCommit_PythonDocker_Commit@3 1002M beam_PreCommit_PythonDocker_Cron 877M beam_PreCommit_PythonFormatter_Commit 988M beam_PreCommit_PythonFormatter_Cron 986M beam_PreCommit_PythonFormatter_Phrase 1.7G beam_PreCommit_PythonLint_Commit 2.1G beam_PreCommit_PythonLint_Cron 7.5G beam_PreCommit_Python_Phrase 346M beam_PreCommit_RAT_Commit 341M beam_PreCommit_RAT_Cron 338M beam_PreCommit_Spotless_Commit 339M beam_PreCommit_Spotless_Cron 5.5G beam_PreCommit_SQL_Commit 5.5G beam_PreCommit_SQL_Cron 5.5G beam_PreCommit_SQL_Java11_Commit 750M beam_PreCommit_Website_Commit 750M beam_PreCommit_Website_Commit@2 750M beam_PreCommit_Website_Cron 764M beam_PreCommit_Website_Stage_GCS_Commit 771M beam_PreCommit_Website_Stage_GCS_Cron 336M beam_Prober_CommunityMetrics 693M beam_python_mongoio_load_test 339M beam_SeedJob 333M beam_SeedJob_Standalone 334M beam_sonarqube_report 556M beam_SQLBigQueryIO_Batch_Performance_Test_Java 175G total
On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <tyso...@google.com> wrote: > Ya looks like something in the workspaces is taking up room: > > @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc . > 191G . > 191G total > > > On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <tyso...@google.com> > wrote: > >> Node 8 is also full. The partition that /tmp is on is here: >> >> Filesystem Size Used Avail Use% Mounted on >> /dev/sda1 485G 482G 2.9G 100% / >> >> however after cleaning up tmp with the crontab command, there is only 8G >> usage yet it still remains 100% full: >> >> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp >> 8.0G /tmp >> 8.0G total >> >> The workspaces are in the /home/jenkins/jenkins-slave/workspace >> directory. When I run a du on that, it takes really long. I'll let it keep >> running for a while to see if it ever returns a result but so far this >> seems suspect. >> >> >> >> >> >> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <tyso...@google.com> >> wrote: >> >>> Everything I've been looking at is in the /tmp dir. Where are the >>> workspaces, or what are the named? >>> >>> >>> >>> >>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com> wrote: >>> >>>> I'm curious to what you find. Was it /tmp or the workspaces using up >>>> the space? >>>> >>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <tyso...@google.com> >>>> wrote: >>>> >>>>> Bleck. I just realized that it is 'offline' so that won't work. I'll >>>>> clean up manually on the machine using the cron command. >>>>> >>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <tyso...@google.com> >>>>> wrote: >>>>> >>>>>> Something isn't working with the current set up because node 15 >>>>>> appears to be out of space and is currently 'offline' according to >>>>>> Jenkins. >>>>>> Can someone run the cleanup job? The machine is full, >>>>>> >>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h >>>>>> Filesystem Size Used Avail Use% Mounted on >>>>>> udev 52G 0 52G 0% /dev >>>>>> tmpfs 11G 265M 10G 3% /run >>>>>> */dev/sda1 485G 484G 880M 100% /* >>>>>> tmpfs 52G 0 52G 0% /dev/shm >>>>>> tmpfs 5.0M 0 5.0M 0% /run/lock >>>>>> tmpfs 52G 0 52G 0% /sys/fs/cgroup >>>>>> tmpfs 11G 0 11G 0% /run/user/1017 >>>>>> tmpfs 11G 0 11G 0% /run/user/1037 >>>>>> >>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . | sort -rhk 1,1 >>>>>> | head -n 20 >>>>>> 20G 2020-07-24 17:52 . >>>>>> 580M 2020-07-22 17:31 ./junit1031982597110125586 >>>>>> 517M 2020-07-22 17:31 >>>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof >>>>>> 517M 2020-07-22 17:31 >>>>>> ./junit1031982597110125586/junit8739924829337821410 >>>>>> 263M 2020-07-22 12:23 ./pip-install-2GUhO_ >>>>>> 263M 2020-07-20 09:30 ./pip-install-sxgwqr >>>>>> 263M 2020-07-17 13:56 ./pip-install-bWSKIV >>>>>> 242M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T >>>>>> 242M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK >>>>>> 242M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ >>>>>> 236M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T/tmpOWj3Yr >>>>>> 236M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK/tmppbQHB3 >>>>>> 236M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ/tmpgOXPKW >>>>>> 111M 2020-07-23 00:57 ./pip-install-1JnyNE >>>>>> 105M 2020-07-23 00:17 ./beam-artifact1374651823280819755 >>>>>> 105M 2020-07-23 00:16 ./beam-artifact5050755582921936972 >>>>>> 105M 2020-07-23 00:16 ./beam-artifact1834064452502646289 >>>>>> 105M 2020-07-23 00:15 ./beam-artifact682561790267074916 >>>>>> 105M 2020-07-23 00:15 ./beam-artifact4691304965824489394 >>>>>> 105M 2020-07-23 00:14 ./beam-artifact4050383819822604421 >>>>>> >>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw <rober...@google.com> >>>>>> wrote: >>>>>> >>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton <tyso...@google.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Ah I see, thanks Kenn. I found some advice from the Apache infra >>>>>>>> wiki that also suggests using a tmpdir inside the workspace [1]: >>>>>>>> >>>>>>>> Procedures Projects can take to clean up disk space >>>>>>>> >>>>>>>> Projects can help themselves and Infra by taking some basic steps >>>>>>>> to help clean up their jobs after themselves on the build nodes. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> 1. Use a ./tmp dir in your jobs workspace. That way it gets >>>>>>>> cleaned up when job workspaces expire. >>>>>>>> >>>>>>>> >>>>>>> Tests should be (able to be) written to use the standard temporary >>>>>>> file mechanisms, and the environment set up on Jenkins such that that >>>>>>> falls >>>>>>> into the respective workspaces. Ideally this should be as simple as >>>>>>> setting >>>>>>> the TMPDIR (or similar) environment variable (and making sure it >>>>>>> exists/is >>>>>>> writable). >>>>>>> >>>>>>>> >>>>>>>> 1. Configure your jobs to wipe workspaces on start or finish. >>>>>>>> 2. Configure your jobs to only keep 5 or 10 previous builds. >>>>>>>> 3. Configure your jobs to only keep 5 or 10 previous artifacts. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> [1]: >>>>>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes >>>>>>>> >>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <k...@apache.org> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Those file listings look like the result of using standard temp >>>>>>>>> file APIs but with TMPDIR set to /tmp. >>>>>>>>> >>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton <tyso...@google.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Jobs are hermetic as far as I can tell and use unique >>>>>>>>>> subdirectories inside of /tmp. Here is a quick look into two >>>>>>>>>> examples: >>>>>>>>>> >>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . | sort -rhk >>>>>>>>>> 1,1 | head -n 20 >>>>>>>>>> 1.6G 2020-07-21 02:25 . >>>>>>>>>> 242M 2020-07-17 18:48 ./beam-pipeline-temp3ybuY4 >>>>>>>>>> 242M 2020-07-17 18:46 ./beam-pipeline-tempuxjiPT >>>>>>>>>> 242M 2020-07-17 18:44 ./beam-pipeline-tempVpg1ME >>>>>>>>>> 242M 2020-07-17 18:42 ./beam-pipeline-tempJ4EpyB >>>>>>>>>> 242M 2020-07-17 18:39 ./beam-pipeline-tempepea7Q >>>>>>>>>> 242M 2020-07-17 18:35 ./beam-pipeline-temp79qot2 >>>>>>>>>> 236M 2020-07-17 18:48 >>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmpy_Ytzz >>>>>>>>>> 236M 2020-07-17 18:46 >>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpN5_UfJ >>>>>>>>>> 236M 2020-07-17 18:44 >>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpxSm8pX >>>>>>>>>> 236M 2020-07-17 18:42 >>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpMZJU76 >>>>>>>>>> 236M 2020-07-17 18:39 >>>>>>>>>> ./beam-pipeline-tempepea7Q/tmpWy1vWX >>>>>>>>>> 236M 2020-07-17 18:35 >>>>>>>>>> ./beam-pipeline-temp79qot2/tmpvN7vWA >>>>>>>>>> 3.7M 2020-07-17 18:48 >>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmprlh_di >>>>>>>>>> 3.7M 2020-07-17 18:46 >>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpLmVWfe >>>>>>>>>> 3.7M 2020-07-17 18:44 >>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpvrxbY7 >>>>>>>>>> 3.7M 2020-07-17 18:42 >>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj >>>>>>>>>> 3.7M 2020-07-17 18:39 >>>>>>>>>> ./beam-pipeline-tempepea7Q/tmptYF1v1 >>>>>>>>>> 3.7M 2020-07-17 18:35 >>>>>>>>>> ./beam-pipeline-temp79qot2/tmplfV0Rg >>>>>>>>>> 2.7M 2020-07-17 20:10 ./pip-install-q9l227ef >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time . | sort -rhk >>>>>>>>>> 1,1 | head -n 20 >>>>>>>>>> 817M 2020-07-21 02:26 . >>>>>>>>>> 242M 2020-07-19 12:14 ./beam-pipeline-tempUTXqlM >>>>>>>>>> 242M 2020-07-19 12:11 ./beam-pipeline-tempx3Yno3 >>>>>>>>>> 242M 2020-07-19 12:05 ./beam-pipeline-tempyCrMYq >>>>>>>>>> 236M 2020-07-19 12:14 >>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpstXoL0 >>>>>>>>>> 236M 2020-07-19 12:11 >>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpnnVn65 >>>>>>>>>> 236M 2020-07-19 12:05 >>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpRF0iNs >>>>>>>>>> 3.7M 2020-07-19 12:14 >>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpbJjUAQ >>>>>>>>>> 3.7M 2020-07-19 12:11 >>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpsmmzqe >>>>>>>>>> 3.7M 2020-07-19 12:05 >>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmp5b3ZvY >>>>>>>>>> 2.0M 2020-07-19 12:14 >>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpoj3orz >>>>>>>>>> 2.0M 2020-07-19 12:11 >>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmptng9sZ >>>>>>>>>> 2.0M 2020-07-19 12:05 >>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpWp6njc >>>>>>>>>> 1.2M 2020-07-19 12:14 >>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmphgdj35 >>>>>>>>>> 1.2M 2020-07-19 12:11 >>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmp8ySXpm >>>>>>>>>> 1.2M 2020-07-19 12:05 >>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpNVEJ4e >>>>>>>>>> 992K 2020-07-12 12:00 ./junit642086915811430564 >>>>>>>>>> 988K 2020-07-12 12:00 ./junit642086915811430564/beam >>>>>>>>>> 984K 2020-07-12 12:00 >>>>>>>>>> ./junit642086915811430564/beam/nodes >>>>>>>>>> 980K 2020-07-12 12:00 >>>>>>>>>> ./junit642086915811430564/beam/nodes/0 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <eh...@google.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> You're right, job workspaces should be hermetic. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles <k...@apache.org> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> I'm probably late to this discussion and missing something, but >>>>>>>>>>>> why are we writing to /tmp at all? I would expect TMPDIR to point >>>>>>>>>>>> somewhere >>>>>>>>>>>> inside the job directory that will be wiped by Jenkins, and I >>>>>>>>>>>> would expect >>>>>>>>>>>> code to always create temp files via APIs that respect this. Is >>>>>>>>>>>> Jenkins not >>>>>>>>>>>> cleaning up? Do we not have the ability to set this up? Do we have >>>>>>>>>>>> bugs in >>>>>>>>>>>> our code (that we could probably find by setting TMPDIR to >>>>>>>>>>>> somewhere >>>>>>>>>>>> not-/tmp and running the tests without write permission to /tmp, >>>>>>>>>>>> etc) >>>>>>>>>>>> >>>>>>>>>>>> Kenn >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay <al...@google.com> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri >>>>>>>>>>>>> <eh...@google.com> filed a relevant issue previously ( >>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865) for cleaning >>>>>>>>>>>>> up workspace directory after successful jobs. Alternatively, we >>>>>>>>>>>>> can >>>>>>>>>>>>> consider periodically cleaning up the /src directories. >>>>>>>>>>>>> >>>>>>>>>>>>> I would suggest moving the cron task from internal cron >>>>>>>>>>>>> scripts to the inventory job ( >>>>>>>>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51). >>>>>>>>>>>>> That way, we can see all the cron jobs as part of the source >>>>>>>>>>>>> tree, adjust >>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not know how >>>>>>>>>>>>> internal cron >>>>>>>>>>>>> scripts are created, maintained, and how would they be recreated >>>>>>>>>>>>> for new >>>>>>>>>>>>> worker instances. >>>>>>>>>>>>> >>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com> >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski < >>>>>>>>>>>>> damian.gadom...@polidea.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hey, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I've recently created a solution for the growing /tmp >>>>>>>>>>>>>> directory. Part of it is the job mentioned by Tyson: >>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not triggered >>>>>>>>>>>>>> by cron and should be a last resort solution for some strange >>>>>>>>>>>>>> cases. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Along with that job, I've also updated every worker with an >>>>>>>>>>>>>> internal cron script. It's being executed once a week and >>>>>>>>>>>>>> deletes all the >>>>>>>>>>>>>> files (and only files) that were not accessed for at least three >>>>>>>>>>>>>> days. >>>>>>>>>>>>>> That's designed to be as safe as possible for the running jobs >>>>>>>>>>>>>> on the >>>>>>>>>>>>>> worker (not to delete the files that are still in use), and also >>>>>>>>>>>>>> to be >>>>>>>>>>>>>> insensitive to the current workload on the machine. The cleanup >>>>>>>>>>>>>> will always >>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are blocking the >>>>>>>>>>>>>> machine. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I also think that currently the "No space left" errors may >>>>>>>>>>>>>> be a consequence of growing workspace directory rather than >>>>>>>>>>>>>> /tmp. I didn't >>>>>>>>>>>>>> do any detailed analysis but e.g. currently, on >>>>>>>>>>>>>> apache-beam-jenkins-7 the >>>>>>>>>>>>>> workspace directory size is 158 GB while /tmp is only 16 GB. We >>>>>>>>>>>>>> should >>>>>>>>>>>>>> either guarantee the disk size to hold workspaces for all jobs >>>>>>>>>>>>>> (because >>>>>>>>>>>>>> eventually, every worker will execute each job) or clear also the >>>>>>>>>>>>>> workspaces in some way. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>> Damian >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian Michels < >>>>>>>>>>>>>> m...@apache.org> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't lead to test >>>>>>>>>>>>>>> failures >>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe there is the >>>>>>>>>>>>>>> notion of >>>>>>>>>>>>>>> running exclusively while no other tasks are running? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -Max >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote: >>>>>>>>>>>>>>> > FYI there was a job introduced to do this in Jenkins: >>>>>>>>>>>>>>> beam_Clean_tmp_directory >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm seeing some out >>>>>>>>>>>>>>> of disk related errors in precommit tests currently, perhaps we >>>>>>>>>>>>>>> should >>>>>>>>>>>>>>> schedule this job with cron? >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <heej...@google.com> >>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>> >> Still seeing no space left on device errors on jenkins-7 >>>>>>>>>>>>>>> (for example: >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/ >>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold < >>>>>>>>>>>>>>> amyrv...@google.com> wrote: >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by jenkins >>>>>>>>>>>>>>> older than 3 days. >>>>>>>>>>>>>>> >>> Agree that we need a longer term solution. >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> >>> Passing recent tests on all executors except jenkins-12, >>>>>>>>>>>>>>> which has not >>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 days. Not >>>>>>>>>>>>>>> scheduling: >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> Recent passing builds: >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds >>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay < >>>>>>>>>>>>>>> al...@google.com> wrote: >>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is doing a one >>>>>>>>>>>>>>> time cleanup. I agree >>>>>>>>>>>>>>> >>>> that we need to have a solution to automate this task >>>>>>>>>>>>>>> or address the root >>>>>>>>>>>>>>> >>>> cause of the buildup. >>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał Walenia < >>>>>>>>>>>>>>> michal.wale...@polidea.com> >>>>>>>>>>>>>>> >>>> wrote: >>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>> >>>>> Hi there, >>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins workers again. >>>>>>>>>>>>>>> Nodes 1 and 7 >>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on device". >>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these cases >>>>>>>>>>>>>>> (someone with access >>>>>>>>>>>>>>> >>>>> permissions to the workers). >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> I also noticed that such errors are becoming more and >>>>>>>>>>>>>>> more frequent >>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can this be >>>>>>>>>>>>>>> remedied. Can a cleanup >>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow? >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> Regards >>>>>>>>>>>>>>> >>>>> Michal >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> -- >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> Michał Walenia >>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> | Software Engineer >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002> < >>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>> >>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>>> Unique Tech >>>>>>>>>>>>>>> >>>>> Check out our projects! < >>>>>>>>>>>>>>> https://www.polidea.com/our-work> >>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>