The additional slots (@3 directories) take up even more space now than before.
I'm testing out https://github.com/apache/beam/pull/12326 which could help by cleaning up workspaces after a run (just started a seed job). On Fri, Jul 24, 2020 at 12:13 PM Tyson Hamilton <tyso...@google.com> wrote: > 664M beam_PreCommit_JavaPortabilityApi_Commit > 656M beam_PreCommit_JavaPortabilityApi_Commit@2 > 611M beam_PreCommit_JavaPortabilityApi_Cron > 616M beam_PreCommit_JavaPortabilityApiJava11_Commit > 598M beam_PreCommit_JavaPortabilityApiJava11_Commit@2 > 662M beam_PreCommit_JavaPortabilityApiJava11_Cron > 2.9G beam_PreCommit_Portable_Python_Commit > 2.9G beam_PreCommit_Portable_Python_Commit@2 > 1.7G beam_PreCommit_Portable_Python_Commit@3 > 3.4G beam_PreCommit_Portable_Python_Cron > 1.9G beam_PreCommit_Python2_PVR_Flink_Commit > 1.4G beam_PreCommit_Python2_PVR_Flink_Cron > 1.3G beam_PreCommit_Python2_PVR_Flink_Phrase > 6.2G beam_PreCommit_Python_Commit > 7.5G beam_PreCommit_Python_Commit@2 > 7.5G beam_PreCommit_Python_Cron > 1012M beam_PreCommit_PythonDocker_Commit > 1011M beam_PreCommit_PythonDocker_Commit@2 > 1011M beam_PreCommit_PythonDocker_Commit@3 > 1002M beam_PreCommit_PythonDocker_Cron > 877M beam_PreCommit_PythonFormatter_Commit > 988M beam_PreCommit_PythonFormatter_Cron > 986M beam_PreCommit_PythonFormatter_Phrase > 1.7G beam_PreCommit_PythonLint_Commit > 2.1G beam_PreCommit_PythonLint_Cron > 7.5G beam_PreCommit_Python_Phrase > 346M beam_PreCommit_RAT_Commit > 341M beam_PreCommit_RAT_Cron > 338M beam_PreCommit_Spotless_Commit > 339M beam_PreCommit_Spotless_Cron > 5.5G beam_PreCommit_SQL_Commit > 5.5G beam_PreCommit_SQL_Cron > 5.5G beam_PreCommit_SQL_Java11_Commit > 750M beam_PreCommit_Website_Commit > 750M beam_PreCommit_Website_Commit@2 > 750M beam_PreCommit_Website_Cron > 764M beam_PreCommit_Website_Stage_GCS_Commit > 771M beam_PreCommit_Website_Stage_GCS_Cron > 336M beam_Prober_CommunityMetrics > 693M beam_python_mongoio_load_test > 339M beam_SeedJob > 333M beam_SeedJob_Standalone > 334M beam_sonarqube_report > 556M beam_SQLBigQueryIO_Batch_Performance_Test_Java > 175G total > > On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <tyso...@google.com> > wrote: > >> Ya looks like something in the workspaces is taking up room: >> >> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc . >> 191G . >> 191G total >> >> >> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <tyso...@google.com> >> wrote: >> >>> Node 8 is also full. The partition that /tmp is on is here: >>> >>> Filesystem Size Used Avail Use% Mounted on >>> /dev/sda1 485G 482G 2.9G 100% / >>> >>> however after cleaning up tmp with the crontab command, there is only 8G >>> usage yet it still remains 100% full: >>> >>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp >>> 8.0G /tmp >>> 8.0G total >>> >>> The workspaces are in the /home/jenkins/jenkins-slave/workspace >>> directory. When I run a du on that, it takes really long. I'll let it keep >>> running for a while to see if it ever returns a result but so far this >>> seems suspect. >>> >>> >>> >>> >>> >>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <tyso...@google.com> >>> wrote: >>> >>>> Everything I've been looking at is in the /tmp dir. Where are the >>>> workspaces, or what are the named? >>>> >>>> >>>> >>>> >>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com> wrote: >>>> >>>>> I'm curious to what you find. Was it /tmp or the workspaces using up >>>>> the space? >>>>> >>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <tyso...@google.com> >>>>> wrote: >>>>> >>>>>> Bleck. I just realized that it is 'offline' so that won't work. I'll >>>>>> clean up manually on the machine using the cron command. >>>>>> >>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <tyso...@google.com> >>>>>> wrote: >>>>>> >>>>>>> Something isn't working with the current set up because node 15 >>>>>>> appears to be out of space and is currently 'offline' according to >>>>>>> Jenkins. >>>>>>> Can someone run the cleanup job? The machine is full, >>>>>>> >>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h >>>>>>> Filesystem Size Used Avail Use% Mounted on >>>>>>> udev 52G 0 52G 0% /dev >>>>>>> tmpfs 11G 265M 10G 3% /run >>>>>>> */dev/sda1 485G 484G 880M 100% /* >>>>>>> tmpfs 52G 0 52G 0% /dev/shm >>>>>>> tmpfs 5.0M 0 5.0M 0% /run/lock >>>>>>> tmpfs 52G 0 52G 0% /sys/fs/cgroup >>>>>>> tmpfs 11G 0 11G 0% /run/user/1017 >>>>>>> tmpfs 11G 0 11G 0% /run/user/1037 >>>>>>> >>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . | sort -rhk 1,1 >>>>>>> | head -n 20 >>>>>>> 20G 2020-07-24 17:52 . >>>>>>> 580M 2020-07-22 17:31 ./junit1031982597110125586 >>>>>>> 517M 2020-07-22 17:31 >>>>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof >>>>>>> 517M 2020-07-22 17:31 >>>>>>> ./junit1031982597110125586/junit8739924829337821410 >>>>>>> 263M 2020-07-22 12:23 ./pip-install-2GUhO_ >>>>>>> 263M 2020-07-20 09:30 ./pip-install-sxgwqr >>>>>>> 263M 2020-07-17 13:56 ./pip-install-bWSKIV >>>>>>> 242M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T >>>>>>> 242M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK >>>>>>> 242M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ >>>>>>> 236M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T/tmpOWj3Yr >>>>>>> 236M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK/tmppbQHB3 >>>>>>> 236M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ/tmpgOXPKW >>>>>>> 111M 2020-07-23 00:57 ./pip-install-1JnyNE >>>>>>> 105M 2020-07-23 00:17 ./beam-artifact1374651823280819755 >>>>>>> 105M 2020-07-23 00:16 ./beam-artifact5050755582921936972 >>>>>>> 105M 2020-07-23 00:16 ./beam-artifact1834064452502646289 >>>>>>> 105M 2020-07-23 00:15 ./beam-artifact682561790267074916 >>>>>>> 105M 2020-07-23 00:15 ./beam-artifact4691304965824489394 >>>>>>> 105M 2020-07-23 00:14 ./beam-artifact4050383819822604421 >>>>>>> >>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw < >>>>>>> rober...@google.com> wrote: >>>>>>> >>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton <tyso...@google.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Ah I see, thanks Kenn. I found some advice from the Apache infra >>>>>>>>> wiki that also suggests using a tmpdir inside the workspace [1]: >>>>>>>>> >>>>>>>>> Procedures Projects can take to clean up disk space >>>>>>>>> >>>>>>>>> Projects can help themselves and Infra by taking some basic steps >>>>>>>>> to help clean up their jobs after themselves on the build nodes. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> 1. Use a ./tmp dir in your jobs workspace. That way it gets >>>>>>>>> cleaned up when job workspaces expire. >>>>>>>>> >>>>>>>>> >>>>>>>> Tests should be (able to be) written to use the standard temporary >>>>>>>> file mechanisms, and the environment set up on Jenkins such that that >>>>>>>> falls >>>>>>>> into the respective workspaces. Ideally this should be as simple as >>>>>>>> setting >>>>>>>> the TMPDIR (or similar) environment variable (and making sure it >>>>>>>> exists/is >>>>>>>> writable). >>>>>>>> >>>>>>>>> >>>>>>>>> 1. Configure your jobs to wipe workspaces on start or finish. >>>>>>>>> 2. Configure your jobs to only keep 5 or 10 previous builds. >>>>>>>>> 3. Configure your jobs to only keep 5 or 10 previous artifacts. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> [1]: >>>>>>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes >>>>>>>>> >>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <k...@apache.org> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Those file listings look like the result of using standard temp >>>>>>>>>> file APIs but with TMPDIR set to /tmp. >>>>>>>>>> >>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton < >>>>>>>>>> tyso...@google.com> wrote: >>>>>>>>>> >>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique >>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look into two >>>>>>>>>>> examples: >>>>>>>>>>> >>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . | sort -rhk >>>>>>>>>>> 1,1 | head -n 20 >>>>>>>>>>> 1.6G 2020-07-21 02:25 . >>>>>>>>>>> 242M 2020-07-17 18:48 ./beam-pipeline-temp3ybuY4 >>>>>>>>>>> 242M 2020-07-17 18:46 ./beam-pipeline-tempuxjiPT >>>>>>>>>>> 242M 2020-07-17 18:44 ./beam-pipeline-tempVpg1ME >>>>>>>>>>> 242M 2020-07-17 18:42 ./beam-pipeline-tempJ4EpyB >>>>>>>>>>> 242M 2020-07-17 18:39 ./beam-pipeline-tempepea7Q >>>>>>>>>>> 242M 2020-07-17 18:35 ./beam-pipeline-temp79qot2 >>>>>>>>>>> 236M 2020-07-17 18:48 >>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmpy_Ytzz >>>>>>>>>>> 236M 2020-07-17 18:46 >>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpN5_UfJ >>>>>>>>>>> 236M 2020-07-17 18:44 >>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpxSm8pX >>>>>>>>>>> 236M 2020-07-17 18:42 >>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpMZJU76 >>>>>>>>>>> 236M 2020-07-17 18:39 >>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmpWy1vWX >>>>>>>>>>> 236M 2020-07-17 18:35 >>>>>>>>>>> ./beam-pipeline-temp79qot2/tmpvN7vWA >>>>>>>>>>> 3.7M 2020-07-17 18:48 >>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmprlh_di >>>>>>>>>>> 3.7M 2020-07-17 18:46 >>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpLmVWfe >>>>>>>>>>> 3.7M 2020-07-17 18:44 >>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpvrxbY7 >>>>>>>>>>> 3.7M 2020-07-17 18:42 >>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj >>>>>>>>>>> 3.7M 2020-07-17 18:39 >>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmptYF1v1 >>>>>>>>>>> 3.7M 2020-07-17 18:35 >>>>>>>>>>> ./beam-pipeline-temp79qot2/tmplfV0Rg >>>>>>>>>>> 2.7M 2020-07-17 20:10 ./pip-install-q9l227ef >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time . | sort >>>>>>>>>>> -rhk 1,1 | head -n 20 >>>>>>>>>>> 817M 2020-07-21 02:26 . >>>>>>>>>>> 242M 2020-07-19 12:14 ./beam-pipeline-tempUTXqlM >>>>>>>>>>> 242M 2020-07-19 12:11 ./beam-pipeline-tempx3Yno3 >>>>>>>>>>> 242M 2020-07-19 12:05 ./beam-pipeline-tempyCrMYq >>>>>>>>>>> 236M 2020-07-19 12:14 >>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpstXoL0 >>>>>>>>>>> 236M 2020-07-19 12:11 >>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpnnVn65 >>>>>>>>>>> 236M 2020-07-19 12:05 >>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpRF0iNs >>>>>>>>>>> 3.7M 2020-07-19 12:14 >>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpbJjUAQ >>>>>>>>>>> 3.7M 2020-07-19 12:11 >>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpsmmzqe >>>>>>>>>>> 3.7M 2020-07-19 12:05 >>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmp5b3ZvY >>>>>>>>>>> 2.0M 2020-07-19 12:14 >>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpoj3orz >>>>>>>>>>> 2.0M 2020-07-19 12:11 >>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmptng9sZ >>>>>>>>>>> 2.0M 2020-07-19 12:05 >>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpWp6njc >>>>>>>>>>> 1.2M 2020-07-19 12:14 >>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmphgdj35 >>>>>>>>>>> 1.2M 2020-07-19 12:11 >>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmp8ySXpm >>>>>>>>>>> 1.2M 2020-07-19 12:05 >>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpNVEJ4e >>>>>>>>>>> 992K 2020-07-12 12:00 ./junit642086915811430564 >>>>>>>>>>> 988K 2020-07-12 12:00 ./junit642086915811430564/beam >>>>>>>>>>> 984K 2020-07-12 12:00 >>>>>>>>>>> ./junit642086915811430564/beam/nodes >>>>>>>>>>> 980K 2020-07-12 12:00 >>>>>>>>>>> ./junit642086915811430564/beam/nodes/0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <eh...@google.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> You're right, job workspaces should be hermetic. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles < >>>>>>>>>>>> k...@apache.org> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I'm probably late to this discussion and missing something, >>>>>>>>>>>>> but why are we writing to /tmp at all? I would expect TMPDIR to >>>>>>>>>>>>> point >>>>>>>>>>>>> somewhere inside the job directory that will be wiped by Jenkins, >>>>>>>>>>>>> and I >>>>>>>>>>>>> would expect code to always create temp files via APIs that >>>>>>>>>>>>> respect this. >>>>>>>>>>>>> Is Jenkins not cleaning up? Do we not have the ability to set >>>>>>>>>>>>> this up? Do >>>>>>>>>>>>> we have bugs in our code (that we could probably find by setting >>>>>>>>>>>>> TMPDIR to >>>>>>>>>>>>> somewhere not-/tmp and running the tests without write permission >>>>>>>>>>>>> to /tmp, >>>>>>>>>>>>> etc) >>>>>>>>>>>>> >>>>>>>>>>>>> Kenn >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay <al...@google.com> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri >>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue previously ( >>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865) for >>>>>>>>>>>>>> cleaning up workspace directory after successful jobs. >>>>>>>>>>>>>> Alternatively, we >>>>>>>>>>>>>> can consider periodically cleaning up the /src directories. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I would suggest moving the cron task from internal cron >>>>>>>>>>>>>> scripts to the inventory job ( >>>>>>>>>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51). >>>>>>>>>>>>>> That way, we can see all the cron jobs as part of the source >>>>>>>>>>>>>> tree, adjust >>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not know how >>>>>>>>>>>>>> internal cron >>>>>>>>>>>>>> scripts are created, maintained, and how would they be recreated >>>>>>>>>>>>>> for new >>>>>>>>>>>>>> worker instances. >>>>>>>>>>>>>> >>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski < >>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hey, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I've recently created a solution for the growing /tmp >>>>>>>>>>>>>>> directory. Part of it is the job mentioned by Tyson: >>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not >>>>>>>>>>>>>>> triggered by cron and should be a last resort solution for some >>>>>>>>>>>>>>> strange >>>>>>>>>>>>>>> cases. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Along with that job, I've also updated every worker with an >>>>>>>>>>>>>>> internal cron script. It's being executed once a week and >>>>>>>>>>>>>>> deletes all the >>>>>>>>>>>>>>> files (and only files) that were not accessed for at least >>>>>>>>>>>>>>> three days. >>>>>>>>>>>>>>> That's designed to be as safe as possible for the running jobs >>>>>>>>>>>>>>> on the >>>>>>>>>>>>>>> worker (not to delete the files that are still in use), and >>>>>>>>>>>>>>> also to be >>>>>>>>>>>>>>> insensitive to the current workload on the machine. The cleanup >>>>>>>>>>>>>>> will always >>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are blocking the >>>>>>>>>>>>>>> machine. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I also think that currently the "No space left" errors may >>>>>>>>>>>>>>> be a consequence of growing workspace directory rather than >>>>>>>>>>>>>>> /tmp. I didn't >>>>>>>>>>>>>>> do any detailed analysis but e.g. currently, on >>>>>>>>>>>>>>> apache-beam-jenkins-7 the >>>>>>>>>>>>>>> workspace directory size is 158 GB while /tmp is only 16 GB. We >>>>>>>>>>>>>>> should >>>>>>>>>>>>>>> either guarantee the disk size to hold workspaces for all jobs >>>>>>>>>>>>>>> (because >>>>>>>>>>>>>>> eventually, every worker will execute each job) or clear also >>>>>>>>>>>>>>> the >>>>>>>>>>>>>>> workspaces in some way. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>> Damian >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian Michels < >>>>>>>>>>>>>>> m...@apache.org> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't lead to >>>>>>>>>>>>>>>> test failures >>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe there is the >>>>>>>>>>>>>>>> notion of >>>>>>>>>>>>>>>> running exclusively while no other tasks are running? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -Max >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote: >>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in Jenkins: >>>>>>>>>>>>>>>> beam_Clean_tmp_directory >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm seeing some >>>>>>>>>>>>>>>> out of disk related errors in precommit tests currently, >>>>>>>>>>>>>>>> perhaps we should >>>>>>>>>>>>>>>> schedule this job with cron? >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <heej...@google.com> >>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>> >> Still seeing no space left on device errors on jenkins-7 >>>>>>>>>>>>>>>> (for example: >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/ >>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold < >>>>>>>>>>>>>>>> amyrv...@google.com> wrote: >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by jenkins >>>>>>>>>>>>>>>> older than 3 days. >>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution. >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except >>>>>>>>>>>>>>>> jenkins-12, which has not >>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 days. Not >>>>>>>>>>>>>>>> scheduling: >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> Recent passing builds: >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds >>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay < >>>>>>>>>>>>>>>> al...@google.com> wrote: >>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is doing a one >>>>>>>>>>>>>>>> time cleanup. I agree >>>>>>>>>>>>>>>> >>>> that we need to have a solution to automate this task >>>>>>>>>>>>>>>> or address the root >>>>>>>>>>>>>>>> >>>> cause of the buildup. >>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał Walenia < >>>>>>>>>>>>>>>> michal.wale...@polidea.com> >>>>>>>>>>>>>>>> >>>> wrote: >>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>> >>>>> Hi there, >>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins workers >>>>>>>>>>>>>>>> again. Nodes 1 and 7 >>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on device". >>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these cases >>>>>>>>>>>>>>>> (someone with access >>>>>>>>>>>>>>>> >>>>> permissions to the workers). >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are becoming more and >>>>>>>>>>>>>>>> more frequent >>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can this be >>>>>>>>>>>>>>>> remedied. Can a cleanup >>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow? >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> Regards >>>>>>>>>>>>>>>> >>>>> Michal >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> -- >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> Michał Walenia >>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> | Software >>>>>>>>>>>>>>>> Engineer >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002> < >>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>> >>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>>> Unique Tech >>>>>>>>>>>>>>>> >>>>> Check out our projects! < >>>>>>>>>>>>>>>> https://www.polidea.com/our-work> >>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>
smime.p7s
Description: S/MIME Cryptographic Signature