I forgot Docker images: ehudm@apache-ci-beam-jenkins-3:~$ sudo docker system df TYPE TOTAL ACTIVE SIZE RECLAIMABLE Images 88 9 125.4GB 124.2GB (99%) Containers 40 4 7.927GB 7.871GB (99%) Local Volumes 47 0 3.165GB 3.165GB (100%) Build Cache 0 0 0B 0B
There are about 90 images on that machine, with all but 1 less than 48 hours old. I think the docker test jobs need to try harder at cleaning up their leftover images. (assuming they're already doing it?) On Fri, Jul 24, 2020 at 12:31 PM Udi Meiri <eh...@google.com> wrote: > The additional slots (@3 directories) take up even more space now than > before. > > I'm testing out https://github.com/apache/beam/pull/12326 which could > help by cleaning up workspaces after a run (just started a seed job). > > On Fri, Jul 24, 2020 at 12:13 PM Tyson Hamilton <tyso...@google.com> > wrote: > >> 664M beam_PreCommit_JavaPortabilityApi_Commit >> 656M beam_PreCommit_JavaPortabilityApi_Commit@2 >> 611M beam_PreCommit_JavaPortabilityApi_Cron >> 616M beam_PreCommit_JavaPortabilityApiJava11_Commit >> 598M beam_PreCommit_JavaPortabilityApiJava11_Commit@2 >> 662M beam_PreCommit_JavaPortabilityApiJava11_Cron >> 2.9G beam_PreCommit_Portable_Python_Commit >> 2.9G beam_PreCommit_Portable_Python_Commit@2 >> 1.7G beam_PreCommit_Portable_Python_Commit@3 >> 3.4G beam_PreCommit_Portable_Python_Cron >> 1.9G beam_PreCommit_Python2_PVR_Flink_Commit >> 1.4G beam_PreCommit_Python2_PVR_Flink_Cron >> 1.3G beam_PreCommit_Python2_PVR_Flink_Phrase >> 6.2G beam_PreCommit_Python_Commit >> 7.5G beam_PreCommit_Python_Commit@2 >> 7.5G beam_PreCommit_Python_Cron >> 1012M beam_PreCommit_PythonDocker_Commit >> 1011M beam_PreCommit_PythonDocker_Commit@2 >> 1011M beam_PreCommit_PythonDocker_Commit@3 >> 1002M beam_PreCommit_PythonDocker_Cron >> 877M beam_PreCommit_PythonFormatter_Commit >> 988M beam_PreCommit_PythonFormatter_Cron >> 986M beam_PreCommit_PythonFormatter_Phrase >> 1.7G beam_PreCommit_PythonLint_Commit >> 2.1G beam_PreCommit_PythonLint_Cron >> 7.5G beam_PreCommit_Python_Phrase >> 346M beam_PreCommit_RAT_Commit >> 341M beam_PreCommit_RAT_Cron >> 338M beam_PreCommit_Spotless_Commit >> 339M beam_PreCommit_Spotless_Cron >> 5.5G beam_PreCommit_SQL_Commit >> 5.5G beam_PreCommit_SQL_Cron >> 5.5G beam_PreCommit_SQL_Java11_Commit >> 750M beam_PreCommit_Website_Commit >> 750M beam_PreCommit_Website_Commit@2 >> 750M beam_PreCommit_Website_Cron >> 764M beam_PreCommit_Website_Stage_GCS_Commit >> 771M beam_PreCommit_Website_Stage_GCS_Cron >> 336M beam_Prober_CommunityMetrics >> 693M beam_python_mongoio_load_test >> 339M beam_SeedJob >> 333M beam_SeedJob_Standalone >> 334M beam_sonarqube_report >> 556M beam_SQLBigQueryIO_Batch_Performance_Test_Java >> 175G total >> >> On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <tyso...@google.com> >> wrote: >> >>> Ya looks like something in the workspaces is taking up room: >>> >>> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc . >>> 191G . >>> 191G total >>> >>> >>> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <tyso...@google.com> >>> wrote: >>> >>>> Node 8 is also full. The partition that /tmp is on is here: >>>> >>>> Filesystem Size Used Avail Use% Mounted on >>>> /dev/sda1 485G 482G 2.9G 100% / >>>> >>>> however after cleaning up tmp with the crontab command, there is only >>>> 8G usage yet it still remains 100% full: >>>> >>>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp >>>> 8.0G /tmp >>>> 8.0G total >>>> >>>> The workspaces are in the /home/jenkins/jenkins-slave/workspace >>>> directory. When I run a du on that, it takes really long. I'll let it keep >>>> running for a while to see if it ever returns a result but so far this >>>> seems suspect. >>>> >>>> >>>> >>>> >>>> >>>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <tyso...@google.com> >>>> wrote: >>>> >>>>> Everything I've been looking at is in the /tmp dir. Where are the >>>>> workspaces, or what are the named? >>>>> >>>>> >>>>> >>>>> >>>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com> wrote: >>>>> >>>>>> I'm curious to what you find. Was it /tmp or the workspaces using up >>>>>> the space? >>>>>> >>>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <tyso...@google.com> >>>>>> wrote: >>>>>> >>>>>>> Bleck. I just realized that it is 'offline' so that won't work. I'll >>>>>>> clean up manually on the machine using the cron command. >>>>>>> >>>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <tyso...@google.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Something isn't working with the current set up because node 15 >>>>>>>> appears to be out of space and is currently 'offline' according to >>>>>>>> Jenkins. >>>>>>>> Can someone run the cleanup job? The machine is full, >>>>>>>> >>>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h >>>>>>>> Filesystem Size Used Avail Use% Mounted on >>>>>>>> udev 52G 0 52G 0% /dev >>>>>>>> tmpfs 11G 265M 10G 3% /run >>>>>>>> */dev/sda1 485G 484G 880M 100% /* >>>>>>>> tmpfs 52G 0 52G 0% /dev/shm >>>>>>>> tmpfs 5.0M 0 5.0M 0% /run/lock >>>>>>>> tmpfs 52G 0 52G 0% /sys/fs/cgroup >>>>>>>> tmpfs 11G 0 11G 0% /run/user/1017 >>>>>>>> tmpfs 11G 0 11G 0% /run/user/1037 >>>>>>>> >>>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . | sort -rhk >>>>>>>> 1,1 | head -n 20 >>>>>>>> 20G 2020-07-24 17:52 . >>>>>>>> 580M 2020-07-22 17:31 ./junit1031982597110125586 >>>>>>>> 517M 2020-07-22 17:31 >>>>>>>> ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof >>>>>>>> 517M 2020-07-22 17:31 >>>>>>>> ./junit1031982597110125586/junit8739924829337821410 >>>>>>>> 263M 2020-07-22 12:23 ./pip-install-2GUhO_ >>>>>>>> 263M 2020-07-20 09:30 ./pip-install-sxgwqr >>>>>>>> 263M 2020-07-17 13:56 ./pip-install-bWSKIV >>>>>>>> 242M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T >>>>>>>> 242M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK >>>>>>>> 242M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ >>>>>>>> 236M 2020-07-21 20:25 ./beam-pipeline-tempmByU6T/tmpOWj3Yr >>>>>>>> 236M 2020-07-21 20:21 ./beam-pipeline-tempV85xeK/tmppbQHB3 >>>>>>>> 236M 2020-07-21 20:15 ./beam-pipeline-temp7dJROJ/tmpgOXPKW >>>>>>>> 111M 2020-07-23 00:57 ./pip-install-1JnyNE >>>>>>>> 105M 2020-07-23 00:17 ./beam-artifact1374651823280819755 >>>>>>>> 105M 2020-07-23 00:16 ./beam-artifact5050755582921936972 >>>>>>>> 105M 2020-07-23 00:16 ./beam-artifact1834064452502646289 >>>>>>>> 105M 2020-07-23 00:15 ./beam-artifact682561790267074916 >>>>>>>> 105M 2020-07-23 00:15 ./beam-artifact4691304965824489394 >>>>>>>> 105M 2020-07-23 00:14 ./beam-artifact4050383819822604421 >>>>>>>> >>>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw < >>>>>>>> rober...@google.com> wrote: >>>>>>>> >>>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton < >>>>>>>>> tyso...@google.com> wrote: >>>>>>>>> >>>>>>>>>> Ah I see, thanks Kenn. I found some advice from the Apache infra >>>>>>>>>> wiki that also suggests using a tmpdir inside the workspace [1]: >>>>>>>>>> >>>>>>>>>> Procedures Projects can take to clean up disk space >>>>>>>>>> >>>>>>>>>> Projects can help themselves and Infra by taking some basic steps >>>>>>>>>> to help clean up their jobs after themselves on the build nodes. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 1. Use a ./tmp dir in your jobs workspace. That way it gets >>>>>>>>>> cleaned up when job workspaces expire. >>>>>>>>>> >>>>>>>>>> >>>>>>>>> Tests should be (able to be) written to use the standard temporary >>>>>>>>> file mechanisms, and the environment set up on Jenkins such that that >>>>>>>>> falls >>>>>>>>> into the respective workspaces. Ideally this should be as simple as >>>>>>>>> setting >>>>>>>>> the TMPDIR (or similar) environment variable (and making sure it >>>>>>>>> exists/is >>>>>>>>> writable). >>>>>>>>> >>>>>>>>>> >>>>>>>>>> 1. Configure your jobs to wipe workspaces on start or finish. >>>>>>>>>> 2. Configure your jobs to only keep 5 or 10 previous builds. >>>>>>>>>> 3. Configure your jobs to only keep 5 or 10 previous >>>>>>>>>> artifacts. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> [1]: >>>>>>>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes >>>>>>>>>> >>>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <k...@apache.org> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Those file listings look like the result of using standard temp >>>>>>>>>>> file APIs but with TMPDIR set to /tmp. >>>>>>>>>>> >>>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton < >>>>>>>>>>> tyso...@google.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique >>>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look into two >>>>>>>>>>>> examples: >>>>>>>>>>>> >>>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . | sort >>>>>>>>>>>> -rhk 1,1 | head -n 20 >>>>>>>>>>>> 1.6G 2020-07-21 02:25 . >>>>>>>>>>>> 242M 2020-07-17 18:48 ./beam-pipeline-temp3ybuY4 >>>>>>>>>>>> 242M 2020-07-17 18:46 ./beam-pipeline-tempuxjiPT >>>>>>>>>>>> 242M 2020-07-17 18:44 ./beam-pipeline-tempVpg1ME >>>>>>>>>>>> 242M 2020-07-17 18:42 ./beam-pipeline-tempJ4EpyB >>>>>>>>>>>> 242M 2020-07-17 18:39 ./beam-pipeline-tempepea7Q >>>>>>>>>>>> 242M 2020-07-17 18:35 ./beam-pipeline-temp79qot2 >>>>>>>>>>>> 236M 2020-07-17 18:48 >>>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmpy_Ytzz >>>>>>>>>>>> 236M 2020-07-17 18:46 >>>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpN5_UfJ >>>>>>>>>>>> 236M 2020-07-17 18:44 >>>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpxSm8pX >>>>>>>>>>>> 236M 2020-07-17 18:42 >>>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpMZJU76 >>>>>>>>>>>> 236M 2020-07-17 18:39 >>>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmpWy1vWX >>>>>>>>>>>> 236M 2020-07-17 18:35 >>>>>>>>>>>> ./beam-pipeline-temp79qot2/tmpvN7vWA >>>>>>>>>>>> 3.7M 2020-07-17 18:48 >>>>>>>>>>>> ./beam-pipeline-temp3ybuY4/tmprlh_di >>>>>>>>>>>> 3.7M 2020-07-17 18:46 >>>>>>>>>>>> ./beam-pipeline-tempuxjiPT/tmpLmVWfe >>>>>>>>>>>> 3.7M 2020-07-17 18:44 >>>>>>>>>>>> ./beam-pipeline-tempVpg1ME/tmpvrxbY7 >>>>>>>>>>>> 3.7M 2020-07-17 18:42 >>>>>>>>>>>> ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj >>>>>>>>>>>> 3.7M 2020-07-17 18:39 >>>>>>>>>>>> ./beam-pipeline-tempepea7Q/tmptYF1v1 >>>>>>>>>>>> 3.7M 2020-07-17 18:35 >>>>>>>>>>>> ./beam-pipeline-temp79qot2/tmplfV0Rg >>>>>>>>>>>> 2.7M 2020-07-17 20:10 ./pip-install-q9l227ef >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time . | sort >>>>>>>>>>>> -rhk 1,1 | head -n 20 >>>>>>>>>>>> 817M 2020-07-21 02:26 . >>>>>>>>>>>> 242M 2020-07-19 12:14 ./beam-pipeline-tempUTXqlM >>>>>>>>>>>> 242M 2020-07-19 12:11 ./beam-pipeline-tempx3Yno3 >>>>>>>>>>>> 242M 2020-07-19 12:05 ./beam-pipeline-tempyCrMYq >>>>>>>>>>>> 236M 2020-07-19 12:14 >>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpstXoL0 >>>>>>>>>>>> 236M 2020-07-19 12:11 >>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpnnVn65 >>>>>>>>>>>> 236M 2020-07-19 12:05 >>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpRF0iNs >>>>>>>>>>>> 3.7M 2020-07-19 12:14 >>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpbJjUAQ >>>>>>>>>>>> 3.7M 2020-07-19 12:11 >>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmpsmmzqe >>>>>>>>>>>> 3.7M 2020-07-19 12:05 >>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmp5b3ZvY >>>>>>>>>>>> 2.0M 2020-07-19 12:14 >>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmpoj3orz >>>>>>>>>>>> 2.0M 2020-07-19 12:11 >>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmptng9sZ >>>>>>>>>>>> 2.0M 2020-07-19 12:05 >>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpWp6njc >>>>>>>>>>>> 1.2M 2020-07-19 12:14 >>>>>>>>>>>> ./beam-pipeline-tempUTXqlM/tmphgdj35 >>>>>>>>>>>> 1.2M 2020-07-19 12:11 >>>>>>>>>>>> ./beam-pipeline-tempx3Yno3/tmp8ySXpm >>>>>>>>>>>> 1.2M 2020-07-19 12:05 >>>>>>>>>>>> ./beam-pipeline-tempyCrMYq/tmpNVEJ4e >>>>>>>>>>>> 992K 2020-07-12 12:00 ./junit642086915811430564 >>>>>>>>>>>> 988K 2020-07-12 12:00 ./junit642086915811430564/beam >>>>>>>>>>>> 984K 2020-07-12 12:00 >>>>>>>>>>>> ./junit642086915811430564/beam/nodes >>>>>>>>>>>> 980K 2020-07-12 12:00 >>>>>>>>>>>> ./junit642086915811430564/beam/nodes/0 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <eh...@google.com> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> You're right, job workspaces should be hermetic. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles < >>>>>>>>>>>>> k...@apache.org> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> I'm probably late to this discussion and missing something, >>>>>>>>>>>>>> but why are we writing to /tmp at all? I would expect TMPDIR to >>>>>>>>>>>>>> point >>>>>>>>>>>>>> somewhere inside the job directory that will be wiped by >>>>>>>>>>>>>> Jenkins, and I >>>>>>>>>>>>>> would expect code to always create temp files via APIs that >>>>>>>>>>>>>> respect this. >>>>>>>>>>>>>> Is Jenkins not cleaning up? Do we not have the ability to set >>>>>>>>>>>>>> this up? Do >>>>>>>>>>>>>> we have bugs in our code (that we could probably find by setting >>>>>>>>>>>>>> TMPDIR to >>>>>>>>>>>>>> somewhere not-/tmp and running the tests without write >>>>>>>>>>>>>> permission to /tmp, >>>>>>>>>>>>>> etc) >>>>>>>>>>>>>> >>>>>>>>>>>>>> Kenn >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay < >>>>>>>>>>>>>> al...@google.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri >>>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue previously ( >>>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865) for >>>>>>>>>>>>>>> cleaning up workspace directory after successful jobs. >>>>>>>>>>>>>>> Alternatively, we >>>>>>>>>>>>>>> can consider periodically cleaning up the /src directories. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I would suggest moving the cron task from internal cron >>>>>>>>>>>>>>> scripts to the inventory job ( >>>>>>>>>>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51). >>>>>>>>>>>>>>> That way, we can see all the cron jobs as part of the source >>>>>>>>>>>>>>> tree, adjust >>>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not know how >>>>>>>>>>>>>>> internal cron >>>>>>>>>>>>>>> scripts are created, maintained, and how would they be >>>>>>>>>>>>>>> recreated for new >>>>>>>>>>>>>>> worker instances. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski < >>>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hey, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I've recently created a solution for the growing /tmp >>>>>>>>>>>>>>>> directory. Part of it is the job mentioned by Tyson: >>>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not >>>>>>>>>>>>>>>> triggered by cron and should be a last resort solution for >>>>>>>>>>>>>>>> some strange >>>>>>>>>>>>>>>> cases. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Along with that job, I've also updated every worker with an >>>>>>>>>>>>>>>> internal cron script. It's being executed once a week and >>>>>>>>>>>>>>>> deletes all the >>>>>>>>>>>>>>>> files (and only files) that were not accessed for at least >>>>>>>>>>>>>>>> three days. >>>>>>>>>>>>>>>> That's designed to be as safe as possible for the running jobs >>>>>>>>>>>>>>>> on the >>>>>>>>>>>>>>>> worker (not to delete the files that are still in use), and >>>>>>>>>>>>>>>> also to be >>>>>>>>>>>>>>>> insensitive to the current workload on the machine. The >>>>>>>>>>>>>>>> cleanup will always >>>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are blocking the >>>>>>>>>>>>>>>> machine. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I also think that currently the "No space left" errors may >>>>>>>>>>>>>>>> be a consequence of growing workspace directory rather than >>>>>>>>>>>>>>>> /tmp. I didn't >>>>>>>>>>>>>>>> do any detailed analysis but e.g. currently, on >>>>>>>>>>>>>>>> apache-beam-jenkins-7 the >>>>>>>>>>>>>>>> workspace directory size is 158 GB while /tmp is only 16 GB. >>>>>>>>>>>>>>>> We should >>>>>>>>>>>>>>>> either guarantee the disk size to hold workspaces for all jobs >>>>>>>>>>>>>>>> (because >>>>>>>>>>>>>>>> eventually, every worker will execute each job) or clear also >>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>> workspaces in some way. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>> Damian >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian Michels < >>>>>>>>>>>>>>>> m...@apache.org> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't lead to >>>>>>>>>>>>>>>>> test failures >>>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe there is the >>>>>>>>>>>>>>>>> notion of >>>>>>>>>>>>>>>>> running exclusively while no other tasks are running? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -Max >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote: >>>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in Jenkins: >>>>>>>>>>>>>>>>> beam_Clean_tmp_directory >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm seeing some >>>>>>>>>>>>>>>>> out of disk related errors in precommit tests currently, >>>>>>>>>>>>>>>>> perhaps we should >>>>>>>>>>>>>>>>> schedule this job with cron? >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <heej...@google.com> >>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>> >> Still seeing no space left on device errors on >>>>>>>>>>>>>>>>> jenkins-7 (for example: >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/ >>>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold < >>>>>>>>>>>>>>>>> amyrv...@google.com> wrote: >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by jenkins >>>>>>>>>>>>>>>>> older than 3 days. >>>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution. >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except >>>>>>>>>>>>>>>>> jenkins-12, which has not >>>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 days. Not >>>>>>>>>>>>>>>>> scheduling: >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> Recent passing builds: >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds >>>>>>>>>>>>>>>>> >>> < >>>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay < >>>>>>>>>>>>>>>>> al...@google.com> wrote: >>>>>>>>>>>>>>>>> >>> >>>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is doing a one >>>>>>>>>>>>>>>>> time cleanup. I agree >>>>>>>>>>>>>>>>> >>>> that we need to have a solution to automate this task >>>>>>>>>>>>>>>>> or address the root >>>>>>>>>>>>>>>>> >>>> cause of the buildup. >>>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał Walenia < >>>>>>>>>>>>>>>>> michal.wale...@polidea.com> >>>>>>>>>>>>>>>>> >>>> wrote: >>>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>>> >>>>> Hi there, >>>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins workers >>>>>>>>>>>>>>>>> again. Nodes 1 and 7 >>>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on device". >>>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these cases >>>>>>>>>>>>>>>>> (someone with access >>>>>>>>>>>>>>>>> >>>>> permissions to the workers). >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are becoming more >>>>>>>>>>>>>>>>> and more frequent >>>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can this be >>>>>>>>>>>>>>>>> remedied. Can a cleanup >>>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow? >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> Regards >>>>>>>>>>>>>>>>> >>>>> Michal >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> -- >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> Michał Walenia >>>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> | Software >>>>>>>>>>>>>>>>> Engineer >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002> < >>>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>> >>>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>>> Unique Tech >>>>>>>>>>>>>>>>> >>>>> Check out our projects! < >>>>>>>>>>>>>>>>> https://www.polidea.com/our-work> >>>>>>>>>>>>>>>>> >>>>> >>>>>>>>>>>>>>>>> >>>> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>
smime.p7s
Description: S/MIME Cryptographic Signature