664M    beam_PreCommit_JavaPortabilityApi_Commit
656M    beam_PreCommit_JavaPortabilityApi_Commit@2
611M    beam_PreCommit_JavaPortabilityApi_Cron
616M    beam_PreCommit_JavaPortabilityApiJava11_Commit
598M    beam_PreCommit_JavaPortabilityApiJava11_Commit@2
662M    beam_PreCommit_JavaPortabilityApiJava11_Cron
2.9G    beam_PreCommit_Portable_Python_Commit
2.9G    beam_PreCommit_Portable_Python_Commit@2
1.7G    beam_PreCommit_Portable_Python_Commit@3
3.4G    beam_PreCommit_Portable_Python_Cron
1.9G    beam_PreCommit_Python2_PVR_Flink_Commit
1.4G    beam_PreCommit_Python2_PVR_Flink_Cron
1.3G    beam_PreCommit_Python2_PVR_Flink_Phrase
6.2G    beam_PreCommit_Python_Commit
7.5G    beam_PreCommit_Python_Commit@2
7.5G    beam_PreCommit_Python_Cron
1012M   beam_PreCommit_PythonDocker_Commit
1011M   beam_PreCommit_PythonDocker_Commit@2
1011M   beam_PreCommit_PythonDocker_Commit@3
1002M   beam_PreCommit_PythonDocker_Cron
877M    beam_PreCommit_PythonFormatter_Commit
988M    beam_PreCommit_PythonFormatter_Cron
986M    beam_PreCommit_PythonFormatter_Phrase
1.7G    beam_PreCommit_PythonLint_Commit
2.1G    beam_PreCommit_PythonLint_Cron
7.5G    beam_PreCommit_Python_Phrase
346M    beam_PreCommit_RAT_Commit
341M    beam_PreCommit_RAT_Cron
338M    beam_PreCommit_Spotless_Commit
339M    beam_PreCommit_Spotless_Cron
5.5G    beam_PreCommit_SQL_Commit
5.5G    beam_PreCommit_SQL_Cron
5.5G    beam_PreCommit_SQL_Java11_Commit
750M    beam_PreCommit_Website_Commit
750M    beam_PreCommit_Website_Commit@2
750M    beam_PreCommit_Website_Cron
764M    beam_PreCommit_Website_Stage_GCS_Commit
771M    beam_PreCommit_Website_Stage_GCS_Cron
336M    beam_Prober_CommunityMetrics
693M    beam_python_mongoio_load_test
339M    beam_SeedJob
333M    beam_SeedJob_Standalone
334M    beam_sonarqube_report
556M    beam_SQLBigQueryIO_Batch_Performance_Test_Java
175G    total

On Fri, Jul 24, 2020 at 12:04 PM Tyson Hamilton <tyso...@google.com> wrote:

> Ya looks like something in the workspaces is taking up room:
>
> @apache-ci-beam-jenkins-8:/home/jenkins$ sudo du -shc .
> 191G    .
> 191G    total
>
>
> On Fri, Jul 24, 2020 at 11:44 AM Tyson Hamilton <tyso...@google.com>
> wrote:
>
>> Node 8 is also full. The partition that /tmp is on is here:
>>
>> Filesystem      Size  Used Avail Use% Mounted on
>> /dev/sda1       485G  482G  2.9G 100% /
>>
>> however after cleaning up tmp with the crontab command, there is only 8G
>> usage yet it still remains 100% full:
>>
>> @apache-ci-beam-jenkins-8:/tmp$ sudo du -shc /tmp
>> 8.0G    /tmp
>> 8.0G    total
>>
>> The workspaces are in the /home/jenkins/jenkins-slave/workspace
>> directory. When I run a du on that, it takes really long. I'll let it keep
>> running for a while to see if it ever returns a result but so far this
>> seems suspect.
>>
>>
>>
>>
>>
>> On Fri, Jul 24, 2020 at 11:19 AM Tyson Hamilton <tyso...@google.com>
>> wrote:
>>
>>> Everything I've been looking at is in the /tmp dir. Where are the
>>> workspaces, or what are the named?
>>>
>>>
>>>
>>>
>>> On Fri, Jul 24, 2020 at 11:03 AM Udi Meiri <eh...@google.com> wrote:
>>>
>>>> I'm curious to what you find. Was it /tmp or the workspaces using up
>>>> the space?
>>>>
>>>> On Fri, Jul 24, 2020 at 10:57 AM Tyson Hamilton <tyso...@google.com>
>>>> wrote:
>>>>
>>>>> Bleck. I just realized that it is 'offline' so that won't work. I'll
>>>>> clean up manually on the machine using the cron command.
>>>>>
>>>>> On Fri, Jul 24, 2020 at 10:56 AM Tyson Hamilton <tyso...@google.com>
>>>>> wrote:
>>>>>
>>>>>> Something isn't working with the current set up because node 15
>>>>>> appears to be out of space and is currently 'offline' according to 
>>>>>> Jenkins.
>>>>>> Can someone run the cleanup job? The machine is full,
>>>>>>
>>>>>> @apache-ci-beam-jenkins-15:/tmp$ df -h
>>>>>> Filesystem      Size  Used Avail Use% Mounted on
>>>>>> udev             52G     0   52G   0% /dev
>>>>>> tmpfs            11G  265M   10G   3% /run
>>>>>> */dev/sda1       485G  484G  880M 100% /*
>>>>>> tmpfs            52G     0   52G   0% /dev/shm
>>>>>> tmpfs           5.0M     0  5.0M   0% /run/lock
>>>>>> tmpfs            52G     0   52G   0% /sys/fs/cgroup
>>>>>> tmpfs            11G     0   11G   0% /run/user/1017
>>>>>> tmpfs            11G     0   11G   0% /run/user/1037
>>>>>>
>>>>>> apache-ci-beam-jenkins-15:/tmp$ sudo du -ah --time . | sort -rhk 1,1
>>>>>> | head -n 20
>>>>>> 20G     2020-07-24 17:52        .
>>>>>> 580M    2020-07-22 17:31        ./junit1031982597110125586
>>>>>> 517M    2020-07-22 17:31
>>>>>>  ./junit1031982597110125586/junit8739924829337821410/heap_dump.hprof
>>>>>> 517M    2020-07-22 17:31
>>>>>>  ./junit1031982597110125586/junit8739924829337821410
>>>>>> 263M    2020-07-22 12:23        ./pip-install-2GUhO_
>>>>>> 263M    2020-07-20 09:30        ./pip-install-sxgwqr
>>>>>> 263M    2020-07-17 13:56        ./pip-install-bWSKIV
>>>>>> 242M    2020-07-21 20:25        ./beam-pipeline-tempmByU6T
>>>>>> 242M    2020-07-21 20:21        ./beam-pipeline-tempV85xeK
>>>>>> 242M    2020-07-21 20:15        ./beam-pipeline-temp7dJROJ
>>>>>> 236M    2020-07-21 20:25        ./beam-pipeline-tempmByU6T/tmpOWj3Yr
>>>>>> 236M    2020-07-21 20:21        ./beam-pipeline-tempV85xeK/tmppbQHB3
>>>>>> 236M    2020-07-21 20:15        ./beam-pipeline-temp7dJROJ/tmpgOXPKW
>>>>>> 111M    2020-07-23 00:57        ./pip-install-1JnyNE
>>>>>> 105M    2020-07-23 00:17        ./beam-artifact1374651823280819755
>>>>>> 105M    2020-07-23 00:16        ./beam-artifact5050755582921936972
>>>>>> 105M    2020-07-23 00:16        ./beam-artifact1834064452502646289
>>>>>> 105M    2020-07-23 00:15        ./beam-artifact682561790267074916
>>>>>> 105M    2020-07-23 00:15        ./beam-artifact4691304965824489394
>>>>>> 105M    2020-07-23 00:14        ./beam-artifact4050383819822604421
>>>>>>
>>>>>> On Wed, Jul 22, 2020 at 12:03 PM Robert Bradshaw <rober...@google.com>
>>>>>> wrote:
>>>>>>
>>>>>>> On Wed, Jul 22, 2020 at 11:57 AM Tyson Hamilton <tyso...@google.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Ah I see, thanks Kenn. I found some advice from the Apache infra
>>>>>>>> wiki that also suggests using a tmpdir inside the workspace [1]:
>>>>>>>>
>>>>>>>> Procedures Projects can take to clean up disk space
>>>>>>>>
>>>>>>>> Projects can help themselves and Infra by taking some basic steps
>>>>>>>> to help clean up their jobs after themselves on the build nodes.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>    1. Use a ./tmp dir in your jobs workspace. That way it gets
>>>>>>>>    cleaned up when job workspaces expire.
>>>>>>>>
>>>>>>>>
>>>>>>> Tests should be (able to be) written to use the standard temporary
>>>>>>> file mechanisms, and the environment set up on Jenkins such that that 
>>>>>>> falls
>>>>>>> into the respective workspaces. Ideally this should be as simple as 
>>>>>>> setting
>>>>>>> the TMPDIR (or similar) environment variable (and making sure it 
>>>>>>> exists/is
>>>>>>> writable).
>>>>>>>
>>>>>>>>
>>>>>>>>    1. Configure your jobs to wipe workspaces on start or finish.
>>>>>>>>    2. Configure your jobs to only keep 5 or 10 previous builds.
>>>>>>>>    3. Configure your jobs to only keep 5 or 10 previous artifacts.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> [1]:
>>>>>>>> https://cwiki.apache.org/confluence/display/INFRA/Disk+Space+cleanup+of+Jenkins+nodes
>>>>>>>>
>>>>>>>> On Wed, Jul 22, 2020 at 8:06 AM Kenneth Knowles <k...@apache.org>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Those file listings look like the result of using standard temp
>>>>>>>>> file APIs but with TMPDIR set to /tmp.
>>>>>>>>>
>>>>>>>>> On Mon, Jul 20, 2020 at 7:55 PM Tyson Hamilton <tyso...@google.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Jobs are hermetic as far as I can tell and use unique
>>>>>>>>>> subdirectories inside of /tmp. Here is a quick look into two 
>>>>>>>>>> examples:
>>>>>>>>>>
>>>>>>>>>> @apache-ci-beam-jenkins-4:/tmp$ sudo du -ah --time . | sort -rhk
>>>>>>>>>> 1,1 | head -n 20
>>>>>>>>>> 1.6G    2020-07-21 02:25        .
>>>>>>>>>> 242M    2020-07-17 18:48        ./beam-pipeline-temp3ybuY4
>>>>>>>>>> 242M    2020-07-17 18:46        ./beam-pipeline-tempuxjiPT
>>>>>>>>>> 242M    2020-07-17 18:44        ./beam-pipeline-tempVpg1ME
>>>>>>>>>> 242M    2020-07-17 18:42        ./beam-pipeline-tempJ4EpyB
>>>>>>>>>> 242M    2020-07-17 18:39        ./beam-pipeline-tempepea7Q
>>>>>>>>>> 242M    2020-07-17 18:35        ./beam-pipeline-temp79qot2
>>>>>>>>>> 236M    2020-07-17 18:48
>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmpy_Ytzz
>>>>>>>>>> 236M    2020-07-17 18:46
>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpN5_UfJ
>>>>>>>>>> 236M    2020-07-17 18:44
>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpxSm8pX
>>>>>>>>>> 236M    2020-07-17 18:42
>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpMZJU76
>>>>>>>>>> 236M    2020-07-17 18:39
>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmpWy1vWX
>>>>>>>>>> 236M    2020-07-17 18:35
>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmpvN7vWA
>>>>>>>>>> 3.7M    2020-07-17 18:48
>>>>>>>>>>  ./beam-pipeline-temp3ybuY4/tmprlh_di
>>>>>>>>>> 3.7M    2020-07-17 18:46
>>>>>>>>>>  ./beam-pipeline-tempuxjiPT/tmpLmVWfe
>>>>>>>>>> 3.7M    2020-07-17 18:44
>>>>>>>>>>  ./beam-pipeline-tempVpg1ME/tmpvrxbY7
>>>>>>>>>> 3.7M    2020-07-17 18:42
>>>>>>>>>>  ./beam-pipeline-tempJ4EpyB/tmpLTb6Mj
>>>>>>>>>> 3.7M    2020-07-17 18:39
>>>>>>>>>>  ./beam-pipeline-tempepea7Q/tmptYF1v1
>>>>>>>>>> 3.7M    2020-07-17 18:35
>>>>>>>>>>  ./beam-pipeline-temp79qot2/tmplfV0Rg
>>>>>>>>>> 2.7M    2020-07-17 20:10        ./pip-install-q9l227ef
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> @apache-ci-beam-jenkins-11:/tmp$ sudo du -ah --time . | sort -rhk
>>>>>>>>>> 1,1 | head -n 20
>>>>>>>>>> 817M    2020-07-21 02:26        .
>>>>>>>>>> 242M    2020-07-19 12:14        ./beam-pipeline-tempUTXqlM
>>>>>>>>>> 242M    2020-07-19 12:11        ./beam-pipeline-tempx3Yno3
>>>>>>>>>> 242M    2020-07-19 12:05        ./beam-pipeline-tempyCrMYq
>>>>>>>>>> 236M    2020-07-19 12:14
>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpstXoL0
>>>>>>>>>> 236M    2020-07-19 12:11
>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpnnVn65
>>>>>>>>>> 236M    2020-07-19 12:05
>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpRF0iNs
>>>>>>>>>> 3.7M    2020-07-19 12:14
>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpbJjUAQ
>>>>>>>>>> 3.7M    2020-07-19 12:11
>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmpsmmzqe
>>>>>>>>>> 3.7M    2020-07-19 12:05
>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmp5b3ZvY
>>>>>>>>>> 2.0M    2020-07-19 12:14
>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmpoj3orz
>>>>>>>>>> 2.0M    2020-07-19 12:11
>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmptng9sZ
>>>>>>>>>> 2.0M    2020-07-19 12:05
>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpWp6njc
>>>>>>>>>> 1.2M    2020-07-19 12:14
>>>>>>>>>>  ./beam-pipeline-tempUTXqlM/tmphgdj35
>>>>>>>>>> 1.2M    2020-07-19 12:11
>>>>>>>>>>  ./beam-pipeline-tempx3Yno3/tmp8ySXpm
>>>>>>>>>> 1.2M    2020-07-19 12:05
>>>>>>>>>>  ./beam-pipeline-tempyCrMYq/tmpNVEJ4e
>>>>>>>>>> 992K    2020-07-12 12:00        ./junit642086915811430564
>>>>>>>>>> 988K    2020-07-12 12:00        ./junit642086915811430564/beam
>>>>>>>>>> 984K    2020-07-12 12:00
>>>>>>>>>>  ./junit642086915811430564/beam/nodes
>>>>>>>>>> 980K    2020-07-12 12:00
>>>>>>>>>>  ./junit642086915811430564/beam/nodes/0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Jul 20, 2020 at 6:46 PM Udi Meiri <eh...@google.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> You're right, job workspaces should be hermetic.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Jul 20, 2020 at 1:24 PM Kenneth Knowles <k...@apache.org>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I'm probably late to this discussion and missing something, but
>>>>>>>>>>>> why are we writing to /tmp at all? I would expect TMPDIR to point 
>>>>>>>>>>>> somewhere
>>>>>>>>>>>> inside the job directory that will be wiped by Jenkins, and I 
>>>>>>>>>>>> would expect
>>>>>>>>>>>> code to always create temp files via APIs that respect this. Is 
>>>>>>>>>>>> Jenkins not
>>>>>>>>>>>> cleaning up? Do we not have the ability to set this up? Do we have 
>>>>>>>>>>>> bugs in
>>>>>>>>>>>> our code (that we could probably find by setting TMPDIR to 
>>>>>>>>>>>> somewhere
>>>>>>>>>>>> not-/tmp and running the tests without write permission to /tmp, 
>>>>>>>>>>>> etc)
>>>>>>>>>>>>
>>>>>>>>>>>> Kenn
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Jul 20, 2020 at 11:39 AM Ahmet Altay <al...@google.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Related to workspace directory growth, +Udi Meiri
>>>>>>>>>>>>> <eh...@google.com> filed a relevant issue previously (
>>>>>>>>>>>>> https://issues.apache.org/jira/browse/BEAM-9865) for cleaning
>>>>>>>>>>>>> up workspace directory after successful jobs. Alternatively, we 
>>>>>>>>>>>>> can
>>>>>>>>>>>>> consider periodically cleaning up the /src directories.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I would suggest moving the cron task from internal cron
>>>>>>>>>>>>> scripts to the inventory job (
>>>>>>>>>>>>> https://github.com/apache/beam/blob/master/.test-infra/jenkins/job_Inventory.groovy#L51).
>>>>>>>>>>>>> That way, we can see all the cron jobs as part of the source 
>>>>>>>>>>>>> tree, adjust
>>>>>>>>>>>>> frequencies and clean up codes with PRs. I do not know how 
>>>>>>>>>>>>> internal cron
>>>>>>>>>>>>> scripts are created, maintained, and how would they be recreated 
>>>>>>>>>>>>> for new
>>>>>>>>>>>>> worker instances.
>>>>>>>>>>>>>
>>>>>>>>>>>>> /cc +Tyson Hamilton <tyso...@google.com>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 4:50 AM Damian Gadomski <
>>>>>>>>>>>>> damian.gadom...@polidea.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hey,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I've recently created a solution for the growing /tmp
>>>>>>>>>>>>>> directory. Part of it is the job mentioned by Tyson:
>>>>>>>>>>>>>> *beam_Clean_tmp_directory*. It's intentionally not triggered
>>>>>>>>>>>>>> by cron and should be a last resort solution for some strange 
>>>>>>>>>>>>>> cases.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Along with that job, I've also updated every worker with an
>>>>>>>>>>>>>> internal cron script. It's being executed once a week and 
>>>>>>>>>>>>>> deletes all the
>>>>>>>>>>>>>> files (and only files) that were not accessed for at least three 
>>>>>>>>>>>>>> days.
>>>>>>>>>>>>>> That's designed to be as safe as possible for the running jobs 
>>>>>>>>>>>>>> on the
>>>>>>>>>>>>>> worker (not to delete the files that are still in use), and also 
>>>>>>>>>>>>>> to be
>>>>>>>>>>>>>> insensitive to the current workload on the machine. The cleanup 
>>>>>>>>>>>>>> will always
>>>>>>>>>>>>>> happen, even if some long-running/stuck jobs are blocking the 
>>>>>>>>>>>>>> machine.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I also think that currently the "No space left" errors may
>>>>>>>>>>>>>> be a consequence of growing workspace directory rather than 
>>>>>>>>>>>>>> /tmp. I didn't
>>>>>>>>>>>>>> do any detailed analysis but e.g. currently, on 
>>>>>>>>>>>>>> apache-beam-jenkins-7 the
>>>>>>>>>>>>>> workspace directory size is 158 GB while /tmp is only 16 GB. We 
>>>>>>>>>>>>>> should
>>>>>>>>>>>>>> either guarantee the disk size to hold workspaces for all jobs 
>>>>>>>>>>>>>> (because
>>>>>>>>>>>>>> eventually, every worker will execute each job) or clear also the
>>>>>>>>>>>>>> workspaces in some way.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>> Damian
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Jul 20, 2020 at 10:43 AM Maximilian Michels <
>>>>>>>>>>>>>> m...@apache.org> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> +1 for scheduling it via a cron job if it won't lead to test
>>>>>>>>>>>>>>> failures
>>>>>>>>>>>>>>> while running. Not a Jenkins expert but maybe there is the
>>>>>>>>>>>>>>> notion of
>>>>>>>>>>>>>>> running exclusively while no other tasks are running?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> -Max
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 17.07.20 21:49, Tyson Hamilton wrote:
>>>>>>>>>>>>>>> > FYI there was a job introduced to do this in Jenkins:
>>>>>>>>>>>>>>> beam_Clean_tmp_directory
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> > Currently it needs to be run manually. I'm seeing some out
>>>>>>>>>>>>>>> of disk related errors in precommit tests currently, perhaps we 
>>>>>>>>>>>>>>> should
>>>>>>>>>>>>>>> schedule this job with cron?
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> > On 2020/03/11 19:31:13, Heejong Lee <heej...@google.com>
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> >> Still seeing no space left on device errors on jenkins-7
>>>>>>>>>>>>>>> (for example:
>>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>>> https://builds.apache.org/job/beam_PreCommit_PythonLint_Commit/2754/
>>>>>>>>>>>>>>> )
>>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>>> >> On Fri, Mar 6, 2020 at 7:11 PM Alan Myrvold <
>>>>>>>>>>>>>>> amyrv...@google.com> wrote:
>>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>>> >>> Did a one time cleanup of tmp files owned by jenkins
>>>>>>>>>>>>>>> older than 3 days.
>>>>>>>>>>>>>>> >>> Agree that we need a longer term solution.
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> >>> Passing recent tests on all executors except jenkins-12,
>>>>>>>>>>>>>>> which has not
>>>>>>>>>>>>>>> >>> scheduled recent builds for the past 13 days. Not
>>>>>>>>>>>>>>> scheduling:
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-12/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-12/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>> Recent passing builds:
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-1/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-1/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-2/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-2/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-3/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-3/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-4/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-4/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-5/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-5/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-6/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-6/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-7/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-7/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-8/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-8/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-9/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-9/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-10/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-10/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-11/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-11/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-13/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-13/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-14/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-14/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-15/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-15/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> https://builds.apache.org/computer/apache-beam-jenkins-16/builds
>>>>>>>>>>>>>>> >>> <
>>>>>>>>>>>>>>> https://www.google.com/url?q=https://builds.apache.org/computer/apache-beam-jenkins-16/builds&sa=D
>>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> >>> On Fri, Mar 6, 2020 at 11:54 AM Ahmet Altay <
>>>>>>>>>>>>>>> al...@google.com> wrote:
>>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>>> >>>> +Alan Myrvold <amyrv...@google.com> is doing a one
>>>>>>>>>>>>>>> time cleanup. I agree
>>>>>>>>>>>>>>> >>>> that we need to have a solution to automate this task
>>>>>>>>>>>>>>> or address the root
>>>>>>>>>>>>>>> >>>> cause of the buildup.
>>>>>>>>>>>>>>> >>>>
>>>>>>>>>>>>>>> >>>> On Thu, Mar 5, 2020 at 2:47 AM Michał Walenia <
>>>>>>>>>>>>>>> michal.wale...@polidea.com>
>>>>>>>>>>>>>>> >>>> wrote:
>>>>>>>>>>>>>>> >>>>
>>>>>>>>>>>>>>> >>>>> Hi there,
>>>>>>>>>>>>>>> >>>>> it seems we have a problem with Jenkins workers again.
>>>>>>>>>>>>>>> Nodes 1 and 7
>>>>>>>>>>>>>>> >>>>> both fail jobs with "No space left on device".
>>>>>>>>>>>>>>> >>>>> Who is the best person to contact in these cases
>>>>>>>>>>>>>>> (someone with access
>>>>>>>>>>>>>>> >>>>> permissions to the workers).
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> I also noticed that such errors are becoming more and
>>>>>>>>>>>>>>> more frequent
>>>>>>>>>>>>>>> >>>>> recently and I'd like to discuss how can this be
>>>>>>>>>>>>>>> remedied. Can a cleanup
>>>>>>>>>>>>>>> >>>>> task be automated on Jenkins somehow?
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> Regards
>>>>>>>>>>>>>>> >>>>> Michal
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> --
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> Michał Walenia
>>>>>>>>>>>>>>> >>>>> Polidea <https://www.polidea.com/> | Software Engineer
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> M: +48 791 432 002 <+48%20791%20432%20002> <
>>>>>>>>>>>>>>> +48791432002 <+48%20791%20432%20002>>
>>>>>>>>>>>>>>> >>>>> E: michal.wale...@polidea.com
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>> Unique Tech
>>>>>>>>>>>>>>> >>>>> Check out our projects! <
>>>>>>>>>>>>>>> https://www.polidea.com/our-work>
>>>>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>>>> >>>>
>>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>

Reply via email to