A job/pipeline is using custom workspace. The job is set-up in such a way that every build creates there own workspace inside the main/default workspace. So, if the actual workspace is *"D:\jenkins\workspace\<job_name>"*, each run of the job/pipeline is creating there own workspace like
*"D:\jenkins\workspace\<job_name>\1""D:\jenkins\workspace\<job_name>\2""D:\jenkins\workspace\<job_name>\3"* . . . This lock down of workspace is implemented as per a requirement. If the job is successful, these workspaces are deleted in a stage using the *deletedir(). *But if the job fails, i leave them for postmortem. Is there any standard/recommended way to clean up the entire workspace in a scheduled way (like once a month)? I have a pool of slaves and the scheduled job should clean in all slaves. -- You received this message because you are subscribed to the Google Groups "Jenkins Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to jenkinsci-users+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/jenkinsci-users/e4b51b51-1e4c-4ee0-9795-abd99efc3dd6%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.