On Mon, Jan 10, 2022 at 4:23 PM Roy Smith <r...@panix.com> wrote:
>
> I'm starting on a project which will process every edit comment in the enwiki 
> dumps using k8s jobs on toolforge. There's 736
> enwiki-20211201-pages-meta-history*.xml*bz2 files.  Would kicking off 736 
> jobs (one per file) be a reasonable thing to do from the resource consumption 
> point of view, or would that lead to a WTF response from the sysadmins?

The Toolforge Kubernetes cluster tries to protect itself from runaway
resource consumption via quota limits. You can see the quotas for your
tool by running `kubectl describe quota` from a bastion. The default
concurrent jobs quota is 15, but you are probably going to run out of
other quota limited resources (cpu, memory, pods) long before you get
to the point of 15 concurrent jobs.

To your specific question, yes if you managed to spawn 736
simultaneous dump processing jobs it would almost certainly lead to
resource starvation across the entirety of the Toolforge Kubernetes
cluster and make many sad SREs.

Bryan
-- 
Bryan Davis              Technical Engagement      Wikimedia Foundation
Principal Software Engineer                               Boise, ID USA
[[m:User:BDavis_(WMF)]]                                      irc: bd808
_______________________________________________
Cloud mailing list -- cloud@lists.wikimedia.org
List information: 
https://lists.wikimedia.org/postorius/lists/cloud.lists.wikimedia.org/

Reply via email to