You can stop pulp_resource_manager, pulp_workers, and qpidd. Then start qpidd, pulp_resource_manager, and pulp_workers. The workers will mark the tasks as canceled in the database when they start and qpid should drain the queue when it is restarted.
On Fri, Nov 10, 2017 at 6:33 PM, Dustin McNabb <[email protected]> wrote: > Anyone ever encountered anything like this? I found 67K pulp tasks pending > on one of my pulp servers today. I’m running pulp 2.10.3-1 (I know I know > I’m fixing that) with mongoDB 2.6 on RHEL7 with a couple hundred yum repos > and one python repo. We recently developed a config mgmt. state to manage > the pulp repos on all our pulp servers, and that appears to have caused > this issue by submitting a large number of pulp tasks in a short period and > then trying again 30 minutes later. My question now, is how can I kill so > many pulp tasks in a more efficient manner than the for loop I’m using? > > # A whole lotta pulp tasks > [root@pulp-server :~]# pulp-admin tasks list |grep 'Task Id' |wc -l > 67016 > > # A for loop to generate a list of the task IDs and cancel one at a time. > for n in `pulp-admin tasks list |grep 'Task Id' |awk '{print $NF}'`; do > echo $n; pulp-admin tasks cancel --task-id $n; done > > Thanks > Dustin > > _______________________________________________ > Pulp-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/pulp-list >
_______________________________________________ Pulp-list mailing list [email protected] https://www.redhat.com/mailman/listinfo/pulp-list
