Slight correction: (in the middle of my email):
I did grep /var/log/httpd/ssl_access_log *for my ta**sk**-id
(**f3a4a828-1a98-4d30-989f-a8f62acc29f9)* in /var/log/* and
/var/log/*/*, but it only found it in /var/log/httpd/ssl_access_log and
/var/log/httpd/ssl_request_log (lines like:
/var/log/httpd/ssl_access_log:135.1.215.52 - - [12/Apr/2016:15:04:17
-0500] "GET /pulp/api/v2/tasks/f3a4a828-1a98-4d30-989f-a8f62acc29f9/
HTTP/1.1" 200 579
regards,
Paul
On 04/12/2016 03:09 PM, EXT Paul Jochum wrote:
Hi Brian:
I am running into the same problem of "waiting", so here are my
outputs (let me know if you need anything else):
-sh-4.1# pulp-admin -vv tasks details --task-id
f3a4a828-1a98-4d30-989f-a8f62acc29f9
+----------------------------------------------------------------------+
Task Details
+----------------------------------------------------------------------+
2016-04-12 14:59:46,539 - DEBUG - sending GET request to
/pulp/api/v2/tasks/f3a4a828-1a98-4d30-989f-a8f62acc29f9/
2016-04-12 14:59:46,602 - INFO - GET request to
/pulp/api/v2/tasks/f3a4a828-1a98-4d30-989f-a8f62acc29f9/ with
parameters None
2016-04-12 14:59:46,602 - INFO - Response status : 200
2016-04-12 14:59:46,602 - INFO - Response body :
{
"exception": null,
"task_type": "pulp.server.managers.repo.sync.sync",
"_href": "/pulp/api/v2/tasks/f3a4a828-1a98-4d30-989f-a8f62acc29f9/",
"task_id": "f3a4a828-1a98-4d30-989f-a8f62acc29f9",
"tags": [
"pulp:repository:rhel-i386-server-optional-6.5.z-debuginfo",
"pulp:action:sync"
],
"finish_time": null,
"_ns": "task_status",
"start_time": null,
"traceback": null,
"spawned_tasks": [],
"progress_report": {},
"queue": "None.dq",
"state": "waiting",
"worker_name": null,
"result": null,
"error": null,
"_id": {
"$oid": "570d1727e75b4adec6bd0d58"
},
"id": "570d1727e75b4adec6bd0d58"
}
Operations: sync
Resources: rhel-i386-server-optional-6.5.z-debuginfo (repository)
State: Waiting
Start Time: Unstarted
Finish Time: Incomplete
Result: Incomplete
Task Id: f3a4a828-1a98-4d30-989f-a8f62acc29f9
Progress Report:
-sh-4.1# ps -awfux | grep celery
Warning: bad syntax, perhaps a bogus '-'? See
/usr/share/doc/procps-3.2.8/FAQ
root 16147 0.0 0.0 6444 704 pts/2 S+ 15:01 0:00 \_ grep
celery
apache 977 0.1 0.1 561608 25908 ? Sl Apr09 6:15
/usr/bin/python /usr/bin/celery beat
--app=pulp.server.async.celery_instance.celery
--scheduler=pulp.server.async.scheduler.Scheduler
--workdir=/var/run/pulp/ -f /var/log/pulp/celerybeat.log -l INFO
--detach --pidfile=/var/run/pulp/celerybeat.pid
apache 9558 0.0 0.4 553760 58972 ? Sl 10:40 0:05
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n resource_mana...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO -Q resource_manager
--logfile=/var/log/pulp/resource_manager.log
--pidfile=/var/run/pulp/resource_manager.pid
apache 9602 0.4 0.3 472928 50512 ? Sl 10:40 1:16 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n resource_mana...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO -Q resource_manager
--logfile=/var/log/pulp/resource_manager.log
--pidfile=/var/run/pulp/resource_manager.pid
apache 9792 0.0 0.4 553976 58924 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-0.log
--pidfile=/var/run/pulp/reserved_resource_worker-0.pid
apache 9865 0.0 0.3 472916 49692 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-0.log
--pidfile=/var/run/pulp/reserved_resource_worker-0.pid
apache 9807 0.0 0.4 553976 58920 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-1.log
--pidfile=/var/run/pulp/reserved_resource_worker-1.pid
apache 9887 0.0 0.3 472916 49688 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-1.log
--pidfile=/var/run/pulp/reserved_resource_worker-1.pid
apache 9826 0.0 0.4 553564 58896 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-2.log
--pidfile=/var/run/pulp/reserved_resource_worker-2.pid
apache 9893 0.0 0.3 472924 49684 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-2.log
--pidfile=/var/run/pulp/reserved_resource_worker-2.pid
apache 9847 0.0 0.4 553976 58912 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-3.log
--pidfile=/var/run/pulp/reserved_resource_worker-3.pid
apache 9917 0.0 0.3 472916 49692 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-3.log
--pidfile=/var/run/pulp/reserved_resource_worker-3.pid
apache 9871 0.0 0.4 553752 58896 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-4.log
--pidfile=/var/run/pulp/reserved_resource_worker-4.pid
apache 9945 0.0 0.3 472920 49684 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-4.log
--pidfile=/var/run/pulp/reserved_resource_worker-4.pid
apache 9896 0.0 0.4 553988 58932 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-5.log
--pidfile=/var/run/pulp/reserved_resource_worker-5.pid
apache 9982 0.0 0.3 472924 49700 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-5.log
--pidfile=/var/run/pulp/reserved_resource_worker-5.pid
apache 9918 0.0 0.4 553972 58916 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-6.log
--pidfile=/var/run/pulp/reserved_resource_worker-6.pid
apache 9990 0.0 0.3 472912 49680 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-6.log
--pidfile=/var/run/pulp/reserved_resource_worker-6.pid
apache 9940 0.0 0.4 553756 58892 ? Sl 10:41 0:06
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-7.log
--pidfile=/var/run/pulp/reserved_resource_worker-7.pid
apache 10013 0.0 0.3 472920 49684 ? S 10:41 0:00 \_
/usr/bin/python -m celery.__main__ worker --heartbeat-interval=30 -c 1
-n reserved_resource_worke...@lss-pulp01.ih.lucent.com --events
--app=pulp.server.async.app --loglevel=INFO
--logfile=/var/log/pulp/reserved_resource_worker-7.log
--pidfile=/var/run/pulp/reserved_resource_worker-7.pid
-sh-4.1#
I did grep /var/log/httpd/ssl_access_log in /var/log/* and
/var/log/*/*, but it only found it in /var/log/httpd/ssl_access_log
and /var/log/httpd/ssl_request_log (lines like:
/var/log/httpd/ssl_access_log:135.1.215.52 - - [12/Apr/2016:15:04:17
-0500] "GET /pulp/api/v2/tasks/f3a4a828-1a98-4d30-989f-a8f62acc29f9/
HTTP/1.1" 200 579
-sh-4.1# rpm -q -a | grep -i pulp | sort
mod_wsgi-3.4-2.pulp.el6.x86_64
pulp-admin-client-2.8.1-1.el6.noarch
pulp-puppet-admin-extensions-2.8.1-1.el6.noarch
pulp-puppet-plugins-2.8.1-1.el6.noarch
pulp-rpm-admin-extensions-2.8.1-1.el6.noarch
pulp-rpm-plugins-2.8.1-1.el6.noarch
pulp-selinux-2.8.1-1.el6.noarch
pulp-server-2.8.1-1.el6.noarch
python-isodate-0.5.0-4.pulp.el6.noarch
python-kombu-3.0.33-4.pulp.el6.noarch
python-pulp-bindings-2.8.1-1.el6.noarch
python-pulp-client-lib-2.8.1-1.el6.noarch
python-pulp-common-2.8.1-1.el6.noarch
python-pulp-oid_validation-2.8.1-1.el6.noarch
python-pulp-puppet-common-2.8.1-1.el6.noarch
python-pulp-repoauth-2.8.1-1.el6.noarch
python-pulp-rpm-common-2.8.1-1.el6.noarch
-sh-4.1# uname -a
Linux lss-pulp01.ih.lucent.com 2.6.32-042stab113.21 #1 SMP Wed Mar 23
11:05:25 MSK 2016 x86_64 x86_64 x86_64 GNU/Linux
-sh-4.1# more /etc/*release*
::::::::::::::
/etc/redhat-release
::::::::::::::
Scientific Linux release 6.7 (Carbon)
::::::::::::::
/etc/system-release
::::::::::::::
Scientific Linux release 6.7 (Carbon)
::::::::::::::
/etc/system-release-cpe
::::::::::::::
cpe:/o:redhat:enterprise_linux:6.7:ga
-sh-4.1#
On 04/11/2016 03:32 PM, Brian Bouterse wrote:
To anyone who has a stuck task running Pulp 2.6 or newer, can you open a
bug or e-mail the list with the following. The full output from:
pulp-admin -vv tasks details --task-id <the_stuck_task_UUID>
Post the full output, but observe the value of "worker_name" similar to
what was suggested here[0].
Can you also include the `sudo ps -awfux | grep celery` output to show
your running process?
Also can you grep the logs for that UUID and include any lines that you
see as output too.
[0]: https://www.redhat.com/archives/pulp-list/2016-April/msg00026.html
Thanks,
Brian
On 04/11/2016 01:32 PM, Mihai Ibanescu wrote:
I have a similar problem on 2.7.1 with rabbitmq.
Occasionally a task will get "lost":
pulp-admin tasks details --task-id 622041ac-e9e4-4a15-bd7c-7c98a17782e0
+----------------------------------------------------------------------+
Task Details
+----------------------------------------------------------------------+
Operations: publish
Resources: thirdparty-snapshot-rpm-latest (repository)
State: Waiting
Start Time: Unstarted
Finish Time: Incomplete
Result: Incomplete
Task Id: 622041ac-e9e4-4a15-bd7c-7c98a17782e0
Progress Report:
I have seen more than one getting in this state.
I can cancel the task, but I'd like to find out if I can tickle
something to re-process the queue. Other tasks do get correctly
processed.
If there's something you would like me to do in order to debug this
further, please let me know.
Thank you!
Mihai
On Fri, Apr 8, 2016 at 9:27 AM, Brian Bouterse <bbout...@redhat.com
<mailto:bbout...@redhat.com>> wrote:
There is a known issue [0] introduced in 2.8.0 which will cause
your
tasks to become stuck at waiting. This occurs after Pulp runs
normally
for hours or days.
A fix is completed but won't be released until Pulp 2.8.3 is
released.
The good news is, you can resolve that issue now by upgrading
python-kombu on your existing 2.8.0, 2.8.1, or 2.8.2 system.
Manually
install python-kombu-3.0.33-5.pulp for your distro on your system,
restart Pulp services and you'll be all set. All of the links
below come
out of the Pulp build system[1]
EL6: http://tinyurl.com/j5sy5zh
EL7: http://tinyurl.com/jppwcdv
F22: http://tinyurl.com/hp7z4sq
F23: http://tinyurl.com/zplzua3
F24: update is in -testing repo [2]
rawhide: already available in stable
These are the same versions of python-kombu that you'll receive
when
2.8.3 comes out so upgrading this package will keep your
packages on the
normal upgrade path. These newer python-kombu versions are
compatible
with 2.8.0, 2.8.1, and 2.8.2 versions of Pulp.
[0]: https://pulp.plan.io/issues/1801
[1]: http://koji.katello.org/koji/packageinfo?packageID=665
[2]:
https://bodhi.fedoraproject.org/updates/FEDORA-2016-ec038bbf19
-Brian
_______________________________________________
Pulp-list mailing list
Pulp-list@redhat.com <mailto:Pulp-list@redhat.com>
https://www.redhat.com/mailman/listinfo/pulp-list
_______________________________________________
Pulp-list mailing list
Pulp-list@redhat.com
https://www.redhat.com/mailman/listinfo/pulp-list
_______________________________________________
Pulp-list mailing list
Pulp-list@redhat.com
https://www.redhat.com/mailman/listinfo/pulp-list
_______________________________________________
Pulp-list mailing list
Pulp-list@redhat.com
https://www.redhat.com/mailman/listinfo/pulp-list