> 
> hi,
Hi Gordon,

> 
> this seems to be similar to a bug we were tracking in earlier[1].
> basically, any service with a listener never seemed to idle properly.
> 
> based on earlier investigation, we found it relates to the heartbeat
> functionality in oslo.messaging. i'm not entirely sure if it's because of it 
> or
> some combination of things including it. the short answer, is to disable
> heartbeat by setting heartbeat_timeout_threshold = 0 and see if it fixes your
> cpu usage. you can track the comments in bug.

As I see in the bug report, you mention that the problem is only with the 
notification agent, 
and the collector is fine. I'm in an entirely opposite else situtation.

starce-ing the two processes:

Notification agent:
----------------------
epoll_wait(4, {}, 1023, 43)             = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_ctl(4, EPOLL_CTL_DEL, 8, 
{EPOLLWRNORM|EPOLLMSG|EPOLLERR|EPOLLHUP|EPOLLRDHUP|EPOLLONESHOT|EPOLLET|0x1ec88000,
 {u32=32738, u64=24336577484324834}}) = 0
recvfrom(8, 0x7fe2da3a4084, 7, 0, 0, 0) = -1 EAGAIN (Resource temporarily 
unavailable)
epoll_ctl(4, EPOLL_CTL_ADD, 8, {EPOLLIN|EPOLLPRI|EPOLLERR|EPOLLHUP, {u32=8, 
u64=40046962262671368}}) = 0
epoll_wait(4, {}, 1023, 1)              = 0
epoll_ctl(4, EPOLL_CTL_DEL, 24, 
{EPOLLWRNORM|EPOLLMSG|EPOLLERR|EPOLLHUP|EPOLLRDHUP|EPOLLONESHOT|EPOLLET|0x1ec88000,
 {u32=32738, u64=24336577484324834}}) = 0
recvfrom(24, 0x7fe2da3a4084, 7, 0, 0, 0) = -1 EAGAIN (Resource temporarily 
unavailable)
epoll_ctl(4, EPOLL_CTL_ADD, 24, {EPOLLIN|EPOLLPRI|EPOLLERR|EPOLLHUP, {u32=24, 
u64=40046962262671384}}) = 0
epoll_wait(4, {}, 1023, 0)              = 0

ceilometer-collector:
-------------------------
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0
epoll_wait(4, {}, 1023, 0)              = 0

So the notification agent do something at least between the crazy epoll()s.

It is the same with or without the heartbeat_timeout_threshold = 0 in 
[oslo_messaging_rabbit].
Then something must be still wrong with the listeners, the bug[1] should not be 
closed, I think.

Br,
György

> 
> [1] https://bugs.launchpad.net/oslo.messaging/+bug/1478135
> 
> On 17/02/2016 4:14 AM, Gyorgy Szombathelyi wrote:
> > Hi!
> >
> > Excuse me, if the following question/problem is a basic one, already
> > known problem, or even a bad setup on my side.
> >
> > I just noticed that the most CPU consuming process in an idle
> > OpenStack cluster is ceilometer-collector. When there are only
> > 10-15 samples/minute, it just constantly eats about 15-20% CPU.
> >
> > I started to debug, and noticed that it epoll()s constantly with a
> > zero timeout, so it seems it just polls for events in a tight loop.
> > I found out that the _maybe_ the python side of the problem is
> > oslo_messaging.get_notification_listener() with the eventlet executor.
> > A quick search showed that this function is only used in aodh_listener
> > and ceilometer_collector, and both are using relatively high CPU even
> > if they're just 'listening'.
> >
> > My skills for further debugging is limited, but I'm just curious why
> > this listener uses so much CPU, while other executors, which are using
> > eventlet, are not that bad. Excuse me, if it was a basic question,
> > already known problem, or even a bad setup on my side.
> >
> > Br,
> > György
> >
> >
> __________________________________________________________
> ____________
> > ____ OpenStack Development Mailing List (not for usage questions)
> > Unsubscribe:
> > openstack-dev-requ...@lists.openstack.org?subject:unsubscribe
> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> >
> 
> --
> gord
> 
> __________________________________________________________
> ________________
> OpenStack Development Mailing List (not for usage questions)
> Unsubscribe: OpenStack-dev-
> requ...@lists.openstack.org?subject:unsubscribe
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to