I've done some further digging, and managed to simplify the system a little to 
reproduce the problem. The system is now an external process that posts 
messages to the default headers exchange on my machine, which has a ring queue 
to receive effectively all messages from the default headers exchange, process 
them, and post to another headers exchange. There is now nothing listening on 
the subsequent headers exchange, and all exchanges are non-durable. I've also 
tried Fraser's suggestion of marking the link as unreliable on the queue which 
seems to have no effect (is there any way in the qpid utilities to confirm the 
link has been set to unreliable?)

So essentially what happens is the system happily processes away, normally with 
an empty ring queue, sometimes it spikes up a bit and goes back down again, 
with my ingest process using ~70% CPU and qpidd ~50% CPU, on a machine with 8 
CPU cores. However sometimes the queue spikes up to 2GB (the max), starts 
throwing messages away, and qpid hits 100%+ CPU and the ingest process goes to 
about 3% CPU. I can see messages are being very slowly processed.

I've tried attaching to qpidd with gdb a few times, and all threads apart from 
one seem to be idle in epoll_wait or pthread_cond_wait. The running thread 
always seems to be somewhere under DispatchHandle::processEvent.

I'm at a bit of a loss for what I can do to fix this!

Jimmy
  
----- Original Message -----
From: Fraser Adams
Sent: 08/23/13 09:09 AM
To: [email protected]
Subject: Re: System stalling
 Hi Jimmy, hope you are well!
As an experiment one thing that you could try is messing with the link 
"reliability". As you know in the normal mode of operation it's 
necessary to periodically send acknowledgements from the consumer client 
application which then get passed back ultimately to the broker.

I'm no expert on this but from my recollection if you are in a position 
particularly where circular queues are overflowing and you are 
continually trying to produce and consume and you have some fair level 
of prefetch/capacity on the consumer the mechanism for handling the 
acknowledgements on the broker is "sub-optimal" - I think it's a linear 
search or some such and there are conditions where catching up with 
acknowledgements becomes a bit "N squared".

Gordon would be able to explain this way better than me - that's 
assuming this hypothesis is even relevant :-)

Anyway if you try having a link: {reliability: unreliable} stanza in 
your consumer address string (as an example one of mine looks like the 
following - the address sting syntax isn't exactly trivial :-)).

string address = "test_consumer; {create: receiver, node: {x-declare: 
{auto-delete: True, exclusive: True, arguments: {'qpid.policy_type': 
ring, 'qpid.max_size': 100000000}}, x-bindings: [{exchange: 'amq.match', 
queue: 'test_consumer', key: 'test1', arguments: {x-match: all, 
data-format: test}}]}, link: {reliability: unreliable}}";

Clearly your arguments would be different but hopefully it'll give you a 
kick start.


The main down side of disabling link reliability is that if you have 
enabled prefetch and the consumer unexpectedly dies then all of the 
messages on the prefetch queue will be lost, whereas with reliable 
messaging the broker maintains references to all unacknowledged messages 
so would resent them (I *think* that's how it works.....)


At the very least it's a fairly simple tweak to your consumer addresses 
that might rule out (or point to) acknowledgement shenanigans as being 
the root of your problem. From my own experience I always end up blaming 
this first if I hit performance weirdness with ring queues :-)

HTH,
Frase



On 21/08/13 17:08, Jimmy Jones wrote:
>>>>> I've got an simple processing system using the 0.22 C++ broker, all
>>>>> on one box, where an external system posts messages to the default
>>>>> headers exchange, and an ingest process receives them using a ring
>>>>> queue, transforms them and outputs to a different headers exchange.
>>>>> Various other processes pick messages of interest off that exchange
>>>>> using ring queues. Recently however the system has been stalling -
>>>>> I'm still receiving lots of data from the other system, but the
>>>>> ingest process suddenly goes to <5% CPU usage and its queue fills up
>>>>> and messages start getting discarded from the ring, the follow on
>>>>> processes go to practically 0% CPU and qpidd hovers around 95-120%
>>>>> CPU (normally its ~75%) and the rest of the system pretty much goes
>>>>> idle (no swapping, there is free memory)
>>>>>
>>>>> I attached to the ingest process with gdb and it was stuck in send
>>>>> (waitForCapacity/waitForCompletionImpl) - I notice this can block.
>>>> Is there any queue bound to the second headers exchange, i.e. to the one
>>>> this ingest process is sending to, that is not a ring queue? (If you run
>>>> qpid-config queue -r, you get a quick listing of the queues and their
>>>> bindings).
>>> I've run qpid-config queue, and all my queues have --limit-policy=ring, 
>>> apart
>>> from a UUID one which I presume is qpid-config itself. Are there any other 
>>> useful
>>> debugging things I can do?
>> What does qpid-stat -q show? Is it possible to test whether the broker
>> is still responsive e,g, by sending and receiving messages through a
>> test queue/exchange? Are there any errors in the logs? Are any of the
>> queues durable (and messages persistent)?
> qpid-stat -q is all zero's in the msg & bytes column, apart from the ingest 
> queue,
> and another overflowing ring queue I have.
>
> I did run qpid-tool when the system was broken to dump some stats. 
> msgTotalDequeues
> was slowly incremeneting on the ingest queue, so I presume messages were 
> still being
> delivered and the broker was responsive?
>
> The only logging I've got is syslog, and I just see a warning about unsent 
> data,
> presumably when the ingest process receives a SIGALARM. I'm happy to swich on 
> more
> logging, what would you recommend?
>
> None of my queues are durable, but I think incoming messages from the other 
> system
> are marked as durable. The exchange that the ingest process sends to is 
> durable,
> but I'm not setting any durable flags on outgoing messages (I presume the 
> default
> is off).
>
>> Another thing might be a ptrace of the broker process. Maybe two or
>> three with a short delay between them.
> I'll try this next time it goes haywire.
>
>> For some reason it seems like the broker is not sending back
>> confirmation to the sender in the ingest process, causing that to block.
>> Ring queues shouldn't be subject to producer flow control so we need to
>> figure out what other reason there could be for that.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [email protected]
> For additional commands, e-mail: [email protected]
>


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected] 

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to