Hi

We have experienced this problem again. A couple of our binds to one
particular smsc (the rest were okay) had connectivity issues last night at
12 AM. The binds were re-established and reported as being online from the
status pages. However a queue for one of the binds built up on the
bearerbox. Only when I had run a stop-smsc and start-smsc for that bind did
the queue for that bind start processing again.

In the logs at 12AM we have a bunch of Errors:

2010-09-16 23:59:35 [32641] [44] ERROR: Error reading from fd 57:
2010-09-16 23:59:35 [32641] [44] ERROR: System error 104: Connection reset
by peer
2010-09-16 23:59:35 [32641] [44] ERROR: SMPP[XXX]: Couldn't connect to SMS
center (retrying in 10 seconds).
2010-09-16 23:59:38 [32641] [38] ERROR: Error reading from fd 52:
2010-09-16 23:59:38 [32641] [38] ERROR: System error 104: Connection reset
by peer
2010-09-16 23:59:38 [32641] [38] ERROR: SMPP[YYY]: Couldn't connect to SMS
center (retrying in 10 seconds).
2010-09-16 23:59:39 [32641] [46] ERROR: Error reading from fd 50:
2010-09-16 23:59:39 [32641] [46] ERROR: System error 104: Connection reset
by peer
2010-09-16 23:59:39 [32641] [46] ERROR: SMPP[AAA]: Couldn't connect to SMS
center (retrying in 10 seconds).
2010-09-16 23:59:47 [32641] [39] ERROR: Error reading from fd 49:
2010-09-16 23:59:47 [32641] [39] ERROR: System error 104: Connection reset
by peer
2010-09-16 23:59:47 [32641] [39] ERROR: SMPP[YYY]: Couldn't connect to SMS
center (retrying in 10 seconds).
2010-09-16 23:59:51 [32641] [48] ERROR: Error reading from fd 61:
2010-09-16 23:59:51 [32641] [48] ERROR: System error 104: Connection reset
by peer
2010-09-16 23:59:51 [32641] [48] ERROR: SMPP[BBB]: Couldn't connect to SMS
center (retrying in 10 seconds).
2010-09-17 00:00:00 [32641] [47] ERROR: Error reading from fd 40:
2010-09-17 00:00:00 [32641] [47] ERROR: System error 104: Connection reset
by peer
2010-09-17 00:00:00 [32641] [47] ERROR: SMPP[XXX]: Couldn't connect to SMS
center (retrying in 10 seconds).


I am not sure how Kannel works internally, but it is almost as if when the
bind is re-established, the old one is disposed and a new one is created but
the queue and the pointers are still sticking around for the old one and
have not been updated. This results in messages sitting in the queue and not
being routed to the bind which reports as being online.

I see that there might have been similar issues in the past:
http://www.kannel.org/pipermail/users/2009-May/007166.html. It might be
related maybe not.

<http://www.kannel.org/pipermail/users/2009-May/007166.html>We have already
set our binds up in a transmitter and receiver. We are not running
transceiver.

Regards,

On Thu, Sep 9, 2010 at 3:42 PM, brett skinner <tatty.dishcl...@gmail.com>wrote:

> Thanks Alvaro for your response.
>
> I am running a build from SVN from about a 2 weeks ago. I am bit weary of
> turning the loggers to debug mode because we are doing a lot of traffic and
> debug mode is very verbose we will eat through our disk in no time. It would
> be different if it was reproducible or if we could anticipate the problem
> because we could just turn on the loggers at the right time. This happens so
> sporadically we would have to leave the loggers in debug mode. The last time
> this happened was last week.
>
> I will go check out that tool you mentioned.
>
> I am not that interested in the extra TLVs. They were just making a bit of
> noise in our logs :)
>
> Thanks again for your help.
>
>
>
> On Thu, Sep 9, 2010 at 3:35 PM, Alvaro Cornejo 
> <cornejo.alv...@gmail.com>wrote:
>
>> Have you checked what does the system logs in debug mode?
>>
>> Regarding the queue, there is a tool created by Alejandro Guerreri
>> that allows you to view the queue content and delete messages... well
>> kannel does have several queues, so I don't know if it does wirk for
>> the one you mention. I don't remember the details but you can check
>> his blog. http://www.blogalex.com/archives/72
>>
>> About the TLV's you are receiving, you should ask yoru provider to see
>> what does they mean and what info are they sending. If its of your
>> interest, you can configure meta-data so you can capture that info;
>> otherwise you can safely ignore. As the PDU type is deliver_sm, I
>> suspect that it might be the dlr status... and that is why you have
>> that queue.
>>
>> Also if you upgrade to a recent version, the status page was improved
>> and it shows now separate counters for MT and dlrs. in older versions
>> MT/dlr counters were mixed
>>
>>
>> Hope helps
>>
>> Alvaro
>>
>> |-----------------------------------------------------------------------------------------------------------------|
>> Envíe y Reciba Datos y mensajes de Texto (SMS) hacia y desde cualquier
>> celular y Nextel
>> en el Perú, México y en mas de 180 paises. Use aplicaciones 2 vias via
>> SMS y GPRS online
>>               Visitenos en www.perusms.NET www.smsglobal.com.mx y
>> www.pravcom.com
>>
>>
>>
>> On Thu, Sep 9, 2010 at 2:42 AM, brett skinner <tatty.dishcl...@gmail.com>
>> wrote:
>> > Hi everyone
>> > Just wondering if anyone has had a chance to look at this yet?
>> > Thanks and appreciate any help.
>> >
>> > ---------- Forwarded message ----------
>> > From: brett skinner <tatty.dishcl...@gmail.com>
>> > Date: Tue, Sep 7, 2010 at 10:47 AM
>> > Subject: Weird binding issue that causes queues to build up.
>> > To: Users <users@kannel.org>
>> >
>> >
>> > Hi
>> > We are experiencing a rather weird occasional issue with Kannel. We have
>> two
>> > different boxes each with a Kannel installation. Every now and then one
>> of
>> > the boxes stops processing SMS queues and the queues just build up. This
>> > happens to both boxes (just not at the same time) When we have a look at
>> the
>> > status page we can see the queue and there are sms queued to the
>> bearerbox.
>> > I assume that it is the bearerbox queue. It looks as followed (from the
>> > status page)
>> > SMS: received 123 (0 queued), sent 123 (456 queued), store size -1
>> > It is the 456 queued part that we are concerned about. All the binds
>> report
>> > as being online with 0 in the queues but that 456 queue does not
>> disappear.
>> > If I sit trying to restart bind after bind one of them usually does the
>> > trick and queue disappears. The problem is we usually have no idea which
>> > bind it is and they are all reporting as being online. I have noticed
>> > looking through our logs from upstream applications that it appears that
>> > there was a network outage at round about the same time. I have not yet
>> > confirmed this with the hosting company. Also this is what appears in
>> the
>> > syslog.
>> > Sep  6 23:02:46 123-123-123-123 avahi-daemon[17943]: Received response
>> from
>> > host 64.150.181.120 with invalid source port 53756 on interface 'eth0.0'
>> > Sep  6 23:17:01 123-123-123-123 CRON[16934]: (root) CMD (   cd / &&
>> > run-parts --report /etc/cron.hourly)
>> > Sep  6 23:32:46 123-123-123-123 avahi-daemon[17943]: Received response
>> from
>> > host 64.150.181.120 with invalid source port 33895 on interface 'eth0.0'
>> > Sep  7 00:02:45 123-123-123-123 avahi-daemon[17943]: Received response
>> from
>> > host 64.150.181.120 with invalid source port 55945 on interface 'eth0.0'
>> > Sep  7 00:17:01 123-123-123-123 CRON[17231]: (root) CMD (   cd / &&
>> > run-parts --report /etc/cron.hourly)
>> > Sep  7 00:32:45 123-123-123-123 avahi-daemon[17943]: Received response
>> from
>> > host 64.150.181.120 with invalid source port 45291 on interface 'eth0.0'
>> > Sep  7 01:02:45 123-123-123-123 avahi-daemon[17943]: Received response
>> from
>> > host 64.150.181.120 with invalid source port 39067 on interface 'eth0.0'
>> > Sep  7 01:17:01 123-123-123-123 CRON[17479]: (root) CMD (   cd / &&
>> > run-parts --report /etc/cron.hourly)
>> > That IP address is not in our kannel.conf file. I am not sure what these
>> > errors are about. I might need to investigate this further. I am not
>> > security expert so I have no idea if this is malicious or not.
>> > This is what appears in the bearerbox logs at about the same time as the
>> > outage:
>> > 2010-09-06 23:02:46 [32641] [12] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032503580) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:03:07 [32641] [12] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032605180) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:08:04 [32641] [10] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032113180) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:14:32 [32641] [9] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032711480) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:26:12 [32641] [6] ERROR: SMPP[AAA]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:26:12 [32641] [6] ERROR: SMPP[AAA]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:26:12 [32641] [18] WARNING: SMPP[BBB]: Not ACKED message
>> > found, will retransmit. SENT<94>sec. ago, SEQ<423861>,
>> DST<+xxxxxxxxxxxxx>
>> > 2010-09-06 23:26:12 [32641] [18] WARNING: SMPP[BBB]: Not ACKED message
>> > found, will retransmit. SENT<94>sec. ago, SEQ<423862>,
>> DST<+xxxxxxxxxxxxx>
>> > 2010-09-06 23:26:12 [32641] [18] ERROR: SMPP[BBB]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:26:12 [32641] [18] ERROR: SMPP[BBB]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:26:12 [32641] [24] WARNING: Time-out waiting for
>> concatenated
>> > message ''+xxxxxxxxxxxxx>' '+yyyyyyyyyyy' 'EEE' '10' '2' '''. Send
>> message
>> > parts as is.
>> > 2010-09-06 23:26:12 [32641] [24] WARNING: Time-out waiting for
>> concatenated
>> > message ''+xxxxxxxxxxxxx>' '+yyyyyyyyyyy' 'EEE' '85' '2' '''. Send
>> message
>> > parts as is.
>> > 2010-09-06 23:26:12 [32641] [24] WARNING: Time-out waiting for
>> concatenated
>> > message ''+xxxxxxxxxxxxx>' '+yyyyyyyyyyy' 'EEE' '152' '2' '''. Send
>> message
>> > parts as is.
>> > 2010-09-06 23:26:42 [32641] [17] ERROR: SMPP[CCC]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:26:42 [32641] [17] ERROR: SMPP[CCC]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:27:08 [32641] [13] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032035180) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:27:12 [32641] [19] ERROR: SMPP[BBB]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:27:12 [32641] [19] ERROR: SMPP[BBB]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:27:14 [32641] [12] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032031280) for PDU type (deliver_sm) received!
>> > 2010-09-06 23:27:25 [32641] [16] ERROR: SMPP[CCC]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:27:25 [32641] [16] ERROR: SMPP[CCC]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:28:11 [32641] [6] WARNING: SMPP: Unknown
>> > TLV(0x140e,0x000c,323738373735303030303200) for PDU type (deliver_sm)
>> > received!
>> > 2010-09-06 23:55:18 [32641] [8] ERROR: SMPP[DDD]: I/O error or other
>> error.
>> > Re-connecting.
>> > 2010-09-06 23:55:18 [32641] [8] ERROR: SMPP[DDD]: Couldn't connect to
>> SMS
>> > center (retrying in 10 seconds).
>> > 2010-09-06 23:55:21 [32641] [11] WARNING: SMPP: Unknown
>> > TLV(0x1406,0x0007,01906032858280) for PDU type (deliver_sm) received!
>> > I am looking for any sort of guidance of where to start to resolve this
>> > issue. Also any comments will be most welcome. In general I would like
>> to
>> > know:
>> >
>> > Is there anyway that I can see what is in that queue of 456, I would
>> like to
>> > know which bind is down. I thought store-status might be it does not
>> appear
>> > to be.
>> > What could be causing this issue? (If you suspect that is something to
>> do
>> > with configuration I will post the configuration file)
>> > I notice some unknown TLV warnings. Is this something we should be
>> concerned
>> > about?
>> > It seems that there was some sort of network problem and all the
>> connections
>> > (to different smscs) disconnected and reconnected. Why does the queue
>> not
>> > disappear after they reconnect?
>> >
>> > I greatly appreciate your time and effort. Thanks
>> > Regards,
>> >
>>
>
>
>

Reply via email to