Hi Bogdan,
Yup its under load.
There aren't any errors apart from the WARNING:core:utimer_ticker:
utimer task <tm-utimer> already scheduled for
The live script doesn't have the failover on the 404/50X responses and
copes with the load no problem and never crashes.
It's only when I try and do the failover with the use_next_gw and the
load ramps up to about 1/4 of the normal load. So in testing a made a
few calls and its works fine, but when I put it live its starts crashing
at about 09:15 when the users start coming on line, but our load is
highest after 11:00am so there is load, but not a large amount. It
starts crashing at around 50 concurrent calls and maybe 5 or 6 cps.
I can reproduce it on our test server, but it will disrupt traffic, so
i'd rather do that out of hours, but if I sipp to an invalid number I
can reproduce it but all the cores look similar to me too and have the
reply message in them but i'm no expert at decoding the back traces.
I can get some more Cores for you but I suspect that they will all be
similar. I would have thought though, that would make the debuging easier?
Let me know id you need anything else from me.
Regards,
Richard
On 07/03/2017 21:10, Bogdan-Andrei Iancu wrote:
Hi Richard,
Sorry for the slow reply - is this crash happing only under some
++load ? do you see any errors from OpenSIPS prior to the crash ?
I noticed that the backtraces in the corefiles are similar - how easy
is for you to reproduce it ?
Regards,
Bogdan-Andrei Iancu
OpenSIPS Founder and Developer
http://www.opensips-solutions.com
OpenSIPS Summit May 2017 Amsterdam
http://www.opensips.org/events/Summit-2017Amsterdam.html
On 03/07/2017 12:28 PM, Richard Robson wrote:
Hi,
I've gone over the script and as far as I can see its working as
expected until the traffic remps up and then opensips crashes.
cores:
http://pastebin.com/CgN0h40K
http://pastebin.com/ay5TS8zD
http://pastebin.com/PGn3AqmU
Regards,
Richard
On 06/03/2017 12:14, Richard Robson wrote:
Hi<
I've tested this on the latest 2.2.3 with the same results.
http://pastebin.com/Uixb3v8G
there were a few of these in the logsd too just before the crash:
Mar 5 22:02:27 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079270 ms), it may overlap..
Mar 5 22:02:27 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079360 ms), it may overlap..
Mar 5 22:02:27 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079460 ms), it may overlap..
Mar 5 22:02:27 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079560 ms), it may overlap..
Mar 5 22:02:27 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079660 ms), it may overlap..
Mar 5 22:02:28 gl-sip-03 /usr/sbin/opensips[29875]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already
scheduled for 204079170 ms (now 204079760 ms), it may overlap..
Regards,
Richard
On 03/03/2017 13:15, Richard Robson wrote:
More cores
http://pastebin.com/MXW2VBhi
http://pastebin.com/T7JFAP2U
http://pastebin.com/u44aaVpWquit
http://pastebin.com/SFKKcGxE
http://pastebin.com/dwSgMsJi
http://pastebin.com/9HdGLm96
I've put 2.2.3 on the dev box now and will try to replicate on that
box, but its difficult to replicate the traffic artificially. I'll
try to replicate the fault on the dev box over the weekend. I cant
do it on the live gateways because it will affect customer traffic.
Regards,
Richard
On 03/03/2017 11:28, Richard Robson wrote:
I've revisited the gateway failover mechanism I had developed in
order to re route calls to the next gateway on 500's due to
capacity on the gateways we are using.
we have 3 gateways from one carrier and one from another. The 3
have 4 cps and will return a 503 or 500 if we breach this. The
single gateway from the other carrier has plenty of capacity and
should not be a problem so we want to catch this . and route to
the next gateway.
We are counting the CPS and channel limits and are routing to the
next gateway if we exceed the limit set, but There are still
occasions where a 5XX is generated, which results in a rejected call.
We want to stop these rejected calls and therefore want to
implement the failover mechanism for the 5XX responses. For 6
months we have been failing over if we think the counts are to
high on any one gateway without a problem. But when I implement a
failover on a 5XX response opensips starts crashing.
It's difficult to generate artificial traffic to mimic the real
traffic, but I've not had a problem with the script in testing.
Last night I rolled out the new script but by 09:15 this morning
opensips started crashing 10 times in 5 minutes. This was as the
traffic ramped up. I rolled back the script and it restarted OK
and has not crashed since. Therefore the Failover Mechanism in the
script is where the crash is happening
Core dump: http://pastebin.com/CqnESCm4
I'll add more dumps later
Regards,
Richard
this is the failure route catching the 5XX
failure_route[dr_fo] {
xlog (" [dr] Recieved reply to method $rm From: $fd, $fn,
$ft, $fu, $fU, $si, $sp, To: $ru");
if (t_was_cancelled()) {
xlog("[dr]call cancelled by internal caller");
rtpengine_manage();
do_accounting("db", "cdr|missed");
exit;
}
if ( t_check_status("[54]03")) {
route(relay_failover);
}
if ( t_check_status("500")) {
route(relay_failover);
}
do_accounting("db", "cdr|missed");
rtpengine_manage();
exit;
}
This is the route taken on the failure
route[relay_failover]{
if (use_next_gw()) {
xlog("[relay_failover-route] Selected Gateway is
$rd");
$avp(trunkratelimit)=$(avp(attrs){s.select,0,:});
$avp(trunkchannellimit)=$(avp(attrs){s.select,1,:});
####### check channel limit ######
get_profile_size("outbound","$rd","$var(size)");
xlog("[relay_failover-route] Selected Gateway is
$rd var(size) = $var(size)");
xlog("[relay_failover-route] Selected Gateway is
$rd avp(trunkcalllimit) = $avp(trunkchannellimit)");
xlog("[relay_failover-route] Selected Gateway is
$rd result = ( $var(size) > $avp(trunkchannellimit))");
if ( $(var(size){s.int}) >
$(avp(trunkchannellimit){s.int})) {
xlog("[relay_failover-route] Trunk $rd
exceeded $avp(trunkchannellimit) concurrent calls $var(size)");
route(relay_failover);
}
} else {
send_reply("503", "Gateways Exhusted");
exit;
}
##### We need to check Rate Limiting #######
if (!rl_check("$rd", "$(avp(trunkratelimit){s.int})",
"TAILDROP")) { # Check Rate limit $avp needs changing
rl_dec_count("$rd"); # decrement the counter since
we've not "used" one
xlog("[ratelimiter-route] [Max CPS:
$(avp(trunkratelimit){s.int}) Current CPS: $rl_count($rd)] Call
to: $rU from: $fU CPS exceeded, delaying");
$avp(initial_time)=($Ts*1000)+($Tsm/1000);
async(usleep("200000"),relay_failover_delay);
xlog ("Should not get here!!!! after async requst");
} else {
xlog ("[relay_outbound-route] [Max CPS:
$avp(trunkratelimit) Current CPS: $rl_count($rd)] Call to: $rU
from: $fU not ratelimited");
}
t_on_failure("dr_fo");
do_accounting("db", "cdr|missed");
rtpengine_manage();
if (!t_relay()) {
xlog("[relay-route] ERROR: Unable to relay");
send_reply("500","Internal Error");
exit;
}
}
--
Richard Robson
Greenlight Support
01382 843843
supp...@greenlightcrm.com
_______________________________________________
Users mailing list
Users@lists.opensips.org
http://lists.opensips.org/cgi-bin/mailman/listinfo/users