Just wanted to report that we are back up and running for the most part as well, BUT quite a number of our servers are still experiencing some BlackOut periods where twitter fails to respond and connections time out. They seem to last about 5-10 minutes each. We are running quite a few servers during this time to mitigate the issue propagating to our users, but it is a costly proposition to continue running this many servers for extended periods of time.

Here is a traceroute to twitter.com on one of the servers during the "Blackout" when it is not getting any response from twitter

traceroute to twitter.com (168.143.162.100), 30 hops max, 40 byte packets 1 67-207-128-2.slicehost.net (67.207.128.2) 0.000 ms 0.000 ms 0.000 ms 2 209-20-79-2.slicehost.net (209.20.79.2) 0.000 ms 0.000 ms 0.000 ms 3 ge-6-10-193.car1.StLouis1.Level3.net (4.53.160.189) 0.000 ms 0.000 ms 0.000 ms 4 ae-11-11.car2.StLouis1.Level3.net (4.69.132.186) 0.000 ms 0.000 ms 0.000 ms 5 ae-4-4.ebr2.Chicago1.Level3.net (4.69.132.190) 7.999 ms 7.999 ms 7.999 ms 6 ae-2-54.edge3.Chicago3.Level3.net (4.68.101.116) 7.999 ms ae-2-52.edge3.Chicago3.Level3.net (4.68.101.52) 7.999 ms ae-2-54.edge3.Chicago3.Level3.net (4.68.101.116) 7.999 ms
 7  4.68.63.198 (4.68.63.198)  7.999 ms  7.999 ms  8.000 ms
8 ae-1.r21.chcgil09.us.bb.gin.ntt.net (129.250.3.8) 8.000 ms 8.000 ms 8.000 ms 9 as-5.r20.snjsca04.us.bb.gin.ntt.net (129.250.3.77) 51.996 ms 51.996 ms 51.996 ms 10 xe-1-3.r02.mlpsca01.us.bb.gin.ntt.net (129.250.5.61) 55.995 ms 55.995 ms 55.995 ms 11 mg-1.c00.mlpsca01.us.da.verio.net (129.250.24.202) 55.995 ms 55.995 ms 59.995 ms
12  128.121.150.245 (128.121.150.245)  55.995 ms  51.995 ms  51.995 ms
13  128.121.150.245 (128.121.150.245)  51.996 ms !X * *



On Aug 10, 2009, at 2:57 PM, Ryan Sarver wrote:

Wanted to send out a status update and let everyone know where the situation stands as of today at noon.

- Most developers are reporting being back in operation as of noon on Sunday - We have changed our defenses to make sure API developers are better supported. As such the system has more general strain on it and thus will produce some more 502/503 errors. If you see them, you should do a geometric back off instead of just sending a new request.
- OAuth should be fully operational
- If you continue to have unexpected errors, please produce a packet trace so we can help debug and define the issue.

I will continue to give periodic updates throughout the day as we know more, but as most apps are back in action the updates will be more based on new news. Please continue to let us know of any unexpected issues you may have.

Thanks again for your continued patience and support.

Best, Ryan


Reply via email to