I'm afraid I can't use iostat on a VPS. Shows me a "Cannot find disk data"
warning and dies. Googled it a bit and found a bunch of "iostat doesn't
work from inside containers" and "it would be a security issue if you could
have that kkind of access to physical drivers from inside a virtual
machine".

It's a shame dedicated servers are so overpriced and aren't really an
option. The cheapest ones cost at least 25x (yes, I'm not exaggerating) the
VPS I rent. Brazil's network is a piece of shit.

Still, gotta get the best I can from these VPSes. If I could solidly link
the stuttering to i/o, I can argue with the host and see if they can place
me with calmer neighbours.

Would a SSD, still using a virtual machine, help me somehow?


_pilger


On 9 April 2014 23:23, Yun Huang Yong <[email protected]> wrote:

> Try watching:
>
> $ iostat -x 1
>
> You want to pay attention to the r_await/w_await which are the average
> read and write times respectively, in milliseconds.
>
> If any of those numbers goes above 15 you will get the stuttering you
> describe.
>
> Basically it indicates that the underlying disk(s) can not serve your
> read/write requests within 15 milliseconds. As mentioned previously
> 1000ms/66 ticket = time budget of 15ms per tick.
>
> If that is the case talk to the host about possibly migrating to another
> physical host, or change providers. There's nothing you can do inside the
> VM.
>
> With VPS hosting there are a lot of anecdotes about various providers. At
> the end of the day your *consistent* performance level depends on what your
> provider does in terms of locking down resource usage between VMs, how many
> VMs they try to cram onto each box, and what your neighbours are doing.
>
> In an ideal world providers would provide specifications down to things
> like disk latency, IOPS, CPU scheduling etc but in reality most customers
> won't understand it anyway, or prefer to pay a low price for "burst"
> ability even if that means massive oversubscription
>
> Unfortunately that means VPS is basically a crapshoot, especially within
> the budget market. You try VPSes until you land with a provider & host that
> works for your budget. On the bright side at least with the server
> registration now you can migrate between VPS providers without losing all
> of your players :]
>
>
> On 10/04/2014 3:48 AM, pilger wrote:
>
>> Hey guys,
>>
>> Switched to Ubuntu 13 x86 and things got a little better but the hiccups
>> are still there.
>>
>> I've noticed the "wa" value of top varies a bit during the hiccups. Here's
>> what I have observed:
>>
>> [image: Inline images 1]
>>
>> Could that be related to the problem? I've googled it briefly and found it
>> has something to do with disk i/o. Is that correct!? The problem is caused
>> by slow disk!?
>>
>> Thanks in advance!
>>
>>
>> _pilger
>>
>>
>> On 8 April 2014 11:14, pilger <[email protected]> wrote:
>>
>>  The VPS is running OpenVZ virtualization, so I guess it should be fine.
>>>
>>> I also have a friend on the same hosting company with the same plan and
>>> Debian 7 x64 (which was the SO at the time I first posted about the
>>> problem) that works way better than my VPS. Not sure why.
>>>
>>>
>>> _pilger
>>>
>>>
>>> On 8 April 2014 11:06, pilger <[email protected]> wrote:
>>>
>>>  Weasels, whew. Long text. But it was a good share of experience there.
>>>> My
>>>> VPS host offers Ubuntu 13 as the latest SO there, so I switched to it.
>>>> On
>>>> preliminary tests it did well, let's see how it goes when the server
>>>> gets
>>>> full and we reach peak time of day were the lousy neighbours are all
>>>> screaming around.
>>>>
>>>> I've had previous experiences with Linux but for other purposes and,
>>>> since I do some programming around, the learning curve isn't too steep
>>>> for
>>>> me. Networking is still a mystery to me, though. So I went here for your
>>>> help.
>>>>
>>>> Many thanks for all the info!
>>>>
>>>> Thanks for all the input, Yun. Seems interesting but a rather long read,
>>>> so I'll take my time to digest the whole information. I'll also have to
>>>> produce another server to host the monitor, which I don't have available
>>>> right now, so I guess collectd will be postponed for a while, despite
>>>> of it
>>>> being a good method to see what the hell is going on.
>>>>
>>>>
>>>>
>>>>
>>>> _pilger
>>>>
>>>>
>>>> On 7 April 2014 21:55, Yun Huang Yong <[email protected]> wrote:
>>>>
>>>>  collectd is fairly light which is why it is popular as a collection
>>>>> agent.
>>>>>
>>>>> I'm going to assume you are comfortable installing packages and
>>>>> configuring software. I don't have the time to write copy pasta
>>>>> instructions. I *strongly* recommend you read all of this & the links
>>>>> before you begin, and make sure you understand what is required of you.
>>>>>
>>>>> The key components are:
>>>>>    - https://collectd.org/
>>>>>    - http://graphite.wikidot.com/
>>>>>
>>>>> You need to get collectd running on each of your TF2 hosts. Basically
>>>>> apt-get but see note below regarding collectd versions.
>>>>>
>>>>> You'll then want to setup Graphite on another machine. You *could* run
>>>>> it on your TF2 host but Carbon can get I/O hungry (it is tunable) and
>>>>> that
>>>>> will create more problems for you so I strongly recommend running
>>>>> Graphite
>>>>> on another machine.
>>>>>
>>>>> Also, having Graphite on another machine (with the collectd collector,
>>>>> below) makes it easy for you to have multiple TF2 hosts, or migrate TF2
>>>>> hosts.
>>>>>
>>>>> In my setup I have my Graphite host running on an Ubuntu VM at home
>>>>> with
>>>>> 6 external servers reporting to it.
>>>>>
>>>>> Here's a picture of the overall setup:
>>>>> https://dl.dropboxusercontent.com/u/8110989/2014/collectd-graphite.png
>>>>>
>>>>> Back to setup... with collectd running on your TF2 host, and Graphite
>>>>> on
>>>>> another host, how do you connect them?
>>>>>
>>>>> https://collectd.org/wiki/index.php/Networking_introduction
>>>>>
>>>>> Your Graphite host *also* needs to run collectd in order to act as a
>>>>> collection server for your TF2 host's collectd to send it data. Edit
>>>>> config
>>>>> on both TF2 host & Graphite host -- both sides need to run the collectd
>>>>> network plugin, Graphite host as server, TF2 host as client.
>>>>>
>>>>> Your Graphite host's collectd also needs to run the write_graphite
>>>>> plugin to write the network collected data to Graphite.
>>>>>
>>>>> https://collectd.org/wiki/index.php/Plugin:Write_Graphite
>>>>>
>>>>> <Plugin write_graphite>
>>>>>    <Carbon>
>>>>>      Host "localhost"
>>>>>      Port "2003"
>>>>>      EscapeCharacter "_"
>>>>>    </Carbon>
>>>>> </Plugin>
>>>>>
>>>>> Note: if you Google collectd + graphite you may be confused by many
>>>>> blog
>>>>> posts refer to custom written plugins which were necessary before
>>>>> collectd
>>>>> had its write_graphite plugin.
>>>>>
>>>>> Note 2: since you're on Debian note that the write_graphite plugin was
>>>>> added with collectd 5.1. You may need to get it from backports or
>>>>> something.
>>>>>
>>>>> For Graphite...
>>>>>
>>>>> This is a reasonable overview but may be out of date:
>>>>> http://graphite.wikidot.com/installation
>>>>>
>>>>> Read ^ as an overview but maybe follow the current instructions here:
>>>>> https://graphite.readthedocs.org/en/latest/install.html
>>>>>
>>>>> You need to pay attention to the storage-schemas.conf but you can more
>>>>> or less ignore other instructions about feeding data into Graphite.
>>>>> With
>>>>> the collectd write_graphite plugin your data will automagically be fed
>>>>> from
>>>>> collectd -> localhost:2003 which is Carbon (Graphite's collector).
>>>>>
>>>>> Good luck :]
>>>>>
>>>>> PS: I am happy to answer specific questions about the collectd/graphite
>>>>> setup but if you ask general sysadmin stuff I probably won't respond.
>>>>>
>>>>>
>>>>> On 8/04/2014 12:50 AM, pilger wrote:
>>>>>
>>>>>  I've noticed the yellow bars mainly on the Mem field. Don't know if
>>>>>> that
>>>>>> might be related. Could it?
>>>>>>
>>>>>> About collectd, it seems very nice and a lot easier to visualize but
>>>>>> you
>>>>>> talked greek to me up there. Would you point me to some tutorial or
>>>>>> show
>>>>>> me some ropes on how to get it running so I can find the bottlenecks?
>>>>>> Does it use a lot of resource!?
>>>>>>
>>>>>>
>>>>>> _pilger
>>>>>>
>>>>>>
>>>>>> On 7 April 2014 11:35, Yun Huang Yong <[email protected]
>>>>>> <mailto:[email protected]>> wrote:
>>>>>>
>>>>>>      Your concern about noisy VPS neighbours will show up as CPU
>>>>>> steal -
>>>>>>      htop shows this as yellow bars by default.
>>>>>>
>>>>>>      Disk latency could also be an issue.
>>>>>>
>>>>>>      66 tick means each tick has a time budget of around 15ms
>>>>>> (1000/66).
>>>>>>      If disk latency exceeds 15ms you will get stuttering - I had this
>>>>>>      happen on servers in the past.
>>>>>>
>>>>>>      e.g.
>>>>>>      https://dl.dropboxusercontent.__com/u/8110989/2013/np1-disk-
>>>>>> __latency.png
>>>>>>
>>>>>>      <https://dl.dropboxusercontent.com/u/
>>>>>> 8110989/2013/np1-disk-latency.
>>>>>> png>
>>>>>>
>>>>>>      Stuttery server leading up to 08/03 (US style month/day, August
>>>>>> last
>>>>>>      year). Host migrated my server to another less loaded machine,
>>>>>> great
>>>>>>      for a few weeks then as that machine also became more heavily
>>>>>>      utilised (by other customers) it started to stutter again.
>>>>>>
>>>>>>      FWIW I use collectd to gather these metrics on each host, feeding
>>>>>>      into a single collectd collector which then uses collectd's
>>>>>>      write_graphite plugin to write all the data into graphite for
>>>>>>      storage & graphing. collectd's default 10s polling is great for
>>>>>>      picking up transient issues, and graphite_web makes the
>>>>>>      visualisation easy.
>>>>>>
>>>>>>
>>>>>>      On 7/04/2014 10:26 PM, pilger wrote:
>>>>>>
>>>>>>          Hey guys, thanks for the replies.
>>>>>>
>>>>>>             * The RAM seems all right when I look at it with htop;
>>>>>>             * We tried CentOS but the network was behaving poorly
>>>>>> with it
>>>>>>          so we
>>>>>>
>>>>>>               switched to Debian x64 and it became a lot better;
>>>>>>             * net_splitpacket_maxrate was set to 50000 while the rates
>>>>>>          were from
>>>>>>
>>>>>>               30000 to 60000. I've now set the splitpacket to 100000
>>>>>> and
>>>>>>          the rates
>>>>>>               to 50000 to 100000 as you guys suggested. Gotta wait a
>>>>>> bit
>>>>>>          for the
>>>>>>               server to get full so I can check if it worked;
>>>>>>
>>>>>>          Wouldn't the htop or any other monitoring tool show something
>>>>>>          wrong even
>>>>>>          it being a VPS!?
>>>>>>
>>>>>>          But, anyway, as I mentioned before, the problem occurs with
>>>>>> the
>>>>>>          server
>>>>>>          practically empty. So I don't think it is related to CPU
>>>>>> being
>>>>>>          overloaded... could I be wrong on this? Could my VPS
>>>>>> neighbours
>>>>>> be
>>>>>>          leeching on my CPU even it being supposedly reserved to my
>>>>>> service?
>>>>>>
>>>>>>          Thanks!
>>>>>>
>>>>>>
>>>>>>
>>>>>>          _pilger
>>>>>>
>>>>>>
>>>>>>          On 7 April 2014 02:10, John <lists.valve@nuclearfallout.__
>>>>>> net
>>>>>>          <mailto:[email protected]>
>>>>>>          <mailto:lists.valve@__nuclearfallout.net
>>>>>>
>>>>>>          <mailto:[email protected]>>> wrote:
>>>>>>
>>>>>>                   Its not the RAM. Its packet loss from server side -
>>>>>> you
>>>>>>          won't
>>>>>>                   see it on net graph as its only client side.
>>>>>>
>>>>>>
>>>>>>               Packet loss should show in net_graph output either way.
>>>>>>          But, to be
>>>>>>               safe, certainly run MTR tests.
>>>>>>
>>>>>>
>>>>>>                   I've had this happen to me lots of times. Been
>>>>>> running
>>>>>>          servers
>>>>>>                   since the 1.5 days. Ditch your host and also ditch
>>>>>>          Debian BS.
>>>>>>
>>>>>>
>>>>>>               Recent versions of Debian work well for game servers, so
>>>>>>          ditching it
>>>>>>               would not be necessary.
>>>>>>
>>>>>>               You should confer with your host on the status of your
>>>>>>          hardware and
>>>>>>               whether a performance limitation is involved, such as
>>>>>> I/O
>>>>>>          delays.
>>>>>>               You should also double-check server-side rates,
>>>>>> including
>>>>>>          by making
>>>>>>               sure that net_splitpacket_maxrate is set sufficiently
>>>>>> high
>>>>>>          (such as
>>>>>>               100000). These symptoms seem along the lines of what I
>>>>>>          would expect
>>>>>>               from net_splitpacket_maxrate being low.
>>>>>>
>>>>>>
>>>>>>                   Ask ant corporation or enterprise, all use CentOS.
>>>>>>
>>>>>>
>>>>>>               CentOS is marketed to enterprise and works well for such
>>>>>>               applications because of its older, stable, well-tested
>>>>>> software
>>>>>>               packages and extended RHEL support for those older
>>>>>>          packages. For
>>>>>>               game servers, it is not ideal, since those older
>>>>>> packages
>>>>>>          often lack
>>>>>>               useful features and performance tweaks. Debian is
>>>>>> usually a
>>>>>>          better
>>>>>>               choice for game servers.
>>>>>>
>>>>>>
>>>>>>                   If you're interested in hosting DDoS protected
>>>>>> servers,
>>>>>>          email me
>>>>>>                   - I can help you.
>>>>>>
>>>>>>
>>>>>>               Be very careful with hosts that claim to offer DDoS
>>>>>> protection.
>>>>>>               There is an extremely limited number who do it right,
>>>>>> and a
>>>>>>          very
>>>>>>               large number who do not.
>>>>>>
>>>>>>               -John
>>>>>>
>>>>>>
>>>>>>               ___________________________________________________
>>>>>>
>>>>>>
>>>>>>               To unsubscribe, edit your list preferences, or view the
>>>>>> list
>>>>>>               archives, please visit:
>>>>>>          https://list.valvesoftware.____com/cgi-bin/mailman/listinfo/
>>>>>> ____hlds
>>>>>>
>>>>>>          <https://list.valvesoftware.__com/cgi-bin/mailman/listinfo/_
>>>>>> _hlds
>>>>>> <https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>          _________________________________________________
>>>>>>          To unsubscribe, edit your list preferences, or view the list
>>>>>>          archives, please visit:
>>>>>>          https://list.valvesoftware.__com/cgi-bin/mailman/listinfo/_
>>>>>> _hlds
>>>>>>          <https://list.valvesoftware.com/cgi-bin/mailman/listinfo/
>>>>>> hlds>
>>>>>>
>>>>>>
>>>>>>
>>>>>>      _________________________________________________
>>>>>>      To unsubscribe, edit your list preferences, or view the list
>>>>>>      archives, please visit:
>>>>>>      https://list.valvesoftware.__com/cgi-bin/mailman/listinfo/__hlds
>>>>>>      <https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> To unsubscribe, edit your list preferences, or view the list archives,
>>>>>> please visit:
>>>>>> https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds
>>>>>>
>>>>>>
>>>>>>
>>>>> _______________________________________________
>>>>> To unsubscribe, edit your list preferences, or view the list archives,
>>>>> please visit:
>>>>> https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds
>>>>>
>>>>>
>>>>
>>>>
>>>  _______________________________________________
>> To unsubscribe, edit your list preferences, or view the list archives,
>> please visit:
>> https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds_linux
>>
>>
>
> _______________________________________________
> To unsubscribe, edit your list preferences, or view the list archives,
> please visit:
> https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds_linux
>
_______________________________________________
To unsubscribe, edit your list preferences, or view the list archives, please 
visit:
https://list.valvesoftware.com/cgi-bin/mailman/listinfo/hlds_linux

Reply via email to