On Mon, 05 Sep 2011 20:23:02 +0200, David Touzeau wrote:
Dear

I would like to create a kind of law calculation in order to quickly
calculate server performances to store squid...

I know there is a lot of parameters that should make e this calculation
more complex, but this is just to be generic.


"generic" is not possible. It boils down to what _your_ users are doing is different to _my_ users.


For example :

I have 150 users :
-----------------------------------

Memory : 750K per users = 150x750=112 Mb memory + 300Mb for the system =
512 Mb minimal.
Hard disk cache : 50Mb/user = 150*50 = 7.5Go minimal stored disk size
for cache..

Is it make sense ?

Your aim makes sense in a way. The metric of "per user" does not relate to resource usage though.

This is solely because one user could be making no requests at all or several thousand per second. And they can switch between these behaviours and random values in between without notice. I have seen a network happily serving 15K users with one Squid on a 50MBit uplink, and also a GigE network brought to its knees by just a few users. "Normal" web traffic in both cases.


The Squid relevant metrics are closer tied to requests per second, or the internal Mbps line speed of HTTP requirements.


Minimal stored disk size is always zero. Maximal is best at <80% of the size of the disk you plug in. Moderated by at most 2^24 objects per cache_dir. This is a fixed limit, so performance there is relative to your average cacheable object size. Disks are best sized along the lines of: 2^24 * avg object size. Overall disk count multiply that out by req-per-sec to the time you want things cached by or run out of $$ to buy disks.


Memory consumption is dominated by the disk index (10-15 MB index / 1 GB of storage). And by cache_mem storage, which will suck up every spare byte you can give it until itself starts to suffer those 2^24 object limits. per-User/connection requirements are measured in KB, so these days not really worth worrying over.


Amos

Reply via email to