How about VOD streams ? We're about to implement another VOD system and 
am not sure yet if it will be able to handle it, how is it possible to 
test say 3000 streams at once ? I have noticed the jdk garbage 
collection isnt optimized as much as it could be maybe ?  Even setting a 
max of 1.5GB of ram which leaves only 500MB available for other services 
its reaching 800MB when ive seen it usually only do about 200MB ? CPU is 
OK for now, the ram is a concern though, should we be putting 4GB in all 
of them ?


Interalab wrote:
> Rob Schoenaker and I ran a little stress test this morning and wanted to 
> share our results.  Rob, feel free to add to or correct me if you want.
>
> This was a test of one publishing live stream client and many 
> subscribing clients.
>
> Here's the server config:
>
> Xubuntu Linux
> AMD 64 3500+ processor
> 4 GB RAM
> Red 5 trunk ver 1961
> Gbit Internet connection
>
> Client side:
>
>  From the other side of the world . . .
> Lots of available bandwidth
>
> The first run choked the server at 256 simultaneous connections.  They 
> were 250k - 450k live streams.
>
> After a re-boot, we got up into the 300 + connections.  This time the 
> resolution was lower, so the average bandwidth per stream was about 150k
>
> Server looked like this: 
> Cpu(s): 12.0%us,  2.0%sy,  0.0%ni, 84.0%id,  0.0%wa,  0.3%hi,  1.7%si,  
> 0.0%st
>  Mem:   3976784k total,  1085004k used,  2891780k free,     7896k buffers
>  Swap:  2819368k total,        0k used,  2819368k free,   193740k cached
>
> After about 15 minutes, and over 400 connections, Red5 quit without any 
> log errors.  The Java PID just went away.  Had a bunch of these in 
> dmesg:  e1000: eth1: e1000_clean_tx_irq: Detected Tx Unit Hang
>
> Started Red5 by running red5.sh without re-booting the server.  It came 
> right back up and started streaming again.
>
> This time, we set the resolution to 80x60, or about 60-80 kbps per stream.
>
> Rob tried to crash it by launching about 200 connections in about 10 
> seconds, but it kept running.  It didn't die again.
>
> Final outcome of the last test:
>
> 627 concurrent connections peak
> approx 1100 connections total (some dropped when browsers crashed under 
> the load, etc.)
>
> At the peak, player buffers started to get big.  Some as high as 70, 
> most of mine were in the 30's. 
>
> So, my observation is that even though the server and available 
> bandwidth didn't seem to be stressed too much - lots of memory and cpu % 
> in the teens, the larger the individual streams, the fewer total 
> connections we could make.
>
> Not very scientific, but we thought it was worth sharing with the list.
>
> Regards,
> Bill
>
>
>
>
>
>
>
>
>
>
>
>
>
> _______________________________________________
> Red5 mailing list
> [email protected]
> http://osflash.org/mailman/listinfo/red5_osflash.org
>
>   


_______________________________________________
Red5 mailing list
[email protected]
http://osflash.org/mailman/listinfo/red5_osflash.org

Reply via email to