On 09/12/05, Peter Lin <[EMAIL PROTECTED]> wrote:
> thanks for explaining. that makes sense now. given the application is
> caching, having different requests would be crucial for valid measurement.
> chances are, you'll need to use atleast 4 clients and split the test plan
> into 4 smaller test plans.  this way, it increases the chances that the
> threads will have a shorter delay between each thread.
>
> in the past, when I've had to test applications with cache, we made it so
> the cache can be turned off.  This way, we can test the impact of concurrent
> queries, versus the webserver's ability to handle 100 concurrent requests.
> If your application doesn't have the capability, it's really going to be
> hard to effectively test the impact of traffic spike.

Unless you can add some variability to the URL to ensure that the
cache does not contain the request.

> peter
>
>
> On 12/9/05, Iago Toral Quiroga <[EMAIL PROTECTED]> wrote:
> >
> > El vie, 09-12-2005 a las 18:49, Peter Lin escribió:
> > > honestly, I don't understand why first request needs to be different for
> > all
> > > threads.  if the point is to measure an application's ability to handle
> > a
> > > sudden spike, it's better to pick a very heavy page and set 1
> > threadGroup
> > > with 100 threads and hit it.
> >
> > Because the web server is serving a GIS application that has a cache
> > system. So I need all the requests to be different in order to avoid
> > cached responses.
> >
> > > using different thread groups just means you have to ramp for a longer
> > > period. I can't stress enough how hard it is to really get 100
> > concurrent
> > > requests.  From my experience, what matters more is the system is able
> > to
> > > handle a sudden spike gracefully without brinding down the website and
> > > return to normal operation once the spike has passed.  100 concurrent
> > > requests for an average size webpage 10Kb, means that in an ideal
> > situation
> > > one would need a full 100mbit bandwidth. On a 10mbit bandwidth, it's
> > never
> > > going to reach that. It's physically impossible.
> > >
> > > unless the hosting facility has a dedicated OC12, it won't be able to
> > handle
> > > 100 concurrent.  for some perspective, 40 concurrent requests for 18hrs
> > a
> > > day translates to 10million pages views.  I know this from first hand
> > > experience working at superpages.com.  98% of the sites out there don't
> > get
> > > any where near this kind of traffic.
> > >
> >
> > I'm not talking about of being able to serve 100 requests in one second
> > to the clients. What I want to know is what happens at the server when
> > 100 requests appear simultaneouly. Surely I need a huge bandwidth to
> > give response to all those requests, but not to get the requests
> > themselves, which is the point. A http request is very short in size,
> > let's say 500 bytes, so you don't need a huge bandwidth in order to
> > receive them.
> >
> > So, if I can receive 100 simultaneous requests, what will happen to my
> > server? will it crash? will it refuse connections? will it be able to
> > continue working? at which performce? etc... that is what I want to
> > know.
> >
> > > On 12/9/05, Iago Toral Quiroga <[EMAIL PROTECTED]> wrote:
> > > >
> > > > Thanks for your comment sebb,
> > > >
> > > > if I have more than one thread in each thread group my problem is
> > > > ensuring that each thread launches a different request, because each
> > > > thread will send the same sequence of requests under the threadgroup.
> > > > I've tried using an interleave controller, but it deals the requests
> > for
> > > > each thread and not for all the threads in the threadgroup :(
> > > >
> > > > Iago.
> > > >
> > > > El vie, 09-12-2005 a las 18:01, sebb escribió:
> > > > > I suspect part of the problem is that all the threads start at once,
> > > > > and having 100 thread groups with only 1 thread in each will make it
> > > > > tedious to fix - you'll need to add a gradually increasing delay to
> > > > > each of the thread groups.
> > > > > What happens if you have fewer thread groups and more threads in
> > each
> > > > group?
> > > > > You can set the ramp-up for each thread-group to ensure that the
> > > > > threads start more evenly.
> > > > >
> > > > > S.
> > > > > On 09/12/05, Iago Toral Quiroga <[EMAIL PROTECTED]> wrote:
> > > > > > Hi,
> > > > > >
> > > > > > I've configured a test with 100 thread groups (one thread per
> > thread
> > > > > > group) and added a constant throughput timer to get a 10 requests
> > per
> > > > > > second performance. To do so, I configured target throughput to
> > 600
> > > > > > (samples per minute) and selected to compute performance based on
> > all
> > > > > > active threads.
> > > > > >
> > > > > > The result is as expected, I get an average throughput of 10
> > requests
> > > > > > per second, but they are not uniform along the time. What I get is
> > > > > > something like this:
> > > > > >
> > > > > > At second 0, jmeter launches 100 requests to the server. At second
> > 4,
> > > > > > jmeter has received all the responses, but because it has lauched
> > 100
> > > > > > requests at second 0, it must wait till second 10 to start another
> > > > bunch
> > > > > > of 100 requests. What I expect from this kind of tests is getting
> > 10
> > > > > > requests per second *each second*.
> > > > > >
> > > > > > This kind of behaviour is much more like a repeated peak test than
> > a
> > > > > > constant troughput test. I know I can get a more uniform test by
> > > > droping
> > > > > > the thread count so jmeter would have to wait less time to launch
> > the
> > > > > > next bunch of requests, but that is weird and still a trick that
> > does
> > > > > > not solve the point of problem at all ¿I'm missing something?, ¿is
> > > > there
> > > > > > a way to get a more uniform behaviour for this kind of tests?
> > > > > >
> > > > > > Thanks in advance for your help!
> > > > > > --
> > > > > > Abel Iago Toral Quiroga
> > > > > > Igalia http://www.igalia.com
> > > > > >
> > > > > >
> > ---------------------------------------------------------------------
> > > > > > To unsubscribe, e-mail: [EMAIL PROTECTED]
> > > > > > For additional commands, e-mail:
> > [EMAIL PROTECTED]
> > > > > >
> > > > > >
> > > > >
> > > > >
> > ---------------------------------------------------------------------
> > > > > To unsubscribe, e-mail: [EMAIL PROTECTED]
> > > > > For additional commands, e-mail: [EMAIL PROTECTED]
> > > > --
> > > > Abel Iago Toral Quiroga
> > > > Igalia http://www.igalia.com
> > > >
> > > > ---------------------------------------------------------------------
> > > > To unsubscribe, e-mail: [EMAIL PROTECTED]
> > > > For additional commands, e-mail: [EMAIL PROTECTED]
> > > >
> > > >
> > --
> > Abel Iago Toral Quiroga
> > Igalia http://www.igalia.com
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: [EMAIL PROTECTED]
> > For additional commands, e-mail: [EMAIL PROTECTED]
> >
> >
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to