Mac OSX 10.6.8. I suggest you open an issue with your attached test plan (exactly the one that has the issue).
Regards Philippe On Mon, Feb 13, 2012 at 8:28 PM, Robin D. Wilson <rwils...@gmail.com> wrote: > Just out of curiosity - what type of OS was your JMeter GUI running on? > I've been running my tests on Windows XP. > > -- > Robin D. Wilson > Sr. Director of Web Development > KingsIsle Entertainment, Inc. > VOICE: 512-777-1861 > www.KingsIsle.com > > > -----Original Message----- > From: Philippe Mouawad [mailto:philippe.moua...@gmail.com] > Sent: Friday, February 10, 2012 7:41 AM > To: JMeter Users List > Subject: Re: Average times are confusing me... > > Hello, > I have just made the same test with a JSP (doing a Thread.sleep(2000) => > 2s) behind a Tomcat 6 with JMeter 2.6 > > I put 100 threads that loop 100 times in GUI mode (one JMETER): > All results shown in Aggregate report are around 2s . > > > > > > > > > > > > sampler_label,aggregate_report_count,average,aggregate_report_median,aggregate_report_90%_line,aggregate_report_min,aggregate_report > > _max,aggregate_report_error%,aggregate_report_rate,aggregate_report_bandwidth > HTTP > > Request,10000,2006,2004,2012,2001,2195,0.0,49.5793194742609,1224.082013722936 > TOTAL,10000,2006,2004,2012,2001,2195,0.0,49.5793194742609,1224.082013722936 > > > Summary listener shows this: > 2012/02/10 14:36:00 INFO - jmeter.reporters.Summariser: Generate Summary > Results + 2655 in 54,8s = 48,4/s Avg: 2010 Min: 2001 Max: 2195 > Err: 0 (0,00%) > 2012/02/10 14:38:26 INFO - jmeter.reporters.Summariser: Generate Summary > Results + 7345 in 148,9s = 49,3/s Avg: 2005 Min: 2001 Max: 2088 > Err: 0 (0,00%) > 2012/02/10 14:38:26 INFO - jmeter.reporters.Summariser: Generate Summary > Results = 10000 in 201,7s = 49,6/s Avg: 2006 Min: 2001 Max: 2195 > Err: 0 (0,00%) > > So as you can see, nothing abnormal . > > JSP Code: > > <%@ page contentType="text/html; charset=UTF-8" %> > <html> > <body> > <% > Thread.sleep(2000); > %> > <!-- GENERATE around 25Ko page --> > <%=org.apache.commons.lang.RandomStringUtils.random(25000, > "abcdefghijklmnopqrstuvwxyz01234567890")%> > <p>Test</p> > </body> > </html> > > > Regards > Philippe > http://www.ubik-ingenierie.com > > On Thu, Feb 9, 2012 at 4:52 PM, sebb <seb...@gmail.com> wrote: > > > On 9 February 2012 15:38, Robin D. Wilson <rwils...@gmail.com> wrote: > > > Thanks sebb for the replies... > > > > > > Here's the deal, I am running the same test script on JM2.4 and JM2.6. > I > > am running in GUI mode. The test script has 3 thread groups > > > - but the first and the last thread group is just a 'timer' I created > to > > log the total elapsed time of the test (the first and last > > > group has 1 thread, and 1 request, and take less than 1 second each to > > run). The 'real' test is the middle thread group. It has 100 > > > threads (0 ramp), and runs 100 iterations (10,000 total samples). It > > simply does a 'POST' to a URL, with 15 > > > > > > So the 'elapsed time' I referring to in my test is actually the > > timestamp taken in the first thread group (in ms since epoch) > > > subtracted from the timestamp taken in the 3rd (last) thread group. > That > > part of my test may only add 2 total seconds to the test, > > > so while it may skew my results slightly - it doesn't explain the vast > > difference in the 'average' sample duration. According to the > > > Summary Report docs, the "Average" is supposed to be "the average > > elapsed time of a set of samples". But clearly, if the minimum > > > time it takes to actually get the page is 2 seconds (due to the > built-in > > delay in the cgi-script), there is no way I could have an > > > 'average' elapsed time of less than 2 seconds, yet I'm showing an > > average elapsed time of ~750 ms... (My "Max" elapsed time shows as > > > only 1198!). When I request the page in Firefox, it takes ~2104ms > (using > > a status bar timer), so I think the cgi script is working > > > correctly.) > > > > > > Sebb asked: > > > > > >>Again, the throughput calculations are based on total test time. Are > you > > sure the test run times are comparable? > > > > > > The test run times are automatically calculated by the 1st and 3rd > > thread groups. The ~210 seconds total elapsed time is accurate > > > based on my external measurement too (e.g., it is close to what I can > > observe with my stopwatch). > > > > > > Both the JM2.4 test and the JM2.6 test are using the exact same ".jmx" > > test file. > > > > > >>There's clearly something else going on here. > > > > > > I don't believe that the Summary Report is accurately calculating > > anything except the total number of samples and the Avg. Bytes... > > > > What makes you say that? > > Are the Min and Max really incorrect? > > Error %? > > > > It's easy enough to check the Summary Results if you can provide the > > CSV sample result files. > > > > > The cgi-script I'm using definitely takes 2+ seconds to respond after > it > > gets the request (I've measured this with Firefox directly, > > > and it _never_ gets a response in less than 2 seconds). I even changed > > the 'sleep' to 9 seconds, and JMeter pauses for that long in > > > recording results (e.g., it shows 100 threads run, then waits 9 > seconds, > > shows another 100 threads, etc.), but the numbers just go > > > up to '1758' Average, and '2415' Max (which is impossible since it is > > taking 9+ seconds to respond to each request!). It takes over > > > 15 minutes to complete 10,000 samples (and that seems about right - > > 10000 samples/100 threads * 9 seconds each = 900 seconds). > > > > > > I even went so far as to inject a 2 second sleep in the middle of the > > response (e.g., pause 2 seconds - send part of the response - > > > pause 2 more seconds - send the rest), I'm still getting average times > > of ~1000 ms. (That's with 4 seconds of built-in delays, and 2 > > > of those seconds are in the middle of the response.) The browser shows > > this delay properly, but JMeter isn't calculating it > > > properly. > > > > > >>Please recheck the individual sample response times and see how they > > compare to the average. > > > > > > I'm not sure how to do that in JMeter. I can manually hit the page, and > > it takes about 100ms longer than the built-in delay I have. > > > > Add a Table View Listener, or just check the CSV sample result files. > > > > >>If there still appears to be a problem, create a Bugzilla issue and > > attach: > > >>- JMX test case > > > > > > I'm trying to simplify the test case to the bare minimum case - so the > > results will be indisputable. I will also include the > > > 'cgi-bin' script that I'm using, so someone else can easily setup the > > same test. > > > > Thanks. > > > > > > > >>- log files for JMeter 2.4 and 2.6 > > > > > > Which log files are these? Is it just the 'jmeter.log' that gets > created > > in the 'bin' folder when I run the GUI mode, or do you need > > > another log file? > > > > jmeter.log > > > > >>- CSV result files for 2.4 and 2.6 > > > > > > I can do this. > > > > > > -- > > > Robin D. Wilson > > > Sr. Director of Web Development > > > KingsIsle Entertainment, Inc. > > > VOICE: 512-777-1861 > > > www.KingsIsle.com > > > > > > > > > > > > > > > --------------------------------------------------------------------- > > > To unsubscribe, e-mail: user-unsubscr...@jmeter.apache.org > > > For additional commands, e-mail: user-h...@jmeter.apache.org > > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: user-unsubscr...@jmeter.apache.org > > For additional commands, e-mail: user-h...@jmeter.apache.org > > > > > > > -- > Cordialement. > Philippe Mouawad. > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@jmeter.apache.org > For additional commands, e-mail: user-h...@jmeter.apache.org > > -- Cordialement. Philippe Mouawad.