You originally wrote that "single user work station ... results come back fantastically".
Is that true for all the nodes used individually? Or do some nodes struggle? If all nodes work OK individually, try adding more nodes to the test, i.e instead of one node with 300 users, use 2 nodes with 150 users each. Try these both completely separately - in non-GUI mode - and in client-server mode. If you can run all the nodes together in non-GUI mode and get the same performance as a single node with the same total of users, but not when you run in client-server mode then there must be a bottleneck in the client. You also wrote: "We're load testing a web application, so primarily, the only work we're doing is http requests ... However, when we view the transactions in the database, they are extremely low. (frighteningly low)." This does not make sense. If the http requests directly relate to database transactions, then either some requests are not succeeding, or there is an error in the test data or the application. You'll need to trace the progress of the http request in the server to see why some requests are not generating database transactions. On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote: > Actually, that's about the exact configuration across all 7 of them. > > > On Fri, Jun 6, 2008 at 9:58 AM, sebb <[EMAIL PROTECTED]> wrote: > > > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote: > > > One would think so, but my response times are terrible, I'm wondering if > > it > > > has anything to do with the fact that its a 1st gen p4. > > > (Ubuntu Hardy, 1.5 Gig PC3200 RAM) > > > > > > > What node is that? > > > > You wrote: > > > > "Then we run the test from a single user work station (same test, 300 users > > doing work) and our results come back fantastically!" > > > > > > > > On Fri, Jun 6, 2008 at 9:44 AM, sebb <[EMAIL PROTECTED]> wrote: > > > > > > > Yes, 300 users should not be a problem as a single JMeter client > > > > should be able to handle that. > > > > > > > > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote: > > > > > That makes sense. I'll give it a go. (We're pretty sure there's no > > bottle > > > > > neck passing things, they do it every 100 samples, and this is over > > a > > > > 100 > > > > > MB/s net. I'm only trying to run 300 users, so they should be able > > to > > > > > perform well over a 10 MB/s > > > > > > > > > > > > > > > On Fri, Jun 6, 2008 at 9:27 AM, sebb <[EMAIL PROTECTED]> wrote: > > > > > > > > > > > In client-server mode, only the test plan is sent from the client > > to > > > > > > the server(s). > > > > > > > > > > > > Any additional files - e.g. CSV input files - need to be present > > on > > > > > > the server host in the location specified by the test plan. > > > > > > > > > > > > Sample data is returned to the client, and processed/stored by > > the > > > > client. > > > > > > This can become a bottleneck at the client - both for JMeter > > itself, > > > > > > and for the network connection - under high loads. > > > > > > > > > > > > Data files are best randomised before use. > > > > > > Likewise, if you want to run with different data on different > > hosts, > > > > > > then create different data files for each host (but you can use > > the > > > > > > same name). > > > > > > > > > > > > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote: > > > > > > > How did you randomize the data from the CSVs? (if I may ask) > > > > > > > > > > > > > > Also, I'm dealing with a lot of optimistic locking issues > > which > > > > would > > > > > > only > > > > > > > occur if each csv is doing the EXACT same thing at the exact > > same > > > > time > > > > > > > (which is completely likely) > > > > > > > > > > > > > > > > > > > > > On Thu, Jun 5, 2008 at 9:54 PM, Ryan Dooley < > > [EMAIL PROTECTED] > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > I had a similar experience the first time. Turns out that > > the > > > > data I > > > > > > > > wanted > > > > > > > > to test with (HTTP POSTs) has to be put on each remote. I > > also > > > > had a > > > > > > > > process to randomize the data when transferred to the > > remotes. I > > > > > > finally > > > > > > > > got the load up high enough across 10 machines like yours. > > > > > > > > > > > > > > > > The test harness I had was pretty simple: post these things > > to > > > > this > > > > > > url. > > > > > > > > > > > > > > > > On Thu, Jun 5, 2008 at 5:19 PM, Michael McDonnell < > > > > > > [EMAIL PROTECTED]> > > > > > > > > wrote: > > > > > > > > > > > > > > > > > We're running a distributed test (roughly 7 remote > > > > workstations) on > > > > > > a > > > > > > > > > pretty > > > > > > > > > hefty box (8 cores, 32 gigs ram.... etc...) > > > > > > > > > > > > > > > > > > However, something seems to be going wrong... perhaps its > > > > because > > > > > > I'm > > > > > > > > > crossing linux and windows platforms to try to do the > > testing? > > > > > > > > > > > > > > > > > > We're load testing a web application, so primarily, the > > only > > > > work > > > > > > we're > > > > > > > > > doing is http requests (there are a few "java requests" > > that > > > > > > actually is > > > > > > > > an > > > > > > > > > app I created to make webservice calls, but we'll get to > > that > > > > later) > > > > > > > > > > > > > > > > > > However, when we view the transactions in the database, > > they > > > > are > > > > > > > > extremely > > > > > > > > > low. (frighteningly low). > > > > > > > > > > > > > > > > > > Then we run the test from a single user work station (same > > > > test, 300 > > > > > > > > users > > > > > > > > > doing work) and our results come back fantastically! > > > > > > > > > > > > > > > > > > Now granted: I guess the big deal is this: when the app > > uses a > > > > csv > > > > > > in > > > > > > > > > distributed mode, does each slave utilize the the same csv > > in > > > > the > > > > > > same > > > > > > > > > order > > > > > > > > > ? or is there a sort of "break up" so that no two slaves > > are > > > > using > > > > > > the > > > > > > > > same > > > > > > > > > line in the csv? > > > > > > > > > > > > > > > > > > I'm sorry for what may be dumb questions... but we're > > coming > > > > down to > > > > > > a > > > > > > > > > tight > > > > > > > > > deadline, and the distributed testing is not giving us > > good > > > > results > > > > > > where > > > > > > > > > as > > > > > > > > > the local testing is. > > > > > > > > > > > > > > > > > > Thanks for all your help in advance. > > > > > > > > > > > > > > > > > > Michael > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --------------------------------------------------------------------- > > > > > > To unsubscribe, e-mail: > > [EMAIL PROTECTED] > > > > > > For additional commands, e-mail: > > [EMAIL PROTECTED] > > > > > > > > > > > > > > > > > > > > > > > > > --------------------------------------------------------------------- > > > > To unsubscribe, e-mail: [EMAIL PROTECTED] > > > > For additional commands, e-mail: [EMAIL PROTECTED] > > > > > > > > > > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: [EMAIL PROTECTED] > > For additional commands, e-mail: [EMAIL PROTECTED] > > > > > --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]

