On 07/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote:
> I haven't had the opportunity to test each of the nodes in single mode, just
>  due to the nature of my work place environment (Most of the computers are
>  locked, I was just given an IP and told that JMeter-server is running.) I
>  know all the data is in sync because they're all synchronized using
>  subversion.
>
>  I ran some tests yesterday, comparing server response time (using JAMon) vs
>  Jmeter's response time, and I found that my results start to diverge at 20
>  users (spread out amongst all 7 of the nodes).
>
>  Regardless, I'm almost positive (as I said before) that its because of all
>  the optimistic locking (because of the situation with subversion, its
>  impossible right now to adjust the datafiles for everyone) exceptions being
>  thrown all the time.

If you include the hostname in the filename, then you can have
different files for each host.

>  Since we've aquired a faster machine (dual core, 2 gigs ram...) and set the
>  heap up on jmeter to like 1024m, that can run it beautifully in single node
>  mode, I unfortunately have to abandon the distributed test environment.
>  (Corporate...)
>
>  Thanks much for your time!
>
>
>  Michael
>
>
>  On Sat, Jun 7, 2008 at 5:35 AM, sebb <[EMAIL PROTECTED]> wrote:
>
>  > You originally wrote that "single user work station ... results come
>  > back fantastically".
>  >
>  > Is that true for all the nodes used individually?
>  > Or do some nodes struggle?
>  >
>  > If all nodes work OK individually, try adding more nodes to the test,
>  > i.e instead of one node with 300 users, use 2 nodes with 150 users
>  > each. Try these both completely separately - in non-GUI mode - and in
>  > client-server mode.
>  >
>  > If you can run all the nodes together in non-GUI mode and get the same
>  > performance as
>  > a single node with the same total of users, but not when you run in
>  > client-server mode then there must be a bottleneck in the client.
>  >
>  > You also wrote:
>  >
>  > "We're load testing a web application, so primarily, the only work we're
>  > doing is http requests ...
>  >
>  > However, when we view the transactions in the database, they are extremely
>  > low. (frighteningly low)."
>  >
>  > This does not make sense. If the http requests directly relate to
>  > database transactions, then either some requests are not succeeding,
>  > or there is an error in the test data or the application. You'll need
>  > to trace the progress of the http request in the server to see why
>  > some requests are not generating database transactions.
>  >
>  > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote:
>  > > Actually, that's about the exact configuration across all 7 of them.
>  > >
>  > >
>  > >  On Fri, Jun 6, 2008 at 9:58 AM, sebb <[EMAIL PROTECTED]> wrote:
>  > >
>  > >  > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote:
>  > >  > > One would think so, but my response times are terrible, I'm
>  > wondering if
>  > >  > it
>  > >  > >  has anything to do with the fact that its a 1st gen p4.
>  > >  > >  (Ubuntu Hardy, 1.5 Gig PC3200 RAM)
>  > >  > >
>  > >  >
>  > >  > What node is that?
>  > >  >
>  > >  > You wrote:
>  > >  >
>  > >  > "Then we run the test from a single user work station (same test, 300
>  > users
>  > >  > doing work) and our results come back fantastically!"
>  > >  >
>  > >  > >
>  > >  > >  On Fri, Jun 6, 2008 at 9:44 AM, sebb <[EMAIL PROTECTED]> wrote:
>  > >  > >
>  > >  > >  > Yes, 300 users should not be a problem as a single JMeter client
>  > >  > >  > should be able to handle that.
>  > >  > >  >
>  > >  > >  > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]> wrote:
>  > >  > >  > > That makes sense. I'll give it a go. (We're pretty sure there's
>  > no
>  > >  > bottle
>  > >  > >  > >  neck passing things, they do it every 100 samples, and this is
>  > over
>  > >  > a
>  > >  > >  > 100
>  > >  > >  > >  MB/s net. I'm only trying to run 300 users, so they should be
>  > able
>  > >  > to
>  > >  > >  > >  perform well over a 10 MB/s
>  > >  > >  > >
>  > >  > >  > >
>  > >  > >  > >  On Fri, Jun 6, 2008 at 9:27 AM, sebb <[EMAIL PROTECTED]>
>  > wrote:
>  > >  > >  > >
>  > >  > >  > >  > In client-server mode, only the test plan is sent from the
>  > client
>  > >  > to
>  > >  > >  > >  > the server(s).
>  > >  > >  > >  >
>  > >  > >  > >  > Any additional files - e.g. CSV input files - need to be
>  > present
>  > >  > on
>  > >  > >  > >  > the server host in the location specified by  the test plan.
>  > >  > >  > >  >
>  > >  > >  > >  > Sample data is returned to the client, and processed/stored
>  > by
>  > >  > the
>  > >  > >  > client.
>  > >  > >  > >  > This can become a bottleneck at the client - both for JMeter
>  > >  > itself,
>  > >  > >  > >  > and for the network connection - under high loads.
>  > >  > >  > >  >
>  > >  > >  > >  > Data files are best randomised before use.
>  > >  > >  > >  > Likewise, if you want to run with different data on
>  > different
>  > >  > hosts,
>  > >  > >  > >  > then create different data files for each host (but you can
>  > use
>  > >  > the
>  > >  > >  > >  > same name).
>  > >  > >  > >  >
>  > >  > >  > >  > On 06/06/2008, Michael McDonnell <[EMAIL PROTECTED]>
>  > wrote:
>  > >  > >  > >  > > How did you randomize the data from the CSVs? (if I may
>  > ask)
>  > >  > >  > >  > >
>  > >  > >  > >  > >  Also, I'm dealing with a lot of optimistic locking issues
>  > >  > which
>  > >  > >  > would
>  > >  > >  > >  > only
>  > >  > >  > >  > >  occur if each csv is doing the EXACT same thing at the
>  > exact
>  > >  > same
>  > >  > >  > time
>  > >  > >  > >  > >  (which is completely likely)
>  > >  > >  > >  > >
>  > >  > >  > >  > >
>  > >  > >  > >  > >  On Thu, Jun 5, 2008 at 9:54 PM, Ryan Dooley <
>  > >  > [EMAIL PROTECTED]
>  > >  > >  > >
>  > >  > >  > >  > wrote:
>  > >  > >  > >  > >
>  > >  > >  > >  > >  > I had a similar experience the first time.  Turns out
>  > that
>  > >  > the
>  > >  > >  > data I
>  > >  > >  > >  > >  > wanted
>  > >  > >  > >  > >  > to test with (HTTP POSTs) has to be put on each remote.
>  >  I
>  > >  > also
>  > >  > >  > had a
>  > >  > >  > >  > >  > process to randomize the data when transferred to the
>  > >  > remotes.  I
>  > >  > >  > >  > finally
>  > >  > >  > >  > >  > got the load up high enough across 10 machines like
>  > yours.
>  > >  > >  > >  > >  >
>  > >  > >  > >  > >  > The test harness I had was pretty simple:  post these
>  > things
>  > >  > to
>  > >  > >  > this
>  > >  > >  > >  > url.
>  > >  > >  > >  > >  >
>  > >  > >  > >  > >  > On Thu, Jun 5, 2008 at 5:19 PM, Michael McDonnell <
>  > >  > >  > >  > [EMAIL PROTECTED]>
>  > >  > >  > >  > >  > wrote:
>  > >  > >  > >  > >  >
>  > >  > >  > >  > >  > > We're running a distributed test (roughly 7 remote
>  > >  > >  > workstations) on
>  > >  > >  > >  > a
>  > >  > >  > >  > >  > > pretty
>  > >  > >  > >  > >  > > hefty box (8 cores, 32 gigs ram.... etc...)
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > However, something seems to be going wrong... perhaps
>  > its
>  > >  > >  > because
>  > >  > >  > >  > I'm
>  > >  > >  > >  > >  > > crossing linux and windows platforms to try to do the
>  > >  > testing?
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > We're load testing a web application, so primarily,
>  > the
>  > >  > only
>  > >  > >  > work
>  > >  > >  > >  > we're
>  > >  > >  > >  > >  > > doing is http requests (there are a few "java
>  > requests"
>  > >  > that
>  > >  > >  > >  > actually is
>  > >  > >  > >  > >  > an
>  > >  > >  > >  > >  > > app I created to make webservice calls, but we'll get
>  > to
>  > >  > that
>  > >  > >  > later)
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > However, when we view the transactions in the
>  > database,
>  > >  > they
>  > >  > >  > are
>  > >  > >  > >  > >  > extremely
>  > >  > >  > >  > >  > > low. (frighteningly low).
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > Then we run the test from a single user work station
>  > (same
>  > >  > >  > test, 300
>  > >  > >  > >  > >  > users
>  > >  > >  > >  > >  > > doing work) and our results come back fantastically!
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > Now granted: I guess the big deal is this: when the
>  > app
>  > >  > uses a
>  > >  > >  > csv
>  > >  > >  > >  > in
>  > >  > >  > >  > >  > > distributed mode, does each slave utilize the the
>  > same csv
>  > >  > in
>  > >  > >  > the
>  > >  > >  > >  > same
>  > >  > >  > >  > >  > > order
>  > >  > >  > >  > >  > > ? or is there a sort of "break up" so that no two
>  > slaves
>  > >  > are
>  > >  > >  > using
>  > >  > >  > >  > the
>  > >  > >  > >  > >  > same
>  > >  > >  > >  > >  > > line in the csv?
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > I'm sorry for what may be dumb questions... but we're
>  > >  > coming
>  > >  > >  > down to
>  > >  > >  > >  > a
>  > >  > >  > >  > >  > > tight
>  > >  > >  > >  > >  > > deadline, and the distributed testing is not giving
>  > us
>  > >  > good
>  > >  > >  > results
>  > >  > >  > >  > where
>  > >  > >  > >  > >  > > as
>  > >  > >  > >  > >  > > the local testing is.
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > Thanks for all your help in advance.
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  > > Michael
>  > >  > >  > >  > >  > >
>  > >  > >  > >  > >  >
>  > >  > >  > >  > >
>  > >  > >  > >  >
>  > >  > >  > >
>  > >  > >  > > >
>  > >  > ---------------------------------------------------------------------
>  > >  > >  > >  > To unsubscribe, e-mail:
>  > >  > [EMAIL PROTECTED]
>  > >  > >  > >  > For additional commands, e-mail:
>  > >  > [EMAIL PROTECTED]
>  > >  > >  > >  >
>  > >  > >  > >  >
>  > >  > >  > >
>  > >  > >  >
>  > >  > >  >
>  > ---------------------------------------------------------------------
>  > >  > >  > To unsubscribe, e-mail:
>  > [EMAIL PROTECTED]
>  > >  > >  > For additional commands, e-mail:
>  > [EMAIL PROTECTED]
>  > >  > >  >
>  > >  > >  >
>  > >  > >
>  > >  >
>  > >  > ---------------------------------------------------------------------
>  > >  > To unsubscribe, e-mail: [EMAIL PROTECTED]
>  > >  > For additional commands, e-mail: [EMAIL PROTECTED]
>  > >  >
>  > >  >
>  > >
>  >
>  > ---------------------------------------------------------------------
>  > To unsubscribe, e-mail: [EMAIL PROTECTED]
>  > For additional commands, e-mail: [EMAIL PROTECTED]
>  >
>  >
>

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to