On 15/09/2009, Shanmugam Ganeshkumar <[email protected]> wrote: > List, > This was posted in one of the list that does special benchmarking. > > Can some expert on jmeter help us to get it right. > > What we need is a table output when we run it from command prompt. How does > this gets executed ? > > basically, how does the listener summary_report is saved while running the > test on command?
Sorry, not possible. The GUI part of Listeners is deliberately not invoked in command-line (non-GUI) mode, and that is where the summarising is done. > > http://jakarta.apache.org/jmeter/usermanual/component_reference.html#Summary_Report > > Thanks for the help > > Cheers > Ganesh > > > > > > ---------- Forwarded message ---------- > From: Andrea Aime > Date: 2009/9/14 > Subject: [Benchmarking] Running the tests on the command line and > summarizing results > > > Hi, > here are some quick instructions to get started doing benchmarking on the > test box. It's a tentative setup and I'm looking for someone > to finish the job ;-) > > In order to run jmeter you first need a jmeter script and possibly a > .csv file that drives its requests. I've attached a sample with > bluearble_ecw.jmx and bluemarble.csv > > The .jmx file has to be generated using the JMeter GUI (it's an > xml file, if you want to hurt yourself you can also write it by hand), > the .csv file can be generated using Frank's bbox generator (wms_request.py, > attached). > Of course the .jmx file contains the url that will need to be > hit, that is something relatively easy to change even in the > .jmx file. > > Once you have those you can run a test using: > jmeter -p jmeter.properties -n -t script.jmx -l script_results.jtl > > The jtl file is actually just a csv file with details of all requests, > you can run the summarizer on it to get a table summary with average time, > throughput and so on, for example: > > ./summarizer.py states.jtl > Label Avg Min Max Throughput > 1 19 13 41 45.0 > 10 85 12 376 80.9 > 20 179 13 851 84.7 > 40 435 14 2175 77.1 > > Now, it would be cool if we had a little wrapper that does > run both, something where one can call: > > runbench script.jmx > > and get the table as an output in one shot. > Any taker? ;-) > > Cheers > Andrea > > > #--------------------------------------------------------------------------- > # Results file configuration > #--------------------------------------------------------------------------- > > # This section helps determine how result data will be saved. > # The commented out values are the defaults. > > # legitimate values: xml, csv, db. Only xml and csv are currently > supported. > jmeter.save.saveservice.output_format=csv > > > # true when field should be saved; false otherwise > > # assertion_results_failure_message only affects CSV output > #jmeter.save.saveservice.assertion_results_failure_message=false > # > #jmeter.save.saveservice.data_type=true > #jmeter.save.saveservice.label=true > #jmeter.save.saveservice.response_code=true > # response_data is not currently supported for CSV output > #jmeter.save.saveservice.response_data=false > # Save ResponseData for failed samples > #jmeter.save.saveservice.response_data.on_error=false > #jmeter.save.saveservice.response_message=true > #jmeter.save.saveservice.successful=true > #jmeter.save.saveservice.thread_name=true > #jmeter.save.saveservice.time=true > #jmeter.save.saveservice.subresults=true > #jmeter.save.saveservice.assertions=true > #jmeter.save.saveservice.latency=true > #jmeter.save.saveservice.samplerData=false > #jmeter.save.saveservice.responseHeaders=false > #jmeter.save.saveservice.requestHeaders=false > #jmeter.save.saveservice.encoding=false > #jmeter.save.saveservice.bytes=true > jmeter.save.saveservice.url=true > #jmeter.save.saveservice.filename=false > #jmeter.save.saveservice.hostname=false > #jmeter.save.saveservice.thread_counts=false > #jmeter.save.saveservice.sample_count=false > > # Timestamp format > # legitimate values: none, ms, or a format suitable for SimpleDateFormat > #jmeter.save.saveservice.timestamp_format=ms > #jmeter.save.saveservice.timestamp_format=MM/dd/yy HH:mm:ss > > # Put the start time stamp in logs instead of the end > sampleresult.timestamp.start=true > > > # legitimate values: none, first, all > #jmeter.save.saveservice.assertion_results=none > > # For use with Comma-separated value (CSV) files or other formats > # where the fields' values are separated by specified delimiters. > # Default: > #jmeter.save.saveservice.default_delimiter=, > # For TAB, since JMeter 2.3 one can use: > #jmeter.save.saveservice.default_delimiter=\t > > #jmeter.save.saveservice.print_field_names=false > > # Optional list of JMeter variable names whose values are to be saved in the > result data files. > # Use commas to separate the names. For example: > #sample_variables=SESSION_ID,REFERENCE > # N.B. The current implementation saves the values in XML as attributes, > # so the names must be valid XML names. > # Versions of JMeter after 2.3.2 send the variable to all servers > # to ensure that the correct data is available at the client. > > # Optional xml processing instruction for line 2 of the file: > #jmeter.save.saveservice.xml_pi=<?xml-stylesheet type="text/xsl" > href="sample.xsl"?> > --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]

