On 08/09/2009, Deepak Shetty <[email protected]> wrote:
> >I need to log the time taken by each request when 100/200/300/400/500
>  >concurrent requests are made. Hope the logger can do that.
>
> Yes.
>
>
>  >when i have 5 users (threads) and 50 users (threads), the througput is
>  >same 12/sec. Now how do I explain the user concurrency,  load /
>  >stress?
>
> See explanation on throughput curves.
>  
> http://books.google.com/books?id=HTX8DyD0WzkC&pg=PA12&lpg=PA12&dq=throughput+curve&source=bl&ots=7qYRIZiPX9&sig=7UoxT-8gpbmqWwwUcu0aROe_QWA&hl=en&ei=1X6mSqLyHpDK_gbMgvC-CQ&sa=X&oi=book_result&ct=result&resnum=5#v=onepage&q=throughput%20curve&f=false
>
>  You have reached your throughput 'plateau' and need to check response times
>  as well..
>

This could be due to:
* network saturation (unlikely at this throughput unless the response
are huge, though using virtual hosts on a single physical system may
be relevant)
* JMeter limit (unlikely with only 50 threads - assuming you have not
added a throughput timer!)
* host resource exhaustion - possible, given that everything is
running on the same host
* server resource exhaustion - again possible, if not configured with
enough sessions.

>  >How to measure the load / stress on the server?
>
> Thats server specific, your O.S. will give you tools to do this. (e.g.
>  perfmon on windows, vmstat on unix , other tools ).
>  regards
>
> deepak
>
>
>
>  On Tue, Sep 8, 2009 at 3:35 AM, Bruce Foster <[email protected]> wrote:
>
>  > Thanks All,
>  >
>  > I will try the options and let you know. Got distracted with some
>  > other work and will spend some time on the benchmarking next week.
>  >
>  > I need to log the time taken by each request when 100/200/300/400/500
>  > concurrent requests are made. Hope the logger can do that.
>  >
>  > I have some  basic question being newbie;
>  >
>  > when i have 5 users (threads) and 50 users (threads), the througput is
>  > same 12/sec. Now how do I explain the user concurrency,  load /
>  > stress?
>  >
>  > I need to find out if the system can handle 500 concurrent users.
>  >
>  > Throughput is the response time right which turns out to be around
>  > 85ms (12/sec), since there are no change from 5 to 50, how do I test
>  > for 500 concurrent users ( or 300 or 200)?
>  >
>  > How to measure the load / stress on the server?
>  >
>  > Thanks a lot
>  >
>  > Bruce
>  >
>  >
>  >
>  >
>  >
>  >
>  >
>  > Thanks
>  > Bruce
>  >
>  >
>  >
>  > On Sat, Sep 5, 2009 at 6:09 PM, sebb<[email protected]> wrote:
>  > > On 05/09/2009, Bruce Foster <[email protected]> wrote:
>  > >> Hi Deepak and others,
>  > >>
>  > >>  Thanks for quick response and help.
>  > >>
>  > >>  Yes, the listener Save_Responses_to_a_file did the trick for me. Just
>  > >>  ran a test with 1000 request to see the response and got all the
>  > >>  images saved in directory. Well, the purpose was to check the response
>  > >>  and not the performance (response time). After making sure that the
>  > >>  image are correct, I ran the actual test to get the performance
>  > >>  results.
>  > >>
>  > >>  Well, I'm using the random function and it worked well to generate
>  > >>  random bound box request. Also, I adapted the osgeo test method of
>  > >>  using pre generated csv file.
>  > >>
>  > >>  got a good result of 12 user per second in one method for total
>  > >>  random, and 20 users per second for 800x600px random bbox request.
>  > >>  need further more to test.
>  > >>
>  > >>  now i have to find out how to log the 10000 request time. jmeter gives
>  > >>  only summary/average.
>  > >
>  > > In the GUI, that depends on the Listener - e.g. the Table View
>  > > Listener shows response times. But don't use this for a performance
>  > > test as it will use lots of memory.
>  > >
>  > > Just save the responses to a file, and you have all the details there,
>  > > depending on what you have configured. Probably easiest to use CSV
>  > > output.
>  > >
>  > >>  Cheers
>  > >>
>  > >> bruce.
>  > >>
>  > >>
>  > >>
>  > >>
>  > >>
>  > >>  On Thu, Sep 3, 2009 at 7:16 PM, sebb<[email protected]> wrote:
>  > >>  > On 03/09/2009, Adrian Speteanu <[email protected]> wrote:
>  > >>  >> true, you can use either method for what you said you need, but in
>  > >>  >>  this case, saving the file on the test machine will significantly
>  > >>  >>  increase the stress on the test environment (quality image files
>  > mean
>  > >>  >>  lots of space and that means disk usage).
>  > >>  >>
>  > >>  >>  if you run the test with fewer requests and see that you get the
>  > >>  >>  responses you expect, then you will also get these responses in a
>  > load
>  > >>  >>  / stress test even if you don't save the files locally.
>  > >>  >
>  > >>  > Not necessarily; the server may degrade under load.
>  > >>  >
>  > >>  > For checking responses such as images, consider using
>  > >>  >
>  > >>  >
>  > 
> http://jakarta.apache.org/jmeter/usermanual/component_reference.html#MD5Hex_Assertion
>  > >>  >
>  > >>  > Or you can use the HTTP sampler option "Save response as MD5 hash?"
>  > >>  > and check that.
>  > >>  >
>  > >>  >>  this is
>  > >>  >>  recommended.
>  > >>  >>
>  > >>  >>
>  > >>  >>  On Tue, Sep 1, 2009 at 2:04 AM, Deepak Shetty<[email protected]>
>  > wrote:
>  > >>  >>  > Hi
>  > >>  >>  > you can add
>  > >>  >>  >
>  > 
> http://jakarta.apache.org/jmeter/usermanual/component_reference.html#Save_Responses_to_a_file
>  > >>  >>  > OR you can add a BeanShell Post Assertion  that can read the
>  > bytes and save
>  > >>  >>  > it to whatever you want or run comparisons
>  > >>  >>  > OR
>  > >>  >>  >
>  > 
> http://jakarta.apache.org/jmeter/usermanual/component_reference.html#Sample_Result_Save_Configuration
>  > >>  >>  > (Check Save Response Data) - I wouldnt do this though because
>  > some binary
>  > >>  >>  > can cause the xml to break
>  > >>  >>  >
>  > >>  >>  >
>  > >>  >>  > regards
>  > >>  >>  > deepak
>  > >>  >>  >
>  > >>  >>  > On Mon, Aug 31, 2009 at 3:57 PM, Bruce Foster <
>  > [email protected]> wrote:
>  > >>  >>  >
>  > >>  >>  >> Hi List,
>  > >>  >>  >>
>  > >>  >>  >> I'm totally new to jmeter and also benchmarking.
>  > >>  >>  >>
>  > >>  >>  >> I'm testing a WMS (web map service) service performance of three
>  > >>  >>  >> server softwares. Basically, they are GET request of images from
>  > a
>  > >>  >>  >> server.
>  > >>  >>  >>
>  > >>  >>  >> Is there a way to SAVE the requested images? I have the mandate
>  > to
>  > >>  >>  >> make sure that the response from the servers are exactly the
>  > same
>  > >>  >>  >> image (in resolution, quality) that we request for.
>  > >>  >>  >>
>  > >>  >>  >> When I did a test, I put a network monitor. I could see 70mb of
>  > data
>  > >>  >>  >> is transfered. Now, where to look for that, does jmeter save
>  > them in
>  > >>  >>  >> cache?
>  > >>  >>  >>
>  > >>  >>  >> Note, I'm doing everything on a vmware machine running on my
>  > notebook.
>  > >>  >>  >>
>  > >>  >>  >>
>  > >>  >>  >> Thanks
>  > >>  >>  >> Bruce
>  > >>  >>  >>
>  > >>  >>  >>
>  > ---------------------------------------------------------------------
>  > >>  >>  >> To unsubscribe, e-mail:
>  > [email protected]
>  > >>  >>  >> For additional commands, e-mail:
>  > [email protected]
>  > >>  >>  >>
>  > >>  >>  >>
>  > >>  >>  >
>  > >>  >>
>  > >>  >>
>  >  ---------------------------------------------------------------------
>  > >>  >>  To unsubscribe, e-mail: [email protected]
>  > >>  >>  For additional commands, e-mail:
>  > [email protected]
>  > >>  >>
>  > >>  >>
>  > >>  >
>  > >>  > ---------------------------------------------------------------------
>  > >>  > To unsubscribe, e-mail: [email protected]
>  > >>  > For additional commands, e-mail: [email protected]
>  > >>  >
>  > >>  >
>  > >>
>  > >>  ---------------------------------------------------------------------
>  > >>  To unsubscribe, e-mail: [email protected]
>  > >>  For additional commands, e-mail: [email protected]
>  > >>
>  > >>
>  > >
>  > > ---------------------------------------------------------------------
>  > > To unsubscribe, e-mail: [email protected]
>  > > For additional commands, e-mail: [email protected]
>  > >
>  > >
>  >
>  > ---------------------------------------------------------------------
>  > To unsubscribe, e-mail: [email protected]
>  > For additional commands, e-mail: [email protected]
>  >
>  >
>

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to