both systems must be insync. That's fundamental to all distributed applications, including distributed testing.
peter On Tue, Oct 20, 2009 at 2:13 PM, Carl Shaulis <[email protected]> wrote: > > The difference appears to be about 10 seconds between the clock on my > machine and the slave server. I added a constant timer and that made no > difference. > > Do the two machines really have to be set down to the exact second? > > I would think we are measuring the delta between start and stop on the same > machine, so the clocks should not matter. > > Thanks, > > Carl > > On 10/20/09 1:06 PM, "Deepak Shetty" <[email protected]> wrote: > >> are the time clocks on both machines in sync? >> >> On Tue, Oct 20, 2009 at 11:02 AM, Carl Shaulis <[email protected]>wrote: >> >>> Hello, >>> >>> We have recently set up a distributed JMeter environment. I am using my >>> MacBook Pro as the Master and a Linux machine as the slave. I executed a >>> very simple test for 5 minutes, where 500 concurrent users access a static >>> html page. The results showed an average response time of 0 ms. Looking >>> more closely at the data there are numerous transactions that look like >>> this. >>> >>> Thread Name: SorryPageTest 1-97 >>> Sample Start: 2009-10-20 12:42:29 CDT >>> Load time: -897 >>> Latency: -897 >>> Size in bytes: 1723 >>> Sample Count: 1 >>> Error Count: 0 >>> Response code: 200 >>> Response message: OK >>> >>> How can you get a negative load time and negative latency with a 200 >>> response code? >>> >>> Help! >>> >>> Carl >>> > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [email protected] > For additional commands, e-mail: [email protected] > > --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]

