Hi Folks,

I am trying to verify that I am looking at my "problem" from the correct perspective. I have a single thread group and a http sampler that requests the same file repeatedly for some user-defined duration (say 300 seconds). The sampler has a Gaussian Random Timer as a child with offset and std dev set manually (e.g., 500ms offset and 50ms deviation), in order to rule out any weirdness using variables. When I run the test with one thread, I obtain my desired behavior: the delay between requests is normally distributed with mean and std dev very close to the timer settings. However when I use 2 or more threads, I get delay distributions that are not fitting any particular distribution and are certainly not normal. As I go higher with the number of threads (and correspondingly increase the offset and std dev, just to make sure the center mass is sufficiently distant from zero), the distribution starts to look exponential. Am I missing something? If it is true that the sum of two or more normal distributions is itself a normal distribution, then I don't understand why using multiple threads in this fashion results in this behavior.

Any help is greatly appreciated.

Regards,
Marc

PS I reproduced the generation of normally distributed delay times using a Beanshell timer, but obtain similar results.


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to