Unfortunately, the results are almost exactly the same with haproxy 1.4 and
those changes you recommended. I'm so confused...

Thanks for your help!

On Sat, Jan 29, 2011 at 3:25 PM, Sean Hess <[email protected]> wrote:

> Ok, here are the results from apache benchmark *before* making any other
> changes to the system (1.4, timeouts, etc).
>
> The Test - https://gist.github.com/802251
>
> The Results against the 1*256 haproxy -> 4*512 node cluster -
> https://gist.github.com/802268
> Here's the haproxy status after the test -
> http://dl.dropbox.com/u/1165308/ab_haproxy.png
>
> The Results against the 1*512 node instance -
> https://gist.github.com/802271
>
> So, it looks about the same. The single instance outperforms the cluster,
> which doesn't make any sense. I'll try those changes and see if it gets any
> better.
>
>
>
> On Sat, Jan 29, 2011 at 2:44 PM, Sean Hess <[email protected]> wrote:
>
>> Thanks Joel,
>>
>> I'm working on converting the test to ab (shouldn't take long) and trying
>> out 1.4, but to answer your questions. RSTavg is average response time.
>> There's a 500ms timer in the http response, and some serialization. It's
>> over the local network. So that should be about 550ms under no load.
>>
>> Users per second, yes.
>>
>> I didn't use ab to start because I'm not interested in response time, per
>> se, but at what load response time starts to fail. I don't know an effective
>> way to do this with ab, partially because it doesn't support stepping (my
>> test steps through the concurrency levels specified by "users", I should
>> rename Usersps to sessions per second, because if a "user" takes less
>> than 1 second they start again right away). My testing harness allows me to
>> write tests in my application language, blah blah.. you get the idea. But
>> yes, I'll run ab and see if I get the same results.
>>
>> I'll also try your changes to the timeouts. Thanks for your help!
>>
>>
>>
>> On Sat, Jan 29, 2011 at 1:00 PM, Joel Krauska <[email protected]> wrote:
>>
>>> Speculation, but using a newer version of haproxy (1.4) might also
>>> improve performance for you.
>>>
>>> --Joel
>>>
>>>
>>> On 1/29/11 10:53 AM, Sean Hess wrote:
>>>
>>>> I'm performing real-world load tests for the first time, and my results
>>>> aren't making a lot of sense.
>>>>
>>>> Just to make sure I have the test harness working, I'm not testing
>>>> "real" application code yet, I'm just hitting a web page that simulates
>>>> an IO delay (500 ms), and then serializes out some json (about 85 bytes
>>>> of content). It's not accessing the database, or doing anything other
>>>> than printing out that data. My application servers are written in
>>>> node.js, on 512MB VPSes on rackspace (centos55).
>>>>
>>>> Here are the results that don't make sense:
>>>>
>>>> https://gist.github.com/802082
>>>>
>>>> When I run this test against a single application server (bottom one),
>>>> You can see that it stays pretty flat (about 550ms response time) until
>>>> it gets to 1500 simultaneous users, when it starts to error out and get
>>>> slow.
>>>>
>>>> When I run it against an haproxy instance in front of 4 of the same
>>>> nodes (top one), my performance is worse. It doesn't drop any
>>>> connections, but the response time edges up much earlier than against a
>>>> single node.
>>>>
>>>> Does this make any sense to you? Does haproxy need more RAM? I was
>>>> watching the box while the test was running and the haproxy process
>>>> didn't get higher than 20% CPU and 10% RAM.
>>>>
>>>> Please help, thanks!
>>>>
>>>
>>>
>>
>

Reply via email to