Chris Mason wrote:
On Fri, 2009-03-13 at 17:52 -0500, Steven Pratt wrote:
Chris Mason wrote:
Hello everyone,

I've rebased the experimental branch to include most of the
optimizations I've been working on.

The two major changes are doing all extent tree operations in delayed
processing queues and removing many of the blocking points with btree
locks held.

In addition to smoothing out IO performance, these changes really cut
down on the amount of stack btrfs is using, which is especially
important for kernels with 4k stacks enabled (fedora).

Well, no drastic changes. On Raid, creates got better, but random write got worse. Mail server was mixed. For single disk, pretty much the same story, although CPU savings is noticeable on write, although at the expense of performance.


Thanks for running this, but the main performance fixes for your test
are still in testing locally.  One thing that makes a huge difference on
the random write run is to mount -o ssd.

Tried a run with -o ssd on the raid system. It made some minor improvements in random write performance. Helps more on odirect, but mainly at the 16thread count. Single and 128 threads it doesn't make much difference.

Results syncing now to history boxacle
http://btrfs.boxacle.net/repository/raid/history/History.html

Steve

-chris


--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to