tj wrote:
Bill Davidsen wrote:
tj wrote:
Thomas Jager wrote:
Hi list.

I run a file server on MD raid-5.
If a client reads one big file and at the same time another client tries to write a file, the thread writing just sits in uninterruptible sleep until the reader has finished. Only very small amount of writes get trough while the reader is still working.
I'm having some trouble pinpointing the problem.
It's not consistent either sometimes it works as expected both the reader and writer gets some transactions. On huge reads I've seen the writer blocked for 30-40 minutes without any significant writes happening (Maybe a few megabytes, of several gigs waiting). It happens with NFS, SMB and FTP, and local with dd. And seems to be connected to raid-5. This does not happen on block devices without raid-5. I'm also wondering if it can have anything to do with loop-aes? I use loop-aes on top of the md, but then again i have not observed this problem on loop-devices with disk backend. I do know that loop-aes degrades performance but i didn't think it would do something like this?

I've seen this problem in 2.6.16-2.6.21

All disks in the array is connected to a controller with a SiI 3114 chip.

I just noticed something else. A couple of slow readers where running on my raid-5 array. Then i started a copy from another local disk to the array. Then i got the extremely long wait. I noticed something in iostat:

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
          3.90    0.00   48.05   31.93    0.00   16.12

Device:            tps    kB_read/s    kB_wrtn/s    kB_read    kB_wrtn
....
sdg               0.80        25.55         0.00        128          0
sdh             154.89       632.34         0.00       3168          0
sdi               0.20        12.77         0.00         64          0
sdj               0.40        25.55         0.00        128          0
sdk               0.40        25.55         0.00        128          0
sdl               0.80        25.55         0.00        128          0
sdm               0.80        25.55         0.00        128          0
sdn               0.60        23.95         0.00        120          0
md0             199.20       796.81         0.00       3992          0

All disks are member of the same raid array (md0). One of the disks has a ton of transactions compared to the other disks. Read operations as far as i can tell. Why? May be connected with my problem?
Two thoughts on that, if you are doing a lot of directory operations, it's possible that the inodes being used most are all in one chunk.
Hi thanks for the reply.

It's not directory operations AFAIK. Reading a few files (3 in this case) and writing one.

The other possibility is that these a journal writes and reflect updates to the atime. The way to see if this is in some way related is to mount (remount) with noatime: "mount -o remount,noatime /dev/md0 /wherever" and retest. If this is journal activity you can do several things to reduce the problem, which I'll go into (a) if it seems to be the problem, and (b) if someone else doesn't point you to an existing document or old post on the topic. Oh, you could also try mounting the filesystem as etc2, assuming that it's ext3 now. I wouldn't run that way, but it's useful as a diagnostic tool.
I don't use ext3 i use ReiserFS. ( It seemed like a good idea at the time. ) It's mounted with -o noatime. I've done some more testing and i seems like it might be connected to mount --bind. If i write to a binded mount i get the slow writes. But if i write directly to the real mount i don't. It might just be a random occurrence, as the problem always has been inconsistent. Thoughts?

I said I would test, and I did. I don't see a difference with ext3 in reads at all. I don't see a difference in bind vs. direct for write, either, but all of my space large enough to have room for a few GB write had internal bitmaps.

Other info: block size made no consistent difference, changing the stripe_cache_size helped but was very non-linear in effect, and direct raid over partitions had the same performance as LVM over raid on other partitions of the same disks.

Neil: is there a reason (other than ease of coding) why the bitmap isn't distributed to minimize seeks? ie. put the bitmap for given stripes at the end of those strips rather than the end of the space.

I have added to my tests-to-do list partitioning a disk such that I have a small partition and a large, making RAID-10 no bitmap on the small (multiple drive, obviously) and then RAID-5 on the large, with the bitmap for RAID-5 on the RAID-10 raw array. The only reason I have any interest is that I did something like this with JFS, putting the journal on a dedicated partition with a different chunk size, and it really helped. If this gives any useful information I'll report, but I'm building a few analysis tools first, so it will be several weeks.

--
bill davidsen <[EMAIL PROTECTED]>
 CTO TMR Associates, Inc
 Doing interesting things with small computers since 1979

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to