On Nov 19, 2007 9:41 AM, Roch - PAE <[EMAIL PROTECTED]> wrote:
>
> Neil Perrin writes:
>  >
>  >
>  > Joe Little wrote:
>  > > On Nov 16, 2007 9:13 PM, Neil Perrin <[EMAIL PROTECTED]> wrote:
>  > >> Joe,
>  > >>
>  > >> I don't think adding a slog helped in this case. In fact I
>  > >> believe it made performance worse. Previously the ZIL would be
>  > >> spread out over all devices but now all synchronous traffic
>  > >> is directed at one device (and everything is synchronous in NFS).
>  > >> Mind you 15MB/s seems a bit on the slow side - especially is
>  > >> cache flushing is disabled.
>  > >>
>  > >> It would be interesting to see what all the threads are waiting
>  > >> on. I think the problem maybe that everything is backed
>  > >> up waiting to start a transaction because the txg train is
>  > >> slow due to NFS requiring the ZIL to push everything synchronously.
>  > >>
>  > >
>  > > I agree completely. The log (even though slow) was an attempt to
>  > > isolate writes away from the pool. I guess the question is how to
>  > > provide for async access for NFS. We may have 16, 32 or whatever
>  > > threads, but if a single writer keeps the ZIL pegged and prohibiting
>  > > reads, its all for nought. Is there anyway to tune/configure the
>  > > ZFS/NFS combination to balance reads/writes to not starve one for the
>  > > other. Its either feast or famine or so tests have shown.
>  >
>  > No there's no way currently to give reads preference over writes.
>  > All transactions get equal priority to enter a transaction group.
>  > Three txgs can be outstanding as we use a 3 phase commit model:
>  > open; quiescing; and syncing.
>
> That makes me wonder if this is not just the lack of write
> throttling issue. If one txg is syncing and the other is
> quiesced out, I think it means we have let in too many
> writes. We do need a better balance.
>
> Neil is  it correct that  reads never hit txg_wait_open(), but
> they just need an I/O scheduler slot ?
>
> If so seems to me just a matter of
>
>         6429205 each zpool needs to monitor it's throughput and throttle 
> heavy writers
>
> However, if this is it, disabling the zil would not solve the
> issue (it might even make it worse). So I am lost as to
> what could be blocking the reads other than lack of I/O
> slots. As another way to improve I/O scheduler we have :
>
>
>         6471212 need reserved I/O scheduler slots to improve I/O latency of 
> critical ops
>
>

So, when are each of these solutions hitting or expected to hit Nevada?

I think the second is the winner, as I see in the iostat either one or
the other getting i/o at a time, and its just feels like under duress
the system simply doesn't have balance anymore between reads/writes.
Its like a smooth moving road with traffic in both directions. Then
the construction crew arrives with flag men who only allow one
direction at a time to go.


>
> -r
>
>  >
>  > Neil.
>  >
>  > _______________________________________________
>  > zfs-discuss mailing list
>  > zfs-discuss@opensolaris.org
>  > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to