Roman,

Ok, thanks. That's what I was thinking about. I just wan't sure if it's ok to add volumes into existent groups from AVS point of view

Only one question about commands you suggested. Is this about adding a disk queue?
sndradm -g -q a :

Is this kind a compulsory?

It will be for any filesystem like ZFS, QFS, also volume managers, or databases, each of which uses multiple physical disks to represent a single consistent collection of pooled storage. Failure to do this, will allow the replicated volumes to be updated inconsistently.

Grouping two or more replicated volumes into an I/O consistency group, is compulsory if the data on the secondary node is deemed valuable at any point in time.

I mean I don't use disk queues for this particular installation.

Since ZFS does a very good job of distributing I/Os across all constituent vdevs in a single ZFS storage pool, you have been either lucky or fortunate that ZFS has not seen inconsistencies, especially since ZFS is very good at this. Based on this, it is likely that the ZFS storage pool has been idle prior to accessing the replicated ZFS storage pool.

And let me ask you Jim about the famous post "AVS and ZFS, seamless". It was yours, right? Even if not, would you mind commenting it?

It it my posting.

In hindsight, I should have described the "goals" of this write up, as in hindsight it is too complex. The goals of the write up were to take the available 46 disks of a Sun Fire 4500 (Thumper), and replicate them to another SunFire 4500. They complexity came about due to the fact that there were additional requirements for the bitmap volumes.

- On RAID-1, not RAIDz storage, allowing for both redundancy, but not at the cost of a RAIDz write I/O, when SNDR is just flipping one or more bits in a bitmap volume.

- Distributed across as many disks as possible, so that a distributed ZFS write across many vdevs, would also incur a distributed bitmap write across many bitmap volumes. Having a multi-disk ZFS write, resulting in a single disk update of multiple bitmap volumes, would create an I/O bottleneck.

FWIW: This article was written before the general availability of SSDs. Placing bitmap volumes on write SSDs is highly optimal, as there is no seek cost, which is what was trying to be avoided when doing multi-disk updates. Years ago, AVS had a optional SPARC only feature called Fast Write Cache (FWC). It was software that utilized an S-BUS based non-volatile memory module. Due to its high cost and ties to S- BUS architectures, this technology became obsolete and was dropped out of AVS. Bitmaps on SSDs, replace this functionality very well.

The first question is about the issue with RDC timeouts and disks queues that I asked some times ago.
http://www.mail-archive.com/[email protected]/msg05497.html
I wonder if that configuration described  in blog was successful?

I recall that there was a cut-n-paste error, from my notes to the blog, but my notes were based on a working configuration.

How did the link look like?

This was a 1 GigE with less the 5 ms latency between sites. Measured link latency is the #1 issues in data replication, in that the local- to-remote, plus remote-to-local, round trip time is import. SNDR replication is done using RPC over TCP/IP, and then amount of time it takes to move data, even bitmap data, plays a key roles in replication. The RPC timeout value, a feature of RPC, not SNDR, has become a problem as volume sizes, thus bitmap sizes have increased.

Was it something like dedicated circuit with usual latency?

Link latency averaged about 5 ms, with spikes no more than 10 ms. If link latency is higher than this, the RPC timeouts will be common, and problematic.

How did it work out eventually?

Well, it was an experiment, not a production environment, as the equipment was on loan for the purposes of SNDR and ZFS testing.


And the second question about slices for bitmap. What was the reason for putting it not on dedicated pair of disk as documentation suggests?

If ZFS does a 46 volume write I/O, one would not want 46 bitmap updates, to happen to a single disk. The example had mirrored bitmaps volumes for each replica, minimizing a single bitmap as being an I/O bottleneck.


(avs administration guide: raw devices must be stored on a disk separate from the disk that contains the data from the replicated volumes.... The bitmap must not be stored on the same disk as replicated volumes)

From the simplistic point of view of the administration guide, if one had a single data disk and was doing sequential write I/O, these write I/Os, plus bitmap updates, would turn the optimal sequential write, into random write I/Os, if both replica and bitmap were on the same disk. Placing them on separate disks, would allow the sequence write I/ Os and sequential bitmap I/Os, to be sequential I/Os. Ideally, one would want to carry this theme across any sized replication configuration, but maximizing a large ZFS configuration of 46 disks, requires 46 bitmaps, and some tradeoffs where taken.


- Jim Dunham

Roman Naumenko
Frontline Technologies
--
This message posted from opensolaris.org
_______________________________________________
storage-discuss mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/storage-discuss

_______________________________________________
storage-discuss mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/storage-discuss

Reply via email to