Marcus,

Holy moly, that is beautiful.

So glad to understand better what's in the box.

Also please note that I'm not trying to suggest to implement lots of crap, am perfectly clear that high security is correlated with low complexity.


On 2016-02-21 00:29, Marcus MERIGHI wrote:
ti...@openmailbox.org (Tinker), 2016.02.20 (Sat) 16:43 (CET):
..
You appear to mean bioctl(8). Thats the only place I could find the word
'patrol'. bioctl(8) can control more than softraid(4) devices.

bio(4):
     The following device drivers register with bio for volume
        management:

           ami(4)         American Megatrends Inc. MegaRAID
                          PATA/SATA/SCSI RAID controller
           arc(4)         Areca Technology Corporation SAS/SATA RAID
                          controller
           cac(4)         Compaq Smart Array 2/3/4 SCSI RAID controller
           ciss(4)        Compaq Smart Array SAS/SATA/SCSI RAID
                                controller
           ips(4)         IBM SATA/SCSI ServeRAID controller
           mfi(4)         LSI Logic & Dell MegaRAID SAS RAID controller
mpi(4) LSI Logic Fusion-MPT Message Passing Interface mpii(4) LSI Logic Fusion-MPT Message Passing Interface
                                II
           softraid(4)    Software RAID

It is talking about controlling a HW raid controller, in that 'patrol'
paragraph, isn't it?

So by this you mean that patrolling is really implemented for softraid?? (Karel and Constantine don't agree??)

So I just do.. "bioctl -t start sdX" wher sdX is the name of my softraid device, and it'll do the "scrub" as in reading through all underlying physical media to check its internal integrity so for RAID1C that will be data readability and that checksums are correct, and "doas bioctl softraid0" will show me the % status, and if I don't get any errors before it goes back to normal it means the patrol was successful right?

(And as usual patrol is implemented to have the lowest priority, so it should not interfere extreemely much with ordinary SSD softraid operation.)

* Rebuild - I think I saw some console dump of the status of a rebuild
process on the net, so MAYBE or NO..?

That's what it looks like:

$ doas bioctl softraid0
Volume      Status               Size Device
softraid0 0 Rebuild    12002360033280 sd6     RAID5 35% done
          0 Rebuild     4000786726912 0:0.0   noencl <sd2a>
          1 Online      4000786726912 0:1.0   noencl <sd3a>
          2 Online      4000786726912 0:2.0   noencl <sd4a>
          3 Online      4000786726912 0:3.0   noencl <sd5a>

Yey!!

Wait, can you explain to me what I would write instead of "device" and "channel:target[.lun]" in "bioctl -R device" and "bioctl -R channel:target[.lun]", AND what effect those would have?

Say that my sd0 and sd1 SSD:s run a RAID1C already, can I then make softraid extend my RAID1C with my sd2 SSD by "rebuilding" it, as a way to live-copy in all my data to sd2, so this would work as a kind of live attach even if expensive?

Does it work for a softraid that's live already?

* Hotspare - MAYBE, "man softraid" says "Currently there is no automated
mechanism to recover from failed disks.", but that is not so specific
wording, and I think I read a hint somewhere that there is hotspare
functionality.

bioctl(8)
     -H channel:target[.lun]
             If the device at channel:target[.lun] is currently marked
             ``Unused'', promote it to being a ``Hot Spare''.

That's the only mention of 'hot spare'. And again talking about
controlling a hardware RAID controller, isn't it?

What is 'not so specific' about 'no' (as in "Currently there is *no*
automated mechanism to recover from failed disks")?

Awesome.

I guess "bioctl softraid0" will list which hotspares there are currently, and that "-d" will drop a hotspare.


The fact that there is hotspare functionality, means that there are cases when softraid will take a disk out of use.

That will be when that disk reports itself as COMPLETELY out of use ALL BY ITSELF, such as self-detaching itself on the level of the SATA controller or reporting failure via some SMART command?

A disk just half-breaking with broken sectors and 99% IO slowdown will not cause it to go offline though so I guess I should buy enterprise drives with IO access time guarantees then.

* Hotswap - MAYBE, this would depend on if there's rebuild. Only disconnect
("bioctl -O" I think; "bioctl -d" is to.. unmount or self-destruct a
softraid?)

bioctl -O should fail the chunk specified, simulating hardware failure.
After this command you have an 'Offline' chunk in the 'bioctl' output.

bioctl -d 'detach', not 'destroy'; just as sdX appears when you assamble
  a softraid volume, this makes it go away. better unmount before...

So "-d" is to take down a whole softraid. "-O" could work to take out a single physical disk but it's unclean.


So then, there is a very unrefined hotswapping functionality in that "-O" can be used to take out a single physical drive, and "-R" (if I understood it correctly above) can be used to plug in a drive.

Preferable would be to "hotswap" the whole softraid by simply taking it offline altogether ("bioctl -d [raid dev]") and then taking it online altogether ("bioctl -c 1 -l [comma-separated devs] softraid0")


The man pages are sometimes over-minimalistic with respect to an individual user who's trying to learn, this is why I'm asking for your clarification.

I am quite sure the man pages are kept as condensed as they are on
purpose.

You can always read mplayer(1) if you want something lengthy ;-)

So your clarifications would still be much appreciated.

Nothing authoritative from me!
I am just trying to flatten your learning curve.

Bye, Marcus


Awesome. Thank you so much!

Reply via email to