Can anyone explain to me what the openindiana-1 filesystem is all about?  I 
thought it was the "backup" copy of the openindiana filesystem, when you apply 
OS updates, but that doesn't seem to be the case...

I have time-slider enabled for rpool/ROOT/openindiana.  It has a daily snapshot 
(amongst others).  But every day when the new daily snap is taken, the old 
daily snap rotates into the rpool/ROOT/openindiana-1 filesystem.  This is 
messing up my cron-scheduled "zfs send" script - which detects that the 
rpool/ROOT/openindiana filesystem no longer has the old daily snapshot, and 
therefore has no snapshot in common with the receiving system, and therefore 
sends a new full backup every night.

To make matters more confusing, when I run "mount" and when I zfs get all | 
grep -i mount, I see / on rpool/ROOT/openindiana-1

It would seem, I shouldn't be backing up openindiana, but instead, backup 
openindiana-1?  I would have sworn, out-of-the-box, there was no openindiana-1. 
 Am I simply wrong?

My expectation is that rpool/ROOT/openindiana should have lots of snaps 
available...  3 frequent: one every 15 mins, 23 hourly: one every hour, 6 
daily: one every day, 4 weekly: one every 7 days, etc.

I checked to ensure auto-snapshot service is enabled.  I checked svccfg to 
ensure I understood the correct interval, keep, and period (as described above.)

I have the expected behavior (as I described, the expected behavior according 
to my expectations) on rpool/export/home/eharvey...  But the behavior is 
different on rpool/ROOT/openindiana, even though, as far as I can tell, I have 
the same settings for both.  That is, simply, com.sun:auto-snapshot=true

One more comment - I recall, when I first configured time-slider, they have a 
threshold, default 80% pool used before they automatically bump off old 
snapshots (or stop taking new snaps, I'm not sure what the behavior is).  I 
don't see that setting anywhere I look, using svccfg or zfs get.

My pools are pretty much empty right now.  Nowhere near the 80% limit.
zfs-discuss mailing list

Reply via email to