On 26/04/2018 10:28, Eugene M. Zheganin wrote: > Hello, > > > I have some active vdev disk members that used to be in pool that clearly have > not beed destroyed properly, so I'm seeing in a "zpool import" output > something > like > > > # zpool import > pool: zroot > id: 14767697319309030904 > state: UNAVAIL > status: The pool was last accessed by another system. > action: The pool cannot be imported due to damaged devices or data. > see: http://illumos.org/msg/ZFS-8000-EY > config: > > zroot UNAVAIL insufficient replicas > mirror-0 UNAVAIL insufficient replicas > 5291726022575795110 UNAVAIL cannot open > 2933754417879630350 UNAVAIL cannot open > > pool: esx > id: 8314148521324214892 > state: UNAVAIL > status: The pool was last accessed by another system. > action: The pool cannot be imported due to damaged devices or data. > see: http://illumos.org/msg/ZFS-8000-EY > config: > > esx UNAVAIL insufficient replicas > mirror-0 UNAVAIL insufficient replicas > 10170732803757341731 UNAVAIL cannot open > 9207269511643803468 UNAVAIL cannot open > > > is there any _safe_ way to get rid of this ? I'm asking because a gptzfsboot > loader in recent -STABLE stumbles upon this and refuses to boot the system > (https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=227772). The workaround is > to > use the 11.1 loader, but I'm afraid this behavior will now be the intended > one.
You can try to use zdb -l to find the stale labels. And then zpool labelclear to clear them. -- Andriy Gapon _______________________________________________ freebsd-stable@freebsd.org mailing list https://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"