On 02/ 8/10 06:38 PM, Lori Alt wrote:

Can you please send a complete list of the actions taken: The commands you used to create the send stream, the commands used to receive the stream. Also the output of `zfs list -t all` on both the sending and receiving sides. If you were able to collect a core dump (it should be in /var/crash/<hostname>), it would be good to upload it.

The panic you're seeing is in the code that is specific to receiving a dedup'ed stream. It's possible that you could do the migration if you turned off dedup (i.e. didn't specify -D) when creating the send stream.. However, then we wouldn't be able to diagnose and fix what appears to be a bug.

The best way to get us the crash dump is to upload it here:

https://supportfiles.sun.com/upload

We need either both vmcore.X and unix.X OR you can just send us vmdump.X.

Sometimes big uploads have mixed results, so if there is a problem some helpful hints are on http://wikis.sun.com/display/supportfiles/Sun+Support+Files+-+Help+and+Users+Guide,
specifically in section 7.

It's best to include your name or your initials or something in the name of the file you upload. As
you might imagine we get a lot of files uploaded named vmcore.1

You might also create a defect report at http://defect.opensolaris.org/bz/

Lori


On 02/08/10 09:41, Bruno Damour wrote:
<copied from opensolaris-dicuss as this probably belongs here.>

I kept on trying to migrate my pool with children (see previous threads) and had the (bad) idea to try the -d option on the receive part.
The system reboots immediately.

Here is the log in /var/adm/messages

Feb 8 16:07:09 amber unix: [ID 836849 kern.notice]
Feb 8 16:07:09 amber ^Mpanic[cpu1]/thread=ffffff014ba86e40:
Feb 8 16:07:09 amber genunix: [ID 169834 kern.notice] avl_find() succeeded inside avl_add()
Feb 8 16:07:09 amber unix: [ID 100000 kern.notice]
Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4660 genunix:avl_add+59 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46c0 zfs:find_ds_by_guid+b9 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46f0 zfs:findfunc+23 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c47d0 zfs:dmu_objset_find_spa+38c () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4810 zfs:dmu_objset_find+40 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4a70 zfs:dmu_recv_stream+448 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4c40 zfs:zfs_ioc_recv+41d () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4cc0 zfs:zfsdev_ioctl+175 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d00 genunix:cdev_ioctl+45 () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d40 specfs:spec_ioctl+5a () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4dc0 genunix:fop_ioctl+7b () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4ec0 genunix:ioctl+18e () Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4f10 unix:brand_sys_syscall32+1ca ()
Feb 8 16:07:09 amber unix: [ID 100000 kern.notice]
Feb 8 16:07:09 amber genunix: [ID 672855 kern.notice] syncing file systems...
Feb 8 16:07:09 amber genunix: [ID 904073 kern.notice] done
Feb 8 16:07:10 amber genunix: [ID 111219 kern.notice] dumping to /dev/zvol/dsk/rpool/dump, offset 65536, content: kernel Feb 8 16:07:10 amber ahci: [ID 405573 kern.info] NOTICE: ahci0: ahci_tran_reset_dport port 3 reset port
Feb 8 16:07:35 amber genunix: [ID 100000 kern.notice]
Feb 8 16:07:35 amber genunix: [ID 665016 kern.notice] ^M100% done: 107693 pages dumped,
Feb 8 16:07:35 amber genunix: [ID 851671 kern.notice] dump succeeded

Hello,
I'll try to do my best.

Here are the commands :

   amber ~ # zfs unmount data
   amber ~ # zfs snapshot -r d...@prededup
   amber ~ # zpool destroy ezdata
   amber ~ # zpool create ezdata c6t1d0
   amber ~ # zfs set dedup=on ezdata
   amber ~ # zfs set compress=on ezdata
   amber ~ # zfs send -RD d...@prededup |zfs receive ezdata/data
   cannot receive new filesystem stream: destination 'ezdata/data' exists
   must specify -F to overwrite it
   amber ~ # zpool destroy ezdata
   amber ~ # zpool create ezdata c6t1d0
   amber ~ # zfs set compression=on ezdata
   amber ~ # zfs set dedup=on ezdata
   amber ~ # zfs send -RD d...@prededup |zfs receive -F ezdata/data
   cannot receive new filesystem stream: destination has snapshots (eg.
   ezdata/d...@prededup)
   must destroy them to overwrite it


Each time the send/receive command took some hours and transferred 151G data before issuing the message

   amber ~ # zfs list
   NAME                         USED  AVAIL  REFER  MOUNTPOINT
   data                         295G   621G   161G  /data
   data/archive                 134G   621G  67.7G  /data/archive
   data/archive/scanrisk       9.45G   621G  8.03G  /data/archive/scanrisk
   data/archive/slbp_g         56.6G   621G  13.5G  /data/archive/slbp_g
   data/cyrus23                 219M   621G   219M  /data/cyrus23
   data/postgres84_64           373M   621G   199M  /data/postgres84_64
   data/postgres84_64/8k        174M   621G   174M  /data/postgres84_64/8k
   ezdata                       151G   144G    21K  /ezdata
   ezdata/data                  151G   144G   151G  /ezdata/data
   rpool                       16.2G  98.0G    87K  /rpool
   ...

And the complete :

   amber ~ # zfs list -t all
   NAME                            USED  AVAIL  REFER  MOUNTPOINT
   data                            295G   621G   161G  /data
   data/archive                    134G   621G  67.7G  /data/archive
   data/arch...@20090521          32.9M      -  36.4G  -
data/archive/scanrisk 9.45G 621G 8.03G /data/archive/scanrisk
   data/archive/scanr...@2008pre  1.42G      -  9.45G  -
   data/archive/slbp_g            56.6G   621G  13.5G  /data/archive/slbp_g
   data/archive/slb...@20081129   2.35G      -  9.02G  -
   data/archive/slb...@20081212   2.33G      -  9.70G  -
   data/archive/slb...@20090110   9.98M      -  9.34G  -
   data/archive/slb...@20090521   15.3M      -  9.35G  -
   data/archive/slb...@20090702   1.58G      -  14.1G  -
   data/archive/slb...@20090809     67K      -  18.4G  -
   data/archive/slb...@20090912     68K      -  18.4G  -
   data/archive/slb...@20090915   4.25G      -  22.1G  -
   data/archive/slb...@20091128   97.7M      -  19.0G  -
   data/archive/slb...@20091130    438M      -  19.2G  -
   data/cyrus23                    219M   621G   219M  /data/cyrus23
   data/postgres84_64              373M   621G   199M  /data/postgres84_64
data/postgres84_64/8k 174M 621G 174M /data/postgres84_64/8k
   ezdata                          151G   144G    24K  /ezdata
   ezd...@now                       29K      -    31K  -
   ezdata/data                     151G   144G   151G  /ezdata/data
   ezdata/d...@prededup             46K      -   151G  -
ezdata/test 31K 144G 31K /ezdata/testamber ~ #
   ezdata/t...@now                    0      -    31K  -
   rpool                          16.3G  97.8G    87K  /rpool
   rpool/ROOT                     8.29G  97.8G    21K  legacy
   rpool/ROOT/snv_132             8.29G  97.8G  7.19G  /
   rpool/ROOT/snv_...@install     1.10G      -  3.24G  -
   rpool/dump                     4.00G  97.8G  4.00G  -
   rpool/export                   49.9M  97.8G    23K  /export
   rpool/export/home              49.8M  97.8G    23K  /export/home
   rpool/export/home/admin        49.8M  97.8G  49.8M  /export/home/admin
   rpool/swap                     4.00G   102G   109M  -
   tank                            292G   165G   166G  /tank
   t...@20090517                  6.75G      -  63.6G  -
   tank/corwin.raw                  15G   166G  14.2G  -
   tank/dara.raw                    15G   168G  12.2G  -
   tank/deirdre.raw                 15G   172G  8.36G  -
   tank/fiona.raw                   20G   179G  6.49G  -
   tank/oberon.raw                  15G   180G    16K  -
   tank/rebma.raw                 22.9G   173G  7.93G  -
   tank/rebma....@20100202        7.34G      -  7.93G  -
   tank/soas.raw                    15G   180G   494M  -
   tank/test                        79K   165G    31K  /tank/test
   tank/t...@now                    18K      -    31K  -
   tank/test/child                  30K   165G    30K  /tank/test/child
   tank/test/ch...@now                0      -    30K  -
   tank/zones                      994M   165G    36K  /tank/zones
   tank/zones/avalon               994M   165G    24K  /tank/zones/avalon
   tank/zones/avalon/ROOT          994M   165G    21K  legacy
   tank/zones/avalon/ROOT/zbe      994M   165G   994M  legacy

I will upload the core as vmdump.amber.0.7z
Good luck

Bruno
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to