Just an observation: panic occurs in avl_add when called from
find_ds_by_guid that tries to add existing snapshot id to the avl tree
(http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/fs/zfs/dmu_send.c#find_ds_by_guid).

HTH,
Andrey




On Tue, Feb 9, 2010 at 1:37 AM, Bruno Damour <br...@ruomad.net> wrote:
> On 02/ 8/10 06:38 PM, Lori Alt wrote:
>
> Can you please send a complete list of the actions taken:  The commands you
> used to create the send stream, the commands used to receive the stream.
> Also the output of `zfs list -t all` on both the sending and receiving
> sides.  If you were able to collect a core dump (it should be in
> /var/crash/<hostname>), it would be good to upload it.
>
> The panic you're seeing is in the code that is specific to receiving a
> dedup'ed stream.  It's possible that you could do the migration if you
> turned off dedup (i.e. didn't specify -D) when creating the send stream..
> However, then we wouldn't be able to diagnose and fix what appears to be a
> bug.
>
> The best way to get us the crash dump is to upload it here:
>
> https://supportfiles.sun.com/upload
>
> We need either both vmcore.X and unix.X OR you can just send us vmdump.X.
>
> Sometimes big uploads have mixed results, so if there is a problem some
> helpful hints are
> on
> http://wikis.sun.com/display/supportfiles/Sun+Support+Files+-+Help+and+Users+Guide,
> specifically in section 7.
>
> It's best to include your name or your initials or something in the name of
> the file you upload.  As
> you might imagine we get a lot of files uploaded named vmcore.1
>
> You might also create a defect report at http://defect.opensolaris.org/bz/
>
> Lori
>
>
> On 02/08/10 09:41, Bruno Damour wrote:
>
> <copied from opensolaris-dicuss as this probably belongs here.>
>
> I kept on trying to migrate my pool with children (see previous threads) and
> had the (bad) idea to try the -d option on the receive part.
> The system reboots immediately.
>
> Here is the log in /var/adm/messages
>
> Feb 8 16:07:09 amber unix: [ID 836849 kern.notice]
> Feb 8 16:07:09 amber ^Mpanic[cpu1]/thread=ffffff014ba86e40:
> Feb 8 16:07:09 amber genunix: [ID 169834 kern.notice] avl_find() succeeded
> inside avl_add()
> Feb 8 16:07:09 amber unix: [ID 100000 kern.notice]
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4660
> genunix:avl_add+59 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46c0
> zfs:find_ds_by_guid+b9 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46f0
> zfs:findfunc+23 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c47d0
> zfs:dmu_objset_find_spa+38c ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4810
> zfs:dmu_objset_find+40 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4a70
> zfs:dmu_recv_stream+448 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4c40
> zfs:zfs_ioc_recv+41d ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4cc0
> zfs:zfsdev_ioctl+175 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d00
> genunix:cdev_ioctl+45 ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d40
> specfs:spec_ioctl+5a ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4dc0
> genunix:fop_ioctl+7b ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4ec0
> genunix:ioctl+18e ()
> Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4f10
> unix:brand_sys_syscall32+1ca ()
> Feb 8 16:07:09 amber unix: [ID 100000 kern.notice]
> Feb 8 16:07:09 amber genunix: [ID 672855 kern.notice] syncing file
> systems...
> Feb 8 16:07:09 amber genunix: [ID 904073 kern.notice] done
> Feb 8 16:07:10 amber genunix: [ID 111219 kern.notice] dumping to
> /dev/zvol/dsk/rpool/dump, offset 65536, content: kernel
> Feb 8 16:07:10 amber ahci: [ID 405573 kern.info] NOTICE: ahci0:
> ahci_tran_reset_dport port 3 reset port
> Feb 8 16:07:35 amber genunix: [ID 100000 kern.notice]
> Feb 8 16:07:35 amber genunix: [ID 665016 kern.notice] ^M100% done: 107693
> pages dumped,
> Feb 8 16:07:35 amber genunix: [ID 851671 kern.notice] dump succeeded
>
>
> Hello,
> I'll try to do my best.
>
> Here are the commands :
>
> amber ~ # zfs unmount data
> amber ~ # zfs snapshot -r d...@prededup
> amber ~ # zpool destroy ezdata
> amber ~ # zpool create ezdata c6t1d0
> amber ~ # zfs set dedup=on ezdata
> amber ~ # zfs set compress=on ezdata
> amber ~ # zfs send -RD d...@prededup |zfs receive ezdata/data
> cannot receive new filesystem stream: destination 'ezdata/data' exists
> must specify -F to overwrite it
> amber ~ # zpool destroy ezdata
> amber ~ # zpool create ezdata c6t1d0
> amber ~ # zfs set compression=on ezdata
> amber ~ # zfs set dedup=on ezdata
> amber ~ # zfs send -RD d...@prededup |zfs receive -F ezdata/data
> cannot receive new filesystem stream: destination has snapshots (eg.
> ezdata/d...@prededup)
> must destroy them to overwrite it
>
> Each time the send/receive command took some hours and transferred 151G data
> before issuing the message
>
> amber ~ # zfs list
> NAME                         USED  AVAIL  REFER  MOUNTPOINT
> data                         295G   621G   161G  /data
> data/archive                 134G   621G  67.7G  /data/archive
> data/archive/scanrisk       9.45G   621G  8.03G  /data/archive/scanrisk
> data/archive/slbp_g         56.6G   621G  13.5G  /data/archive/slbp_g
> data/cyrus23                 219M   621G   219M  /data/cyrus23
> data/postgres84_64           373M   621G   199M  /data/postgres84_64
> data/postgres84_64/8k        174M   621G   174M  /data/postgres84_64/8k
> ezdata                       151G   144G    21K  /ezdata
> ezdata/data                  151G   144G   151G  /ezdata/data
> rpool                       16.2G  98.0G    87K  /rpool
> ...
>
> And the complete :
>
> amber ~ # zfs list -t all
> NAME                            USED  AVAIL  REFER  MOUNTPOINT
> data                            295G   621G   161G  /data
> data/archive                    134G   621G  67.7G  /data/archive
> data/arch...@20090521          32.9M      -  36.4G  -
> data/archive/scanrisk          9.45G   621G  8.03G  /data/archive/scanrisk
> data/archive/scanr...@2008pre  1.42G      -  9.45G  -
> data/archive/slbp_g            56.6G   621G  13.5G  /data/archive/slbp_g
> data/archive/slb...@20081129   2.35G      -  9.02G  -
> data/archive/slb...@20081212   2.33G      -  9.70G  -
> data/archive/slb...@20090110   9.98M      -  9.34G  -
> data/archive/slb...@20090521   15.3M      -  9.35G  -
> data/archive/slb...@20090702   1.58G      -  14.1G  -
> data/archive/slb...@20090809     67K      -  18.4G  -
> data/archive/slb...@20090912     68K      -  18.4G  -
> data/archive/slb...@20090915   4.25G      -  22.1G  -
> data/archive/slb...@20091128   97.7M      -  19.0G  -
> data/archive/slb...@20091130    438M      -  19.2G  -
> data/cyrus23                    219M   621G   219M  /data/cyrus23
> data/postgres84_64              373M   621G   199M  /data/postgres84_64
> data/postgres84_64/8k           174M   621G   174M  /data/postgres84_64/8k
> ezdata                          151G   144G    24K  /ezdata
> ezd...@now                       29K      -    31K  -
> ezdata/data                     151G   144G   151G  /ezdata/data
> ezdata/d...@prededup             46K      -   151G  -
> ezdata/test                      31K   144G    31K  /ezdata/testamber ~ #
> ezdata/t...@now                    0      -    31K  -
> rpool                          16.3G  97.8G    87K  /rpool
> rpool/ROOT                     8.29G  97.8G    21K  legacy
> rpool/ROOT/snv_132             8.29G  97.8G  7.19G  /
> rpool/ROOT/snv_...@install     1.10G      -  3.24G  -
> rpool/dump                     4.00G  97.8G  4.00G  -
> rpool/export                   49.9M  97.8G    23K  /export
> rpool/export/home              49.8M  97.8G    23K  /export/home
> rpool/export/home/admin        49.8M  97.8G  49.8M  /export/home/admin
> rpool/swap                     4.00G   102G   109M  -
> tank                            292G   165G   166G  /tank
> t...@20090517                  6.75G      -  63.6G  -
> tank/corwin.raw                  15G   166G  14.2G  -
> tank/dara.raw                    15G   168G  12.2G  -
> tank/deirdre.raw                 15G   172G  8.36G  -
> tank/fiona.raw                   20G   179G  6.49G  -
> tank/oberon.raw                  15G   180G    16K  -
> tank/rebma.raw                 22.9G   173G  7.93G  -
> tank/rebma....@20100202        7.34G      -  7.93G  -
> tank/soas.raw                    15G   180G   494M  -
> tank/test                        79K   165G    31K  /tank/test
> tank/t...@now                    18K      -    31K  -
> tank/test/child                  30K   165G    30K  /tank/test/child
> tank/test/ch...@now                0      -    30K  -
> tank/zones                      994M   165G    36K  /tank/zones
> tank/zones/avalon               994M   165G    24K  /tank/zones/avalon
> tank/zones/avalon/ROOT          994M   165G    21K  legacy
> tank/zones/avalon/ROOT/zbe      994M   165G   994M  legacy
>
> I will upload the core as vmdump.amber.0.7z
> Good luck
>
> Bruno
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to