zfs import panic
biancalana at gmail.com
Tue Nov 13 14:36:44 PST 2007
On Nov 13, 2007 12:58 PM, Pawel Jakub Dawidek <pjd at freebsd.org> wrote:
> On Tue, Nov 13, 2007 at 03:55:06PM +0100, Pawel Jakub Dawidek wrote:
> > On Mon, Nov 12, 2007 at 06:52:15PM -0200, Alexandre Biancalana wrote:
> > > Hi list,
> > >
> > > After an spontaneous reboot this weekend our backup server is
> > > panic'ing on zfs import. I've the vmcore file but I can't debug this
> > > with kgdb
> > >
> > > :/var/crash # more info.3
> > > Dump header from device /dev/ad0s1b
> > > Architecture: amd64
> > > Architecture Version: 2
> > > Dump Length: 103477248B (98 MB)
> > > Blocksize: 512
> > > Dumptime: Mon Nov 12 14:56:12 2007
> > > Hostname:
> > > Magic: FreeBSD Kernel Dump
> > > Version String: FreeBSD 7.0-BETA2 #0: Mon Nov 12 11:49:07 BRST 2007
> > > root@:/usr/src/sys/amd64/compile/MANNY.debug
> > > Panic String: solaris assert: ss == NULL, file:
> > > /usr/src/sys/modules/zfs/../../contrib/opensolaris/uts/common/fs/zfs/space_map.c,
> > > line: 110
> > > Dump Parity: 2217569595
> > > Bounds: 3
> > > Dump Status: good
> > This looks like http://bugs.opensolaris.org/view_bug.do?bug_id=6458218,
> > but fix for this bug is already in FreeBSD's version...
> > To recover your data you can try removing /boot/zfs/zpool.cache file
> > before loading zfs.ko module, and then try to import your pool with:
> > # zpool import -o ro <pool_name>
> BTW. You could try booting from Indiana CD (Solaris) and see if you can
> import the pool from there.
Bad lucky.... OpenSolaris does not recognize my Areca ARC-1231 SATA
googling around I find some reports of corrupted zpools, pointing to
bug_id 6458218 (http://www.opensolaris.org/jive/thread.jspa?messageID=105474𙰂)
and the proposed solutions was update the system (solaris) to build 60
or later, and put 'set zfs:zfs_recover=1' in /etc/system.
We have a vfs.zfs.recover then I tried to set this to 1 on the
/boot/loader.conf, boot the machine and try to import the pool again,
but I have no success the panic is triggered in the same way...
When using zfs at this server, I expected some panics but not loose data :-(
Any other hint ?
Thank you for your time.
More information about the freebsd-current