Corrupted zpool import -f FAILS state FAULTED
Outback Dingo
outbackdingo at gmail.com
Sat May 11 11:16:46 UTC 2013
On Sat, May 11, 2013 at 4:44 AM, Ronald Klop <ronald-freebsd8 at klop.yi.org>wrote:
> On Fri, 10 May 2013 16:07:36 +0200, Outback Dingo <outbackdingo at gmail.com>
> wrote:
>
> On Fri, May 10, 2013 at 9:45 AM, Volodymyr Kostyrko <c.kworr at gmail.com
>> >wrote:
>>
>> 09.05.2013 15:31, Outback Dingo:
>>>
>>> ok zfsgurus, FreeBSD 9.1-STABLE box zpool import -f reports pool status
>>>
>>>> Faulted, one of more devices contains corrupted data, however its
>>>> showing
>>>> the guid as faulted in the poll, and not the actual disk device
>>>> /dev/daX,
>>>> the pool is a single vdev 24 disk raidz3. Essentially the hardward
>>>> platform
>>>> is a dual node system, with 8 enclosures connected to 24 SAS drives via
>>>> 4
>>>> LSI cards. I am not currently using geom_multipath, but the box is zoned
>>>> so
>>>> that each node can see 50% of the drives,
>>>> in case of Failure, carp kicks in and migrates "zpool import -af" the
>>>> pools
>>>> onto the other node. it seems as though somehow the pool is now seeing
>>>> guid
>>>> and not devices, not sure if they have switched devices ids due to a
>>>> reboot.
>>>>
>>>>
>>> Am not a zfs guru, but I'll try to help.
>>>
>>> Any console log snippets are welcome. What does "showing the guid as
>>> faulted in the pool" looks like.
>>>
>>> What are the guids for all partitions? Do they interlap for different
>>> nodes?
>>>
>>> ZFS recognizes devices by tasting they vdev labels and not by their
>>> logical location and naming. It can safely report any vdev location - but
>>> it requires the same set vdevs to bring pool online.
>>>
>>
>>
>> zdb shows valid data on the drives, no drives have been removed from the
>> box
>> whats confusing is why its using guids and not devices daX is what puzzles
>> me
>> camcontrol devlist and dmesg clearly show the devices are there. The SAS
>> bus is shared
>> so both nodes with 2 LSI controllers each see all drives. We were
>> utilizing
>> a failover script
>> if nodeA dies, carp would kick the script to import the pool to nodeB,
>> both
>> nodes are in the
>> same chassis and see all the enclosures and all the drives
>>
>
> Are the machines configured the same? As in _exactly_ the same. Glabel
> modules, hint files, sysctls, etc.
>
>
yes, both nodes are identical, from sysctl.conf to loader.conf, ive also
noticed that playing around with enclosure zoning on the system i can now
see which strikes me as quite odd..... now im wondering if i have a
controller flaking out. right now according to the zoning, gmultipath
should see 24+ LUNS however it sees nothing.
zpool import -f
pool: backup
id: 8548776274175948174
state: UNAVAIL
status: The pool was last accessed by another system.
action: The pool cannot be imported due to damaged devices or data.
see: http://illumos.org/msg/ZFS-8000-EY
config:
backup UNAVAIL insufficient replicas
raidz3-0 UNAVAIL insufficient replicas
da32 ONLINE
da30 ONLINE
da29 ONLINE
da3 ONLINE
da4 ONLINE
da5 ONLINE
da6 ONLINE
da7 ONLINE
da8 ONLINE
label/big4 ONLINE
18084052867377310822 UNAVAIL cannot open
2641768775090614171 UNAVAIL cannot open
8083525846528480855 UNAVAIL cannot open
8200855950201180014 UNAVAIL cannot open
da37 ONLINE
da11 ONLINE
4678398398699137944 UNAVAIL cannot open
18315550984013241979 UNAVAIL cannot open
da22 ONLINE
da23 ONLINE
label/backup ONLINE
da25 ONLINE
da26 ONLINE
da27 ONLINE
> Ronald.
>
> ______________________________**_________________
> freebsd-fs at freebsd.org mailing list
> http://lists.freebsd.org/**mailman/listinfo/freebsd-fs<http://lists.freebsd.org/mailman/listinfo/freebsd-fs>
> To unsubscribe, send any mail to "freebsd-fs-unsubscribe@**freebsd.org<freebsd-fs-unsubscribe at freebsd.org>
> "
>
More information about the freebsd-fs
mailing list