gvinum raid10 stale
Dimitri Aivaliotis
aglarond at gmail.com
Thu Dec 18 03:40:55 PST 2008
Hi,
I created a raid10 using gvinum with the following config:
drive a device /dev/da2
drive b device /dev/da3
volume raid10
plex org striped 512k
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
plex org striped 512k
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
sd length 4374m drive b
sd length 4374m drive a
I wanted to add two additional disks to this raid10, so I shutdown the
server, inserted the disks and brought it back up. When the system
booted, it reported the filesystem as needing a check. Doing a gvinum
list, I saw that all subdisks were stale, so both plexes were down.
After rebooting again (to remove the additional disks), the problem
persisted. My assumption that the new disks caused the old subdisks
to be stale wasn't true, as I later noticed that a different server
with the same config has a plex down as well because all subdisks on
that plex are stale. The servers are running 6.3-RELEASE-p1 and
6.2-RELEASE-p9, respectively.
(I wound up doing a 'gvinum setstate -f up raid10.p1.s<num>' 32 times
to bring one plex back up on the server that had both down.)
My questions:
- Why would these subdisks be set stale?
- How can I recover the other plex, such that the data continues to be
striped+mirrored correctly?
- How can I extend this raid10 by adding two additional disks?
These servers are both in production, so I unfortunately can't do
things like move the data, re-create the RAID, and move the data back.
Any help, tips, advice would be greatly appreciated. Below are
messages from dmesg on server1, as well as gvinum list output for
both.
- Dimitri
server1 (6.3-RELEASE-p1)
=======================
dmesg|grep -i geom
GEOM_VINUM: subdisk raid10.p0.s1 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s3 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s5 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s7 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s9 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s11 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s13 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s15 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s17 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s19 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s21 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s23 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s25 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s27 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s29 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s31 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s0 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s2 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s4 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s6 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s8 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s10 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s12 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s14 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s16 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s18 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s20 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s22 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s24 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s26 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s28 state change: down -> stale
GEOM_VINUM: subdisk raid10.p0.s30 state change: down -> stale
(So, I figured it would be best to bring plex 1 up, and executed the
above-mentioned setstate commands.)
GEOM_VINUM: plex raid10.p1 state change: down -> up
gvinum list
-------------
2 drives:
D a State: up /dev/da2 A: 10/139978 MB (0%)
D b State: up /dev/da3 A: 10/139978 MB (0%)
1 volume:
V raid10 State: up Plexes: 2 Size: 136 GB
2 plexes:
P raid10.p0 S State: down Subdisks: 32 Size: 136 GB
P raid10.p1 S State: up Subdisks: 32 Size: 136 GB
64 subdisks:
S raid10.p0.s0 State: stale D: a Size: 4374 MB
S raid10.p0.s1 State: stale D: b Size: 4374 MB
S raid10.p0.s2 State: stale D: a Size: 4374 MB
S raid10.p0.s3 State: stale D: b Size: 4374 MB
S raid10.p0.s4 State: stale D: a Size: 4374 MB
S raid10.p0.s5 State: stale D: b Size: 4374 MB
S raid10.p0.s6 State: stale D: a Size: 4374 MB
S raid10.p0.s7 State: stale D: b Size: 4374 MB
S raid10.p0.s8 State: stale D: a Size: 4374 MB
S raid10.p0.s9 State: stale D: b Size: 4374 MB
S raid10.p0.s10 State: stale D: a Size: 4374 MB
S raid10.p0.s11 State: stale D: b Size: 4374 MB
S raid10.p0.s12 State: stale D: a Size: 4374 MB
S raid10.p0.s13 State: stale D: b Size: 4374 MB
S raid10.p0.s14 State: stale D: a Size: 4374 MB
S raid10.p0.s15 State: stale D: b Size: 4374 MB
S raid10.p0.s16 State: stale D: a Size: 4374 MB
S raid10.p0.s17 State: stale D: b Size: 4374 MB
S raid10.p0.s18 State: stale D: a Size: 4374 MB
S raid10.p0.s19 State: stale D: b Size: 4374 MB
S raid10.p0.s20 State: stale D: a Size: 4374 MB
S raid10.p0.s21 State: stale D: b Size: 4374 MB
S raid10.p0.s22 State: stale D: a Size: 4374 MB
S raid10.p0.s23 State: stale D: b Size: 4374 MB
S raid10.p0.s24 State: stale D: a Size: 4374 MB
S raid10.p0.s25 State: stale D: b Size: 4374 MB
S raid10.p0.s26 State: stale D: a Size: 4374 MB
S raid10.p0.s27 State: stale D: b Size: 4374 MB
S raid10.p0.s28 State: stale D: a Size: 4374 MB
S raid10.p0.s29 State: stale D: b Size: 4374 MB
S raid10.p0.s30 State: stale D: a Size: 4374 MB
S raid10.p0.s31 State: stale D: b Size: 4374 MB
S raid10.p1.s0 State: up D: b Size: 4374 MB
S raid10.p1.s1 State: up D: a Size: 4374 MB
S raid10.p1.s2 State: up D: b Size: 4374 MB
S raid10.p1.s3 State: up D: a Size: 4374 MB
S raid10.p1.s4 State: up D: b Size: 4374 MB
S raid10.p1.s5 State: up D: a Size: 4374 MB
S raid10.p1.s6 State: up D: b Size: 4374 MB
S raid10.p1.s7 State: up D: a Size: 4374 MB
S raid10.p1.s8 State: up D: b Size: 4374 MB
S raid10.p1.s9 State: up D: a Size: 4374 MB
S raid10.p1.s10 State: up D: b Size: 4374 MB
S raid10.p1.s11 State: up D: a Size: 4374 MB
S raid10.p1.s12 State: up D: b Size: 4374 MB
S raid10.p1.s13 State: up D: a Size: 4374 MB
S raid10.p1.s14 State: up D: b Size: 4374 MB
S raid10.p1.s15 State: up D: a Size: 4374 MB
S raid10.p1.s16 State: up D: b Size: 4374 MB
S raid10.p1.s17 State: up D: a Size: 4374 MB
S raid10.p1.s18 State: up D: b Size: 4374 MB
S raid10.p1.s19 State: up D: a Size: 4374 MB
S raid10.p1.s20 State: up D: b Size: 4374 MB
S raid10.p1.s21 State: up D: a Size: 4374 MB
S raid10.p1.s22 State: up D: b Size: 4374 MB
S raid10.p1.s23 State: up D: a Size: 4374 MB
S raid10.p1.s24 State: up D: b Size: 4374 MB
S raid10.p1.s25 State: up D: a Size: 4374 MB
S raid10.p1.s26 State: up D: b Size: 4374 MB
S raid10.p1.s27 State: up D: a Size: 4374 MB
S raid10.p1.s28 State: up D: b Size: 4374 MB
S raid10.p1.s29 State: up D: a Size: 4374 MB
S raid10.p1.s30 State: up D: b Size: 4374 MB
S raid10.p1.s31 State: up D: a Size: 4374 MB
server2 (6.2-RELEASE-p9)
=======================
(no clues in the logs as to why the subdisks are stale)
gvinum list
-------------
2 drives:
D b State: up /dev/da3 A: 10/139978 MB (0%)
D a State: up /dev/da2 A: 10/139978 MB (0%)
1 volume:
V raid10 State: up Plexes: 2 Size: 136 GB
2 plexes:
P raid10.p0 S State: up Subdisks: 32 Size: 136 GB
P raid10.p1 S State: down Subdisks: 32 Size: 136 GB
64 subdisks:
S raid10.p0.s0 State: up D: a Size: 4374 MB
S raid10.p0.s1 State: up D: b Size: 4374 MB
S raid10.p0.s2 State: up D: a Size: 4374 MB
S raid10.p0.s3 State: up D: b Size: 4374 MB
S raid10.p0.s4 State: up D: a Size: 4374 MB
S raid10.p0.s5 State: up D: b Size: 4374 MB
S raid10.p0.s6 State: up D: a Size: 4374 MB
S raid10.p0.s7 State: up D: b Size: 4374 MB
S raid10.p0.s8 State: up D: a Size: 4374 MB
S raid10.p0.s9 State: up D: b Size: 4374 MB
S raid10.p0.s10 State: up D: a Size: 4374 MB
S raid10.p0.s11 State: up D: b Size: 4374 MB
S raid10.p0.s12 State: up D: a Size: 4374 MB
S raid10.p0.s13 State: up D: b Size: 4374 MB
S raid10.p0.s14 State: up D: a Size: 4374 MB
S raid10.p0.s15 State: up D: b Size: 4374 MB
S raid10.p0.s16 State: up D: a Size: 4374 MB
S raid10.p0.s17 State: up D: b Size: 4374 MB
S raid10.p0.s18 State: up D: a Size: 4374 MB
S raid10.p0.s19 State: up D: b Size: 4374 MB
S raid10.p0.s20 State: up D: a Size: 4374 MB
S raid10.p0.s21 State: up D: b Size: 4374 MB
S raid10.p0.s22 State: up D: a Size: 4374 MB
S raid10.p0.s23 State: up D: b Size: 4374 MB
S raid10.p0.s24 State: up D: a Size: 4374 MB
S raid10.p0.s25 State: up D: b Size: 4374 MB
S raid10.p0.s26 State: up D: a Size: 4374 MB
S raid10.p0.s27 State: up D: b Size: 4374 MB
S raid10.p0.s28 State: up D: a Size: 4374 MB
S raid10.p0.s29 State: up D: b Size: 4374 MB
S raid10.p0.s30 State: up D: a Size: 4374 MB
S raid10.p0.s31 State: up D: b Size: 4374 MB
S raid10.p1.s0 State: stale D: b Size: 4374 MB
S raid10.p1.s1 State: stale D: a Size: 4374 MB
S raid10.p1.s2 State: stale D: b Size: 4374 MB
S raid10.p1.s3 State: stale D: a Size: 4374 MB
S raid10.p1.s4 State: stale D: b Size: 4374 MB
S raid10.p1.s5 State: stale D: a Size: 4374 MB
S raid10.p1.s6 State: stale D: b Size: 4374 MB
S raid10.p1.s7 State: stale D: a Size: 4374 MB
S raid10.p1.s8 State: stale D: b Size: 4374 MB
S raid10.p1.s9 State: stale D: a Size: 4374 MB
S raid10.p1.s10 State: stale D: b Size: 4374 MB
S raid10.p1.s11 State: stale D: a Size: 4374 MB
S raid10.p1.s12 State: stale D: b Size: 4374 MB
S raid10.p1.s13 State: stale D: a Size: 4374 MB
S raid10.p1.s14 State: stale D: b Size: 4374 MB
S raid10.p1.s15 State: stale D: a Size: 4374 MB
S raid10.p1.s16 State: stale D: b Size: 4374 MB
S raid10.p1.s17 State: stale D: a Size: 4374 MB
S raid10.p1.s18 State: stale D: b Size: 4374 MB
S raid10.p1.s19 State: stale D: a Size: 4374 MB
S raid10.p1.s20 State: stale D: b Size: 4374 MB
S raid10.p1.s21 State: stale D: a Size: 4374 MB
S raid10.p1.s22 State: stale D: b Size: 4374 MB
S raid10.p1.s23 State: stale D: a Size: 4374 MB
S raid10.p1.s24 State: stale D: b Size: 4374 MB
S raid10.p1.s25 State: stale D: a Size: 4374 MB
S raid10.p1.s26 State: stale D: b Size: 4374 MB
S raid10.p1.s27 State: stale D: a Size: 4374 MB
S raid10.p1.s28 State: stale D: b Size: 4374 MB
S raid10.p1.s29 State: stale D: a Size: 4374 MB
S raid10.p1.s30 State: stale D: b Size: 4374 MB
S raid10.p1.s31 State: stale D: a Size: 4374 MB
More information about the freebsd-geom
mailing list