geom_multipath and zfs doesn't work
Nagy, Attila
bra at fsn.hu
Thu Dec 18 17:12:36 UTC 2014
Hi,
Running stable/10 at r273159 on FC disks (through isp) I can't create a zfs
pool.
What I have is a simple device, accessible on /dev/multipath/sas0,
backed by da0 and da4:
# gmultipath status
Name Status Components
multipath/sas0 OPTIMAL da0 (ACTIVE)
da4 (READ)
When I issue a
zpool create data /dev/multipath/sas0
command, zpool starts to eat 100% CPU:
# procstat -k 3924
PID TID COMM TDNAME KSTACK
3924 100128 zpool - <running>
gstat shows that there is one uncompleted read IO on multipath/sas0 and
the queue length constantly grows on da0:
# gstat -b
dT: 1.030s w: 1.000s
L(q) ops/s r/s kBps ms/r w/s kBps ms/w %busy Name
124402 146 0 0 0.0 0 0 0.0 100.5 da0
0 0 0 0 0.0 0 0 0.0 0.0 da1
0 0 0 0 0.0 0 0 0.0 0.0 da2
0 0 0 0 0.0 0 0 0.0 0.0 da3
1 0 0 0 0.0 0 0 0.0 0.0 multipath/sas0
0 0 0 0 0.0 0 0 0.0 0.0 da4
0 0 0 0 0.0 0 0 0.0 0.0 da5
0 0 0 0 0.0 0 0 0.0 0.0 da6
0 0 0 0 0.0 0 0 0.0 0.0 da7
0 0 0 0 0.0 0 0 0.0 0.0 da8
0 0 0 0 0.0 0 0 0.0 0.0 da9
0 0 0 0 0.0 0 0 0.0 0.0 multipath/sas1
0 0 0 0 0.0 0 0 0.0 0.0 multipath/sata0
0 0 0 0 0.0 0 0 0.0 0.0 multipath/sata1
0 0 0 0 0.0 0 0 0.0 0.0 md0
0 0 0 0 0.0 0 0 0.0 0.0 md1
I can use these devices finely with dd.
What's going on here?
More information about the freebsd-stable
mailing list