quantifying zpool performance with number of vdevs

Graham Allan allan at physics.umn.edu
Fri Jan 29 18:27:58 UTC 2016


In many of the storage systems I built to date I was slightly 
conservative (?) in wanting to keep any one pool confined to a single 
JBOD chassis. In doing this I've generally been using the Supermicro 
45-drive chassis with pools made of 4x (8+2) raidz2, other slots being 
kept for spares, ZIL and L2ARC.

Now I have several servers with 3-4 such chassis, and reliability has 
also been such that I'd feel more comfortable about spanning chassis, if 
there was worthwhile performance benefit.

Obviously theory says that iops should scale with number of vdevs but it 
would be nice to try and quantify.

Getting relevant data out of iperf seems problematic on machines with 
128GB+ RAM - it's hard to blow out the ARC.

It does seem like I get possibly more valid-looking results if I set 
"zfs set primarycache=metadata" on my test dataset - it seems like this 
should mostly disable the ARC (seems to be borne out by arcstat output, 
though there could still be L2ARC effects).

Wonder if anyone has any thoughts on this, and also on benefits/risks of 
moving from 40-drive to 80- or 120-drive pools.

Graham
-- 


More information about the freebsd-fs mailing list