mirror vdevs with different sizes
- Reply: Alan Somers : "Re: mirror vdevs with different sizes"
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Fri, 25 Mar 2022 15:16:13 UTC
Hello, I have an existing zpool with 12 mirrors of 8 TB disks. It is currently about 60% full and we expect to fill the remaining space fairly quickly. I would like to expand it, preferably using 12 mirrors of 16 TB disks. Any reason I shouldn't do this? Using plain files created with truncate(1) like these: [root@ibex] # ls -lh /vd/vd* -rw-r--r-- 1 root wheel 8.0G Mar 25 08:49 /vd/vd0 -rw-r--r-- 1 root wheel 8.0G Mar 25 08:49 /vd/vd1 -rw-r--r-- 1 root wheel 16G Mar 25 08:49 /vd/vd2 -rw-r--r-- 1 root wheel 16G Mar 25 08:49 /vd/vd3 I can first do this: [root@ibex] # zpool create ztest mirror /vd/vd{0,1} [root@ibex] # zpool list ztest NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT ztest 7.50G 384K 7.50G - - 0% 0% 1.00x ONLINE - And then do this: [root@ibex] # zpool add ztest mirror /vd/vd{2,3} [root@ibex] # zpool list ztest NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT ztest 23G 528K 23.0G - - 0% 0% 1.00x ONLINE - And FWIW, everything works as expected. But I've never constructed a real zpool with vdevs of different sizes and I don't know whether there might be any expected problems. I could just create a new zpool with new disks, but most of the existing data and most of the expected new data is in just two file systems and for simplicity's sake from the perspective of those users, it would be nicer to just make the existing file systems larger than to give them access to a new, different one. Any comments, suggestions, warnings, etc. much appreciated. Thanks.