Different size after zfs send receive

kc atgb kisscoolandthegangbang at hotmail.fr
Thu May 18 19:36:57 UTC 2017


Hi,

Some days ago I had a need to backup my current pool and restore it after pool destroy and create. 

The pool in my home server is a raidz1 with 4 disks. To backup this pool I grabbed two 4TB disks (single disk pools) to have a double backup (I have just one
sata port left I can use to plug a disk). 

The whole process of backup and restore went well as I can say. But looking at the size reported by zfs list make me a little bit curious. 

storage/datas/ISO                                                        35420869824  381747995136    35420726976  /datas/ISO
storage/datas/ISO at backup_send                                                 142848             -    35420726976  -
storage/datas/ISO at backup_sync                                                      0             -    35420726976  -

b1/datas/ISO                                                        35439308800  2176300351488    35439210496  /datas/ISO
b1/datas/ISO at backup_send                                                  98304              -    35439210496  -
b1/datas/ISO at backup_sync                                                      0              -    35439210496  -

b2/datas/ISO                                                        35439308800  2176298991616    35439210496  /datas/ISO
b2/datas/ISO at backup_send                                                  98304              -    35439210496  -
b2/datas/ISO at backup_sync                                                      0              -    35439210496  -

storage/datas/ISO                                                        35421024576  381303470016    35420715072  /datas/ISO
storage/datas/ISO at backup_send                                                 142848             -    35420715072  -
storage/datas/ISO at backup_sync                                                  11904             -    35420715072  -


storage/usrobj                                                            5819085888  381747995136     5816276544  legacy
storage/usrobj at create                                                         166656             -         214272  -
storage/usrobj at backup_send                                                   2642688             -     5816228928  -
storage/usrobj at backup_sync                                                         0             -     5816276544  -

b1/usrobj                                                            5675081728  2176300351488     5673222144  legacy
b1/usrobj at create                                                         114688              -         147456  -
b1/usrobj at backup_send                                                   1744896              -     5673222144  -
b1/usrobj at backup_sync                                                         0              -     5673222144  -

b2/usrobj                                                            5675188224  2176298991616     5673328640  legacy
b2/usrobj at create                                                         114688              -         147456  -
b2/usrobj at backup_send                                                   1744896              -     5673328640  -
b2/usrobj at backup_sync                                                         0              -     5673328640  -

storage/usrobj                                                            5820359616  381303470016     5815098048  legacy
storage/usrobj at create                                                         166656             -         214272  -
storage/usrobj at backup_send                                                   2535552             -     5815098048  -
storage/usrobj at backup_sync                                                     11904             -     5815098048  -

As you can see the numbers are different for each pool (the initial raidz1, backup1 disk, backup2 disk and new raidz1). I mean in the USED column. I have
nearly all my datasets in the same situation (those with fixed data that have not changed between the beginning of the process and now). backup1 and backup2
are identical disks with exactly the same configurations and have different numbers. I used the same commands for all my transfers except the name of the
destination pool. 

So, I wonder what can cause these differences ? Is it something I have to worry about ? Can I consider this as a normal behavior ? 

Thanks for your enlightments,
K.


More information about the freebsd-fs mailing list