zfs receive stalls whole system

Rainer Duffner rainer at ultra-secure.de
Thu Aug 4 21:51:41 UTC 2016


> Am 26.05.2016 um 12:48 schrieb Fabian Keil <freebsd-listen at fabiankeil.de>:
> 
>> 
>> It can cause deadlocks and poor performance when paging.
>> 
>> This was recently fixed in ElectroBSD and I intend to submit
>> the patch in a couple of days after a bit more stress testing.
> 
> Done: https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=209759 <https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=209759>

I’ve now been able to implement this patch.
It doesn’t seem to do any harm but doesn’t fix the problem, either.


(sorry for the long post)

pool: datapool
state: ONLINE
scan: none requested
config:

NAME                      STATE     READ WRITE CKSUM
datapool                  ONLINE       0     0     0
  raidz2-0                ONLINE       0     0     0
    gpt/S0M1ESLL_C1S03    ONLINE       0     0     0
    gpt/S0M1F8V0_C1S04    ONLINE       0     0     0
    gpt/S0M1EQPR_C1S05    ONLINE       0     0     0
    gpt/S0M19J9D_C1S06    ONLINE       0     0     0
    gpt/S0M1ES7R_C1S07    ONLINE       0     0     0
    gpt/S0M1DXJR_C1S08    ONLINE       0     0     0
  raidz2-1                ONLINE       0     0     0
    gpt/S0M1EQHL_C2S01    ONLINE       0     0     0
    gpt/S0M1EQSL_C2S02    ONLINE       0     0     0
    gpt/S0M1F7CG_C2S03    ONLINE       0     0     0
    gpt/S0M1F2B1_C2S04    ONLINE       0     0     0
    gpt/S0M1ER7Y_C2S05    ONLINE       0     0     0
    gpt/S0M1F9B0_C2S06    ONLINE       0     0     0
  raidz2-2                ONLINE       0     0     0
    gpt/S3L29R3L_EC1_S01  ONLINE       0     0     0
    gpt/S3L29XFQ_EC1_S02  ONLINE       0     0     0
    gpt/S3L29QTK_EC1_S03  ONLINE       0     0     0
    gpt/S3L28ZFA_EC1_S04  ONLINE       0     0     0
    gpt/S3L29PG9_EC1_S05  ONLINE       0     0     0
    gpt/S3L29TAA_EC1_S06  ONLINE       0     0     0
  raidz2-3                ONLINE       0     0     0
    gpt/S3L29RHR_EC1_S07  ONLINE       0     0     0
    gpt/S3L29VQT_EC1_S08  ONLINE       0     0     0
    gpt/S3L2A7WM_EC1_S09  ONLINE       0     0     0
    gpt/S3L29GXN_EC1_S10  ONLINE       0     0     0
    gpt/S3L29TPT_EC1_S11  ONLINE       0     0     0
    gpt/S3L2A4EJ_EC1_S12  ONLINE       0     0     0

errors: No known data errors

pool: zroot
state: ONLINE
scan: none requested
config:

NAME        STATE     READ WRITE CKSUM
zroot       ONLINE       0     0     0
  mirror-0  ONLINE       0     0     0
    da8p3   ONLINE       0     0     0
    da9p3   ONLINE       0     0     0

errors: No known data errors

Machine 1 creates hourly, daily and weekly snapshots (with zfSnap) and sends them hourly to Machine 2.

/usr/local/sbin/zxfer -dF -o sharenfs="-maproot=1003 -network 10.10.91.224 -mask 255.255.255.240" -T root at 10.168.91.231 -R datapool/nfs datapool/backup 

The network is Gbit.

The filesystems aren’t that big (IMO):

NAME       SIZE  ALLOC   FREE  EXPANDSZ   FRAG    CAP  DEDUP  HEALTH  ALTROOT
datapool  19.5T  7.08T  12.4T         -    13%    36%  1.00x  ONLINE  -
zroot      556G  7.78G   548G         -     1%     1%  1.00x  ONLINE  -

Nor are the snapshots really that large:

NAME                                                               USED  AVAIL  REFER  MOUNTPOINT
datapool/nfs/bla1-archives at weekly-2016-07-23_04.44.27--2w          43.3M      -  49.5G  -
datapool/nfs/bla1-archives at daily-2016-07-28_03.35.25--7d           42.0M      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-07-29_03.33.40--7d           42.0M      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-07-30_03.22.18--7d               0      -  49.9G  -
datapool/nfs/bla1-archives at weekly-2016-07-30_04.15.01--2w              0      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-07-31_03.14.47--7d           42.0M      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-08-01_05.03.36--7d           42.0M      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-08-02_05.02.39--7d           42.0M      -  49.9G  -
datapool/nfs/bla1-archives at daily-2016-08-03_03.57.46--7d           42.2M      -  49.9G  -
datapool/nfs/bla1-archives at hourly-2016-08-03_12.04.00--12h             0      -  19.8G  -
datapool/nfs/bla1-archives at hourly-2016-08-03_13.04.00--12h             0      -  19.8G  -
datapool/nfs/bla1-archives at hourly-2016-08-03_14.04.00--12h          192K      -   575K  -
datapool/nfs/bla1-archives at hourly-2016-08-03_15.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_16.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_17.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_18.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_19.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_20.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_21.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_22.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-03_23.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-archives at hourly-2016-08-04_00.04.00--12h             0      -  12.1M  -
datapool/nfs/bla1-documents at weekly-2016-07-23_04.44.27--2w         6.02G      -  4.51T  -
datapool/nfs/bla1-documents at daily-2016-07-28_03.35.25--7d          5.85G      -  4.54T  -
datapool/nfs/bla1-documents at daily-2016-07-29_03.33.40--7d          5.82G      -  4.55T  -
datapool/nfs/bla1-documents at daily-2016-07-30_03.22.18--7d              0      -  4.56T  -
datapool/nfs/bla1-documents at weekly-2016-07-30_04.15.01--2w             0      -  4.56T  -
datapool/nfs/bla1-documents at daily-2016-07-31_03.14.47--7d          5.80G      -  4.56T  -
datapool/nfs/bla1-documents at daily-2016-08-01_05.03.36--7d          5.80G      -  4.56T  -
datapool/nfs/bla1-documents at daily-2016-08-02_05.02.39--7d          5.81G      -  4.56T  -
datapool/nfs/bla1-documents at daily-2016-08-03_03.57.46--7d          70.6M      -  4.56T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_12.04.00--12h        6.85M      -  4.57T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_13.04.00--12h        3.42M      -  4.57T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_14.04.00--12h        9.88M      -  4.57T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_15.04.00--12h        12.6M      -  4.57T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_16.04.00--12h        12.4M      -  4.57T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_17.04.00--12h        11.5M      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_18.04.00--12h        4.64M      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_19.04.00--12h         464K      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_20.04.00--12h         352K      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_21.04.00--12h         384K      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_22.04.00--12h        79.9K      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-03_23.04.00--12h            0      -  4.58T  -
datapool/nfs/bla1-documents at hourly-2016-08-04_00.04.00--12h            0      -  4.58T  -
datapool/nfs/bla1-project_layouts at weekly-2016-07-23_04.44.27--2w    176K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-07-28_03.35.25--7d     144K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-07-29_03.33.40--7d     144K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-07-30_03.22.18--7d        0      -  1.85M  -
datapool/nfs/bla1-project_layouts at weekly-2016-07-30_04.15.01--2w       0      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-07-31_03.14.47--7d     128K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-08-01_05.03.36--7d     128K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-08-02_05.02.39--7d     176K      -  1.85M  -
datapool/nfs/bla1-project_layouts at daily-2016-08-03_03.57.46--7d     176K      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_12.04.00--12h   144K      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_13.04.00--12h   112K      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_14.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_15.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_16.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_17.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_18.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_19.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_20.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_21.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_22.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-03_23.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-project_layouts at hourly-2016-08-04_00.04.00--12h      0      -  1.85M  -
datapool/nfs/bla1-wkhtml at weekly-2016-07-23_04.44.27--2w             128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-07-28_03.35.25--7d              128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-07-29_03.33.40--7d              128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-07-30_03.22.18--7d                 0      -   208K  -
datapool/nfs/bla1-wkhtml at weekly-2016-07-30_04.15.01--2w                0      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-07-31_03.14.47--7d              128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-08-01_05.03.36--7d              128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-08-02_05.02.39--7d              128K      -   208K  -
datapool/nfs/bla1-wkhtml at daily-2016-08-03_03.57.46--7d                 0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_12.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_13.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_14.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_15.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_16.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_17.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_18.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_19.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_20.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_21.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_22.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-03_23.04.00--12h               0      -   208K  -
datapool/nfs/bla1-wkhtml at hourly-2016-08-04_00.04.00--12h               0      -   208K  -



I even went as far as „cross-flashing“ LSI’s 20.00.xy firmware to the HP cards, which came with very old firmware (which really only works with the older MS-DOS versions of the software).
HP refuses to provide updated firmware.

However, zfs receive stalls the system even if there is virtually no data to be transferred.
The stalls takes longer at 03 and 04 AM, which is when I assume the filesystems that are deleted on the master are also deleted on this zfs receive target.


...
-----
Wed Aug 3 01:17:00 CEST 2016
Starting sync...
Wed Aug 3 01:17:17 CEST 2016
-----
Wed Aug 3 02:17:00 CEST 2016
Starting sync...
Wed Aug 3 02:17:17 CEST 2016
-----
Wed Aug 3 03:17:00 CEST 2016
Starting sync...
Wed Aug 3 03:23:16 CEST 2016
-----
Wed Aug 3 04:17:00 CEST 2016
Starting sync...
Wed Aug 3 04:20:12 CEST 2016
-----
Wed Aug 3 05:17:00 CEST 2016
Starting sync...
Wed Aug 3 05:17:22 CEST 2016

…
Thu Aug 4 01:17:00 CEST 2016
Starting sync...
Thu Aug 4 01:17:24 CEST 2016
-----
Thu Aug 4 02:17:00 CEST 2016
Starting sync...
Thu Aug 4 02:17:20 CEST 2016
-----
Thu Aug 4 03:17:00 CEST 2016
Starting sync...
Thu Aug 4 03:23:14 CEST 2016
-----
Thu Aug 4 04:17:00 CEST 2016
Starting sync...
Thu Aug 4 04:19:53 CEST 2016
-----
Thu Aug 4 05:17:00 CEST 2016
Starting sync...
Thu Aug 4 05:17:29 CEST 2016




I had this problem with 9.x with the old HP PA4x0 controller (and maybe with 10.0) - but it went away with 10.1
I switched controllers when I had to attach an external disk-shelf to the servers because the customer needed more space. Also, it’s a real PITA exchanging broken disks when you have no HPACUCLI for FreeBSD...

The first 12 disks in the pool are 600GB SAS disks, the other 12 disks are 900 GB SAS in an external HP enclosure.
I have no L2ARC, no separate log-device.

The system is really completely frozen. Besides just being a warm-standby device, this server also acts as a Read-Only MySQL Slave that the application uses.
When it hangs, the whole application hangs and Netbackup stops backing up.
The zfs sender has no problems.

The vdevs/pools were created with vfs.zfs.min_auto_ashift=12


What else is there to look for?


------------------------------------------------------------------------
ZFS Subsystem Report				Thu Aug  4 00:09:56 2016
------------------------------------------------------------------------

System Information:

	Kernel Version:				1003000 (osreldate)
	Hardware Platform:			amd64
	Processor Architecture:			amd64

	ZFS Storage pool Version:		5000
	ZFS Filesystem Version:			5

FreeBSD 10.3-RELEASE #0 r297264: Fri Mar 25 02:10:02 UTC 2016 root
12:09AM  up 1 day,  6:59, 1 user, load averages: 0.01, 0.07, 0.07

------------------------------------------------------------------------

System Memory:

	0.32%	615.41	MiB Active,	11.44%	21.39	GiB Inact
	50.72%	94.86	GiB Wired,	0.00%	252.00	KiB Cache
	37.52%	70.18	GiB Free,	0.00%	64.00	KiB Gap

	Real Installed:				192.00	GiB
	Real Available:			99.97%	191.94	GiB
	Real Managed:			97.44%	187.03	GiB

	Logical Total:				192.00	GiB
	Logical Used:			52.31%	100.43	GiB
	Logical Free:			47.69%	91.57	GiB

Kernel Memory:					1.45	GiB
	Data:				98.17%	1.43	GiB
	Text:				1.83%	27.14	MiB

Kernel Memory Map:				187.03	GiB
	Size:				32.73%	61.22	GiB
	Free:				67.27%	125.81	GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
	Memory Throttle Count:			0

ARC Misc:
	Deleted:				15
	Recycle Misses:				0
	Mutex Misses:				0
	Evict Skips:				436

ARC Size:				30.04%	55.89	GiB
	Target Size: (Adaptive)		100.00%	186.03	GiB
	Min Size (Hard Limit):		12.50%	23.25	GiB
	Max Size (High Water):		8:1	186.03	GiB

ARC Size Breakdown:
	Recently Used Cache Size:	50.00%	93.01	GiB
	Frequently Used Cache Size:	50.00%	93.01	GiB

ARC Hash Breakdown:
	Elements Max:				1.65m
	Elements Current:		99.94%	1.65m
	Collisions:				358.52k
	Chain Max:				3
	Chains:					37.77k

------------------------------------------------------------------------

ARC Efficiency:					87.79m
	Cache Hit Ratio:		64.95%	57.01m
	Cache Miss Ratio:		35.05%	30.77m
	Actual Hit Ratio:		60.48%	53.09m

	Data Demand Efficiency:		96.42%	21.65m
	Data Prefetch Efficiency:	58.89%	4.98m

	CACHE HITS BY CACHE LIST:
	  Anonymously Used:		6.88%	3.92m
	  Most Recently Used:		30.90%	17.62m
	  Most Frequently Used:		62.22%	35.48m
	  Most Recently Used Ghost:	0.00%	0
	  Most Frequently Used Ghost:	0.00%	0

	CACHE HITS BY DATA TYPE:
	  Demand Data:			36.61%	20.87m
	  Prefetch Data:		5.14%	2.93m
	  Demand Metadata:		56.47%	32.20m
	  Prefetch Metadata:		1.78%	1.02m

	CACHE MISSES BY DATA TYPE:
	  Demand Data:			2.52%	775.75k
	  Prefetch Data:		6.65%	2.05m
	  Demand Metadata:		89.40%	27.51m
	  Prefetch Metadata:		1.42%	438.12k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:					180.34m
	Hit Ratio:			0.31%	554.89k
	Miss Ratio:			99.69%	179.78m

	Colinear:				0
	  Hit Ratio:			100.00%	0
	  Miss Ratio:			100.00%	0

	Stride:					0
	  Hit Ratio:			100.00%	0
	  Miss Ratio:			100.00%	0

DMU Misc:
	Reclaim:				0
	  Successes:			100.00%	0
	  Failures:			100.00%	0

	Streams:				0
	  +Resets:			100.00%	0
	  -Resets:			100.00%	0
	  Bogus:				0

------------------------------------------------------------------------

VDEV cache is disabled

------------------------------------------------------------------------

ZFS Tunables (sysctl):
	kern.maxusers                           12620
	vm.kmem_size                            200818708480
	vm.kmem_size_scale                      1
	vm.kmem_size_min                        0
	vm.kmem_size_max                        1319413950874
	vfs.zfs.trim.max_interval               1
	vfs.zfs.trim.timeout                    30
	vfs.zfs.trim.txg_delay                  32
	vfs.zfs.trim.enabled                    1
	vfs.zfs.vol.unmap_enabled               1
	vfs.zfs.vol.mode                        1
	vfs.zfs.version.zpl                     5
	vfs.zfs.version.spa                     5000
	vfs.zfs.version.acl                     1
	vfs.zfs.version.ioctl                   5
	vfs.zfs.debug                           0
	vfs.zfs.super_owner                     0
	vfs.zfs.sync_pass_rewrite               2
	vfs.zfs.sync_pass_dont_compress         5
	vfs.zfs.sync_pass_deferred_free         2
	vfs.zfs.zio.exclude_metadata            0
	vfs.zfs.zio.use_uma                     1
	vfs.zfs.cache_flush_disable             0
	vfs.zfs.zil_replay_disable              0
	vfs.zfs.min_auto_ashift                 12
	vfs.zfs.max_auto_ashift                 13
	vfs.zfs.vdev.trim_max_pending           10000
	vfs.zfs.vdev.bio_delete_disable         0
	vfs.zfs.vdev.bio_flush_disable          0
	vfs.zfs.vdev.write_gap_limit            4096
	vfs.zfs.vdev.read_gap_limit             32768
	vfs.zfs.vdev.aggregation_limit          131072
	vfs.zfs.vdev.trim_max_active            64
	vfs.zfs.vdev.trim_min_active            1
	vfs.zfs.vdev.scrub_max_active           2
	vfs.zfs.vdev.scrub_min_active           1
	vfs.zfs.vdev.async_write_max_active     10
	vfs.zfs.vdev.async_write_min_active     1
	vfs.zfs.vdev.async_read_max_active      3
	vfs.zfs.vdev.async_read_min_active      1
	vfs.zfs.vdev.sync_write_max_active      10
	vfs.zfs.vdev.sync_write_min_active      10
	vfs.zfs.vdev.sync_read_max_active       10
	vfs.zfs.vdev.sync_read_min_active       10
	vfs.zfs.vdev.max_active                 1000
	vfs.zfs.vdev.async_write_active_max_dirty_percent60
	vfs.zfs.vdev.async_write_active_min_dirty_percent30
	vfs.zfs.vdev.mirror.non_rotating_seek_inc1
	vfs.zfs.vdev.mirror.non_rotating_inc    0
	vfs.zfs.vdev.mirror.rotating_seek_offset1048576
	vfs.zfs.vdev.mirror.rotating_seek_inc   5
	vfs.zfs.vdev.mirror.rotating_inc        0
	vfs.zfs.vdev.trim_on_init               1
	vfs.zfs.vdev.cache.bshift               16
	vfs.zfs.vdev.cache.size                 0
	vfs.zfs.vdev.cache.max                  16384
	vfs.zfs.vdev.metaslabs_per_vdev         200
	vfs.zfs.txg.timeout                     5
	vfs.zfs.space_map_blksz                 4096
	vfs.zfs.spa_slop_shift                  5
	vfs.zfs.spa_asize_inflation             24
	vfs.zfs.deadman_enabled                 1
	vfs.zfs.deadman_checktime_ms            5000
	vfs.zfs.deadman_synctime_ms             1000000
	vfs.zfs.recover                         0
	vfs.zfs.spa_load_verify_data            1
	vfs.zfs.spa_load_verify_metadata        1
	vfs.zfs.spa_load_verify_maxinflight     10000
	vfs.zfs.check_hostid                    1
	vfs.zfs.mg_fragmentation_threshold      85
	vfs.zfs.mg_noalloc_threshold            0
	vfs.zfs.condense_pct                    200
	vfs.zfs.metaslab.bias_enabled           1
	vfs.zfs.metaslab.lba_weighting_enabled  1
	vfs.zfs.metaslab.fragmentation_factor_enabled1
	vfs.zfs.metaslab.preload_enabled        1
	vfs.zfs.metaslab.preload_limit          3
	vfs.zfs.metaslab.unload_delay           8
	vfs.zfs.metaslab.load_pct               50
	vfs.zfs.metaslab.min_alloc_size         33554432
	vfs.zfs.metaslab.df_free_pct            4
	vfs.zfs.metaslab.df_alloc_threshold     131072
	vfs.zfs.metaslab.debug_unload           0
	vfs.zfs.metaslab.debug_load             0
	vfs.zfs.metaslab.fragmentation_threshold70
	vfs.zfs.metaslab.gang_bang              16777217
	vfs.zfs.free_bpobj_enabled              1
	vfs.zfs.free_max_blocks                 -1
	vfs.zfs.no_scrub_prefetch               0
	vfs.zfs.no_scrub_io                     0
	vfs.zfs.resilver_min_time_ms            3000
	vfs.zfs.free_min_time_ms                1000
	vfs.zfs.scan_min_time_ms                1000
	vfs.zfs.scan_idle                       50
	vfs.zfs.scrub_delay                     4
	vfs.zfs.resilver_delay                  2
	vfs.zfs.top_maxinflight                 32
	vfs.zfs.zfetch.array_rd_sz              1048576
	vfs.zfs.zfetch.max_distance             8388608
	vfs.zfs.zfetch.min_sec_reap             2
	vfs.zfs.zfetch.max_streams              8
	vfs.zfs.prefetch_disable                0
	vfs.zfs.delay_scale                     500000
	vfs.zfs.delay_min_dirty_percent         60
	vfs.zfs.dirty_data_sync                 67108864
	vfs.zfs.dirty_data_max_percent          10
	vfs.zfs.dirty_data_max_max              4294967296
	vfs.zfs.dirty_data_max                  4294967296
	vfs.zfs.max_recordsize                  1048576
	vfs.zfs.mdcomp_disable                  0
	vfs.zfs.nopwrite_enabled                1
	vfs.zfs.dedup.prefetch                  1
	vfs.zfs.l2c_only_size                   0
	vfs.zfs.mfu_ghost_data_lsize            0
	vfs.zfs.mfu_ghost_metadata_lsize        0
	vfs.zfs.mfu_ghost_size                  0
	vfs.zfs.mfu_data_lsize                  40921600
	vfs.zfs.mfu_metadata_lsize              2360084992
	vfs.zfs.mfu_size                        4470225920
	vfs.zfs.mru_ghost_data_lsize            0
	vfs.zfs.mru_ghost_metadata_lsize        0
	vfs.zfs.mru_ghost_size                  0
	vfs.zfs.mru_data_lsize                  49482637824
	vfs.zfs.mru_metadata_lsize              4404856320
	vfs.zfs.mru_size                        53920903168
	vfs.zfs.anon_data_lsize                 0
	vfs.zfs.anon_metadata_lsize             0
	vfs.zfs.anon_size                       106496
	vfs.zfs.l2arc_norw                      1
	vfs.zfs.l2arc_feed_again                1
	vfs.zfs.l2arc_noprefetch                1
	vfs.zfs.l2arc_feed_min_ms               200
	vfs.zfs.l2arc_feed_secs                 1
	vfs.zfs.l2arc_headroom                  2
	vfs.zfs.l2arc_write_boost               8388608
	vfs.zfs.l2arc_write_max                 8388608
	vfs.zfs.arc_meta_limit                  49936241664
	vfs.zfs.arc_free_target                 339922
	vfs.zfs.arc_shrink_shift                7
	vfs.zfs.arc_average_blocksize           8192
	vfs.zfs.arc_min                         24968120832
	vfs.zfs.arc_max                         199744966656

————————————————————————————————————





More information about the freebsd-fs mailing list