plenty of memory, but system us intensively swapping
Eugene M. Zheganin
eugene at zhegan.in
Tue Nov 20 09:54:02 UTC 2018
Hello,
I have a recent FreeBSD 11-STABLE which is mainly used as an iSCSI
target. The system has 64G of RAM but is swapping intensively. Yup,
about of half of the memory is used as ZFS ARC (isn't capped in
loader.conf), and another half is eaten by the kernel, but it oly uses
only about half of it (thus 25% of the total amount).
Could this be tweaked by some sysctl oids (I suppose not, but worth asking).
top, vmstat 1 snapshots and zfs-stats -a are listed below.
Thanks.
[root at san01:nginx/vhost.d]# vmstat 1
procs memory page disks faults cpu
r b w avm fre flt re pi po fr sr da0 da1 in sy cs us sy id
0 0 38 23G 609M 1544 68 118 64 895 839 0 0 3644 2678 649
0 13 87
0 0 53 23G 601M 1507 185 742 315 1780 33523 651 664 56438 785
476583 0 28 72
0 0 53 23G 548M 1727 330 809 380 2377 33256 758 763 55555 1273
468545 0 26 73
0 0 53 23G 528M 1702 239 660 305 1347 32335 611 631 59962 1025
490365 0 22 78
0 0 52 23G 854M 2409 309 693 203 97943 16944 525 515 64309 1570
540533 0 29 71
3 0 54 23G 1.1G 2756 639 641 149 124049 19531 542 538 64777 1576
553946 0 35 65
0 0 53 23G 982M 1694 236 680 282 2754 35602 597 603 66540 1385
583687 0 28 72
0 0 41 23G 867M 1882 223 767 307 1162 34936 682 638 67284 780
568818 0 33 67
0 0 39 23G 769M 1542 167 673 336 1187 35123 646 610 65925 1176
551623 0 23 77
2 0 41 23G 700M 3602 535 688 327 2192 37109 622 594 65862 4256
518934 0 33 67
0 0 54 23G 650M 2957 219 726 464 4838 36464 852 868 65384 4110
558132 1 37 62
0 0 54 23G 641M 1576 245 730 344 1139 33681 740 679 67216 970
560379 0 31 69
[root at san01:nginx/vhost.d]# top
last pid: 55190; load averages: 11.32, 12.15, 10.76
up 10+16:05:14 14:38:58
101 processes: 1 running, 100 sleeping
CPU: 0.2% user, 0.0% nice, 28.9% system, 1.6% interrupt, 69.3% idle
Mem: 85M Active, 1528K Inact, 12K Laundry, 62G Wired, 540M Free
ARC: 31G Total, 19G MFU, 6935M MRU, 2979M Anon, 556M Header, 1046M Other
25G Compressed, 34G Uncompressed, 1.39:1 Ratio
Swap: 32G Total, 1186M Used, 31G Free, 3% Inuse, 7920K In, 3752K Out
PID USERNAME THR PRI NICE SIZE RES STATE C TIME WCPU
COMMAND
40132 root 131 52 0 3152M 75876K uwait 14 36:59 6.10% java
55142 root 1 20 0 7904K 2728K CPU20 20 0:00 0.72% top
20026 root 1 20 0 106M 5676K nanslp 28 1:23 0.60% gstat
53642 root 1 20 0 7904K 2896K select 14 0:03 0.58% top
977 zfsreplica 1 20 0 30300K 3568K kqread 21 4:00 0.42%
uwsgi
968 zfsreplica 1 20 0 30300K 2224K swread 11 2:03 0.21%
uwsgi
973 zfsreplica 1 20 0 30300K 2264K swread 13 12:26 0.13%
uwsgi
53000 www 1 20 0 23376K 1372K kqread 24 0:00 0.05% nginx
1292 root 1 20 0 6584K 2040K select 29 0:23 0.04%
blacklistd
776 zabbix 1 20 0 12408K 4236K nanslp 26 4:42 0.03%
zabbix_agentd
1289 root 1 20 0 67760K 5148K select 13 9:50 0.03%
bsnmpd
777 zabbix 1 20 0 12408K 1408K select 25 5:06 0.03%
zabbix_agentd
785 zfsreplica 1 20 0 27688K 3960K kqread 28 2:04 0.02%
uwsgi
975 zfsreplica 1 20 0 30300K 464K kqread 18 2:33 0.02%
uwsgi
974 zfsreplica 1 20 0 30300K 480K kqread 30 3:39 0.02%
uwsgi
965 zfsreplica 1 20 0 30300K 464K kqread 4 3:23 0.02%
uwsgi
976 zfsreplica 1 20 0 30300K 464K kqread 14 2:59 0.01%
uwsgi
972 zfsreplica 1 20 0 30300K 464K kqread 10 2:57 0.01%
uwsgi
963 zfsreplica 1 20 0 30300K 460K kqread 3 2:45 0.01%
uwsgi
971 zfsreplica 1 20 0 30300K 464K kqread 13 3:16 0.01%
uwsgi
69644 emz 1 20 0 13148K 4596K select 24 0:05 0.01% sshd
18203 vryabov 1 20 0 13148K 4624K select 9 0:02 0.01% sshd
636 root 1 20 0 6412K 1884K select 17 4:10 0.01%
syslogd
51266 emz 1 20 0 13148K 4576K select 5 0:00 0.01% sshd
964 zfsreplica 1 20 0 30300K 460K kqread 18 11:02 0.01%
uwsgi
962 zfsreplica 1 20 0 30300K 460K kqread 28 6:56 0.01%
uwsgi
969 zfsreplica 1 20 0 30300K 464K kqread 12 2:07 0.01%
uwsgi
967 zfsreplica 1 20 0 30300K 464K kqread 27 5:18 0.01%
uwsgi
970 zfsreplica 1 20 0 30300K 464K kqread 0 4:25 0.01%
uwsgi
966 zfsreplica 1 22 0 30300K 468K kqread 14 4:29 0.01%
uwsgi
53001 www 1 20 0 23376K 1256K kqread 10 0:00 0.01% nginx
791 zfsreplica 1 20 0 27664K 4244K kqread 17 1:34 0.01%
uwsgi
52431 root 1 20 0 17132K 4492K select 21 0:00 0.01% mc
70013 root 1 20 0 17132K 4492K select 4 0:03 0.01% mc
870 root 1 20 0 12448K 12544K select 19 0:51 0.01% ntpd
[root at san01:nginx/vhost.d]# zfs-stats -a
------------------------------------------------------------------------
ZFS Subsystem Report Tue Nov 20 14:39:05 2018
------------------------------------------------------------------------
System Information:
Kernel Version: 1102503 (osreldate)
Hardware Platform: amd64
Processor Architecture: amd64
ZFS Storage pool Version: 5000
ZFS Filesystem Version: 5
FreeBSD 11.2-STABLE #0 r340287M: Fri Nov 9 22:23:22 +05 2018 emz
14:39 up 10 days, 16:05, 5 users, load averages: 10,96 12,05 10,74
------------------------------------------------------------------------
System Memory:
0.14% 90.20 MiB Active, 0.01% 8.62 MiB Inact
98.97% 61.57 GiB Wired, 0.00% 0 Cache
0.88% 560.02 MiB Free, -0.00% -184320 Bytes Gap
Real Installed: 64.00 GiB
Real Available: 99.77% 63.85 GiB
Real Managed: 97.43% 62.21 GiB
Logical Total: 64.00 GiB
Logical Used: 99.13% 63.44 GiB
Logical Free: 0.87% 568.64 MiB
Kernel Memory: 22.77 GiB
Data: 99.84% 22.73 GiB
Text: 0.16% 36.52 MiB
Kernel Memory Map: 62.21 GiB
Size: 46.85% 29.15 GiB
Free: 53.15% 33.06 GiB
------------------------------------------------------------------------
ARC Summary: (HEALTHY)
Memory Throttle Count: 0
ARC Misc:
Deleted: 5.61b
Recycle Misses: 0
Mutex Misses: 64.25m
Evict Skips: 98.33m
ARC Size: 50.02% 30.62 GiB
Target Size: (Adaptive) 50.02% 30.62 GiB
Min Size (Hard Limit): 12.50% 7.65 GiB
Max Size (High Water): 8:1 61.21 GiB
ARC Size Breakdown:
Recently Used Cache Size: 50.69% 15.52 GiB
Frequently Used Cache Size: 49.31% 15.10 GiB
ARC Hash Breakdown:
Elements Max: 8.35m
Elements Current: 30.08% 2.51m
Collisions: 2.18b
Chain Max: 10
Chains: 308.03k
------------------------------------------------------------------------
ARC Efficiency: 48.52b
Cache Hit Ratio: 84.49% 40.99b
Cache Miss Ratio: 15.51% 7.53b
Actual Hit Ratio: 84.19% 40.85b
Data Demand Efficiency: 83.84% 13.06b
Data Prefetch Efficiency: 40.66% 1.42b
CACHE HITS BY CACHE LIST:
Most Recently Used: 15.11% 6.19b
Most Frequently Used: 84.54% 34.66b
Most Recently Used Ghost: 0.97% 396.23m
Most Frequently Used Ghost: 0.18% 75.49m
CACHE HITS BY DATA TYPE:
Demand Data: 26.70% 10.95b
Prefetch Data: 1.41% 576.89m
Demand Metadata: 70.83% 29.04b
Prefetch Metadata: 1.06% 434.19m
CACHE MISSES BY DATA TYPE:
Demand Data: 28.04% 2.11b
Prefetch Data: 11.18% 841.81m
Demand Metadata: 60.37% 4.54b
Prefetch Metadata: 0.40% 30.27m
------------------------------------------------------------------------
L2ARC is disabled
------------------------------------------------------------------------
File-Level Prefetch: (HEALTHY)
DMU Efficiency: 9.20b
Hit Ratio: 5.80% 533.41m
Miss Ratio: 94.20% 8.66b
Colinear: 0
Hit Ratio: 100.00% 0
Miss Ratio: 100.00% 0
Stride: 0
Hit Ratio: 100.00% 0
Miss Ratio: 100.00% 0
DMU Misc:
Reclaim: 0
Successes: 100.00% 0
Failures: 100.00% 0
Streams: 0
+Resets: 100.00% 0
-Resets: 100.00% 0
Bogus: 0
------------------------------------------------------------------------
VDEV cache is disabled
------------------------------------------------------------------------
ZFS Tunables (sysctl):
kern.maxusers 4422
vm.kmem_size 66799345664
vm.kmem_size_scale 1
vm.kmem_size_min 0
vm.kmem_size_max 1319413950874
vfs.zfs.trim.max_interval 1
vfs.zfs.trim.timeout 30
vfs.zfs.trim.txg_delay 32
vfs.zfs.trim.enabled 0
vfs.zfs.vol.immediate_write_sz 131072
vfs.zfs.vol.unmap_sync_enabled 0
vfs.zfs.vol.unmap_enabled 1
vfs.zfs.vol.recursive 0
vfs.zfs.vol.mode 1
vfs.zfs.version.zpl 5
vfs.zfs.version.spa 5000
vfs.zfs.version.acl 1
vfs.zfs.version.ioctl 7
vfs.zfs.debug 0
vfs.zfs.super_owner 0
vfs.zfs.immediate_write_sz 32768
vfs.zfs.sync_pass_rewrite 2
vfs.zfs.sync_pass_dont_compress 5
vfs.zfs.sync_pass_deferred_free 2
vfs.zfs.zio.dva_throttle_enabled 1
vfs.zfs.zio.exclude_metadata 0
vfs.zfs.zio.use_uma 1
vfs.zfs.zil_slog_bulk 786432
vfs.zfs.cache_flush_disable 0
vfs.zfs.zil_replay_disable 0
vfs.zfs.standard_sm_blksz 131072
vfs.zfs.dtl_sm_blksz 4096
vfs.zfs.min_auto_ashift 9
vfs.zfs.max_auto_ashift 13
vfs.zfs.vdev.trim_max_pending 10000
vfs.zfs.vdev.bio_delete_disable 0
vfs.zfs.vdev.bio_flush_disable 0
vfs.zfs.vdev.def_queue_depth 32
vfs.zfs.vdev.queue_depth_pct 1000
vfs.zfs.vdev.write_gap_limit 4096
vfs.zfs.vdev.read_gap_limit 32768
vfs.zfs.vdev.aggregation_limit 1048576
vfs.zfs.vdev.trim_max_active 64
vfs.zfs.vdev.trim_min_active 1
vfs.zfs.vdev.scrub_max_active 2
vfs.zfs.vdev.scrub_min_active 1
vfs.zfs.vdev.async_write_max_active 10
vfs.zfs.vdev.async_write_min_active 1
vfs.zfs.vdev.async_read_max_active 3
vfs.zfs.vdev.async_read_min_active 1
vfs.zfs.vdev.sync_write_max_active 10
vfs.zfs.vdev.sync_write_min_active 10
vfs.zfs.vdev.sync_read_max_active 10
vfs.zfs.vdev.sync_read_min_active 10
vfs.zfs.vdev.max_active 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc 0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc 5
vfs.zfs.vdev.mirror.rotating_inc 0
vfs.zfs.vdev.trim_on_init 1
vfs.zfs.vdev.cache.bshift 16
vfs.zfs.vdev.cache.size 0
vfs.zfs.vdev.cache.max 16384
vfs.zfs.vdev.default_ms_shift 29
vfs.zfs.vdev.min_ms_count 16
vfs.zfs.vdev.max_ms_count 200
vfs.zfs.txg.timeout 5
vfs.zfs.space_map_ibs 14
vfs.zfs.spa_allocators 4
vfs.zfs.spa_min_slop 134217728
vfs.zfs.spa_slop_shift 5
vfs.zfs.spa_asize_inflation 24
vfs.zfs.deadman_enabled 1
vfs.zfs.deadman_checktime_ms 5000
vfs.zfs.deadman_synctime_ms 1000000
vfs.zfs.debug_flags 0
vfs.zfs.debugflags 0
vfs.zfs.recover 0
vfs.zfs.spa_load_verify_data 1
vfs.zfs.spa_load_verify_metadata 1
vfs.zfs.spa_load_verify_maxinflight 10000
vfs.zfs.max_missing_tvds_scan 0
vfs.zfs.max_missing_tvds_cachefile 2
vfs.zfs.max_missing_tvds 0
vfs.zfs.spa_load_print_vdev_tree 0
vfs.zfs.ccw_retry_interval 300
vfs.zfs.check_hostid 1
vfs.zfs.mg_fragmentation_threshold 85
vfs.zfs.mg_noalloc_threshold 0
vfs.zfs.condense_pct 200
vfs.zfs.metaslab_sm_blksz 4096
vfs.zfs.metaslab.bias_enabled 1
vfs.zfs.metaslab.lba_weighting_enabled 1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled 1
vfs.zfs.metaslab.preload_limit 3
vfs.zfs.metaslab.unload_delay 8
vfs.zfs.metaslab.load_pct 50
vfs.zfs.metaslab.min_alloc_size 33554432
vfs.zfs.metaslab.df_free_pct 4
vfs.zfs.metaslab.df_alloc_threshold 131072
vfs.zfs.metaslab.debug_unload 0
vfs.zfs.metaslab.debug_load 0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.force_ganging 16777217
vfs.zfs.free_bpobj_enabled 1
vfs.zfs.free_max_blocks -1
vfs.zfs.zfs_scan_checkpoint_interval 7200
vfs.zfs.zfs_scan_legacy 0
vfs.zfs.no_scrub_prefetch 0
vfs.zfs.no_scrub_io 0
vfs.zfs.resilver_min_time_ms 3000
vfs.zfs.free_min_time_ms 1000
vfs.zfs.scan_min_time_ms 1000
vfs.zfs.scan_idle 50
vfs.zfs.scrub_delay 4
vfs.zfs.resilver_delay 2
vfs.zfs.top_maxinflight 32
vfs.zfs.zfetch.array_rd_sz 1048576
vfs.zfs.zfetch.max_idistance 67108864
vfs.zfs.zfetch.max_distance 8388608
vfs.zfs.zfetch.min_sec_reap 2
vfs.zfs.zfetch.max_streams 8
vfs.zfs.prefetch_disable 0
vfs.zfs.delay_scale 500000
vfs.zfs.delay_min_dirty_percent 60
vfs.zfs.dirty_data_sync 67108864
vfs.zfs.dirty_data_max_percent 10
vfs.zfs.dirty_data_max_max 4294967296
vfs.zfs.dirty_data_max 4294967296
vfs.zfs.max_recordsize 1048576
vfs.zfs.default_ibs 17
vfs.zfs.default_bs 9
vfs.zfs.send_holes_without_birth_time 1
vfs.zfs.mdcomp_disable 0
vfs.zfs.per_txg_dirty_frees_percent 30
vfs.zfs.nopwrite_enabled 1
vfs.zfs.dedup.prefetch 1
vfs.zfs.dbuf_cache_lowater_pct 10
vfs.zfs.dbuf_cache_hiwater_pct 10
vfs.zfs.dbuf_metadata_cache_overflow 0
vfs.zfs.dbuf_metadata_cache_shift 6
vfs.zfs.dbuf_cache_shift 5
vfs.zfs.dbuf_metadata_cache_max_bytes 1026962560
vfs.zfs.dbuf_cache_max_bytes 2053925120
vfs.zfs.arc_min_prescient_prefetch_ms 6
vfs.zfs.arc_min_prefetch_ms 1
vfs.zfs.l2c_only_size 0
vfs.zfs.mfu_ghost_data_esize 1910587392
vfs.zfs.mfu_ghost_metadata_esize 5158840832
vfs.zfs.mfu_ghost_size 7069428224
vfs.zfs.mfu_data_esize 17620227072
vfs.zfs.mfu_metadata_esize 950300160
vfs.zfs.mfu_size 20773338624
vfs.zfs.mru_ghost_data_esize 6989578240
vfs.zfs.mru_ghost_metadata_esize 18479132160
vfs.zfs.mru_ghost_size 25468710400
vfs.zfs.mru_data_esize 4455460352
vfs.zfs.mru_metadata_esize 70236672
vfs.zfs.mru_size 7413314560
vfs.zfs.anon_data_esize 0
vfs.zfs.anon_metadata_esize 0
vfs.zfs.anon_size 3040037888
vfs.zfs.l2arc_norw 1
vfs.zfs.l2arc_feed_again 1
vfs.zfs.l2arc_noprefetch 1
vfs.zfs.l2arc_feed_min_ms 200
vfs.zfs.l2arc_feed_secs 1
vfs.zfs.l2arc_headroom 2
vfs.zfs.l2arc_write_boost 8388608
vfs.zfs.l2arc_write_max 8388608
vfs.zfs.arc_meta_limit 16431400960
vfs.zfs.arc_free_target 113124
vfs.zfs.arc_kmem_cache_reap_retry_ms 0
vfs.zfs.compressed_arc_enabled 1
vfs.zfs.arc_grow_retry 60
vfs.zfs.arc_shrink_shift 7
vfs.zfs.arc_average_blocksize 8192
vfs.zfs.arc_no_grow_shift 5
vfs.zfs.arc_min 8215700480
vfs.zfs.arc_max 65725603840
vfs.zfs.abd_chunk_size 4096
vfs.zfs.abd_scatter_enabled 1
------------------------------------------------------------------------
More information about the freebsd-stable
mailing list