ZFS exhausts kernel memory just by importing zpools
Nagy, Attila
bra at fsn.hu
Wed Jul 3 14:34:16 UTC 2019
Hi,
Oh, should've written about that: no I don't use (and never used) dedup.
On 2019. 07. 02. 18:13, Sam Fourman Jr. wrote:
> Hello,
>
> My initial guess is that you may have de-duplication enabled on one
> (or more) of the underlying datasets.
> **if** this is the case, a simple solution is to add more memory to
> the machine. (64GB of memory is not sufficient for dedup to be enabled )
>
> -- Sam Fourman Jr.
>
> On Tue, Jul 2, 2019 at 10:59 AM Nagy, Attila <bra at fsn.hu
> <mailto:bra at fsn.hu>> wrote:
>
> Hi,
>
> Running latest stable/12 on amd64 with 64 GiB memory on a machine
> with
> 44 4T disks. Each disks have its own zpool on it (because I solve the
> redundancy between machines and not locally with ZFS).
>
> One example zpool holds 2.2 TiB of data (according to df) and have
> around 75 million files in hashed directories, this is the typical
> usage
> on them.
>
> When I import these zpools, top says around 50 GiB wired memory
> (ARC is
> minimal, files weren't yet touched) and after I start to use (heavy
> reads/writes) the pools, the free memory quickly disappears (ARC
> grows)
> until all memory is gone and the machine starts to kill processes,
> ends
> up in a deadlock, where nothing helps.
>
> If I import the pools one by one, each of them adds around 1-1.5
> GiB of
> wired memory.
>
> Top shows this, right after it came to a halt and nothing else
> works (I
> can't log in even on the console):
>
> last pid: 61878; load averages: 5.05, 4.42, 2.50 up 0+01:07:23
> 15:45:17
> 171 processes: 1 running, 162 sleeping, 1 stopped, 1 zombie, 6 waiting
> CPU: 0.0% user, 0.0% nice, 0.2% system, 0.0% interrupt, 99.8% idle
> Mem: 7716K Active, 8192 Inact, 84K Laundry, 57G Wired, 180M Buf,
> 14M Free
> ARC: 21G Total, 10G MFU, 4812M MRU, 4922M Anon, 301M Header, 828M
> Other
> 5739M Compressed, 13G Uncompressed, 2.35:1 Ratio
> Swap:
>
> PID USERNAME THR PRI NICE SIZE RES STATE C TIME WCPU
> COMMAND
> 61412 root 1 20 0 14M 3904K CPU14 14 0:06 1.55% top
> 57569 redis 57 20 0 1272M 64M uwait 22 4:28 0.24%
> consul
> 5574 root 1 20 0 13M 3440K nanslp 10 0:02
> 0.05% gstat
> 5557 root 1 20 0 20M 7808K select 20 0:00
> 0.01% sshd
> 5511 root 1 20 0 20M 7808K select 4 0:01
> 0.01% sshd
> 4955 root 1 20 0 10M 1832K select 9 0:00 0.01%
> supervis
> 5082 root 1 20 0 25M 14M select 0 0:00
> 0.00% perl
> 4657 _pflogd 1 20 0 12M 2424K bpf 1 0:00
> 0.00% pflogd
> 5059 elasticsea 2 20 -20 6983M 385M STOP 5 1:29
> 0.00% java
> 61669 root 1 26 0 23M 0 pfault 4 0:14 0.00%
> <python3
> 61624 root 1 20 -20 24M 14M buf_ha 9 0:09 0.00%
> python3.
> 61626 root 1 20 -20 23M 16K pfault 0 0:08 0.00%
> python3.
> 61651 root 1 20 -20 23M 14M buf_ha 10 0:08 0.00%
> python3.
> 61668 root 1 20 -20 23M 13M buf_ha 20 0:08 0.00%
> python3.
>
> I've already tried to shrink ARC and vm.kmem_size without too much
> success.
>
> Any ideas what causes this?
>
> _______________________________________________
> freebsd-fs at freebsd.org <mailto:freebsd-fs at freebsd.org> mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-fs
> To unsubscribe, send any mail to
> "freebsd-fs-unsubscribe at freebsd.org
> <mailto:freebsd-fs-unsubscribe at freebsd.org>"
>
>
>
> --
>
> Sam Fourman Jr.
More information about the freebsd-fs
mailing list