-
Notifications
You must be signed in to change notification settings - Fork 1.9k
Description
Distribution Name | debian
Distribution Version | 10.9 (buster)
Linux Kernel | 5.4.106-1
Architecture | x86_64
ZFS Version | 2.0.4-pve1
SPL Version | 2.0.4-pve1
Describe the problem you're observing
i have metadata intensive workload, so i set primarycache=metadata and secondarycache=metadata because otherwise, data would evict metadata too soon.
when doing "find /dir/tree/with/1mio/dirs/files" or "rsync -av --dry-run /dir/tree/with/1mio/dirs/files /tmp" i see arc and l2arc filling, but after a while, arc is collapsing and file metadata is not even served from l2arc, or at least only being served particularly from there.
even doing find / rsync a multiple times does not reliable fix this, i see lot's of reads being served from disk instead from arc or l2arc on ssd, so this renders using arc/l2arc for metadata caching pretty useless for me.
in arcstat, i see typical arc miss% at 5 and dm% at 4-5 - and suddenly pm% goes up from 0 to 100 and is see read i/o hitting the clasic harddisks instead being served from ram/ssd . that's very frustrating.
system has 4gb of ram, btw.
i'm trying to optimize two different system (rsync backup machine and proxmox backup server) for metadata centric type of workload, because otherwise backup job runtime gets too long. i'm aware of special vdev feature for relocating metadata, but i do not see the point why i should invest into enterprise quality, mirrored ssd when darn cheap consumer ssd should be able to act as a metadata read caching device, which is not not causing any data loss when it breaks.