Hi List, I have read the Gotcha[1] page: Files with a lot of random writes can become heavily fragmented (10000+ extents) causing trashing on HDDs and excessive multi-second spikes of CPU load on systems with an SSD or **large amount a RAM**. Why could large amount of memory worsen the problem? If **too much** memory is a problem, is it possible to limit the memory btrfs use? Background info: I am running a heavy-write database server with 96GB ram. In the worse case it cause multi minutes of high cpu loads. Systemd keeping kill and restarting services, and old job don't die because they stuck in uninterruptable wait... etc. Tried with nodatacow, but it seems only affect new file. It is not an subvolume option either... Regards, Daniel [1] https://btrfs.wiki.kernel.org/index.php/Gotchas#Fragmentation -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html
