A year ago I set up Ubuntu server with 3 ZFS pools on my server, normally I don’t make copies of very large files but today I was making a copy of a ~30GB directory and I saw in rsync that the transfer doesn’t exceed 3mb/s (cp is also very slow).
What is the best file system that “just works”? I’m thinking of migrating everything to ext4
EDIT: I really like the automatic pool recovery feature in ZFS, has saved me from 1 hard drive failure so far
ZFS is a very robust choice for a NAS. Many people, myself included, as well as hundreds of businesses across the globe, have used ZFS at scale for over a decade.
Attack the problem. Check your system logs, htop, zpool status.
When was the last time you ran a zpool scrub? Is there a scrub, or other zfs operation in progress? How many snapshots do you have? How much RAM vs disk space? Are you using ZFS deduplication? Compression?
Ext4 does not have snapshots, COW or similar features. I am very happy with BTRFS. It just “works” out of the box.
I use BTRFS on everything too nowadays. The thing that made me switch everything to BTRFS was filesystem compression.
FWIW lvm can give you snapshots and other features. And mdadm can be used for a raid. All very robust tools.
Yes but BTRFS can this out of the box without extra tools. Both ways have their own advantage, but I would still prefer BTRFS
I’m in BTRFS, and wish I wasn’t.
Booting into a failed mdadm RAID1 is normal,
whereas booting into a failed BTRFS RAID1 requires competent manual intervention, and special parameters given to the boot-kernel.
mdadm & lvm, with a fixed version of ZFS would be my preference.
ZFS recently had a bug discovered that was silently corrupting data, and I HOPE a fix has been got in.
Lemme see if I can find something on both of these points…
https://linuxnatives.net/2015/using-raid-btrfs-recovering-broken-disks
https://www.theregister.com/2023/11/27/openzfs_2_2_0_data_corruption/
_ /\ _
ZFS will perform better on a NAS
Make sure you don’t have SMR drives, if they are spinning drives. CMR drives are the I ly ones that should be used in a NAS, especially with ZFS. https://vermaden.wordpress.com/2022/05/08/zfs-on-smr-drives/
From the article it looks like zfs is the perfect file system for smr drives as it would try to cache random writes
Possibly, with tuning. Op would just have to be careful about reslivering. In my experience SMR drives really slow down when the CMR buffer is full.
Acronyms, initialisms, abbreviations, contractions, and other phrases which expand to something larger, that I’ve seen in this thread:
Fewer Letters More Letters LVM (Linux) Logical Volume Manager for filesystem mapping NAS Network-Attached Storage PSU Power Supply Unit SSD Solid State Drive mass storage ZFS Solaris/Linux filesystem focusing on data integrity
5 acronyms in this thread; the most compressed thread commented on today has 9 acronyms.
[Thread #486 for this sub, first seen 5th Feb 2024, 15:05] [FAQ] [Full list] [Contact] [Source code]
ZFS is by far the best just use TrueNAS, Ubuntu is crap at supporting ZFS, also only set your pool’s VDEV 6-8 wide.
I was thinking about switching to debian (all that I host is in docker so that’s why), but the weird thing is that it was working perfectly 1 month ago
Maybe your HBA is having issues? or a Drive is Failing? have you done a memtest? you may need to do system wide tests, it can even be a PSU failing or a software Bug.
also TrueNAS is built with Docker they use it heavily something like 106 apps, Debian has good ZFS support, but you will end up doing a lot of unneeded work using Debian unless you keep it simple.