I may as well make myself unpopular with some context…
Some here have compared NTFS with ZFS, which is unfair as ZFS is over 12 years younger. In 1993 machines had an average of less than 4Mb of RAM, and the average disk size was probably somewhere in the 80-100Mb range. NTFS required more RAM - if you wanted to run it I think you had to have 12Mb of RAM minimum, maybe even 16Mb. If you didn’t have that you had to install your Windows NT 3.1 copy with FAT…
A better comparison filesystem would be XFS, which was developed at around the same time and saw its first release in 1994.
XFS has had a lot more development of late than NTFS has, and it could be argued that because of that it now has the edge. But both are venerable survivors of that era. Both are reliable, robust, feature-rich and widely deployed.
A lot of problems that people have with NTFS are to do with the way Windows handles disk access rather than the filesystem itself. A filesystem is more than just an on-disk layout and a bit of code to read or write from it, it also has to interact with OS disk buffering systems, security systems, caching mechanisms, and possibly even things like file locking and notification mechanisms.
Windows has a concept of the “installable file system” - these days it’s primarily a way to load filter drivers that can inspect all I/O operations. It’s how Windows security programs like antivirus work, but also how Windows prevents writes to its own folders by ordinary users. As you can guess, that slows things down. On the boot/OS drive of a Windows machine there are a lot of filter drivers. Android developers know this from how long some build operations take, and have often cursed at NTFS for it. Yet if you move the project onto a non-OS NTFS drive, suddenly it’s much faster - because that drive lacks many of those filter drivers, as there is no OS to protect on that drive.
The point here being that NTFS often gets slammed for issues which aren’t its fault, and it has no control over.
NTFS is probably in the top ten most-installed filesystems ever. And high on that top ten. (I wonder what that top ten would look like? I think that embedded use of ext2 probably places it near the top, but then you have wildcards like the Minix file system… anyway, back on track!)
Filesystems are one of those things that everyone takes for granted, yet are incredibly important. NTFS may not be native to Linux, and may come from somewhere that many see as “the enemy”, but I think 30 years of tireless work deserves some recognition.
Happy birthday, NTFS. You’ve done well.
My local backups are handled by rdiff-backup to a mirror set of disks. That means my data is versioned but easily accessible for immediate restore, and now on three disks (my SSD, and two rotating rust drives). It also makes restores as simple as copying a file if I want the latest version, or an easy command if I want an older version. And testing backups is as easy as a diff command to compare the backup version with the live version.
Having your files just be files in your backup solution is very handy. At work I don’t mind having to use an application like Veeam, because I’m being paid to do that. At home I want to see my backups quickly and easily, because I’d rather be working on my files than wrestling with backup software…
Remote backups are handled by SpiderOak, who have been fine for me for almost a decade. I also use them to synchronise my desktop and laptop computer. On my desktop SpiderOak also backs up some files in an archive area on the rotating rust mirror set - stuff that’s large and I don’t access often, so don’t need to put on my laptop but do want backed up.
I also have a USB thumbdrive that’s encrypted and used when I’m travelling to back up changes on my laptop via a simple rsync copy - just in case I have limited internet access and SpiderOak can’t do its thing…
I did also have a NAS in the mix once, but I realised that it was a waste of energy - both mine and electricity. In normal circumstances my data is on 5 locations (desktop SSD, laptop SSD, desktop mirror set, SpiderOak’s storage) and in the very worst case it’s in two locations (laptop SSD, USB thumbdrive). Rdiff-backup to the NAS was simply overkill once I’d added the local mirror set into my desktop, so I retired it.
I’d added the local mirror set because I was working with large files - data sets and VM images - and backups over the network to the NAS were taking an age. A local set of cheap disks in my desktop tower was faster and yet still fairly cheap.
Here’s my advice for your consideration:
Good luck with your backup strategy!
Absolutely - rdiff-backup onto a local mirror set of disks. As you say, the big advantage is that the last “current” entry in the backup is available just by browsing, but I have a full history just a command away. Backups are no use if you can’t access them, and people really under-rate ease of access when evaluating their backup strategy.
It’s fine.
Like any CMS, it has a seemingly constant low level of patching to be applied. The more third party modules and themes you have, the worse that gets.
Remove unused modules that aren’t core. Same with themes. That’ll make things easier.
Otherwise it’s overheads are just Apache/nginx, MySQL/MariaDb, and maintenance of the TLS certificate, plus OS patching. All fairly well understood stuff that you should have no issues with.