[Discuss] what are pros/cons of different (Linux) filesystems for use with SMR (shingled) hard drives?
Kent Borg
kentborg at borg.org
Fri Nov 13 19:12:26 EST 2020
On 11/13/20 3:35 PM, Dan Ritter wrote:
> https://www.snia.org/sites/default/files/SDC15_presentations/smr/HannesReinecke_Strategies_for_running_unmodified_FS_SMR.pdf
I'll have to look at that.
> Super-short TL;DR: rsync is bad. Compressed tar files are good.
> Big writes are good. Small writes are bad. ext4 can be coerced
> into sorta doing a better job; btrfs ought to be better, but is
> worse.
>
> It's not the longevity of the disk which is the problem, it's
> the fact that once you start updating data on the disk, it can
> cause horrendously amazing write amplification and the disk will
> be ridiculously slow as it tries to handle its housekeeping
> tasks while writing your data.
I got into an argument recently with a (silly) friend who thinks
rotating media is old fashioned and far too slow and SSDs have come down
in price, I should buy an SSD for backup!
No, I am sure you would not be that silly, but my response still holds:
I use these disks for off-line backups. Their prime job is to sit there
most of the time doing nothing, unused. When I do one of my encrypted
backups I think it takes 20-minutes minimum, (if only a few days worth)
sometimes several times longer (if it has been weeks). So what? I'm
likely off doing something else (yes, some risk of version skew between
different files during the backup if I am using my computer while the
backup is happening).
Also, realize that this technique writes whole files or makes a hard
link to an old file. Yes, writing new directory structures is little
writes, but when I tail -f the log file, that's not what seems to take
time. Writing big files takes time.
If I had a lot of data streaming in constantly, my solution would be
completely different. If I had database files, my solution would be
completely different. But this is just me, not doing much on my computer.
The ease and speed of random access to the link-dest backup contents is
so nice—particularly if I might be frantic over some other failure. Each
backup is its own tree, intact and complete. Want a newer version of
some file, look at a newer tree. Want older, look in an older tree. No
"How do I access the backup again?" The technique to access the backup
is largely the same as doing a new backup: plugin, enter key. For the
backup case, cd into the disk and run my script. For the restore case cd
deeper, to whatever data I want. My custom script is needed to make a
backup, but no specialized software is needed to access the backup
beyond a reasonably recent Linux machine with an appropriate available
USB port.
-kb
More information about the Discuss
mailing list