Originally posted by caligula
View Post
Announcement
Collapse
No announcement yet.
Zstd 1.5.1 Released With Even More Performance Improvements
Collapse
X
-
Originally posted by timofonic View PostI never use Facebook nor Instagram or any of their products and consider them very evil (and fatal for teenagers), but Zstd is their only good product.
Btrfs is a total failure.
Congratulations, Facebook = META. You did something good, finally
I can also boot btrfs root snapshots from GRUB which is very useful - no more booting from USB to fix issues.
The only issue I've had in 6-7 years on btrfs was due to dropbox messing up a /home subvol & filling up the disk (the only time I ever did not have a separate /home partition)
- Likes 2
Comment
-
Originally posted by useless View PostIf anything, having automatic and trouble-free checksumming is an understatement.
Firstly, if you need a filesystem to do checksums then only because the drives and the bus fail at doing so. This is not impossible, but a drive not reporting a corruption, but the filesystem detecting one is extremely rare. It is more likely for the filesystem to have a bug, or to have lost its checksum due to a RAM corruption, and as a result reports a false positive. Do keep in mind a checksum is not evidence of corrupted data, but only of a mismatch between checksum and data, and evidence of only a possibility of a corruption. An event where a drive's error detection has failed and did not discover a corruption, but the filesystem does, is the extreme case. It does not matter if this is BTRFS, EXT4, F2FS or any other filesystem. It is more likely for the data on the drive to still be correct.
Secondly, putting a checksum on all your data means you are giving everything the same priority, from the most useless bits in some long-forgotten file to the content of /etc/shadow, with the result that a single, meaningless bit flip makes you halt your system, reinstall the file, and question your entire hardware. This is not what every user wants. Many are happy with letting bits-flip until it actually causes a problem, and not only a checksum error. Many users do not wish to give in to a paranoia of a data corruption always leading to the worst-case scenario. It just is not always the most important bits on your system that get corrupted. Nor do these people suffer false positives. Only a few people need this much checksumming and they know what they are doing.
Thirdly, of course, I do run backups, but again do I have no use for checksums here, but only for the actual backup data. No checksum is going to replace the data and even if I did use checksums here would I still look at the data itself to see if it actually is corrupt. Again, checksums do not fix data losses, only redundancy does, and whether data is still usable is not determined by a checksum but by the software that uses the data, which I will have to test anyway, with or without there being a checksum.
For the rare cases where I do want a checksum for a file, i.e. for security reasons, do I use SHA256 and better, but this I admit is also more based on paranoia and perhaps the need to be an over-achiever than on actually having a good reason for it. To me does a checksum literally only add a check into my workflow, thereby only adding to my work, but not removing any. Thus do I prefer redundancy and only enable checksums where these are useful to me. To say having an automatic and trouble-free checksumming just makes little sense to me. It better be automatic, because I sure do not want to calculate the checksum myself, and trouble-free just does not exist. Once you have a mismatch do you also have trouble.Last edited by sdack; 21 December 2021, 03:12 PM.
- Likes 1
Comment
-
Originally posted by sdack View PostIs it though?
Firstly, if you need a filesystem to do checksums then only because the drives and the bus fail at doing so. This is not impossible, but a drive not reporting a corruption, but the filesystem detecting one is extremely rare.
Originally posted by sdack View PostIt is more likely for the filesystem to have a bug, or to have lost its checksum due to a RAM corruption, and as a result reports a false positive.
Originally posted by sdack View PostDo keep in mind a checksum is not evidence of corrupted data, but only of a mismatch between checksum and data, and evidence of only a possibility of a corruption.
Originally posted by sdack View PostAn event where a drive's error detection has failed and did not discover a corruption, but the filesystem does, is the extreme case. It does not matter if this is BTRFS, EXT4, F2FS or any other filesystem. It is more likely for the data on the drive to still be correct.
Originally posted by sdack View PostSecondly, putting a checksum on all your data means you are giving everything the same priority, from the most useless bits in some long-forgotten file to the content of /etc/shadow, with the result that a single, meaningless bit flip makes you halt your system, reinstall the file, and question your entire hardware.
Originally posted by sdack View PostThis is not what every user wants. Many are happy with letting bits-flip until it actually causes a problem, and not only a checksum error. Many users do not wish to give in to a paranoia of a data corruption always leading to the worst-case scenario. It just is not always the most important bits on your system that get corrupted. Nor do these people suffer false positives. Only a few people need this much checksumming and they know what they are doing.
Originally posted by sdack View PostThirdly, of course, I do run backups, but again do I have no use for checksums here, but only for the actual backup data. No checksum is going to replace the data and even if I did use checksums here would I still look at the data itself to see if it actually is corrupt.
Originally posted by sdack View PostAgain, checksums do not fix data losses, only redundancy does, and whether data is still usable is not determined by a checksum but by the software that uses the data, which I will have to test anyway, with or without there being a checksum.
Originally posted by sdack View PostFor the rare cases where I do want a checksum for a file, i.e. for security reasons, do I use SHA256 and better, but this I admit is also more based on paranoia and perhaps the need to be an over-achiever than on actually having a good reason for it. To me does a checksum literally only add a check into my workflow, thereby only adding to my work, but not removing any.
Originally posted by sdack View PostThus do I prefer redundancy and only enable checksums where these are useful to me.
- Likes 2
Comment
-
Originally posted by DrYak View PostI would like to point out that it is also possible to periodically run both short (I do nightly) and long (used to be weekly on all drives, but now I only run them monthly on mechanical HDD, given that I have a weekly btrfs scrub) using the smartd service that comes with smarttools.
Thanks!
Comment
-
Originally posted by useless View PostA ha! You souldn't generalize the usefulness of some feature by looking at your specific workflow.
Comment
-
Originally posted by DrYak View PostAlthough I too run "btrfs scrub" periodically to detect such problems (and correct them if a usable RAID1 copy exists), I would like to point out that it is also possible to periodically run both short (I do nightly) and long (used to be weekly on all drives, but now I only run them monthly on mechanical HDD, given that I have a weekly btrfs scrub) using the smartd service that comes with smarttools.
Comment
-
Originally posted by sdack View PostThirdly, of course, I do run backups, but again do I have no use for checksums here, but only for the actual backup data. No checksum is going to replace the data and even if I did use checksums here would I still look at the data itself to see if it actually is corrupt. Again, checksums do not fix data losses, only redundancy does, and whether data is still usable is not determined by a checksum but by the software that uses the data, which I will have to test anyway, with or without there being a checksum.
- Likes 1
Comment
-
useless and sdack , I wonder if you two are arguing past each other somewhat. You seem to disagree about whether a checksum error is evidence of bad hardware and bit rot, and then everything after that you seem to be talking past each other rather than to each other. I really don't mean to criticize, because I appreciate and am learning from both of your responses.
- Likes 1
Comment
-
Originally posted by sdack View PostI did not say it was useless. I said one should not overrate it. I still think you do overrate its usefulness because you are not really focused on arguing on why you think you would need "automatic and trouble-free checksumming", but instead are you now only interested in what I wrote. This tells me you want to be right, but not that you are. Frankly, I think you like checksumming simply because it exists.
One simple example would be personal videos and photos: snapshot, backup, test once, scrub frequently. Any change will be evident without going photo after photo, video after video (you can't automate this process a lot more). New data? Repeat. You can test at longer intervals.
Cold storage is another example. You can sample and test them a lot less frequently than you can scrub.
Another one would be some system daemon misbehaving because corruption. With btrfs you will get EIO and prevent all kind of quirks (which can be important if that daemon touches important data).
Even more is the fact that drives have bugs. If your tests are throwing errors then you will search for the cause. What is it? Did your hardware crap itself or your app corrupted the data somehow? If you add filesystem checksumming then you know that you need to replace your drive or controller and you can rule out bugs in your app (your app has bugs too). What if there has been a change in you app and it can no longer open your backups for whatever reason? Do you discard your 'corrupted' backups?
In my last reply I gave you one scenario in which having checksum not only is helpful but necessary: your app can fail too. But you decided to ignore it. I don't have the need to be right, you issued some statements in that reply that was worth to argue about, nothing more.
All in all, trouble-free checksumming allows regular users to notice corruption and take action. If you have automatic scrubs (which you should) you can prevent more damage soon enough and lose less recently generated data.
- Likes 1
Comment
Comment