Unable to mount hammer file system Undo failed
Wojciech Puchar
wojtek at wojtek.tensor.gdynia.pl
Thu Jul 19 07:50:27 PDT 2012
>
> OK, understood now, i think: you agree with temporarily loosing a bit of
> unreclaimed free-space on disk until time permits cleaning things up
> properly, afaiu softupdates (+journalling ? not really clear).
That it. And that's how original softupdates document describe it.
You may run quite safely without fsck, just not abuse that feature for too
long!
No journalling. I am currently FreeBSD user, FreeBSD 9 added softupdates
journalling, but REALLY it doesn't change much except extra writes on disk.
I found that you actually have to run full fsck now and then even with
journalling. In theory it shouldn't find any inconsistences, in practice
it always find minor ones.
As to end that topic my practices are:
- do not make huge filesystems or create large "RAID" arrays.
2 disk, one mirror from them, one filesystem.
- it takes like 30 minutes or less to fsck it, and the same time for 10
such filesystems as it can go in parallel.
in case of crash i do fsck manually when pause isn't a problem.
at reboot i check only root filesystem, and (if it's separate) /usr, so i
could execute all other checks remotely without rebooting.
>> Assuming hardware never fails is certainly wrong
>
> And there's no practical point assuming it *always* fails, is there ?
Just that it fails sometimes is enough to assume it can.
>>> could any FS be trustworthy then ??? IMHO, that's nonsense.
>> No it isn't. Sorry if i wasn't clear enough to explain it.
>
> Well, if the thing that you try to defend against is plain hardware failure
> (memory bits flipping, CPU going mad, whatever), i just doubt that any kind
> of software layer could definitely solve it (checksums of checksums of? i/o
You are completely right.
What i point out that flat data layout makes chance of recovery far higher
and chance of bad destruction far lower.
Any Tree-like structure produces a huge risk of losing much more data that
was corrupted at first place.
That rule already prove true for UFS filesystem, as well for eg. fixed
size database tables like .DBF format which i still use, not "modern"
ones.
Using DBF files as example - you have indexes in separate files, but
indexes are not crucial and can be rebuild.
So if any tree like structure (or hash type or whatever) would be invented
to speed up filesystem access - great. But only as extra index, with
crucial data (INODES!!) written as flat fixed record table at known place.
I don't say HAMMER is bad - contrary to ZFS which is 100% PURE SHIT(R) -
but i don't agree it is (or will be) a total replacement of older UFS.
Hammer pseudo-filesystems, snapshots and on like replication are useful
features but actually not that needed for everyone, and not without cost
of extra complexity. No matter how smart Matthew Dillon is, it still be
far more complex, and more risky.
That's why it's not good that swapcache doesn't support efficient caching
of UFS as there are no vfs.ufs.double_buffer feature just like hammer.
-----------------------------------------------------------------
Disclaimer ;): None of my practices, my ideas about safe filesystems,
mirroring or anything else are not replacement of proper backup
strategy!!! Please do not interpret anything i write about ideas against
backups.
More information about the Users
mailing list