ZFS...
Walter Cramer
wfc at mintsol.com
Tue Apr 30 16:08:03 UTC 2019
Brief "Old Man" summary/perspective here...
Computers and hard drives are complex, sensitive physical things. They,
or the data on them, can be lost to fire, flood, lightning strikes, theft,
transportation screw-ups, and more. Mass data corruption by faulty
hardware or software is mostly rare, but does happen. Then there's the
users - authorized or not - who are inept or malicious.
You can spent a fortune to make loss of the "live" data in your home
server / server room / data center very unlikely. Is that worth the time
and money? Depends on the business case. At any scale, it's best to have
a manager - who understands both computers and the bottom line - keep a
close eye on this.
"Real" protection from data loss means multiple off-site and generally
off-line backups. You could spend a fortune on that, too...but for your
use case (~21TB in an array that could hold ~39TB, and what sounds like a
"home power user" budget), I'd say to put together two "backup servers" -
cheap little (aka transportable) FreeBSD systems with, say 7x6GB HD's,
raidz1. With even a 1Gbit ethernet connection to your main system, savvy
use of (say) rsync (net/rsync in Ports), and the sort of "know your data /
divide & conquer" tactics that Karl mentions, you should be able to
complete initial backups (on both backup servers) in <1 month. After that
- rsync can generally do incremental backups far, far faster. How often
you gently haul the backup servers to/from your off-site location(s)
depends on a bunch of factors - backup frequency, cost of bandwidth, etc.
Never skimp on power supplies.
-Walter
[Credits: Nothing above is original. Others have already made most of my
points in this thread. It's pretty much all decades-old computer wisdom
in any case.]
On Tue, 30 Apr 2019, Michelle Sullivan wrote:
> Karl Denninger wrote:
>> On 4/30/2019 05:14, Michelle Sullivan wrote:
>>>> On 30 Apr 2019, at 19:50, Xin LI <delphij at gmail.com> wrote:
>>>>> On Tue, Apr 30, 2019 at 5:08 PM Michelle Sullivan <michelle at sorbs.net>
> wrote:
>>>>> but in my recent experience 2 issues colliding at the same time results
> in disaster
>>>> Do we know exactly what kind of corruption happen to your pool? If you
> see it twice in a row, it might suggest a software bug that should be
> investigated.
>>>>
>>>> All I know is itâs a checksum error on a meta slab (122) and from what
> I can gather itâs the spacemap that is corrupt... but I am no expert. I
> donât believe itâs a software fault as such, because this was cause by a
> hard outage (damaged UPSes) whilst resilvering a single (but completely
> failed) drive. ...and after the first outage a second occurred (same as the
> first but more damaging to the power hardware)... the host itself was not
> damaged nor were the drives or controller.
>> .....
>>>> Note that ZFS stores multiple copies of its essential metadata, and in my
> experience with my old, consumer grade crappy hardware (non-ECC RAM, with
> several faulty, single hard drive pool: bad enough to crash almost monthly
> and damages my data from time to time),
>>> This was a top end consumer grade mb with non ecc ram that had been
> running for 8+ years without fault (except for hard drive platter failures.).
> Uptime would have been years if it wasnât for patching.
>> Yuck.
>>
>> I'm sorry, but that may well be what nailed you.
>>
>> ECC is not just about the random cosmic ray. It also saves your bacon
>> when there are power glitches.
>
> No. Sorry no. If the data is only half to disk, ECC isn't going to save
> you at all... it's all about power on the drives to complete the write.
>>
>> Unfortunately however there is also cache memory on most modern hard
>> drives, most of the time (unless you explicitly shut it off) it's on for
>> write caching, and it'll nail you too. Oh, and it's never, in my
>> experience, ECC.
>
> No comment on that - you're right in the first part, I can't comment if
> there are drives with ECC.
>
>>
>> In addition, however, and this is something I learned a LONG time ago
>> (think Z-80 processors!) is that as in so many very important things
>> "two is one and one is none."
>>
>> In other words without a backup you WILL lose data eventually, and it
>> WILL be important.
>>
>> Raidz2 is very nice, but as the name implies it you have two
>> redundancies. If you take three errors, or if, God forbid, you *write*
>> a block that has a bad checksum in it because it got scrambled while in
>> RAM, you're dead if that happens in the wrong place.
>
> Or in my case you write part data therefore invalidating the checksum...
>>
>>> Yeah.. unlike UFS that has to get really really hosed to restore from
> backup with nothing recoverable it seems ZFS can get hosed where issues occur
> in just the wrong bit... but mostly it is recoverable (and my experience has
> been some nasty shit that always ended up being recoverable.)
>>>
>>> Michelle
>> Oh that is definitely NOT true.... again, from hard experience,
>> including (but not limited to) on FreeBSD.
>>
>> My experience is that ZFS is materially more-resilient but there is no
>> such thing as "can never be corrupted by any set of events."
>
> The latter part is true - and my blog and my current situation is not
> limited to or aimed at FreeBSD specifically, FreeBSD is my experience.
> The former part... it has been very resilient, but I think (based on
> this certain set of events) it is easily corruptible and I have just
> been lucky. You just have to hit a certain write to activate the issue,
> and whilst that write and issue might be very very difficult (read: hit
> and miss) to hit in normal every day scenarios it can and will
> eventually happen.
>
>> Backup
>> strategies for moderately large (e.g. many Terabytes) to very large
>> (e.g. Petabytes and beyond) get quite complex but they're also very
>> necessary.
>>
> and there in lies the problem. If you don't have a many 10's of
> thousands of dollars backup solutions, you're either:
>
> 1/ down for a looooong time.
> 2/ losing all data and starting again...
>
> ..and that's the problem... ufs you can recover most (in most
> situations) and providing the *data* is there uncorrupted by the fault
> you can get it all off with various tools even if it is a complete
> mess.... here I am with the data that is apparently ok, but the
> metadata is corrupt (and note: as I had stopped writing to the drive
> when it started resilvering the data - all of it - should be intact...
> even if a mess.)
>
> Michelle
>
> --
> Michelle Sullivan
> http://www.mhix.org/
>
> _______________________________________________
> freebsd-stable at freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscribe at freebsd.org"
>
More information about the freebsd-stable
mailing list