[Dovecot] RAID1+md concat+XFS as mailstorage

Charles Marcus CMarcus at Media-Brokers.com
Sun Jul 1 13:34:14 EEST 2012


On 2012-06-29 12:07 PM, Ed W <lists at wildgooses.com> wrote:
> On 29/06/2012 12:15, Charles Marcus wrote:
>> Depends on what you mean exactly by 'incorrect'...

> I'm sorry, this wasn't meant to be an attack on you,

No worries - it wasn't taken that way - I simply disagreed with the main 
point you were making, and still do. While I do agree there is some 
truth to the issue you have raised, I just don't see it as quite the 
disaster-in-waiting that you do. I have been running small RAID setups 
for quite a while, and while I had one older RAID5 (with NO hot spare) 
that I inherited (this was many years ago) that gave me fits for about a 
month once (had drives randomly 'failing', but a rebuild - which took a 
few HOURS, and this was with small (by today's standards - 120GB drives) 
would fix it, then another one would do drop out 2 or 3 days later, etc. 
I finally found an identical replacement controller on ebay (old 3ware 
card) and once it was replaced it fixed the problem). I also had one 
instance in a RAID10 setup I configured myself a few years ago where one 
of the pairs had some errors on an unclean shutdown (this was after 
about 3 years of 24/7 operation on a mail server) and went into 
automatic rebuild, which went smoothly (and was mucho faster than the 
RAID5 rebuilds were even though the drives were much bigger).

So, yes, while I acknowledge the risk, it is the risk we all run storing 
data on hard drives.

> I thought I was pointing out what is now fairly obvious stuff, but
> it's only recently that the maths has been popularised by the common
> blogs on the interwebs. Whilst I guess not everyone read the flurry
> of blog articles about this last year, I think it's due to be
> repeated in increasing frequency as we go forward:
>
> The most recent article which prompted all of the above is I think this
> one:
> http://queue.acm.org/detail.cfm?id=1670144
> More here (BARF = Battle Against Raid 5/4)
> http://www.miracleas.com/BAARF/

I'll find time to read these over the next week or two, thanks...

> Intel have a whitepaper which says:
>
> Intelligent RAID 6 Theory Overview And Implementation
>
> RAID 5 systems are commonly deployed for data protection in most
> business environments.

While maybe true many years ago, I don't think this is true today. I 
wouldn't touch RAID5 with a ten foot pole, but yes, maybe there are 
still people who use it for some reason - and maybe there are some 
corner cases where it is even desirable?

> However, RAID 5 systems only tolerate a single drive failure, and the
> probability of encountering latent defects [i.e. UREs, among other
> problems] of drives approaches 100 percent as disk capacity and array
> width increase.

Well, this is definitely true, but I wouldn't touch RAID5 today.

> And to be clear - RAID5/RAID1 has a very significant probability that
> once your first disk has failed, in the process of replacing that disk
> you will discover an unrecoverable error on your remaining drive and
> hence you have lost some data...

Well, this is true, but the part of your comment that I was responding 
to and challenging was that the entire RAID just 'died' and you lost ALL 
of your data.

That is simply not true on modern systems.

>>> So the vulnerability is not the first failed disk, but discovering
>>> subsequent problems during the rebuild.

>> True, but this applies to every RAID mode (RAID6 included).

> No, see RAID6 has a dramatically lower chance of this happening than
> RAID1/5. See this is the real insight and I think it's important that
> this fairly (obvious in retrospect) idea becomes widely known and
> understood to those who manage arrays.

> RAID6 needs a failed drive and *two* subsequent errors *per stripe* to
> lose data. RAID5/1 simply need one subsequent error *per array* to lose
> data. Quite a large difference!

Interesting... I'll look at this more closely then, thanks.

>> Also, one big disadvantage of RAID5/6 is the rebuild times

> Hmm, at least theoretically both need a full linear read of the other
> disks. The time for an idle array should be similar in both cases. Agree
> though that for an active array the raid5/6 generally causes more drives
> to read/write, hence yes, the impact is probably greater.

No 'probably' to it. It is definitely greater, even comparing  the 
smallest possible RAID setups (4 drives are minimum for each). But, as 
the size of (number of disks in) the array increases, the difference 
increases dramatically. With RAID10, when a drive fails and a rebuild 
occurs, only ONE drive must be read (remirrored) - in a RAID5/6, most if 
not *all* of the drives must be read from (depends on how it is 
configured I guess).

> However, don't miss the big picture, your risk is a second error
> occurring anywhere on the array with raid1/5, but with raid 6 your risk
> is *two* errors per stripe, ie you can fail a whole second drive and
> still continue rebuilding with raid6

And is the same with a RAID10, as long as the second drive failure isn't 
the one currently being remirrored.

I think you have proven your case that a RAID6 is statistically a little 
less likely to suffer a catastrophic cascading disk failure scenario 
than RAID10.

> I personally feel that raid arrays *are* very fragile. Backups are often
> the option when you get multi-drive failures (even if theoretically the
> array is repairable). However, it's about the best option we have right
> now, so all we can do is be aware of the limitations...

And since backups are stored on drives (well, mine are, I stopped using 
tape long ago), they have the same associated risks... but of course I 
agree with you that they are absolutely essential.

> Additionally I have very much suffered this situation of a failing RAID5
> which was somehow hanging together with just the odd uncorrectable read
> error reported here and there (once a month say). I copied off all the
> data and then as an experiment replaced one disk in this otherwise
> working array, which then triggered a cascade of discovered errors all
> over the disk and rebuilding was basically impossible.

Sounds like you had a bad controller to me... and yes, when a controller 
goes bad, lots of weirdness and 'very bad things' can occur.

> Roll on btrfs I say...

+1000 ;)

-- 

Best regards,

Charles



More information about the dovecot mailing list