>
> > I would argue that file system stability is paramount for backups and
> > performance is really a distant second.  who cares how fast the system
>
> If you can't perform your backups in a 24-hour period, performance
> becomes more important.
>
> if you backup data just to lose it, performance is never even a factor.
Performance needs can be measured.  I need to get this much done in this
amount of time.  reliability is a need.  a 99% reliability is the equivalent
of saying its ok to lose data 1 out of every 100 days.  So performance is
almost completely irrelevant if you dont have reliable storage.  of coarse
it makes sense to identify your data safety need and then buy/build to
perform as well as possible for your workload while achieving your
reliability needs.  I personally think through some experimenting of my own
that JFS is not an ideal filesystem on *linux*.  So even if JFS is very
fast, I dont trust it with my data so it wont get used.  I would chose XFS
if it were second place because it is reliable.  stability and security
first, then performance in this specific environment.

On my desktop I use ext4 and formerly used reiserfs.  ext4 is your and
reiserfs doesnt like to get killed hard but I keep backups of my data and
the loss of the system is a minor inconvenience while my backup server MUST
be stable and I wouldnt dare use ext4(yet) or reiserfs on it.  ext3 may be
slow but it is bulletproof.


> > will be substantially slower than raid10 or raid1 that have no parity
> > calculation.
>
> He is using a hardware RAID card.
>
 By hardware raid I really mean dedicated processor raid controller.  I dont
know the 3ware model and dont know its capabilities.  3ware makes some cards
that are little more than software raid with a physical interface.  Im
thinking higher end Adaptec or LSI that has a processor that does that
instead of handing it off to the cpu.


> > That being said, if raid5(or6) is fast enough for you it is a mature and
> > stable option and a good choice, but certainly comes with a performance
> > penalty.
>
> You make it sounds like RAID-5 is incapable of saturating drive
> bandwidth.  I haven't seen this on any modern (2-yr-old or newer)
> machine with more than one CPU.  And he's using hardware RAID so the
> point is moot anyway.
>
sure raid5 can saturate the bus.  specifically with backuppc latency is the
issue, not bandwidth.  the latency added by the parity calculation
negatively effects IO performance.  SSD drives are the best example.  they
can have lower, sometimes much lower, bandwidth than a HD but the
drastically lower latency translates directly to massively improved IO.
backuppc lives on hardlinks which are tiny writes where bandwidth is of
little importance and IO performance owns the day.  bandwidth is important,
but backuppc really likes IO performance.



>
> > examples:
> > raid10 with 6 drives in a raid0(r1-1+2, r1-3+4, r1-5+6), is 3 active
> > spindles because the other three are mirrors but has a worst case safety
> > of just 1 drive
> > raid10 in raid0(raid1-1+2+3, raid1-4+5+6) is just 2 spindles but is more
> > resilient because you can loose 2 or more drives and keep the array up.
>
> I think you have it backwards.  A stripe of three mirrors (your first
> example) means that you could lose up to three drives and still have
> data (as long as you're not losing both drives in a mirror).
>
> by worst case, you could lose just one drive but best case 3 drives.  if
you lose drive 1 and 2 the array is dead.


> > These are round numbers, kind of a rule of thumb.  6 volumes is about
> > where raid5 actually catches up.  with a 4 drive set the raid5 penalty
> > brings is to 2 active spindles and has a large 33% latency penalty
> > because the array has to wait for all writes to complete while a raid10
> > is 2 active spindles without a latency hit.
>
> I think you're overestimating the performance penalty of latency.  This
> is worked around in most systems with caching (including a large amount
> of RAM on hardware controller cards), and only for writes.
>
I would agree with you on a lot of workloads, but not backuppc.  creating
hardlinks at the level of backuppc is very IO intensive.  For bandwidth
intensive servers or desktops the raid5(or better 6) is likely very close in
overall speed and offers more storage for the money/drive count.  backuppc
is a very unusual workload because of the hardlinks and that is why latency
is so much more important than in the vast majority of other workloads.
------------------------------------------------------------------------------
Come build with us! The BlackBerry® Developer Conference in SF, CA
is the only developer event you need to attend this year. Jumpstart your
developing skills, take BlackBerry mobile applications to market and stay 
ahead of the curve. Join us from November 9-12, 2009. Register now!
http://p.sf.net/sfu/devconf
_______________________________________________
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:    https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:    http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/

Reply via email to