Re: need SSD RAID controller advice

2015-05-07 Thread ToddAndMargo

On 04/10/2015 05:21 PM, ToddAndMargo wrote:

Hi All,

I have a customer interested in high end workstation. She wants RAID 1
to protect her drives. She is also interested in Solid State Drives.
(And she knows the difference between RAID and Backup.)

What host controller would you guys recommend for a pair of Intel
SSDSC2BW480A401, 480 GB 2.5 Internal Solid State Drives in RAID 1?

Many thanks,
-T




Not to ask too stupid a question, but would a hybrid drive,
such as

http://www.seagate.com/internal-hard-drives/solid-state-hybrid/desktop-solid-state-hybrid-drive/?sku=ST2000DX001

not require TRIM and work with an hardware controller?

-T

--
~~
Computers are like air conditioners.
They malfunction when you open windows
~~


Re: need SSD RAID controller advice

2015-05-07 Thread Vladimir Mosgalin
Hi Kevin K!

 On 2015.05.06 at 18:52:42 -0500, Kevin K wrote next:

  On May 6, 2015, at 6:31 PM, Vladimir Mosgalin mosga...@vm10124.spb.edu 
  wrote:
  
  Alternative to TRIM-aided GC is Idle Time Garbage Collection (which you
  could see on Plextor), which actively scans data in background, finds
  empty blocks (zeroes) and marks them as free. Well probably more
  complicated than that, but it's the idea. 
 
 A quick search indicates that Idle TIme Garbage Collection is just another 
 term for garbage collection.  And I don’t see how it can work as well without 
 TRIM as with TRIM, unless massively over provisioned.  There is still the 
 issue that, at the file system level, you may now that you just deleted a 1GB 
 file that fully fills a whole lot of EBs.  Without trim, the controller can’t 
 know that the blocks aren’t in use until the OS attempts to write to those 
 blocks.  if the controller needs to make space, it will have to copy the 
 stale data that it doesn’t know isn’t used anymore.  Now, a continous write 
 to a whole EB’s worth of data (still stored in DRAM on the drive), it can see 
 that the whole block can be marked as unused, and freed later.  

I've posted the graph, you've seen it. You can find similar graphs on
other sites exploring the matter (anandtech, tom's hardware and so on).
There are drives which restore performance to original or near original
without aid of TRIM. They are consumer-grade with typical 7-15%
over provisioning, just like any other drive. Not enterprise models with
40-50% over provisioning.

If you want graph with more drives there is one here
http://www.xbitlabs.com/articles/storage/display/plextor-m5-pro-m5s_6.html
You can clearly see that some drives can't restore performance without
TRIM, and some restore most of it (up to 100% in case of M5P, which
isn't here) after some idle even without TRIM.

Of course, massive over provisioning is still a great feature of
enteprise drives. There is performance degradation on any kind of drive
even with the use of TRIM before GC is performed (explored here in detail:
http://www.tomshardware.com/reviews/m5s-256gb-ssd-benchmark,3252-7.html)
However it isn't so on massively over provisioned drives (Anandtech has
excellent graphs of this in articles about enterprise SSD drives, how
over provisioning lets drive to show high write IOPS numbers at any
moment).

Here is another example of drive restoring most of its performance
without TRIM:
http://www.anandtech.com/show/6337/samsung-ssd-840-250gb-review/11
(as you can see, it behaves in pretty complicated way, however it's
clear that GC works very well even without TRIM).

 
 I’m not sure what the zero blocks you are referring to means.  If you write a 
 whole bunch of zeros, is it marked as unused?  Even if contained in a file?  
 You really can’t mark obsolete data as zeros since you can’t rewrite until 
 you erase first.

I think I used the wrong wording. I'm sure it would mark zeroes as
unused (yes, even in a file - it just erases these blocks and unmaps
them, when you read from unmapped LBA the drive returns you zeroes), but
generally these algorithms are more complicated. Writing zeroes isn't
recommended (http://www.lowlevelformat.info/zero-filling.aspx) - it
might work for some SSDs, of course, but still risky.

I don't really know the details of how TRIM-less GC works in complicated
cases, but it's obvious from performance graphs that it really does work
on certain drivers and doesn't work too good on others.

Regarding the term Idle Time GC - yes it is used for describing any
kind of background GC, whether TRIM-aided or not by some manufacturers.
There is nothing I can do about that. I used it the way it's often used on
some hardware sites and forums, but there are manufacturers who use this
term in a different way...

-- 

Vladimir


Re: need SSD RAID controller advice

2015-05-07 Thread David Sommerseth
On 07/05/15 03:51, Jeff Siddall wrote:
 On 05/06/2015 09:33 PM, ToddAndMargo wrote:
 Thank you.  I will have to get over my phobia of software
 RAID.
 
 FWIW I have used SW RAID1 for years (maybe decades by now) on all sorts
 of systems.  It has saved me many times and never burned me.
 
 If you are going to do anything in SW RAID1 is as easy/good as it gets.

+1 ... Same experience here too.

I've had mdadm RAID setups with both RAID1 and RAID5 where harddrive
failures have happened.  Recovered without any issues and no data loss.

Performance wise ... I've not done anything testing in particular.  But
none of my setups have felt sluggish on normal usage.


David S.


Re: need SSD RAID controller advice

2015-05-07 Thread Vladimir Mosgalin
Hi ToddAndMargo!

 On 2015.05.06 at 23:45:41 -0700, ToddAndMargo wrote next:

 Many thanks,
 -T
 
 
 
 Not to ask too stupid a question, but would a hybrid drive,
 such as
 
 http://www.seagate.com/internal-hard-drives/solid-state-hybrid/desktop-solid-state-hybrid-drive/?sku=ST2000DX001
 
 not require TRIM and work with an hardware controller?

I don't recommend using these at all, if you check the reviews, their
performance isn't that much different from regular hard drive in many
use cases.

They make sense for notebooks but hardly for desktop.

Regarding your question, it doesn't make much sense because you aren't
going to get the kind of performance you get from SSD with degraded
performance from this drive. For OS, it's regular hard drive (without
TRIM support) with extra cache to speed up some operations.

-- 

Vladimir


Re: need SSD RAID controller advice

2015-05-06 Thread ToddAndMargo

On 04/14/2015 02:29 AM, Vladimir Mosgalin wrote:

Hi ToddAndMargo!

  On 2015.04.13 at 18:26:49 -0700, ToddAndMargo wrote next:


-T

Wonder if there are any enterprise level SSD's that don't
need TRIM?


Tons. All Hitachi SSDs, OCZ Interpid and Saber, all Intel DC series SSDs
(based on Intel controllers) and many more. Though I'm not sure that
1) you'll like the prices of enteprise SSDs and 2) for workstation (and
even for many kinds of servers) there is absolutely no point in using
these enterprise SSDs. The controller itself isn't more reliable, just
flash chips endure more wear, but the difference (e.g. 100 TB endurance
of normal SSD vs 20 PB endurance of enterprise SSD) makes a difference
only for certain databases and cache (when data on SSD is constantly
refreshed) workloads. So if you exclude the wear factor, the chance of
failure depends on the controller/firmware and is about the same whether
SSD was marked for enterprise or not. And you already got RAID in mind
against these kind of failures anyway.

For nearly all other cases, avoiding SandForce-based SSDs is about all you
need to care about when picking SSD for systems without TRIM. I listed
some model examples in the previous mail.




Hi Vladimir,

I have look all over Intel's specs for their new SSD DC S3510
series SSDs (older drives are harder to find):

http://www.intel.com/content/www/us/en/solid-state-drives/ssd-dc-s3510-spec.html

They do not say a thing about Garbage Collection and not having to use 
TRIM.  They do say Data set management Trim command, which means they

support TRIM.

I have eMailed Intel's tech support twice over two weeks and
have been ignored twice by them (which is typical).

Also: in my research, I did find mention that TRIM did not effect
lifespan -- TRIM only affected write speed, which would slow down
to a crawl without it.  Can you verify this?

Thank you for helping me with this.

-T


--
~~
Computers are like air conditioners.
They malfunction when you open windows
~~


Re: need SSD RAID controller advice

2015-05-06 Thread ToddAndMargo

On 05/06/2015 04:31 PM, Vladimir Mosgalin wrote:

Hi ToddAndMargo!

  On 2015.05.05 at 23:05:36 -0700, ToddAndMargo wrote next:


Hi Vladimir,

I have look all over Intel's specs for their new SSD DC S3510
series SSDs (older drives are harder to find):

http://www.intel.com/content/www/us/en/solid-state-drives/ssd-dc-s3510-spec.html

They do not say a thing about Garbage Collection and not having to use TRIM.
They do say Data set management Trim command, which means they
support TRIM.


Well, every modern SSD (or even eMMC device) supports TRIM.

I looked at review of this SSD - it's somewhat strange, unlike Intel
SSDs of previous generation it doesn't support Idle Garbage Collection.
So, indeed, TRIM support would be very useful.

Take a look at this graph
http://img.fcenter.ru/imgmat/article/SSD/Intel_DC_S3500/207103.png

You can see performance degradation right after writes (20-40% of
original performance typically - note that this is pretty heavy test,
amount of data written is twice the size of SSD). This is expected. Now,
there are drives that support Idle Time Garbage Collection. Among these,
Plextor M5 are most amazing - they restore to original performance
without even need of TRIM at all. As for the other drives, after TRIM
they restore performance. There are, however, SandForce controllers
(Intel 530 in this example), where performance not just degrades a lot
without TRIM, it doesn't restore to original values with TRIM (though,
mind you, this is pretty heavy test, in real life this isn't that
likely).


Also: in my research, I did find mention that TRIM did not effect
lifespan -- TRIM only affected write speed, which would slow down
to a crawl without it.  Can you verify this?


TRIM does indirectly affects lifespan: the write speed degradation
doesn't appear out of thin air just because you haven't TRIMMED it. It
happens because free space becomes fragmented and SSD hardly has any
free space left (in its opinion, because without TRIM the OS doesn't
report which blocks are actually free and can be reused). But data on
SSD can't be written (or, rather, erased) in small blocks, so in order
to write new data on heavy fragmented device SSD controller has to move
data (read from half-filled blocks and rearrange, writing to different
place, to free whole blocks for new data). E.g. SSD might be able to
read 4KB blocks but erase only 512KB blocks.

Because of that, when SSD is used without TRIM for a long time, this
process greatly increases amount of written data - you write 100MB, but
SSD had to rearrange data to free bigger blocks and internally had
written 300MB. This is called write amplification and it decreases
lifespan of SSD much quicker. Also it makes your writes slow (as seen on
graph) because at the time you think you've written 100MB SSD had to
write 3 times more.

Proper TRIM usage makes SSD aware of all the real free blocks (all free
space in your filesystem), so it doesn't have to move data anymore and
just writes new data there. It reduces write amplification coefficient -
and thus prolongs lifespan. Not that I would worry about SSD lifespan
even with huge write amplification, mind you.  IMHO performance problem
is more important.

It doesn't TRIM itself that affects write speed or lifespan, it only
aids Garbage Collection algorithms of SSD controller. Just keep in mind
that getting bad linear write performance usually means big write
amplification because SSD doesn't have much free blocks.

Alternative to TRIM-aided GC is Idle Time Garbage Collection (which you
could see on Plextor), which actively scans data in background, finds
empty blocks (zeroes) and marks them as free. Well probably more
complicated than that, but it's the idea. So given some idle time (SSD
runs this type of GC only when it isn't too busy serving user requests,
obviously), it is possible to free blocks for future writings even
without OS reporting them using TRIM.

There are other alternatives, e.g. SandForce controllers internally
compress data to write less, you can say that they might have less-than
1 write amplification coefficient. However, this doesn't help when
incompressible data is written.

I'm sorry about initial confusion suggesting that Intel DC drive works
well without TRIM; I knew that DC drives use good controller and it will
give you much better experience than SF-based drive but I didn't
check that they actually have Idle GC. Previous generations Intel drives
did, but not these. S3500 relies on TRIM for GC.




Hi Vladimir,

Thank you for all the time you have spend with me on this.
Seems to me that I either get over my phobia of software
RAID of just get the most reliable SSD I can find and talk
the customer out of RAID.

-T


--
~~
Computers are like air conditioners.
They malfunction when you open windows
~~


Re: need SSD RAID controller advice

2015-05-06 Thread Kevin K

 On May 6, 2015, at 6:31 PM, Vladimir Mosgalin mosga...@vm10124.spb.edu 
 wrote:
 
 Alternative to TRIM-aided GC is Idle Time Garbage Collection (which you
 could see on Plextor), which actively scans data in background, finds
 empty blocks (zeroes) and marks them as free. Well probably more
 complicated than that, but it's the idea. 

A quick search indicates that Idle TIme Garbage Collection is just another term 
for garbage collection.  And I don’t see how it can work as well without TRIM 
as with TRIM, unless massively over provisioned.  There is still the issue 
that, at the file system level, you may now that you just deleted a 1GB file 
that fully fills a whole lot of EBs.  Without trim, the controller can’t know 
that the blocks aren’t in use until the OS attempts to write to those blocks.  
if the controller needs to make space, it will have to copy the stale data that 
it doesn’t know isn’t used anymore.  Now, a continous write to a whole EB’s 
worth of data (still stored in DRAM on the drive), it can see that the whole 
block can be marked as unused, and freed later.  

I’m not sure what the zero blocks you are referring to means.  If you write a 
whole bunch of zeros, is it marked as unused?  Even if contained in a file?  
You really can’t mark obsolete data as zeros since you can’t rewrite until you 
erase first.

Re: need SSD RAID controller advice

2015-05-06 Thread ToddAndMargo

On 05/06/2015 04:32 PM, Vladimir Mosgalin wrote:

Hi ToddAndMargo!

  On 2015.05.06 at 00:48:30 -0700, ToddAndMargo wrote next:


http://www.intel.ie/content/dam/www/public/us/en/documents/white-papers/ssd-dc-s3500-workload-raid-paper.pdf


  The system used for RAID 1 testing include the following:
  • LSI MegaRAID 9265-8i* controller card

  To summarize:
  • In both RAID 1 and RAID 5, the Intel SSD DC S3500 Series
drive shows excellent scalability, performance, and consistency.

  • Very little latency was introduced by the RAID controller
in RAID 1. In RAID 5, the overhead and latency are slightly
 higher.

  • In random, mixed read/write workloads, SSDs perform
significantly (as much as 100 times) better than HDDs
in a similar situation


Maybe the S3500 is what I want?  They are still available
at my distributors.


They are fine drives. In fact, I got servers running them; excellent
performance, no problem with incompressible data (unlike drives with SF
controllers) - this matters to me because I'm using zfs on linux with
compression. Sustained write speed 500 MB/s (480GB model).

However, like I said before, if long-term performance is needed,


It is, but it will be in a high end workstation.  99% of the
writes will be OS related and not user related.


I
advise against using them in hardware RAID controller. This document
only shows after-install performance; they didn't do any endurance
testing. Without TRIM, performance of this drive will degrade.
Buy Plextor M5/M6 if your environment doesn't support TRIM.


I'm curious, what stops you from using software RAID with mdadm?
For two-disk RAID 1 it provides excellent performance, high
compatibility and TRIM support. I could understand if you needed
hardware controller for RAID6 with tons of disks, in RAID6 or something,
but for two-disk RAID1?..

--

Vladimir



Hi Vladimir,

I am allergic to software RAID.  I adore the way an independent
controller works.

Thank you for helping me with this,

-T

--
~~
Computers are like air conditioners.
They malfunction when you open windows
~~


Re: need SSD RAID controller advice

2015-05-06 Thread Konstantin Olchanski
On Wed, May 06, 2015 at 04:46:27PM -0700, ToddAndMargo wrote:

 ... that I ... get the most reliable SSD I can find and talk
 the customer out of RAID.
 

This is bad thinking. most reliable does not mean will never fail.

(leaving aside the question on how you can tell which brand is more
reliable, other than by vendor rosy promises or by counting of stars at newegg).

If you are building a system for a customer, you have to have a reasonable
answer to the question if this SSD fails, do I lose all of everything?.

This means you have to have backups of everything and well tested instructions
on how to restore the full working system from these backups.

In my experience, mdadm RAID1 is the simplest way to build a system
that survives single device failure (SSD or HDD, does not matter),
but you still have to have backups and restoration instructions
because RAID1 does not protect against filesystem corruption,
against accidentally or maliciously deleted or modified files, etc.

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-17 Thread Konstantin Olchanski
On Mon, Apr 13, 2015 at 03:28:39PM -0700, jdow wrote:
 The 3Ware RAID cards I have vastly outstrip the motherboard built in
 Intel RAID implementations for a RAID 5 setup. (I don't consider
 RAID 1 to be economically sensible for most uses.) A four disk RAID
 5 SSD configuration can be breathtaking fast, too.

In my experience, all raid cards have rather limited total bandwidth.

For example-
- consider 6TB disks, each can read data at 160 Mbytes/sec when connected to an 
on-board (intel) SATA port.
- connect 8 of them to an 8-port RAID card.
- sure enough, through the raid card you can read each disk at this speed.
- now try to read all disks at the same time (8 programs reading from 8 disks)
- what I see is the grand total data coming in is around 500-600-700 Mbytes/sec 
instead of 8*160=1280 Mbytes/sec.
- (observe the same with writing instead of reading)
- what I have measured is the raid card internal bandwidth (which includes the 
PCIe bandwidth,
  you should better be connected at PCIe x4 or better x8 or better x16 link 
lanes).
- some vendors are honest and this bandwidth is written down in spec sheets 
(making it
  abvious that the raid cards were designed in the days of 1TB disks that could 
never
  reach 100 Mbytes/sec read or write).
- now if you put RAID6 on top of this, the rate goes down again.
- if instead of software RAID6 you use hardware RAID6, maybe you gain some 
bandwidth back,
  I have no data for this.
- if instead, I connect all the disks to onboard ports (no raid card), using 
the sadly
  discontinued ASUS Z87 mobo with 10 SATA ports (6 intel + 4 marvell), I do see
  total bandwidths in the 1000 Mbytes/sec range (those measurements were done 
with 8x4TB disks).

All this proves is that intel SATA ports on the intel internal interconnect
are much faster than SATA ports connected to a not-super-fast microprocessor
on a raid card connected to the system by an x4 or x8 PCIe bus (have not see 
x16 raid cards yet).

Speaking of a theoretical 4xSSD hardware RAID5 configuration, I would be 
surprised
if it reaches 1000 Mbytes/sec read/write speeds (assuming 500 Mbytes/sec SSDs).
Software RAID5 with 4xSSDs connected to onboard SATA (with a 4GHz CPU)
would probably get all the way to 2000 Mbytes/sec (assuming 500 Mbytes/sec 
SSDs).

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-14 Thread Ian Murray

 From: Nico Kadel-Garcia nka...@gmail.com
To: Nico Kadel-Garcia nka...@gmail.com; James M. Pulver 
jmp...@cornell.edu; scientific-linux-users@fnal.gov 
scientific-linux-users@fnal.gov 
Sent: Tuesday, 14 April 2015, 3:02
Subject: Re: need SSD RAID controller advice
 

On Mon, Apr 13, 2015 at 8:16 PM, Konstantin Olchanski




olcha...@triumf.ca wrote:
 On Mon, Apr 13, 2015 at 07:29:18PM -0400, Nico Kadel-Garcia wrote:
  I am not sure what you refer to. With SL5 and SL6 you have 2 disks,
  put / on a software RAID1 partitions put grub on both disks ...

 Until it fails. Grubby, along with the kernel installlation RPM,
 doesn't know how to update, manage, or keep synchronized this second
 /boot partition. Hilarity can then ensue, along with making sure that
 your /etc/fstab doesn't detect the wrong disk and mount it incorrectly
 as /dev/sda. See, if your first disk dies, unless you're very cautious
 with /etc/fstab, it's very much a crapshoot if hte right partition
 will mount as /boot.

 Been there, done that, gave up on the silliness.


 You talk about /boot, /dev/sda, grubby, etc, I doubt you have been there, 
 done that. Gave up, sure.

 With mirrored disks, there is only /dev/md0 (/), /dev/md1 (swap), etc (if 
 you have /home, /data, etc),
 in /etc/fstab, / is mounted by UUID of the /dev/md0 RAID1 device,
 which in turn is assembled by UUID of component disks using mdadm magic.

Been there, done that. Don't have to make this stuff up, I spent some
time *designing* Linux based storage servers, but it was more than 5
years ago.

However, software RAID may have improved to the point where /boot
doesn't have to be its own partition. So I just tried it on a VM
runing SL 6.6, and got unhandled exception, and git it to work
flawlessly when I made /boot its own, non-RAID partition.. So I'm
not filled with confidence that putting /boot on its partition is not

still necessary.


Are you finding that it actually works? Would you please post an
/etc/fstab from a working system to help verify that it works?



More cutting edge linux distributions such as Ubuntu can have /boot as part of 
a RAID system, and even on an LVM. I think this is a feature of GRUB2. If you 
wish to RAID /boot on older dists (AFAIR this applies to SL5 and SL6), you can 
RAID it but it must be mirrored. AFAIK GRUB doesn't understand the RAID but it 
looks sufficiently like a regular partition/filesystem that it is able to find 
what it needs. I don't have a fstab to show you from an older config as I 
migrated to Ubuntu Server for my Xen machines. I thoroughly tested both these 
configurations and the machine is able to fully boot with any given drive 
removed.

Here is an Ubuntu one, plus supporting information, FWIW...


# cat /etc/fstab
# /etc/fstab: static file system information.
#
# Use 'blkid' to print the universally unique identifier for a
# device; this may be used with UUID= as a more robust way to name devices
# that works even if disks are added and removed. See fstab(5).
#
# file system mount point   type  options   dump  pass
proc/proc   procnodev,noexec,nosuid 0   0
/dev/mapper/xen2-root /   ext4errors=remount-ro 0   1
/dev/mapper/xen2-swap noneswapsw  0   0



# cat /proc/mdstat 
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] 
[raid10] 
md0 : active raid5 sdc1[2] sda1[0] sdb1[1]
1953258496 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/3] [UUU]

unused devices: none
root@xen2:~# pvs
PV VG   Fmt  Attr PSize PFree 
/dev/md0   xen2 lvm2 a-   1.82t 53.18g

root@xen2:~# lvs
LV  VG   Attr   LSize   Origin Snap%  Move Log Copy%  Convert
--- snip ---
rootxen2 -wi-ao 279.39g 
swapxen2 -wi-ao   7.45g
--- snip ---










Re: need SSD RAID controller advice

2015-04-14 Thread David Sommerseth
On 11/04/15 02:21, ToddAndMargo wrote:
 Hi All,
 
 I have a customer interested in high end workstation. She wants RAID 1
 to protect her drives. She is also interested in Solid State Drives.
 (And she knows the difference between RAID and Backup.)
 
 What host controller would you guys recommend for a pair of Intel
 SSDSC2BW480A401, 480 GB 2.5 Internal Solid State Drives in RAID 1?

I've followed this discussion for a little while, and there are a lot of
good information here - so thanks to everyone sharing their opinions and
experiences.

What I've not seen discussed here is the purpose of using RAID1.  I
guess it is related to avoid data loss in case one drive dies.  The
other thing I'm guessing is that SSD is wanted due to the performance.

Considering that the RAID technology is old which was designed for
rotating media, maybe that isn't the the best solution for SSD?
I'm wondering ... maybe another (possibly better?) approach would be
dm-cache?  This has been added as a tech-preview in RHEL7.

The purpose of dm-cache is to have a faster unit (such as an SSD) to
cache contents from a slower device (such as a ordinary hard drive).  To
have better data security, the ordinary hard drive(s) can be an md-raid
device.

Maybe dm-cache can provide the best from SDD, HDD and RAID combined?


Just my 2 cents.


-- 
kind regards,

David Sommerseth


Re: need SSD RAID controller advice

2015-04-14 Thread Konstantin Olchanski
On Tue, Apr 14, 2015 at 06:39:52PM +0300, Vladimir Mosgalin wrote:
 
  Considering that the RAID technology is old which was designed for
  rotating media, maybe that isn't the the best solution for SSD?


With SSD, there is still possibility of single-disk failure (SSD bricks itself).

When this happens, to protect against data loss you have 2 choices:

a) daily (hourly) rsync of SSD to some other storage (on SSD failure, machine 
will die, require manual recovery)
b) RAID1 with another SSD (or HDD, use --write-mostly) (on SSD failure, 
machine will continue running)

SL7 comes with btrfs, which is intended to be better RAID1 (I am presently 
testing this).

But SL7 cannot boot from btrfs, so you still need a /boot partition and RAID1 
to protect it against single-disk failure.

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-14 Thread Vladimir Mosgalin
Hi David Sommerseth!

 On 2015.04.14 at 12:55:14 +0200, David Sommerseth wrote next:

 Considering that the RAID technology is old which was designed for
 rotating media, maybe that isn't the the best solution for SSD?
 I'm wondering ... maybe another (possibly better?) approach would be
 dm-cache?  This has been added as a tech-preview in RHEL7.

Heh, RAID is VERY old and doesn't play that good with modern rotating
media as well, which is why RAID2 and RAID3 are obsolete for a long
time, and RAID5 is obsolete / too dangerous for a few years by now; not
to mention write hole problems of RAID5/6, no solution for data
inconsistency problems between drives for RAID1/10 and other issues.

The proper modern era RAID replacements would be software volume-based
solutions, preferable integrated into filesystem (zfs, btrfs) and
distributed filesystems like ceph. That aside, using RAID1 or 10 for
very small servers to increase reliability somewhat and RAID6 and
similar for big file storages and backup servers still does make some
sense..

 The purpose of dm-cache is to have a faster unit (such as an SSD) to
 cache contents from a slower device (such as a ordinary hard drive).  To
 have better data security, the ordinary hard drive(s) can be an md-raid
 device.
 
 Maybe dm-cache can provide the best from SDD, HDD and RAID combined?

Not really. It was designed for completely different usage (think of it
as of ability to greatly increase file cache size, cheaper than adding
more RAM). It doesn't solve problems which are solved by RAID and
doesn't have its performance characteristics as well.

Using second drive for increased reliability to protect against device
failure: RAID1 can do this, dm-cache cannot (hard drives breaks - you
lose data even if it was somewhat cached on SSD).

Using multiple drives for creating storage of a big size - somewhat
slow, but with good reliability: RAID5/6/50/60 can do this, dm-cache
cannot.

Using multiple drives for getting guaranteed high bandwidth (required
for video recording from multiple stream sources, video editing and
such): RAID0 can do this, dm-cache cannot.

But yes, for course it makes sense to use dm-cache with RAID, e.g. on
database server where you can't afford to go all-SSD or on file storage
where you need extra performance for popular accessed files.

-- 

Vladimir


Re: need SSD RAID controller advice

2015-04-14 Thread Konstantin Olchanski
On Tue, Apr 14, 2015 at 12:55:14PM +0200, David Sommerseth wrote:
 
 What I've not seen discussed here is the purpose of using RAID1.  I
 guess it is related to avoid data loss in case one drive dies.  The
 other thing I'm guessing is that SSD is wanted due to the performance.
 

Correcto. SSD, HDD, punch cards, you still need protection against single 
device failure.

These days, hardware is free, compared to cost of down time, cost of manual 
crash recovery, etc.

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-14 Thread Konstantin Olchanski
On Tue, Apr 14, 2015 at 12:29:41PM +0300, Vladimir Mosgalin wrote:
 
 ... (e.g. 100 TB endurance of normal SSD vs 20 PB endurance of enterprise SSD)


These are numbers from vendors. Data on actual performance as seen by end users
is much harder to come by.

For HDDs, we have periodic reports from backblaze (i.e. 
https://www.backblaze.com/blog/best-hard-drive/)
For SSDs, we have 
http://techreport.com/review/27909/the-ssd-endurance-experiment-theyre-all-dead

Not much else, in recent years.

Both reports do not inspire much confidence in vendor-advertised failure rates 
or endurance numbers.

And direct studies of failure rates between enterprise and consumer storage
do not seem to exist (reports of my WD green drive died, but my WD black drive
is still ok need not apply).

The vendors themselves of course do have this data in the form of warranty 
return rates,
but they are not telling.

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-13 Thread Konstantin Olchanski
On Fri, Apr 10, 2015 at 05:21:07PM -0700, ToddAndMargo wrote:
 
 I have a customer interested in high end workstation. She wants RAID
 1 to protect her drives. She is also interested in Solid State
 Drives.
 (And she knows the difference between RAID and Backup.)
 
 What host controller would you guys recommend for a pair of Intel
 SSDSC2BW480A401, 480 GB 2.5 Internal Solid State Drives in RAID 1?
 

Since you do not say that hardware raid is required, I assume linux software
raid is acceptable?

We have been using dual Kingston 120GB SSDs attached to on-mobo sata ports
and linux mdadm raid1 for a few years now without any special problems.

Some web sites suggest special settings for SSDs in RAID1 configurations,
suggest the importance of TRIM, etc, but (as a test) I ignore all that
and use them as if they were normal HDDs. See no ill effects.

One thing to remember with SSDs - unlike HDDs that seem to soldier on forever 
(just
retired some still working perfectly HDDs from 2004), the SSDs do have finite
write capacity and they *will* die eventually (see the eye-opener test of SSD
capacity that just completed in February or so).

-- 
Konstantin Olchanski
Data Acquisition Systems: The Bytes Must Flow!
Email: olchansk-at-triumf-dot-ca
Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada


Re: need SSD RAID controller advice

2015-04-13 Thread jdow
The 3Ware RAID cards I have vastly outstrip the motherboard built in Intel RAID 
implementations for a RAID 5 setup. (I don't consider RAID 1 to be economically 
sensible for most uses.) A four disk RAID 5 SSD configuration can be 
breathtaking fast, too.


{^_^}   Joanne

On 2015-04-13 05:10, James M. Pulver wrote:

I would point out that I'm not sure I've ever really seen the benefit of Real 
Raid except  for the vendor making more money. The only place I've used it is in 
iSCSI boxes that run everything in firmware.

On all computers / servers, I've always used MDADM on Linux and ZFS on FreeNAS. 
Both have been excellent for my intended use, though FreeNAS is only at home 
for ~3 concurrent users, so take that whole thing with a grain of salt. Neither 
has lost data due to power outages or drive failures.

--
James Pulver
CLASSE Computer Group
Cornell University


-Original Message-
From: owner-scientific-linux-us...@listserv.fnal.gov 
[mailto:owner-scientific-linux-us...@listserv.fnal.gov] On Behalf Of Vladimir 
Mosgalin
Sent: Monday, April 13, 2015 5:30 AM
To: scientific-linux-users@fnal.gov
Subject: Re: need SSD RAID controller advice

Hi ToddAndMargo!

  On 2015.04.12 at 17:35:04 -0700, ToddAndMargo wrote next:


On 04/12/2015 10:54 AM, Nico Kadel-Garcia wrote:

and I*loved*  3Ware


Me too.  LSI gobbled them up.


Well, consolidation is often a good thing.

You can still buy best performing 3Ware 9750 (2011 model!) from LSI, they are 
selling them for those who are fine with 6 Gbps speeds. Don't think they'll be 
upgrading it to 12Gbps (not that many people are interested in real 
RAID-supporting cards at such speeds, these are mostly for connecting external 
storages...)



Re: need SSD RAID controller advice

2015-04-13 Thread ToddAndMargo

On 04/10/2015 05:21 PM, ToddAndMargo wrote:

Hi All,

I have a customer interested in high end workstation. She wants RAID 1
to protect her drives. She is also interested in Solid State Drives.
(And she knows the difference between RAID and Backup.)

What host controller would you guys recommend for a pair of Intel
SSDSC2BW480A401, 480 GB 2.5 Internal Solid State Drives in RAID 1?

Many thanks,
-T




Hi All,

This from LSI tech support:

  The deal breaker is TRIM.

  Once you put drives into the array, TRIM is not
  supported.  So you will have to sacrifice TRIM
  for RAID and or visa versa.

  You can use an HBA such as 9207-4i and set up
  software RAID 1 and use TRIM, but there is no
  controller cache.

  Or

  You can use  9261-4i 9266-4i or 9271-4i with RAID, 1 gig
  of controller cache but no TRIM.


So ... poop!

-T

Wonder if there are any enterprise level SSD's that don't
need TRIM?


--
~~
Computers are like air conditioners.
They malfunction when you open windows
~~


Re: need SSD RAID controller advice

2015-04-13 Thread Nico Kadel-Garcia
On Mon, Apr 13, 2015 at 8:16 PM, Konstantin Olchanski
olcha...@triumf.ca wrote:
 On Mon, Apr 13, 2015 at 07:29:18PM -0400, Nico Kadel-Garcia wrote:
  I am not sure what you refer to. With SL5 and SL6 you have 2 disks,
  put / on a software RAID1 partitions put grub on both disks ...

 Until it fails. Grubby, along with the kernel installlation RPM,
 doesn't know how to update, manage, or keep synchronized this second
 /boot partition. Hilarity can then ensue, along with making sure that
 your /etc/fstab doesn't detect the wrong disk and mount it incorrectly
 as /dev/sda. See, if your first disk dies, unless you're very cautious
 with /etc/fstab, it's very much a crapshoot if hte right partition
 will mount as /boot.

 Been there, done that, gave up on the silliness.


 You talk about /boot, /dev/sda, grubby, etc, I doubt you have been there, 
 done that. Gave up, sure.

 With mirrored disks, there is only /dev/md0 (/), /dev/md1 (swap), etc (if 
 you have /home, /data, etc),
 in /etc/fstab, / is mounted by UUID of the /dev/md0 RAID1 device,
 which in turn is assembled by UUID of component disks using mdadm magic.

Been there, done that. Don't have to make this stuff up, I spent some
time *designing* Linux based storage servers, but it was more than 5
years ago.

However, software RAID may have improved to the point where /boot
doesn't have to be its own partition. So I just tried it on a VM
runing SL 6.6, and got unhandled exception, and git it to work
flawlessly when I made /boot its own, non-RAID partition.. So I'm
not filled with confidence that putting /boot on its partition is not
still necessary.

Are you finding that it actually works? Would you please post an
/etc/fstab from a working system to help verify that it works?


RE: need SSD RAID controller advice

2015-04-13 Thread James M. Pulver
I would point out that I'm not sure I've ever really seen the benefit of Real 
Raid except  for the vendor making more money. The only place I've used it is 
in iSCSI boxes that run everything in firmware.

On all computers / servers, I've always used MDADM on Linux and ZFS on FreeNAS. 
Both have been excellent for my intended use, though FreeNAS is only at home 
for ~3 concurrent users, so take that whole thing with a grain of salt. Neither 
has lost data due to power outages or drive failures.

--
James Pulver
CLASSE Computer Group
Cornell University


-Original Message-
From: owner-scientific-linux-us...@listserv.fnal.gov 
[mailto:owner-scientific-linux-us...@listserv.fnal.gov] On Behalf Of Vladimir 
Mosgalin
Sent: Monday, April 13, 2015 5:30 AM
To: scientific-linux-users@fnal.gov
Subject: Re: need SSD RAID controller advice

Hi ToddAndMargo!

 On 2015.04.12 at 17:35:04 -0700, ToddAndMargo wrote next:

 On 04/12/2015 10:54 AM, Nico Kadel-Garcia wrote:
 and I*loved*  3Ware
 
 Me too.  LSI gobbled them up.

Well, consolidation is often a good thing.

You can still buy best performing 3Ware 9750 (2011 model!) from LSI, they are 
selling them for those who are fine with 6 Gbps speeds. Don't think they'll be 
upgrading it to 12Gbps (not that many people are interested in real 
RAID-supporting cards at such speeds, these are mostly for connecting external 
storages...)

-- 

Vladimir


Re: need SSD RAID controller advice

2015-04-12 Thread Nico Kadel-Garcia
On Sat, Apr 11, 2015 at 5:36 PM, ToddAndMargo toddandma...@zoho.com wrote:
 On 04/11/2015 11:22 AM, Nico Kadel-Garcia wrote:

 RAID is not backup.


 Hi Nico,

 She is well aware.

 And you are correct about goofing a file being
 far more prevalent.

 Also, backup needs history, otherwise you can't restore
 after the goofed file gets backed up.  (Often times,
 the user doesn't know they have goofed it up until they
 open it again.)

Yeah, that's why I'm a big user of rsnapshot. Lightweight,
configurable, and doesn't require complex software packages: I
especially use it for backing up /etc for systems where the owners
believe in reconfiguring things locally with no source control.

 You have any favorite hardware RAID controller that is SSD
 friendly?

 -T

I've been dealing more with industrial grade virtualization
environments lately, and the vendors or in house support staff for
those got really, really touchy when I asked them for part numbers. I
had good success, and continue to see good reviews, for Adaptec
controllers with *real* RAID, not kind-of-sort-of-software-RAID like a
lot of the older, cheaper LSI and Promise cards I ran into more than 5
years back. And I *loved* 3Ware for performance, but rarely had the
budget for them.

Check the latest reviews for more information, it's been more  than 10
years since I had a test rack for this sort of thing.


Re: need SSD RAID controller advice

2015-04-11 Thread ToddAndMargo

On 04/11/2015 11:06 AM, Brandon Vincent wrote:

IR mode


What is IR mode?


Re: need SSD RAID controller advice

2015-04-11 Thread ToddAndMargo

On 04/11/2015 11:22 AM, Nico Kadel-Garcia wrote:

RAID is not backup.


Hi Nico,

She is well aware.

And you are correct about goofing a file being
far more prevalent.

Also, backup needs history, otherwise you can't restore
after the goofed file gets backed up.  (Often times,
the user doesn't know they have goofed it up until they
open it again.)

You have any favorite hardware RAID controller that is SSD
friendly?

-T


Re: need SSD RAID controller advice

2015-04-11 Thread Vladimir Mosgalin
Hi ToddAndMargo!

 On 2015.04.10 at 17:21:07 -0700, ToddAndMargo wrote next:

 I have a customer interested in high end workstation. She wants RAID 1 to
 protect her drives. She is also interested in Solid State Drives.
 (And she knows the difference between RAID and Backup.)
 
 What host controller would you guys recommend for a pair of Intel
 SSDSC2BW480A401, 480 GB 2.5 Internal Solid State Drives in RAID 1?

If you really have no choice in picking any non-SF based SSD, then I
would only recommend software RAID with mdadm. It supports TRIM.
Otherwise you're going to live with eventual performance degradation,
because hardware controllers (at least LSI and Adaptec) do not support
TRIM for drives in RAID. You can google yourself on many reports on how
performance drops for SandForce SSDs filled with data in absence of
TRIM (you may also google about difference between background GC and
foreground GC). It's not pretty.

If you *do* have a choice in this (way, way more important than picking
controller) matter, pick an SSD from other vendor which supports
background garbage collection without TRIM and performance doesn't
degrade even in absence of TRIM. There are so many options and most
vendors put in description that these SSDs are OK to use in RAID or
other environment without TRIM support. These include Marvell-based SSDs
(Plextor M series, Crucial M/MX series and many others), Samsung SSDs
and really many others. If you want to pick SSD for RAID make sure to
check some review which covers performance degradation without TRIM.

It's not to say that Intel SSDs are bad (some are among very best, but
they are expensive enterprise-focused models), but these particular
consumer models based on SandForce controllers do not play that good
with random RAID controller. They are okay to use in software
(mdadm)-based RAID or on Intel software RAID (in Windows) because these
support TRIM, also there were specialized RAID-on-board PCIe SSDs based
on them which supported TRIM also, but you aren't going to get it with
any random controller.

Now, as long as you are using non-SF based SSD, you can among various
RAID controllers. It's probably easier to list ones which aren't
supported.. Good vendors (proven with years) include LSI and Adaptec,
basically you can pick nearly any controller with RAID support from
these vendors and it will work fine. Note that many other vendors like
Intel or Dell sell controllers based on LSI models, often cheaper than
originals. While there are occasional glitches (e.g. my favorite low-end
RAID controller LSI 9211-8i with buggy P20 firmware, had to downgrade),
usually they work great. But really there are tons of options, e.g. if
you are using Dell or HP workstation you can go with controller from
these vendors (Dell will likely to be LSI-based), they work great as
well. It's your choice.

(myself, for simple cases like two-disk RAID1 or four-disk RAID10 I
always just use software RAID :P)

-- 

Vladimir


Re: need SSD RAID controller advice

2015-04-11 Thread Brandon Vincent
On Sat, Apr 11, 2015 at 11:03 AM, Brandon Vincent
brandon.vinc...@asu.edu wrote:
 This is not correct. LSI SAS2008 I/O controllers do support TRIM with
 drives that support both deterministic trim (DRAT) and deterministic
 read zero after trim (DZAT).

My bad. I didn't see that this thread was about using the controller
in IR mode. Vladimir is correct about IR mode not supporting RAID.
Sorry for any confusion.

Brandon Vincent


Re: need SSD RAID controller advice

2015-04-11 Thread Kevin K
 On Apr 11, 2015, at 4:39 PM, ToddAndMargo toddandma...@zoho.com wrote:
 
 On 04/11/2015 11:06 AM, Brandon Vincent wrote:
 IR mode
 
 What is IR mode?

That appears to be when the controller is running in RAID mode.  Especially 
with integrated controllers, there can be the option to run the drives in raid 
mode or normal mode.

In the past when I bothered with this in computers I ran at home, the advantage 
of raid mode was you could have, say, 2 hard drives treated as 1 by the BIOS 
and operating system.  So you could boot off of the raid system, and have an OS 
driver take over from it later.  

In normal mode, if you wanted to RAID, you probably needed at least 3 hard 
drives.  A boot drive, and the others raided together using the OS software.  I 
kind of stopped doing this with a combination of OS software being sufficient 
to combine drives, and with drives (even SSDs) getting much larger.