Anthony, I'd also add to Jim's comments that once you have one big central
drive you can use someting like XBMC to have a very nice interface on all
your HTPCs in the house and accessibility to all your content.



---------
Brian



On Mon, Feb 24, 2014 at 9:42 AM, James Maki <jwm_maill...@comcast.net>wrote:

> That's how I started! :) But the desire for ease of use for my family (if
> it's not in plain sight, they can't find the drive, folder or location of a
> desired movie or TV show) and it just got "out of control!" A couple of
> drives here. A Sans Digital tower there. A new HTPC in the family room.
> Gigabit network hooking upstairs bedroom to the main computer downstairs.
> You name it, it got added.  I ended up spending lots of time "cataloging,"
> especially when adding drives. The pooling aspect of FlexRAID allows me to
> have one BIG drive with a folder for Blu-rays, one for DVDs, and another
> for
> recorded TV shows. Previously, a desired file might have been on one of 4
> computers and any one of the approximately 30 drives. I did compromise
> awhile back and create 8 and 10 TB JBODs on the Sans Digital towers and
> internal in the main HTPC. This made it slightly easier to catalog.
>
> Of course, all of this ignores the "building computers, etc." is fun factor
> of this hobby. :)
>
> If nothing else, I have learned lots about SAS (which had intimidated me
> before), building my own NAS, and a little about Server software. Always a
> fun (if not occasionally, frustrating) experience.
>
> To Brian: I am doing exactly that-One big drive with 3 shared folders. The
> multiple pool idea was to facilitate doing smaller Updates/validates that
> could be done overnight rather than over 3 or 4 days. Once I get the drive
> set up as desired, I will give the parity backup another try and see if
> once
> it is set if the periodic updates of a static pool are quick. Thanks for
> the
> input and feedback.
>
> Jim
>
> > -----Original Message-----
> > From: hardware-boun...@lists.hardwaregroup.com [mailto:hardware-
> > boun...@lists.hardwaregroup.com] On Behalf Of Anthony Q. Martin
>
> > You guys are so sophisticated!  I'm just stringing all my drives off a PC
> with
> > external enclosures (10 drives inside the box, 8 more in two four-bay
> > enclosures).  Using 3 and 4 TB drives (greens, mostly, from WD and
> seagate).
> > Mine or just NTFS mount volumes all shared over my GB network.  That way,
> > I can just navigate to any drive and any folder to play my rips from my
> other
> > HTPCs.  Easy setup.  If a drive goes down, I just re-rip as I have all
> the
> optical
> > discs as backup.  Poor man's setup.  Lazy man's setup. :) Raid is too
> > complicated for my brain and I don't see my use as super critical.
> Ripping to
> > mkv is mostly done in the background while working on other stuff.
> >
> > On 2/24/2014 8:30 AM, Brian Weeden wrote:
> > > Jim, have you thought about setting up multiple shares instead of
> > > multiple pools?  For example, you could have one big drive pool with
> > > all your data but share out any folder on that pool as a separate
> network
> > share.
> > >
> > >
> > >
> > > ---------
> > > Brian
> > >
> > >
> > >
> > > On Sun, Feb 23, 2014 at 2:46 PM, Brian Weeden
> > <brian.wee...@gmail.com>wrote:
> > >
> > >> If you're doing an initialization and building parity for 23 TB of
> > >> data, I can expect that to take quite a while. The update I'm not so
> > >> sure about. It should only need up update parity for whatever files
> > >> were changed. So if the update needs just as long, that indicates
> maybe
> > all your data changed.
> > >> But if it's just video files then it shouldn't.
> > >>
> > >> I do know people have talked about exempting things like nfo files
> > >> and thumbnails from the RAID so the parity process will skip them.
> > >>
> > >>
> > >>
> > >> ---------
> > >> Brian
> > >>
> > >>
> > >>
> > >> On Sun, Feb 23, 2014 at 2:17 PM, James Maki
> > <jwm_maill...@comcast.net>wrote:
> > >>
> > >>> Hi Brian,
> > >>>
> > >>> I switched to FlexRAID to combine a total of 23 2tb drives spread
> > >>> over 5 Sans Digital port multiplier towers plus extra drives on
> > >>> several PCs used as HTPCs. I have ripped all my Blu-ray, DVDs and
> > >>> recorded TV to the various arrays and over time had just gotten too
> > >>> large to easily manage. I wanted to centralize everything on one
> > >>> system. The system I started with utilized a AMD FM2 motherboard
> > >>> with 8 onboard SATA ports, 2 SAS ports on an add-on card (for a
> > >>> total of 8 additional SATA ports, and 3 of the Sans Digital towers
> > >>> (5
> > >>> disks each) for a total of 31 drives distributed as 1 OS drive, 4
> > >>> parity drives and 26 data drives (several were empty). When this
> > >>> continued to fail on creation, I moved the Sans Digital based drives
> > >>> to a 6 port SAS controller card.
> > >>>
> > >>> When I still had problems, I found that several drives were bad
> > >>> (scan disk), including the 1st parity drive. Replacing the drives
> > >>> gave me a successful creation but it took 4 days. The Update took
> > >>> another 4 days. That's when I started having second thoughts on
> > >>> using the Parity backup option. I guess I am just expected too much
> > >>> from the software. That's when I thought creating several pools
> > >>> would reduce the strain for each update/validate.
> > >>>
> > >>> I am using a modestly powered AMD dual core 3.2 GHz processor and
> > >>> mostly consumer drives (mixed with a few WD reds). I went with
> > >>> Windows Home Server for economy reasons ($50 vs. $90-130 for
> > Windows
> > >>> 7 Home Premium/Professional). I utilized a HighPoint RocketRAID
> > >>> 2760A SAS RAID controller card. I am using RAID over File System
> > >>> 2.0u12, SnapRAID 1.4 Stable and Storage Pool 1.0 Stable (although
> > >>> not using the SnapRAID at this point).
> > >>>
> > >>> Overall, I am happy with the pooling facility of the software. I
> > >>> just wish my large setup would not choke the parity option. Thanks
> > >>> for all the input.
> > >>>
> > >>> Not sure if there is an answer to my problem. More powerful hardware?
> > >>> Reading the forums seems to indicate that hardware should NOT be the
> > >>> bottleneck. There seems to be the option of Updating/Validating only
> > >>> portions of the RAID each night. More research is needed on that
> > >>> front. My current plan at this point is to fill the RAID in the
> > >>> pooling only mode, make sure all names and organization is correct,
> > >>> then commit to a stable, unchanging file system that I will then
> > >>> commit to the SnapRAID parity option. That way I will only need to
> > Validate/Verify periodically.
> > >>>
> > >>> Thanks,
> > >>>
> > >>> Jim
> > >>>
> > >>>> -----Original Message-----
> > >>>> From: hardware-boun...@lists.hardwaregroup.com [mailto:hardware-
> > >>>> boun...@lists.hardwaregroup.com] On Behalf Of Brian Weeden
> > >>>> Sent: Sunday, February 23, 2014 6:06 AM
> > >>>> To: hardware
> > >>>> Cc: hwg
> > >>>> Subject: Re: [H] What are we up to (Was-Are we alive?)
> > >>>>
> > >>>> Hi Jim. Sorry to hear you're having such troubles, especially since
> > >>>> I
> > >>> think I'm
> > >>>> the one who introduced FlexRAID to the list.
> > >>>>
> > >>>> I've been running it on my HTPC for several years now and (knock on
> > >>> wood)
> > >>>> it's been running fine. Not sure how big your setup is, I'm running
> > >>>> 7
> > >>> DRUs
> > >>> and
> > >>>> 2 PRUs of 2 TB each. I have them mounted as a single pool that is
> > >>>> shared
> > >>> on
> > >>>> my LAN. I run nightly parity updates.
> > >>>>
> > >>>> Initilaizing my setup did take several hours, but my updates don't
> > >>>> take
> > >>> very
> > >>>> long. Sometimes when I add several ripped HD movies at once it
> > >>>> might
> > >>> take
> > >>> a
> > >>>> few hours but that's it. How much data are you calcluating parity
> > >>>> for at
> > >>> the
> > >>>> initialization? Do you have a lot of little files (like thousand of
> > >>> pictures) or lots
> > >>>> of files that change often? Either of those could greatly increase
> > >>>> the
> > >>> time it
> > >>>> takes to calcluate parity.
> > >>>>
> > >>>> I'm running it under Win7, and unfortunately I don't have any
> > >>>> experience with Server 2011 or any of the Windows Server builds.
> > >>>>
> > >>>>  From what I've gathered you can only have one pool per system. I
> > >>>> think that's a limit of how things work. But I've never needed more
> > >>>> than one
> > >>> pool,
> > >>>> so it hasn't bothered me.
> > >>>>
> > >>>> For hardware, I'm running the following based largely on a HTPC
> > >>>> hardware guide I found online. It's based on a server chipset to
> > >>>> maximize the bandwidth to the drives.
> > >>>>
> > >>>> Intel Xeon E3-1225
> > >>>> Asus P8B WS LGA 1155 Intel C206
> > >>>> 8 GB DDR3 SDRAM
> > >>>> Corsair TX750 V2 750W
> > >>>> 2x Intel RAID Controller Card SATA/SAS PCI-E x8 Antec 1200 V3 Case
> > >>>> 3x
> > >>> 5in1
> > >>>> hot swap HDD cages
> > >>>>
> > >>>> Part of the key is the controller cards. I'm not actually using the
> > >>> on-board
> > >>>> RAID, just using it for the ports and the bandwidth. I've  got two
> > >>>> SAS
> > >>> to
> > >>> SATA
> > >>>> cables plugged into each card, which gives me a total of 16 SATA
> ports.
> > >>> The
> > >>>> cards are each on an 8x PCIe bus that gives them a lot of bandwidth.
> > >>> Boot
> > >>>> drive is an older SSD that is attached to one of the SATA ports on
> > >>>> the
> > >>> mobo.
> > >>>> Once trick I figured out early on was to initialize your array with
> > >>>> the
> > >>> biggest
> > >>>> number of DRUs you think you'll eventually have, even if you don't
> > >>> actually
> > >>>> have that many drives at the start. That way you can add new DRUs
> > >>>> and
> > >>> not
> > >>>> have to reinitialize the array.
> > >>>>
> > >>>> When I started using FlexRAID it was basically a part-time project
> > >>>> being
> > >>> run
> > >>>> by Brahim. He's now created a fully-fledged business out of it and
> > >>>> has
> > >>> gone
> > >>>> way beyond just FlexRAID. Apparently he now has two products. I
> > >>>> think
> > >>> the
> > >>>> classic FlexRAID system I'm still using has become RAID-F (RAID
> > >>>> over
> > >>>> filesystem) and he's got a new Transparent RAID product as well:
> > >>>> http://www.flexraid.com/faq/
> > >>>>
> > >>>> I'm still running 2.0u8 (snapshot 1.4 stable) so I guess at some
> > >>>> point
> > >>> I'll need
> > >>>> to move over to the commercial version. But for now it's working
> > >>>> fine
> > >>> so I
> > >>>> don't want to disturb it.
> > >>>>
> > >>>> Hope all this helps, and happy to answer any other questions
> > >>>> however I
> > >>> can.
> > >>>>
> > >>>>
> > >>>> ---------
> > >>>> Brian
> > >>>
>
>

Reply via email to