Thanks Rudi, that helps as you have good luck with all of them. I see they have some boards that go up to 192gb (but not DDR3), but some do 144gb as well. I just need to find out if the POD supports extended ATX and I see others have just used regular ATX boards.
-- Jason On Sunday, May 8, 2011 at 12:12 PM, Rudi Ahlers wrote: > > > On Sun, May 8, 2011 at 9:06 PM, Jason <slackmoehrle.li...@gmail.com> wrote: > > Rudy, > > > > > > Do you have a recommendation of a motherboard? > > Well, choose one here: http://www.supermicro.com/products/motherboard/matrix/ > > I don't have specific recommendations but we've had great success with all > our SuperMicro servers, both with single & dual CPU configurations, ranging > from 4GB - 128GB RAM > > > I am still reading the rest of your post. Thanks! > > > > -Jason > > > > -- > > Jason > > > > On Sunday, May 8, 2011 at 11:26 AM, Rudi Ahlers wrote: > > > > > > > > > On Sun, May 8, 2011 at 8:03 PM, Jason <slackmoehrle.li...@gmail.com> > > > wrote: > > > > Hi All, > > > > > > > > I am about to embark on a project that deals with allowing information > > > > archival, over time and seeing change over time as well. I can explain > > > > it a lot better, but I would certainly talk your ear off. I really > > > > don't have a lot of money to throw at the initial concept, but I have > > > > some. This device will host all of the operations for the first few > > > > months until I can afford to build a duplicate device. I already had a > > > > few parts of the idea done and ready to get live. > > > > > > > > I am contemplating building a BackBlaze Style POD. The goal of the > > > > device is to start acting as a place to have the crawls store > > > > information, massage it, get it into db's and then notify the user the > > > > task is done so they can start looking at the results. > > > > > > > > For reference here are a few links: > > > > > > > > http://blog.backblaze.com/2009/09/01/petabytes-on-a-budget-how-to-build-cheap-cloud-storage/ > > > > > > > > and > > > > > > > > http://cleanenergy.harvard.edu/index.php?ira=Jabba&tipoContenido=sidebar&sidebar=science > > > > > > > > There is room for 45 drives in the case (technically a few more). > > > > > > > > 45 x 1tb 7200rpm drives is really cheap, about $60 each. > > > > > > > > 45 x 1.5tb 7200rpm drives are about $70 each. > > > > > > > > 45 x 2tb 7200rpm drives are about $120 each > > > > > > > > 45 x 3tb 7200rpm drives are about $180-$230 each (or more, some are > > > > almost $400) > > > > > > > > I have question before I commit to building one and I was hoping to get > > > > advice. > > > > > > > > 1. Can anyone recommend a mobo/processor setup that can hold lots of > > > > RAM? Like 24gb or 64gb or more? > > > > > > Any brand server motherboard will do. I prefer supermicro, but you can > > > use Dell, HP, Intell, etc, etc. > > > > > > > > > > > 2. Hardware RAID or Software RAID for this? > > > > > > Hardware RAID will be expensive on 45 drives. IF you can, split the 45 > > > drives into a few smaller RAID arrays. To rebuild 1x large 45TB RAID > > > array, with either hardware or software would probably take a week, or > > > more, depending on which RAID type you use - i.e. RAID 5, or 6, or 10. I > > > prefer RAID 10 since it's best for speed and the rebuilds are the > > > quickest. But you loose half the space, i.e. 45TB drives will give you > > > about 22TB space. 45x 2TB HDD's would give you about 44TB space though. > > > > > > > > > > > 3. Would CentOS be a good choice? I have never used CentOS on a device > > > > so massive. Just ordinary servers, so to speak. I assume that it could > > > > handle so many drives, a large, expanding file system. > > > > > > Yes it would be fine. > > > > > > > > > > 4. Someone recommended ZFS but I dont recall that being available on > > > > CentOS, but it is on FreeBSD which I have little experience with. > > > > > > I would also prefer to use ZFS for this type of setup. use one 128GB SL > > > type SSD drive as a cache drive to speed up things and 2x log drives to > > > help with drive recovery. With ZFS you would be able to use one large > > > RAID array if you have the log drives since it was recover from driver > > > failure much better than other file systems. Although you can install ZFS > > > as user-land tools, which will be slower than running it via the kernel. > > > But, it would be better to use Solaris or FreeBSD for this - look @ > > > Nexenta / FreeNAS / OpenIndia for this. > > > > > > > > > > > 5. How would someone realistically back something like this up? > > > > > > To another one as large :) > > > > > > OR, more realistically, if you already have some backup servers, and the > > > full 45TB isn't full of data yet, then simply backup what you have. By > > > the sounds of it your project is still new so your data won't be that > > > much. I would simply build a gluster / CLVM cluster of smaller cheaper > > > servers - which basically allows you to add say 4TB / 8TB (depending on > > > what chassis you use and how many drives it can take) at a time to the > > > backup cluster, which will be cheaper than buying another one identical > > > to this right now. > > > > > > > > > > > Ultimately I know over time I need to distribute my architecture out > > > > and have a number of web-servers, balancing, etc but to get started I > > > > think this device with good backups might fit the bill. > > > > > > If this device will be used for web + mail + SQL, then you may probably > > > look at using 4 quad core CPU's + 128GB RAM. With this many drives (or > > > rather, this much data) you'll probably run out of RAM / CPU / Network > > > resources before you run out of HDD space. > > > > > > > > > > > > With a device this big (in terms of storage) I would rather have 2 > > > separate "processing" servers which just mounts LUN's from this POD > > > (exported as NFS / iSCSI / FCoE / etc) and then have a few faster SAS / > > > SSD drives for SQL / log processing. > > > > > > > > > > > I can be way more detailed if it helps, I just didn't want to clutter > > > > with information that might not be relevant. > > > > -- > > > > Jason > > > > > > > > _______________________________________________ > > > > CentOS mailing list > > > > CentOS@centos.org > > > > http://lists.centos.org/mailman/listinfo/centos > > > > > > > > > -- > > > Kind Regards > > > Rudi Ahlers > > > SoftDux > > > > > > Website: http://www.SoftDux.com > > > Technical Blog: http://Blog.SoftDux.com > > > Office: 087 805 9573 > > > Cell: 082 554 7532 > > > _______________________________________________ > > > CentOS mailing list > > > CentOS@centos.org > > > http://lists.centos.org/mailman/listinfo/centos > > > > _______________________________________________ > > CentOS mailing list > > CentOS@centos.org > > http://lists.centos.org/mailman/listinfo/centos > > > -- > Kind Regards > Rudi Ahlers > SoftDux > > Website: http://www.SoftDux.com > Technical Blog: http://Blog.SoftDux.com > Office: 087 805 9573 > Cell: 082 554 7532 > _______________________________________________ > CentOS mailing list > CentOS@centos.org > http://lists.centos.org/mailman/listinfo/centos > _______________________________________________ CentOS mailing list CentOS@centos.org http://lists.centos.org/mailman/listinfo/centos