On Dec 2, 2010, at 15:43 , Sven Almgren wrote: > What Raid controller do you use, and what kernel version? (Assuming > Linux). We hade problems during high load with a 3Ware raid controller > and the current kernel for Ubuntu 10.04, we hade to downgrade the > kernel... > > The problem was a bug in the driver that only showed up with very high > disk load (as is the case when doing imports) >
We're running freebsd: RaidController 3ware 9500S-8 Corrupt unit: Raid-10 3725.27GB 256K Stripe Size without BBU Freebsd 7.2, UFS Filesystem. > /Sven > > 2010/12/2 Robert Gründler <rob...@dubture.com>: >>> The very first thing I'd ask is "how much free space is on your disk >>> when this occurs?" Is it possible that you're simply filling up your >>> disk? >> >> no, i've checked that already. all disks have plenty of space (they have >> a capacity of 2TB, and are currently filled up to 20%. >> >>> >>> do note that an optimize may require up to 2X the size of your index >>> if/when it occurs. Are you sure you aren't optimizing as you add >>> items to your index? >>> >> >> index size is not a problem in our case. Our index currently has about 3GB. >> >> What do you mean with "optimizing as you add items to your index"? >> >>> But I've never heard of Solr causing hard disk crashes, >> >> neither did we, and google is the same opinion. >> >> One thing that i've found is the mergeFactor value: >> >> http://wiki.apache.org/solr/SolrPerformanceFactors#mergeFactor >> >> Our sysadmin speculates that maybe the chunk size of our raid/harddisks >> and the segment size of the lucene index does not play well together. >> >> Does the lucene segment size affect how the data is written to the disk? >> >> >> thanks for your help. >> >> >> -robert >> >> >> >> >> >> >> >>> >>> Best >>> Erick >>> >>> 2010/12/2 Robert Gründler <rob...@dubture.com> >>> >>>> Hi, >>>> >>>> we have a serious harddisk problem, and it's definitely related to a >>>> full-import from a relational >>>> database into a solr index. >>>> >>>> The first time it happened on our development server, where the >>>> raidcontroller crashed during a full-import >>>> of ~ 8 Million documents. This happened 2 weeks ago, and in this period 2 >>>> of the harddisks where the solr >>>> index files are located stopped working (we needed to replace them). >>>> >>>> After the crash of the raid controller, we decided to move the development >>>> of solr/index related stuff to our >>>> local development machines. >>>> >>>> Yesterday i was running another full-import of ~10 Million documents on my >>>> local development machine, >>>> and during the import, a harddisk failure occurred. Since this failure, my >>>> harddisk activity seems to >>>> be around 100% all the time, even if no solr server is running at all. >>>> >>>> I've been googling the last 2 days to find some info about solr related >>>> harddisk problems, but i didn't find anything >>>> useful. >>>> >>>> Are there any steps we need to take care of in respect to harddisk failures >>>> when doing a full-import? Right now, >>>> our steps look like this: >>>> >>>> 1. Delete the current index >>>> 2. Restart solr, to load the updated schemas >>>> 3. Start the full import >>>> >>>> Initially, the solr index and the relational database were located on the >>>> same harddisk. After the crash, we moved >>>> the index to a separate harddisk, but nevertheless this harddisk crashed >>>> too. >>>> >>>> I'd really appreciate any hints on what we might do wrong when importing >>>> data, as we can't release this >>>> on our production servers when there's the risk of harddisk failures. >>>> >>>> >>>> thanks. >>>> >>>> >>>> -robert >>>> >>>> >>>> >>>> >>>> >>>> >> >>