We have very large indexes, almost a terabyte for a single index, and normally it takes overnight to run a checkindex. I started a CheckIndex on Friday and today (Monday) it seems to be stuck testing vectors although we haven't got vectors turned on. (See below) The output file was last written Jul 27 02:28, Note that in this 750 GB segment we have about 83 million docs with about 2.4 billion unique terms and about 110 trillion tokens.
Have we hit a new CheckIndex limit? Tom ----------------------- Opening index @ /htsolr/lss-dev/solrs/4.2/3/core/data/index Segments file=segments_e numSegments=2 version=4.2.1 format= userData={commitTimeMSec=1374712392103} 1 of 2: name=_bch docCount=82946896 codec=Lucene42 compound=false numFiles=12 size (MB)=752,005.689 diagnostics = {timestamp=1374657630506, os=Linux, os.version=2.6.18-348.12.1.el5, mergeFactor=16, source=merge, lucene.version=4.2.1 1461071 - mark - 2013-03-26 08:23:34, os.arch=amd64, mergeMaxNumSegments=2, java.version=1.6.0_16, java.vendor=Sun Microsystems Inc.} no deletions test: open reader.........OK test: fields..............OK [12 fields] test: field norms.........OK [3 fields] test: terms, freq, prox...OK [2442919802 terms; 73922320413 terms/docs pairs; 109976572432 tokens] test: stored fields.......OK [960417844 total field count; avg 11.579 fields per doc] test: term vectors........ ~