We have very large indexes, almost a terabyte for a single index, and
normally it takes overnight to run a checkindex. I started a CheckIndex
on Friday and today (Monday) it seems to be stuck testing vectors although
we haven't got vectors turned on. (See below)
The output file was last written Jul 27 02:28,
Note that in this 750 GB segment we have about 83 million docs with about
2.4 billion unique terms and about 110 trillion tokens.
Have we hit a new CheckIndex limit?
Tom
-----------------------
Opening index @ /htsolr/lss-dev/solrs/4.2/3/core/data/index
Segments file=segments_e numSegments=2 version=4.2.1 format=
userData={commitTimeMSec=1374712392103}
1 of 2: name=_bch docCount=82946896
codec=Lucene42
compound=false
numFiles=12
size (MB)=752,005.689
diagnostics = {timestamp=1374657630506, os=Linux,
os.version=2.6.18-348.12.1.el5, mergeFactor=16, source=merge,
lucene.version=4.2.1 1461071 - mark - 2013-03-26 08:23:34, os.arch=amd64,
mergeMaxNumSegments=2, java.version=1.6.0_16, java.vendor=Sun Microsystems
Inc.}
no deletions
test: open reader.........OK
test: fields..............OK [12 fields]
test: field norms.........OK [3 fields]
test: terms, freq, prox...OK [2442919802 terms; 73922320413 terms/docs
pairs; 109976572432 tokens]
test: stored fields.......OK [960417844 total field count; avg 11.579
fields per doc]
test: term vectors........
~