RE: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-11 Thread Toke Eskildsen
Erick Erickson [erickerick...@gmail.com] wrote: I guess my $0.02 is that you'd have to have strong evidence that extending Lucene to 64 bit is even useful. Or more generally, useful enough to pay the penalty. All the structures that allocate maxDoc id arrays would suddenly require twice the

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-11 Thread Erick Erickson
bq: Are there any such structures? Well, I thought there were, but I've got to admit I can't call any to mind immediately. bq: 2b is just the hard limit Yeah, I'm always a little nervous as to when Moore's Law will make everything I know about current systems' performance obsolete. At any

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-11 Thread Jack Krupansky
Over the past few years, it is only on the Solr user mailing list that I have read requests for greater than 2B documents in a single index - nobody has requested this on the Lucene user list. The point is that this is primarily a Lucene issue that just happens to get passed through to Solr users,

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-10 Thread Erick Erickson
I guess my $0.02 is that you'd have to have strong evidence that extending Lucene to 64 bit is even useful. Or more generally, useful enough to pay the penalty. All the structures that allocate maxDoc id arrays would suddenly require twice the memory for instance, plus all the coding effort that

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-10 Thread Shawn Heisey
On 2/4/2015 3:31 PM, Arumugam, Suresh wrote: We are trying to do a POC for searching our log files with a single node Solr(396 GB RAM with 14 TB Space). Since the server is powerful, added 2 Billion records successfully search is working fine without much issues. Due to the restriction of

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-06 Thread Upayavira
Given that it is log entries, you might find it works to use a collection per day, and then use collection aliasing to query over them all. This way, you can have a different aliases that specify certain ranges (e.g. week is an alias for the last 7 or 8 day's collections). Upayavira On Thu, Feb

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-05 Thread Toke Eskildsen
On Wed, 2015-02-04 at 23:31 +0100, Arumugam, Suresh wrote: We are trying to do a POC for searching our log files with a single node Solr(396 GB RAM with 14 TB Space). We're running 7 billion larger-than-typical-log-entries documents from a machine of similar size and it serves our needs well:

RE: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Arumugam, Suresh
of it. Regards, Suresh.A From: Arumugam, Suresh [mailto:suresh.arumu...@emc.com] Sent: Wednesday, February 04, 2015 1:54 PM To: solr-user@lucene.apache.org Cc: Habeeb, Anwar Subject: Exception while loading 2 Billion + Documents in Solr 4.8.0 Hi All, We are trying to load 14+ Billion documents into Solr

Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Arumugam, Suresh
Hi All, We are trying to load 14+ Billion documents into Solr. But we are failing to load them into Solr. Solr version: 4.8.0 Analyzer used: ClassicTokenizer for index as well as query. Can someone help me in getting into the core of this issue? For 14+ Billion document load, we are loading

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Jack Krupansky
What's your cluster size? The 2 billion limit is per-node. My personal recommendation is that you don't load more than 100 million documents per node. You need to do a proof of concept test to verify whether your particular data would support a higher number or not. Ultimately, it will not be a

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Shawn Heisey
On 2/4/2015 2:54 PM, Arumugam, Suresh wrote: Hi All, We are trying to load 14+ Billion documents into Solr. But we are failing to load them into Solr. Solr version: *4.8.0* Analyzer used: *ClassicTokenizer for index as well as query.* Can someone help me in getting into the

RE: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Chris Hostetter
: We are trying to do a POC for searching our log files with a single node Solr(396 GB RAM with 14 TB Space). : Since the server is powerful, added 2 Billion records successfully search is working fine without much issues. how much CPU? Assuming it's comparable to the amount of RAM you've

Re: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Walter Underwood
You can only put 2 billion documents in one core. This error message is the clue: Too many documents, composite IndexReaders cannot exceed 2147483647 You will need to shard the collection. You might have multiple shards per node, but you will probably need 50-100 shards and lots of servers.

RE: Exception while loading 2 Billion + Documents in Solr 4.8.0

2015-02-04 Thread Arumugam, Suresh
: Re: Exception while loading 2 Billion + Documents in Solr 4.8.0 On 2/4/2015 2:54 PM, Arumugam, Suresh wrote: Hi All, We are trying to load 14+ Billion documents into Solr. But we are failing to load them into Solr. Solr version: *4.8.0* Analyzer used: *ClassicTokenizer for index