Re: unified highlighter performance in solr 8.5.1

2020-07-04 Thread David Smiley
Here's my PR, which includes some edits to the ref guide docs where I tried
to clarify these settings a little too.
https://github.com/apache/lucene-solr/pull/1651
~ David


On Sat, Jul 4, 2020 at 8:44 AM Nándor Mátravölgyi 
wrote:

> I guess that's fair. Let's have hl.fragsizeIsMinimum=true as default.
>
> On 7/4/20, David Smiley  wrote:
> > I doubt that WORD mode is impacted much by hl.fragsizeIsMinimum in terms
> of
> > quality of the highlight since there are vastly more breaks to pick from.
> > I think that setting is more useful in SENTENCE mode if you can stand the
> > perf hit.  If you agree, then why not just let this one default to
> "true"?
> >
> > We agree on better documenting the perf trade-off.
> >
> > Thanks again for working on these settings, BTW.
> >
> > ~ David
> >
> >
> > On Fri, Jul 3, 2020 at 1:25 PM Nándor Mátravölgyi <
> nandor.ma...@gmail.com>
> > wrote:
> >
> >> Since the issue seems to be affecting the highlighter differently
> >> based on which mode it is using, having different defaults for the
> >> modes could be explored.
> >>
> >> WORD may have the new defaults as it has little effect on performance
> >> and it creates nicer highlights.
> >> SENTENCE should have the defaults that produce reasonable performance.
> >> The docs could document this while also mentioning that the UH's
> >> performance is highly dependent on the underlying Java String/Text?
> >> Iterator.
> >>
> >> One can argue that having different defaults based on mode is
> >> confusing. In this case I think the defaults should be changed to have
> >> the SENTENCE mode perform better. Maybe the options for nice
> >> highlights with WORD mode could be put into the docs in this case as
> >> some form of an example.
> >>
> >> As long as I can use the UH with nicely aligned snippets in WORD mode
> >> I'm fine with any defaults. I explicitly set them in the config and in
> >> the queries most of the time anyways.
> >>
> >
>


Re: Time-out errors while indexing (Solr 7.7.1)

2020-07-04 Thread Mad have
Thank a lot for your inputs and suggestions, even I was thinking similar like 
creating  another collection of the same ( hot and cold), and moving documents 
which are older than certain days like 180 days from original collection (hot) 
to new collection(cold). 

Thanks,
Madhava

Sent from my iPhone

> On 4 Jul 2020, at 14:37, Erick Erickson  wrote:
> 
> You need more shards. And, I’m pretty certain, more hardware.
> 
> You say you have 13 billion documents and 6 shards. Solr/Lucene has an 
> absolute upper limit of 2B (2^31) docs per shard. I don’t quite know how 
> you’re running at all unless that 13B is a round number. If you keep adding 
> documents, your installation will shortly, at best, stop accepting new 
> documents for indexing. At worst you’ll start seeing weird errors and 
> possibly corrupt indexes and have to re-index everything from scratch.
> 
> You’ve backed yourself in to a pretty tight corner here. You either have to 
> re-index to a properly-sized cluster or use SPLITSHARD. This latter will 
> double the index-on-disk size (it creates two child indexes per replica and 
> keeps the old one for safety’s sake that you have to clean up later). I 
> strongly recommend you stop ingesting more data while you do this.
> 
> You say you have 6 VMs with 2 nodes running on each. If those VMs are 
> co-located with anything else, the physical hardware is going to be stressed. 
> VMs themselves aren’t bad, but somewhere there’s physical hardware that runs 
> it…
> 
> In fact, I urge you to stop ingesting data immediately and address this 
> issue. You have a cluster that’s mis-configured, and you must address that 
> before Bad Things Happen.
> 
> Best,
> Erick
> 
>> On Jul 4, 2020, at 5:09 AM, Mad have  wrote:
>> 
>> Hi Eric,
>> 
>> There are total 6 VM’s in Solr clusters and 2 nodes are running on each VM. 
>> Total number of shards are 6 with 3 replicas. I can see the index size is 
>> more than 220GB on each node for the collection where we are facing the 
>> performance issue.
>> 
>> The more documents we add to the collection the indexing become slow and I 
>> also have same impression that the size of the collection is creating this 
>> issue. Appreciate if you can suggests any solution on this.
>> 
>> 
>> Regards,
>> Madhava 
>> Sent from my iPhone
>> 
 On 3 Jul 2020, at 23:30, Erick Erickson  wrote:
>>> 
>>> Oops, I transposed that. If your index is a terabyte and your RAM is 128M, 
>>> _that’s_ a red flag.
>>> 
 On Jul 3, 2020, at 5:53 PM, Erick Erickson  wrote:
 
 You haven’t said how many _shards_ are present. Nor how many replicas of 
 the collection you’re hosting per physical machine. Nor how large the 
 indexes are on disk. Those are the numbers that count. The latter is 
 somewhat fuzzy, but if your aggregate index size on a machine with, say, 
 128G of memory is a terabyte, that’s a red flag.
 
 Short form, though is yes. Subject to the questions above, this is what 
 I’d be looking at first.
 
 And, as I said, if you’ve been steadily increasing the total number of 
 documents, you’ll reach a tipping point sometime.
 
 Best,
 Erick
 
>> On Jul 3, 2020, at 5:32 PM, Mad have  wrote:
> 
> Hi Eric,
> 
> The collection has almost 13billion documents with each document around 
> 5kb size, all the columns around 150 are the indexed. Do you think that 
> number of documents in the collection causing this issue. Appreciate your 
> response.
> 
> Regards,
> Madhava 
> 
> Sent from my iPhone
> 
>> On 3 Jul 2020, at 12:42, Erick Erickson  wrote:
>> 
>> If you’re seeing low CPU utilization at the same time, you probably
>> just have too much data on too little hardware. Check your
>> swapping, how much of your I/O is just because Lucene can’t
>> hold all the parts of the index it needs in memory at once? Lucene
>> uses MMapDirectory to hold the index and you may well be
>> swapping, see:
>> 
>> https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
>> 
>> But my guess is that you’ve just reached a tipping point. You say:
>> 
>> "From last 2-3 weeks we have been noticing either slow indexing or 
>> timeout errors while indexing”
>> 
>> So have you been continually adding more documents to your
>> collections for more than the 2-3 weeks? If so you may have just
>> put so much data on the same boxes that you’ve gone over
>> the capacity of your hardware. As Toke says, adding physical
>> memory for the OS to use to hold relevant parts of the index may
>> alleviate the problem (again, refer to Uwe’s article for why).
>> 
>> All that said, if you’re going to keep adding document you need to
>> seriously think about adding new machines and moving some of
>> your replicas to them.
>> 
>> Best,
>> Erick
>> 
>>> On Jul 3, 2020, at 7:

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Tran Van Hoan
 
*total thread is 25.6k when solr hang.
On Sunday, July 5, 2020, 2:55:26 AM GMT+7, Tran Van Hoan 
 wrote:  
 
  All server only run Solr, zookeeper, exporters (node-exporter, 
process-exporter, solr-exporter, zoo-exporter). 
- network: no package loss, TCP no issue before incident, TCP drop around 
100-200/s when incident and overflow ~100 in somaxcon.
- total mem Avalilable is greater 25G (Solr's XMX = 30G)- CPU < 10%, no high 
load average 5, 15 (except when solr hang, high load average is high)- Normally 
total process is around ~2k, when incident occurs, total thread is 25.6 (solr 
out of cluster but still running)- low TCP contrack- Swapiness = 0 and no use 
swap;- Solr cloud has 2 collections, only one collection down causes node down 
(the remain collection is still green)

On Sunday, July 5, 2020, 1:30:59 AM GMT+7, Rodrigo Oliveira 
 wrote:  
 
 Network it's ok? Between nodes? The use? Swap it's disabled? Swapiness rhe
value it's 0?

Em sáb, 4 de jul de 2020 15:19, Tran Van Hoan
 escreveu:

>  I used physical servers, and IO wait is small :(!!!I saw that iptables
> dropped all ACK message from clients (not only client solr, prometheus
> scape metric from exporter was dropped too).all when i check netstat
> -anp|grep 8983, all socket are TIME_WAIT state.Only restart solrs, the
> incident was resolved. Total request around 2.5k request per second per
> node.
>
>    On Sunday, July 5, 2020, 1:11:38 AM GMT+7, Rodrigo Oliveira <
> adamantina.rodr...@gmail.com> wrote:
>
>  Hi,
>
> I had this problem. In my case was the wait/io in vm. I migrate my
> environment to another place and solved.
>
> Actually it's problem wirh wait/io at host physical (until backup it's a
> problem over veeam).
>
> Regards
>
> Em sáb, 4 de jul de 2020 12:30, Tran Van Hoan
>  escreveu:
>
> > The problem reoccurs repeatly in recent days.
> > To day i tried dump heap and thread. Only dumping thread, heap can not
> > because solr instance was hang.
> > Almost thread was blocked.
> >
> > On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan
> >  wrote:
> >
> >
> > I checked node exporter metrics and saw network no problem
> >
> > On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan <
> > tranvanhoan...@yahoo.com> wrote:
> >
> >
> > I check node exporter, no problem with OS, hardware and network.
> > I attached images about solr metrics 7 days and 12h.
> >
> > On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin <
> > dario.rigo...@comperio.it> wrote:
> >
> >
> > What about a network issue?
> >
> > Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
> >  ha scritto:
> >
> >
> > >
> > > dear all,
> > >
> > >  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> > > instance has xmx = xms = 30G.
> > >
> > > Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> > > nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One
> node
> > > were down. almost metrics didn't increase too much except threads.
> > >
> > > Performance in one week ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > performace 12h ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > I go to the admin UI, some node dead some node too long to response.
> When
> > > checking logfile, they generate too much (log level warning), here are
> > logs
> > > which appears in the solr cloud:
> > >
> > > Log before server 4 and 6 down
> > >
> > > - Server 4 before it dead:
> > >
> > >    + o.a.s.h.RequestHandlerBase java.io.IOException:
> > > java.util.concurrent.TimeoutException: Idle timeout expired:
> > 12/12
> > > ms
> > >
> > >  +org.apache.solr.client.solrj.SolrServerException: Timeout occured
> while
> > > waiting response from server at:
> > > http://server6:8983/solr/mycollection_shard3_replica_n5/select
> > >
> > >
> > >
> > > at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
> > >
> > >                at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
> > >
> > >                at
> > > org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
> > >
> > >                at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
> > >
> > >                at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
> > >
> > >                at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >                at
> > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> > >
> > >                at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >                at

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Tran Van Hoan
 All server only run Solr, zookeeper, exporters (node-exporter, 
process-exporter, solr-exporter, zoo-exporter). 
- network: no package loss, TCP no issue before incident, TCP drop around 
100-200/s when incident and overflow ~100 in somaxcon.
- total mem Avalilable is greater 25G (Solr's XMX = 30G)- CPU < 10%, no high 
load average 5, 15 (except when solr hang, high load average is high)- Normally 
total process is around ~2k, when incident occurs, total thread is 25.6 (solr 
out of cluster but still running)- low TCP contrack- Swapiness = 0 and no use 
swap;- Solr cloud has 2 collections, only one collection down causes node down 
(the remain collection is still green)

On Sunday, July 5, 2020, 1:30:59 AM GMT+7, Rodrigo Oliveira 
 wrote:  
 
 Network it's ok? Between nodes? The use? Swap it's disabled? Swapiness rhe
value it's 0?

Em sáb, 4 de jul de 2020 15:19, Tran Van Hoan
 escreveu:

>  I used physical servers, and IO wait is small :(!!!I saw that iptables
> dropped all ACK message from clients (not only client solr, prometheus
> scape metric from exporter was dropped too).all when i check netstat
> -anp|grep 8983, all socket are TIME_WAIT state.Only restart solrs, the
> incident was resolved. Total request around 2.5k request per second per
> node.
>
>    On Sunday, July 5, 2020, 1:11:38 AM GMT+7, Rodrigo Oliveira <
> adamantina.rodr...@gmail.com> wrote:
>
>  Hi,
>
> I had this problem. In my case was the wait/io in vm. I migrate my
> environment to another place and solved.
>
> Actually it's problem wirh wait/io at host physical (until backup it's a
> problem over veeam).
>
> Regards
>
> Em sáb, 4 de jul de 2020 12:30, Tran Van Hoan
>  escreveu:
>
> > The problem reoccurs repeatly in recent days.
> > To day i tried dump heap and thread. Only dumping thread, heap can not
> > because solr instance was hang.
> > Almost thread was blocked.
> >
> > On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan
> >  wrote:
> >
> >
> > I checked node exporter metrics and saw network no problem
> >
> > On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan <
> > tranvanhoan...@yahoo.com> wrote:
> >
> >
> > I check node exporter, no problem with OS, hardware and network.
> > I attached images about solr metrics 7 days and 12h.
> >
> > On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin <
> > dario.rigo...@comperio.it> wrote:
> >
> >
> > What about a network issue?
> >
> > Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
> >  ha scritto:
> >
> >
> > >
> > > dear all,
> > >
> > >  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> > > instance has xmx = xms = 30G.
> > >
> > > Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> > > nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One
> node
> > > were down. almost metrics didn't increase too much except threads.
> > >
> > > Performance in one week ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > performace 12h ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > I go to the admin UI, some node dead some node too long to response.
> When
> > > checking logfile, they generate too much (log level warning), here are
> > logs
> > > which appears in the solr cloud:
> > >
> > > Log before server 4 and 6 down
> > >
> > > - Server 4 before it dead:
> > >
> > >    + o.a.s.h.RequestHandlerBase java.io.IOException:
> > > java.util.concurrent.TimeoutException: Idle timeout expired:
> > 12/12
> > > ms
> > >
> > >  +org.apache.solr.client.solrj.SolrServerException: Timeout occured
> while
> > > waiting response from server at:
> > > http://server6:8983/solr/mycollection_shard3_replica_n5/select
> > >
> > >
> > >
> > > at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
> > >
> > >                at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
> > >
> > >                at
> > > org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
> > >
> > >                at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
> > >
> > >                at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
> > >
> > >                at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >                at
> > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> > >
> > >                at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >                at
> > >
> >
> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.ja

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Rodrigo Oliveira
Network it's ok? Between nodes? The use? Swap it's disabled? Swapiness rhe
value it's 0?

Em sáb, 4 de jul de 2020 15:19, Tran Van Hoan
 escreveu:

>  I used physical servers, and IO wait is small :(!!!I saw that iptables
> dropped all ACK message from clients (not only client solr, prometheus
> scape metric from exporter was dropped too).all when i check netstat
> -anp|grep 8983, all socket are TIME_WAIT state.Only restart solrs, the
> incident was resolved. Total request around 2.5k request per second per
> node.
>
> On Sunday, July 5, 2020, 1:11:38 AM GMT+7, Rodrigo Oliveira <
> adamantina.rodr...@gmail.com> wrote:
>
>  Hi,
>
> I had this problem. In my case was the wait/io in vm. I migrate my
> environment to another place and solved.
>
> Actually it's problem wirh wait/io at host physical (until backup it's a
> problem over veeam).
>
> Regards
>
> Em sáb, 4 de jul de 2020 12:30, Tran Van Hoan
>  escreveu:
>
> > The problem reoccurs repeatly in recent days.
> > To day i tried dump heap and thread. Only dumping thread, heap can not
> > because solr instance was hang.
> > Almost thread was blocked.
> >
> > On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan
> >  wrote:
> >
> >
> > I checked node exporter metrics and saw network no problem
> >
> > On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan <
> > tranvanhoan...@yahoo.com> wrote:
> >
> >
> > I check node exporter, no problem with OS, hardware and network.
> > I attached images about solr metrics 7 days and 12h.
> >
> > On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin <
> > dario.rigo...@comperio.it> wrote:
> >
> >
> > What about a network issue?
> >
> > Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
> >  ha scritto:
> >
> >
> > >
> > > dear all,
> > >
> > >  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> > > instance has xmx = xms = 30G.
> > >
> > > Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> > > nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One
> node
> > > were down. almost metrics didn't increase too much except threads.
> > >
> > > Performance in one week ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > performace 12h ago:
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > I go to the admin UI, some node dead some node too long to response.
> When
> > > checking logfile, they generate too much (log level warning), here are
> > logs
> > > which appears in the solr cloud:
> > >
> > > Log before server 4 and 6 down
> > >
> > > - Server 4 before it dead:
> > >
> > >+ o.a.s.h.RequestHandlerBase java.io.IOException:
> > > java.util.concurrent.TimeoutException: Idle timeout expired:
> > 12/12
> > > ms
> > >
> > >  +org.apache.solr.client.solrj.SolrServerException: Timeout occured
> while
> > > waiting response from server at:
> > > http://server6:8983/solr/mycollection_shard3_replica_n5/select
> > >
> > >
> > >
> > > at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
> > >
> > >at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
> > >
> > >at
> > > org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
> > >
> > >at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
> > >
> > >at
> > >
> >
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
> > >
> > >at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >at
> > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> > >
> > >at
> > java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >
> > >at
> > >
> >
> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:181)
> > >
> > >at
> > >
> >
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
> > >
> > >at
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> > >
> > >at
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> > >
> > >... 1 more
> > >
> > > Caused by: java.util.concurrent.TimeoutException
> > >
> > >at
> > >
> >
> org.eclipse.jetty.client.util.InputStreamResponseListener.get(InputStreamResponseListener.java:216)
> > >
> > >at
> > >
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Tran Van Hoan
 I used physical servers, and IO wait is small :(!!!I saw that iptables dropped 
all ACK message from clients (not only client solr, prometheus scape metric 
from exporter was dropped too).all when i check netstat -anp|grep 8983, all 
socket are TIME_WAIT state.Only restart solrs, the incident was resolved. Total 
request around 2.5k request per second per node.

On Sunday, July 5, 2020, 1:11:38 AM GMT+7, Rodrigo Oliveira 
 wrote:  
 
 Hi,

I had this problem. In my case was the wait/io in vm. I migrate my
environment to another place and solved.

Actually it's problem wirh wait/io at host physical (until backup it's a
problem over veeam).

Regards

Em sáb, 4 de jul de 2020 12:30, Tran Van Hoan
 escreveu:

> The problem reoccurs repeatly in recent days.
> To day i tried dump heap and thread. Only dumping thread, heap can not
> because solr instance was hang.
> Almost thread was blocked.
>
> On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan
>  wrote:
>
>
> I checked node exporter metrics and saw network no problem
>
> On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan <
> tranvanhoan...@yahoo.com> wrote:
>
>
> I check node exporter, no problem with OS, hardware and network.
> I attached images about solr metrics 7 days and 12h.
>
> On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin <
> dario.rigo...@comperio.it> wrote:
>
>
> What about a network issue?
>
> Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
>  ha scritto:
>
>
> >
> > dear all,
> >
> >  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> > instance has xmx = xms = 30G.
> >
> > Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> > nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One node
> > were down. almost metrics didn't increase too much except threads.
> >
> > Performance in one week ago:
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > performace 12h ago:
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > I go to the admin UI, some node dead some node too long to response. When
> > checking logfile, they generate too much (log level warning), here are
> logs
> > which appears in the solr cloud:
> >
> > Log before server 4 and 6 down
> >
> > - Server 4 before it dead:
> >
> >    + o.a.s.h.RequestHandlerBase java.io.IOException:
> > java.util.concurrent.TimeoutException: Idle timeout expired:
> 12/12
> > ms
> >
> >  +org.apache.solr.client.solrj.SolrServerException: Timeout occured while
> > waiting response from server at:
> > http://server6:8983/solr/mycollection_shard3_replica_n5/select
> >
> >
> >
> > at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
> >
> >                at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
> >
> >                at
> > org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
> >
> >                at
> >
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
> >
> >                at
> >
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
> >
> >                at
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> >
> >                at
> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> >
> >                at
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> >
> >                at
> >
> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:181)
> >
> >                at
> >
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
> >
> >                at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> >
> >                at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> >
> >                ... 1 more
> >
> > Caused by: java.util.concurrent.TimeoutException
> >
> >                at
> >
> org.eclipse.jetty.client.util.InputStreamResponseListener.get(InputStreamResponseListener.java:216)
> >
> >                at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:397)
> >
> >                ... 12 more
> >
> >
> >
> > + o.a.s.s.HttpSolrCall invalid return code: -1
> >
> > + o.a.s.s.PKIAuthenticationPlugin Invalid key request timestamp:
> > 1592803662746 , received timestamp: 1592803796152 , TTL: 12
> >
> > + o.a.s.s.PKIAuthenticationPlugin Decryption failed , key must be wrong
> =>
> > java.security.InvalidKeyException: No installed provider supports this
> key:
> > (null)
> >
> > +  o.a.s.u.ErrorReportingConcurrentUpdateSolrClient Error when calling
> > SolrCmdDistributor$Req: cmd=delete{,commitWithin=-1}; node=ForwardNode:

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Rodrigo Oliveira
Hi,

I had this problem. In my case was the wait/io in vm. I migrate my
environment to another place and solved.

Actually it's problem wirh wait/io at host physical (until backup it's a
problem over veeam).

Regards

Em sáb, 4 de jul de 2020 12:30, Tran Van Hoan
 escreveu:

> The problem reoccurs repeatly in recent days.
> To day i tried dump heap and thread. Only dumping thread, heap can not
> because solr instance was hang.
> Almost thread was blocked.
>
> On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan
>  wrote:
>
>
> I checked node exporter metrics and saw network no problem
>
> On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan <
> tranvanhoan...@yahoo.com> wrote:
>
>
> I check node exporter, no problem with OS, hardware and network.
> I attached images about solr metrics 7 days and 12h.
>
> On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin <
> dario.rigo...@comperio.it> wrote:
>
>
> What about a network issue?
>
> Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
>  ha scritto:
>
>
> >
> > dear all,
> >
> >  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> > instance has xmx = xms = 30G.
> >
> > Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> > nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One node
> > were down. almost metrics didn't increase too much except threads.
> >
> > Performance in one week ago:
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > performace 12h ago:
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > I go to the admin UI, some node dead some node too long to response. When
> > checking logfile, they generate too much (log level warning), here are
> logs
> > which appears in the solr cloud:
> >
> > Log before server 4 and 6 down
> >
> > - Server 4 before it dead:
> >
> >+ o.a.s.h.RequestHandlerBase java.io.IOException:
> > java.util.concurrent.TimeoutException: Idle timeout expired:
> 12/12
> > ms
> >
> >  +org.apache.solr.client.solrj.SolrServerException: Timeout occured while
> > waiting response from server at:
> > http://server6:8983/solr/mycollection_shard3_replica_n5/select
> >
> >
> >
> > at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
> >
> >at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
> >
> >at
> > org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
> >
> >at
> >
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
> >
> >at
> >
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
> >
> >at
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> >
> >at
> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> >
> >at
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> >
> >at
> >
> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:181)
> >
> >at
> >
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
> >
> >at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> >
> >at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> >
> >... 1 more
> >
> > Caused by: java.util.concurrent.TimeoutException
> >
> >at
> >
> org.eclipse.jetty.client.util.InputStreamResponseListener.get(InputStreamResponseListener.java:216)
> >
> >at
> >
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:397)
> >
> >... 12 more
> >
> >
> >
> > + o.a.s.s.HttpSolrCall invalid return code: -1
> >
> > + o.a.s.s.PKIAuthenticationPlugin Invalid key request timestamp:
> > 1592803662746 , received timestamp: 1592803796152 , TTL: 12
> >
> > + o.a.s.s.PKIAuthenticationPlugin Decryption failed , key must be wrong
> =>
> > java.security.InvalidKeyException: No installed provider supports this
> key:
> > (null)
> >
> > +  o.a.s.u.ErrorReportingConcurrentUpdateSolrClient Error when calling
> > SolrCmdDistributor$Req: cmd=delete{,commitWithin=-1}; node=ForwardNode:
> > http://server6:8983/solr/mycollection_shard3_replica_n5/ to
> > http://server6:8983/solr/mycollection_shard3_replica_n5/ =>
> > java.util.concurrent.TimeoutException
> >
> > + o.a.s.s.HttpSolrCall
> >
> null:org.apache.solr.update.processor.DistributedUpdateProcessor$DistributedUpdatesAsyncException:
> > Async exception during distributed update: null
> >
> >
> >
> > Server 2:
> >
> >  + Max requests queued per destination 3000 exceede

Re: Almost nodes in Solrcloud dead suddently

2020-07-04 Thread Tran Van Hoan
 The problem reoccurs repeatly in recent days. 
To day i tried dump heap and thread. Only dumping thread, heap can not because 
solr instance was hang.Almost thread was blocked.

On Tuesday, June 23, 2020, 10:42:36 PM GMT+7, Tran Van Hoan 
 wrote:  
 
  I checked node exporter metrics and saw network no problem

On Tuesday, June 23, 2020, 8:37:41 PM GMT+7, Tran Van Hoan 
 wrote:  
 
  I check node exporter, no problem with OS, hardware and network.I attached 
images about solr metrics 7 days and 12h.

On Tuesday, June 23, 2020, 2:23:05 PM GMT+7, Dario Rigolin 
 wrote:  
 
 What about a network issue?

Il giorno mar 23 giu 2020 alle ore 01:37 Tran Van Hoan
 ha scritto:

>
> dear all,
>
>  I have a solr cloud 8.2.0 with 6 instance per 6 server (64G RAM), each
> instance has xmx = xms = 30G.
>
> Today almost nodes in the solrcloud were dead 2 times from 8:00AM (5/6
> nodes were down) and 1:00PM (2/6 nodes  were down). yesterday,  One node
> were down. almost metrics didn't increase too much except threads.
>
> Performance in one week ago:
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> performace 12h ago:
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> I go to the admin UI, some node dead some node too long to response. When
> checking logfile, they generate too much (log level warning), here are logs
> which appears in the solr cloud:
>
> Log before server 4 and 6 down
>
> - Server 4 before it dead:
>
>    + o.a.s.h.RequestHandlerBase java.io.IOException:
> java.util.concurrent.TimeoutException: Idle timeout expired: 12/12
> ms
>
>  +org.apache.solr.client.solrj.SolrServerException: Timeout occured while
> waiting response from server at:
> http://server6:8983/solr/mycollection_shard3_replica_n5/select
>
>
>
> at
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:406)
>
>                at
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:746)
>
>                at
> org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
>
>                at
> org.apache.solr.handler.component.HttpShardHandler.request(HttpShardHandler.java:238)
>
>                at
> org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
>
>                at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
>                at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>
>                at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
>                at
> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:181)
>
>                at
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
>
>                at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
>                at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
>                ... 1 more
>
> Caused by: java.util.concurrent.TimeoutException
>
>                at
> org.eclipse.jetty.client.util.InputStreamResponseListener.get(InputStreamResponseListener.java:216)
>
>                at
> org.apache.solr.client.solrj.impl.Http2SolrClient.request(Http2SolrClient.java:397)
>
>                ... 12 more
>
>
>
> + o.a.s.s.HttpSolrCall invalid return code: -1
>
> + o.a.s.s.PKIAuthenticationPlugin Invalid key request timestamp:
> 1592803662746 , received timestamp: 1592803796152 , TTL: 12
>
> + o.a.s.s.PKIAuthenticationPlugin Decryption failed , key must be wrong =>
> java.security.InvalidKeyException: No installed provider supports this key:
> (null)
>
> +  o.a.s.u.ErrorReportingConcurrentUpdateSolrClient Error when calling
> SolrCmdDistributor$Req: cmd=delete{,commitWithin=-1}; node=ForwardNode:
> http://server6:8983/solr/mycollection_shard3_replica_n5/ to
> http://server6:8983/solr/mycollection_shard3_replica_n5/ =>
> java.util.concurrent.TimeoutException
>
> + o.a.s.s.HttpSolrCall
> null:org.apache.solr.update.processor.DistributedUpdateProcessor$DistributedUpdatesAsyncException:
> Async exception during distributed update: null
>
>
>
> Server 2:
>
>  + Max requests queued per destination 3000 exceeded for
> HttpDestination[http://server4:8983
> ]@7d7ec93c,queue=3000,pool=MultiplexConnectionPool@73b938e3
> [c=4/4,b=4,m=0,i=0]
>
>  +  Max requests queued per destination 3000 exceeded for
> HttpDestination[http://server5:8983
> ]@7d7ec93c,queue=3000,pool=MultiplexConnectionPool@73b938e3
> [c=4/4,b=4,m=0,i=0]
>
>
>
> + Timeout occured while waiting response from server at:
> http://server4:8983/solr/mycollection_shard6_replica_n23/select
>
> + Timeout occured while waiting response from server at:
> http://server6:8983/solr/mycollection_shard2_replica_n15/select
>
> +  o.a.s.s.HttpSolrCall null:org.apache.solr.common.SolrException:
> org.apache.solr.client.solrj.SolrServerException: IOExcepti

Re: Questions about Solr Search

2020-07-04 Thread Nicolas Franck
Short answer: no

Neither Solr nor ElasticSearch have such capabilities out of the box.

Solr does have a plugin infrastructure that enables you to provide
better tokenization based on language rules, and some are better
than others.

I saw for example integration of openNLP here: 
https://lucene.apache.org/solr/guide/7_3/language-analysis.html
but that requires additional data to provide language rules,
but not a lot of languages have been tackled 
(http://opennlp.sourceforge.net/models-1.5/).

There are of course more general solutions like provided
in the first link, but the tokenization are a bit "rough" and aggressive (that 
is why it is so fast),
delivering tokens that do not exist (debugging shows you that).

Anyway: do not overdo it. There is no way
you can beat large sized company like Google.
Language support is more than a set of rules to apply.

More important is this: make sure Google can decently index
your webpages. More than 90% of your users reach your website through Google,
so do not invest too much into this.


On 2 Jul 2020, at 16:19, Gautam K mailto:gka...@gmail.com>> 
wrote:

Dear Team,

Hope you all are doing well.

Can you please help with the following question? We are using Solr search
in our Organisation and now checking whether Solr provides search
capabilities like Google Enterprise search(Google Knowledge Graph Search).

1, Does Solr Search provide Voice Search like Google?
2. Does Solar Search provide NLP Search(Natural Language Processing)?
3. Does Solr have all the capabilities which Google Knowledge Graph
provides like below?


  - Getting a ranked list of the most notable entities that match certain
  criteria.
  - Predictively completing entities in a search box.
  - Annotating/organizing content using the Knowledge Graph entities.


*Your help will be appreciated highly.*

Many thanks
Gautam Kanaujia
India



Re: Time-out errors while indexing (Solr 7.7.1)

2020-07-04 Thread Erick Erickson
You need more shards. And, I’m pretty certain, more hardware.

You say you have 13 billion documents and 6 shards. Solr/Lucene has an absolute 
upper limit of 2B (2^31) docs per shard. I don’t quite know how you’re running 
at all unless that 13B is a round number. If you keep adding documents, your 
installation will shortly, at best, stop accepting new documents for indexing. 
At worst you’ll start seeing weird errors and possibly corrupt indexes and have 
to re-index everything from scratch.

You’ve backed yourself in to a pretty tight corner here. You either have to 
re-index to a properly-sized cluster or use SPLITSHARD. This latter will double 
the index-on-disk size (it creates two child indexes per replica and keeps the 
old one for safety’s sake that you have to clean up later). I strongly 
recommend you stop ingesting more data while you do this.

You say you have 6 VMs with 2 nodes running on each. If those VMs are 
co-located with anything else, the physical hardware is going to be stressed. 
VMs themselves aren’t bad, but somewhere there’s physical hardware that runs it…

In fact, I urge you to stop ingesting data immediately and address this issue. 
You have a cluster that’s mis-configured, and you must address that before Bad 
Things Happen.

Best,
Erick

> On Jul 4, 2020, at 5:09 AM, Mad have  wrote:
> 
> Hi Eric,
> 
> There are total 6 VM’s in Solr clusters and 2 nodes are running on each VM. 
> Total number of shards are 6 with 3 replicas. I can see the index size is 
> more than 220GB on each node for the collection where we are facing the 
> performance issue.
> 
> The more documents we add to the collection the indexing become slow and I 
> also have same impression that the size of the collection is creating this 
> issue. Appreciate if you can suggests any solution on this.
> 
> 
> Regards,
> Madhava 
> Sent from my iPhone
> 
>> On 3 Jul 2020, at 23:30, Erick Erickson  wrote:
>> 
>> Oops, I transposed that. If your index is a terabyte and your RAM is 128M, 
>> _that’s_ a red flag.
>> 
>>> On Jul 3, 2020, at 5:53 PM, Erick Erickson  wrote:
>>> 
>>> You haven’t said how many _shards_ are present. Nor how many replicas of 
>>> the collection you’re hosting per physical machine. Nor how large the 
>>> indexes are on disk. Those are the numbers that count. The latter is 
>>> somewhat fuzzy, but if your aggregate index size on a machine with, say, 
>>> 128G of memory is a terabyte, that’s a red flag.
>>> 
>>> Short form, though is yes. Subject to the questions above, this is what I’d 
>>> be looking at first.
>>> 
>>> And, as I said, if you’ve been steadily increasing the total number of 
>>> documents, you’ll reach a tipping point sometime.
>>> 
>>> Best,
>>> Erick
>>> 
> On Jul 3, 2020, at 5:32 PM, Mad have  wrote:
 
 Hi Eric,
 
 The collection has almost 13billion documents with each document around 
 5kb size, all the columns around 150 are the indexed. Do you think that 
 number of documents in the collection causing this issue. Appreciate your 
 response.
 
 Regards,
 Madhava 
 
 Sent from my iPhone
 
> On 3 Jul 2020, at 12:42, Erick Erickson  wrote:
> 
> If you’re seeing low CPU utilization at the same time, you probably
> just have too much data on too little hardware. Check your
> swapping, how much of your I/O is just because Lucene can’t
> hold all the parts of the index it needs in memory at once? Lucene
> uses MMapDirectory to hold the index and you may well be
> swapping, see:
> 
> https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
> 
> But my guess is that you’ve just reached a tipping point. You say:
> 
> "From last 2-3 weeks we have been noticing either slow indexing or 
> timeout errors while indexing”
> 
> So have you been continually adding more documents to your
> collections for more than the 2-3 weeks? If so you may have just
> put so much data on the same boxes that you’ve gone over
> the capacity of your hardware. As Toke says, adding physical
> memory for the OS to use to hold relevant parts of the index may
> alleviate the problem (again, refer to Uwe’s article for why).
> 
> All that said, if you’re going to keep adding document you need to
> seriously think about adding new machines and moving some of
> your replicas to them.
> 
> Best,
> Erick
> 
>> On Jul 3, 2020, at 7:14 AM, Toke Eskildsen  wrote:
>> 
>>> On Thu, 2020-07-02 at 11:16 +, Kommu, Vinodh K. wrote:
>>> We are performing QA performance testing on couple of collections
>>> which holds 2 billion and 3.5 billion docs respectively.
>> 
>> How many shards?
>> 
>>> 1.  Our performance team noticed that read operations are pretty
>>> more than write operations like 100:1 ratio, is this expected during
>>> indexing or solr nodes are doing any other operations lik

Re: unified highlighter performance in solr 8.5.1

2020-07-04 Thread Nándor Mátravölgyi
I guess that's fair. Let's have hl.fragsizeIsMinimum=true as default.

On 7/4/20, David Smiley  wrote:
> I doubt that WORD mode is impacted much by hl.fragsizeIsMinimum in terms of
> quality of the highlight since there are vastly more breaks to pick from.
> I think that setting is more useful in SENTENCE mode if you can stand the
> perf hit.  If you agree, then why not just let this one default to "true"?
>
> We agree on better documenting the perf trade-off.
>
> Thanks again for working on these settings, BTW.
>
> ~ David
>
>
> On Fri, Jul 3, 2020 at 1:25 PM Nándor Mátravölgyi 
> wrote:
>
>> Since the issue seems to be affecting the highlighter differently
>> based on which mode it is using, having different defaults for the
>> modes could be explored.
>>
>> WORD may have the new defaults as it has little effect on performance
>> and it creates nicer highlights.
>> SENTENCE should have the defaults that produce reasonable performance.
>> The docs could document this while also mentioning that the UH's
>> performance is highly dependent on the underlying Java String/Text?
>> Iterator.
>>
>> One can argue that having different defaults based on mode is
>> confusing. In this case I think the defaults should be changed to have
>> the SENTENCE mode perform better. Maybe the options for nice
>> highlights with WORD mode could be put into the docs in this case as
>> some form of an example.
>>
>> As long as I can use the UH with nicely aligned snippets in WORD mode
>> I'm fine with any defaults. I explicitly set them in the config and in
>> the queries most of the time anyways.
>>
>


Re: Time-out errors while indexing (Solr 7.7.1)

2020-07-04 Thread Mad have
Hi Eric,

There are total 6 VM’s in Solr clusters and 2 nodes are running on each VM. 
Total number of shards are 6 with 3 replicas. I can see the index size is more 
than 220GB on each node for the collection where we are facing the performance 
issue.

The more documents we add to the collection the indexing become slow and I also 
have same impression that the size of the collection is creating this issue. 
Appreciate if you can suggests any solution on this.


Regards,
Madhava 
Sent from my iPhone

> On 3 Jul 2020, at 23:30, Erick Erickson  wrote:
> 
> Oops, I transposed that. If your index is a terabyte and your RAM is 128M, 
> _that’s_ a red flag.
> 
>> On Jul 3, 2020, at 5:53 PM, Erick Erickson  wrote:
>> 
>> You haven’t said how many _shards_ are present. Nor how many replicas of the 
>> collection you’re hosting per physical machine. Nor how large the indexes 
>> are on disk. Those are the numbers that count. The latter is somewhat fuzzy, 
>> but if your aggregate index size on a machine with, say, 128G of memory is a 
>> terabyte, that’s a red flag.
>> 
>> Short form, though is yes. Subject to the questions above, this is what I’d 
>> be looking at first.
>> 
>> And, as I said, if you’ve been steadily increasing the total number of 
>> documents, you’ll reach a tipping point sometime.
>> 
>> Best,
>> Erick
>> 
 On Jul 3, 2020, at 5:32 PM, Mad have  wrote:
>>> 
>>> Hi Eric,
>>> 
>>> The collection has almost 13billion documents with each document around 5kb 
>>> size, all the columns around 150 are the indexed. Do you think that number 
>>> of documents in the collection causing this issue. Appreciate your response.
>>> 
>>> Regards,
>>> Madhava 
>>> 
>>> Sent from my iPhone
>>> 
 On 3 Jul 2020, at 12:42, Erick Erickson  wrote:
 
 If you’re seeing low CPU utilization at the same time, you probably
 just have too much data on too little hardware. Check your
 swapping, how much of your I/O is just because Lucene can’t
 hold all the parts of the index it needs in memory at once? Lucene
 uses MMapDirectory to hold the index and you may well be
 swapping, see:
 
 https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
 
 But my guess is that you’ve just reached a tipping point. You say:
 
 "From last 2-3 weeks we have been noticing either slow indexing or timeout 
 errors while indexing”
 
 So have you been continually adding more documents to your
 collections for more than the 2-3 weeks? If so you may have just
 put so much data on the same boxes that you’ve gone over
 the capacity of your hardware. As Toke says, adding physical
 memory for the OS to use to hold relevant parts of the index may
 alleviate the problem (again, refer to Uwe’s article for why).
 
 All that said, if you’re going to keep adding document you need to
 seriously think about adding new machines and moving some of
 your replicas to them.
 
 Best,
 Erick
 
> On Jul 3, 2020, at 7:14 AM, Toke Eskildsen  wrote:
> 
>> On Thu, 2020-07-02 at 11:16 +, Kommu, Vinodh K. wrote:
>> We are performing QA performance testing on couple of collections
>> which holds 2 billion and 3.5 billion docs respectively.
> 
> How many shards?
> 
>> 1.  Our performance team noticed that read operations are pretty
>> more than write operations like 100:1 ratio, is this expected during
>> indexing or solr nodes are doing any other operations like syncing?
> 
> Are you saying that there are 100 times more read operations when you
> are indexing? That does not sound too unrealistic as the disk cache
> might be filled with the data that the writers are flushing.
> 
> In that case, more RAM would help. Okay, more RAM nearly always helps,
> but such massive difference in IO-utilization does indicate that you
> are starved for cache.
> 
> I noticed you have at least 18 replicas. That's a lot. Just to sanity
> check: How many replicas are each physical box handling? If they are
> sharing resources, fewer replicas would probably be better.
> 
>> 3.  Our client timeout is set to 2mins, can they increase further
>> more? Would that help or create any other problems?
> 
> It does not hurt the server to increase the client timeout as the
> initiated query will keep running until it is finished, independent of
> whether or not there is a client to receive the result.
> 
> If you want a better max time for query processing, you should look at 
> 
> https://lucene.apache.org/solr/guide/7_7/common-query-parameters.html#timeallowed-parameter
> but due to its inherent limitations it might not help in your
> situation.
> 
>> 4.  When we created an empty collection and loaded same data file,
>> it loaded fine without any issues so having more documents in a
>> collection would 

***URGENT***Re: Questions about Solr Search

2020-07-04 Thread Gautam K
Since it's a bit of an urgent request so if could please help me on this by
today it will be highly appreciated.

Thanks & Regards,
Gautam Kanaujia

On Thu, Jul 2, 2020 at 7:49 PM Gautam K  wrote:

> Dear Team,
>
> Hope you all are doing well.
>
> Can you please help with the following question? We are using Solr search
> in our Organisation and now checking whether Solr provides search
> capabilities like Google Enterprise search(Google Knowledge Graph Search).
>
> 1, Does Solr Search provide Voice Search like Google?
> 2. Does Solar Search provide NLP Search(Natural Language Processing)?
> 3. Does Solr have all the capabilities which Google Knowledge Graph
> provides like below?
>
>
>- Getting a ranked list of the most notable entities that match
>certain criteria.
>- Predictively completing entities in a search box.
>- Annotating/organizing content using the Knowledge Graph entities.
>
>
> *Your help will be appreciated highly.*
>
> Many thanks
> Gautam Kanaujia
> India
>


Questions about Solr Search

2020-07-04 Thread Gautam K
Dear Team,

Hope you all are doing well.

Can you please help with the following question? We are using Solr search
in our Organisation and now checking whether Solr provides search
capabilities like Google Enterprise search(Google Knowledge Graph Search).

1, Does Solr Search provide Voice Search like Google?
2. Does Solar Search provide NLP Search(Natural Language Processing)?
3. Does Solr have all the capabilities which Google Knowledge Graph
provides like below?


   - Getting a ranked list of the most notable entities that match certain
   criteria.
   - Predictively completing entities in a search box.
   - Annotating/organizing content using the Knowledge Graph entities.


*Your help will be appreciated highly.*

Many thanks
Gautam Kanaujia
India