Re: Solr 8.2 docker image in cloud mode not connecting to Zookeeper on startup

2019-10-17 Thread Rajeswari Natarajan
Are you running zookeeper as container too. If yes , port 2181 needs to be
exposed.

-Rajeswari

On Thu, Oct 17, 2019 at 4:25 PM Drew Kidder  wrote:

> Hello! I'm new to the list and I have a bit of an issue that I could use
> some help with.
>
> I'm in the process of upgrading our Solr installation from legacy to cloud.
> I'm new to the idea of Solr Cloud, so I've been wading through the
> documentation and trying to get a basic cluster up and running. I've got my
> Zookeeper ensemble set up, talking to each other,  and accessible to my
> network via DNS hostnames. I'm using the official Solr 8.2 docker image
> from docker hub. Please see more environment information following the
> gigantic stack trace below. I've tried to run the docker image both locally
> and in the same Amazon VPC as the ZK ensemble, but every time it tries to
> start up I get this message in both cases:
>
> 2019-10-17 22:30:03.443 INFO  (main) [   ] o.a.s.c.c.ConnectionManager
> Waiting for client to connect to ZooKeeper
> 2019-10-17 22:30:23.539 WARN  (main-SendThread(zk1:2181)) [   ]
> o.a.z.ClientCnxn Client session timed out, have not heard from server in
> 20095ms for sessionid 0x0
> 2019-10-17 22:30:43.612 WARN  (main-SendThread(zk3:2181)) [   ]
> o.a.z.ClientCnxn Client session timed out, have not heard from server in
> 20005ms for sessionid 0x0
> 2019-10-17 22:30:43.724 ERROR (main-EventThread) [   ] o.a.z.ClientCnxn
> Error while calling watcher  =>
> java.util.concurrent.RejectedExecutionException: Task
>
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
> rejected from
>
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea
> [Terminated,
> pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
> at
>
> java.base/java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
> java.util.concurrent.RejectedExecutionException: Task
>
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
> rejected from
>
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea
> [Terminated,
> pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
> at
>
> java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
> ~[?:?]
> at
> java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:825)
> ~[?:?]
> at
>
> java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1355)
> ~[?:?]
> at
>
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.execute(ExecutorUtil.java:194)
> ~[?:?]
> at
>
> java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:118)
> ~[?:?]
> at
>
> org.apache.solr.common.cloud.SolrZkClient$ProcessWatchWithExecutor.process(SolrZkClient.java:843)
> ~[?:?]
> at
>
> org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:535)
> ~[?:?]
> at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:510)
> ~[?:?]
> 2019-10-17 22:30:43.742 ERROR (main) [   ] o.a.s.s.SolrDispatchFilter Could
> not start Solr. Check solr/home property and the logs
> 2019-10-17 22:30:43.818 ERROR (main) [   ] o.a.s.c.SolrCore
> null:org.apache.solr.common.SolrException: Error occurred while loading
> solr.xml from zookeeper
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.loadNodeConfig(SolrDispatchFilter.java:289)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.createCoreContainer(SolrDispatchFilter.java:259)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.init(SolrDispatchFilter.java:181)
> at org.eclipse.jetty.servlet.FilterHolder.initialize(FilterHolder.java:136)
> at
>
> org.eclipse.jetty.servlet.ServletHandler.lambda$initialize$0(ServletHandler.java:750)
> at
>
> java.base/java.util.Spliterators$ArraySpliterator.forEachRemaining(Spliterators.java:948)
> at
>
> java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
> at
>
> java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
> at
>
> java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:658)
> at
>
> org.eclipse.jetty.servlet.ServletHandler.initialize(ServletHandler.java:744)
> at
>
> org.eclipse.jetty.servlet.ServletContextHandler.startContext(ServletContextHandler.java:369)
> at
> org.eclipse.jetty.webapp.WebAppContext.startWebapp(WebAppContext.java:1497)
> at
>
> org.eclipse.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1459)
> at
>
> org.eclipse.jetty.server.handler.ContextHandler.doStart(ContextHandler.java:854)
> at
>
> org.eclipse.jetty.servlet.ServletContextHandler.doStart(ServletContextHandler.java:278)
> at org.eclipse.jetty.webapp.WebAppContext.doStart(WebAppContext.java:545)
> at
>
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
> at
>
> 

Re: Solr 8.2 docker image in cloud mode not connecting to Zookeeper on startup

2019-10-17 Thread Jörn Franke
Could it be that you start the Solr command too early, ie before the network is 
setup in the Docker container? 

Normally I would also expect that a zkRoot Is specified.
Can the Zknodes talk to each other?
Have you tried to specify it in the Solr config?
Normally, I would expect that the Solr config is external to the container, 
especially later when you secure it. Eg you would not put certificates etc 
directly in the container as this is not a secure practice.

> Am 18.10.2019 um 01:25 schrieb Drew Kidder :
> 
> Hello! I'm new to the list and I have a bit of an issue that I could use
> some help with.
> 
> I'm in the process of upgrading our Solr installation from legacy to cloud.
> I'm new to the idea of Solr Cloud, so I've been wading through the
> documentation and trying to get a basic cluster up and running. I've got my
> Zookeeper ensemble set up, talking to each other,  and accessible to my
> network via DNS hostnames. I'm using the official Solr 8.2 docker image
> from docker hub. Please see more environment information following the
> gigantic stack trace below. I've tried to run the docker image both locally
> and in the same Amazon VPC as the ZK ensemble, but every time it tries to
> start up I get this message in both cases:
> 
> 2019-10-17 22:30:03.443 INFO  (main) [   ] o.a.s.c.c.ConnectionManager
> Waiting for client to connect to ZooKeeper
> 2019-10-17 22:30:23.539 WARN  (main-SendThread(zk1:2181)) [   ]
> o.a.z.ClientCnxn Client session timed out, have not heard from server in
> 20095ms for sessionid 0x0
> 2019-10-17 22:30:43.612 WARN  (main-SendThread(zk3:2181)) [   ]
> o.a.z.ClientCnxn Client session timed out, have not heard from server in
> 20005ms for sessionid 0x0
> 2019-10-17 22:30:43.724 ERROR (main-EventThread) [   ] o.a.z.ClientCnxn
> Error while calling watcher  =>
> java.util.concurrent.RejectedExecutionException: Task
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
> rejected from
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea[Terminated,
> pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
> at
> java.base/java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
> java.util.concurrent.RejectedExecutionException: Task
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
> rejected from
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea[Terminated,
> pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
> at
> java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
> ~[?:?]
> at
> java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:825)
> ~[?:?]
> at
> java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1355)
> ~[?:?]
> at
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.execute(ExecutorUtil.java:194)
> ~[?:?]
> at
> java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:118)
> ~[?:?]
> at
> org.apache.solr.common.cloud.SolrZkClient$ProcessWatchWithExecutor.process(SolrZkClient.java:843)
> ~[?:?]
> at
> org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:535)
> ~[?:?]
> at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:510)
> ~[?:?]
> 2019-10-17 22:30:43.742 ERROR (main) [   ] o.a.s.s.SolrDispatchFilter Could
> not start Solr. Check solr/home property and the logs
> 2019-10-17 22:30:43.818 ERROR (main) [   ] o.a.s.c.SolrCore
> null:org.apache.solr.common.SolrException: Error occurred while loading
> solr.xml from zookeeper
> at
> org.apache.solr.servlet.SolrDispatchFilter.loadNodeConfig(SolrDispatchFilter.java:289)
> at
> org.apache.solr.servlet.SolrDispatchFilter.createCoreContainer(SolrDispatchFilter.java:259)
> at
> org.apache.solr.servlet.SolrDispatchFilter.init(SolrDispatchFilter.java:181)
> at org.eclipse.jetty.servlet.FilterHolder.initialize(FilterHolder.java:136)
> at
> org.eclipse.jetty.servlet.ServletHandler.lambda$initialize$0(ServletHandler.java:750)
> at
> java.base/java.util.Spliterators$ArraySpliterator.forEachRemaining(Spliterators.java:948)
> at
> java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
> at
> java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
> at
> java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:658)
> at
> org.eclipse.jetty.servlet.ServletHandler.initialize(ServletHandler.java:744)
> at
> org.eclipse.jetty.servlet.ServletContextHandler.startContext(ServletContextHandler.java:369)
> at
> org.eclipse.jetty.webapp.WebAppContext.startWebapp(WebAppContext.java:1497)
> at
> org.eclipse.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1459)
> at
> 

Solr 8.2 docker image in cloud mode not connecting to Zookeeper on startup

2019-10-17 Thread Drew Kidder
Hello! I'm new to the list and I have a bit of an issue that I could use
some help with.

I'm in the process of upgrading our Solr installation from legacy to cloud.
I'm new to the idea of Solr Cloud, so I've been wading through the
documentation and trying to get a basic cluster up and running. I've got my
Zookeeper ensemble set up, talking to each other,  and accessible to my
network via DNS hostnames. I'm using the official Solr 8.2 docker image
from docker hub. Please see more environment information following the
gigantic stack trace below. I've tried to run the docker image both locally
and in the same Amazon VPC as the ZK ensemble, but every time it tries to
start up I get this message in both cases:

2019-10-17 22:30:03.443 INFO  (main) [   ] o.a.s.c.c.ConnectionManager
Waiting for client to connect to ZooKeeper
2019-10-17 22:30:23.539 WARN  (main-SendThread(zk1:2181)) [   ]
o.a.z.ClientCnxn Client session timed out, have not heard from server in
20095ms for sessionid 0x0
2019-10-17 22:30:43.612 WARN  (main-SendThread(zk3:2181)) [   ]
o.a.z.ClientCnxn Client session timed out, have not heard from server in
20005ms for sessionid 0x0
2019-10-17 22:30:43.724 ERROR (main-EventThread) [   ] o.a.z.ClientCnxn
Error while calling watcher  =>
java.util.concurrent.RejectedExecutionException: Task
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
rejected from
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea[Terminated,
pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
at
java.base/java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
java.util.concurrent.RejectedExecutionException: Task
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$186/0x000100328440@5b1d0665
rejected from
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@64e89eea[Terminated,
pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
at
java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2055)
~[?:?]
at
java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:825)
~[?:?]
at
java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1355)
~[?:?]
at
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.execute(ExecutorUtil.java:194)
~[?:?]
at
java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:118)
~[?:?]
at
org.apache.solr.common.cloud.SolrZkClient$ProcessWatchWithExecutor.process(SolrZkClient.java:843)
~[?:?]
at
org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:535)
~[?:?]
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:510)
~[?:?]
2019-10-17 22:30:43.742 ERROR (main) [   ] o.a.s.s.SolrDispatchFilter Could
not start Solr. Check solr/home property and the logs
2019-10-17 22:30:43.818 ERROR (main) [   ] o.a.s.c.SolrCore
null:org.apache.solr.common.SolrException: Error occurred while loading
solr.xml from zookeeper
at
org.apache.solr.servlet.SolrDispatchFilter.loadNodeConfig(SolrDispatchFilter.java:289)
at
org.apache.solr.servlet.SolrDispatchFilter.createCoreContainer(SolrDispatchFilter.java:259)
at
org.apache.solr.servlet.SolrDispatchFilter.init(SolrDispatchFilter.java:181)
at org.eclipse.jetty.servlet.FilterHolder.initialize(FilterHolder.java:136)
at
org.eclipse.jetty.servlet.ServletHandler.lambda$initialize$0(ServletHandler.java:750)
at
java.base/java.util.Spliterators$ArraySpliterator.forEachRemaining(Spliterators.java:948)
at
java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
at
java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
at
java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:658)
at
org.eclipse.jetty.servlet.ServletHandler.initialize(ServletHandler.java:744)
at
org.eclipse.jetty.servlet.ServletContextHandler.startContext(ServletContextHandler.java:369)
at
org.eclipse.jetty.webapp.WebAppContext.startWebapp(WebAppContext.java:1497)
at
org.eclipse.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1459)
at
org.eclipse.jetty.server.handler.ContextHandler.doStart(ContextHandler.java:854)
at
org.eclipse.jetty.servlet.ServletContextHandler.doStart(ServletContextHandler.java:278)
at org.eclipse.jetty.webapp.WebAppContext.doStart(WebAppContext.java:545)
at
org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
at
org.eclipse.jetty.deploy.bindings.StandardStarter.processBinding(StandardStarter.java:46)
at org.eclipse.jetty.deploy.AppLifeCycle.runBindings(AppLifeCycle.java:192)
at
org.eclipse.jetty.deploy.DeploymentManager.requestAppGoal(DeploymentManager.java:510)
at
org.eclipse.jetty.deploy.DeploymentManager.addApp(DeploymentManager.java:153)
at

Japanese Query Unexpectedly Misses

2019-10-17 Thread Stephen Lewis Bianamara
Hi SOLR Community,

I have an example of a basic Japanese indexing/recall scenario which I am
trying to support, but cannot get to work.

The scenario is: I would like for 日本人 (Japanese Person) to be matched by
either 日本 (Japan) or 人 (Person). Currently, I am not seeing this work. My
Japanese text field currently has the tokenizer

> 
>
What is most surprising to me is that I though this is what mode="search"
was made for. From the docs, I see

> Use search mode to get a noun-decompounding effect useful for search.
> search mode improves segmentation for search at the expense of
> part-of-speech accuracy
>

I analyzed the breakdown, and I can see that the tokenizer is not
generating three tokens (one for Japan, one for person, and one for
Japanese Person) as I would have expected. Interestingly, the tokenizer
does recognize that  日本人 is a compound noun, so it would seem to be that it
should decompound it (see image below).

Can you help me figure out if my configuration is incorrect, or if there is
some way to fix this scenario?

Thanks!
Stephen


[image: image.png]


Re: Help with Stream Graph

2019-10-17 Thread Rajeswari Natarajan
My config is from

https://github.com/apache/lucene-solr/tree/branch_7_6/solr/solrj/src/test-files/solrj/solr/configsets/streaming/conf














Thanks,

Rajeswari

On Thu, Oct 17, 2019 at 8:16 AM Rajeswari Natarajan 
wrote:

> I tried below query  and it returns o results
>
>
> http://localhost:8983/solr/knr/export?{!terms+f%3Dproduct_s}product1=false=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2
> 
>
>
> {
>   "responseHeader":{"status":0},
>   "response":{
> "numFound":0,
> "docs":[]}}
>
> Regards,
> Rajeswari
> On Thu, Oct 17, 2019 at 8:05 AM Rajeswari Natarajan 
> wrote:
>
>> Thanks Joel.
>>
>> Here is the logs for below request
>>
>> curl --data-urlencode
>> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
>> http://localhost:8983/solr/knr/stream
>>
>> 2019-10-17 15:02:06.969 INFO  (qtp952486988-280) [c:knr s:shard1
>> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
>> [knr_shard1_replica_n1]  webapp=/solr path=/stream
>> params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s")}
>> status=0 QTime=0
>>
>> 2019-10-17 15:02:06.975 INFO  (qtp952486988-192) [c:knr s:shard1
>> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
>> [knr_shard1_replica_n1]  webapp=/solr path=/export
>> params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
>> hits=0 status=0 QTime=1
>>
>>
>>
>> Here is the logs for
>>
>>
>>
>> curl --data-urlencode
>> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
>> leaves")' http://localhost:8983/solr/knr/stream
>>
>>
>> 2019-10-17 15:03:57.068 INFO  (qtp952486988-356) [c:knr s:shard1
>> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
>> [knr_shard1_replica_n1]  webapp=/solr path=/stream
>> params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s",scatter%3D"branches,+leaves")}
>> status=0 QTime=0
>>
>> 2019-10-17 15:03:57.071 INFO  (qtp952486988-400) [c:knr s:shard1
>> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
>> [knr_shard1_replica_n1]  webapp=/solr path=/export
>> params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
>> hits=0 status=0 QTime=0
>>
>>
>>
>>
>> Thank you,
>>
>> Rajeswari
>>
>> On Thu, Oct 17, 2019 at 5:23 AM Joel Bernstein 
>> wrote:
>>
>>> Can you show the logs from this request. There will be a Solr query that
>>> gets sent with product1 searched against the product_s field. Let's see
>>> how
>>> many documents that query returns.
>>>
>>>
>>> Joel Bernstein
>>> http://joelsolr.blogspot.com/
>>>
>>>
>>> On Thu, Oct 17, 2019 at 1:41 AM Rajeswari Natarajan 
>>> wrote:
>>>
>>> > Hi,
>>> >
>>> > Since the stream graph query for my use case , didn't work as  i took
>>> the
>>> > data from solr source code test and also copied the schema and
>>> > solrconfig.xml from solr 7.6 source code.  Had to substitute few
>>> variables.
>>> >
>>> > Posted below data
>>> >
>>> > curl -X POST http://localhost:8983/solr/knr/update -H
>>> > 'Content-type:text/csv' -d '
>>> > id, basket_s, product_s, prics_f
>>> > 90,basket1,product1,20
>>> > 91,basket1,product3,30
>>> > 92,basket1,product5,1
>>> > 93,basket2,product1,2
>>> > 94,basket2,product6,5
>>> > 95,basket2,product7,10
>>> > 96,basket3,product4,20
>>> > 97,basket3,product3,10
>>> > 98,basket3,product1,10
>>> > 99,basket4,product4,40
>>> > 110,basket4,product3,10
>>> > 111,basket4,product1,10'
>>> > After this I committed and made sure the data got published. to solr
>>> >
>>> > curl --data-urlencode
>>> > 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
>>> > http://localhost:8983/solr/knr/stream
>>> >
>>> > {
>>> >
>>> >   "result-set":{
>>> >
>>> > "docs":[{
>>> >
>>> > "EOF":true,
>>> >
>>> > "RESPONSE_TIME":4}]}}
>>> >
>>> >
>>> > and if I add *scatter="branches, leaves" , there is one doc.*
>>> >
>>> >
>>> >
>>> > curl --data-urlencode
>>> >
>>> >
>>> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
>>> > leaves")' http://localhost:8983/solr/knr/stream
>>> >
>>> > {
>>> >
>>> >   "result-set":{
>>> >
>>> > "docs":[{
>>> >
>>> > "node":"product1",
>>> >
>>> > "collection":"knr",
>>> >
>>> > "field":"node",
>>> >
>>> > "level":0}
>>> >
>>> >   ,{
>>> >
>>> > "EOF":true,
>>> >
>>> > "RESPONSE_TIME":4}]}}
>>> >
>>> >
>>> >
>>> >
>>> > Below is the data I got from
>>> >
>>> >
>>> https://github.com/apache/lucene-solr/blob/branch_7_6/solr/solrj/src/test/org/apache/solr/client/solrj/io/graph/GraphExpressionTest.java#L271
>>> >
>>> >
>>> >
>>> > According to this test 4 docs are expected.
>>> >
>>> >
>>> > I am not sure what I am missing. Any pointers, please
>>> >
>>> >
>>> > Thanks you,
>>> >
>>> > Rajeswari
>>> >

Re: Help with Stream Graph

2019-10-17 Thread Rajeswari Natarajan
I tried below query  and it returns o results

http://localhost:8983/solr/knr/export?{!terms+f%3Dproduct_s}product1=false=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2


{
  "responseHeader":{"status":0},
  "response":{
"numFound":0,
"docs":[]}}

Regards,
Rajeswari
On Thu, Oct 17, 2019 at 8:05 AM Rajeswari Natarajan 
wrote:

> Thanks Joel.
>
> Here is the logs for below request
>
> curl --data-urlencode
> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
> http://localhost:8983/solr/knr/stream
>
> 2019-10-17 15:02:06.969 INFO  (qtp952486988-280) [c:knr s:shard1
> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
> [knr_shard1_replica_n1]  webapp=/solr path=/stream
> params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s")}
> status=0 QTime=0
>
> 2019-10-17 15:02:06.975 INFO  (qtp952486988-192) [c:knr s:shard1
> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
> [knr_shard1_replica_n1]  webapp=/solr path=/export
> params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
> hits=0 status=0 QTime=1
>
>
>
> Here is the logs for
>
>
>
> curl --data-urlencode
> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
> leaves")' http://localhost:8983/solr/knr/stream
>
>
> 2019-10-17 15:03:57.068 INFO  (qtp952486988-356) [c:knr s:shard1
> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
> [knr_shard1_replica_n1]  webapp=/solr path=/stream
> params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s",scatter%3D"branches,+leaves")}
> status=0 QTime=0
>
> 2019-10-17 15:03:57.071 INFO  (qtp952486988-400) [c:knr s:shard1
> r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
> [knr_shard1_replica_n1]  webapp=/solr path=/export
> params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
> hits=0 status=0 QTime=0
>
>
>
>
> Thank you,
>
> Rajeswari
>
> On Thu, Oct 17, 2019 at 5:23 AM Joel Bernstein  wrote:
>
>> Can you show the logs from this request. There will be a Solr query that
>> gets sent with product1 searched against the product_s field. Let's see
>> how
>> many documents that query returns.
>>
>>
>> Joel Bernstein
>> http://joelsolr.blogspot.com/
>>
>>
>> On Thu, Oct 17, 2019 at 1:41 AM Rajeswari Natarajan 
>> wrote:
>>
>> > Hi,
>> >
>> > Since the stream graph query for my use case , didn't work as  i took
>> the
>> > data from solr source code test and also copied the schema and
>> > solrconfig.xml from solr 7.6 source code.  Had to substitute few
>> variables.
>> >
>> > Posted below data
>> >
>> > curl -X POST http://localhost:8983/solr/knr/update -H
>> > 'Content-type:text/csv' -d '
>> > id, basket_s, product_s, prics_f
>> > 90,basket1,product1,20
>> > 91,basket1,product3,30
>> > 92,basket1,product5,1
>> > 93,basket2,product1,2
>> > 94,basket2,product6,5
>> > 95,basket2,product7,10
>> > 96,basket3,product4,20
>> > 97,basket3,product3,10
>> > 98,basket3,product1,10
>> > 99,basket4,product4,40
>> > 110,basket4,product3,10
>> > 111,basket4,product1,10'
>> > After this I committed and made sure the data got published. to solr
>> >
>> > curl --data-urlencode
>> > 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
>> > http://localhost:8983/solr/knr/stream
>> >
>> > {
>> >
>> >   "result-set":{
>> >
>> > "docs":[{
>> >
>> > "EOF":true,
>> >
>> > "RESPONSE_TIME":4}]}}
>> >
>> >
>> > and if I add *scatter="branches, leaves" , there is one doc.*
>> >
>> >
>> >
>> > curl --data-urlencode
>> >
>> >
>> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
>> > leaves")' http://localhost:8983/solr/knr/stream
>> >
>> > {
>> >
>> >   "result-set":{
>> >
>> > "docs":[{
>> >
>> > "node":"product1",
>> >
>> > "collection":"knr",
>> >
>> > "field":"node",
>> >
>> > "level":0}
>> >
>> >   ,{
>> >
>> > "EOF":true,
>> >
>> > "RESPONSE_TIME":4}]}}
>> >
>> >
>> >
>> >
>> > Below is the data I got from
>> >
>> >
>> https://github.com/apache/lucene-solr/blob/branch_7_6/solr/solrj/src/test/org/apache/solr/client/solrj/io/graph/GraphExpressionTest.java#L271
>> >
>> >
>> >
>> > According to this test 4 docs are expected.
>> >
>> >
>> > I am not sure what I am missing. Any pointers, please
>> >
>> >
>> > Thanks you,
>> >
>> > Rajeswari
>> >
>>
>


Re: Help with Stream Graph

2019-10-17 Thread Rajeswari Natarajan
Thanks Joel.

Here is the logs for below request

curl --data-urlencode
'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
http://localhost:8983/solr/knr/stream

2019-10-17 15:02:06.969 INFO  (qtp952486988-280) [c:knr s:shard1
r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
[knr_shard1_replica_n1]  webapp=/solr path=/stream
params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s")}
status=0 QTime=0

2019-10-17 15:02:06.975 INFO  (qtp952486988-192) [c:knr s:shard1
r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
[knr_shard1_replica_n1]  webapp=/solr path=/export
params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
hits=0 status=0 QTime=1



Here is the logs for



curl --data-urlencode
'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
leaves")' http://localhost:8983/solr/knr/stream


2019-10-17 15:03:57.068 INFO  (qtp952486988-356) [c:knr s:shard1
r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
[knr_shard1_replica_n1]  webapp=/solr path=/stream
params={expr=gatherNodes(knr,walk%3D"product1->product_s",gather%3D"basket_s",scatter%3D"branches,+leaves")}
status=0 QTime=0

2019-10-17 15:03:57.071 INFO  (qtp952486988-400) [c:knr s:shard1
r:core_node2 x:knr_shard1_replica_n1] o.a.s.c.S.Request
[knr_shard1_replica_n1]  webapp=/solr path=/export
params={q={!terms+f%3Dproduct_s}product1=false=off=basket_s,product_s=basket_s+asc,product_s+asc=json=2.2}
hits=0 status=0 QTime=0




Thank you,

Rajeswari

On Thu, Oct 17, 2019 at 5:23 AM Joel Bernstein  wrote:

> Can you show the logs from this request. There will be a Solr query that
> gets sent with product1 searched against the product_s field. Let's see how
> many documents that query returns.
>
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
>
> On Thu, Oct 17, 2019 at 1:41 AM Rajeswari Natarajan 
> wrote:
>
> > Hi,
> >
> > Since the stream graph query for my use case , didn't work as  i took the
> > data from solr source code test and also copied the schema and
> > solrconfig.xml from solr 7.6 source code.  Had to substitute few
> variables.
> >
> > Posted below data
> >
> > curl -X POST http://localhost:8983/solr/knr/update -H
> > 'Content-type:text/csv' -d '
> > id, basket_s, product_s, prics_f
> > 90,basket1,product1,20
> > 91,basket1,product3,30
> > 92,basket1,product5,1
> > 93,basket2,product1,2
> > 94,basket2,product6,5
> > 95,basket2,product7,10
> > 96,basket3,product4,20
> > 97,basket3,product3,10
> > 98,basket3,product1,10
> > 99,basket4,product4,40
> > 110,basket4,product3,10
> > 111,basket4,product1,10'
> > After this I committed and made sure the data got published. to solr
> >
> > curl --data-urlencode
> > 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
> > http://localhost:8983/solr/knr/stream
> >
> > {
> >
> >   "result-set":{
> >
> > "docs":[{
> >
> > "EOF":true,
> >
> > "RESPONSE_TIME":4}]}}
> >
> >
> > and if I add *scatter="branches, leaves" , there is one doc.*
> >
> >
> >
> > curl --data-urlencode
> >
> >
> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
> > leaves")' http://localhost:8983/solr/knr/stream
> >
> > {
> >
> >   "result-set":{
> >
> > "docs":[{
> >
> > "node":"product1",
> >
> > "collection":"knr",
> >
> > "field":"node",
> >
> > "level":0}
> >
> >   ,{
> >
> > "EOF":true,
> >
> > "RESPONSE_TIME":4}]}}
> >
> >
> >
> >
> > Below is the data I got from
> >
> >
> https://github.com/apache/lucene-solr/blob/branch_7_6/solr/solrj/src/test/org/apache/solr/client/solrj/io/graph/GraphExpressionTest.java#L271
> >
> >
> >
> > According to this test 4 docs are expected.
> >
> >
> > I am not sure what I am missing. Any pointers, please
> >
> >
> > Thanks you,
> >
> > Rajeswari
> >
>


Re: Query regarding positionIncrementGap

2019-10-17 Thread Erick Erickson
First, it only counts if you add multiple entries for the field. Consider the 
following

   a b c
   def


where the field has a positionIncrementGap of 100. The term positions of the 
entries are
a:1
b:2
c:3
d:103
e:104
f:105

Now consider the doc where there’s only one field:

   a b c d e f


The term positions are
a:1
b:2
c:3
d:4
e:5
f:6

The use-case is if you, say, index individual sentences and want to match two 
or more words in the _same_ sentence. You can specify a phrase query where the 
slop is < the positionIncrementGap. So in the first case, if I search for “a 
b”~99 I’d get a match. But if I searched for “a d”~99 I wouldn’t.

Best,
Erick

> On Oct 17, 2019, at 2:09 AM, Shubham Goswami  
> wrote:
> 
> Hi Community
> 
> I am a beginner in solr and i am trying to understand the working of
> positionIncrementGap but i am still not clear how it exactly works for the
> phrase queries and general queires.
>   Can somebody please help me to understand this with the help fo an
> example ?
> Any help will be appreciated. Thanks in advance.
> 
> -- 
> *Thanks & Regards*
> Shubham Goswami
> Enterprise Software Engineer
> *HotWax Systems*
> *Enterprise open source experts*
> cell: +91-7803886288
> office: 0731-409-3684
> http://www.hotwaxsystems.com



Re: Help with Stream Graph

2019-10-17 Thread Joel Bernstein
Can you show the logs from this request. There will be a Solr query that
gets sent with product1 searched against the product_s field. Let's see how
many documents that query returns.


Joel Bernstein
http://joelsolr.blogspot.com/


On Thu, Oct 17, 2019 at 1:41 AM Rajeswari Natarajan 
wrote:

> Hi,
>
> Since the stream graph query for my use case , didn't work as  i took the
> data from solr source code test and also copied the schema and
> solrconfig.xml from solr 7.6 source code.  Had to substitute few variables.
>
> Posted below data
>
> curl -X POST http://localhost:8983/solr/knr/update -H
> 'Content-type:text/csv' -d '
> id, basket_s, product_s, prics_f
> 90,basket1,product1,20
> 91,basket1,product3,30
> 92,basket1,product5,1
> 93,basket2,product1,2
> 94,basket2,product6,5
> 95,basket2,product7,10
> 96,basket3,product4,20
> 97,basket3,product3,10
> 98,basket3,product1,10
> 99,basket4,product4,40
> 110,basket4,product3,10
> 111,basket4,product1,10'
> After this I committed and made sure the data got published. to solr
>
> curl --data-urlencode
> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s")'
> http://localhost:8983/solr/knr/stream
>
> {
>
>   "result-set":{
>
> "docs":[{
>
> "EOF":true,
>
> "RESPONSE_TIME":4}]}}
>
>
> and if I add *scatter="branches, leaves" , there is one doc.*
>
>
>
> curl --data-urlencode
>
> 'expr=gatherNodes(knr,walk="product1->product_s",gather="basket_s",scatter="branches,
> leaves")' http://localhost:8983/solr/knr/stream
>
> {
>
>   "result-set":{
>
> "docs":[{
>
> "node":"product1",
>
> "collection":"knr",
>
> "field":"node",
>
> "level":0}
>
>   ,{
>
> "EOF":true,
>
> "RESPONSE_TIME":4}]}}
>
>
>
>
> Below is the data I got from
>
> https://github.com/apache/lucene-solr/blob/branch_7_6/solr/solrj/src/test/org/apache/solr/client/solrj/io/graph/GraphExpressionTest.java#L271
>
>
>
> According to this test 4 docs are expected.
>
>
> I am not sure what I am missing. Any pointers, please
>
>
> Thanks you,
>
> Rajeswari
>


Re: Error in Initializing plugin

2019-10-17 Thread Erick Erickson
There is very little information to go on here. What is the full stack trace 
from the Solr log file for this?

Best,
Erick

> On Oct 17, 2019, at 5:36 AM, Shubham Goswami  
> wrote:
> 
> Hi Community
> 
> I am working with currencies and exchange rates and getting this following
> error.
> error: Plugin Initializing failure for [schema.xml] fieldType
> Can somebody please help me out regarding this issue ?
> schema changes :  defaultCurrency="USD" currencyConfig="currency.xml"/>
> Thanks in adavance.
> 
> -- 
> *Thanks & Regards*
> Shubham Goswami
> Enterprise Software Engineer
> *HotWax Systems*
> *Enterprise open source experts*
> cell: +91-7803886288
> office: 0731-409-3684
> http://www.hotwaxsystems.com



Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Jon Kjær Amundsen
We have restores that take longer but do not have problems with timeouts.
But we use the async parameter
https://lucene.apache.org/solr/guide/7_2/collections-api.html#CollectionsAPI-restore

It could seem from the code you provided that it will timeout after 180
seconds if you do not make the restore async,

Venlig hilsen/Best regards

*Jon Kjær Amundsen*
Developer


Phone: +45 7023 9080
E-mail: j...@udbudsvagten.dk
Web: www.udbudsvagten.dk
Parken - Tårn D - 5. Sal
Øster Allé 48 | DK - 2100 København



Intelligent Offentlig Samhandel
*Før, under og efter udbud*

*Følg UdbudsVagten og markedet her Linkedin
 *


Den tor. 17. okt. 2019 kl. 13.36 skrev Koen De Groote <
koen.degro...@limecraft.com>:

> Is "it" a hard limit, sorry.
>
> On Thu, Oct 17, 2019 at 1:35 PM Koen De Groote <
> koen.degro...@limecraft.com>
> wrote:
>
> > The error was a zookeeper connect timeout. Which apparently is hardcoded
> > to 180 seconds.
> >
> > I've recently succeeded in the restore. It may well have been a
> connection
> > issue, since the environment is a shared VM environment. Outside pressure
> > is possible.
> >
> > The timeout source code(in version 7.6.0):
> >
> https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L301
> >
> > And eventually it gets used here:
> >
> https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L359
> >
> > This makes me wonder: is there a hard limit? The restore can only take
> 180
> > seconds or it fails? Or is that timeout per connection attempt to
> zookeeper?
> >
> >
> >
> > On Thu, Oct 17, 2019 at 11:16 AM Jon Kjær Amundsen 
> > wrote:
> >
> >> As a restore is server agnostic (i.e. you can restore to a totally
> >> different host than the backup was taken from) hat shouldn't be the
> >> problem.
> >> Also a collection as such has no leader, only shards have.
> >>
> >> Do you have any kind of logs stating the errors encountered?
> >>
> >> Venlig hilsen/Best regards
> >>
> >> *Jon Kjær Amundsen*
> >> Developer
> >>
> >>
> >> Phone: +45 7023 9080
> >> E-mail: j...@udbudsvagten.dk
> >> Web: www.udbudsvagten.dk
> >> Parken - Tårn D - 5. Sal
> >> Øster Allé 48 | DK - 2100 København
> >>
> >> 
> >>
> >> Intelligent Offentlig Samhandel
> >> *Før, under og efter udbud*
> >>
> >> *Følg UdbudsVagten og markedet her Linkedin
> >>  *
> >>
> >>
> >> Den tor. 17. okt. 2019 kl. 10.06 skrev Koen De Groote <
> >> koen.degro...@limecraft.com>:
> >>
> >> > Yes, both servers back up to a network drive.
> >> >
> >> > However, that is not the point of my question.
> >> >
> >> > The point of my question is: If I execute the curl command, that
> >> contacts
> >> > the collections API, to perform the backup, does it matter that the
> >> leader
> >> > is on a different host from the one where the backup command was
> >> executed?
> >> >
> >> >
> >> >
> >> > On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen  >
> >> > wrote:
> >> >
> >> > > Hi Koen
> >> > >
> >> > > A quick sanity check:
> >> > > Do you use a network drive accessible from both servers to make the
> >> > backup
> >> > > to?
> >> > > If you've backed up server2's collection to a local disk, then when
> >> > you're
> >> > > trying to restore it via server1 it does not know anything about the
> >> > > backup.
> >> > >
> >> > > Venlig hilsen/Best regards
> >> > >
> >> > > *Jon Kjær Amundsen*
> >> > > Developer
> >> > >
> >> > >
> >> > > Phone: +45 7023 9080
> >> > > E-mail: j...@udbudsvagten.dk
> >> > > Web: www.udbudsvagten.dk
> >> > > Parken - Tårn D - 5. Sal
> >> > > Øster Allé 48 | DK - 2100 København
> >> > >
> >> > > 
> >> > >
> >> > > Intelligent Offentlig Samhandel
> >> > > *Før, under og efter udbud*
> >> > >
> >> > > *Følg UdbudsVagten og markedet her Linkedin
> >> > >  *
> >> > >
> >> > >
> >> > > Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
> >> > > koen.degro...@limecraft.com>:
> >> > >
> >> > > > I'm trying to restore a couple of collections, and 1 keeps
> feeling.
> >> > This
> >> > > > happens to be the only one who's leader isn't on the host that the
> >> > backup
> >> > > > was taken from.
> >> > > >
> >> > > >
> >> > > > The backup was done on server1, for all collections.
> >> > > >
> >> > > > For this collection that is failing, the Leader was on server2.
> All
> >> > other
> >> > > > collections had their leader on server1. All collections had 1
> >> replica,
> >> > > on
> >> > > > the other server.
> >> > > >
> >> > > > I would think that having the replica there would be enough to
> >> perform
> >> > a
> >> > > > restore.

Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Koen De Groote
Is "it" a hard limit, sorry.

On Thu, Oct 17, 2019 at 1:35 PM Koen De Groote 
wrote:

> The error was a zookeeper connect timeout. Which apparently is hardcoded
> to 180 seconds.
>
> I've recently succeeded in the restore. It may well have been a connection
> issue, since the environment is a shared VM environment. Outside pressure
> is possible.
>
> The timeout source code(in version 7.6.0):
> https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L301
>
> And eventually it gets used here:
> https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L359
>
> This makes me wonder: is there a hard limit? The restore can only take 180
> seconds or it fails? Or is that timeout per connection attempt to zookeeper?
>
>
>
> On Thu, Oct 17, 2019 at 11:16 AM Jon Kjær Amundsen 
> wrote:
>
>> As a restore is server agnostic (i.e. you can restore to a totally
>> different host than the backup was taken from) hat shouldn't be the
>> problem.
>> Also a collection as such has no leader, only shards have.
>>
>> Do you have any kind of logs stating the errors encountered?
>>
>> Venlig hilsen/Best regards
>>
>> *Jon Kjær Amundsen*
>> Developer
>>
>>
>> Phone: +45 7023 9080
>> E-mail: j...@udbudsvagten.dk
>> Web: www.udbudsvagten.dk
>> Parken - Tårn D - 5. Sal
>> Øster Allé 48 | DK - 2100 København
>>
>> 
>>
>> Intelligent Offentlig Samhandel
>> *Før, under og efter udbud*
>>
>> *Følg UdbudsVagten og markedet her Linkedin
>>  *
>>
>>
>> Den tor. 17. okt. 2019 kl. 10.06 skrev Koen De Groote <
>> koen.degro...@limecraft.com>:
>>
>> > Yes, both servers back up to a network drive.
>> >
>> > However, that is not the point of my question.
>> >
>> > The point of my question is: If I execute the curl command, that
>> contacts
>> > the collections API, to perform the backup, does it matter that the
>> leader
>> > is on a different host from the one where the backup command was
>> executed?
>> >
>> >
>> >
>> > On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen 
>> > wrote:
>> >
>> > > Hi Koen
>> > >
>> > > A quick sanity check:
>> > > Do you use a network drive accessible from both servers to make the
>> > backup
>> > > to?
>> > > If you've backed up server2's collection to a local disk, then when
>> > you're
>> > > trying to restore it via server1 it does not know anything about the
>> > > backup.
>> > >
>> > > Venlig hilsen/Best regards
>> > >
>> > > *Jon Kjær Amundsen*
>> > > Developer
>> > >
>> > >
>> > > Phone: +45 7023 9080
>> > > E-mail: j...@udbudsvagten.dk
>> > > Web: www.udbudsvagten.dk
>> > > Parken - Tårn D - 5. Sal
>> > > Øster Allé 48 | DK - 2100 København
>> > >
>> > > 
>> > >
>> > > Intelligent Offentlig Samhandel
>> > > *Før, under og efter udbud*
>> > >
>> > > *Følg UdbudsVagten og markedet her Linkedin
>> > >  *
>> > >
>> > >
>> > > Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
>> > > koen.degro...@limecraft.com>:
>> > >
>> > > > I'm trying to restore a couple of collections, and 1 keeps feeling.
>> > This
>> > > > happens to be the only one who's leader isn't on the host that the
>> > backup
>> > > > was taken from.
>> > > >
>> > > >
>> > > > The backup was done on server1, for all collections.
>> > > >
>> > > > For this collection that is failing, the Leader was on server2. All
>> > other
>> > > > collections had their leader on server1. All collections had 1
>> replica,
>> > > on
>> > > > the other server.
>> > > >
>> > > > I would think that having the replica there would be enough to
>> perform
>> > a
>> > > > restore.
>> > > >
>> > > > Or does the backup need to happen on the actual leader?
>> > > >
>> > > > Kind regards,
>> > > > Koen De Groote
>> > > >
>> > >
>> >
>>
>


Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Koen De Groote
The error was a zookeeper connect timeout. Which apparently is hardcoded to
180 seconds.

I've recently succeeded in the restore. It may well have been a connection
issue, since the environment is a shared VM environment. Outside pressure
is possible.

The timeout source code(in version 7.6.0):
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L301

And eventually it gets used here:
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/handler/admin/CollectionsHandler.java#L359

This makes me wonder: is there a hard limit? The restore can only take 180
seconds or it fails? Or is that timeout per connection attempt to zookeeper?



On Thu, Oct 17, 2019 at 11:16 AM Jon Kjær Amundsen 
wrote:

> As a restore is server agnostic (i.e. you can restore to a totally
> different host than the backup was taken from) hat shouldn't be the
> problem.
> Also a collection as such has no leader, only shards have.
>
> Do you have any kind of logs stating the errors encountered?
>
> Venlig hilsen/Best regards
>
> *Jon Kjær Amundsen*
> Developer
>
>
> Phone: +45 7023 9080
> E-mail: j...@udbudsvagten.dk
> Web: www.udbudsvagten.dk
> Parken - Tårn D - 5. Sal
> Øster Allé 48 | DK - 2100 København
>
> 
>
> Intelligent Offentlig Samhandel
> *Før, under og efter udbud*
>
> *Følg UdbudsVagten og markedet her Linkedin
>  *
>
>
> Den tor. 17. okt. 2019 kl. 10.06 skrev Koen De Groote <
> koen.degro...@limecraft.com>:
>
> > Yes, both servers back up to a network drive.
> >
> > However, that is not the point of my question.
> >
> > The point of my question is: If I execute the curl command, that contacts
> > the collections API, to perform the backup, does it matter that the
> leader
> > is on a different host from the one where the backup command was
> executed?
> >
> >
> >
> > On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen 
> > wrote:
> >
> > > Hi Koen
> > >
> > > A quick sanity check:
> > > Do you use a network drive accessible from both servers to make the
> > backup
> > > to?
> > > If you've backed up server2's collection to a local disk, then when
> > you're
> > > trying to restore it via server1 it does not know anything about the
> > > backup.
> > >
> > > Venlig hilsen/Best regards
> > >
> > > *Jon Kjær Amundsen*
> > > Developer
> > >
> > >
> > > Phone: +45 7023 9080
> > > E-mail: j...@udbudsvagten.dk
> > > Web: www.udbudsvagten.dk
> > > Parken - Tårn D - 5. Sal
> > > Øster Allé 48 | DK - 2100 København
> > >
> > > 
> > >
> > > Intelligent Offentlig Samhandel
> > > *Før, under og efter udbud*
> > >
> > > *Følg UdbudsVagten og markedet her Linkedin
> > >  *
> > >
> > >
> > > Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
> > > koen.degro...@limecraft.com>:
> > >
> > > > I'm trying to restore a couple of collections, and 1 keeps feeling.
> > This
> > > > happens to be the only one who's leader isn't on the host that the
> > backup
> > > > was taken from.
> > > >
> > > >
> > > > The backup was done on server1, for all collections.
> > > >
> > > > For this collection that is failing, the Leader was on server2. All
> > other
> > > > collections had their leader on server1. All collections had 1
> replica,
> > > on
> > > > the other server.
> > > >
> > > > I would think that having the replica there would be enough to
> perform
> > a
> > > > restore.
> > > >
> > > > Or does the backup need to happen on the actual leader?
> > > >
> > > > Kind regards,
> > > > Koen De Groote
> > > >
> > >
> >
>


Re: Solaris Install Package

2019-10-17 Thread Jason Gerlowski
Hi Andrew,

I believe that yes, Solr should work on Solaris.  I've never done so
personally, but very occasionally I hear of someone doing so.
Additionally, Uwe runs a Jenkins server that runs tests on Solaris
(among other OSs), and the results for Solaris look to be pretty
standard for our test suite.  I'm not sure what Solaris version these
tests run on, that might be worth double checking here:
https://jenkins.thetaphi.de/view/Lucene-Solr/job/Lucene-Solr-8.x-Solaris/
. If you find any particular issues on Solaris (especially in the
scripts accompanying Solr, e.g. bin/solr, bin/post), it'd be
appropriate to open up JIRA tickets for those.

That saidwhile it seems to work and receives at least some
testing, it's definitely not common in terms of what the community
uses, and tests, and codes for on a daily basis.  As with any
open-source project, there's always a certain amount of risk in
diverging from the commonly used/tested environments and usage
patterns.  So, YMMV.

Best,

Jason

On Mon, Oct 7, 2019 at 5:34 PM Andrew Corbett  wrote:
>
> I have been trying to research the possibility of adding Solr to servers 
> running the Solaris 10 and 11 operating systems. Solaris isn't mentioned in 
> the documentation. Would adding Solr to these servers be possible? Would I 
> need to make a feature request?


Error in Initializing plugin

2019-10-17 Thread Shubham Goswami
Hi Community

I am working with currencies and exchange rates and getting this following
error.
error: Plugin Initializing failure for [schema.xml] fieldType
Can somebody please help me out regarding this issue ?
schema changes : 
Thanks in adavance.

-- 
*Thanks & Regards*
Shubham Goswami
Enterprise Software Engineer
*HotWax Systems*
*Enterprise open source experts*
cell: +91-7803886288
office: 0731-409-3684
http://www.hotwaxsystems.com


Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Jon Kjær Amundsen
As a restore is server agnostic (i.e. you can restore to a totally
different host than the backup was taken from) hat shouldn't be the problem.
Also a collection as such has no leader, only shards have.

Do you have any kind of logs stating the errors encountered?

Venlig hilsen/Best regards

*Jon Kjær Amundsen*
Developer


Phone: +45 7023 9080
E-mail: j...@udbudsvagten.dk
Web: www.udbudsvagten.dk
Parken - Tårn D - 5. Sal
Øster Allé 48 | DK - 2100 København



Intelligent Offentlig Samhandel
*Før, under og efter udbud*

*Følg UdbudsVagten og markedet her Linkedin
 *


Den tor. 17. okt. 2019 kl. 10.06 skrev Koen De Groote <
koen.degro...@limecraft.com>:

> Yes, both servers back up to a network drive.
>
> However, that is not the point of my question.
>
> The point of my question is: If I execute the curl command, that contacts
> the collections API, to perform the backup, does it matter that the leader
> is on a different host from the one where the backup command was executed?
>
>
>
> On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen 
> wrote:
>
> > Hi Koen
> >
> > A quick sanity check:
> > Do you use a network drive accessible from both servers to make the
> backup
> > to?
> > If you've backed up server2's collection to a local disk, then when
> you're
> > trying to restore it via server1 it does not know anything about the
> > backup.
> >
> > Venlig hilsen/Best regards
> >
> > *Jon Kjær Amundsen*
> > Developer
> >
> >
> > Phone: +45 7023 9080
> > E-mail: j...@udbudsvagten.dk
> > Web: www.udbudsvagten.dk
> > Parken - Tårn D - 5. Sal
> > Øster Allé 48 | DK - 2100 København
> >
> > 
> >
> > Intelligent Offentlig Samhandel
> > *Før, under og efter udbud*
> >
> > *Følg UdbudsVagten og markedet her Linkedin
> >  *
> >
> >
> > Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
> > koen.degro...@limecraft.com>:
> >
> > > I'm trying to restore a couple of collections, and 1 keeps feeling.
> This
> > > happens to be the only one who's leader isn't on the host that the
> backup
> > > was taken from.
> > >
> > >
> > > The backup was done on server1, for all collections.
> > >
> > > For this collection that is failing, the Leader was on server2. All
> other
> > > collections had their leader on server1. All collections had 1 replica,
> > on
> > > the other server.
> > >
> > > I would think that having the replica there would be enough to perform
> a
> > > restore.
> > >
> > > Or does the backup need to happen on the actual leader?
> > >
> > > Kind regards,
> > > Koen De Groote
> > >
> >
>


Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Koen De Groote
Both to the same network drive, sorry.

On Thu, Oct 17, 2019 at 10:05 AM Koen De Groote 
wrote:

> Yes, both servers back up to a network drive.
>
> However, that is not the point of my question.
>
> The point of my question is: If I execute the curl command, that contacts
> the collections API, to perform the backup, does it matter that the leader
> is on a different host from the one where the backup command was executed?
>
>
>
> On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen 
> wrote:
>
>> Hi Koen
>>
>> A quick sanity check:
>> Do you use a network drive accessible from both servers to make the backup
>> to?
>> If you've backed up server2's collection to a local disk, then when you're
>> trying to restore it via server1 it does not know anything about the
>> backup.
>>
>> Venlig hilsen/Best regards
>>
>> *Jon Kjær Amundsen*
>> Developer
>>
>>
>> Phone: +45 7023 9080
>> E-mail: j...@udbudsvagten.dk
>> Web: www.udbudsvagten.dk
>> Parken - Tårn D - 5. Sal
>> Øster Allé 48 | DK - 2100 København
>>
>> 
>>
>> Intelligent Offentlig Samhandel
>> *Før, under og efter udbud*
>>
>> *Følg UdbudsVagten og markedet her Linkedin
>>  *
>>
>>
>> Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
>> koen.degro...@limecraft.com>:
>>
>> > I'm trying to restore a couple of collections, and 1 keeps feeling. This
>> > happens to be the only one who's leader isn't on the host that the
>> backup
>> > was taken from.
>> >
>> >
>> > The backup was done on server1, for all collections.
>> >
>> > For this collection that is failing, the Leader was on server2. All
>> other
>> > collections had their leader on server1. All collections had 1 replica,
>> on
>> > the other server.
>> >
>> > I would think that having the replica there would be enough to perform a
>> > restore.
>> >
>> > Or does the backup need to happen on the actual leader?
>> >
>> > Kind regards,
>> > Koen De Groote
>> >
>>
>


Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Koen De Groote
Yes, both servers back up to a network drive.

However, that is not the point of my question.

The point of my question is: If I execute the curl command, that contacts
the collections API, to perform the backup, does it matter that the leader
is on a different host from the one where the backup command was executed?



On Thu, Oct 17, 2019 at 9:30 AM Jon Kjær Amundsen 
wrote:

> Hi Koen
>
> A quick sanity check:
> Do you use a network drive accessible from both servers to make the backup
> to?
> If you've backed up server2's collection to a local disk, then when you're
> trying to restore it via server1 it does not know anything about the
> backup.
>
> Venlig hilsen/Best regards
>
> *Jon Kjær Amundsen*
> Developer
>
>
> Phone: +45 7023 9080
> E-mail: j...@udbudsvagten.dk
> Web: www.udbudsvagten.dk
> Parken - Tårn D - 5. Sal
> Øster Allé 48 | DK - 2100 København
>
> 
>
> Intelligent Offentlig Samhandel
> *Før, under og efter udbud*
>
> *Følg UdbudsVagten og markedet her Linkedin
>  *
>
>
> Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
> koen.degro...@limecraft.com>:
>
> > I'm trying to restore a couple of collections, and 1 keeps feeling. This
> > happens to be the only one who's leader isn't on the host that the backup
> > was taken from.
> >
> >
> > The backup was done on server1, for all collections.
> >
> > For this collection that is failing, the Leader was on server2. All other
> > collections had their leader on server1. All collections had 1 replica,
> on
> > the other server.
> >
> > I would think that having the replica there would be enough to perform a
> > restore.
> >
> > Or does the backup need to happen on the actual leader?
> >
> > Kind regards,
> > Koen De Groote
> >
>


Re: Do backups of collections need to be taken on the Leader?

2019-10-17 Thread Jon Kjær Amundsen
Hi Koen

A quick sanity check:
Do you use a network drive accessible from both servers to make the backup
to?
If you've backed up server2's collection to a local disk, then when you're
trying to restore it via server1 it does not know anything about the backup.

Venlig hilsen/Best regards

*Jon Kjær Amundsen*
Developer


Phone: +45 7023 9080
E-mail: j...@udbudsvagten.dk
Web: www.udbudsvagten.dk
Parken - Tårn D - 5. Sal
Øster Allé 48 | DK - 2100 København



Intelligent Offentlig Samhandel
*Før, under og efter udbud*

*Følg UdbudsVagten og markedet her Linkedin
 *


Den ons. 16. okt. 2019 kl. 17.42 skrev Koen De Groote <
koen.degro...@limecraft.com>:

> I'm trying to restore a couple of collections, and 1 keeps feeling. This
> happens to be the only one who's leader isn't on the host that the backup
> was taken from.
>
>
> The backup was done on server1, for all collections.
>
> For this collection that is failing, the Leader was on server2. All other
> collections had their leader on server1. All collections had 1 replica, on
> the other server.
>
> I would think that having the replica there would be enough to perform a
> restore.
>
> Or does the backup need to happen on the actual leader?
>
> Kind regards,
> Koen De Groote
>


Query regarding positionIncrementGap

2019-10-17 Thread Shubham Goswami
Hi Community

I am a beginner in solr and i am trying to understand the working of
positionIncrementGap but i am still not clear how it exactly works for the
phrase queries and general queires.
   Can somebody please help me to understand this with the help fo an
example ?
Any help will be appreciated. Thanks in advance.

-- 
*Thanks & Regards*
Shubham Goswami
Enterprise Software Engineer
*HotWax Systems*
*Enterprise open source experts*
cell: +91-7803886288
office: 0731-409-3684
http://www.hotwaxsystems.com


Re: Query on autoGeneratePhraseQueries

2019-10-17 Thread Shubham Goswami
Hi Michael/Shawn

Thanks for the response.
Michael you are right, autoGeneratePhraseQueries works for the query like
Black-company
with the setting of Sow=true.
Thanks for your great support.

Best
Shubham

On Wed, Oct 16, 2019 at 9:22 PM Michael Gibney 
wrote:

> Going to back to the initial question, the wording is a little ambiguous
> and it occurs to me that it's possible there's a misunderstanding of what
> autoGeneratePhraseQueries does. It really only auto-generates phrase
> *subqueries*. To use the example from the initial request, a query like
> (black company) would always generate a non-phrase query (respecting mm,
> q.op, etc. -- but in any case not a top-level phrase query), regardless of
> the setting of autoGeneratePhraseQueries.
>
> autoGeneratePhraseQueries (when set to true) only kicks in (in different
> ways depending on analysis chain, and setting of "sow") for a query like
> (the black-company manufactures), which would be transformed to something
> more like (the "black company" manufactures). The idea is that there's some
> extra indication that the two words should be bundled together for purposes
> of querying.
>
> If you want to auto-generate a top-level phrase query, some other approach
> would be called for.
>
> Apologies if this is obvious and/or not helpful, Shubham!
>
> On Wed, Oct 16, 2019 at 10:10 AM Shawn Heisey  wrote:
>
> > On 10/16/2019 7:14 AM, Shubham Goswami wrote:
> > > I have implemented the sow=false property with eDismax Query parser but
> > > still it does not has any effect
> > > on the query as it is still parsing as separate terms instead of
> phrased
> > > one.
> >
> > We have seen reports that when sow=false, which is the default setting
> > since Solr 7.0, autoGeneratePhraseQueries does not work.  Try setting
> > sow=true and see whether you get the results you expect.
> >
> > I do not know whether this behavior is a bug or if it is expected.
> >
> > Thanks,
> > Shawn
> >
>


-- 
*Thanks & Regards*
Shubham Goswami
Enterprise Software Engineer
*HotWax Systems*
*Enterprise open source experts*
cell: +91-7803886288
office: 0731-409-3684
http://www.hotwaxsystems.com