Hello guys!
I need an advise. My task is to delete some documents in collection.
Del algorithm is following:
Group docs by field1 with sort by field2 and delete every 3 and following
occurrences in every group.
Unfortunately I didn't find easy way to do so.
Closest approach was to use
Hi ,
I am working on SOLR upgrade from my current version to SOLR 8.4.1 version
and I am unable to delete indexed data from solr collection .
I have tried this URL
It would be better if you give us an example.
On Fri, 14 Feb 2020 at 17:20, Vadim Ivanov
wrote:
> Hello guys!
> I need an advise. My task is to delete some documents in collection.
> Del algorithm is following:
> Group docs by field1 with sort by field2 and delete every 3 and following
>
I don’t see anything in your description that requires a large heap. This is a
terrible JVM configuration.
Do this:
* Use the GC configuration I recommended, with an 8 GB heap.
* Run one copy of Solr. That hosts both shard1 and shard1.
That increases the RAM available for OS and file buffers
Hi,
Yes. I meant facet.mincount only.
Thanks
Saurabh
On Fri, Feb 14, 2020, 8:51 PM Vadim Ivanov <
vadim.iva...@spb.ntk-intourist.ru> wrote:
> group.mincount ? Never heard of it. It exists?
> May be you have in mind facet.mincount and second approach mentioned
> earlier:
>
> > > > > Next
I'm trying to use highlighting on a multivalued text field (analysis not so
important) ..
{ text: [ "hello", "world" ], id: 1 }
but I want to match across the string boundaries:
q=text:"hello world"
This works by setting the attribute
positionIncrementGap to 0, but then the hightlighting
We've seen this type of deadlock pretty often. Our recourse is to restart
solr on only one of the nodes, this seems to force the leader election to
take place and it soon stars rebuilding.
Let me know if you try that and it works... Wouldn't mind another
validation point that this happens to
Example of gtouping with empty groups in results:
Filed1 = rr_group, field2 = rr_updatedate
Problem is that I have tens of million groups in result and only several
thousand with "numFound" >2
"params":{
"q":"*:* ",
"group.sort":"rr_updatedate desc ",
"group.limit":"-1",
Try using HTTP POST to send your … command in the body instead
of using stream.body. E.g. using the cURL utility:
curl -XPOST http://localhost:8983/solr/TakTech/update?commit=true -d
'*:*'
Or go to Admin UI, find your collection, select «Documents» sub meny, change
the dropdown from JSON to
Hi,
If you want to sort on your field and want to put a count restriction too
then you have to use mincount. That seems to be best approach for your
problem.
Thanks
Saurabh
On Fri, Feb 14, 2020, 6:24 PM Vadim Ivanov <
vadim.iva...@spb.ntk-intourist.ru> wrote:
> Example of gtouping with empty
Hi all,
How do you all evaluate the success of your query autocomplete (i.e. suggester)
component if you use it?
We cannot use MRR for various reasons (I can go into them if you're
interested), so we're thinking of using nDCG since we already use that for
relevance eval of our system as a
Probably solution is here
https://stackoverflow.com/questions/51416042/solr-error-stream-body-is-disab
led/51420987
> -Original Message-
> From: Nitish Kumar [mailto:nnitishku...@firstam.com]
> Sent: Friday, February 14, 2020 10:28 AM
> To: solr-user@lucene.apache.org
> Subject: Deleting
group.mincount ? Never heard of it. It exists?
May be you have in mind facet.mincount and second approach mentioned earlier:
> > > > Next approach was to use facet first with facet.mincount=3, then
> > > > find docs ids by every facet result and then delete docs by id.
> > > > That way seems to
Actually I should clarify: we stop solr on one of the nodes, wait for the
other node to become the leader, and then start solr back up on the one
that was stopped.
On Fri, 14 Feb 2020 at 09:41, lstusr 5u93n4 wrote:
> We've seen this type of deadlock pretty often. Our recourse is to restart
>
Yes, you can manually manipulate the data in Zookeeper, but as you
say that’s a “heroic” option. But even if it’s totally messed up, you’re no
worse off. You can use bin/solr zk… to copy individual znodes up and
down, or there are various tools to let you do the same if you have
them.
It’s also
Overnight the replicas with a state of "down" changed to "recovery_failed".
Nothing I did. So I brought down both nodes, then started one and waited 5
min. A leader was born then I started the other node. So luckily no heroics
were needed.
I'll remember your advice about creating a parallel
You are asking on solr-user but your scenario seems pure Lucene.
For Lucene and indexing point-data, I strongly recommend LatLonPoint. For
Solr, same scenario, the Solr adaptation of the same functionality is
LatLonPointSpatialField. I know this doesn't directly address your
question. Just
Yeah, that is pretty outdated. At Netflix, I was running an 8 GB heap with Solr
1.3. :-)
Every GC I know about has a stop-the-world collector as a last ditch measure.
G1GC limits the time that the world will stop. It gives up after
MaxGCPauseMillis
milliseconds and leaves the rest of the
Try also looking at the HunspellFilter:
https://lucene.apache.org/solr/guide/6_6/filter-descriptions.html
dictionaries ( .dic and .aff ) can be found here:
https://cgit.freedesktop.org/libreoffice/dictionaries
or via the git repo:
Hi,
I'm trying to replace pprter stemmer with an english lemmatizer in my
analysis chain. Just wondering what
is the recommended way of achieving this. I've come across few different
implementation which are listed below;
Open NLP -->
Hello,
In the section on JVM tuning in the Solr 8.3 documentation (
https://lucene.apache.org/solr/guide/8_3/jvm-settings.html#jvm-settings)
there is a paragraph which cautions about setting heap sizes over 2 GB:
"The larger the heap the longer it takes to do garbage collection. This can
mean
I just read that page over and it looks way out of date. I’ll raise
a JIRA.
> On Feb 14, 2020, at 2:55 PM, Walter Underwood wrote:
>
> Yeah, that is pretty outdated. At Netflix, I was running an 8 GB heap with
> Solr 1.3. :-)
>
> Every GC I know about has a stop-the-world collector as a last
Glad it worked out, I like to avoid heroics whenever possible ;)…
It can take quite some time for solr to finally and for good give up,
waiting 10-15 minutes for something to change seems like an eternity.
What’s happening here is the node attempts to recover but fails for some
reason. So it
Hi All,
In our project we have to use multiple graph queries with AND and OR
conditions but graph query parser does not work for the below scenario, can
any one suggest how to overcome this kind of problem? this is stopping our
pre prod release .
we are also using traversalFilter but our usecase
Yes I did Erick, and that didn't do it. What about manual manipulation of the
zookeeper data? Rather than telling the customer they need to rebuild from
scratch, I'd prefer to attempt some last minute heroics.
--
Sent from: https://lucene.472066.n3.nabble.com/Solr-User-f472068.html
Hi team,
I am using Lucene 6.6.2, Spatial4j 0.7, lucene-spatial-extras 6.6.2. I am
trying to create a Spatial Query string for a given longitude, latitude &
radius in miles.
The query string generated using SpatialHelper (code as attached ) for
long: -122.8515139 & lat: 45.5099231 in .25 miles
26 matches
Mail list logo