Hi Kumar
> Emir , i need all tokens of query in incrementToken() function not only
> current token
That was just an example - the point was that you need to set attributes - you
can read all tokens from previous stream, do whatever needed with them and when
ready, set attributes and return
Hi guys,
Solr Version :: 6.6.1
I am able to import the pdf files into the Solr system using the DIH and
performs the indexing as expected. But i wish to clear the folder
C:/solr-6.6.1/server/solr/core_K2_Depot*/Depot* after the successful finish
of the indexing process.
Please suggest, if there
Hi Sundeep,
The simplified explanation is that terms are indexed to be more prefix search
friendly (and that is why Amrit suggested that you index term reversed if you
want leading wildcard). If you use leading wildcard, there is no structure to
limit terms that can be matched and engine has to
Yes, that works. Thanks.
-
--Ilay
--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
Hi Raja,
we are using solrcloud as a statefulset and every pod has its own storage
attached to it.
Thanks
Björn
> On 20. Nov 2017, at 05:59, rajasaur wrote:
>
> Hi Bjorn,
>
> Im trying a similar approach now (to get solrcloud working on kubernetes). I
> have run
@rick
I see many indexing config, but i don't see any config related to query
(i.e., number of threads etc.) in solrconfig. What will be the relevant
part for this area? In jetty threadpool is set to 1.
@Toke:
I have a webserver which uses solr for querying, this i guess is pretty
typical.
Pad
Read the CVE. Do you have an affected version of Solr? Do you have the
replication feature enabled in solrconfig.xml? Note that it might be enabled by
default. Test directory traversal on your system: can you read files remotely?
No? Then you are finished.
A better plan: upgrade to a newer
Hi Team,
I am facing issue for string containing hyphen when searched in spell field.
My solr core is solr-6.6.0
Points to reproduce:-
Eg:- 1. My search string is "spider-man".
2. When I do a search in solr with query spell:*spider-*. It shows numDocs=0
even though content is present.
3 . But
Please help me here
-- Forwarded message --
From: padmanabhan gonesani
Date: Mon, Nov 13, 2017 at 5:12 PM
Subject: CVE-2017-3163 - SOLR-5.2.1 version
To: gene...@lucene.apache.org
Hi Team,
*Description:* Apache Solr could allow a remote attacker to
Hi,
We have several indexed string fields which is not tokenized and does not
have docValues enabled.
When we do trailing wildcard searches on these fields they are running very
slow. We were thinking that since this field is indexed, such queries
should be running pretty quickly. We are using
You already asked that question and got several answers, did you not
see them? If you did see them, what is unclear?
Best,
Erick
On Mon, Nov 20, 2017 at 9:33 AM, Sundeep T wrote:
> Hi,
>
> We have several indexed string fields which is not tokenized and does not
> have
Hi,
Does anyone knows how long usually the merging in Solr will take?
I am currently merging about 3.5TB of data, and it has been running for
more than 28 hours and it is not completed yet. The merging is running on
SSD disk.
I am using Solr 6.5.1.
Regards,
Edwin
I am developing an application that uses cursorMark deep paging. It's a
java client using solrj client.
Currently the client is created with Solr 6.2 solrj jars, but the test
server is a solr 7.1 server
I am getting this error:
Error from server at http://XX:8983/solr/sial-catalog-product:
Nawab
Why it would be good to share the solrconfigs: I had a suspicion that you might
be using the same solrconfig for version 7 and 4.5. That is unlikely to work
well. But I could be way off base.
Rick
--
Sorry for being brief. Alternate email is rickleir at yahoo dot com
Hi All,
According to
http://lucene.apache.org/core/7_1_0/core/org/apache/lucene/util/automaton/RegExp.html.
Lucene supports repeat expressions.
repeatexp ::= repeatexp ? (zero or one occurrence)
| repeatexp * (zero or more occurrences)
| repeatexp +
Nawab Zada Asad Iqbal wrote:
> I have a webserver which uses solr for querying, this i guess is pretty
> typical. At times, there are 50 users sending queries at a given second.
> Sometimes, the queries take a few second to finish (i.e., if the max across
> all shards is 5
Hi Erick.
I initially asked this question regarding leading wildcards. This was a
typo, and what I meant was trailing wild card queries were slow. So queries
like text:'hello*" are slow. We were expecting since the string field is
already indexed, the searches should be fast, but that seems to be
At first glance you have a mis-configured setup. The most glaring
issue is that you're trying to search a 150G index in 1G of memory.
bq: String field (not tokenized) is docValues=true, indexed=true and stored=true
OK, this is kind of unusual to query but if the field just contains
single tokens
Similarity is query time.
wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/ (my blog)
> On Nov 20, 2017, at 4:57 PM, Nawab Zada Asad Iqbal wrote:
>
> Hi,
>
> I want to switch to Classic similarity instead of BM25 (default in solr7).
> Do I need
Hi,
I want to switch to Classic similarity instead of BM25 (default in solr7).
Do I need to reindex all cores after this? Or is it only a query time
setting?
Thanks
Nawab
Chirag
Some scattered clues:
StandardTokenizer splits on punctuation, so your spell field might not contain
spider-man.
When you do a wildcard search, the analysis chain can be different from what
you expected.
Cheers -- Rick
On November 20, 2017 9:58:54 AM EST, Chirag Garg
When I ran load benchmarks with 6.3.0, an overloaded cluster would get super
slow but keep functioning. With 6.5.1, we hit 100% CPU, then start getting
OOMs. That is really bad, because it means we need to reboot every node in the
cluster.
Also, the JVM OOM hook isn’t running the process
Hi,
Any updates regarding the error?
Regards,
Edwin
On 16 November 2017 at 10:21, Zheng Lin Edwin Yeo
wrote:
> Hi Karthik,
>
> Thanks for the update.
>
> I see from the JIRA that it is still unresolved, meaning we can't index
> EML files to Solr 7.1.0 for the time
Hi Erick,
Thanks for the reply. Here are more details on our setup -
*Setup/schema details -*
100 million doc solr core
String field (not tokenized) is docValues=true, indexed=true and stored=true
Field is almost unique in the index, around 80 million are unique
no commits on index
all
As I suspected this was a bug in my code. We use KIE Drools to configure
our queries, and there was a conflict between two rules.
On Mon, Nov 20, 2017 at 4:09 PM, Webster Homer
wrote:
> I am developing an application that uses cursorMark deep paging. It's a
> java client
Hi Rick,
Actually my spell field also contains text with hyphen i.e. it contains
"spider-man" even then also i am not able to search it.
Regards,
Chirag
--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
Hi Walter,
you can check if the JVM OOM hook is acknowledged by JVM
and setup in the JVM. The options are "-XX:+PrintFlagsFinal -version"
You can modify your bin/solr script and tweak the function "launch_solr"
at the end of the script. Replace "-jar start.jar" with "-XX:+PrintFlagsFinal
Hello, Chuming.
It doesn't. The closest thing is to create TermAutomatonQuery.
On Mon, Nov 20, 2017 at 11:03 PM, Chuming Chen
wrote:
> Hi All,
>
> According to http://lucene.apache.org/core/7_1_0/core/org/apache/lucene/
> util/automaton/RegExp.html. Lucene supports repeat
Well, define "slow". Conceptually a large OR clause is created that
contains all the terms that start with the indicated text. (actually a
PrefixQuery should be formed).
That said, I'd expect hello* to be reasonably fast as not many terms
_probably_ start with 'hello'. Not the same at all for,
Hi Shawn,
I did as you suggested and created the core by hand - I copied the files
from the existing core, including the index files (data directory) and
changed the core.properties file to the new core name (core_new) and
restarted. Now I'm having a different issue - it says it is Optimized but
Dear Shawn and Chris,
Thanks very much for your replies and helps.
And so sorry for my mistakes of first-time use of Mailing Lists.
On 11/9/2017 5:13 PM, Shawn wrote:
> Where did this information originate?
My SHA data come from the paper On the Naturalness of Buggy Code(Baishakhi Ray,
et al.
Hey Ankit,
Try this tool for a better view of your debug output, and then if you have
any specific question, do let me know :
http://splainer.io/
On Sun, Oct 29, 2017 at 2:34 AM, Ankit Shah wrote:
> Hi,
> I am new to the solr community, and have this weird problem with
32 matches
Mail list logo