Hi,
Maybe you could have a look at this :
http://wiki.apache.org/solr/DisMaxQParserPlugin#tie_.28Tie_breaker.29
Marc.
On Wed, Dec 7, 2011 at 5:48 PM, John fatmanc...@gmail.com wrote:
I have a complex edismax query:
I tried that, didn't seem to affect anything.
I think it only works between fields and not within a field.
On Thu, Dec 8, 2011 at 10:00 AM, Marc SCHNEIDER
marc.schneide...@gmail.comwrote:
Hi,
Maybe you could have a look at this :
When I start solr indexing RAM taken by MS SQL Server 2008 R2 also keeps on
increasing initially from some 1GB it went on to 3.6 GB (when indexing was
completed for just 1 Million records/5GB). I have set the responseBuffering
parameter to adaptive in data-config.xml however it didn't help me
Hi,
I have a solr core named collection1 running in a tomcat instance
modified with a DIH.
It's working fine and everything seems to be in order, but there was
discovered some minor flaws in
the DIH setup, so I have corrected it in the data-config.xml file that
is used by the DIH
Hello,
Usually, when such an error occur, there are some good hints of what's
wrong with your new configuration in solr logs.
Depending on how you setup your solr instance and configured logging for
solr (http://wiki.apache.org/solr/SolrLogging), log files may be located
at different places.
Hi Otis,
I can't find the download for the free SPM.
What Hardware and OS do I need for installing SPM to monitor my servers?
Regards
Bernd
Am 07.12.2011 18:47, schrieb Otis Gospodnetic:
Hi Dmitry,
You should use SPM for Solr - it exposes all Solr metrics and more (JVM, system
info, etc.)
Hello Yavar,
thanks for your reply.
I have two entries in the log files.
one entry is in the localhost_access_log.txt:
my ip adress - - [08/Dec/2011:08:59:50 +] GET
/solr/select/?q=sicherheitsschuhversion=2.2start=0rows=10indent=on
HTTP/1.1 200 388
and the other entry is in catalina.txt:
Hi All,
My autocommit settings are
max docs - 1000
max time - 86 secs
We have put newrelic agent so as to monitor our solr performance.
In that we see a continous curve for autocommit. It is as good as
autocommit is continuously being fired.
Is it that if Autocommit for certain documents takes
Hello,
I'm using Solr 3.4, and I'm having a problem with a request returning
different results if I have or not a space after a coma.
The request name, number rue taine paris returns results with 4 words out
of 5 matching (name, number, rue, paris)
The request name,number rue taine paris (no
same problem with Solr 4.0
2011/12/8 elisabeth benoit elisaelisael...@gmail.com
Hello,
I'm using Solr 3.4, and I'm having a problem with a request returning
different results if I have or not a space after a coma.
The request name, number rue taine paris returns results with 4 words
out
I have a text field, using stopwords...
Index and query analysers setup as follows:
SynonymFilterFactory
StopFilterFactory
WordDelimiterFilterFactory
LowerCaseFilterFactory
SnowballPorterFilterFactory
Searching for front of house brings back perfect matches, but
doesn't highlight the of.
I
Go ahead with SOLR based text search. Thats what it is meant for and does it
great.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-or-SQL-fultext-search-tp3566654p3569894.html
Sent from the Solr - User mailing list archive at Nabble.com.
Can u share more info: like what is your H/W infra, CPU, RAM, HDD??
From where you pick the records/documents to index; RDBMS, Files, Network??
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-using-very-high-I-O-tp3567076p3569903.html
Sent from the Solr
I have a query and faceting on one field. i want to improve the speed of
getting this facet please advise me the solutions what are all the options
available to get this facet faster.
--
View this message in context:
What is the type of the field on which you are getting facets (string, Text,
int, date etc.). Is it multivalued or not?
How many unique values do you have for the field?
What is your filtercache setting in your solrconfig.xml?
Regards
Pravesh
--
View this message in context:
On 08/12/2011 05:00, Mark Miller wrote:
Replication just copies the index, so I'm not sure how this would help offhand?
With SolrCloud this is a breeze - just fire up another replica for a shard and
the current index will replicate to it.
If you where willing to export the data to some
After migrating to Solr 3.5, i restart tomcat and i get the error below.
Any ideas what i am doing wrong?
SEVERE: org.apache.solr.common.SolrException: Invalid luceneMatchVersion
'LUCENE_35', valid values are: [LUCENE_20, LUCENE_21, LUCENE_22, LU
CENE_23, LUCENE_24, LUCENE_29, LUCENE_30,
filed type is either long or string.
Solr Statistics are:
*name: * filterCache *class: * org.apache.solr.search.FastLRUCache *
version: * 1.0 *description: * Concurrent LRU Cache(maxSize=512,
initialSize=512, minSize=460, acceptableSize=486, cleanupThread=false) *
stats: * lookups : 6679
I just migrated to Solr 3.5 and whenever i start it up i get the error
below. Any ideas what might be wrong? Previously i didn't have to do
anything special to get it to work. HAs anything changed in solr 3.5?
08-Dec-2011 10:45:03 org.apache.solr.common.SolrException log
SEVERE:
Hi, all
I'm using surround query parser.
The request A B returns ParseException.
But A OR B returns correct results.
I think this is the problem of default query operator.
Anyone know how to set?
Thanks,
Jason
--
View this message in context:
change the default operator in schema
Thanks and Regards,
S SYED ABDUL KATHER
On Thu, Dec 8, 2011 at 4:44 PM, Jason, Kim [via Lucene]
ml-node+s472066n3570034...@n3.nabble.com wrote:
Hi, all
I'm using surround query parser.
The request A B returns ParseException.
But A
From the changelog:
187
* SOLR-2588: Moved VelocityResponseWriter back to contrib module in order to
188
remove it as a mandatory core dependency. (Erik Hatcher)
http://svn.apache.org/viewvc/lucene/dev/branches/branch_3x/solr/CHANGES.txt?view=markup
I just migrated to Solr 3.5 and
How many unique terms do you have in the faceting field
Since there are lot of evictions, consider increasing the size of the
filtercache. Try to keep evictions to min.
BTW how much is your index size (GB/MB??) How much RAM is allocated?
Above All: Have you benchmarked your search? Is
I'm using surround query parser.
The request A B returns ParseException.
But A OR B returns correct results.
I think this is the problem of default query operator.
Anyone know how to set?
There is no room for default operator in surround query parser. This is a
limitation of surround.
Oh. That's bad to me.
Thanks anyway.
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-set-default-query-operator-in-surround-query-parser-tp3570034p3570088.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thanks Marcus. To resolve this problem i just added a shared lib folder for
my cores and added the velocity jars in this folder and that resolved the
error. I hope it was the right thing to do though.
Thanks.
On Thu, Dec 8, 2011 at 11:20 AM, Markus Jelsma
markus.jel...@openindex.iowrote:
From
Hi,
I tried to implement language detection in SOLR 3.5.
I added the processor class for langDetect in solrconfig.xml :
updateRequestProcessorChain name=langid
processor
class=org.apache.solr.update.processor.LangDetectLanguageIdentifierUpdateProcessorFactory
str
Did you also install apache-solr-langid-4.0.jar ?
And the jars in contrib/langid/lib ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Language-Detection-in-SOLR-3-5-tp3570177p3570202.html
Sent from the Solr - User mailing list archive at Nabble.com.
I have apache-solr-langid-3.5.jar and also the jars in contrib/langid/lib
Ankita
On 8 December 2011 18:20, O. Klein kl...@octoweb.nl wrote:
Did you also install apache-solr-langid-4.0.jar ?
And the jars in contrib/langid/lib ?
--
View this message in context:
What about q=my_field1:my_search_string~2 my_search_string
?
True, the clause edismax created would search for your
my_search_string in my_field1 twice, but it might suffice..
Best
Erick
On Wed, Dec 7, 2011 at 3:15 AM, Marc SCHNEIDER
marc.schneide...@gmail.com wrote:
Hello,
I'm using edismax
I have increased the filtercache values as to as filterCache
class=solr.FastLRUCache size=16384 initialSize=4096
autowarmCount=4096/ and documentCache class=solr.LRUCache
size=16384 initialSize=16384/ and queryResultCache
class=solr.LRUCache size=16384 initialSize=4096
Hmmm, this is unusual. Can we see the code you use to delete?
And your solrconfig file? You're not doing something odd like
optimizing on commit or anything, right?
You shouldn't have to commit after deletes. The fact that you're
hanging is very odd (BTW, does hanging mean you're system
is locked
Have you, for instance, used the admin/browse schema or Luke
to examine your indes and see if you're indexing your text data
at all?
How are you indexing with SolrJ? Using the extracting request
handler or parsing the docs with the Tika libs yourself?
Details matter.
Best
Erick
On Wed, Dec 7,
Hi,
I was just testing field collapsing in my solr admin on solr 3.5.0. I have
observed that the results of field collapsing are not being cached unlike
other solr query results. Am doing the same query multiple times and the
time taken still remains approximately the same. Is there something i
Nope, they're the same. The original name was Field Collapsing,
but it was changed to Grouping later.
But note that the functionality has changed over time, so you might
be seeing documents from different incarnations of the code.
Best
Erick
On Wed, Dec 7, 2011 at 10:41 AM, Kissue Kissue
This is all controlled by Solr via the uniqueKey field in your schema. Just
remove that entry.
But then it's all up to you to handle the fact that there will be multiple
documents with the same ID all returned as a result of querying. And
it won't matter what program adds data, *nothing* will be
Mark,
Agreed that Replication wouldn't help, I was dreaming that there was
some intermediate format used in replication.
Ideally you are right, I could just reindex the data and go on with
life, but my case is not so simple. Currently we have some set of
processes which is run against the raw
Good Afternoon,
Im looking at Deltas via a DeltaImportHandler. I was running Solr 1.4.1
but just upgraded to 3.5. Previously I was able to run debug and verbose
from:
http://localhost:8080/solr/admin/dataimport.jsp?handler=/advert
But since upgrading when choosing these options the
Thanks for the response Erick. I actually turned up logging yesterday and
noticed spellchecker builds were causing the delays. Setting buildOnCommit to
false solved the problem. Our plan is to schedule a nightly timer task that
sends a 'spellcheck.build=true' to trigger it.
Mike
Date:
I read this response, but it lacks the quoted text so I have no clue what
your advice is in reference to. This makes it hard for others to benefit
from the advice. Just a thought.
Go ahead with SOLR based text search. Thats what it is meant for and does
it
great.
Regards
Pravesh
--
View
This would seem to indicate that you are using a whitespace analyzer on
the default search field. I believe other analyzers will properly tokenize
around the comma.
same problem with Solr 4.0
2011/12/8 elisabeth benoit elisaelisael...@gmail.com
Hello,
I'm using Solr 3.4, and I'm having a
Lets please keep the conversation on list.
If you are using auto commit and soft auto commit, it makes no sense to be
committing with solrj. The commits happen automatically. You couldn't measure
them with solrj.
- mark
On Dec 8, 2011, at 12:44 AM, yu shen wrote:
Thanks for the help.
For
On Dec 8, 2011, at 8:50 AM, Jamie Johnson wrote:
Isn't the codec stuff merged with trunk now?
Robert merged this recently AFAIK.
- Mark Miller
lucidimagination.com
Now when i start my solr it gives an exception :
org.apache.solr.common.SolrException: Error loading class
'org.apache.solr.update.processor.LangDetectLanguageIdentifierUpdateProcessorFactory'
Caused by: java.lang.ClassNotFoundException:
Hi,
I am trying to provide a means to search our corpus of nearly 2
million fulltext astronomy and physics articles using regular
expressions. A small percentage of our users need to be able to
locate, for example, certain types of identifiers that are present
within the fulltext (grant numbers,
On Thu, Dec 8, 2011 at 11:01 AM, Jay Luker lb...@reallywow.com wrote:
Hi,
I am trying to provide a means to search our corpus of nearly 2
million fulltext astronomy and physics articles using regular
expressions. A small percentage of our users need to be able to
locate, for example, certain
On Thu, Dec 8, 2011 at 10:46 AM, Mark Miller markrmil...@gmail.com wrote:
On Dec 8, 2011, at 8:50 AM, Jamie Johnson wrote:
Isn't the codec stuff merged with trunk now?
Robert merged this recently AFAIK.
true but that issue only moved the majority of the rest of the index
(stored fields,
Thanks Robert. I'll continue to watch the Jira and try not to bother
folks about this. Again greatly appreciate the insight.
On Thu, Dec 8, 2011 at 11:31 AM, Robert Muir rcm...@gmail.com wrote:
On Thu, Dec 8, 2011 at 10:46 AM, Mark Miller markrmil...@gmail.com wrote:
On Dec 8, 2011, at 8:50
Hello!!!
I have a trouble with searching russian words in Solr. And also i'm used
alfresco with solr. I'm add a content in my custom field on the russian
language. But when i search this words in Solr - i don't found anyone. I'm
search topics with this problems and did the following things, but
Have you looked at this page?
http://wiki.apache.org/solr/SolrTomcat
In particular the installing Solr instances under Tomcat.
You haven't told us much at all about how you start up Tomcat,
what configurations you have set, not even what error you get
when you try to hit the admin page. Please
On 08/12/2011 14:50, Jamie Johnson wrote:
Mark,
Agreed that Replication wouldn't help, I was dreaming that there was
some intermediate format used in replication.
Ideally you are right, I could just reindex the data and go on with
life, but my case is not so simple. Currently we have some set
Thanks Andrzej. I'll continue to follow the portable format JIRA
along with 3622, are there any others that you're aware of that are
blockers that would be useful to watch?
On Thu, Dec 8, 2011 at 10:49 AM, Andrzej Bialecki a...@getopt.org wrote:
On 08/12/2011 14:50, Jamie Johnson wrote:
Mark,
On Thu, Dec 8, 2011 at 12:55 PM, Jamie Johnson jej2...@gmail.com wrote:
Thanks Andrzej. I'll continue to follow the portable format JIRA
along with 3622, are there any others that you're aware of that are
blockers that would be useful to watch?
There is a lot to be done, particularly norms
Ah. Thanks Erick.
I see now that my question is different from sabman's.
Is there a way to use the DataImportHandler's full-import command so that
it does not delete the existing material before it begins?
Thanks,
Tricia
On Thu, Dec 8, 2011 at 6:35 AM, Erick Erickson
Hi Bernd,
Check this:
SPM for Solr is the enterprise-class, cloud-based, System/OS and Solr
Performance Monitoring SaaS.
So it's a SaaS - you simply sign up for it. During the signup you'll get to
download a small agent that works on RedHat, CentOS, Debian, Ubuntu, and maybe
other OSes.
: If I check in the solr.admin.analyzer, I get the same analysis for the two
: different requests. But it seems, if fact, that the lacking space after
: coma prevents name and number from matching.
query analysis is only part of hte picture ... Did you look at the
debuqQuery output? ... i
: But there is a workaround:
: 1) Do a normal query without facets (you only need to request doc ids
: at this point)
: 2) Collect all the IDs of the documents returned
: 3) Do a second query for all fields and facets, adding a filter to
: restrict result to those IDs collected in step 2.
an
Thanks Robert. I'll watch them all. Any others that are good to keep track of?
On Thu, Dec 8, 2011 at 1:25 PM, Robert Muir rcm...@gmail.com wrote:
On Thu, Dec 8, 2011 at 12:55 PM, Jamie Johnson jej2...@gmail.com wrote:
Thanks Andrzej. I'll continue to follow the portable format JIRA
along
: 3) When the user clicks into a single video in the search result,
: retrieve from the corresponding doc in Solr the timestamps of all
: words matching the keyword(s) (including stemming).
...
: Ok, so now for the harder part. For #3 it would seem I need something
: roughly like the
Hi,
Is there a specific reason why it is hard-coded to use the lucene
QParser? I was looking at JoinQParserPlugin.java and here it is in
createParser:
QParser fromQueryParser = subQuery(v, lucene);
I could pass another param named fromQueryParser and use it instead of
lucene. But again, is
If you look at the admin/stats page, does it show anything
for numDocs and maxDocs?
Have you tried looking at the index with Luke to see what's
in there?
Have you tried just looking at your data/index directory and
seeing if there are files there?
Are you sure you commit the changes?
Did you
OK, migrating to Solr 3.5 from what?
But what it *looks* like is that you are running a 3.5
config (see luceneMatchVersion in solrconfig.xml)
against a 3.3 Solr installation. In other words
it might be that you're running the old code against
a new schema.
But I'll freely admit that the
These are really horrible numbers:
hitratio : 0.18
inserts : 5488
evictions : 5011
Your cache isn't doing you much good, part of the problem
is you're faceting on so many values and it's chewing through
your cache. If your full data set has significantly more terms,
you're going to have trouble
: *pk*: The primary key for the entity. It is*optional*and only needed when
: using delta-imports. It has no relation to the uniqueKey defined in schema.xml
: but they both can be the same.
:
: When using in a nested entity is the PK the primary key column of the join
: table or the key used for
I believe all you need to do is add a ?clean=false to your query string.
If you have a unique key setup as your ID in solr then it should update
the existing documents instead of delete and re-indexing.
Cody
-Original Message-
From: P Williams [mailto:williams.tricia.l...@gmail.com]
Hello,
It should do the job, thanks!
Marc.
On Thu, Dec 8, 2011 at 2:06 PM, Erick Erickson erickerick...@gmail.comwrote:
What about q=my_field1:my_search_string~2 my_search_string
?
True, the clause edismax created would search for your
my_search_string in my_field1 twice, but it might
Greetings,
I see that we can query multiple facets for a search with a syntax like
fq=grade:A OR grade:B. However, I only know how to do this by modifying the
URL parameter. Is there a UI component that allows you to select multiple
facet values? I'm thinking something like a checkbox next
Hello,
I've a long field defined in my schema:
fieldType name=long class=solr.TrieLongField precisionStep=0
omitNorms=true positionIncrementGap=0 /
field name=ts type=long indexed=true stored=true required=true /
Before r1201855 I could use stats.facet=ts which allowed me to have a
: I've a long field defined in my schema:
:
: fieldType name=long class=solr.TrieLongField precisionStep=0
: omitNorms=true positionIncrementGap=0 /
:
: field name=ts type=long indexed=true stored=true required=true /
:
: Before r1201855 I could use stats.facet=ts which allowed me to have a
:
We're seeing the same thing (though we're not using replication). Based
on the trace, it looks like it would happen when Solr's response is too
slow for the client, and it's trying to send a response back to someone
who's no longer listening for one.
So, I think this isn't an error in itself
Hi Mark,
I did not fully catch you.
I have a dataImportHandler which import data from a database. Are you
suggesting autoCommit/autoSoftCommit can automaticall pull data from
database using dataImportHandler?
Otherwise, how can I commit user modification of search result from my web
page back to
On Thu, Dec 8, 2011 at 6:21 PM, Tom Lianza t...@wishpot.com wrote:
We're seeing the same thing (though we're not using replication). Based on
the trace, it looks like it would happen when Solr's response is too slow
for the client, and it's trying to send a response back to someone who's no
Hi Spark,
Is it possible to store a modified time on the database record which would help
you to get the updated documents?
That way you can avoid reindexing the documents that have not been updated.
Best wishes,
Siva on 3GS
On Dec 8, 2011, at 15:40, yu shen shenyu...@gmail.com wrote:
Hi
That sounds like a good idea. I will check my schema configuration, and see
what I can do.
Thanks Mark and Siva for all the information, and see what I can do, and
update the result here.
Spark
2011/12/9 Siva Kommuri snv.komm...@gmail.com
Hi Spark,
Is it possible to store a modified time on
I don't know about DataImportHandler.
But the whole point of AutoCommit and AutoSoftCommit is that they happen
'Auto'matically. You couldn't measure how long they take from solrj, and if
you are using solrj to do the commit, it has nothing to do with auto commit.
If you want to commit from solrj
Kinda off topic why not directly use /solr/update/json?commitWithin=3000,
ofcourse with autocommit.maxtime and autoSoftCommit in configs.
I am not using the java client but send the docs using the http client.
It is doing fairly good for me.
Zoie is another option worth looking at. Had very good
Hi Mark,
My question inline.
2011/12/9 Mark Miller markrmil...@gmail.com
I don't know about DataImportHandler.
But the whole point of AutoCommit and AutoSoftCommit is that they happen
'Auto'matically. You couldn't measure how long they take from solrj, and if
you are using solrj to do the
I will try to use your methods.
2011/12/9 Sharath Jagannath shotsonclo...@gmail.com
Kinda off topic why not directly use /solr/update/json?commitWithin=3000,
ofcourse with autocommit.maxtime and autoSoftCommit in configs.
I am not using the java client but send the docs using the http client.
pseudo code:
SolrServer client = ...
UpdateRequest ureq = new UpdateRequest();
ureq.add(doc);
ureq.setParam(param, value);
ureq.setAction( UpdateRequest.ACTION.COMMIT, waitFlush, waitSearcher);
ureq.process(client);
It will say setAction is deprecated, but internal
[?], thanks. Try in a minute.
Spark
2011/12/9 Mark Miller markrmil...@gmail.com
pseudo code:
SolrServer client = ...
UpdateRequest ureq = new UpdateRequest();
ureq.add(doc);
ureq.setParam(param, value);
ureq.setAction( UpdateRequest.ACTION.COMMIT, waitFlush,
Hi guys,
I'm looking for NRT functionality or similar in Solr 3.5. Is that possible?
From what I understand there's NRT in Solr 4, but I can't figure out
whether or not 3.5 can do it as well?
If not, is it feasible to use an autoCommit every 1000ms? We don't
currently process *that* much data so
Thanks for reply. Please find my comments below. although after all the
basic RD, i wrote the issue.
On Fri, Dec 9, 2011 at 1:37 AM, Erick Erickson erickerick...@gmail.comwrote:
If you look at the admin/stats page, does it show anything
for numDocs and maxDocs?
i did look at stats. there are
82 matches
Mail list logo