Looks I am getting exception as below
May 22, 2013 10:52:11 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.NumberFormatException: For input string: [3 TO 9] OR
salary:0
at
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at
Hi,
I checked out from here
http://svn.apache.org/repos/asf/lucene/dev/tags/lucene_solr_4_3_0 and
downloaded the latest patch LUCENE-2899-current.patch.
Applied the patch ok but when I did 'ant compile' I got the following error:
==
[javac]
Hello again,
Is any one could help me, plase
David
Le 22/05/2013 18:09, It-forum a écrit :
Hello to all,
I'm trying to setup solr 4.2 to index and search into french content.
I defined a special fieldtype for french content :
fieldType name=text_fr class=solr.TextField
Hello,
I think you're confusing three different things:
1) schema and fields definition is for precision/recall: treating
differently a field means different search results and results ranking
2) the pomppe a chaler problem is more a spellchecking problem
Consider the following:
Solr 4.3, 2 node test cluster, each is a leader.
During (or immediately after, before hard commit) indexing I shutdown one of
them and restart later.
The tlog is about 200Mb size.
I see recurring 'Reordered DBQs detected' in the log, seems like an endless
loop because THE
Hi Kevenz,
kevenz wrote
...
String sql = indexType:219 AND
geo:Contains(POINT(114.078327401257,22.5424866754136));
...
Then I got an error at java.lang.IllegalArgumentException: missing
parens: Contains. Is there any suggestion?
First of all, if your query shape is a point, then use
a small change: it's not an endless loop, but a painfully slow processing
which includes running a delete query and then insertion. Each document from
the tlog takes tens of seconds to process (more than 100 times slower than
during normal insertion process)
--
View this message in context:
You can also think about using a SynonymFilter if you can list the
misspelled words.
That's a quick and dirty solution.
But it's easier to add a pomppe - pompe in a synonym list than tuning
a phonetic filter.
NB: an indexation is required whenever the synonyms file change
Franck Brisbart
Le
Hi,
Release 3.0.3 was tested with :
* Oracle Java 6 but should work fine with version 7
* Tomcat 5.5 and 6 and 7
* PHP 5.2.x and 5.3.x
* Apache 2.2.x
* MongoDB 64 bits 2.2 (know issue with 2.4)
The new release 4.0.0-alpha-2 is available under Github -
https://github.com/bejean/crawl-anywhere
Hello, guys!
I'm running Solr 4.3.0 and I've notice an strange behavior during
distributed queries execution. Currently I have three Solr servers as
shards and I when I do the following query...
http://localhost:11080/twitter/data/select?q=*:**rows=10*
Oh thank you Chris, this is much clearer, and thank you for updating the
Wiki too.
On 05/22/2013 08:29 PM, Chris Hostetter wrote:
: NOTE: make sure norms are enabled (omitNorms=false in the schema.xml) for
: any fields where the index-time boost should be stored.
:
: In my case where I only
Hello,
Tx Cristian for your details.
I totally agreed with your explanation, this is 2 differents aspect
which I need to solve.
Could you clarify few more thinks :
- SpellchekComponent and Phonetic, should be use while indexing or only
while querying ?
- Does spellcheck component return
In case anyone is interested, I solved my problem using the grouping
feature:
*query* -- filter query (if any)
*field* -- field that you want to count (in my case field B)
SolrQuery solrQuery = new SolrQuery(query);
solrQuery.add(group, true);
solrQuery.add(group.field, B); // Group by the field
Could you clarify few more thinks :
- SpellchekComponent and Phonetic, should be use while indexing or only
while querying ?
SpellCheck: you can define a specific field for spellchecking (in this
sense it's a query/schema time) or you can create a specific vocabulary for
spell-checking. I
*str name=msgorg.apache.solr.search.SyntaxError: Cannot parse
'*mm_state_code:(**TX)*': Encountered : : at line 1, column 14.
Was expecting one of:*
This suggests to me that you kept the df parameter in the query hence it
was forming mm_state_code:mm_state_code:(TX), can you try exactly they
Thanks Chris ,
for my 2nd Query (~1 returns words with 2 editing distance), it may be the
issue.
still m looking for my last issue. hope jira helps to resolve that.
Chris Hostetter-3 wrote
:
: 2) although I set editing distance to 1 in my query (e.g. worde~1), solr
: returns me results
Hello solr team,
I want to index multiple fields into one solr index entity, with the same id.
We are using solr 4.1
I try it with following source fragment:
public void addContentSet(ContentSet contentSet) throws
SearchProviderException {
...
Hi,
This is the situation, I have two sources of data in my dataimport handler,
one is huge, the other is tiny:
Source A: 10-20 records
Source B: 50,000,000 records
I was wondering what happens if I was to do a DIH just on Source A every 10
mins, and only run the DIH on source B every 24 hours.
Hello Folks,
Sorry, my last email was a bit messy, so I am sending it again.
I have a question about coordination factor to ensure my understanding
of this value is correct.
If I have documents that contain some keywords like the following:
Doc1: A, B, C
Doc2: A, C
Doc3: B, C
And my
This looks correct.
On Thu, May 23, 2013 at 7:37 AM, Kazuaki Hiraga kazuaki.hir...@gmail.comwrote:
Hello Folks,
Sorry, my last email was a bit messy, so I am sending it again.
I have a question about coordination factor to ensure my understanding
of this value is correct.
If I have
Please post the results of adding debug=query to the URL.
That'll tell us what the query parser spits out which is much
easier to analyze.
Best
Erick
On Wed, May 22, 2013 at 12:16 PM, Shankar Sundararaju
shan...@ebrary.com wrote:
This query returns 0 documents: *q=(+Title:() +Classification:()
bq: So cant we do fq on non-indexed field
No. By definition the fq clause is a search and
you can only search on indexed fields.
Best
Erick
On Wed, May 22, 2013 at 5:08 PM, gpssolr2020 psgoms...@gmail.com wrote:
Hi
i am trying to apply filtering on non-indexed double field .But its not
What's your max warming searcher value?
About warming queries, that may be _adding_ to your problem.
I'd first try removing many of them, especially if you have
your cache autowarm settings very high, try 16 or so.
Autowarming is all about pre-loading the caches etc, but you
reach diminishing
I need to know when a document is committed in SOLR - i.e. is searchable.
Is there anyone who has a solution on how to do this.
I'm aware of three methods to create hooks for knowing when a doc is added or
a commit is performed, but the doc(id) does not seem to be included for the
commit-hooks
On 23 May 2013 11:19, Kamal Palei palei.ka...@gmail.com wrote:
HI Rafał Kuć
I tried fq=Salary:[5+TO+10]+OR+Salary:0 and as well as fq=Salary:[5 TO 10]
OR Salary:0 both, both the cases I retrieved 0 results.
[...]
Please try the suggested filter query from the
Solr admin. interface, or by
by definition, there is no LUCENE_44 constant in a 4.3
distro! Just change it to LUCENE_43 (or whatever you find
in the Version class that suits your needs) or try this on a
4.x checkout.
Best
Erick
On Thu, May 23, 2013 at 2:08 AM, Patrick Mi
patrick...@touchpointgroup.com wrote:
Hi,
I
Tangential to the issue you raise is that this is a huge tlog. It indicates that
you aren't doing a hard commit (openSearcher=false) very often. That
operation will truncate your tlog which should speed recovery/startup.
You're also chewing up some memory with a tlog that size since pointers
to
A poller really is the most sensible, practical, and easiest route to go. If
you add the versions=true parameter to your update request and have the
transaction log enabled the update response will have the version numbers
for each document id, then the poller can also tell if an update has
That should work. Just watch out for (set value of)
preImportDeleteQuery. Otherwise, when you do full import you may
accidentally delete items from the other set.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is
I just skimmed your post, but I'm responding to the last bit.
If you have uniqueKey defined as id in schema.xml then
no, you cannot have multiple documents with the same ID.
Whenever a new doc comes in it replaces the old doc with that ID.
You can remove the uniqueKey definition and do what you
On 23. mai 2013, at 14:05, Jack Krupansky j...@basetechnology.com wrote:
Hi Jack,
thanks for your answer.
A poller really is the most sensible, practical, and easiest route to go. If
you add the versions=true parameter to your update request and have the
transaction log enabled the update
Thanks Erick..
i hope we cant do q also on non-indexed field.
Whats is the difference between q and fq other than cache .
Thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/fq-facet-on-double-and-non-indexed-field-tp4065457p4065604.html
Sent from the Solr - User
Hi,
in my attempt to migrate for m 3.6.x to 4.3.0 I stumbled upon an issue
loading the MySQL driver from the [instance]/lib dir:
Caused by: java.lang.ClassNotFoundException:
org.apache.solr.handler.dataimport.DataImportHandler
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at
Yes, by definition, a poller retries. But by picking a sensible default for
initial poll and retry (possibly an initial delay tuned to match average
update/commit time) couple with a traditional exponential backoff, that
should not be a problem at all. In other words, an average request would
I was playing around with spellcheck.alternativeTermCount and noticed that
if it is set to zero, Solr gives an exception with certain queries. Maybe
the value isn't supposed to be zero, but I don't think an exception is the
expected behaviour.
Rounak
Hi,
I am are building a website that lists restaurant information and I also
like to include the availability information.
I've created a custom ValueSourceParser and ValueSource that retrieve the
availability information from a MySQL database. An example query is as
follows.
Check the Solr log on startup - it will explicitly state which lib
directories/files will be used. Make sure they agree with where the DIH jars
reside. Keep in mind that the directory structure of Solr changed - use the
lib from 4.3 solrconfig.
Try to use DIH in the standard Solr 4.3 example
Hi
When having a collection with 3 shards en 2 replica's for each shard and
I want to split shard1. Does it matter where to start the splitshard
command in the cloud or should it be started on the master of that shard?
BR,
Arkadi
Huge tlogs seems to be a common problem. Should we make it flush automatically
on huge file size? Could be configurable on the updateLog tag?
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
23. mai 2013 kl. 14:03 skrev Erick Erickson erickerick...@gmail.com:
Regex expressions work on individual terms. Positional information is
irrelevant when it comes to regex matching - it's not matching across terms*.
The syntax allowed is documented here
https://lucene.apache.org/core/4_3_0/core/org/apache/lucene/util/automaton/RegExp.html
- it's not quite the
Hello Erick,
Thank you for your fast answer.
Maybe I don't exclaim my question clearly.
I want index many files to one index entity. I will use the same behavior as
any other multivalued field which can indexed to one unique id.
So I think every ContentStreamUpdateRequest represent one index
the hard commit is set to about 20 minutes, while ram buffer is 256Mb.
We will add more frequent hard commits without refreshing the searcher, that
for the tip.
from what I understood from the code, for each 'add' command there is a test
for a 'delete by query'. if there is an older dbq, it's
Any idea why I got a Broken pipe?
INFO - 2013-05-23 13:37:19.881; org.apache.solr.core.SolrCore;
[messages_shard3_replica1] webapp=/solr path=/select/
On 5/23/2013 1:51 AM, Luis Cappa Banda wrote:
I've query each Solr shard server one by one and the total number of
documents is correct. However, when I change rows parameter from 10 to 100
the total numFound of documents change:
I've seen this problem on the list before and the cause has been
This usually happens when the client sending the request to Solr has given up
waiting for the response (terminated the connection).
In your example, we see that the Solr query time is 81 seconds. Probably the
client issuing the request has a time-out of maybe 30 or 60 seconds.
André
On 5/23/2013 6:25 AM, Christian Köhler wrote:
in my attempt to migrate for m 3.6.x to 4.3.0 I stumbled upon an issue
loading the MySQL driver from the [instance]/lib dir:
Caused by: java.lang.ClassNotFoundException:
org.apache.solr.handler.dataimport.DataImportHandler
The best thing to do
Hi All,
I just started indexing data in my brand new Solr Cloud running on 4.2.1.
Since I am a big user of the grouping feature, I need to route my documents on
the proper shard.
Following the instruction found here:
http://docs.lucidworks.com/display/solr/Shards+and+Indexing+Data+in+SolrCloud
That's strange. The default value of rows param is 10 so you should be
getting 10 results back unless your StandardRequestHandler config in
solrconfig has set rows to 0 or if none of your fields are stored.
On Thu, May 23, 2013 at 7:40 PM, Jean-Sebastien Vachon
Can you give instructions on how to reproduce problem?
James Dyer
Ingram Content Group
(615) 213-4311
-Original Message-
From: Rounak Jain [mailto:rouna...@gmail.com]
Sent: Thursday, May 23, 2013 7:36 AM
To: solr-user@lucene.apache.org
Subject: Bug in spellcheck.alternativeTermCount
I
Also happens (same reason) if you are behind a smart load-balance and
it decides to time out and fail over.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is the quality of nature that keeps events from happening
Check out Gilt's presentation. It might give you some ideas, including
possibly on refactoring your entities around 'availability' as a
document:
http://www.lucenerevolution.org/sites/default/files/Personalized%20Search%20on%20the%20Largest%20Flash%20Sale%20Site%20in%20America.pdf
Regards,
I know. If a stop routing the documents and simply use a standard 'id' field
then I am getting back my fields.
I forgot to tell you how the collection was created.
http://localhost:8201/solr/admin/collections?action=CREATEname=CurrentnumShards=15replicationFactor=3maxShardsPerNode=9
Since I am
It seems to me that the behavior of the Core admin action CREATE has changed
when going from Solr 4.1 to 4.3.
With 4.1, I could re-configure an existing core (changing path/name to
solrconfig.xml for example). In 4.3, I get an error message:
SEVERE: org.apache.solr.common.SolrException:
Hi Arkadi,
It does not matter where you invoke that command because ultimately that
command is executed by the Overseer node. That being said, shard splitting
has some bugs whose fixes will be released with Solr 4.3.1 so I'd suggest
that you wait until then to use this feature.
On Thu, May 23,
Hi Patrick,
I think you should check out and apply the patch to branch_4x, rather than the
lucene_solr_4_3_0 tag:
http://svn.apache.org/repos/asf/lucene/dev/branches/branch_4x
Steve
On May 23, 2013, at 2:08 AM, Patrick Mi patrick...@touchpointgroup.com wrote:
Hi,
I checked out from here
If that can help.. adding distrib=false or shard.keys= is giving back
results.
-Original Message-
From: Jean-Sebastien Vachon [mailto:jean-sebastien.vac...@wantedanalytics.com]
Sent: May-23-13 10:39 AM
To: solr-user@lucene.apache.org
Subject: RE: Problem with document routing with
Yes, this did change - it's actually a protection for a previous change though.
There was a time when you did a core reload by just making a new core with the
same name and closing the old core - that is no longer really supported though
- the proper way to do this is to use SolrCore#reload,
On May 23, 2013, at 14:25 , gpssolr2020 psgoms...@gmail.com wrote:
Thanks Erick..
i hope we cant do q also on non-indexed field.
Whats is the difference between q and fq other than cache .
Thanks.
How do you expect to search on a field that is non-indexed (and thus
Hi Otis,
Thank you for your reply. I'm in the middle of that upgrade and will report
back when testing is complete. I'd like to get some nice set of reproducible
steps so I'm not just ranting on. :)
Regards,
Gilles
-Original Message-
From: Otis Gospodnetic
I think the wiki needs to be updated to reflect this?
http://wiki.apache.org/solr/CoreAdmin
If somebody adds me as an editor (AlanWoodward), I'll do it.
Alan Woodward
www.flax.co.uk
On 23 May 2013, at 16:43, Mark Miller wrote:
Yes, this did change - it's actually a protection for a
Alan, I've added AlanWoodward to the Solr AdminGroup page.
On May 23, 2013, at 12:29 PM, Alan Woodward a...@flax.co.uk wrote:
I think the wiki needs to be updated to reflect this?
http://wiki.apache.org/solr/CoreAdmin
If somebody adds me as an editor (AlanWoodward), I'll do it.
Alan
Thank you for your answer.
Do you mean I should index the availability data as a document in Solr?
Because the availability data in our databases is around 6,509,972 records
and contains the availability per number of seats and per 15 minutes. I also
tried this method, and as far as I know it's
Thanks!
Alan Woodward
www.flax.co.uk
On 23 May 2013, at 17:38, Steve Rowe wrote:
Alan, I've added AlanWoodward to the Solr AdminGroup page.
On May 23, 2013, at 12:29 PM, Alan Woodward a...@flax.co.uk wrote:
I think the wiki needs to be updated to reflect this?
Hi,
thanx for pointing this out to me.
1152 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrConfig
– Adding specified lib dirs to ClassLoader
org.apache.solr.core.SolrResourceLoader – Adding
: in my attempt to migrate for m 3.6.x to 4.3.0 I stumbled upon an issue loading
: the MySQL driver from the [instance]/lib dir:
:
: Caused by: java.lang.ClassNotFoundException:
: org.apache.solr.handler.dataimport.DataImportHandler
one of us is mistaken by what that error means. you say it
Interesting solution. My concern is how to select the most frequent terms
in the story_text field in a way that would make sense to the user. Only
including the X most common non-stopword terms in a document could easily
cause important patterns to be missed. There's a similar issue with only
Actually , It's pretty high end for most of the users. Rishi, u can post
the real h/w details and our typical deployment .
No :of cpus per node
No :of disks per host
Vms per host
Gc params
No :of cores per instance
Noble Paul
Sent from phone
On 21 May 2013 01:47, Rishi Easwaran
Hi
one of us is mistaken by what that error means. you say it means that
the MySQL driver isn't being loaded, but nothing in your mail suggests
to me that there is a problem loading hte MySql driver. what i see is
that Solr can't seem to load the DIH class, suggesting that the
I must add the shard.keys= does not return anything on two on my nodes. But
that is to be expected since I'm using a replication factor of 3 on a cloud of
5 servers
-Original Message-
From: Jean-Sebastien Vachon [mailto:jean-sebastien.vac...@wantedanalytics.com]
Sent: May-23-13
: I only SUSPECT of the MySQL driver being the culprit for the dataimporthandler
: jar is not getting loaded. Not sure!
the dataimporthandler *class* is not getting loaded the
dataimporthandler *jar* is not getting loaded.
: MySql classes, i would check your lib settings for loading the DIH
:
Mark, Alan,
thanks for explaining and updating the wiki.
When reloading the core using action=CREATE with Solr 4.1 I could specify the
path to schema and config. In fact I used this to reconfigure the core to use a
specific one of two prepared config files depending on some external index
Your right - that does seem to be a new limitation. Could you create a JIRA
issue for it?
It would be fairly simple to add another reload method that also took the name
of a new solrconfig/schema file.
- Mark
On May 23, 2013, at 4:11 PM, André Widhani andre.widh...@digicol.de wrote:
Mark,
Hi,
i'm totally confused ... DIH == DataImportHandler ... it's just an
acronym, you say you aren't using DIH, but you are having a problem
loading DIH, so DIH is used in your configs.
sorry for the confusion. I was just trying to say:
I use the example code from
solr-4.3.0/example/solr
and
When I create a core with Core admin handler using these request parameters:
action=CREATE
name=core-tex69bbum21ctk1kq6lmkir-index3
schema=/etc/opt/dcx/solr/conf/schema.xml
instanceDir=/etc/opt/dcx/solr/
config=/etc/opt/dcx/solr/conf/solrconfig.xml
Ok - yes, will do so tomorrow.
Thanks,
André
Von: Mark Miller [markrmil...@gmail.com]
Gesendet: Donnerstag, 23. Mai 2013 22:46
An: solr-user@lucene.apache.org
Betreff: Re: Core admin action CREATE fails for existing core
Your right - that does seem to be
Hi,
I just downloaded Apache Solr 4.3.0 from
http://lucene.apache.org/solr/. I
then got into the /example directory and started Solr with:
java -Djava.util.logging.config.file=etc/logging.properties
-Dsolr.solr.home=./example-DIH/solr/ -jar start.jar
I have not made any changes at
On 5/23/2013 3:50 PM, O. Olson wrote:
I just downloaded Apache Solr 4.3.0 from
http://lucene.apache.org/solr/. I
then got into the /example directory and started Solr with:
java -Djava.util.logging.config.file=etc/logging.properties
-Dsolr.solr.home=./example-DIH/solr/ -jar start.jar
Hossman did a presentation on something similar to this using spatial data
at a Solr meetup some months ago.
http://people.apache.org/~hossman/spatial-for-non-spatial-meetup-20130117/
May be helpful to you.
On Thu, May 23, 2013 at 9:40 AM, rajh ron...@trimm.nl wrote:
Thank you for your
To those of you who may have heard about the Lucene/Solr book that I and two
others are writing on Lucene and Solr, some bad and good news. The bad news:
The book contract with O’Reilly has been canceled. The good news: I’m going to
proceed with self-publishing (possibly on Lulu or even Amazon)
Hi Erick,
Here's the output after turning on the debug flag:
*q=text:()debug=query*
yields
response
lst name=responseHeader
int name=status0/int
int name=QTime17/int
lst name=params
str name=indenttrue/str
str name=qtext:()/str
str name=debugquery/str
/lst
/lst
result name=response
hi, all -
how can I retrieve the value out of a multivalued field in a customized
function query?I want to implement a function query whose first parameter is a
multi-value fileld, from which values are retrieved and manipulated.
however, I used the code but get exceptions - can not use
Yeah, you can't do that. You'll need to keep a copy of whichever value from
the multi-valued field you wish to be considered the value in a separate,
non-multi-valued field. Possibly using an update processor, such as one of:
FirstFieldValueUpdateProcessorFactory,
thanks, jack.
could you please give more details on using update processor?
Thanks.
- frank
From: j...@basetechnology.com
To: solr-user@lucene.apache.org
Subject: Re: howto: get the value from a multivalued field?
Date: Thu, 23 May 2013 20:06:34 -0400
Yeah, you can't do that. You'll need
(+(doc-id:3000 DisjunctionMaxQuery((Publisher:and^2.0 | text:and |
Classification:and^2.0 | Contributors:and^2.0 |
Title:and^3.0/no_coord
You're using edismax, not lucene. So AND is being considered as a search
term, not an operator, and the word 'and' probably exists in 631580
documents.
Okay... sorry I wasn't paying close enough attention. What is happening is
that the empty parentheses are illegal in Lucene query syntax:
str name=msgorg.apache.solr.search.SyntaxError: Cannot parse 'id:* AND
text:()': Encountered ) ) at line 1, column 15.
Was expecting one of:
Thank you for your comment.
Due to historical reasons, Our organization uses trunk version of Solr-4.0,
which is a bit old and unofficial version. And edismax always returns 1/2 as a
coordination value. So I wanted to make sure what this value would be like.
This will be a good reason to
86 matches
Mail list logo