, 17.8.2009, ..., 17.8.879, ...
5 year: ... date of birth 17.8.2006, 17.8.2001, ..., 17.8.1006, ...
10 year: 17.8.2001, 17.8.1991, ..., 17.8.891, ...
My first idea was to index date in mmdd format (20110817) and use
modulo:
1 year: indexed_date % 1 = 0817
5 year: indexed_date % 5 = 10817
Hi All
I have this requirement of indexing and searching files (txt, doc,pdf) on my
disk using Solr Search which I have installed.
I am unable to find a relevant tutorial for the same, I would be thankfull if
anyone of you can actually help me out with the specific steps required.
Thanks
On Wed, Aug 17, 2011 at 12:01 PM, slaava slaav...@gmail.com wrote:
[...]
My first idea was to index date in mmdd format (20110817) and use
modulo:
1 year: indexed_date % 1 = 0817
5 year: indexed_date % 5 = 10817
10 year: indexed_date % 10 = 10817
but I didn't found something
Stupid me. The output file was named something else. I really need to make
a proper servlet mapping. Works now :D
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260843.html
Sent from the Solr - User mailing list
On Tue, Aug 16, 2011 at 03:58:29PM -0400, Grant Ingersoll wrote:
I know you mean well and are probably wondering what to do next [...]
Still, a short heads-up like Johnson's would seem OK?
After all, this is of concern to us all.
Regards
-- tomás
On Wed, Aug 17, 2011 at 12:20 PM, Jagdish Kumar
jagdish.thapar...@hotmail.com wrote:
Hi All
I have this requirement of indexing and searching files (txt, doc,pdf) on my
disk using Solr Search which I have installed.
I am unable to find a relevant tutorial for the same, I would be thankfull
Thanks for quick reply!
Yes, this is my backup solution. But I prefer some one-query approach -
there could be many results so I want use SolrQuery.start() and
SolrQuery.rows() and show persons in table with paging.
Are you sure mod() function is supported now? It isn't included in Math.*
I've got a master set up on a public IP and I'm using my laptop as the slave,
just trying to get a simple replication going. When I'm on my slave machine and
I look at the replication tab of the admin, it says my master is unreachable,
however, I can hit the master's replication handler using
I've been trying (unsuccessfully) to get multicore working for about a day and
a half now I'm nearly at wits end and unsure what to do anymore. **Any** help
would be appreciated.
I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default
Solr install seems to work fine.
Why do you need to collect both primary keys T1_ID_RECORD and
T2_ID_RECORD in your delta query. Isn't T2_ID_RECORD primary key value
enough to get all data from both tables? (you have table1-table2
relation as 1-N, right?)
On Thu, Aug 11, 2011 at 12:52 AM, Eugeny Balakhonov c0f...@gmail.com
Hi we are planning to migrate from solr 1.4.1 to solr 3.3 and I am doing a
manual performance comparison.
We have setup two different solr installations (1.4.1 and 3.3) on different
ports.
1. Both have same index (old lucene format index) of around 20 GB with 10
million documents and 60 fields
can you please try persistent=true in solr tag as per my knowledge it will
solve your 404 - Not found error.
Regards,
Gaurav
Date: Tue, 16 Aug 2011 12:44:45 -0700
From: dnsa...@gmail.com
To: solr-user@lucene.apache.org
Subject: Unable to get multicore working
I've been trying
Hi Nagendra,
Thanks a lot .. i will start working on NRT today.. meanwhile old settings
(increased warmSearcher in Master) have not given me trouble till now ..
but NRT will be more suitable to us ... Will work on that one and will
analyze the performance and share with you.
Thanks
Naveen
On Wed, Aug 17, 2011 at 3:12 AM, Tomas Zerolo
tomas.zer...@axelspringer.de wrote:
On Tue, Aug 16, 2011 at 03:58:29PM -0400, Grant Ingersoll wrote:
I know you mean well and are probably wondering what to do next [...]
Still, a short heads-up like Johnson's would seem OK?
After all, this is of
Le 17 août 2011 à 13:01, Robert Muir a écrit :
On Tue, Aug 16, 2011 at 03:58:29PM -0400, Grant Ingersoll wrote:
I know you mean well and are probably wondering what to do next [...]
Still, a short heads-up like Johnson's would seem OK?
After all, this is of concern to us all.
nothing to
At a first glance I think the problem is in the 'feature' element which is
set to 'title'.
The 'feature' element should contain a UIMA Feature of the type defined in
element 'type'; for example for SentenceAnnotation [1] defined in HMM Tagger
has 'only' the default features of a UIMA Annotation:
Glad to hear it.
BTW, I highly recommend the following documents on the web:
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters
The tutorial at http://lucene.apache.org/solr/tutorial.html
And, of course, the multi-core document at http://wiki.apache.org/solr/CoreAdmin
The book from
I set the defaultOperator at AND in shema.xml : solrQueryParser
defaultOperator=AND/
I use the defType=eDismax in my query. It works very well, but when I want
to use AND or OR operators, it doesn't use AND as the default operator
for the blanks I left without operators.
Examples:
field1:a
Hi,
We have noticed that many index.* directories are appearing on slaves (some
more than others).
e.g. ls shows
index/index.20110101021510/ index.20110105030400/
index.20110106040701/ index.20110130031416/
index.20101222081713/ index.20110101034500/ index.20110105075100/
That is a good suggestion. At the very least I can catch this error and
create a new connection when I see this - thanks.
On Sun, Aug 14, 2011 at 3:46 PM, Erik Hatcher erik.hatc...@gmail.comwrote:
Does instantiating a Solr::Connection for each request make things better?
Erik
On Aug
Paul Libbrecht-4 wrote:
Robert,
I believe, precisely, the objective of such a thread is to be helped by
knowledgeable techies into being able to do what you say.
If Johnson gave only 3 lines of details, such as claimed patent URLs or
dates, we might easily be able to tell him the
Hello,
i get a ClassNotFoundException for JaraWinklerDistance when i start the
solr example server.
I simply copied the server and uncommented the spellchecker in
example/conf/solr-config.xml
I did nothing else.
I already googled but didn't get a hint. Can someone help me please.
Thanks
I'd suggest looking at the logs of the master to see if the request is getting
thru or not, or if there are any errors logged there. If the master has a
replication config error, it might show up there.
We just went thru some master/slave troubleshooting. Here are some things that
you might
On 8/17/2011 6:46 AM, Valentin wrote:
I set the defaultOperator at AND in shema.xml :solrQueryParser
defaultOperator=AND/
I use the defType=eDismax in my query. It works very well, but when I want
to use AND or OR operators, it doesn't use AND as the default operator
for the blanks I left
On Tue, Aug 16, 2011 at 07:56:51AM +, tomas.zer...@axelspringer.de wrote:
Hello, Solrs
we are trying to filter out documents written by (one or more of) the authors
from
a mediumish list (~2K). The document set itself is in the millions.
[...]
Sorry. Forgot to say that we are using
Hi Mike, is your config like this?
Is queryAnalyzerFieldType matching your type of field to be indexed?
Is the field correct?
searchComponent name=spellcheck class=solr.SpellCheckComponent
str name=queryAnalyzerFieldTypetextSpell/str
lst name=spellchecker
str name=namejarowinkler/str
str
I'm having the same problem: I import my data using the DataImportHandler.
When the DIH runs, I see the changes in the index file. However, when I
query the index using SolrJ, the new results don't show up. I have to
restart my server to see the results using SolrJ. This is how I use solrj:
It would perhaps help if you reported what you mean by noticeably less time.
What were your timings? Did you run the tests multiple times?
One thing to watch for in testing: Solr performance is greatly affected by the
OS file system cache. So make sure when testing that you use the same
I'm doing the exact same migration... what I've accomplished so far
1. In solrconfig.xml i
put luceneMatchVersionLUCENE_33/luceneMatchVersion in the first line in
the config branch. Warnings go like crazy if you don't do that.
2. Highlighter shows a deprecated warning, i'm still
Thank you Sujit and Rob for your help; I took the easy way and created a new
field type that is identical to text, but with the stemmer removed. This seems,
so far, to work exactly as needed.
To help anyone else who comes across this issue, this is the field type I used:
fieldType
On Wed, Aug 17, 2011 at 8:51 AM, LaMaze Johnson lam...@zesco.com wrote:
Paul Libbrecht-4 wrote:
Robert,
I believe, precisely, the objective of such a thread is to be helped by
knowledgeable techies into being able to do what you say.
If Johnson gave only 3 lines of details, such
On 17/08/11 11:53, marotosg wrote:
Anyone has any experience with this plugin?.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SIREn-with-Solr-tp3261260p3261260.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
What would you like to know ?
I am
I have no plan to look at the patents, but there is some serious prior art in
faceted search. First, faceted classification for libraries was invented by S.
R. Ranganathan in 1933. Computer search for libraries dates from the 1960's,
probably. Combining the two is obvious, even back then.
patent rights only last 17 years then it is public domain.
On 08/17/2011 11:05 AM, Walter Underwood wrote:
I have no plan to look at the patents, but there is some serious prior art in
faceted search. First, faceted classification for libraries was invented by S.
R. Ranganathan in 1933.
What do you mean you don't want to display it? Generally you'd just
navigate to solr_response['response'] to ignore the header and just deal with
the main body.
But, there is an omitHeader parameter -
http://wiki.apache.org/solr/CommonQueryParameters#omitHeader
Erik
On Aug 17,
Hi,
Is it possible to restrict the /terms component output to the results
of a query?
thanks,
Darren
Hi,
Sorry for the top-quote: On a
mobile.
A discussion on the evils of
patents aside, shirley library
catalogues are prior art. I
remember such systems giving
lists of matches by category, if
maybe not counts.
Will look at the patent applications, but sheesh,
what a waste of time and resources.
I had put mm at 475%, so if i understand it well, for 4 words or less, it
has to match all the words.
In my tests, i did it with 3 words, so i don't understand the results...
thanks for your answer,
Valentin
--
View this message in context:
Hi,
Not sure if this is right forum to ask a question regarding installation.
But the admin page on solr screen pointed to this address.
We have followed the installation procedure mentioned at
http://www.summasolutions.net/blogposts/magento-apache-solr-set.
The solr server seems to work
Thanks for the reply.
I changed it to 'coveredText' as you recommended but that did not help, I
got the same error message.
solrconfig.xml now looks like the following
lst name=fieldMappings
lst name=type
str
name=nameorg.apache.uima.alchemy.ts.concept.ConceptFS/str
Valentin,
There is currently an open issue about this:
https://issues.apache.org/jira/browse/SOLR-2649 . I ran into this also and
ended up telling all of the application developers to always insert AND
between every user keywords. I was using an older version of edismax but the
person who
I've recently started using Solr and I'm stumped by a problem I'm currently
encountering. Given that I can't really find anything close to what I'm
trying to do on Google or the mailing lists, I figured I'd ask if anyone
here had suggestions on how to do it.
I currently have a schema that looks
I think not, but if what you could get a similar result by using faceting on
the field and set the parameter facet.mincount=1. It will be slower than the
TermsComponent.
On Wed, Aug 17, 2011 at 1:19 PM, Darren Govoni dar...@ontrenet.com wrote:
Hi,
Is it possible to restrict the /terms
Daniel,
This looks like a good usecase for FieldCollapsing (see
http://wiki.apache.org/solr/FieldCollapsing). Perhaps try something like:
group=truegroup.field=documentIdgroup.limit=1group.sort=version desc
James Dyer
E-Commerce Systems
Ingram Content Group
(615) 213-4311
-Original
Woah. That looks like exactly what I need. Thanks you very much. Is there
any documentation for how to do that using the SolrJ API?
On Wed, Aug 17, 2011 at 2:26 PM, Dyer, James james.d...@ingrambook.comwrote:
Daniel,
This looks like a good usecase for FieldCollapsing (see
Thanks. I will try that.
For future feature, it would be good to do filter queries on /terms if
the Solr gods are listening! Hehe.
On 08/17/2011 02:25 PM, Tomás Fernández Löbbe wrote:
I think not, but if what you could get a similar result by using faceting on
the field and set the
What is the latest version of Tika that I can use with Solr 1.4.1? it
comes packaged with 0.4. I tried 0.8 and it no workie.
This is the type of problem that's fun to think about...
: As for RandomSortField + function queries... I'm not sure I understand how I
: can use that to achieve what I need :-/
the RandomSortField was designed for simple sorting, ie...
sort=random_1234 desc
...but it can also be used as
For the request end, you can just use something like:
solrquery.add(group, true);
..etc..
For the response, you have 3 options:
1. specify group.main=truegroup.format=simple . (note: When I tested this
on a nightly build from back in February I noticed a significant performance
impact from
What is the latest version of Tika that I can use with Solr 1.4.1? it
comes packaged with 0.4. I tried 0.8 and it no workie.
When I was testing Tika last year, I used Solr build 1271 to get the most
recent Tika I could get my hands on at the time. That was before Solr 3.1, so
I expect it
For response option 1, would I add the group.main=true and
group.format=simple parameters to the SolrQuery object?
On Wed, Aug 17, 2011 at 3:09 PM, Dyer, James james.d...@ingrambook.comwrote:
For the request end, you can just use something like:
solrquery.add(group, true);
..etc..
For the
: I'm looking for an ANTLR parser that consumes solr queries and filters.
: Before I write my own, thought I'd ask if anyone has one they are
: willing to share or can point me to one?
I'm pretty sure that this will be imposisble to do in the general case --
arbitrary QParser instances (that
Yes:
solrquery.add(group.main, true);
solrquery.add(group.format, simple);
James Dyer
E-Commerce Systems
Ingram Content Group
(615) 213-4311
-Original Message-
From: Daniel Skiles [mailto:daniel.ski...@docfinity.com]
Sent: Wednesday, August 17, 2011 2:15 PM
To:
Hello, I am working on creating a auto-complete functionality for my platform
which indexes large ammounts of text (title + contents) - there is too much
data for a dictionary. I am using the latest version of Solr (3.3) and I am
trying to take advantage of the Suggester functionality.
Hi Dan,
I saw this command -
http://localhost:8983/solr/spell?q=ANYTHINGHEREspellcheck=truespellcheck.collate=truespellcheck.build=true
I tried to issue it and got 404 error that I did not have the path
/solr/spell
Should I add this file and what type of file is it.
I got to via he post on
Its not a file, it's a request handler. you add those in the solrconfig.xml
read here plz http://wiki.apache.org/solr/Suggester
2011/8/17 naeluh nae...@gmail.com
Hi Dan,
I saw this command -
so I add spellcheck.build=true to solrconfig.xml just anywhere and that will
wrk?
thks very much for your help
--
View this message in context:
http://lucene.472066.n3.nabble.com/Spell-Checker-tp1914336p3262744.html
Sent from the Solr - User mailing list archive at Nabble.com.
My origination uses Solr 4 because of our geospatial requirements. What
insight can you share (if any) regarding moving forward to a later nightly
build? Or, for those of you using 4.0 in a Production setting, when is it that
you move ahead?
I have the very very very same problem. I could copy+paste your message as
mine. I've discovered so far that bigger dictionaries work better for me,
controlling threshold is much better than avoid indexing one or twio fields.
Of course i'm still polishing this.
At this very moment I was looking
geospatial requirements
Looking at your email address, no surprise there. 8^)
What insight can you share (if any) regarding moving forward to a later
nightly build?
I used build 1271 (Solr 1.4.1, which seemed to be called Solr 4 at the time)
during some testing, and it performed well --
Hi all!
I have configured my schema to use the solr.ASCIIFoldingFilterFactory
filter, this way I'm able to search a word like ferias and get férias,
but when I try to search the exact word férias I got nothing as result.
Is there a way to configure both cases in the search?
Best Regards!
--
For the record, I'm starting work now on moving the Velocity response writer
back to a contrib module so these dependencies won't be embedded in the WAR
file (after I make the commit this week some time, most likely).
Erik
On Aug 17, 2011, at 15:35 , Chris Hostetter wrote:
: Caused
Are you using the same analyzer for both type=query and type=index? Can you
show us the fieldType from your schema?
-Michael
I've been indexing and reindexing stuff here with Shingles. I don't believe
it's the best approach. Results are interesting, but I believe it's not what
the suggester is meant to be.
I tried
fieldType name=textSuggestion class=solr.TextField
positionIncrementGap=10 stored=false multiValued=true
No, if you are trying to build a suggester (what It seems to be) please read
the url I sent you.
You'll need to create the suggester itself searchComponent
class=solr.SpellCheckComponent name=suggest
and the url handler requestHandler
class=org.apache.solr.handler.component.SearchHandler
Config your xml properly, reload your core (or reload solr) then commit.
This spellchecker is configured to build on commit str
name=buildOnCommittrue/str. Everytime you commit something, it will
rebuild your dictionary based on the configuration you selected.
2011/8/17 naeluh nae...@gmail.com
As far as I know, Solr's trunk is pretty stable, so you shoundl't have many
problems with it if you test it correctly. Lucid's search platform is built
upon the trunk (
http://www.lucidimagination.com/products/lucidworks-search-platform/enterprise
).
The one thing I would be concerned is with the
I have a large ec2 instance(7.5 gb ram), it dies every few hours with out of
heap memory issues. I started upping the min memory required, currently I
use -Xms3072M .
I insert about 50k docs an hour and I currently have about 65 million docs
with about 10 fields each. Is this already too much
I should say I'm running: Solr Specification Version: 4.0.0.2010.12.10.08.54.56
and by the looks of the version number I'm running something from Dec 12 of
last year.
Tomas: geofilt and geodist() are supported in 3.3? Along with the location and
point type? Quite frankly, 1.3/1.4, 3.3, 4.0
While I can't be as specific as other here will be, we encountered the
same/similar problem. We simply loaded up our servers with 48GB and life is
good. I too would like to be a bit more proactive on the provisioning front
and hopefully someone will come along and help us out.
FWIW and I'm
After a bit of work, we have 'spellchecking' up and going and we are happy with
the suggestions. I have not; however, ever been able to generate more than one
collation query. Is there something simple that I have overlooked?
Can u show us how is your schema and config?
I believe that's how collation is: the best match, only one.
2011/8/17 Herman Kiefus herm...@angieslist.com
After a bit of work, we have 'spellchecking' up and going and we are happy
with the suggestions. I have not; however, ever been able to
I've just checked my index size with and my data folder is 16GB. So if my
server only has 7.5 gb of ram, does that mean I can't reliably run solr on
this one box and its useless to optimize the box?
If so it look like its time to start using a cluster?
On Wed, Aug 17, 2011 at 2:28 PM, Herman
Keep in mind that a commit warms up another searcher and potentially doubling
RAM consumption in the back ground due to cache warming queries being executed
(newSearcher event). Also, where is your Xmx switch? I don't know how your JVM
will behave if you set Xms Xmx.
65m docs is quite a lot
Herman,
- Specify spellcheck.maxCollations with something higher than one to get more
than 1 collation.
- If you also want the spellchecker to test whether or not a particular
collation will return hits, also specify spellcheck.maxCollationTries
- If you also want to know how many hits each
If you only get one, best, collation then there is no point to my question;
however, since you asked...
The relevant sections:
Solrconfig.xml -
searchComponent name=spellcheck class=solr.SpellCheckComponent
str name=queryAnalyzerFieldTypetextDictionary/str
lst name=spellchecker
str
Thanks James, here are the settings that only yield the one collation:
static int count = 10;
static bool onlyMorePopular = true;
static bool extendedResults = true;
static bool collate = true;
static int maxCollations = 10;
static int maxCollationTries = 100;
static int maxCollationEvaluations =
I've only set set minimum memory and have not set maximum memory. I'm doing
more investigation and I see that I have 100+ dynamic fields for my
documents, not the 10 fields I quoted earlier. I also sort against those
dynamic fields often, I'm reading that this potentially uses a lot of
memory.
On Wed, Aug 17, 2011 at 5:56 PM, Jason Toy jason...@gmail.com wrote:
I've only set set minimum memory and have not set maximum memory. I'm doing
more investigation and I see that I have 100+ dynamic fields for my
documents, not the 10 fields I quoted earlier. I also sort against those
Version number's are confusing, no doubt's. 4.x is currently on trunk (and
was not released yet). 3.x is a maintained branch. There has been three
releases from 3.x branch: 3.1, 3.2 and 3.3.
Most of the spatial search stuff is available since 3.1 (includding geoflit,
geodist and the location
I quickly went through what you've got from your last 2 posts and do not see
any problems. You might want to double-check that your client is translating
the constant variable you've got for spellcheck.maxCollationTries correctly
in your query, or if you've got it in the request handler config
Thank you very much for this awesome config. I'm working on it as we speak.
2011/8/17 Herman Kiefus herm...@angieslist.com
If you only get one, best, collation then there is no point to my question;
however, since you asked...
The relevant sections:
Solrconfig.xml -
searchComponent
What can I do temporarily in this situation? It seems like I must eventually
move to a distributed setup. I am sorting on dynamic float fields.
On Wed, Aug 17, 2011 at 3:01 PM, Yonik Seeley yo...@lucidimagination.comwrote:
On Wed, Aug 17, 2011 at 5:56 PM, Jason Toy jason...@gmail.com wrote:
: we have a fairly complex taxonomy in our search system. I want to store the
: taxonomy revision that was used to built the Solr index. This revision
based on your wording, it sounds like this is an index that you don't ever
update incrementally, and just rebuild and deploy completlye new
: I'm using Solr (with multiple cores) in a Webapp and access the differnt
: cores using CommonsHttpSolrServer. As I would like to know, which cores are
: configured and what there status is I would like to get an instance of
: CoreContainer.
CoreContainer is an internal API inside of Solr -- so
: I am new to solr. Am facing an issue wherein the highlighting of the
: searchresults for matches is not working when I have set a unique field
: as:
:
: uniqueKeyid/uniqueKey
:
: If this is commented then highlighting starts working. I need to have a
: unique field. Could someone please
Hello all,
I was looking into finding a way to do filtering of documents based on
fields of other documents in the index. In particular I have a document
that will update very frequently and hundreds that will very rarely
change, but the rarely changing documents have a field that will change
Hi all,
This may be obvious. My question pertains to use of tokenizerFactory
together with SynonymFilterFactory. Which tokenizerFactory does one use to
treat synonyms with spaces as one token,
Example these two entries are synonyms: lms, learning management system
index time expansion would
Alexei,
From the log, I think Solr couldn't find colored fragmentsBuilder defined in
solrconfig.xml.
Can you check the following fragmentsBuilder/ setting in
searchComponenthighlighting...
/highlighting/searchComponent in solrconfig.xml?
koji
--
Check out Query Log Visualizer
EC2 7.5Gb (large CPU instance, $0.68/hour) sucks. Unpredictably, there are
errors such as
User time: 0 seconds
Kernel time: 0 seconds
Real time: 600 seconds
How can clock time be higher in such extent? Only if _another_ user used
600 seconds CPU: _virtualization_
My client have had constant
I agree with Yonik of course;
But
You should see OOM errors in this case. In case of virtualization
however it is unpredictable and if JVM doesn't have few bytes to output
OOM into log file (because we are catching throwable and trying to
generate HTTP 500 instead !!! Freaky)
I forgot to add: company from UK, something log related (please have a
look at recent LucidImagination -managed Solr Revolution conference blogs;
company provides log analyzer service; http://loggly.com/) they have
16,000 cores per Solr instance (multi-tenancy); of course they have at
least 100k
I've now had several false starts on different versions of Linux. openSUSE
wouldn't load up on my older dell server box because it didn't like my raid
controller and I've now been through CentOS 5.5, 5.6 and 6.0. I really
thought I was just about to get there. I get SolR to load up however the
93 matches
Mail list logo