On Fri, Jan 16, 2009 at 3:37 AM, Chris Hostetter
hossman_luc...@fucit.org wrote:
: I would like to know the advantages of moving from:
: a master-slave system using CollectionDistribution with all their .sh
: scripts
: http://wiki.apache.org/solr/CollectionDistribution
: to:
: use
Hi Otis,
thanks for your input. Although I agree that we may have to go over the
search result once more, I dont think doing so for the first result page
only, is sufficient.
In the first example I showed before, you can see that some of the desired
products (of shops B and C) in fact occur on
Hello all, as the subject says:
DIH XPathEntityProcessor fails with docs containing !DOCTYPE
This is using a solr nightly build from monday.
INFO: Server startup in 3623 ms
Jan 16, 2009 9:54:12 AM org.apache.solr.handler.dataimport.SolrWriter
readIndexerProperties
INFO: Read
stax parser automatically tries to fetch the DTD. How can we disable
that at the parser level?
On Fri, Jan 16, 2009 at 3:34 PM, Fergus McMenemie fer...@twig.me.uk wrote:
Hello all, as the subject says:
DIH XPathEntityProcessor fails with docs containing !DOCTYPE
This is using a solr nightly
Hi,
We are using Solr as a standalone server. And our web application sends a
HTTP request for searching. We receive JSON result back and use the result.
I had initailly asked about Searher
(http://www.nabble.com/What-do-we-mean-by-Searcher--td21436737.html). Now I
understand it better.
As
On Fri, Jan 16, 2009 at 4:17 PM, Manupriya manupriya.si...@gmail.comwrote:
Hi,
We are using Solr as a standalone server. And our web application sends a
HTTP request for searching. We receive JSON result back and use the result.
I had initailly asked about Searher
I gave few new shots today:
- with jetty and nightly build 16 Jan - same problem null pointer exception
- Then I decided I am not using solr multicore but rather tomcat to
handle this. So I get latets tomcat and again with using 1.3.0 solr.war
I setup all
as explained
Shalin is right about cache management, but fot the sake of knowledge,
everytime you sendd a commit to Solr, it will close the old Searcher and
open a new one.
2009/1/16 Shalin Shekhar Mangar shalinman...@gmail.com
On Fri, Jan 16, 2009 at 4:17 PM, Manupriya manupriya.si...@gmail.com
wrote:
Hello,
we are running Solr Implementation Version: 1.2.0 - Yonik - 2007-06-02
17:35:12 with Lucene Implementation Version: build 2007-05-20 in a
Tomcat application server Apache Tomcat/5.5.20 on a 64-bit Ubuntu 7.10.
For some time now (probably due to the continuous growth of the index,
which is
Hi Shalin,
Sorry, my post was unlcear. I am calling snappull from the slave, I
get that part, just obfuscating the domains incorrectly :). The
problem it seems is the index version.
The index was a 1.3 index and I've since moved to 1.4. It's been
working great so far until I hit the
On more thing that might help to identify the problem source (which I've
discovered just now) -- at the time the optimize finished (or broke
off), Tomcat logged the following:
r...@cms004:~# /opt/apache-tomcat-sba1-live/logs/catalina.2009-01-16.log
...
Jan 16, 2009 11:40:14 AM
Dominik,
Form my experiences with Solr, I have seen the index double its size during
an optimize.
I always keep my index directory in a partition that has at least its triple
size, so I have a margin for optimize and for natural doc quantity growth.
2009/1/16 Dominik Schramm
Thanks for the information!!
So can I safely assume that we will not face any memory issue due to caching
even if we do not send commit that frequently? (If we wont send commit, then
new searcher wont be initialized. So I can assume that the current searcher
will correctly manage cache without
I have raised an issue and a patch is provided.
Please confirm if it helps
https://issues.apache.org/jira/browse/SOLR-964
On Fri, Jan 16, 2009 at 3:52 PM, Noble Paul നോബിള് नोब्ळ्
noble.p...@gmail.com wrote:
stax parser automatically tries to fetch the DTD. How can we disable
that at the
No,
You can't assume that. You have to set a good autoCommit value for your
solrconfig.xml, so you don't run out of memory for no commiting to Solr
often, depending on your enviroment, memory share, doc size and update
frequency.
2009/1/16 Manupriya manupriya.si...@gmail.com
Thanks for the
We also make query on date ranges, it works when you use NOW function.
Try using :
?q=dateField:[* TO NOW]
?q=dateField:[NOW-45DAYS TO NOW]
?q=dateField:[NOW TO NOW+45DAYS]
Issue: Current issue which i am facing is with dismaxrequesthandler for date
field.
As soon as I add dateField in
It doesn't really make sense to use a date field in a dismax qf
parameter. Use an fq parameter instead, to filter results by a date
field.
Dismax is aimed for end users textual queries, not for field selection
or more refined typed queries like date or numeric ranges.
Erik
On
On Fri, Jan 16, 2009 at 7:14 PM, Jacob Singh jacobsi...@gmail.com wrote:
Hi Shalin,
Sorry, my post was unlcear. I am calling snappull from the slave, I
get that part, just obfuscating the domains incorrectly :). The
problem it seems is the index version.
The index was a 1.3 index and I've
I'm not sure what else to share here... I can try to code dive a bit
this week, but I imagine it is over my head and at a Lucene level.
Fortunately for us, re-indexing is not an issue, so we can manage it.
If someone can confirm it, I think it would be good to update the wiki
and let people know
Hi
I need a help with boolean queries in my custom RequestHandler. The purpose
of the handler is to translate human readable
date (like January 1990 or 15.2.1983 or 1995) into two date range fields
using internal date representation.
E.g. simple search 'q=chronological:1942' translates to
Julian,
This is with the nightly from jan 12.
I am using mutli core and playing about with DIH. I cant its
Interactive development mode to work properly and suspect
that to use it I need to run in single core mode.
I am still developing, so I have nothing setup within tomcat startup
files, it
I think it's pretty easy to check if SOLR is alive. Even from a shell
script, a simple command like
curl -iIs --url http://solrhost/solr/select?start=0rows=0; | grep -c
HTTP/1.1 200 OK
will return 1 if the response is an HTTP 200. If the return is not 1,
then there is a problem. A
Sorry to all, there was a terrible bug in my code.
I should have checked whether the query was changed by
(q.toString().equals(newQuery.toString()) instead of (q != newQuery)!
Hana wrote:
Hi
I need a help with boolean queries in my custom RequestHandler. The
purpose of the handler
Hey there,
I would like to know if there is any way to customize the output xml that
contains the response. I have been checking the source and looks for me that
should be something close to XMLWriter.java and XMLResponseWriter.java but
not sure about that... Is there any way to handle a pluggin
On Jan 16, 2009, at 12:26 PM, Marc Sturlese wrote:
I would like to know if there is any way to customize the output xml
that
contains the response. I have been checking the source and looks for
me that
should be something close to XMLWriter.java and
XMLResponseWriter.java but
not sure
At a high level, I'm trying to do some more intelligent searching using
an app that will send multiple queries to Solr. My current issue is
around multi-valued fields and determining which entry actually
generated the hit for a particular query.
For example, let's say that I have a
Thanks, I have to study it but I think php serialized format will be the best
for my case...
Erik Hatcher wrote:
On Jan 16, 2009, at 12:26 PM, Marc Sturlese wrote:
I would like to know if there is any way to customize the output xml
that
contains the response. I have been checking the
Solr does have a PHPResponseWriter:
http://wiki.apache.org/solr/SolPHP?highlight=(CategoryQueryResponseWriter)|((CategoryQueryResponseWriter))
http://lucene.apache.org/solr/api/org/apache/solr/request/PHPResponseWriter.html
Bill
On Fri, Jan 16, 2009 at 1:09 PM, Marc Sturlese
Hi
When I parse DateRange query in a custom RequestHandler I get the date in
format -MM-dd'T'HH:mm:ss, but I would like it with the trailling 'Z' for
UTC time. Is there a way how to set the desired date format?
Here is a snippet of the code:
SolrParams p = req.getParams();
String query =
Hi again,
Problem solved. We gave the index space five times its original size, just
to be safe. Then the optimize finished without problems and without
leaving deleted, yet still open files behind. During the optimize the
index directory peaked at almost three times its size, and now,
: How do I find out the status of a slave's index? I have the following
: scenario:
FYI: regardless of where your index comes from (java replication, script
replication, manual mucking, etc...) the stats page of any solr core will
tell you the index version number in use by the current
: Also can can open admin at
:
: http://localhost:8983/solr/core1/admin/
: But then trying to ping
: http://localhost:8983/solr/core1/admin/ping
:
: I get error 500 INTERNAL SERVER ERROR
The ping URL only works if you configure the PingRequestHandler with a
default query, or use the
Hello list.
Are there standardized lists out there for populating synonyms.txt?
Entering the terms manually seems like a bad idea.
Thanks for your help.
Daniel
Synonyms are domain-specific. A food site would list arugula and
rocket as synonyms, but that would be a bad idea for NASA.
wunder
On 1/16/09 1:35 PM, Daniel Lovins daniel.lov...@yale.edu wrote:
Hello list.
Are there standardized lists out there for populating synonyms.txt?
Entering the
Yeah, I saw that... I am wandering if php serialized response use UTF-8
encoding by default...
Thanks
Bill Au wrote:
Solr does have a PHPResponseWriter:
http://wiki.apache.org/solr/SolPHP?highlight=(CategoryQueryResponseWriter)|((CategoryQueryResponseWriter))
This has likely been covered, and I've tried searching through the
archives, but having trouble finding an answer.
On OSVDB.org, if you search for:
title:PHPGroupWare
You get...nothing
if you search for:
title:phpGroupWare
(which is how the entry is indexed originally), you get a match of
36 matches
Mail list logo