Hi to all,
I am using solr for searching in my application.
my problem is ,
for example if i want to serach
"what is java?" means ,
The highly matched result from solr should come (ie java based result ,which
should be a lose search ) .what kind of ANALYSER i have to use and how to
conf
No. DataImportHandler runs inside of Solr itself so it does not need to use
HTTP. It reads databases using Jdbc and uses API calls to create and load
documents into Solr. Benefits over custom solutions would include speed (no
HTTP/XML parsing overhead) and ofcourse all benefits of open source softw
On Jun 6, 2008, at 1:22 PM, Andreas von Hessling wrote:
Thanks to both of you.
I understand from your replies that setting the payloads for terms
(per-document) is easy and the BoostingTermQuery can be used to set
payloads on the query side. Getting this to work in Solr would
require si
Thanks for the advice Yonik.
We have new users at least every few hours so it would be kinda
difficult to maintain the indexes this way. However, we do have a
smaller set of tokens describing the different subscription sets
available (<100). Basically, each folder_id is attached to a cert
I can certainly do: search for the unique key or combination of other fields,
then put rest fields of this document plus new fields back to it.
I know this is not a too smart way, before I do that, is there any solr guru
out there who can think of a better way?
Thanks much,
- Kevin
-Origin
: Subject: Automatic Ranges for Numeric Facets
you might want to peruse this thread...
http://www.nabble.com/Dynamically-calculated-range-facet-to11314725.html#a11314725
...and the other threads i link to in my reply.
-Hoss
On Mon, Jun 9, 2008 at 7:44 PM, Stephen Weiss <[EMAIL PROTECTED]> wrote:
> However, in the plain text search, the user automatically searches through
> *all* of the folders to which they have subscribed. This means, for (good!)
> users who have subscribed to a large (1000+) number of folders, the
Does the DataImportHandler use HTTP to load documents into Solr?
Is there any (run-time) performance benefit to using it rather than a custom
MySQL -> XML solution?
Shalin Shekhar Mangar wrote:
>
> Hi Rusli,
>
> Without going into specifics, I can tell you that we're using
> DataImportHandl
Thank you Shalin,
That works fine. BTW is there an SVN repos just for the
DataImportHandler work somewhere? Or should one just keep referring
to the Jira ticket patches instead?
Thanks.
- Jon
On Jun 10, 2008, at 11:22 AM, Shalin Shekhar Mangar wrote:
Hi Jon,
Yes it is a bug. Please u
Marcus,
It sounds like you may just want to use a good server monitoring package that
collects server data and prints out pretty charts. Then you can show them to
your IT/budget people when the charts start showing increased query latency
times, very little available RAM, swapping, high CPU us
Hi,
Go for option 2. I don't know if Solr will detect document changes, but I
don't think it will. Field updates currently still mean delete+add (see
SOLR-139).
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
> From: wojtekpia <[EMAIL PROTECTE
If your detection algorytims are simple enough, implement them. If you think
they will be more expensive them updating all the index, don't detect
change.
Another thing to take into account is how much from your index changes. If
there are litle docs changing, re-indexing everything will be a wast
Marcus,
2008/6/10 Marcus Herou <[EMAIL PROTECTED]>:
> Well guys you are right... Still I want to have a clue about how much each
> machine stores to predict when we need more machines (measure performance
> degradation per new document). But it's harder to collect that kind of
> data.
> It sure i
Well guys you are right... Still I want to have a clue about how much each
machine stores to predict when we need more machines (measure performance
degradation per new document). But it's harder to collect that kind of data.
It sure is doable no doubt and is a normal sharding "algo" for MySQL.
Th
Hi,
I'm trying to optimize how I update documents in my index. My optimizations
will depend on how Solr behaves in the following scenarios:
1. I update a document with itself (i.e. no changes). Does Solr detect that
the document is unchanged and basically ignore my request?
2. I change only one
I have a question about how punctuation and other special characters are
handled in the Solr index when using the facets toolkit. I have an index of
employees and facets based on their employer. Attempt to constrain the
search based on facets works only as long as the company name doesn't
contain
Thanks Otis!
I'll see what I can figure out and let you know how I do.
Best,
Jacob
Otis Gospodnetic wrote:
> From what I can tell, I think you simply need a coordinator component that is
> aware of both M1 and M2, allows only one of them to be modified at the time,
> and (r)syncs the index from
It is a bug, nice catch
there needs to be a null check there in the method
can us just try replacing the method with the following?
private Node getMatchingChild(XMLStreamReader parser) {
if(childNodes == null) return null;
String localName = parser.getLocalName();
for (Node n :
I just forgot to mention the error related to the description below.
I get the following when running a full-import ( sorry for the
noise .. ) :
SEVERE: Full Import failed
java.lang.RuntimeException: java.lang.NullPointerException
at
org.apache.solr.handler.dataimport.XPathRecordRea
Thanks a lot, it works fine now, fetching subelements properly.
The only issue left is that the XPath syntax passed in the data-
config.xml does not seem to work properly. As an example, processing
the following entity :
This is my title
You can also use a Java IDE like Eclipse, Netbeans or Idea to apply patches.
On Tue, Jun 10, 2008 at 6:03 PM, Jón Helgi Jónsson <[EMAIL PROTECTED]>
wrote:
> Thanks for that. The patch in question is this one:
> http://issues.apache.org/jira/browse/SOLR-469
> I found this patching utility for Wind
Hi Jon,
Yes it is a bug. Please use the TemplateTransformer.java attached to the
following discussion and rebuild solr:
http://www.nabble.com/Re%3A-How-to-describe-2-entities-in-dataConfig-for-the-DataImporter--p17577610.html
I shall give an updated patch containing the fixes soon.
On Tue, Jun 1
>From what I can tell, I think you simply need a coordinator component that is
>aware of both M1 and M2, allows only one of them to be modified at the time,
>and (r)syncs the index from the most recently updated machine/index to the one
>it is about to switch to. I don't think there is a way to
Hi,
For some reason even the simplest template is causing me NPE when
using (Solr trunk) ... ie:
How its being used:
http://www.site.com/path/?id=$
{news.id}"/>
-or-
http://www.site.com/path/?id=123456"/>
Throw ...
WARNING: transformer threw error
java.lang.NullPointerException
at
org
That sounds great!
I'm not launching for the next few of months so I'll wait and hope
that 1.3 comes out before that and just use XML while I develop the
rest of the site.
Thank you!
On Tue, Jun 10, 2008 at 12:57 PM, Shalin Shekhar Mangar
<[EMAIL PROTECTED]> wrote:
> Hi Rusli,
>
> Without going
Hi Rusli,
Without going into specifics, I can tell you that we're using
DataImportHandler successfully in production with many millions of documents
being imported from databases. Try it out and let us know if you face any
problems. DataImportHandler will be part of Solr 1.3 (the next release) and
The configuration is fine but for one detail
The documents are to be created for the entity 'oldsearchcontent' not
for the root entity . so add an attribute rootEntity="false" for the
entity 'oldsearchcontentlist' as follows.
http://localhost/psc/trunk/ezfiles/list_old_content.php?limit=10&urls
for this specific one the binaries are attached
http://wiki.apache.org/solr/DataImportHandler#head-c24dc86472fa50f3e87f744d3c80ebd9c31b791c
--Noble
On Tue, Jun 10, 2008 at 6:14 PM, Noble Paul നോബിള് नोब्ळ्
<[EMAIL PROTECTED]> wrote:
> * install tortoise svn .
> * checkout the code
> * download t
Hello fellow Solr users !
I am in the process of trying to index XML documents in Solr. I went
for the DataImportHandler approach, which seemed to perfectly suit
this need. Due to the large amount of XML documents to be indexed
( ~60MB ), i thought i would hardly be possible to feed solr w
* install tortoise svn .
* checkout the code
* download the patch
* use tortoise svn to apply the patch
--Noble
On Tue, Jun 10, 2008 at 6:03 PM, Jón Helgi Jónsson <[EMAIL PROTECTED]> wrote:
> Thanks for that. The patch in question is this one:
> http://issues.apache.org/jira/browse/SOLR-469
> I f
I wanted to see what people think about using DataImportHandler and
importing directly from mysql rather than with XML.
I have already programmed mysql -> XML in PHP and configured Solr to
import it when I realized I could use DataImportHandler and skip the
XML stage. Are there any negatives of us
Thanks for that. The patch in question is this one:
http://issues.apache.org/jira/browse/SOLR-469
I found this patching utility for Windows, going to give it a go:
http://gnuwin32.sourceforge.net/packages/patch.htm
On Tue, Jun 10, 2008 at 12:11 PM, Jacob Singh <[EMAIL PROTECTED]> wrote:
> Hi Rusli
Hi Rusli,
Is there a URL you'd like to reference for where you got the patch?
That would probably help.
For windows I suppose you'll have to google around to find a version of
"patch" which runs there. Beyond Compare is a windows app which has
patching capabilities. patch is a program for *nix
This is a terribly simple question I bet.
I'm running Solr on Windows and would like to use the Data Import
RequestHandler patch. I have been trying to figure out how to install
this patch but been unsuccessful so far. How would I go about doing
this?
Thanks,
Jon
Hi again :)
I'm also working on a scenario where there is an architecture like this:
(here comes poor man's Visio)
M2
|
M1
|
---
/ \
S1 S2
The catch is M2 isn't always online. The idea being, M1 is online to
take small updates like removing a certain entry from index or one off
ch
Hey folks,
I'm messing around with running multiple indexes on the same server
using Jetty contexts. I've got the running groovy thanks to the
tutorial on the wiki, however I'm a little confused how the collection
distribution stuff will work for replication.
The rsyncd-enable command is simple
36 matches
Mail list logo