Thankyou Tommy. But the real problem here is that the xml is dynamic and the
element names will be different in different docs which means that there will
be a lot of field names to be added in schema if I were to index those xml
nodes separately.
Is it possible to have nested indexing (xml
You could write yourr query like
q=filedname1:searchValue AND fieldName2:value OR fieldName3: Value
Regards,
Manas
From: Suram [mailto:reactive...@yahoo.com]
Sent: Wed 3/17/2010 12:44 AM
To: solr-user@lucene.apache.org
Subject: Issue in search
In solr how
Solr is behaving a bit weirdly for some of the search terms. EG:
co-ownership, co ownership.
It works fine with terms like quasi-delict, non-interference etc.
The issue is, its not return any excerpts in highlighting key of the
result dictionary. My search query is something like this:
Hello,
You can see what happen (which analyzer are used for this field and which is
the output of the analyzers) with this search using the analysis page of the
solr default web page. I assume you are using the same analyzers and
tokenizers in indexing and searching for this field in your schema.
I think I know many sites that ignore this warning... using mod_proxy
is quite an easier method in comparison to this. Maybe you are aiming
at millions of queries per second, then you should consider that. I
wonder if it makes sense before.
paul
Le 17-mars-10 à 04:36, blargy a écrit :
Hi List,
we are running a marketplace which has about a comparable functionality like
ebay (auctions, fixed-price items etc).
The items are placed on the market by users who want to sell their goods.
Currently we are using Sphinx as an indexing engine, but, as Sphinx returns
only document ids
I'm using the SnowballPorterFilterFactory for stemming French words. Some
words are not reconginized by this stemmer; I wonder wether, like synonyms
processing, the stemmers have the option of expansion.
Thanks.
Hi,
Solr is running on top of Lucene and as far as I know Lucene knows only one
approach how to update the document field content: that is delete first and
then (re)index with new values.
However, saying this it does not mean you can not implement what you need.
Take a look at ParallelReader API
Having been thinking about your questions again and I think that if you are
expecting that the price value will be changing a lot, especially when
talking about auctions then you should consider not storing the actual price
into the full text index but into some fast datastore. Some kind of
I was reading Scaling Lucen and Solr
(http://www.lucidimagination.com/Community/Hear-from-the-Experts/Articles/Scaling-Lucene-and-Solr/)
and I came across the section StopWords.
In there it mentioned that its not recommended to remove
stop words at index
time. Why is this the case? Don't
Solr is behaving a bit weirdly for some of the search
terms. EG:
co-ownership, co ownership.
It works fine with terms like quasi-delict,
non-interference etc.
The issue is, its not return any excerpts in highlighting
key of the
result dictionary. My search query is something like this:
On 16.03.2010, at 15:42, Lukas Kahwe Smith wrote:
Hi,
I am trying to use $deleteDocById to delete rows based on an SQL query in my
db-data-config.xml. The following tag is a top level tag in the document
tag.
entity name=company_del query=SELECT e.id AS `$deleteDocById` ROM
Hi Mortiz,
You can take a look on the project ZOIE -
http://code.google.com/p/zoie/. I think it's that what are you looking
for.
br
Krzysztof
On Wed, Mar 17, 2010 at 9:49 AM, Moritz Mädler m...@moritz-maedler.de wrote:
Hi List,
we are running a marketplace which has about a comparable
If you dont' plan on filtering/ sorting and/or faceting on fast-changing
fields it would be better to store them outside of solr/lucene in my
opinion.
If you must: for indexing-performance reasons you will probably end up with
maintaining seperate indices (1 for slow-changing/static fields and 1
On 17.03.2010, at 11:36, Lukas Kahwe Smith wrote:
On 16.03.2010, at 15:42, Lukas Kahwe Smith wrote:
Hi,
I am trying to use $deleteDocById to delete rows based on an SQL query in my
db-data-config.xml. The following tag is a top level tag in the document
tag.
entity
Hi all,
We're meeting up at the Elgin just by Ladbroke Grove on the 6th for a
bit of relaxed chat about search, and related technology. Come along,
we're nice.
http://www.meetup.com/london-search-social/calendar/12781861/
It's a regular event, so if you want prior warning about future
meetups
Have you considered an XML database? Because this is exactly what they are
designed to do.
eXist is open source, or you can use Mark Logic (my employer), which is much
faster and more scalable. We do give out free academic and community licenses
for Mark Logic.
wunder
On Mar 16, 2010, at
The configuration is correct and it works perfectly for French. So far, all
the French words I tried got stemmed correctly; except the word studios.
This is why I thought about expansion, perhaps I might need it for other
words.
Thanks,
-Saïd
2010/3/17 Erick Erickson erickerick...@gmail.com
Hmm. Unfortunately that didn't work. Same problem - Solr doesn't report an
error, but the data doesn't get extracted. Using the same PDF with my previous
/Lib contents works fine.
Any other ideas?
These are the jar files I have in my /Lib
apache-solr-cell-1.4-dev.jar
asm-3.1.jar
That discussion cites a paper via a URL:
http://doc.rero.ch/lm.php?url#16;00,43,4,20091218142456-GY/Dolamic_Ljiljana__When_Stopword_Lists_Make_the_Difference_20091218.pdf
Unfortunately when I go to this URL I get:
L'accès à ce document est limité.
But I tracked down the paper. Here is its
To convert an XMLTYPE to CLOB use the getClobVal() method like this:
SELECT d.XML.getClobVal() FROM DOC d WHERE d.ARCHIVE_ID = '${doc.ARCHIVE_ID}'
-Original Message-
From: Shawn Heisey [mailto:s...@elyograg.org]
Sent: Tuesday, March 16, 2010 7:37 PM
To: solr-user@lucene.apache.org
Very nice. I'd suggest adding a link to the wiki near the tutorial link.
-Grant
On Mar 16, 2010, at 11:44 PM, Juan Pedro Danculovic wrote:
Hi all, we translated the Solr tutorial to Spanish due to a client's
request. For all you Spanish speakers/readers out there, you can have a look
at it:
They apparently moved it .. it's here now:
http://doc.rero.ch/lm.php?url=1000,43,4,20091218142456-GY/Dolamic_Ljiljana_-_When_Stopword_Lists_Make_the_Difference_20091218.pdf
--
From: Glen Newton glen.new...@gmail.com
Sent: Wednesday, March 17, 2010
On Mar 16, 2010, at 9:51 PM, blargy wrote:
I was reading Scaling Lucen and Solr
(http://www.lucidimagination.com/Community/Hear-from-the-Experts/Articles/Scaling-Lucene-and-Solr/)
and I came across the section StopWords.
In there it mentioned that its not recommended to remove stop
On Wed, Mar 17, 2010 at 11:48 AM, Grant Ingersoll gsing...@apache.org wrote:
Yes and no. Putting our historian hat on, stop words were often seen as
contributing very little to scores and also taking up a lot of room on disk
back in the days when disk was very precious. Times, as they say,
On 03/17/2010 12:03 PM, Robert Muir wrote:
On Wed, Mar 17, 2010 at 11:48 AM, Grant Ingersollgsing...@apache.org wrote:
Yes and no. Putting our historian hat on, stop words were often seen as
contributing very little to scores and also taking up a lot of room on disk
back in the days
Hi William,
We are facing the same issue as yourself.. just thought of checking if you
had already resolve this issue?
Thanks,
Barani
William Pierce-3 wrote:
Folks:
I am seeing this exception in my logs that is causing my replication to
fail.I start with a clean slate (empty data
Hi,
One of my collegue back in India is not able to replicate the index present
in the Servers (USA).
I am now thinking if this is due to any proxy related issue? He is getting
the below metioned error message
Is there a way to configure PROXY in SOLR config files?
Server logs
INFO: []
How can i make related search in solr.if i search ipod i need to get answer
like ipodsuffle,ipodnano,ipone with out using morelikethis option
--
View this message in context:
http://old.nabble.com/related-search-tp27933778p27933778.html
Sent from the Solr - User mailing list archive at
Hi Chak
Rather than comparing the overall size of your index to the RAM available
for the OS disk cache, you might want to look at particular files. For
example if you allow phrase queries, than the size of the *prx files is
relevant, if you don't, you can look at the size of your *frq files.
I'm wondering if there's been any progress on an issue described a
year or so ago in More details on my MoreLikeThis mlt.qf boosting
problem http://markmail.org/thread/nmabm5ly3wk2nqyy,
where it was pointed out that the MoreLikeThis handler only queries
one field for each of the interesting terms
I am using the DataImportHandler to index 3 fields in a table: an id, a date,
and the text of a document. This is an Oracle database, and the document is an
XML document stored as Oracle's xmltype data type. Since this is nothing more
than a fancy CLOB, I am using the ClobTransformer to extract
I'm trying to use the Dismax request handler, and thanks to the list, I
fixed one problem, which was the existing configs in solrconfig.xml.
I'm now just not getting any result from the query though. I changed
the dismax section in solrconfig.xml to this:
requestHandler name=dismax
On Mar 17, 2010, at 3:38 PM, Alex Thurlow wrote:
I'm trying to use the Dismax request handler, and thanks to the
list, I fixed one problem, which was the existing configs in
solrconfig.xml. I'm now just not getting any result from the query
though. I changed the dismax section in
Incidentally, I tried adding this:
datasource name=f type=FieldReaderDataSource /
document
entity dataSource=f processor=XPathEntityProcessor
dataField=d.text forEach=/MESSAGE
field column=body xpath=//BODY/
/entity
/document
But this didn't seem to change
The XPath parser in the DIH is a limited implementation. The unit test
program is the only enumeration (that I can find) of what it handles:
http://svn.apache.org/repos/asf/lucene/solr/trunk/contrib/dataimporthandler/src/test/java/org/apache/solr/handler/dataimport/TestXPathRecordReader.java
Another option is the ExternalFileField:
http://www.lucidimagination.com/search/document/CDRG_ch04_4.4.4?q=ExternalFileField
This lets you store the current prices for all items in a separate
file. You can only use it in a function query, that is. But it does
allow you to maintain one Solr
You can use dynamic fields (wildcard field names) to add any and all
element names. You would have to add a suffix to every element name in
your preparation, but you will not have to add all of the element
names to your schema.
On Wed, Mar 17, 2010 at 7:04 AM, Walter Underwood
The localhost URLs have no port numbers.
Is there a more complete error in the logs?
On Wed, Mar 17, 2010 at 9:15 AM, JavaGuy84 bbar...@gmail.com wrote:
Hi William,
We are facing the same issue as yourself.. just thought of checking if you
had already resolve this issue?
Thanks,
Barani
A 5-second connection is not going to work trans-globally. The
replication engine is generally tested in local sites.
If it is possible to set defaults for the Apache Commons http classes
via system properties, that might let this work. This doc does not
seem promising:
Try cutting back Solr's memory - the OS knows how to manage disk
caches better than Solr does.
Another approach is to raise and lower the queryResultCache and see if
the hitratio changes.
On Wed, Mar 17, 2010 at 9:44 AM, Siddhant Goel siddhantg...@gmail.com wrote:
Hi,
Apparently the
This could be the problem: the text field in the example schema is
indexed, but not stored. If you query the index with text:monkeys it
will find records with monkeys, but the text field will not appear
in the returned XML because it was not stored.
On Wed, Mar 17, 2010 at 11:17 AM, Neil
: I want to *search* on title and content, and then, within these results
*boost* by keyword.
...
: str name=bqkeyword:(*.*)^1.0/str
:
: But I'm fairly sure that this is boosting on all keywords (not just ones
matching my search term)
correct.
: Does anyone know how to achieve
: tagskey,value , where key is String and value is Int.
: key is a given tag and value is a count of how many users used this tag for
: a given document.
:
: How can I index and store a key/value type of field? such that one can
: search on the values as well as keys of this field.
It depends
keep in mind that the xpath is case-sensitive. paste a sample xml
what is dataField=d.text it does not seem to refer to anything.
where is the enclosing entity?
did you mean dataField=doc.text.
xpath=//BODY is a supported syntax as long as you are using Solr1.4 or higher
On Thu, Mar 18,
Hi All,
Can some body please explain, What is the use of Solr configuration in
Katta master and nodes after integrating katta into Solr (1395 Patch).
Thanks,
vsreddy
The katta master is set up to act as a solr master server.
The config there is to be setup to distribute requests to the individual shards.
The solr config in the nodes is the default config to use, to start the solr
instance in the node.
On 3/17/10 9:05 PM, V SudershanReddy vsre...@huawei.com
47 matches
Mail list logo