Michael,
Your question is a little bit confusing. Business entities have attributes.
We model entities as documents, and attributes as fields. That's why adding
attributes to a filed is contradictory. Btw there few nearby conceptions in
Lucene, which are Payloads and TermsPositions.
About the
Hi,
Sorry if anyone found my question confusing.
I have an XML document that is nested
file
foo
bar/
thunk/
/foo
/file
And I know that with solr that you have to flatten your data, so I was just
trying to workout the best way to do a search on nested document. I was
looking to see if
It's been two months since I asked about wildcards and phonetic filters, and
finally the task of upgrading Solr to version 4.0 was prioritized in our
project. So the last couple of days I've been working on it. Another team
member upgraded Solr from 3.4 to 4.0, and I've been making changes
I have a doubt about how NRTCachingDirectory works.
As far as I've seen, it receives a delegator Directory and caches newly
created segments. So, if MMapDirectory use to be the default:
1.- Does NRTCachingDirectory works acting sort of as a wrapper of MMap
caching the new segments?
2.- If I have
Depends on exactly what you mean by versioning. But if you mean that
every document in Solr gets a version-number which is increased every
time the document is updated, all you need to do is to add a _version_
field in you schema: http://wiki.apache.org/solr/SolrCloud#Required_Config
Believe
Hi Mark,
Usually I was stopping them with ctrl-c but several times, one of the
servers was hung and had to be stopped with kill -9.
Thanks,
Alain
On Mon, Dec 10, 2012 at 5:09 AM, Mark Miller markrmil...@gmail.com wrote:
Hmmm...EOF on the segments file is odd...
How were you killing the
Lowercasing actually seems to work with Wildcard queries, but not with fuzzy
queries. Are there any reasons why I should experience such a difference?
Regards, Haagen
Den 10. des. 2012 kl. 13:24 skrev Haagen Hasle:
It's been two months since I asked about wildcards and phonetic filters,
Maybe it would? I don't completely get your drift. But you're talking about a
user writing a bunch of custom code to build, save, and query the bitmap
whereas working on top of existing functionality seems to me a lot more
maintainable on the user's part.
~ David
Mikhail,
Join of any nature should be chosen in last resort to using a single index
(when it's possible), especially if there is minimal to no denormalization of
data. In this specific case, if the average document had 200 temporal ranges
to index (100 days out, 2 per day), a Join based
Thanks for the information.
Bill
On Fri, Dec 7, 2012 at 3:04 PM, Mark Miller markrmil...@gmail.com wrote:
Yup, solr.xml is pretty much required - especially if you want to use
solrcloud.
The only reason anything works without is for back compat.
We are working towards removing the need
Its surprising that your full import is working for you. Both your query and
your deltaImportQuery have:
SELECT ID FROM...
...So both your full-import (query attr) and your delta-import
(deltaImportQuery attr) are only getting the ID field from your db.
Shouldn't you be at least be getting
Javi,
The center point of your query circle and the indexed point is just under
49.9km (just under your query radius); this is why it matched. I plugged in
your numbers here:
http://www.movable-type.co.uk/scripts/latlong.html
Perhaps you are misled by the projection you are using to view the
Hi all,
I have a solr instance with one field configured for highlighting as
follows:
str name=hlon/str
str name=hl.flconteudo/str
str name=hl.fragsize500/str
str name=hl.maxAnalyzedChars9/str
str name=hl.simple.prelt;font style=background-color:
yellowgt;/str
but I was willing to
Did you mean that you want multiple snippets?
http://wiki.apache.org/solr/HighlightingParameters#hl.snippets
-Original Message-
From: Rafael Ribeiro [mailto:rafae...@gmail.com]
Sent: Monday, December 10, 2012 11:20 AM
To: solr-user@lucene.apache.org
Subject: highlighting multiple
When you apply your analyzers/filters/tokenizers, the result value is kept in
the indexed; however, the input value is actually stored. For example, from
schema.xml file:
fieldType name=text class=solr.TextField positionIncrementGap=100
analyzer
charFilter
Bit maps can be done with a separate term for each bit. You search for
all of the terms in the bit range you want.
On 12/10/2012 06:34 AM, David Smiley (@MITRE.org) wrote:
Maybe it would? I don't completely get your drift. But you're talking about a
user writing a bunch of custom code to
yep!
I tried enabling this and settings various values bot no success... still
it only shows the first fragment of the search found...
I also saw this
http://lucene.472066.n3.nabble.com/hl-snippets-in-solr-3-1-td2445178.html
but increasing maxAnalyzedChars (that was already huge) produced no
: In logs I can see some UUID is being generated when adding new document:
: INFO: [selekta] webapp=/solr path=/update params={}
: {add=[504a4ea8-7b82-48b6-a2fa-b8dd56376fd7]} 0 27
: but when I query Solr I got:
: Dec 07, 2012 1:52:10 PM org.apache.solr.common.SolrException log
: SEVERE:
Rafael,
Can you share more on how you are rendering the results in your velocity
template? The data is probably being sent to you, but you have to loop through
and actually access the data.
-Original Message-
From: Rafael Ribeiro [mailto:rafae...@gmail.com]
Sent: Monday, December 10,
Sorry to ask a question again, but I want to round date(TireDate) and
TrieLongField, seems they don't support configuring analyzer: charFilter ,
tokenizer or filter.
What I should do? Now I am thinking to write my custom date or long field,
is there any other way? :)
Thanks :)
--
View this
Would http://search-lucene.com/?q=solr+join do it for you?
Otis
--
SOLR Performance Monitoring - http://sematext.com/spm/index.html
Search Analytics - http://sematext.com/search-analytics/index.html
On Mon, Dec 10, 2012 at 1:17 PM, Michael Jones michaelj...@gmail.comwrote:
Hi,
I realise
Hi,
Nope...they don't. Generally, I am not sure if I'd bother rounding this
information to reduce the index size. Have you determined how much index size
space you'll actually be saving? I am not confident that it'd be worth your
time; i.e. I'd just go with indexing/storing the time
How about aggregating all location fields into one searchable multi-Value
field using copyField? It could be an index-only collection. Then, you just
say all_locations:Crosby
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn:
I have stored multiple objects with the values;
uniqueUri
name
timestamp.
There can be multiple object with the same name, but they will have
different timestamps (and different uniqueUri)
I want to retrieve the object of a given name with the latest timestamp. As
an example I might have
1.
Hi All,
I am getting constant OOM errors on a SolrCloud instance. (3 shards, 2 solr
instance in each shard, each server with 22gb Of Memory, Xmx = 12GB for java
) .
Here is a error log:
http://pastie.org/private/dcga3kfatvvamslmtvrp0g
As of now Iam not indexing any more documents. The total
Thanks Shawn,
I am looking at README.txt file and jars/wars that came with Solr 4 binary
distribution. So, if it is out of date, should I do Jira request? Or are
documentation fixes handled differently?
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn:
Hi - the stack trace and preceding log entries look similar to what i've seen
and reported on. A patch has just been attached to the issue, perhaps you can
try it if the description matches your scenario and report back on Jira.
https://issues.apache.org/jira/browse/SOLR-4144
-Original
On 12/10/2012 3:51 PM, Alexandre Rafalovitch wrote:
Thanks Shawn,
I am looking at README.txt file and jars/wars that came with Solr 4 binary
distribution. So, if it is out of date, should I do Jira request? Or are
documentation fixes handled differently?
Yes, filing a jira issue is an
Thanks Markus. Is this issue only on 4.x and 5.x branches? I am currently
running a v recent build of 4.x branch with an applied patch.
I just want to make sure that this is not an issue with 4.0. In which case I
can think of applying my patch to 4.0 instead of 4x or 5x.
--Shreejay
--
: Looking at the dist directory on what I just downloaded, it appears that most
: of the functionality required for writing code related to Solr would actually
: be in apache-solr-core-4.0.0.jar, and depending on what you are doing, you may
: need one or more of the other jars there. It looks
Hi - We're using trunk (5x) but we don't see it on trunk builds from a few
months ago. In the case of the linked issue the oom occurs some time after
start up but i'm not sure this applies to you. You can test the patch if you
think it applies to you, we will test it tomorrow.
If the patch
Hi Chris (Hoss?),
The issue is that README refers to a specific file apache-solr-XXX.jar,
which does not exist. There is apache-solr-4.0.0.war which is referred in a
para before, but not this one. So, maybe the fix is just to say that there
is a bunch of jars now. (apache-solr-component-XXX.jar
Thanks Marcus. I will apply the patch to the 4x branch I have, and report
back.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrCloud-OOM-heap-space-tp4025821p4025858.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi David,
As it happens the points are using the right projection, I can see them in
the same position using the page you just provided.
There is something wrong with the radius of the circle though I need to
investigate that but it is a relief to know that there is nothing wrong
with Solr and
Hi there,
I'm new to Solr and DIH, recently I've been planning to use Solr/DIH to
index some local xml files. Following the DIH example page on solr wiki,
most things work fine, but I found that xml attributes with prefix cannot be
parse. take the following xml file to be indexed for instance:
: However my raw files are stored on some remote storage devices. I am able to
: get an InputStream object for the file to be indexed. To me it may seem
: awkward to have the file temporarily stored locally. Is there a way of
: directly passing the InputStream in (e.g. constructing ContentStream
Seems like a good idea. Could you open a JIRA issue for this task?
Mark
Sent from my iPhone
On Dec 10, 2012, at 6:44 PM, Alexandre Rafalovitch arafa...@gmail.com wrote:
Hello,
I lost good several hours on this, so wanted to check whether this is
fixable.
In the (binary) distribution of
For anyone else looking to run JSPs on solr 4.0, note that supplying
OPTIONS=jsp to the server etc doesn't work (checkout startup config in
start.jar and you'll see why) - don't bother with all that. Instead do the
following:
create a directory ext under: $SOLR_HOME\example\lib
copy the following
I missed this bug report! https://issues.apache.org/jira/browse/SOLR-3912
Will fix this very shortly. It's a problem with numShards=1.
- Mark
On Sun, Dec 9, 2012 at 4:21 PM, sausarkar sausar...@ebay.com wrote:
Thank you very much will wait for the results from your tests.
From: Mark Miller-3
Hi,
The fq one is a FilterQuery that only does matching, but not scoring. It's
results are stored in the filter cache, while the q uses the query cache.
Otis
--
SOLR Performance Monitoring - http://sematext.com/spm/index.html
On Mon, Dec 10, 2012 at 10:11 PM, Floyd Wu floyd...@gmail.com
Thanks Otis.
When talked about query performance(ignore scoring). To use fq is better?
Floyd
2012/12/11 Otis Gospodnetic otis.gospodne...@gmail.com
Hi,
The fq one is a FilterQuery that only does matching, but not scoring. It's
results are stored in the filter cache, while the q uses the
If you don't need scoring on it then yes, just use fq.
Otis
--
SOLR Performance Monitoring - http://sematext.com/spm/index.html
On Mon, Dec 10, 2012 at 10:34 PM, Floyd Wu floyd...@gmail.com wrote:
Thanks Otis.
When talked about query performance(ignore scoring). To use fq is better?
Javier,
I want to expand upon what I said; you might already get this point but
others may come along and read this and might not.
Naturally you are using a 2D map as most applications do (Google Earth is
the stand-out exception), and fundamentally this means the map is projected
-- it has to
You do not need to manage online and offline indexes. Commit when you are done
with your updates and Solr will take care of it for you. The changes are not
live until you commit.
wunder
On Dec 10, 2012, at 9:46 PM, Dikchant Sahi wrote:
Hi,
How can we do delta update of offline indexes?
Hi Walter,
Thanks for the response.
Commit will help to reflect changes on Box1. We are able to achieve this.
We want the changes to reflect in Box2.
We have two indexes. Say
Box1: Master DB has been setup. Data Import runs on this.
Box2: Slave running.
We want all the updates on Box1 to be
45 matches
Mail list logo