if this is edit distance implementation, what is the result apply to CJK
query? For example, 您好~3
Floyd
2011/6/30 entdeveloper cameron.develo...@gmail.com
I'm using Solr trunk.
If it's levenstein/edit distance, that's great, that's what I want. It just
didn't seem to be officially
On Thu, Jun 30, 2011 at 2:06 AM, Yonik Seeley yo...@lucidimagination.comwrote:
On Wed, Jun 29, 2011 at 4:32 PM, eks dev eks...@googlemail.com wrote:
req.getSearcher().getFirstMatch(t) != -1;
Yep, this is currently the fastest option we have.
Just for my understanding, this method won't use
I have a hierarchical taxonomy of documents that I would like users to be
able to search either through search or drill-down faceting. The
documents may appear at multiple points in the hierarchy. I've got a
solution working as follows: a multivalued field labelled category which for
each
my solrconfig.xml configuration is as :
mainIndex
useCompoundFilefalse/useCompoundFile
ramBufferSizeMB32/ramBufferSizeMB
mergeFactor5/mergeFactor
maxMergeDocs10/maxMergeDocs
maxFieldLength1/maxFieldLength
unlockOnStartupfalse/unlockOnStartup
/mainIndex
my
wanted to use clustering in my search results, i configured solr for
clustering and i got following json for clusters. But i am not getting how
to use it to show in search results. as corresponding to one doc i have
number of fields and up till now i am showing name, description and id. now
in
Hi all,
does solar support german phonetic? Searching for how to add german phonetic
to
solr on google does not deliver good results, just lots of JIRA stuff. I
searched for cologne phonetic too. The wikis
Jürgen,
I haven't had the time to deploy it but i heard about Kölner Phonetik that
was to be contributed as part of apache-commons-codec.
It probably still is just a patch in a jira issue.
https://issues.apache.org/jira/browse/CODEC-106
The contribution was posted to commons-dev on
when i run solr/admin page i got this information, it shows optimize=true,
but i have not set optimize=true in configuration file than how it is
optimizing the indexes. and how can i set it to false then .
/Schema Information
Unique Key: UID_PK
Default Search Field: text
numDocs:
That's a good way. How does it perform?
Another way would be to store the parent topics in a field.
Whenever a parent node is drilled-into, simply search for all documents
with that parent. Perhaps not as elegant as your approach though.
I'd be interested in the performance comparison between
when i run solr/admin page i got this
information, it shows optimize=true,
but i have not set optimize=true in configuration file than
how it is
optimizing the indexes. and how can i set it to false then
.
/Schema Information
Unique Key: UID_PK
Default Search Field: text
Good question... I think in Lucene 4.0, the edit distance is (will be)
in Unicode code points, but in past releases, it's UTF16 code units.
Mike McCandless
http://blog.mikemccandless.com
2011/6/30 Floyd Wu floyd...@gmail.com:
if this is edit distance implementation, what is the result apply to
and if i want to set it as optimize=false then what i need to do ??
-
Thanks Regards
Romi
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-optimize-solr-indexes-tp3125293p3125474.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Paul,
thanks for the quick reply. I replaced commons-codec-1.4.jar with
commons-codec-1.5.jar to get the ColognePhonetic. In schema.xml I added
filter class=solr.PhoneticFilterFactory encoder=ColognePhonetic
inject=true/
but then I get
org.apache.solr.common.SolrException: Unknown
Thanks for the suggestion Mike, I will give that a shot. Having no
familiarity with FastVectorHighlighter is there somewhere specific I
should be looking?
On Wed, Jun 29, 2011 at 3:20 PM, Mike Sokolov soko...@ifactory.com wrote:
Does the phonetic analysis preserve the offsets of the original
It's going to be a bit complicated, but I would start by looking at
providing a facility for merging an array of FieldTermStacks. The
constructor for FieldTermStack() takes a fieldName and builds up a list
of TermInfos (terms with positions and offsets): I *think* that if you
make two of
Jürgen,
clearly the Cologne-phonetic was not yet supported, please read:
http://svn.apache.org/repos/asf/lucene/dev/trunk/solr/src/java/org/apache/solr/analysis/PhoneticFilterFactory.java
one would need to add the line about Cologne-phonetic and recompile.
It'd make sense to open a jira issue
--- On Thu, 6/30/11, Romi romijain3...@gmail.com wrote:
From: Romi romijain3...@gmail.com
Subject: Re: How to optimize solr indexes
To: solr-user@lucene.apache.org
Date: Thursday, June 30, 2011, 3:01 PM
and if i want to set it as
optimize=false then what i need to do ??
When calling
Sure, thanks for having a look!
By the way, if I attempt to hit a solr URL, I get this error, followed by
the stacktrace. If I set abortOnConfigurationError to false (I've found you
must put the setting in both solr.xml and solrconfig.xml for both cores
otherwise you keep getting the error), then
On Thu, 2011-06-30 at 11:38 +0200, Russell B wrote:
a multivalued field labelled category which for each document defines
where in the tree it should appear. For example: doc1 has the
category field set to 0/topics, 1/topics/computing,
2/topic/computing/systems.
I then facet on the
I'm not familiar with the CharFilters, I'll look into those now.
Is the solr.LowerCaseFilterFactory not handling wildcards the expected
result or is this a bug?
On Wed, Jun 15, 2011 at 4:34 PM, Mike Sokolov soko...@ifactory.com wrote:
I wonder whether CharFilters are applied to wildcard terms?
I think my answer is here...
On wildcard and fuzzy searches, no text analysis is performed on the
search word.
taken from http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#Analyzers
On Thu, Jun 30, 2011 at 10:23 AM, Jamie Johnson jej2...@gmail.com wrote:
I'm not familiar with the
Hi,
I am using Solr 3.1 and using the SolrJ client. Does anyone know how i can
get the *TOTAL* number of matched documents returned with the QueryResponse?
I am interested in the total documents matched not just the result returned
with the limit applied. Any help will be appreciated.
Thanks.
SolrDocumentList docs = queryResponse.getResults();
long totalMatches = docs.getNumFound();
-Michael
Dear ladies and gentlemen.
Can I ask you to help me with SolrCloud
1) I try to setup a SolrCloud on 2 computers with 3 Zookepers, but it
fails:(
I need to set Zookeper port to 8001, so I change clientPort=8001 in
solr/zoo.cfg.
When I try the command from the example C, to run shard1, it works:
Yes, after posting that response, I read some more and came to the same
conclusion... there seems to be some interest on the dev list in
building a capability to specify an analysis chain for use with wildcard
and related queries, but it doesn't exist now.
-Mike
On 06/30/2011 10:34 AM, Jamie
Hi Romi, after doing the changes, to se the impact you'll have to index some
documents, Solr won't change your index unless you add more documents and
commit them.
It looks like your maxMergeDocs parameter is too small, I would use a grater
value here.
You can see an good explanation on how the
A bug was introduced between Solr 3.1 and 3.2.
With Solr 3.2 we are now getting the follwing error when querying
several pdf and word documents:
SEVERE: org.apache.solr.common.SolrException:
org.apache.lucene.search.highlight.InvalidTokenOffsetsException: Token
17 exceeds length of provided
Jamie - there is a JIRA about this, at least one:
https://issues.apache.org/jira/browse/SOLR-218
Erik
On Jun 15, 2011, at 10:12 , Jamie Johnson wrote:
So simply lower casing the works but can get complex. The query that I'm
executing may have things like ranges which require some
Thanks Michael. Quite helpful.
On Thu, Jun 30, 2011 at 4:06 PM, Michael Ryan mr...@moreover.com wrote:
SolrDocumentList docs = queryResponse.getResults();
long totalMatches = docs.getNumFound();
-Michael
Not that I'm aware of. This is probably something you want to do at the
application layer. If you want to do it in Solr, a good place would be an
UpdateRequestProcessor, but I guess you'll have to implement your own.
On Wed, Jun 29, 2011 at 4:12 PM, Curtis Wilde galv...@gmail.com wrote:
From
Yes, and this too: https://issues.apache.org/jira/browse/SOLR-219
On 06/30/2011 12:46 PM, Erik Hatcher wrote:
Jamie - there is a JIRA about this, at least
one:https://issues.apache.org/jira/browse/SOLR-218
Erik
On Jun 15, 2011, at 10:12 , Jamie Johnson wrote:
So simply lower
Hi everyone,
I'm having some trouble figuring out why a query with an exact word
followed by the * wildcard, eg. teste*, returns no results while a
query for test* returns results that have the word teste in them.
I've created a couple of pasties:
Exact word with wildcard :
It looks like the whole clustering component JAR is not in the classpath. I
remember that I once dealt with a similar issue in Solr 1.4 and the cause
was the relative path of the lib tag being resolved against the core's
instanceDir, which made the path incorrect when directly copying and pasting
I would run that word through the analyzer, I suspect that the word 'teste' is
being stemmed to 'test' in the index, at least that is the first place I would
check.
François
On Jun 30, 2011, at 2:21 PM, Celso Pinto wrote:
Hi everyone,
I'm having some trouble figuring out why a query with
Hi,
I am researching about core administration using Solr. My requirement is to
be able to provision/create/delete indexes dynamically. I have tried it and
it works. Apparently core admin handler will create a new core by
specifying the instance Directory (required), along with data directory,
I am using postgres database and pgpool . Postgres database port : 5432 is
woking fine. But
I am using Pgpool port : is Not Working.
MY importing xml file (*myproduct.xml*)
*Working *
dataSource name=jdbc driver=org.postgresql.Driver
Dear ladies and gentlemen.
Can I ask you to help me with SolrCloud
1) I try to setup a SolrCloud on 2 computers with 3 Zookepers, but it
fails:(
I need to set Zookeper port to 8001, so I change clientPort=8001 in
solr/zoo.cfg.
When I try the command from the example C, to run shard1, it works:
On 6/29/2011 10:16 PM, Shawn Heisey wrote:
I was thinking perhaps I might actually decrease the termIndexInterval
value below the default of 128. I know from reading the Hathi Trust
blog that memory usage for the tii file is much more than the size of
the file would indicate, but if I
I have an idea. I believe I can discover the Properties of an object (C#
reflection) and then code gen schema.xml file based on the field type and
other meta data of that type (possibly from database). After that, I should
be able to ftp the files over to the solr machine. Then I can invoke
Staszek,
That makes sense, but this has always been a multi-core setup, so the paths
have not changed, and the clustering component worked fine for core0. The
only thing new is I have fine tuned core1 (to begin implementing it).
Previously the solrconfig.xml file was very basic. I replaced it
Zarni,
Am 30.06.2011 20:32, schrieb zarni aung:
But I need to know if Solr already handles that case. I wouldn't want to
have to write the tool if Solr already supports creating cores with new
configs on the fly.
there isn't. you have to create the directory structure the related
files
Thank you very much Stefan. This helps.
Zarni
On Thu, Jun 30, 2011 at 4:10 PM, Stefan Matheis
matheis.ste...@googlemail.com wrote:
Zarni,
Am 30.06.2011 20:32, schrieb zarni aung:
But I need to know if Solr already handles that case. I wouldn't want to
have to write the tool if Solr
Perhaps a better question, is this possible?
On Mon, Jun 27, 2011 at 5:15 PM, Jamie Johnson jej2...@gmail.com wrote:
I have a field named content with the following definition
field name=content type=text indexed=true stored=true
multiValued=true termVectors=true termPositions=true
unfortunately the userIdsToScore updates very often. I'd get more Ids almost
every single query (hence why I made the new component). But I see the
problem of not being able to score the whole resultSet. I'd actually need
to do this now that I think about it. I want to get a whole whack of
Sorry for the double post but in this case, is it possible for me to access
the queryResultCache in my component and play with it? Ideally what I want
is this:
1) I have 1 (just a random large number) total results.
2) In my component I access all of these results, score them, and take the
Hello-
I'm looking for a way to find all the links from a set of results. Consider:
doc
id:1
type:X
link:a
link:b
/doc
doc
id:2
type:X
link:a
link:c
/doc
doc
id:3
type:Y
link:a
/doc
Is there a way to search for all the links from stuff of type X -- in
this case (a,b,c)
If I'm
: Lucid Imagination did a webcast on this, as far as I remember?
that was me ... the webcast was a pre-run of my apachecon talk...
http://www.lucidimagination.com/why-lucid/webinars/mastering-power-faceted-search
http://people.apache.org/~hossman/apachecon2010/facets/
...taxonomy stuff comes
Hi All,
How to *uninstall* Solr completely ?
Any help will be appreciated.
Regards,
Gaurav
How'd you install it?
Generally you just delete the directory where you installed it. But you
might be deploying solr.war in a container somewhere besides Solr's example
Jetty setup, in which case you need to undeploy it from those other containers
and remove the remnants.
Curious though...
49 matches
Mail list logo