Hi
I am trying to avoid loading some fields in Lucene's FieldCache.
Is there a way to avoid loading certain failed in Lucene's FieldCache.
One way is to declare them multivalued..
Is there any other way?
./zahoor
I am using Solr 4.2.1
./zahoor
On 20-May-2013, at 11:48 AM, J Mohamed Zahoor zah...@indix.com wrote:
Hi
I am trying to avoid loading some fields in Lucene's FieldCache.
Is there a way to avoid loading certain failed in Lucene's FieldCache.
One way is to declare them multivalued..
Is
On May 20, 2013, at 05:05 , Kamal Palei palei.ka...@gmail.com wrote:
I have put the code to add these fields in document object and index it.
I have not deleted whole indexed data and reindex it. But I expect whatever
new documents are added, for those documents these two fields salary and
Ooops, you didn't say it OK. It is at Timothy's answer.
2013/5/20 Otis Gospodnetic otis.gospodne...@gmail.com
Hm, did I really say that? What was the context? Because I don't see
that in my response below
Otis
--
Search Analytics - http://sematext.com/search-analytics/index.html
OK, I found it, no problem.
2013/5/20 Furkan KAMACI furkankam...@gmail.com
Ooops, you didn't say it OK. It is at Timothy's answer.
2013/5/20 Otis Gospodnetic otis.gospodne...@gmail.com
Hm, did I really say that? What was the context? Because I don't see
that in my response below
From looking at the ReplicationHandler code, it looks like if you hit it
with a 'details' request, it'll show you the details of the most recent
backup, including file count, status and completion time.
Upayavira
On Mon, May 20, 2013, at 08:46 AM, Furkan KAMACI wrote:
Ooops, you didn't say it
Dear All,
I have a requirement to highlight a field only when all keywords entered
match. This also needs to support phrase, operator or wildcard queries.
I'm using Solr 4.0 with edismax because the search needs to be carried out
on multiple fields.
I know with highlighting feature I can
Sandeep,
If you AND all keywords, that should be OK?
Thanks
Jaideep
On Mon, May 20, 2013 at 3:44 PM, Sandeep Mestry sanmes...@gmail.com wrote:
Dear All,
I have a requirement to highlight a field only when all keywords entered
match. This also needs to support phrase, operator or wildcard
Hi,
I am new to Solr. I have a question regarding solrCloud - What is the
difference between solr and solrcloud?
Also please let me know if the complete procedure to setup solr cloud is
mentioned somewhere.
Regards,
Sagar
DISCLAIMER:
On 20 May 2013 16:16, Sagar Chaturvedi
sagar.chaturv...@nectechnologies.in wrote:
Hi,
I am new to Solr. I have a question regarding solrCloud - What is the
difference between solr and solrcloud?
Also please let me know if the complete procedure to setup solr cloud is
mentioned somewhere.
You can start reading from here: http://wiki.apache.org/solr/SolrCloud and
here: http://docs.lucidworks.com/display/solr/SolrCloud
Furkan KAMACI
--
2013/5/20 Sagar Chaturvedi sagar.chaturv...@nectechnologies.in
Hi,
I am new to Solr. I have a question regarding solrCloud - What is the
Hi Jaideep,
The edismax config I have posted mentioned that the default operator is
AND. I am sorry if I was not clear in my previous mail, what I need really
is highlight a field when all search query terms present. The current
highlighter works for *any* of the terms match and not for *all*
If you know all fields that need to be queried, you can rewrite it as -
(assuming, f1, f2 are the fields that you have to search)
(f1:kw1 AND f1:kw2 ... f1:kwn) OR (f2:kw1 AND f2:kw2 ... f2:kwn)
-
Jaideep
On Mon, May 20, 2013 at 4:22 PM, Sandeep Mestry sanmes...@gmail.com wrote:
Hi Jaideep,
Hi all,
Our current project requirement suggests that we should start storing
custom data structures in solr index. The custom data structure would be an
equivalent of C struct.
The task is as follows.
Suppose we have two types of fields, one is FieldName1 and the other
FieldName2.
Suppose
I doubt if that will be the correct approach as it will be hard to generate
the query grammar considering we have support for phrase, operator,
wildcard and group queries.
That's why I have kept it simple and only passing the query text with
minimal parsing (escaping lucene special characters) to
Hi ,
Im uisng solr 3.6.1 version and Im trying to implement delta indexing. Im
using the following configuration in my dataimport handler file
document name=doc
entity name=deltatest dataSource=testsrc pk=id
query=select ID as id, name,LoadDate from NameDetails
If you are saying that you want to change highlighting behaviour, not
query behaviour, then I suspect you are going to have to interact with
the java HighlightComponent. If you can work out how to update that
component to behave as you wish, you could either subclass it, or create
your own
Hi
How do i disable cache (Solr FieldValueCache) for certain queries...
using HTTP it can be done using {!cache=false}...
how can i do it from solrj?
./zahoor
Thanks Upayavira for that valuable suggestion.
I believe overriding highlight component should be the way forward.
Could you tell me if there is any existing example or which methods I
should particularly override?
Thanks,
Sandeep
On 20 May 2013 12:47, Upayavira u...@odoko.co.uk wrote:
If
I have install solr 3.5
I would like to search words(Spanish words) like
- enseñé
- étnico
- castaño
- después
with ascent ñ,é etc.
But solr is not search such words from index.
I have used
-
fieldType name=text_es class=solr.TextField
positionIncrementGap=100
analyzer
(13/05/20 20:53), J Mohamed Zahoor wrote:
Hi
How do i disable cache (Solr FieldValueCache) for certain queries...
using HTTP it can be done using {!cache=false}...
how can i do it from solrj?
./zahoor
How about using facet.method=enum?
koji
--
On 18 May 2013 23:23, jignesh js.vishava...@gmail.com wrote:
I have install solr 3.5
I would like to search words(Spanish words) like
- enseñé
- étnico
- castaño
- después
with ascent ñ,é etc.
But solr is not search such words from index.
[...]
Are you able to set up Solr, and search
Before you dive off the deep end and go crazy with dynamic fields, try a
clean, simple, Solr-oriented static design. Yes, you CAN do an
over-complicated design with dynamic fields, but that doesn't mean you
should.
In a single phrase, denormalize and flatten your design. Sure, that will
lead
Tomcat is notorious for not defaulting to UTF-8 encoding for URLs which is
how the query is passed, which is needed to preserve all these accented
characters.
In Tomcat's server.xml, it should have something like:
Connector port=8080
protocol=HTTP/1.1
Hi All,
I want to override a component from solr-core and for that I need solr-core
jar.
I am using the solr.war that comes from Apache mirror and if I open the
war, I see the solr-core jar is actually named as apache-solr-core.jar.
This is also true about solrj jar.
If I now provide a
Because the same field is split amongst a number of segments. If you
look in the index directory, you should see files like _3fgm.* and
_3ffm.*. Each such group represents one segment. The number of
segments changes with merging etc.
Best
Erick
On Mon, May 20, 2013 at 6:43 AM, J Mohamed Zahoor
I can't give you much advice on the topic. I have reviewed the
HighlightComponent and found it complex and hard to follow, so expect it
to be challenging.
Upayavira
On Mon, May 20, 2013, at 01:28 PM, Sandeep Mestry wrote:
Thanks Upayavira for that valuable suggestion.
I believe overriding
hello,
we are tracking down some performance issues with our DIH process.
not sure if this is related - but i am seeing tons of the messages below in
the logs during re-indexing of the core.
what do these messages mean?
2013-05-18 19:37:30,623 INFO [org.apache.solr.update.UpdateHandler]
On 5/20/2013 8:01 AM, Sandeep Mestry wrote:
And I do remember the discussion on the forum about dropping the name
*apache* from solr jars. If that's what caused this issue, then can you
tell me if the mirrors need updating with solr-core.jar instead of
apache-solr-core.jar?
If it's named
The latter, the schemas must be similar enough to satisfy the query
Best
Erick
On Thu, May 16, 2013 at 5:03 AM, Marcin mar...@workdigital.co.uk wrote:
Hi there,
I am trying to figure out what SOLR means by compatible collection in order
to be able to run the following query:
Query all
On 5/20/2013 5:53 AM, J Mohamed Zahoor wrote:
How do i disable cache (Solr FieldValueCache) for certain queries...
using HTTP it can be done using {!cache=false}...
If you are doing facets, Koji's reply works for those.
The localparam for caching should work just fine if you prepend it to
Whether you add it as a dynamic field or normal field really doesn't
matter from a Solr perspective. Dynamic fields are exactly like normal
fields, you just don't have to fully specify the name. That said, I
prefer normal fields to prevent typo's from messing me up. If you had
a dynamic field like
This is really just parsing the XML using any of several parsers and
putting the results into a SolrInputDocument (assuming a SolrJ
client).
Alternatively, you could perhaps do some XSLT transformations, but I'm
not great on the ins and outs of XSLT...
Best
Erick
On Sun, May 19, 2013 at 11:03
Rishi:
Thanks very much for taking the time to post this, we're always
looking for before/after numbers!
Erick
On Sat, May 18, 2013 at 11:34 PM, Shalin Shekhar Mangar
shalinman...@gmail.com wrote:
Awesome news Rishi! Looking forward to your SolrCloud updates.
On Sat, May 18, 2013 at 12:59
I didn't change it and haven't seen any issues.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Zookeeper-Ensemble-Startup-Parameters-For-SolrCloud-tp4063905p4064654.html
Sent from the Solr - User mailing list archive at Nabble.com.
Just done an upgrade from Solr (cloud) 4.0 to 4.3 and noticed that
clusterstate.json now contains the IP address instead of the hostname for
each shard descriptor.
Was this a conscious change? It caused us some pain when migrating and
breaks our own admin tools, so just checking if this is
What's supposed to happen (not guaranteeing it is completely correct,
mind you) is that the presence of a cores tag defines which checks
are performed. Errors are thrown on old-style constructs when no
cores tag is present and vice-versa.
Best
Erick
On Sun, May 19, 2013 at 7:20 PM, Benson
Lyuba:
Could you go ahead and raise a JIRA and assign it to me to
investigate? You should definitely be able to define cores this way.
Thanks,
Erick
On Sun, May 19, 2013 at 9:27 AM, Lyuba Romanchuk
lyuba.romanc...@gmail.com wrote:
Hi,
It seems like in order to query transient cores they must
We've run into this problem when deploying index jobs that run in Elastic
Mapreduce. We've gotten by with an older version of SolrJ, but some of the
fixes and enhancements with SolrCloud that came out in the 4.x series
aren't available if you go back to an earlier version.
In particular, we're
What is the number at the end?
is it the no of unique elements in each segment?
./zahoor
On 20-May-2013, at 7:37 PM, Erick Erickson erickerick...@gmail.com wrote:
Because the same field is split amongst a number of segments. If you
look in the index directory, you should see files like
Thanks for your reply
I am using jetty for solr search
--
View this message in context:
http://lucene.472066.n3.nabble.com/Not-able-to-search-Spanish-word-with-ascent-in-solr-tp4064404p4064652.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thanks for the reply..
I am send below type of xml to solr
?xml version=1.0 encoding=UTF-8?adddoc
field name=id15/field
field name=id_i15/field
field name=nameMis nuevos colgantes de PRIMARK/field
field name=featuresamp;iquest;Alguna vez os habamp;eacute;is pasado
por la zona de
Every time a commit is done a new searcher is opened. In the solr config file
caches are defined with a parameter called autowarm. Autowarm basically tries
to copy the cache values from previous searcher into the current one. If you
are doing a bulk update and do not care for searching till
i have logging working for the most part with logback 1.0.13 and slf4j
1.7.5 under solr 4.3.0 (or previously under solr 4.2.1)
with two exceptions, i'm very happy with the setup as i can get all the
jetty request logs, and various solr service events logged out with
rotation, etc
BUT i havent
On 5/20/2013 9:25 AM, Daniel Collins wrote:
Just done an upgrade from Solr (cloud) 4.0 to 4.3 and noticed that
clusterstate.json now contains the IP address instead of the hostname for
each shard descriptor.
Was this a conscious change? It caused us some pain when migrating and
breaks our own
Try the Solr Admin UI Analysis page - enter text for both index and query
for your field and see whether the final terms still have their accents.
-- Jack Krupansky
-Original Message-
From: jignesh
Sent: Monday, May 20, 2013 10:46 AM
To: solr-user@lucene.apache.org
Subject: Re: Not
I'm providing a search feature in a web app that searches for documents
that range in size from 1KB to 200MB of varying MIME types (PDF, DOC,
etc). Currently there are about 3000 documents and this will continue to
grow. I'm providing full word search and partial word search. For each
document,
you mean i would add this switch to my script that kicks of the dataimport?
exmaple:
OUTPUT=$(curl -v
http://${SERVER}.intra.searshc.com:${PORT}/solrpartscat/${CORE}/dataimport
-F command=full-import -F clean=${CLEAN} -F commit=${COMMIT} -F
optimize=${OPTIMIZE} -F openSearcher=false)
what
And use the /terms request handler to view what is present in the field:
/solr/terms?terms.fl=text_esterms.prefix=a
You're looking to ensure the index does, in fact, have the accented characters
present. It's just a sanity check, but could possibly save you a little
(sanity, that is).
Jason
: Technically, core Solr does not require a unique key. A lot of features in
nohting in this thread refered to the uniqueKey field, or the lack of a
uniqueKey field in the users schema, at all until you brought it up.
* the user has a field named key
* the user had a question about deleting
On 5/20/2013 10:44 AM, Boogie Shafer wrote:
BUT i havent figured out what i need to do to get the logging events to
display in the SOLR admin ui
e.g. at http://solr-hostname:8983/solr/#/~logging
The logging page in the UI is populated by log watcher classes specific
to the logging
Not sure, never had to worry about what they are..
On Mon, May 20, 2013 at 12:28 PM, J Mohamed Zahoor zah...@indix.com wrote:
What is the number at the end?
is it the no of unique elements in each segment?
./zahoor
On 20-May-2013, at 7:37 PM, Erick Erickson erickerick...@gmail.com
Hello
Here is the output of Solr Admin UI Analysis page
http://awesomescreenshot.com/0ff1ao7347
What should I conclude from this?
Thanks,
Waiting for reply.
--
View this message in context:
Here is the output using
/solr/terms?terms.fl=nameterms.prefix=a
--
responselst name=responseHeaderint name=status0/intint
name=QTime1/int/lstlst name=termslst name=nameint
name=a716/intint name=alt509/intint name=aacute384/intint
name=as260/intint name=amp176/intint name=al95/intint
Most definitely not the number of unique elements in each segment. My 32
document sample index (built from the default example docs data) has the
following:
entry#0:
'StandardDirectoryReader(segments_b:29 _8(4.2.1):C32)'='manu_exact',class
We can conclude that the field type analyzer is NOT the problem. Good
experiment to eliminate one culprit.
-- Jack Krupansky
-Original Message-
From: jignesh
Sent: Monday, May 20, 2013 1:21 PM
To: solr-user@lucene.apache.org
Subject: Re: Not able to search Spanish word with ascent in
My guess is that the problem is those 200M documents.
FastVectorHighlighter is fast at deciding whether a match, especially a
phrase, appears in a document, but it still starts out by walking the
entire list of term vectors, and ends by breaking the document into
candidate-snippet fragments, both
We can conclude that the accents did not get indexed and we know from the
other experiment that the field type analyzer is not at fault.
1. How are you indexing the data? Verify what character encoding it is
using.
2. Try manually indexing some accented data, like with a curl command, and
see
Hi Rishi,
Have you done any tests with Solr 4.3 ?
Regards,
Cordialement,
BOUHLEL Noureddine
On 17 May 2013 21:29, Rishi Easwaran rishi.easwa...@aol.com wrote:
Hi All,
Its Friday 3:00pm, warm sunny outside and it was a good week. Figured
I'd share some good news.
I work for AOL
On 5/20/2013 11:24 AM, jignesh wrote:
responselst name=responseHeaderint name=status0/intint
name=QTime1/int/lstlst name=termslst name=nameint
name=a716/intint name=alt509/intint name=aacute384/intint
name=as260/intint name=amp176/intint name=al95/intint
name=azul67/intint name=ahumado61/intint
Okay. I should have realized from the original email. The input is
XML-encoded HTML. That's fine for a stored field that will be retrieved and
then displayed in a browser, but is NOT searchable.
What you will have to do is maintain two copies of that data, one stored in
HTML (the one your
I suppose you saw my JIRA suggesting that solr.xml should might have
the same repetoire of 'lib' elements as solrconfig.xml, instead of
just a single 'str'.
On Mon, May 20, 2013 at 11:16 AM, Erick Erickson
erickerick...@gmail.com wrote:
What's supposed to happen (not guaranteeing it is
Sure Shalin, hopefully soon.
-Original Message-
From: Shalin Shekhar Mangar shalinman...@gmail.com
To: solr-user solr-user@lucene.apache.org
Sent: Sat, May 18, 2013 11:35 pm
Subject: Re: Upgrading from SOLR 3.5 to 4.2.1 Results.
Awesome news Rishi! Looking forward to your
Hi,
We have an existing Java based enterprise application that is bundled as a WAR
file and runs on Tomcat and uses Spring 3.0.5, Hibernate 3.6.2, and Lucene
3.0.3. We are using annotations in Hibernate that nicely couple it Lucene to
index objects (documents, images, PDFs, etc.) based on key
We use commodity H/W which we procured over the years as our complex grew.
Running on jdk6 with tomcat 5. (Planning to upgrade to jdk7 and tomcat7 soon).
We run them with about 4GB heap. Using CMS GC.
-Original Message-
From: adityab aditya_ba...@yahoo.com
To: solr-user
I am in the process of setting up a core using Solr 4.3. On the Core
Discoveryhttp://wiki.apache.org/solr/Core%20Discovery%20(4.3%20and%20beyond)
wiki
page it states:
As of SOLR-4196, there's a new way of defining cores. Essentially, it is no
longer necessary to define cores in solr.xml. In
On 5/20/2013 1:02 PM, Todd Hunt wrote:
snip
It seems like Solr forces one to expose access to its Cores (indexes) via its
own WAR file. I don't want that. I just want to be able to utilize the Solr Java API to
integrate with our current web services and Hibernate framework to index text
On 5/20/2013 1:28 PM, Shane Perry wrote:
Using the 4.3 war available for download, I attempted to set up my core
using the solr.properties file (in anticipation of moving to 5.0). When I
start the context, logging shows that the process is falling back to the
default solr.xml file (essentially
No, we just upgraded to 4.2.1.
With the size of our complex and effort required apply our patches and rollout,
our upgrades are not that often.
-Original Message-
From: Noureddine Bouhlel nouredd...@ecotour.com
To: solr-user solr-user@lucene.apache.org
Sent: Mon, May 20, 2013 3:36
Hello,
I am submitting rich documents to a SOLR index via Solr Cell. This is all
working well.
The documents are organized in meaningful folders. I would like to capture
the folder names in my index so that I can use the folder names to provide
facets.
I can pass the path data into the
Hello,
I'm having a little problem with a test SolrCloud cluster.
I've set up 3 nodes (SolrCores) to use an external Zookeeper. I use 1 shard
and the other 2 SolrCores are being auto-asigned as replicas.
Let's say I have these 3 nodes: the leader shard A, the replica shard B,
and the (other)
On Mon, May 20, 2013 at 4:21 PM, Sebastián Ramírez
sebastian.rami...@senseta.com wrote:
When I send an update to a non-leader (replica) shard (B), the updated
results are reflected in the leader shard (A) and in the other replica
shard (C), but not in the shard that received the update (B).
I store documents having some meta data that is composed out of multiple
values. Usually an id with a label. A simple example would be the name of a
city and the unique id of that city. The id is needed, because different cities
can have the same name like Berlin in Germany and Berlin in the
Tell us a little more, with examples, of how you really want to search and
facet this information.
One technique is to store the same information in multiple ways, for
different uses, combining the name in different ways, such as Berlin,
Berlin:DE, Berlin, NJ, Berlin:Germany, Berlin GERMANY,
Yes, It's happening with the latest version, 4.2.1
Yes, it's easy to reproduce.
It happened using 3 Virtual Machines and also happened using 3 physical
nodes.
Here are the details:
I installed Hortonworks (a Hadoop distribution) in the 3 nodes. That
installs Zookeeper.
I used the example
Sorry, I think my reference to restriction by country was more confusing than
helpful. Let's say, that the author of the document is one dimension I would
like to use facets for. author would be one field in my document schema. Now
let's take Schmidt, M. as author name, which is quite common in
geeky2 wrote
you mean i would add this switch to my script that kicks of the
dataimport?
exmaple:
OUTPUT=$(curl -v
http://${SERVER}.intra.searshc.com:${PORT}/solrpartscat/${CORE}/dataimport
-F command=full-import -F clean=${CLEAN} -F commit=${COMMIT} -F
optimize=${OPTIMIZE} -F
Hi Cord,
I think you'd do it like this:
1. Add this to schema.xml
!--
Example of using PathHierarchyTokenizerFactory at index time, so
queries for paths match documents at that path, or in descendent paths
--
fieldType name=descendent_path class=solr.TextField
Thank you Brendan,
I had started to read about the tokenizers and couldn't quite piece
together how it would work. I will read about this and post my
implementation if successful.
Cord
On Mon, May 20, 2013 at 4:13 PM, Brendan Grainger
brendan.grain...@gmail.com wrote:
Hi Cord,
I think
: We have master/slave setup. We disabled autocommits/autosoftcommits. So the
: slave only replicates from master and serve query. Master does all the
: indexing and commit every 5 minutes. Slave polls master every 2.5 minutes
: and does replication.
Details matter...
Are you using hte exact
: Hi everybody, I would like to have distinct httpCaching configuration for
: distinct handlers, i.e if a request comes for select, send a cache control
: header of 1 minute ; and if receive a request for mlt then send a cache
: control header of 5 minutes.
: Is there a way to do that in my
Is there a way to query for combinations of two sets of words? For
example, if I had
(java or groovy or scala)
(programming or coding or development)
Is there a query parser that, at query time, would expand that into
combinations like
java programming
groovy programming
scala programming
java
On 21 May 2013 09:12, Mike Hugo m...@piragua.com wrote:
Is there a way to query for combinations of two sets of words? For
example, if I had
(java or groovy or scala)
(programming or coding or development)
Is there a query parser that, at query time, would expand that into
combinations
Yes, with the Solr surround query parser:
q=(java OR groovy OR scala) W (programming OR coding OR development)
BUT... there is the caveat that the surround query parser does no analysis.
So, maybe you need Java OR java etc. Or, if you know that the index is
lower case.
Try this dataset:
Fantastic! Thanks!
On Mon, May 20, 2013 at 11:21 PM, Jack Krupansky j...@basetechnology.comwrote:
Yes, with the Solr surround query parser:
q=(java OR groovy OR scala) W (programming OR coding OR development)
BUT... there is the caveat that the surround query parser does no
analysis. So,
Hello all,
Currently, I used solr for products searching. I used Java web
platform.
fieldType name=text_general class=solr.TextField
positionIncrementGap=100
analyzer type=index
tokenizer
thanks for the pointer on the missing logwatcher for logback...i'll take a
look at that.
on the jetty logging side of things i get nearly all the jetty logging but
the initial startup logs which seem to happen prior to the other logging
jars getting loaded. perhaps i need to add a few more
Hi All,
I am hitting an OOM error while trying to do an hard commit on one of
the cores.
Transaction log dir is Empty and DIH shows indexing going on for 13 hrs..
*Indexing since 13h 22m 22s*
Requests: 5,211,392 (108/s), Fetched: 1,902,792 (40/s), Skipped: 106,853,
Processed: 1,016,696
88 matches
Mail list logo