hl.q works:)
But i have to attach the hl.q to my standard query.
In bigger queries it would by a pain to find out, which terms i need in my hl.q.
My plan: Own query parser in solr, which loops through q, identifies
filter terms(in my case language:de) and append
it as hl.q to the standard query.
Hi,
It depends. If you have a small install with 1-2 servers, the easiest is to
spin up new servers, upgrade features, reindex, and then make the switch in the
frontends.
In very large installs, it's often better to spin up parallell Tomcat instances
on new ports (make sure you got enough
Hi Igor,
i didn't read through the article, but currently I'm not using faceted search.
I just want to ask, for example for all products from the category X name
Samsung
I'll read this article this evening.
Best regards,
Ramo
-Ursprüngliche Nachricht-
Von: Igor MILOVANOVIC
Hi Ramo,
The answer is Yes. You just need to add a specific field category where
you state the category of each item saved and then issue a request like
[text:whatYouWant AND category:smartphone] thus getting all item that
contain whatYouWant and being int he category you pick.
cheers,
gdupont
Research Engineer position
DESCRIPTION
Lucid Imagination, the leading commercial company for Apache Lucene and Solr,
is looking for a research engineer to work on building next generation search,
analytics and machine learning technologies based on Apache Solr, Mahout,
Hadoop and other cutting
Hi Alex,
the dependency tag is used in the Maven project file (pom.xml). If you
are not using Maven to build your project then simply skip that part.
The important thing is that the ICU jar (lucene-icu) and the analysis
extra jar (solr-analysis-extra) are in your classpath.
See also Erick's
Eclipse and IntelliJ have free IDEs, both are good. Personally I prefer
IntelliJ.
Sorry, but I really can't coach you through the whole process from
the very start. I'll be happy to answer some specific questions. SolrJ
is a typical Java application, all the usual rules apply, the only tricky
You can also simply add a filter query as fq=category:smartphones.
Best
Erick
2012/2/1 Gérard Dupont ger.dup...@gmail.com:
Hi Ramo,
The answer is Yes. You just need to add a specific field category where
you state the category of each item saved and then issue a request like
Hello everyone,
I have a question that I imagine has been asked many times before, so I
apologize for the repeat.
I have a basic text field with the following text:
the word ”stemming” in quotes
Uploading the data yields no errors, however when it is indexed, the text looks
like this:
Are you sure the input document is in UTF-8? That looks like classic
ISO-8859-1-treated-as-UTF-8.
How did you confirm the document contains the right quote marks immediately
prior to uploading? If you just visually inspected it, then use whatever
tool you viewed it in to see what the character
Reviving this thread.
You say:
I do wonder...what if (e)dismax had a flag you could set that would tell it
that if any analyzers removed a term, then that term would become optional
for any fields for which it remained? I'm not sure what the development
effort would perhaps it would be a
Hi,
I have tried removing the entry from zookeeper as well as from solr via
admin/cores?action=uload and still the distributed query hits the missing core.
I guess there is no zookeeper watcher in solr to update the core/shard state
used by search.
I got round the problem by doing the above
Hello,
We use Solr to search over a filesystem, so there are a lot of files and
folders indexed, name and path of each file are stored in different fields. The
task is to find folders by name AND containing at least one file of specific
type somewhere inside. For example, we search by phrase
Hi,
I am running apache-solr-3.1.0 and would like to change a field
attribute from stored=false to
stored=true.
I have several hundred cores that have been indexed without storing
the field which is fine
as I only would like to retrieve the value for new data that I plan to
index with the
This should be fine. From my experience, changing a field from stored=false
to stored=true and vice versa is generally safe to do and has no unexpected
behavior.
-Michael
Travis and all,
This is solved and was not directly a Solr issue. I'll note the solution here
in case anyone makes the same mistake. The documents are UTF-8 and the source
documents are converted via XSLT. They look good up to that point.
First off, based off of of some other recommendations
Does anyone have an example of how to use the $deleteDocByQuery and
$deleteDocByID special commands. I am not finding much online. Any help
would be greatly appreciated. Thank you.
Erik
--
View this message in context:
Hi Erik,
Is this what you are looking for?
http://lucene.apache.org/solr/tutorial.html#Deleting+Data
Execute the following command to delete a document
java -Ddata=args -Dcommit=no -jar post.jar
deleteidSP2514N/id/delete
Š
Here is an example of using delete-by-query to delete anything with
Here is an example DIH entity that will delete from solr anything in the
database that is not flagged as 'active'.
entity
name=Deletes
dataSource=ds
query=
SELECT a.id AS '$deleteDocById'
FROM products a
INNER JOIN
Thanks for answering Param, I appreciate it. I am actually looking for a way
to delete records when performing a full import to the index. I am able to
do it when issuing a delta import command by using the deletedPkQuery, but
unfortnately the poor performance of the delta import prohibits me
Hello Erik,
We are running jetty in the background and use curl and the following:
/usr/bin/curl -s http://127.0.0.1:8983/solr/biblio/update?commit=true -H
Content-Type: text/xml --data-binary
'updatedeletequeryebr*/query/delete/update';
/usr/bin/curl -s
Thanks James, I appreciate you taking the time to answer my question. I'll
give this a shot.
Erik
--
View this message in context:
http://lucene.472066.n3.nabble.com/deleteDocByQuery-and-deleteDocByID-tp3707070p3707263.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thanks Alan, going to try this as well. Thanks for taking the time to
answer.
Erik
--
View this message in context:
http://lucene.472066.n3.nabble.com/deleteDocByQuery-and-deleteDocByID-tp3707070p3707266.html
Sent from the Solr - User mailing list archive at Nabble.com.
I have the following requirements :
1. Adds : 20 docs/sec
2. Searches : 100 searches/sec
3. Deletes : (20*3600*24*7 ~ 12 mill ) docs/week ( basically a cron
job which deletes all documents more than 7 days old )
I am thinking of having 6 shards ( with each having 2 million docs )
with 1 master
I think you nailed it, Hoss. What I did is I regenerated the indices and made
sure that they were inline with he schema definitions and it works perfectly
now.
One curious thing is that if there was a mismatch with the schema, why would
a direct query to one of the shards work just fine while the
This topic is either boring or not clear enough...
Another alternative solution would be to add a category field to the
already crawled content.
Any idea how to do it?
Remi
On Tuesday, January 31, 2012, remi tassing tassingr...@gmail.com wrote:
After looking at the Carrot2 introduction, it
Hello Sergey,
if your docs looks like:
PATH:'directory','tree','sements','test1'
FILES:'filename1','ext1','filename2','ext2','filename3','ext3','filename4','ext4'
you can search it:
+PATH:test1 +FILES:jpg
2012/2/1 Sergei Ananko asvs...@gmail.com
Hello,
We use Solr to search over a
2012/2/1 prasenjit mukherjee prasen@gmail.com:
I have the following requirements :
1. Adds : 20 docs/sec
2. Searches : 100 searches/sec
3. Deletes : (20*3600*24*7 ~ 12 mill ) docs/week ( basically a cron
job which deletes all documents more than 7 days old )
I am thinking of having 6
Hi,
I'm trying to get the SolrCloud2 examples to work using a war
deployed solr into glassfish.
The startup properties must be different in this case, because its
having trouble connecting to zookeeper when
I deploy the solr war file.
Perhaps the embedded zookeeper has trouble running in an
Hi all,
In my implementation many fields in documents are having words with
special characters like Company® ,Time™.
Index is created using these fields. However if I make search using
these keywords in solr console, it does not work.
i.e. entering Company® or Time™ in search field box does not
Hello, Mikhail.
Each index record looks like:
DIR:true
PATH:/root/folder1/folder2/
NAME:folder3
SIZE:0
...
This record represents folder /root/folder1/folder2/folder3
DIR:false
PATH:/root/folder1/folder2/folder3/
NAME:image.jpg
SIZE:1234567
...
This is a file
: I m trying to add some weight for popularity in the score returned by solr
: query.
:
http://localhost:10101/solr/syx/select?q={!boost%20b=product(popularity,score)}SIMrows=100fl=score,iddebug=true
:
: I get error undefined field score
that's probably because you have no field named score
Sergey,
Try to employ
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.PatternTokenizerFactory
Regards
On Wed, Feb 1, 2012 at 11:59 PM, asv - gmail asvs...@gmail.com wrote:
**
Hello, Mikhail.
Each index record looks like:
DIR:true
PATH:/root/folder1/folder2/
$doc.getFieldNames() will give you a list of field names as defined in your
schema.xml file. However, the order in which it returns the field names is not
the same order that I defined them in schema.xml. What determines the order
returned by $doc.getFieldNames() ?
: DIR:true
: PATH:/root/folder1/folder2/
: NAME:folder3
: SIZE:0
...
: DIR:false
: PATH:/root/folder1/folder2/folder3/
: NAME:image.jpg
: SIZE:1234567
...
: your solution). Also, in my previous example a file of specified type
: may be deeper than one level: if there are
Thanks for your persistence in tracking this down Mike!
I'm going to start looking into this now...
-Yonik
lucidimagination.com
On Thu, Jan 26, 2012 at 11:06 PM, Mike Hugo m...@piragua.com wrote:
I created issue https://issues.apache.org/jira/browse/SOLR-3062 for this
problem. I was able to
Like the title says, I want to create a page to display a bunch of document
attributes. I accomplished this by creating a new handler and a template for
it. However, I'm having trouble pulling up the details of the document in the
new handler. Here's my code. Is this a good way to do it? I
I've switched my index to use pointtype instead of latlontype of spatial search
queries. Unfortunately I'm seeing much worse performance, and I was wondering
if anybody else knew of any issues between the two types. I would expect a flat
space calculation of pointtype to be better than the
After modifying the schema, I've somehow managed to break the text search
functionality, because the search can't find anything any more.
For example, I defined a field called Entrez ID in my schema.xml file:
field multiValued=false name=Entrez ID type=string index=true
stored=true
I found the answer to my question. The order is determined by the order in
which the fields were defined in the input XML or JSON record for this document.
From: Yuhao nfsvi...@yahoo.com
To: solr-user@lucene.apache.org solr-user@lucene.apache.org
Sent:
For example, I defined a field called Entrez ID in my
schema.xml file:
field multiValued=false name=Entrez ID
type=string index=true stored=true required=true
/
It could be the typo: index=true should be indexed=true
Thanks Yonik!!
The join functionality is proving extremely useful for us in a specific use
case - we're really looking forward to join and other cool features coming
in Solr4!!
Mike
On Wed, Feb 1, 2012 at 3:30 PM, Yonik Seeley yo...@lucidimagination.comwrote:
Thanks for your persistence in
I know I'm kind of reopening a closed thread, but I now have the same
requirement to omitTermFreq only, but still have the ability to run phrase
queries on a field.
Thing is, having a custom Similarity and setting tf=1.0f will turn off term
frequencies globally, which is not what I need; I'd like
Oops, you're right about the typo! However, after I changed it to:
field multiValued=false name=Entrez ID type=string indexed=true
stored=true required=true /
, searching for 335 still returns no result. I did delete the index and
re-index the documents after the change. Interestingly,
Thing is, having a custom Similarity and setting tf=1.0f
will turn off term
frequencies globally, which is not what I need; I'd like to
do it per field.
I think, it is possible to use different similarities for different fields.
https://issues.apache.org/jira/browse/SOLR-2338
Hi Tejinder,
I had this problem yesterday (believe it or not :-)), and the fix for us was to
make Tomcat UTF-8 compliant. In server.xml, there is a Controller tag, we
added the attribute URIEncoding=UTF-8 and restarted Tomcat. Not sure what
container you are using, if its Tomcat this will
In addition to what Emmanuel mentioned, why not consider 7 shards? If
you used one shard/day, your delete problem becomes really easy,
just nuke the oldest shard
Although beware that this solution may affect your TF/IDF calculations
on the new shard (i.e. the one you use for *today's* data)
Well, a lot depends on how you need to query. The simplest
would be to store triplets MSA/Area/Hood, that is denormalize
the data and store each. If you used a multiValued field to store
it, you could do some tricky dancing with positionIncrementGap
to insure that matches occurred as you wish.
Usual question: What did you change last? Because it looks
like you changed your classpath or removed some Jars.
You had to do *something* I think. Possibly removed some
lib/ statements from solrconfig.xml?
Best
Erick
On Wed, Feb 1, 2012 at 1:40 AM, Pranav Prakash pra...@gmail.com wrote:
I really, really, really don't like the fact that you have a space in your
field name. Adding debugQuery=on to your query should show
you the results of parsing the query. What I *expect*, but haven't
tested, is one of two things:
1 the query parser interprets Entrez ID:335 as something like
Sujit's comments are well taken, part of your problem will certainly be
getting the special characters through your container...
But another part of your problem will be having the characters in
your index in the first place. The fact that you can find Time in
the first place suggests that your
: Subject: UTF-8 support during indexing content
: References: 8ce9f966c6f6769-19a0-9e...@webmail-m069.sysops.aol.com
: 1326447127.1952.10.camel@snape
: 8ceade0f7e0ecec-189c-c...@webmail-m069.sysops.aol.com
: 1328105200.2033.33.camel@snape
: In-Reply-To: 1328105200.2033.33.camel@snape
iorixxx wrote
Thing is, having a custom Similarity and setting tf=1.0f
will turn off term
frequencies globally, which is not what I need; I'd like to
do it per field.
I think, it is possible to use different similarities for different
fields.
I'm not following exactly what you're after here in detail, but I think this
will help:
http://www.slideshare.net/erikhatcher/rapid-prototyping-with-solr-5675936
See slides 24 and 25. Note the use of $id in the /document request handler
definition using parameter substitution, a really
Well, sometimes people just copy-paste stuff into the search box probably
because some words (at least in my world) are very hard to spell correctly. We
noticed the problem because the query was getting mangled on its way in and not
returning any search results even though it should have.
Our
I have not yet tried to run SolrCloud in another app server, but it shouldn't
be a problem.
One issue you might have is the fact that we count on hostPort coming from the
system property jetty.port. This is set in the default solr.xml - the hostPort
defaults to jetty.port. You probably want to
Hi, Erick,
Thanks for commenting on this thread, and I think my problem has been
solved. I might start another thread raising technical questions about using
SolrJ.
Thank you again.
Best Regards,
Bing
--
View this message in context:
Appreciate your reply. Have some more follow up questions inline.
On Thu, Feb 2, 2012 at 12:35 AM, Emmanuel Espina
espinaemman...@gmail.com wrote:
1. Adds : 20 docs/sec
2. Searches : 100 searches/sec
3. Deletes : (20*3600*24*7 ~ 12 mill ) docs/week ( basically a cron
job which deletes all
Hi, all,
I am trying to coding Java so that use SolrJ to access Solr, but failed in
the first attempt. I have some experience in Solr, but I am a newbie of
SolrJ. The following are the description of what I set, what I did, and what
I got. I will be grateful if anyone can bring out some
Hi, all,
Following the previous topic, if I abandon my own code and try to build a
project with the original package apache-solr-3.5.0-src, I failed again.
Following are the description of some technical details, and I hope someone
can help to point out my mistakes.
What I Have
Besides the
An offtopic: as some of my questions went unnoticed too, I could recommend
asking them somewhere else in parallel, for example: stackoverflow.com.
But as SOLR and its ecosystem sometimes pose tough questions and
problems, stackoverflow can ignore them as well. Anyhow, just another
opportunity..
High bing,
If you compile your project with solrj jars, you should just make sure they
are explicitly listed in your classpath, not only the directory that
contains them.
You could start with a clean (not maven) java project in NetBeans, create
your class and add the jar dependencies via: right
62 matches
Mail list logo