problem in spellchecker working with dismax requesthandler
Hi, In my solr search previously i was using standard request handler than spellchecker was working fine. now i configured search to dismax request handler. But now spellchecker is not working. would you please tell me what can be the problem Thanks - Thanks Regards Romi -- View this message in context: http://lucene.472066.n3.nabble.com/problem-in-spellchecker-working-with-dismax-requesthandler-tp3143662p3143662.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: @field for child object
no On Mon, Jul 4, 2011 at 3:34 PM, Kiwi de coder kiwio...@gmail.com wrote: hi, i wondering solrj @Field annotation support embedded child object ? e.g. class A { @field string somefield; @emebedded B b; } regards, kiwi -- - Noble Paul
Re: upgrading to Tika 0.9 on Solr 1.4.1
I have upgraded my Solr Distribution to 3.2 and also the referring jars of my application (especially the solr jar was 1.4.1 in my application which calls solr...hence causing javabin exception...) . Also updated the pdfbox/jempbox/fontbox to latest versions and Tika to 0.9 version...which made things up for me! -- Surendranadh
Re: problem in spellchecker working with dismax requesthandler
On Wed, Jul 6, 2011 at 11:36 AM, Romi romijain3...@gmail.com wrote: Hi, In my solr search previously i was using standard request handler than spellchecker was working fine. now i configured search to dismax request handler. But now spellchecker is not working. would you please tell me what can be the problem You may need to add the SpellCheckComponent in the last-components section in the new request handler that you are using. See http://wiki.apache.org/solr/SpellCheckComponent#Configuration -- Regards, Shalin Shekhar Mangar.
Re: Nightly builds
Hi Benson On 07/05/2011 04:29 PM, Benson Margulies wrote: The reason for the email is not that I can't find them, but because the project, I claim, should be advertising them more prominently on the web site than buried in a wiki. Actually they are linked on the hp but unfortunately not directly: http://lucene.apache.org/solr/#21+February+2006%3A+nightly+builds Where I come from, an lmgtfy link is rather hostile. Sorry, I didn't mean to offend you. I obviously misunderstood your question. Oh, and you might want to fix the spelling of 'Author' in your own signature. Thanks, for that! On Tue, Jul 5, 2011 at 10:19 AM, Tom Grossitconse...@gmail.com wrote: On 07/05/2011 04:08 PM, Benson Margulies wrote: The solr download link does not point to or mention nightly builds. Are they out there? http://lmgtfy.com/?q=%2Bsolr+%2Bnightlybuildsl=1 -- Auther of the book Plone 3 Multimedia - http://amzn.to/dtrp0C Tom Gross email.@toms-projekte.de skype.tom_gross web.http://toms-projekte.de blog...http://blog.toms-projekte.de -- Author of the book Plone 3 Multimedia - http://amzn.to/dtrp0C Tom Gross email.@toms-projekte.de skype.tom_gross web.http://toms-projekte.de blog...http://blog.toms-projekte.de
Re:OOM at solr master node while updating document
Seems FieldCache consume all the heap. At 2011-07-05 16:25:27,Chengyang atreey...@163.com wrote: Is there any memory leak when I updating the index at the master node? Here is the stack trace. o.a.solr.servlet.SolrDispatchFilter - java.lang.OutOfMemoryError: Java heap space at org.apache.solr.handler.ReplicationHandler$FileStream.write(ReplicationHandler.java:1000) at org.apache.solr.handler.ReplicationHandler$3.write(ReplicationHandler.java:887) at org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter.java:322) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.jboss.web.tomcat.filters.ReplyHeaderFilter.doFilter(ReplyHeaderFilter.java:96) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:230) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:175) at org.jboss.web.tomcat.security.SecurityAssociationValve.invoke(SecurityAssociationValve.java:179) at org.jboss.web.tomcat.security.JaccContextValve.invoke(JaccContextValve.java:84) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.jboss.web.tomcat.service.jca.CachedConnectionValve.invoke(CachedConnectionValve.java:157) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:262) at org.apache.coyote.ajp.AjpAprProcessor.process(AjpAprProcessor.java:425) at org.apache.coyote.ajp.AjpAprProtocol$AjpConnectionHandler.process(AjpAprProtocol.java:378) at org.apache.tomcat.util.net.AprEndpoint$Worker.run(AprEndpoint.java:1508) at java.lang.Thread.run(Thread.java:619)
Re: faceting on field with two values
Thanks for your advice and for your comments. In fact, we don't use facets to offer a facet UI to user, but to analyze user request, then send a second request to Solr. Lot of requests have lot of answers (often more then a thousand), so we need to filter user request with fq parameter, if possible. Best, Elisabeth 2011/7/5 Chris Hostetter hossman_luc...@fucit.org : I have two fields TOWN and POSTALCODE and I want to concat those two in one : field to do faceting As others have pointed out, copy field doesn't do a concat, it just adds the field values from the source field to the desc field (so with those two copyField/ lines you will typically get two values for each doc in the dest field) if you don't wnat to go the DIH route, and you don't want to change your talend process, you could use a simple UpdateProcessor for this (update processors are used to process add/delete requests no matter what source the come from, before analysis happens) ... but i don't think we have any off the shelf Concat update processors in solr at the moment there is a patch for a a Script based on which might be helpful.. https://issues.apache.org/jira/browse/SOLR-1725 All of that said, based on what you've described about your usecase i would question from a UI standpoint wether this field would actually a good idea... isn't there an extremely large number of postal codes even in a single city? why not let people fact on just the town field first, and then only when they click on one, offer them a facet on Postal code? Otherwise your facet UI is going to have a tendenzy to look like this... Gender: * Male (9000 results) * Female (8000 results) Town/Postal: * paris, 75016 (560 results) * paris, 75015 (490 results) * paris, 75022 (487 results) * boulogne sur mer 62200 (468 results) * paris, 75018 (465 results) * (click to see more) Color: * Red (900 results) * Blue (800 results) ...and many of your users will never find the town they are looking for (let alone the post code) -Hoss
Where does the integration happen inside Solr? (Solr-UIMA)
Hi All I am trying to port and application developed using Lucene based Indexer + UIMA, in to SOlr+UIMA, for the past 3,4 days. I am not able to comprehend where exactly does the UIMA processing get added. Does it get added after I say something like: UpdateResponse response = server.add(a list of docs)...? or do I need to do something else before that? in the pre-UIMA Solr usage, I just took a doc, extracted fields I need...and added them to the SolrInputDocument object. Once in some 2K documents or so, I committed to server. But, now, I am not able to figure out where UIMA comes. I assumed that it comes during the update response. Please correct me if I am wrong. Sowmya. -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
solr.StandardTokenizerFactory: more info needed
Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.StandardTokenizerFactory is very short.. -- Regards, Dmitry Kan
Highlight whole query phrase
i am using solr for search and i implemented highlighting feature for my search results. when i my search string is ring it highlight ring but when search string is gold ring than also it highlight only gold, where i wanted to highlight whole gold ring for highlighting i use description field which i got as * highlighting ={ 8252: { text: [ and emgold/ememRing/em design was finely crafted in Japan. ] }, 8142: { text: [ This emelegant/em emRing/em has an Akoya cultured pearl with a band of bezel-set round diamonds making ] } };* Now i am parsing it as * $.each(newresult.response.docs, function(i,item){ var word = highlight[item[UID_PK]]; var result=; var j=0; for (j=0 ;j=item.text.length;j++) { result = result+item.text[j]+br; } result=result.replace(word,'em' + word + '/em'); }); * Now how should i parse so that i got gold ring highlighted - Thanks Regards Romi -- View this message in context: http://lucene.472066.n3.nabble.com/Highlight-whole-query-phrase-tp3143797p3143797.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Exception when using result grouping and sorting by geodist() with Solr 3.3
My query in the unit test looks like this: q=*:*fq=_query_:{!geofilt sfield=user.location_p pt=48.20927,16.3728 d=50.0} OR _query_:{!geofilt sfield=location_p pt=48.20927,16.3728 d=50.0}group=truegroup.main=truegroup.format=groupedgroup.field=user.uniqueId_ssfield=user.location_ppt=48.20927,16.3728sort=geodist() asc This works without a problem in my trunk build of Solr 4.0 from March 2011. I use the standard schema.xml packaged with the Solr distribution. Thomas On Tue, Jul 5, 2011 at 10:20 AM, Thomas Heigl tho...@umschalt.com wrote: I'm pretty sure my original query contained a distance filter as well. Do I absolutely need to filter by distance in order to sort my results by it? I'll write another unit test including a distance filter as soon as I get a chance. Cheers, Thomas On Tue, Jul 5, 2011 at 9:04 AM, Bill Bell billnb...@gmail.com wrote: Did you add: fq={!geofilt} ?? On 7/3/11 11:14 AM, Thomas Heigl tho...@umschalt.com wrote: Hello, I just tried up(down?)grading our current Solr 4.0 trunk setup to Solr 3.3.0 as result grouping was the only reason for us to stay with the trunk. Everything worked like a charm except for one of our queries, where we group results by the owning user and sort by distance. A simplified example for my query (that still fails) looks like this: q=*:*group=truegroup.field=user.uniqueId_sgroup.main=truegroup.format= groupedsfield=user.location_ppt=48.20927,16.3728sort=geodist() asc The exception thrown is: Caused by: org.apache.solr.common.SolrException: Unweighted use of sort geodist(latlon(user.location_p),48.20927,16.3728) at org.apache.solr.search.function.ValueSource$1.newComparator(ValueSource.j ava:106) at org.apache.lucene.search.SortField.getComparator(SortField.java:413) at org.apache.lucene.search.grouping.AbstractFirstPassGroupingCollector.ini t(AbstractFirstPassGroupingCollector.java:81) at org.apache.lucene.search.grouping.TermFirstPassGroupingCollector.init(T ermFirstPassGroupingCollector.java:56) at org.apache.solr.search.Grouping$CommandField.createFirstPassCollector(Gro uping.java:587) at org.apache.solr.search.Grouping.execute(Grouping.java:256) at org.apache.solr.handler.component.QueryComponent.process(QueryComponent.j ava:237) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchH andler.java:194) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBa se.java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1368) at org.apache.solr.client.solrj.embedded.EmbeddedSolrServer.request(Embedded SolrServer.java:140) ... 39 more Any ideas how to fix this or work around this error for now? I'd really like to move from the trunk to the stable 3.3.0 release and this is the only problem currently keeping me from doing so. Cheers, Thomas
Re: Can I invert the inverted index?
From you patch I see TermFreqVector which provides the information I want. I also found FieldInvertState.getLength() which seems to be exactly what I want. I'm after the word count (sum of tf for every term in the doc). I'm just not sure whether FieldInvertState.getLength() returns just the number of terms (not multiplied by the frequency of each term - word count) or not though. It seems as if it returns word count, but I've not tested it sufficienctly. On Wed, Jul 6, 2011 at 1:39 AM, Trey Grainger the.apache.t...@gmail.comwrote: Gabriele, I created a patch that does this about a year ago. See https://issues.apache.org/jira/browse/SOLR-1837. It was written for Solr 1.4 and is based upon the Document Reconstructor in Luke. The patch adds a link to the main solr admin page to a docinspector page which will reconstruct the document given a uniqueid (required). Keep in mind that you're only looking at what's in the index for non-stored fields, not the original text. If you have any issues using this on the most recent release, let me know and I'd be happy to create a new patch for solr 3.3. One of these days I'll remove the JSP dependency and this may eventually making it into trunk. Thanks, -Trey Grainger Search Technology Development Team Lead, Careerbuilder.com Site Architect, Celiaccess.com On Tue, Jul 5, 2011 at 3:59 PM, Gabriele Kahlout gabri...@mysimpatico.comwrote: Hello, With an inverted index the term is the key, and the documents are the values. Is it still however possible that given a document id I get the terms indexed for that document? -- Regards, K. Gabriele --- unchanged since 20/9/10 --- P.S. If the subject contains [LON] or the addressee acknowledges the receipt within 48 hours then I don't resend the email. subject(this) ∈ L(LON*) ∨ ∃x. (x ∈ MyInbox ∧ Acknowledges(x, this) ∧ time(x) Now + 48h) ⇒ ¬resend(I, this). If an email is sent by a sender that is not a trusted contact or the email does not contain a valid code then the email is not received. A valid code starts with a hyphen and ends with X. ∀x. x ∈ MyInbox ⇒ from(x) ∈ MySafeSenderList ∨ (∃y. y ∈ subject(x) ∧ y ∈ L(-[a-z]+[0-9]X)). -- Regards, K. Gabriele --- unchanged since 20/9/10 --- P.S. If the subject contains [LON] or the addressee acknowledges the receipt within 48 hours then I don't resend the email. subject(this) ∈ L(LON*) ∨ ∃x. (x ∈ MyInbox ∧ Acknowledges(x, this) ∧ time(x) Now + 48h) ⇒ ¬resend(I, this). If an email is sent by a sender that is not a trusted contact or the email does not contain a valid code then the email is not received. A valid code starts with a hyphen and ends with X. ∀x. x ∈ MyInbox ⇒ from(x) ∈ MySafeSenderList ∨ (∃y. y ∈ subject(x) ∧ y ∈ L(-[a-z]+[0-9]X)).
Re: Highlight whole query phrase
On Wed, Jul 6, 2011 at 09:27, Romi romijain3...@gmail.com wrote: Now how should i parse so that i got gold ring highlighted Take a look at the hl.mergeContiguous parameter. http://wiki.apache.org/solr/HighlightingParameters#hl.mergeContiguous
Not getting data using concat in DIH
Hi all, My data-import.xml looks like this: dataConfig dataSource driver=com.ibm.db2.jcc.DB2Driver url=jdbc:db2://10.150.200.69:6/livedb user=property password=property / document name=articlePosting entity name=articleDetails query=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' deltaQuery=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' and (CREATEDATE='${dataimporter.last_index_time}' or LASTMODIDATE='${dataimporter.last_index_time}') field column=MBFARTUNIQUEKEY name=ukey / field column=MBFARTRFNUM name=id / field column=MBFARTUBIRFNUM name=oid / field column=MBFARTCNDCATE name=cg / field column=MBFARTCNDCITY name=ct / field column=MBFARTLMTRFNUM name=lt / field column=MBFARTDESC name=dtldesc / field column=CREATEDATE name=ppd / field column=CREATETIME name=ppt / field column=MODIDATE name=ecd / field column=MODITIME name=tmsrt / field column=MBFARTVIEWED name=viewcount / field column=MBFARTSNIPPET name=ansdesc / field column=POPULARITYSORTCODE name=popularsort / field column=RECENCYSORT name=recencysort / field column=EXPERTADVICE name=expadvice / field column=EXPERTQUES name=expertQues / field column=EXPERTANS name=expertAns / field column=MBFARTTOPIC name=topic / field column=CONTENTTYPE name=contentType / field column=ADTEXT name=ad_text / entity name=setOwnerName query=select UBILOGIN from TPUBI where UBIRFNUM=${articleDetails.MBFARTUBIRFNUM} field column=UBILOGIN name=oname / /entity /entity /document /dataConfig This works fine. But if I comment the line field column=ADTEXT name=ad_text / and uncomment the following It stops working. Does concat not work for fields?? Please help. -- View this message in context: http://lucene.472066.n3.nabble.com/Not-getting-data-using-concat-in-DIH-tp3144006p3144006.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Not getting data using concat in DIH
The XML comments were not taken. This is the data-import.xml that does not work. In bold is the line uncommented. dataConfig dataSource driver=com.ibm.db2.jcc.DB2Driver url=jdbc:db2://10.150.200.69:6/livedb user=property password=property / document name=articlePosting entity name=articleDetails query=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' deltaQuery=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' and (CREATEDATE='${dataimporter.last_index_time}' or LASTMODIDATE='${dataimporter.last_index_time}') field column=MBFARTUNIQUEKEY name=ukey / field column=MBFARTRFNUM name=id / field column=MBFARTUBIRFNUM name=oid / field column=MBFARTCNDCATE name=cg / field column=MBFARTCNDCITY name=ct / field column=MBFARTLMTRFNUM name=lt / field column=MBFARTDESC name=dtldesc / field column=CREATEDATE name=ppd / field column=CREATETIME name=ppt / field column=MODIDATE name=ecd / field column=MODITIME name=tmsrt / field column=MBFARTVIEWED name=viewcount / field column=MBFARTSNIPPET name=ansdesc / field column=POPULARITYSORTCODE name=popularsort / field column=RECENCYSORT name=recencysort / field column=EXPERTADVICE name=expadvice / field column=EXPERTQUES name=expertQues / field column=EXPERTANS name=expertAns / field column=MBFARTTOPIC name=topic / field column=CONTENTTYPE name=contentType / field column=ADTEXT name=ad_text / entity name=setOwnerName query=select UBILOGIN from TPUBI where UBIRFNUM=${articleDetails.MBFARTUBIRFNUM} field column=UBILOGIN name=oname / /entity *entity name=setCityName query=select ${articleDetails.ADTEXT} concat ' ' concat CITYNAME as AD_TEXT from TPCITY where CITY_CNDCITY=${articleDetails.MBFARTCNDCITY} fetch first row only field column=AD_TEXT name=ad_text / /entity * /entity /document /dataConfig -- View this message in context: http://lucene.472066.n3.nabble.com/Not-getting-data-using-concat-in-DIH-tp3144006p3144014.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Highlight whole query phrase
I included hl.mergeContiguous parameter. in my url but no effect on search result highlighting. actually i need to parse highlighting property. please look at http://jsfiddle.net/bhXbh/4/ - Thanks Regards Romi -- View this message in context: http://lucene.472066.n3.nabble.com/Highlight-whole-query-phrase-tp3143797p3144157.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Highlight whole query phrase
On Wed, Jul 6, 2011 at 12:07, Romi romijain3...@gmail.com wrote: I included hl.mergeContiguous parameter. in my url but no effect on search result highlighting. actually i need to parse highlighting property. please look at http://jsfiddle.net/bhXbh/4/ I just noticed that I actually don't understand what problem you are trying to solve. The code snippet in your first mail has one example ( and emgold/ememRing/em design) where both gold and ring are highlighted. So this should be what you need. (Don't know about the second one, where it also highlights elegant.) Marian
Re: Field collapsing on multiple fields and/or ranges?
Have the same requirement. What is your workaround for this? On Thu, May 12, 2011 at 7:40 AM, arian487 akarb...@tagged.com wrote: I'm wondering if there is a way to get the field collapsing to collapse on multiple things? For example, is there a way to get it to collapse on a field (lets say 'domain') but ALSO something else (maybe time or something)? To visualize maybe something like this: Group1 has common field 'www.forum1.com' and ALSO the posts are all from may 11 Group2 has common field 'www.forum2.com' and ALSO the posts are all from may 11 . . . GroupX has common field 'www.forum1.com' and ALSO the posts from may 12 So obviously it's still sorted by date but it won't group the 'www.forum1.com' things together if the document is from a different date, it'll group common date AND common domain field. Thanks! -- View this message in context: http://lucene.472066.n3.nabble.com/Field-collapsing-on-multiple-fields-and-or-ranges-tp2929793p2929793.html Sent from the Solr - User mailing list archive at Nabble.com.
How to match for more than one value in a string using regex match
Hi i want to highlight whole gold ring for this i am doing as *var match = hitem.text[0].match(/em(.*?)\/em/); where hitem.text[0]=emGold/em emRing/em * but the problem is var match is getting only Gold so i am able to highlight only Gold, i want to make it an array so that it contains both gold and ring, how can i do it. look at here http: http://jsfiddle.net/bhXbh/4/ - Thanks Regards Romi -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-match-for-more-than-one-value-in-a-string-using-regex-match-tp3144199p3144199.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Field collapsing on multiple fields and/or ranges?
I believe the underlying grouping module is now technically able to do this, because subclasses of the abstract first/second pass grouping collectors are free to decide what type/value the group key is. But, we have to fix Solr to allow for compound keys by creating the necessary concrete subclasses. Mike McCandless http://blog.mikemccandless.com On Wed, Jul 6, 2011 at 6:22 AM, Rih tanrihae...@gmail.com wrote: Have the same requirement. What is your workaround for this? On Thu, May 12, 2011 at 7:40 AM, arian487 akarb...@tagged.com wrote: I'm wondering if there is a way to get the field collapsing to collapse on multiple things? For example, is there a way to get it to collapse on a field (lets say 'domain') but ALSO something else (maybe time or something)? To visualize maybe something like this: Group1 has common field 'www.forum1.com' and ALSO the posts are all from may 11 Group2 has common field 'www.forum2.com' and ALSO the posts are all from may 11 . . . GroupX has common field 'www.forum1.com' and ALSO the posts from may 12 So obviously it's still sorted by date but it won't group the 'www.forum1.com' things together if the document is from a different date, it'll group common date AND common domain field. Thanks! -- View this message in context: http://lucene.472066.n3.nabble.com/Field-collapsing-on-multiple-fields-and-or-ranges-tp2929793p2929793.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Nested grouping/field collapsing
Hey Martijn, Did you find a good workaround? Rih On Sat, May 28, 2011 at 5:35 AM, Martijn Laarman mpdre...@gmail.com wrote: Thanks Mike, I've opened https://issues.apache.org/jira/browse/SOLR-2553 for this. It's exciting to hear a workable implementation might be possible! On Fri, May 27, 2011 at 6:23 PM, Michael McCandless luc...@mikemccandless.com wrote: Can you open a Lucene issue (against the new grouping module) for this? I think this is a compelling use case that we should try to support. In theory, with the general two-pass grouping collector, this should be possible, but will require three passes, and we also must generalize the 2nd pass collector to accept arbitrary collectors for each group (today it's hardwired to sort-by-SortField collectors). I suspect coupling the single-pass grouping collector (currently still a patch on LUCENE-3129) with the two-pass collector could also work. Also, can you describe more details about the two fields you want to group/collapse by? Mike http://blog.mikemccandless.com On Fri, May 27, 2011 at 6:13 AM, Martijn Laarman mpdre...@gmail.com wrote: Hi, I was wondering if this issue had already been raised. We currently have a use case where nested field collapsing would be really helpful I.e Collapse on field X then Collapse on Field Y within the groups returned by field X The current behavior of specifying multiple fields seem to be returning mutiple result sets. Has this already been feature requested ? Does anybody know of a workaround ? Many thanks, Martijn
Re: Not getting data using concat in DIH
concat is really unrelated to Solr, DIH doesn't understand the SQL at all. So my first guess would be that the SQL isn't doing quite what you expect, can you run it by itself and get your expected results? What doesn't work? What error are you receiving? Or is the data just not searchable? Have you tried using the DIH debug page at ...solr/admin/dataimport.jsp, a little-known debugging facility. What version of Solr are you using? Best Erick On Wed, Jul 6, 2011 at 5:08 AM, tamanjit.bin...@yahoo.co.in tamanjit.bin...@yahoo.co.in wrote: The XML comments were not taken. This is the data-import.xml that does not work. In bold is the line uncommented. dataConfig dataSource driver=com.ibm.db2.jcc.DB2Driver url=jdbc:db2://10.150.200.69:6/livedb user=property password=property / document name=articlePosting entity name=articleDetails query=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' deltaQuery=select 'A' concat cast(MBFARTRFNUM as char(8)) as MBFARTUNIQUEKEY, MBFARTRFNUM, MBFARTUBIRFNUM, MBFARTCNDCATE, MBFARTCNDCITY, MBFARTLMTRFNUM, MBFARTDESC, char(date(CREATEDATE)) as CREATEDATE, time(CREATEDATE) as CREATETIME, char(date(LASTMODIDATE)) as MODIDATE, time(LASTMODIDATE) as MODITIME, MBFARTVIEWED, MBFARTSNIPPET, 2 as POPULARITYSORTCODE, 4 as RECENCYSORT, 1 as EXPERTADVICE, MBFARTTOPIC concat ' ' concat MBFARTSNIPPET concat ' ' concat MBFARTDESC concat ' ' concat ' ' concat cast(MBFARTCNDCATE as char(8)) as ADTEXT, 'Y' as EXPERTQUES, 'Y' as EXPERTANS, MBFARTTOPIC, 'A' as CONTENTTYPE from MBFARTICLE where upper(MBFARTISACTIVE) = 'Y' and (CREATEDATE='${dataimporter.last_index_time}' or LASTMODIDATE='${dataimporter.last_index_time}') field column=MBFARTUNIQUEKEY name=ukey / field column=MBFARTRFNUM name=id / field column=MBFARTUBIRFNUM name=oid / field column=MBFARTCNDCATE name=cg / field column=MBFARTCNDCITY name=ct / field column=MBFARTLMTRFNUM name=lt / field column=MBFARTDESC name=dtldesc / field column=CREATEDATE name=ppd / field column=CREATETIME name=ppt / field column=MODIDATE name=ecd / field column=MODITIME name=tmsrt / field column=MBFARTVIEWED name=viewcount / field column=MBFARTSNIPPET name=ansdesc / field column=POPULARITYSORTCODE name=popularsort / field column=RECENCYSORT name=recencysort / field column=EXPERTADVICE name=expadvice / field column=EXPERTQUES name=expertQues / field column=EXPERTANS name=expertAns / field column=MBFARTTOPIC name=topic / field column=CONTENTTYPE name=contentType / field column=ADTEXT name=ad_text / entity name=setOwnerName query=select UBILOGIN from TPUBI where UBIRFNUM=${articleDetails.MBFARTUBIRFNUM} field column=UBILOGIN name=oname / /entity *entity name=setCityName query=select ${articleDetails.ADTEXT} concat ' ' concat CITYNAME as AD_TEXT from TPCITY where CITY_CNDCITY=${articleDetails.MBFARTCNDCITY} fetch first row only field column=AD_TEXT name=ad_text / /entity * /entity /document /dataConfig -- View this message in context:
RE: solr.StandardTokenizerFactory: more info needed
Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java/org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.StandardTokenizerFactory is very short.. -- Regards, Dmitry Kan
Re: How many fields can SOLR handle?
On Jun 7, 2011, at 06:22 , roySolr wrote: Every product has different facets. I have something like this in my schema: dynamicField name=*_FACET type=facetType indexed=true stored=true multiValued=true/ One optimization, if you don't need the stored values, is to set stored=false. Faceting is driven off the indexed terms, not the stored values. But maybe you need the stored values (but maybe not). In SOLR i have now a lot of fields: CPU_FACET, GPU_FACET etc. How many fields can SOLR handle? Another question: Is it possible to add the FACET fields automatically to my query? facet.field=*_FACET? Now i do first a request to a DB to get the FACET titles and add this to the request: facet.field=cpu_FACET,gpu_FACET. I'm affraid that *_FACET is a overkill solution. Solr currently cannot do this automatically. But One thing that I built in to Solr Flare (long live Solr Flare!) was a startup request to Solr's luke request handler (/admin/luke?numTerms=0) and got the list of field names. Then, and I used *_facet similar to you, kept a list of all *_facet actual field names and used those in all subsequent search requests. Erik
Re: Highlight whole query phrase
it is accidentally became elegant. anyway i got the solution without using hl.mergeContiguous , because its not working. look at http://jsfiddle.net/bhXbh/31/ for the solution - Thanks Regards Romi -- View this message in context: http://lucene.472066.n3.nabble.com/Highlight-whole-query-phrase-tp3143797p3144575.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: solr.StandardTokenizerFactory: more info needed
Hi Steven, This looks very good. Thanks. Do I understand correctly, that I were to change the tokenizer rules, I could go and change e.g. the token class definitions (like NUM) in this file and recompile the code? On Wed, Jul 6, 2011 at 3:45 PM, Steven A Rowe sar...@syr.edu wrote: Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java/org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.StandardTokenizerFactory is very short.. -- Regards, Dmitry Kan -- Regards, Dmitry Kan
RE: solr.StandardTokenizerFactory: more info needed
Yes, you can change the rules and recompile. Before you recompile, you have to run 'ant jflex' to generate the java source. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 9:21 AM To: solr-user@lucene.apache.org Subject: Re: solr.StandardTokenizerFactory: more info needed Hi Steven, This looks very good. Thanks. Do I understand correctly, that I were to change the tokenizer rules, I could go and change e.g. the token class definitions (like NUM) in this file and recompile the code? On Wed, Jul 6, 2011 at 3:45 PM, Steven A Rowe sar...@syr.edu wrote: Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java /org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.Stand ardTokenizerFactory is very short.. -- Regards, Dmitry Kan -- Regards, Dmitry Kan
Re: solr.StandardTokenizerFactory: more info needed
OK, thanks. Do you know if there are tokenizer specific tests to run after compilation? On Wed, Jul 6, 2011 at 4:25 PM, Steven A Rowe sar...@syr.edu wrote: Yes, you can change the rules and recompile. Before you recompile, you have to run 'ant jflex' to generate the java source. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 9:21 AM To: solr-user@lucene.apache.org Subject: Re: solr.StandardTokenizerFactory: more info needed Hi Steven, This looks very good. Thanks. Do I understand correctly, that I were to change the tokenizer rules, I could go and change e.g. the token class definitions (like NUM) in this file and recompile the code? On Wed, Jul 6, 2011 at 3:45 PM, Steven A Rowe sar...@syr.edu wrote: Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java /org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.Stand ardTokenizerFactory is very short.. -- Regards, Dmitry Kan -- Regards, Dmitry Kan -- Regards, Dmitry Kan
indexing but not able to search
Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
Re: indexing but not able to search
Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results
Re: indexing but not able to search
Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
A beginner problem
Thank you for your answer. I downloaded solr from the link you sugested and now it is ok, I can see the administration page. But it is strange that a download from the solr site does not work. Tanks also to Way Cool. I don't know why, but it happened the same to me in the past (with 3.2). Apparently the zip I downloaded was not correct. I think you have to have a solr.war file on the webapps directory, do you have it? Do you know which version of Solr you downloaded? Download this one: http://apache.dattatec.com/lucene/solr/3.3.0/apache-solr-3.3.0.zip I just tried it and it's there. On Mon, Jul 4, 2011 at 1:49 PM, carmme...@qualidade.info wrote: I use nutch, as a search engine. Until now nutch did the crawl and the search functions. The newest version, however, delegated the search to solr. I don't know almost nothing about programming, but i'm able to follow a receipe. So I went to the the solr site, downloaded solr and tried to follow the tutorial. In the example folder of solr, using java -jar start.jar I got: 2011-07-04 13:22:38.439:INFO::Logging to STDERR via org.mortbay.log.StdErrLog 2011-07-04 13:22:38.893:INFO::jetty-6.1-SNAPSHOT 2011-07-04 13:22:38.946:INFO::Started SocketConnector@0.0.0.0:8983 When I tried to go to http://localhost:8983/solr/admin/ I got: HTTP ERROR: 404 Problem accessing /solr/admin/. Reason: NOT_FOUND Can someone help me with this? Tanks
Re: Problems with SolrCloud
I have solved this problem by creating myid file with Zookeper server id in solr/zoo_data folder. However, it's still not working. Now when I try to run it, I have the following in the console output: ... Jul 6, 2011 4:23:11 PM org.apache.zookeeper.ClientCnxn$SendThread startConnect INFO: Opening socket connection to server localhost/127.0.0.1:8022 Jul 6, 2011 4:23:11 PM org.apache.zookeeper.ClientCnxn$SendThread primeConnection INFO: Socket connection established to localhost/127.0.0.1:8022, initiating session Jul 6, 2011 4:23:11 PM org.apache.zookeeper.ClientCnxn$SendThread run INFO: Unable to read additional data from server sessionid 0x0, likely server has closed socket, closing socket connection and attempting reconnect Jul 6, 2011 4:23:11 PM org.apache.zookeeper.server.NIOServerCnxn$Factory run INFO: Accepted socket connection from /127.0.0.1:44432 Jul 6, 2011 4:23:11 PM org.apache.zookeeper.server.NIOServerCnxn doIO WARNING: Exception causing close of session 0x0 due to java.io.IOException: ZooKeeperServer not running Jul 6, 2011 4:23:11 PM org.apache.zookeeper.server.NIOServerCnxn closeSock INFO: Closed socket connection for client /127.0.0.1:44432 (no session established for client) ... And http://localhost:8983/solr/admin/zookeeper.jsp is not available:( My configuration (SolrCloud with 2 shards (and 1 replica for each shard) with 4 Zookeepers): 1) shard1 + Zookeper without -Djetty.port it will use standard port 8983 solr/zoo.cfg changed to use port 8011: clientPort=8011 (shard 1, replica 1) the file myid in the default Zookeeper data folder contains: 11 (shard 1, replica 1) $ screen -S shard1 java -server -Dbootstrap_confdir=./solr/conf -Dcollection.configName=myconf -DzkRun -DzkHost=localhost:8011,localhost:8012,localhost:8021,localhost:8022 -jar start.jar 2) shard1-replica + Zookeper solr/zoo.cfg should changed to use port 8012: clientPort=8012 (shard 1, replica 2) the file myid in the default Zookeeper data folder contains: 12 (shard 1, replica 2) $ screen -S shard1_replica java -server -Djetty.port=8984 -DhostPort=8984 -DzkRun -DzkHost=localhost:8011,localhost:8012,localhost:8021,localhost:8022 -jar start.jar 3) shard2 + Zookeper solr/zoo.cfg should changed to use port 8021: clientPort=8021 (shard 2, replica 1) the file myid in the default Zookeeper data folder contains: 21 (shard 2, replica 1) $ screen -S shard2 java -server -Djetty.port=8985 -DhostPort=8985 -DzkRun -DzkHost=localhost:8011,localhost:8012,localhost:8021,localhost:8022 -jar start.jar 4) shard2-replica + Zookeper: solr/zoo.cfg should changed to use port 8022: clientPort=8022 (shard 2, replica 2) the file myid in the default Zookeeper data folder contains: 22 (shard 2, replica 2) $ screen -S shard2_replica java -server -Djetty.port=8986 -DhostPort=8986 -DzkRun -DzkHost=localhost:8011,localhost:8012,localhost:8021,localhost:8022 -jar start.jar On 06/30/2011 05:04 PM, Andrey Sapegin wrote: Dear ladies and gentlemen. Can I ask you to help me with SolrCloud 1) I try to setup a SolrCloud on 2 computers with 3 Zookepers, but it fails:( I need to set Zookeper port to 8001, so I change clientPort=8001 in solr/zoo.cfg. When I try the command from the example C, to run shard1, it works: java -Dbootstrap_confdir=./solr/conf -Dcollection.configName=myconf -DzkRun -DzkHost=localhost:9983,localhost:8574,localhost:9900 -jar start.jar But if I change it to and try to run shard1: java -Dbootstrap_confdir=./solr/conf -Dcollection.configName=myconf -DzkRun -DzkHost=localhost:8001,localhost:8004 -jar start.jar it fails with the following message: SEVERE: java.lang.IllegalArgumentException: solr/zoo_data/myid file is missing 2) to solve it I tried to set *-Dsolr.solr.home=/data/a.sapegin/SolrCloud/shard1* (without any slashes in the end) But then I receive another exception: Caused by: org.apache.zookeeper.server.quorum.QuorumPeerConfig$ConfigException: Error processing /data/a.sapegin/SolrCloud/shard1//zoo.cfg I think this // is a bug. Could you please help? Thank You in advance, Kind Regards, -- Andrey Sapegin, Software Developer, Unister GmbH Dittrichring 18-20 | 04109 Leipzig +49 (0)341 492885069, +4915778339304, andrey.sape...@unister-gmbh.de www.unister.de
Re: indexing but not able to search
Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
Re: solr.StandardTokenizerFactory: more info needed
See ..src/test/org/apache/solr/analysis. But... you'll be changing the grammar, so I don't know how tests would actually help you. Actually I'd expect them to break. And you'd have to write some new ones of your own to exercise your changes to insure that they do what you want Best Erick On Wed, Jul 6, 2011 at 9:31 AM, Dmitry Kan dmitry@gmail.com wrote: OK, thanks. Do you know if there are tokenizer specific tests to run after compilation? On Wed, Jul 6, 2011 at 4:25 PM, Steven A Rowe sar...@syr.edu wrote: Yes, you can change the rules and recompile. Before you recompile, you have to run 'ant jflex' to generate the java source. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 9:21 AM To: solr-user@lucene.apache.org Subject: Re: solr.StandardTokenizerFactory: more info needed Hi Steven, This looks very good. Thanks. Do I understand correctly, that I were to change the tokenizer rules, I could go and change e.g. the token class definitions (like NUM) in this file and recompile the code? On Wed, Jul 6, 2011 at 3:45 PM, Steven A Rowe sar...@syr.edu wrote: Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java /org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.Stand ardTokenizerFactory is very short.. -- Regards, Dmitry Kan -- Regards, Dmitry Kan -- Regards, Dmitry Kan
Re: indexing but not able to search
Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
Obtaining document IDs used in facet value count
I need to take a new approach on accessing my Solr index. It's indexed where each document is a biological node (drug, gene etc.). Each has multi-valued fields for disease and species and other things. To keep the question as simple as possible let's just worry about disease. I'm faceting on the disease field. My original plan is to present node search results, returning (thus stored) and faceting (indexed) on disease and other fields. That all works great, but now the requirement is to be able to determine which nodes have disease x, y, z... That is, in a graphical viewer, we want to allow the user to click on, say, diabetes, and highlight the set of nodes pertaining to a given disease. The viewer requires information like: diabetes - nodeId1, nodeId2, nodeId3 neoplasia - nodeId1, nodeId123 pancreatic cancer - NodeId3, nodeId5 Where nodeIdx is just the document ID. I can see how I could get all the of diseases for all nodes in the set, and go through the SolrDocuments and compile the disease to node mappings myself. Or, I could query Solr and specify the diseases for which to return nodes. Perhaps that would be a single query, or one per disease? But, faceting on disease gives me the document count for each disease type. Is there a way I can use the faceting functionality and get the document IDs rather than just a count? Solr must know the IDs are some point in calculating the facet count. Thanks! Jeff -- Jeff Schmidt 535 Consulting j...@535consulting.com http://www.535consulting.com (650) 423-1068
Re: indexing but not able to search
I am sorry..I was checking the some other solr instance that ran on this system...when I replied for the previous mail. I still dont get any documents in return to my query...though the index shows a size of some 20K documents. Attaching my solrconfig.xml file with this mail. On Wed, Jul 6, 2011 at 5:32 PM, Erick Erickson erickerick...@gmail.comwrote: Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com ?xml version=1.0 encoding=UTF-8 ? !-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the License); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an AS IS BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. -- !-- For more details about configurations options that may appear in this file, see http://wiki.apache.org/solr/SolrConfigXml. -- config !-- In all configuration below, a prefix of solr. for class names is an alias that causes solr to search appropriate packages, including org.apache.solr.(search|update|request|core|analysis) You may also specify a fully qualified Java classname if you have your own custom plugins. -- abortOnConfigurationError${solr.abortOnConfigurationError:true}/abortOnConfigurationError luceneMatchVersionLUCENE_33/luceneMatchVersion !-- lib directives can be used to instruct Solr to load an Jars identified and use them to resolve any plugins specified in your solrconfig.xml or schema.xml (ie: Analyzers, Request Handlers, etc...). All directories and paths are resolved relative to the instanceDir. If a ./lib directory exists in your instanceDir, all files found in it are included as if you had used the following syntax... lib dir=./lib / -- !-- A dir option by itself adds any files found in the directory to the classpath, this is useful for including all jars in a directory. -- lib dir=../../contrib/extraction/lib / lib dir=../../contrib/uima/lib / !-- When a regex is specified in addition to a directory, only the files in that directory which completely match the regex (anchored on both ends) will be included. -- lib dir=../../dist/ regex=apache-solr-cell-\d.*\.jar
RE: Getting started with Velocity
Thanks. Is there any way to change what fields browse uses / asks for? I've tried changing the code, and I'm clearly missing something. I either get the same fields it was displaying before (and no search results) or I get something that doesn't work at all. -Original Message- From: Way Cool [mailto:way1.wayc...@gmail.com] Sent: Friday, July 01, 2011 5:51 PM To: solr-user@lucene.apache.org Subject: Re: Getting started with Velocity By default, browse is using the following config: requestHandler name=/browse class=solr.SearchHandler lst name=defaults str name=echoParamsexplicit/str !-- VelocityResponseWriter settings -- str name=wtvelocity/str str name=v.templatebrowse/str str name=v.layoutlayout/str str name=titleSolritas/str str name=defTypeedismax/str str name=q.alt*:*/str str name=rows10/str str name=fl*,score/str str name=mlt.qf text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4 /str str name=mlt.fltext,features,name,sku,id,manu,cat/str int name=mlt.count3/int str name=qf text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4 /str str name=faceton/str str name=facet.fieldcat/str str name=facet.fieldmanu_exact/str str name=facet.queryipod/str str name=facet.queryGB/str str name=facet.mincount1/str str name=facet.pivotcat,inStock/str str name=facet.rangeprice/str int name=f.price.facet.range.start0/int int name=f.price.facet.range.end600/int int name=f.price.facet.range.gap50/int str name=f.price.facet.range.otherafter/str str name=facet.rangemanufacturedate_dt/str str name=f.manufacturedate_dt.facet.range.startNOW/YEAR-10YEARS/str str name=f.manufacturedate_dt.facet.range.endNOW/str str name=f.manufacturedate_dt.facet.range.gap+1YEAR/str str name=f.manufacturedate_dt.facet.range.otherbefore/str str name=f.manufacturedate_dt.facet.range.otherafter/str !-- Highlighting defaults -- str name=hlon/str str name=hl.fltext features name/str str name=f.name.hl.fragsize0/str str name=f.name.hl.alternateFieldname/str /lst arr name=last-components strspellcheck/str /arr !-- str name=url-schemehttpx/str -- /requestHandler while the normal search is using the following: requestHandler name=search class=solr.SearchHandler default=true !-- default values for query parameters can be specified, these will be overridden by parameters in the request -- lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler. Just make sure you have those fields defined in browse also in your doc, otherwise change to not use dismax. :-) On Fri, Jul 1, 2011 at 12:51 PM, Chip Calhoun ccalh...@aip.org wrote: I'm a Solr novice, so I hope I'm missing something obvious. When I run a search in the Admin view, everything works fine. When I do the same search in http://localhost:8983/solr/browse , I invariably get 0 results found. What am i missing? Are these not supposed to be searching the same index? Thanks, Chip
StreamingUpdateSolrServer Indexing dates
Hi All I'm indexing a set of xml documents using StreamingUpdateSolrServer but I'm having trouble indexing dates. I get an error like: SEVERE: error java.lang.Exception: Invalid Date Math String:'2011-04-22T05:35:37Z ' Invalid Date Math String:'2011-04-22T05:35:37Z ' I notice the trailing spaces but in my document they are not present: maindocdate_dt![CDATA[2011-04-22T05:35:37Z]]/maindocdate_dt I am adding the field to the document with: document.addField(element, sb); //sb is a string I am using the schema supplied with the example as follows but still no joy. fieldType name=tdate class=solr.TrieDateField omitNorms=true precisionStep=6 positionIncrementGap=0/ dynamicField name=*_dt type=tdateindexed=true stored=true/ Any suggestions, thank you, Matt This email was sent to you by Thomson Reuters, the global news and information company. Any views expressed in this message are those of the individual sender, except where the sender specifically states them to be the views of Thomson Reuters.
the version of a Lucene index changes after an optimize?
Hello everybody, I am new in this forum and I need your expertize on Solr-Lucene. I'm currently develop a new component for Solr for a professional project. This component has to be refreshed when some mofications have been applied in the index. But, if only an optimize has happened (the data in the index didn't change), the component doesn't need to be modified. To do that, I'm testing the version number stored in the index which can be retrieved by the IndexReader class with IndexReader.getCurrentVersion(). But, I need to know if the version number is incremented after an optimize operation. Can you tell me if it is the case? If it is, how can I detect that the data have changed in the index ? Thanks for your help! gquaire - Jouve ITS France -- View this message in context: http://lucene.472066.n3.nabble.com/the-version-of-a-Lucene-index-changes-after-an-optimize-tp3143822p3143822.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: indexing but not able to search
Hi Eric 1)Default field in schema.xml : text, which is the field 2) numDocs = 21501 maxDocs = 21554 3) Attaching debugQuery output with this mail 4) When I search for everything, (*:*)...it shows me all the documents, with their fields. I am new to asking questions on the list..and hence the lack of etiquette. Thanks for the link. :) Sowmya. On Wed, Jul 6, 2011 at 6:32 PM, Erick Erickson erickerick...@gmail.comwrote: OK, there's not much information to go on here. So.. 1 you pasted solrconfig.xml. Schema.xml contains your default field, we need to see that too. 2 you say documents are shown in the stats page. There are two numbers, numDocs and maxDocs. numDocs is the number of documents that have NOT been deleted, what is that number? 3 what results from attaching debugQuery=on to your URL? 4 what shows up in the admin page when you search for everything? It would help a lot if you'd provide some more detailed information, please review: http://wiki.apache.org/solr/UsingMailingLists, Best Erick On Wed, Jul 6, 2011 at 12:10 PM, Sowmya V.B. vbsow...@gmail.com wrote: I am sorry..I was checking the some other solr instance that ran on this system...when I replied for the previous mail. I still dont get any documents in return to my query...though the index shows a size of some 20K documents. Attaching my solrconfig.xml file with this mail. On Wed, Jul 6, 2011 at 5:32 PM, Erick Erickson erickerick...@gmail.com wrote: Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com ?xml version=1.0 encoding=UTF-8? response lst name=responseHeader int name=status0/int int name=QTime11/int /lst result name=response numFound=0 start=0/ lst name=debug str name=rawquerystringquery/str str name=querystringquery/str str name=parsedquerytitle:query/str str name=parsedquery_toStringtitle:query/str lst name=explain/ str name=QParserLuceneQParser/str lst name=timing double name=time6.0/double lst name=prepare double name=time0.0/double lst name=org.apache.solr.handler.component.QueryComponent double name=time0.0/double /lst lst name=org.apache.solr.handler.component.FacetComponent double name=time0.0/double /lst lst name=org.apache.solr.handler.component.MoreLikeThisComponent double name=time0.0/double /lst lst name=org.apache.solr.handler.component.HighlightComponent double name=time0.0/double /lst lst
Re: solr.StandardTokenizerFactory: more info needed
Thanks, Erick. On Wed, Jul 6, 2011 at 6:27 PM, Erick Erickson erickerick...@gmail.comwrote: See ..src/test/org/apache/solr/analysis. But... you'll be changing the grammar, so I don't know how tests would actually help you. Actually I'd expect them to break. And you'd have to write some new ones of your own to exercise your changes to insure that they do what you want Best Erick On Wed, Jul 6, 2011 at 9:31 AM, Dmitry Kan dmitry@gmail.com wrote: OK, thanks. Do you know if there are tokenizer specific tests to run after compilation? On Wed, Jul 6, 2011 at 4:25 PM, Steven A Rowe sar...@syr.edu wrote: Yes, you can change the rules and recompile. Before you recompile, you have to run 'ant jflex' to generate the java source. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 9:21 AM To: solr-user@lucene.apache.org Subject: Re: solr.StandardTokenizerFactory: more info needed Hi Steven, This looks very good. Thanks. Do I understand correctly, that I were to change the tokenizer rules, I could go and change e.g. the token class definitions (like NUM) in this file and recompile the code? On Wed, Jul 6, 2011 at 3:45 PM, Steven A Rowe sar...@syr.edu wrote: Hi Dmitry, The underlying Lucene implementation is here: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_2_9_1/src/java /org/apache/lucene/analysis/standard/ StandardTokenizerImpl.jflex is probably where you should start. Steve -Original Message- From: Dmitry Kan [mailto:dmitry@gmail.com] Sent: Wednesday, July 06, 2011 3:23 AM To: solr-user@lucene.apache.org Subject: solr.StandardTokenizerFactory: more info needed Hi all! solr.StandardTokenizerFactory -- is it possible to see the full description of its behaviour for solr.1.4 somewhere? Wiki http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.Stand ardTokenizerFactory is very short.. -- Regards, Dmitry Kan -- Regards, Dmitry Kan -- Regards, Dmitry Kan -- Regards, Dmitry Kan
Re: the version of a Lucene index changes after an optimize?
I question this point: But, if only an optimize has happened (the data in the index didn't change), the component doesn't need to be modified An optimize may, for instance, change the internal Lucene document IDs. What is your component doing? Also, optimize should be a fairly rare occurrence. I'm wondering if it's worth the hassle to detect it. Best Erick On Wed, Jul 6, 2011 at 3:37 AM, gquaire gquaire@gmail.com wrote: Hello everybody, I am new in this forum and I need your expertize on Solr-Lucene. I'm currently develop a new component for Solr for a professional project. This component has to be refreshed when some mofications have been applied in the index. But, if only an optimize has happened (the data in the index didn't change), the component doesn't need to be modified. To do that, I'm testing the version number stored in the index which can be retrieved by the IndexReader class with IndexReader.getCurrentVersion(). But, I need to know if the version number is incremented after an optimize operation. Can you tell me if it is the case? If it is, how can I detect that the data have changed in the index ? Thanks for your help! gquaire - Jouve ITS France -- View this message in context: http://lucene.472066.n3.nabble.com/the-version-of-a-Lucene-index-changes-after-an-optimize-tp3143822p3143822.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: indexing but not able to search
About being new... no problem, we all have to learn But this part of your output: str name=rawquerystringquery/str str name=querystringquery/str str name=parsedquerytitle:query/str str name=parsedquery_toStringtitle:query/str indicates that something's weird with your query. Can you show the exact URL you use? If you're searching form the admin interface, that will be the URL in the results window. Because this is indicating a couple of things: 1 your query is of the form ?q=query or some such. 2 your default search field is title (see schema.xml).. Best Erick On Wed, Jul 6, 2011 at 12:47 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Eric 1)Default field in schema.xml : text, which is the field 2) numDocs = 21501 maxDocs = 21554 3) Attaching debugQuery output with this mail 4) When I search for everything, (*:*)...it shows me all the documents, with their fields. I am new to asking questions on the list..and hence the lack of etiquette. Thanks for the link. :) Sowmya. On Wed, Jul 6, 2011 at 6:32 PM, Erick Erickson erickerick...@gmail.com wrote: OK, there's not much information to go on here. So.. 1 you pasted solrconfig.xml. Schema.xml contains your default field, we need to see that too. 2 you say documents are shown in the stats page. There are two numbers, numDocs and maxDocs. numDocs is the number of documents that have NOT been deleted, what is that number? 3 what results from attaching debugQuery=on to your URL? 4 what shows up in the admin page when you search for everything? It would help a lot if you'd provide some more detailed information, please review: http://wiki.apache.org/solr/UsingMailingLists, Best Erick On Wed, Jul 6, 2011 at 12:10 PM, Sowmya V.B. vbsow...@gmail.com wrote: I am sorry..I was checking the some other solr instance that ran on this system...when I replied for the previous mail. I still dont get any documents in return to my query...though the index shows a size of some 20K documents. Attaching my solrconfig.xml file with this mail. On Wed, Jul 6, 2011 at 5:32 PM, Erick Erickson erickerick...@gmail.com wrote: Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com
Re: indexing but not able to search
Hi Eric Yes, the schema.xml has both title and text fields... and i was changing between them...perhaps, it was title when I sent you. I am searching from the admin. this is the URL it gives me, after I click search, from the admin window. http://loalhost:8080/apache-solr-3.3.0/select/?q=headversion=2.2start=0rows=10indent=on S On Wed, Jul 6, 2011 at 7:12 PM, Erick Erickson erickerick...@gmail.comwrote: About being new... no problem, we all have to learn But this part of your output: str name=rawquerystringquery/str str name=querystringquery/str str name=parsedquerytitle:query/str str name=parsedquery_toStringtitle:query/str indicates that something's weird with your query. Can you show the exact URL you use? If you're searching form the admin interface, that will be the URL in the results window. Because this is indicating a couple of things: 1 your query is of the form ?q=query or some such. 2 your default search field is title (see schema.xml).. Best Erick On Wed, Jul 6, 2011 at 12:47 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Eric 1)Default field in schema.xml : text, which is the field 2) numDocs = 21501 maxDocs = 21554 3) Attaching debugQuery output with this mail 4) When I search for everything, (*:*)...it shows me all the documents, with their fields. I am new to asking questions on the list..and hence the lack of etiquette. Thanks for the link. :) Sowmya. On Wed, Jul 6, 2011 at 6:32 PM, Erick Erickson erickerick...@gmail.com wrote: OK, there's not much information to go on here. So.. 1 you pasted solrconfig.xml. Schema.xml contains your default field, we need to see that too. 2 you say documents are shown in the stats page. There are two numbers, numDocs and maxDocs. numDocs is the number of documents that have NOT been deleted, what is that number? 3 what results from attaching debugQuery=on to your URL? 4 what shows up in the admin page when you search for everything? It would help a lot if you'd provide some more detailed information, please review: http://wiki.apache.org/solr/UsingMailingLists, Best Erick On Wed, Jul 6, 2011 at 12:10 PM, Sowmya V.B. vbsow...@gmail.com wrote: I am sorry..I was checking the some other solr instance that ran on this system...when I replied for the previous mail. I still dont get any documents in return to my query...though the index shows a size of some 20K documents. Attaching my solrconfig.xml file with this mail. On Wed, Jul 6, 2011 at 5:32 PM, Erick Erickson erickerick...@gmail.com wrote: Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler name=search class=solr.SearchHandler default=true lst name=defaults str name=echoParamsexplicit/str int name=rows10/int /lst /requestHandler - Apart from this, I did not understand what to add inside a search handler. Sowmya. On Wed, Jul 6, 2011 at 3:45 PM, Denis Kuzmenok forward...@ukr.net wrote: Hi All I indexed a set of documents using Solr, which are shown in the stats page on the admin panel. However, the search interface always returns 0 documents to me. When I give the query as *:*, it does return me all the 20K odd documents I tried indexing just a few hours back. Can someone tell me if there is anything I am missing, on the querying config part? Sowmya. Show your solrconfig.xml, and url you are querying to select results -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism is blasphemy! http://vbsowmya.wordpress.com -- Sowmya V.B. Losing optimism
Re: indexing but not able to search
You're giving contradictory information here. This is NOT the query that you submitted when you did the debugQuery=on that you sent before. Look in schema.xml for defaultSearchField I bet its value is title. I'm not talking at all about the fact that the schema has title and text fields, I'm looking at the debug output and inferring the defaultSearchField because: str name=rawquerystringquery/str str name=querystringquery/str str name=parsedquerytitle:query/str str name=parsedquery_toStringtitle:query/str the rawquery string is query. Meaning you typed something like q=query. the parsedquery is title:query implying that your defaultSearchField is title. That means that the URL you gave is searching against the title field for head. Do you really expect that to match? Please make the effort to provide a consistent set of data. Don't give fragmentary pieces from different queries. Your debug data cannot be from a schema that has text as the default field. It just doesn't work that way. So, I suspect you aren't going against the solr instance you think. Or you're looking at configuration data that isn't being used by that solr. Or you're cutting/pasting/copying different fragments. And you still haven't shown us the schema.xml file. So, give us the debug output, and show us the exact query you use to get that output. Best Erick On Wed, Jul 6, 2011 at 1:20 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Eric Yes, the schema.xml has both title and text fields... and i was changing between them...perhaps, it was title when I sent you. I am searching from the admin. this is the URL it gives me, after I click search, from the admin window. http://loalhost:8080/apache-solr-3.3.0/select/?q=headversion=2.2start=0rows=10indent=on S On Wed, Jul 6, 2011 at 7:12 PM, Erick Erickson erickerick...@gmail.comwrote: About being new... no problem, we all have to learn But this part of your output: str name=rawquerystringquery/str str name=querystringquery/str str name=parsedquerytitle:query/str str name=parsedquery_toStringtitle:query/str indicates that something's weird with your query. Can you show the exact URL you use? If you're searching form the admin interface, that will be the URL in the results window. Because this is indicating a couple of things: 1 your query is of the form ?q=query or some such. 2 your default search field is title (see schema.xml).. Best Erick On Wed, Jul 6, 2011 at 12:47 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Eric 1)Default field in schema.xml : text, which is the field 2) numDocs = 21501 maxDocs = 21554 3) Attaching debugQuery output with this mail 4) When I search for everything, (*:*)...it shows me all the documents, with their fields. I am new to asking questions on the list..and hence the lack of etiquette. Thanks for the link. :) Sowmya. On Wed, Jul 6, 2011 at 6:32 PM, Erick Erickson erickerick...@gmail.com wrote: OK, there's not much information to go on here. So.. 1 you pasted solrconfig.xml. Schema.xml contains your default field, we need to see that too. 2 you say documents are shown in the stats page. There are two numbers, numDocs and maxDocs. numDocs is the number of documents that have NOT been deleted, what is that number? 3 what results from attaching debugQuery=on to your URL? 4 what shows up in the admin page when you search for everything? It would help a lot if you'd provide some more detailed information, please review: http://wiki.apache.org/solr/UsingMailingLists, Best Erick On Wed, Jul 6, 2011 at 12:10 PM, Sowmya V.B. vbsow...@gmail.com wrote: I am sorry..I was checking the some other solr instance that ran on this system...when I replied for the previous mail. I still dont get any documents in return to my query...though the index shows a size of some 20K documents. Attaching my solrconfig.xml file with this mail. On Wed, Jul 6, 2011 at 5:32 PM, Erick Erickson erickerick...@gmail.com wrote: Note that you can add that to the URL directly, as fl=* Also, simply querying q=time will look for terms ONLY in the field in your schema.xml defined by defaultSearchField in your schema.xml (text by default)... Best Erick On Wed, Jul 6, 2011 at 11:08 AM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Denis I got my answer. I should have added str name=fl*/str in the defaults list. Sowmya. On Wed, Jul 6, 2011 at 3:57 PM, Sowmya V.B. vbsow...@gmail.com wrote: Hi Url I am using to search: http://localhost:8080/apache-solr-3.3.0/select/?q=timeversion=2.2start=0rows=10indent=on http://localhost:8080/apache-solr-3.3.0/admin/stats.jsp - is my stats page. The part of my SolrConfig file, which said searchHandler - had these, which I found on an example on the web: requestHandler
Re: Nightly builds
: The reason for the email is not that I can't find them, but because : the project, I claim, should be advertising them more prominently on : the web site than buried in a wiki. : : Actually they are linked on the hp but unfortunately not directly: : http://lucene.apache.org/solr/#21+February+2006%3A+nightly+builds http://www.apache.org/dev/release.html#what Do not include any links on the project website that might encourage non-developers to download and use nightly builds, snapshots, release candidates, or any other similar package. The only people who are supposed to know about such packages are the people following the dev list (or searching its archives) and thus aware of the conditions placed on the package. For these reasons, we speicifly do not link to nightly builds off the main website (but we also don't retroactively edit news from 2006) We do however mention/link to the nightly build system on the wiki pages targeted at developers. -Hoss
Re: StreamingUpdateSolrServer Indexing dates
: Invalid Date Math String:'2011-04-22T05:35:37Z ' ... : I notice the trailing spaces but in my document they are not present: ... : maindocdate_dt![CDATA[2011-04-22T05:35:37Z]]/maindocdate_dt ... : I am adding the field to the document with: ... : document.addField(element, sb); //sb is a string There is at least one step missing in your process -- the XML you showed us isn't the Solr XML Message format, so it can't be the poutput of using the StreamingUpdateSolrServer after you build up SOlrInputDocuments. i suspect you have some code that parses your XML, and in that code you call document.addField ... have you verified in that parsing/building code exactly what's going on and what the value of sb is (ie: via debugger) ? -Hoss
DIH / dynamic fields / ...
Hi, I'm currently stuck with a (probably straightforward) problem concerning DIH and dynamic fields. I'm having a DB-Datasource with one of the columns (metaXml) containing an xml string looking something like this: rootmeta title=abc value=123/meta title=def value=456//root The dataConfig looks something like this: dataConfig … entity dataSource=mssql name=doc … … entity name='meta' dataSource='f' processor='XPathEntityProcessor' forEach='/meta' dataField='doc.metaXml' field column='meta_Title' xpath='/meta/@title' / field column='meta_Content' xpath='/meta/@value' / field name=meta_${meta.meta_Title} column=${meta.meta_Content} / /entity /entity /dataConfig In my schema.xml I have a dynamic field looking like: dynamicField name=meta_* type=string indexed=true stored=true/ For whatever reason those meta_* fields are not populated ... Maybe (hopefully) I'm missing something obvious. Any help would be great ;) Thanks, Fred.
Getting the indexed value rather than the stored value
Hi, I am using the following fieldType to change the index of a given field: fieldtype name=profanity stored=false indexed=true class=solr.TextField analyzer type=index tokenizer class=solr.StandardTokenizerFactory/ filter class=solr.KeepWordFilterFactory words=profane-words.txt ignoreCase=true/ filter class=solr.PatternReplaceFilterFactory pattern=^(.+)$ replacement=YES replace=all/ /analyzer analyzer type=query tokenizer class=solr.StandardTokenizerFactory/ /analyzer /fieldtype This is great for finding all things with or without profanity (as separate queries), but I would like to get the value as part of a the query and let the consumer of the call decide what to do with the data. Is there a way to do this w/o having to instantiate a KeepWordFilterFactory in the Java class that is responsible for inserting the document into Solr? For example, I know that I can do this in Java code during the insert, but I would rather get the indexed value (the one that shows up when faceting). Please let me know if this is not clear. Thanks, Christian
ClassCastException launching recent snapshot
Launching solr-4.0-20110705.223601-1.war, I get a class cast exception org.apache.lucene.index.DirectoryReader cannot be cast to org.apache.solr.search.SolrIndexReader with the following backtrace. I'm launching solr-as-a-webapp via an embedded copy of tomcat 7. The location of the index is set up via: System.setProperty(solr.data.dir, solrDataDirectory); Further, the sources in the corresponding -sources .jar doesn't seem to have a cast to SolrIndexReader in it anywhere in SolrIndexSearcher. SolrIndexSearcher.init(SolrCore, IndexSchema, String, IndexReader, boolean, boolean) line: 142 SolrCore.getSearcher(boolean, boolean, Future[]) line: 1085 SolrCore.init(String, String, SolrConfig, IndexSchema, CoreDescriptor) line: 587 CoreContainer.create(CoreDescriptor) line: 660 CoreContainer.load(String, InputStream) line: 412 CoreContainer$Initializer.initialize() line: 246 SolrDispatchFilter.init(FilterConfig) line: 86 ApplicationFilterConfig.initFilter() line: 273 ApplicationFilterConfig.getFilter() line: 254 ApplicationFilterConfig.setFilterDef(FilterDef) line: 372 ApplicationFilterConfig.init(Context, FilterDef) line: 98 StandardContext.filterStart() line: 4584 StandardContext$2.call() line: 5262 StandardContext$2.call() line: 5257 FutureTask$Sync.innerRun() line: 303 FutureTaskV.run() line: 138 ThreadPoolExecutor$Worker.runTask(Runnable) line: 886
Re: Nightly builds
On Wed, Jul 6, 2011 at 3:43 PM, Chris Hostetter hossman_luc...@fucit.org wrote: : The reason for the email is not that I can't find them, but because : the project, I claim, should be advertising them more prominently on : the web site than buried in a wiki. : : Actually they are linked on the hp but unfortunately not directly: : http://lucene.apache.org/solr/#21+February+2006%3A+nightly+builds http://www.apache.org/dev/release.html#what Do not include any links on the project website that might encourage non-developers to download and use nightly builds, snapshots, release candidates, or any other similar package. The only people who are supposed to know about such packages are the people following the dev list (or searching its archives) and thus aware of the conditions placed on the package. For these reasons, we speicifly do not link to nightly builds off the main website (but we also don't retroactively edit news from 2006) Given that prose, I can see why you have done what you've done. I don't agree with the prose, but that's another story for another location. Though I guess the author would probably respond to me by encouraging me to encourage you to vote and formally release beta or alphas or something to give brave people exposure to new features. We do however mention/link to the nightly build system on the wiki pages targeted at developers. -Hoss