Re: Solrits - Problem with facing after restart yetty

2009-05-29 Thread Erik Hatcher
I think you're asking if the (very temporary on trunk) faceting bug is  
fixed.  The answer is yes.


Erik

On May 29, 2009, at 3:10 AM, Jörg Agatz wrote:


i the Bug fixt in the news Nightliy Bilds?




replication solr 1.4

2009-05-29 Thread sunnyfr

Hi guys,

I didnt ennoy you for ages now ... hope everybody is fine ... I've an issue
with my replication
I was wondering ... after a while replication doesnt work anymore ... 
we have a script which enable or not replication every 2hours and this
morning it didnt pull anything 
and it's maybe because the version is too far away ? is it possible ??? 
is it possible to tell it to just take the last one ? 

cuz now it's working but because i clicked on replication.

thanks a lot,
Johanna



Replicatable Index Version:1239958162858, Generation: 1741
Poll Interval   00:00:30
Local Index Index Version: 1239958162819, Generation: 1702
Location: /data/solr/video/data/index
Size: 8.86 GB
Times Replicated Since Startup: 896
Previous Replication Done At: Thu May 28 14:39:51 CEST 2009
Config Files Replicated At:
Config Files Replicated:
Times Config Files Replicated Since Startup:
Next Replication Cycle At: Thu May 28 14:34:31 CEST 2009
Current Replication Status  Start Time: Fri May 29 11:49:50 CEST 2009
Files Downloaded: 0 / 12
Downloaded: 1.66 GB / 8.87 GB [18.0%]
Downloading File: _5io.prx, Downloaded: 1.66 GB / 2.22 GB [74.0%]
Time Elapsed: 122s, Estimated Time Remaining: 528s, Speed: 13.96 MB/s
Controls

Cores:
[video ][user ][group ]

Current Time: Fri May 29 11:51:52 CEST 2009
Server Start At: Thu May 28 10:43:30 CEST 2009

Return to Admin Page 
-- 
View this message in context: 
http://www.nabble.com/replication-solr-1.4-tp23777206p23777206.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: replication solr 1.4

2009-05-29 Thread sunnyfr

what would be the url to ping to replicate 
like http://slave_host:port/solr/replication?command=enablepoll 

thanks 
-- 
View this message in context: 
http://www.nabble.com/replication-solr-1.4-tp23777206p23777272.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: java.lang.RuntimeException: after flush: fdx size mismatch

2009-05-29 Thread Michael McCandless
Very interesting: FieldsWriter thinks it's written 12 bytes to the fdx
file, yet the directory says the file does not exist.

Can you re-run with this new patch?  I'm suspecting that FieldsWriter
wrote to one segment, but somehow we are then looking at the wrong
segment.  The attached patch prints out which segment FieldsWriter
actually wrote to.

What filesystem  underlying IO system/device are you using?

Mike

On Thu, May 28, 2009 at 10:53 PM, James X
hello.nigerian.spamm...@gmail.com wrote:
 My apologies for the delay in running this patched Lucene build - I was
 temporarily pulled onto another piece of work.

 Here is a sample 'fdx size mismatch' exception using the patch Mike
 supplied:

 SEVERE: java.lang.RuntimeException: after flush: fdx size mismatch: 1 docs
 vs 0 length in bytes of _1i.fdx exists=false didInit=false inc=0 dSO=1
 fieldsWriter.doClose=true fieldsWriter.indexFilePointer=12
 fieldsWriter.fieldsFilePointer=2395
        at
 org.apache.lucene.index.StoredFieldsWriter.closeDocStore(StoredFieldsWriter.java:96)
        at
 org.apache.lucene.index.DocFieldConsumers.closeDocStore(DocFieldConsumers.java:83)
        at
 org.apache.lucene.index.DocFieldProcessor.closeDocStore(DocFieldProcessor.java:47)
        at
 org.apache.lucene.index.DocumentsWriter.closeDocStore(DocumentsWriter.java:367)
        at
 org.apache.lucene.index.DocumentsWriter.flush(DocumentsWriter.java:567)
        at
 org.apache.lucene.index.IndexWriter.doFlush(IndexWriter.java:3540)
        at org.apache.lucene.index.IndexWriter.flush(IndexWriter.java:3450)
        at
 org.apache.lucene.index.IndexWriter.closeInternal(IndexWriter.java:1638)
        at org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1602)
        at org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1578)
        at
 org.apache.solr.update.SolrIndexWriter.close(SolrIndexWriter.java:153)


 Will now run with assertions enabled and see how that affects the behaviour!

 Thanks,
 James

 -- Forwarded message --
 From: James X hello.nigerian.spamm...@gmail.com
 Date: Thu, May 21, 2009 at 2:24 PM
 Subject: Re: java.lang.RuntimeException: after flush: fdx size mismatch
 To: solr-user@lucene.apache.org


 Hi Mike,Documents are web pages, about 20 fields, mostly strings, a couple
 of integers, booleans and one html field (for document body content).

 I do have a multi-threaded client pushing docs to Solr, so yes, I suppose
 that would mean I have several active Solr worker threads.

 The only exceptions I have are the RuntimeException flush errors, followed
 by a handful (normally 10-20) of LockObtainFailedExceptions, which i
 presumed were being caused by the faulty threads dying and failing to
 release locks.

 Oh wait, I am getting WstxUnexpectedCharException exceptions every now and
 then:
 SEVERE: com.ctc.wstx.exc.WstxUnexpectedCharException: Illegal character
 ((CTRL-CHAR, code 8))
  at [row,col {unknown-source}]: [1,26070]
        at
 com.ctc.wstx.sr.StreamScanner.throwInvalidSpace(StreamScanner.java:675)
        at
 com.ctc.wstx.sr.BasicStreamReader.readTextSecondary(BasicStreamReader.java:4668)
        at
 com.ctc.wstx.sr.BasicStreamReader.readCoalescedText(BasicStreamReader.java:4126)
        at
 com.ctc.wstx.sr.BasicStreamReader.finishToken(BasicStreamReader.java:3701)
        at
 com.ctc.wstx.sr.BasicStreamReader.safeFinishToken(BasicStreamReader.java:3649)
        at
 com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:809)
        at
 org.apache.solr.handler.XmlUpdateRequestHandler.readDoc(XmlUpdateRequestHandler.java:327)

 I presumed these were caused by character encoding issues, but haven't
 looked into them at all yet.

 Thanks again for your help! I'll make some time this afternoon to build some
 patched Lucene jars and get the results


 On Thu, May 21, 2009 at 5:06 AM, Michael McCandless 
 luc...@mikemccandless.com wrote:

 Another question: are there any other exceptions in your logs?  Eg
 problems adding certain documents, or anything?

 Mike

 On Wed, May 20, 2009 at 11:18 AM, James X
 hello.nigerian.spamm...@gmail.com wrote:
  Hi Mike, thanks for the quick response:
 
  $ java -version
  java version 1.6.0_11
  Java(TM) SE Runtime Environment (build 1.6.0_11-b03)
  Java HotSpot(TM) 64-Bit Server VM (build 11.0-b16, mixed mode)
 
  I hadn't noticed the 268m trigger for LUCENE-1521 - I'm definitely not
  hitting that yet!
 
  The exception always reports 0 length, but the number of of docs varies,
  heavily weighted towards 1 or two docs. Of the last 130 or so exceptions:
      89 1 docs vs 0 length
      20 2 docs vs 0 length
       9 3 docs vs 0 length
       1 4 docs vs 0 length
       3 5 docs vs 0 length
       2 6 docs vs 0 length
       1 7 docs vs 0 length
       1 9 docs vs 0 length
       1 10 docs vs 0 length
 
  The only unusual thing I can think of that we're doing with Solr is
  aggressively CREATE-ing and UNLOAD-ing cores. I've not been able to spot
 a
  pattern between core 

Avoid duplicates in MoreLikeThis using field collapsing

2009-05-29 Thread Marc Sturlese

Hey there, 
I am testing MoreLikeThis feaure (with MoreLikeThis component and with
MoreLikeThis handler) and I am getting lots of duplicates. I have noticed
that lots of the similar documents returned are duplicates. To avoid that I
have tried to use the field collapsing patch but it's not taking effect.

In case of MoreLikeThis handler I think it's normal has I have seen it
extends directly from RequestHandlerBase.java and not from
SearchHandler.java that is the one that in the function handleRequestBody
will deal with components:

  for( SearchComponent c : components ) {
rb.setTimer( subt.sub( c.getName() ) );
c.prepare(rb);
rb.getTimer().stop();
  }

To sort it out I have embbed the collapseFilter in the getMoreLikeThis
method of the MoreLikeThisHandler.java
This is working alrite but would like to know if is there any more polite
way to make MoreLikeThisHandler able to deal with components. I mean via
solrconfig.xml or pluging something instead of hacking it.

Thanks in advance


-- 
View this message in context: 
http://www.nabble.com/Avoid-duplicates-in-MoreLikeThis-using-field-collapsing-tp23778054p23778054.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Solrits - Problem with facing after restart yetty

2009-05-29 Thread Erik Hatcher
Jorg - the rest of that exception would be mighty handy!  Please share  
the entire details.


Erik

On May 29, 2009, at 7:38 AM, Jörg Agatz wrote:


also,

i have after using the Nightly bUILD FROM tODY, 29.05.2009
BUT THE sAME ERROR...

HTTP ERROR 500

null

java.lang.NullPointerException
...
...
...

The url is: .../solr/itas?fq=cat:test
when i tryed .../solr/itas?q=SEARCHWORDcat:test ir worket, links  
dosent

work




Re: Solrits - Problem with facing after restart yetty

2009-05-29 Thread Jörg Agatz
HTTP ERROR: 500

null

java.lang.NullPointerException
at java.io.StringReader.init(StringReader.java:33)
at org.apache.lucene.queryParser.QueryParser.parse(QueryParser.java:169)
at 
org.apache.solr.search.LuceneQParser.parse(LuceneQParserPlugin.java:78)
at org.apache.solr.search.QParser.getQuery(QParser.java:131)
at 
org.apache.solr.handler.component.QueryComponent.prepare(QueryComponent.java:91)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:174)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1328)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:341)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:244)
at 
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089)
at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365)
at 
org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
at 
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)
at 
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712)
at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405)
at 
org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211)
at 
org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139)
at org.mortbay.jetty.Server.handle(Server.java:285)
at 
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502)
at 
org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:821)
at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:513)
at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:208)
at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378)
at 
org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226)
at 
org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442)

RequestURI=/solr/itas

*Powered by Jetty:// http://jetty.mortbay.org/*


goog_1243600484811

My Conf:

requestHandler name=/itas class=solr.SearchHandler
 lst name=defaults
   str name=v.templatebrowse/str
   str name=v.propertiesvelocity.properties/str
   str name=titleSolritas/str

   str name=wtvelocity/str
   str name=q.alt*:*/str
   str name=rows10/str
   str name=fl*,score/str
   str name=faceton/str
   str name=facet.fieldextension/str
   str name=facet.mincount1/str
   str name=qf/str
 /lst



*
http://jetty.mortbay.org/*


Re: Java OutOfmemory error during autowarming

2009-05-29 Thread Yonik Seeley
It's probably not the size of the query cache, but the size of the
FieldCache entries that are used for sorting and function queries
(that's the only thing that should be allocating huge arrays like
that).

What fields do you sort on or use function queries on?  There may be a
way to decrease the memory consumption.

-Yonik
http://www.lucidimagination.com

On Fri, May 29, 2009 at 1:02 AM, Francis Yakin fya...@liquid.com wrote:
 During auto-warming of solr search on QueryResultKey, Our Production 
 solrslaves errors throw OutOfMemory error and application need to be bounced.

 Here is the error logs:

 SEVERE: Error during auto-warming of 
 key:org.apache.solr.search.queryresult...@9acd5d67mailto:org.apache.solr.search.queryresult...@9acd5d67:java.lang.OutOfMemoryError:
  allocLargeObject
 OrArray - Object size: 25291544, Num elements: 6322881
 SEVERE: Error during auto-warming of 
 key:org.apache.solr.search.queryresult...@84554284mailto:org.apache.solr.search.queryresult...@84554284:java.lang.OutOfMemoryError:
  allocLargeObject
 OrArray - Object size: 579368, Num elements: 144837
 SEVERE: Error during auto-warming of 
 key:org.apache.solr.search.queryresult...@8c1e63e4mailto:org.apache.solr.search.queryresult...@8c1e63e4:java.lang.OutOfMemoryError:
  allocLargeObject
 OrArray - Object size: 4280976, Num elements: 1070240


 I read some suggestions from google that increasing the size of 
 queryResultCache in the solrconfig.xml file will help, here is the suggestion:


 This is not directly related to the jvm heap size. During slave replication 
 the current Searcher is used as the source of auto-warming. When a new 
 searcher is opened, its caches may be prepopulated or autowarmed using data 
 from caches in the old searcher. It sounds like caching configuration problem 
 to me as is evident from the error message Error during auto-warming of 
 key. I suggest increasing the size of the query cache on one of the slave 
 servers and monitor it over next week or so and then incrementally roll this 
 change to rest of the slave boxes. This is the change that you need to make 
 in solrconfig.xml

 Right now this what I have:

      queryResultCache
      class=solr.LRUCache
      size=512
      initialSize=512
      autowarmCount=256/

 Should I increase size to 1024 will help?

 Any input?

 Thanks

 Francis





Re: Custom sort

2009-05-29 Thread Marc Sturlese

I have been able to create my custom field. The problem is that I have laoded
in the solr core a couple of HashMapsid_doc,value_influence_sort from a DB
with values that will influence in the sort. My problem is that I don't know
how to let my custom sort have access to this HashMaps.
I am a bit confused now. I think that would be easy to reach my goal using:

CustomSortComponent extends SearchComponent implements SolrCoreAware

This way, I would load the HashMaps in the inform method and would create
de custom sort using the HashMaps in the preprare method.

Don't know how to do that with the CustomField (similar to the
RandomField)... any advice?


Yonik Seeley-2 wrote:
 
 The custom Field type similar to RandomSortField is definitely the way
 to go over a custom component since it's much less invasive - then
 everything else like distributed search should work.
 
 But see SOLR-... I'm in the process of converting Solr to use the
 new FieldComparatorSource.
 
 -Yonik
 http://www.lucidimagination.com
 
 
 
 On Tue, May 26, 2009 at 9:25 AM, Marc Sturlese marc.sturl...@gmail.com
 wrote:

 Hey there,

 I want to implement a custom sort coding my own SortComparatorSource. The
 sorter would sort the results doing some calculations of some document
 attributes witch I have loaded previously from a DB at the solr core
 init. I
 mean, I have a preloaded hash with some attributes per doc id and I want
 to
 use them to implement the sorting.
 I have considered two options:

 First would be to implement a custom field similar to the solr
 RandomSortField and when search requests are done I would tell solr to
 sort
 using this custom field.

 The other option would be to plug a component (wich would extend from
 SearchComponent) and do the custom sort in the prepare method.

 Any advice witch is the most optimal way to do that?

 Thanks in advance.
 --
 View this message in context:
 http://www.nabble.com/Custom-sort-tp23722921p23722921.html
 Sent from the Solr - User mailing list archive at Nabble.com.


 
 

-- 
View this message in context: 
http://www.nabble.com/Custom-sort-tp23722921p23783579.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: index time boosting on multivalued fields

2009-05-29 Thread Erick Erickson
What are you really trying to accomplish here? Because index time boostingis
a way of saying I care about matches in this field of this document
X times more than other documents whereas search time boosting
expresses elevate the relevance of any document where this term matches


From your example, it looks like you're really thinking about search-time
boosting rather than index time.

If I'm out in left field, perhaps an explanation of the problem you're
trying
to solve would help me clarify...

Best
Erick

On Wed, May 27, 2009 at 3:14 PM, Brian Whitman br...@echonest.com wrote:

 I can set the boost of a field or doc at index time using the boost attr in
 the update message, e.g.
 field name=type boost=2.0pet/field

 But that won't work for multivalued fields according to the RelevancyFAQ

 field name=type boost=2.0pet/field
 field name=type boost=1.0animal/field

 ( I assume it applies the last boost parsed to all terms? )

 Now, say I'd like to do index-time boosting of a multivalued field with
 each
 value having a unique boost. I could simply index the field multiple times:

 field name=typepet/field
 field name=typepet/field
 field name=typeanimal/field

 But is there a more exact way?



RE: Java OutOfmemory error during autowarming

2009-05-29 Thread Francis Yakin

There is no FieldCache entries in solrconfig.xml ( BTW we are running version 
1.2.0)


-Original Message-
From: Yonik Seeley [mailto:ysee...@gmail.com]
Sent: Friday, May 29, 2009 9:12 AM
To: solr-user@lucene.apache.org
Subject: Re: Java OutOfmemory error during autowarming

It's probably not the size of the query cache, but the size of the FieldCache 
entries that are used for sorting and function queries (that's the only thing 
that should be allocating huge arrays like that).

What fields do you sort on or use function queries on?  There may be a way to 
decrease the memory consumption.

-Yonik
http://www.lucidimagination.com

On Fri, May 29, 2009 at 1:02 AM, Francis Yakin fya...@liquid.com wrote:
 During auto-warming of solr search on QueryResultKey, Our Production 
 solrslaves errors throw OutOfMemory error and application need to be bounced.

 Here is the error logs:

 SEVERE: Error during auto-warming of
 key:org.apache.solr.search.queryresult...@9acd5d67mailto:org.apache.s
 olr.search.queryresult...@9acd5d67:java.lang.OutOfMemoryError:
 allocLargeObject OrArray - Object size: 25291544, Num elements:
 6322881
 SEVERE: Error during auto-warming of
 key:org.apache.solr.search.queryresult...@84554284mailto:org.apache.s
 olr.search.queryresult...@84554284:java.lang.OutOfMemoryError:
 allocLargeObject OrArray - Object size: 579368, Num elements: 144837
 SEVERE: Error during auto-warming of
 key:org.apache.solr.search.queryresult...@8c1e63e4mailto:org.apache.s
 olr.search.queryresult...@8c1e63e4:java.lang.OutOfMemoryError:
 allocLargeObject OrArray - Object size: 4280976, Num elements: 1070240


 I read some suggestions from google that increasing the size of 
 queryResultCache in the solrconfig.xml file will help, here is the suggestion:


 This is not directly related to the jvm heap size. During slave
 replication the current Searcher is used as the source of
 auto-warming. When a new searcher is opened, its caches may be
 prepopulated or autowarmed using data from caches in the old
 searcher. It sounds like caching configuration problem to me as is
 evident from the error message Error during auto-warming of key. I
 suggest increasing the size of the query cache on one of the slave
 servers and monitor it over next week or so and then incrementally
 roll this change to rest of the slave boxes. This is the change that
 you need to make in solrconfig.xml

 Right now this what I have:

  queryResultCache
  class=solr.LRUCache
  size=512
  initialSize=512
  autowarmCount=256/

 Should I increase size to 1024 will help?

 Any input?

 Thanks

 Francis





Re: Java OutOfmemory error during autowarming

2009-05-29 Thread Yonik Seeley
On Fri, May 29, 2009 at 1:44 PM, Francis Yakin fya...@liquid.com wrote:

 There is no FieldCache entries in solrconfig.xml ( BTW we are running 
 version 1.2.0)

Lucene FieldCache entries are created when you sort on a field or when
you use a field in a function query.

-Yonik


RE: Java OutOfmemory error during autowarming

2009-05-29 Thread Francis Yakin

I know, but the FieldCache is not in the solrconfig.xml


-Original Message-
From: Yonik Seeley [mailto:ysee...@gmail.com]
Sent: Friday, May 29, 2009 10:47 AM
To: solr-user@lucene.apache.org
Subject: Re: Java OutOfmemory error during autowarming

On Fri, May 29, 2009 at 1:44 PM, Francis Yakin fya...@liquid.com wrote:

 There is no FieldCache entries in solrconfig.xml ( BTW we are
 running version 1.2.0)

Lucene FieldCache entries are created when you sort on a field or when you use 
a field in a function query.

-Yonik


Re: java.lang.RuntimeException: after flush: fdx size mismatch

2009-05-29 Thread James X
Hi Mike,I don't see a patch file here?

Could another explanation be that the fdx file doesn't exist yet / has been
deleted from underneath Lucene?

I'm constantly CREATE-ing and UNLOAD-ing Solr cores, and more importantly,
moving the bundled cores around between machines. I find it much more likely
that there's something wrong with my core admin code than there is with the
Lucene internals :) It's possible that I'm occasionally removing files which
are currently in use by a live core...

I'm using an ext3 filesystem on a large EC2 instance's own hard disk. I'm
not sure how Amazon implement the local hard disk, but I assume it's a real
hard disk exposed by the hypervisor.

Thanks,
James

On Fri, May 29, 2009 at 3:41 AM, Michael McCandless 
luc...@mikemccandless.com wrote:

 Very interesting: FieldsWriter thinks it's written 12 bytes to the fdx
 file, yet the directory says the file does not exist.

 Can you re-run with this new patch?  I'm suspecting that FieldsWriter
 wrote to one segment, but somehow we are then looking at the wrong
 segment.  The attached patch prints out which segment FieldsWriter
 actually wrote to.

 What filesystem  underlying IO system/device are you using?

 Mike

 On Thu, May 28, 2009 at 10:53 PM, James X
 hello.nigerian.spamm...@gmail.com wrote:
  My apologies for the delay in running this patched Lucene build - I was
  temporarily pulled onto another piece of work.
 
  Here is a sample 'fdx size mismatch' exception using the patch Mike
  supplied:
 
  SEVERE: java.lang.RuntimeException: after flush: fdx size mismatch: 1
 docs
  vs 0 length in bytes of _1i.fdx exists=false didInit=false inc=0 dSO=1
  fieldsWriter.doClose=true fieldsWriter.indexFilePointer=12
  fieldsWriter.fieldsFilePointer=2395
 at
 
 org.apache.lucene.index.StoredFieldsWriter.closeDocStore(StoredFieldsWriter.java:96)
 at
 
 org.apache.lucene.index.DocFieldConsumers.closeDocStore(DocFieldConsumers.java:83)
 at
 
 org.apache.lucene.index.DocFieldProcessor.closeDocStore(DocFieldProcessor.java:47)
 at
 
 org.apache.lucene.index.DocumentsWriter.closeDocStore(DocumentsWriter.java:367)
 at
  org.apache.lucene.index.DocumentsWriter.flush(DocumentsWriter.java:567)
 at
  org.apache.lucene.index.IndexWriter.doFlush(IndexWriter.java:3540)
 at
 org.apache.lucene.index.IndexWriter.flush(IndexWriter.java:3450)
 at
  org.apache.lucene.index.IndexWriter.closeInternal(IndexWriter.java:1638)
 at
 org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1602)
 at
 org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1578)
 at
  org.apache.solr.update.SolrIndexWriter.close(SolrIndexWriter.java:153)
 
 
  Will now run with assertions enabled and see how that affects the
 behaviour!
 
  Thanks,
  James
 
  -- Forwarded message --
  From: James X hello.nigerian.spamm...@gmail.com
  Date: Thu, May 21, 2009 at 2:24 PM
  Subject: Re: java.lang.RuntimeException: after flush: fdx size mismatch
  To: solr-user@lucene.apache.org
 
 
  Hi Mike,Documents are web pages, about 20 fields, mostly strings, a
 couple
  of integers, booleans and one html field (for document body content).
 
  I do have a multi-threaded client pushing docs to Solr, so yes, I suppose
  that would mean I have several active Solr worker threads.
 
  The only exceptions I have are the RuntimeException flush errors,
 followed
  by a handful (normally 10-20) of LockObtainFailedExceptions, which i
  presumed were being caused by the faulty threads dying and failing to
  release locks.
 
  Oh wait, I am getting WstxUnexpectedCharException exceptions every now
 and
  then:
  SEVERE: com.ctc.wstx.exc.WstxUnexpectedCharException: Illegal character
  ((CTRL-CHAR, code 8))
   at [row,col {unknown-source}]: [1,26070]
 at
  com.ctc.wstx.sr.StreamScanner.throwInvalidSpace(StreamScanner.java:675)
 at
 
 com.ctc.wstx.sr.BasicStreamReader.readTextSecondary(BasicStreamReader.java:4668)
 at
 
 com.ctc.wstx.sr.BasicStreamReader.readCoalescedText(BasicStreamReader.java:4126)
 at
 
 com.ctc.wstx.sr.BasicStreamReader.finishToken(BasicStreamReader.java:3701)
 at
 
 com.ctc.wstx.sr.BasicStreamReader.safeFinishToken(BasicStreamReader.java:3649)
 at
  com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:809)
 at
 
 org.apache.solr.handler.XmlUpdateRequestHandler.readDoc(XmlUpdateRequestHandler.java:327)
 
  I presumed these were caused by character encoding issues, but haven't
  looked into them at all yet.
 
  Thanks again for your help! I'll make some time this afternoon to build
 some
  patched Lucene jars and get the results
 
 
  On Thu, May 21, 2009 at 5:06 AM, Michael McCandless 
  luc...@mikemccandless.com wrote:
 
  Another question: are there any other exceptions in your logs?  Eg
  problems adding certain documents, or anything?
 
  Mike
 
  On Wed, May 20, 2009 at 11:18 AM, James X
  

Refresh synonyms.txt file via replication

2009-05-29 Thread mlathe

Hi All,
I would like to provide an admin interface (in a different system) that
would update the synonyms.txt file and automatically inform a set of Solr
instances that are being replicated to update their synonyms.txt file too.

This discussion shows a possible solution:
http://www.nabble.com/Refresh-of-synonyms.txt-without-reload-td19629361.html
Namely use the CoreAdmin system to do a RELOAD, rather than forcing tomcat
to reload the webapp.

Here is my setup:
1) I made these instances (both are separate webapps running one core named
core, this allows me to use the CoreAdmin):
http://localhost:8080/solr_rep_master/core/admin/
http://localhost:8080/solr_rep_slave1/core/admin/
2) the master's config looks like this:
lst name=master
str name=replicateAfterstartup/str
str name=replicateAftercommit/str
str name=snapshotstartup/str
str name=snapshotcommit/str
str name=confFilesschema.xml,stopwords.txt,synonyms.txt/str
/lst
3) the slave's config looks like this:
lst name=slave
str
name=masterUrlhttp://localhost:8080/solr_rep_master/core/replication/str
str name=pollInterval00:00:20/str  
str name=compressioninternal/str
/lst

First Question:
If i update the solr_rep_master's conf/synonyms.txt file and GET
http://localhost:8080/solr_rep_master/admin/cores?action=RELOADcore=core,
the config is not replicated to the slave. Isn't a RELOAD considered a
startup in regards to the master's replicateAfter setting?

Second Question:
If i force an empty commit, like this:
curl
http://localhost:8080/solr_rep_master/core/update?stream.body=%3Ccommit/%3E
Then the changed synonym.txt config file are replicated to the slave.
Unfortunately now I need to do a core RELOAD on both the master and slave
to get them to see the updated synonym.txt file. 
Is there any way to circumvent this reload? (this is what was effectively
asked in the nabble link at the top)

Thanks
--Matthias
-- 
View this message in context: 
http://www.nabble.com/Refresh-synonyms.txt-file-via-replication-tp23789187p23789187.html
Sent from the Solr - User mailing list archive at Nabble.com.



query issue /special character and case

2009-05-29 Thread revas
Hi ,

When i give a query like the following ,why does it become a phrase query
as shown below?
The field type is the default text field in the schema.

str name=querystringvolker-blanz/str
str name=parsedqueryPhraseQuery(content:volker blanz)/str

Also when i have special characters in the query as SCHÖLKOPF , i am not
able to convert the o with spl character  to lower case on my unix os/it
works fine on windows xp OS .Also if i have a spl character in my  query ,i
would like to search for it wihtout the special character as  SCHOLKOPF
,this works fine in windows with strtr (string translate php fucntion) ,but
again not in windows OS.

Any pointers

Regards
Revas