Re: Question on StreamingUpdateSolrServer
On Sat, Apr 11, 2009 at 3:29 AM, vivek sar vivex...@gmail.com wrote: I also noticed that the Solr app has over 6000 file handles open - lsof | grep solr | wc -l - shows 6455 I've 10 cores (using multi-core) managed by the same Solr instance. As soon as start up the Tomcat the open file count goes up to 6400. Few questions, 1) Why is Solr holding on to all the segments from all the cores - is it because of auto-warmer? You have 10 cores, so Solr opens 10 indexes, each of which contains multiple files. That is one reason. Apart from that, Tomcat will keep some file handles for incoming connections. 2) How can I reduce the open file count? Are they causing a problem? Tomcat will log messages when it cannot accept incoming connections if it runs out of available file handles. But if you experiencing issues, you can increase the file handle limit or you can set useCompoundFile=true in solrconfig.xml. 3) Is there a way to stop the auto-warmer? 4) Could this be related to Tomcat returning blank page for every request? It could be. Check the Tomcat and Solr logs. -- Regards, Shalin Shekhar Mangar.
Re: sorlj search
On Wed, Feb 6, 2008 at 10:51 AM, Tevfik Kiziloren tevfikkizilo...@gmail.com wrote: Caused by: org.apache.solr.common.SolrException: parsing error at org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:138) at org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:99) at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:317) at org.apache.solr.client.solrj.request.QueryRequest.process(QueryRequest.java:84) ... 29 more Caused by: java.lang.RuntimeException: this must be known type! not: int at org.apache.solr.client.solrj.impl.XMLResponseParser.readNamedList(XMLResponseParser.java:217) at org.apache.solr.client.solrj.impl.XMLResponseParser.readNamedList(XMLResponseParser.java:235) at org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:123) Which version of Solr and Solrj client are you using? -- Regards, Shalin Shekhar Mangar.
Re: Help with relevance failure in Solr 1.3
On Apr 10, 2009, at 5:50 PM, Walter Underwood wrote: Normally, both changeling and the changeling work fine. This one server is misbehaving like this for all multi-term queries. Yes, it is VERY weird that the term changeling does not show up in the explain. A server will occasionally go bad and stay in that state. In one case, two servers went bad and both gave the same wrong results. What's the solution for when they go bad? Do you have to restart Solr or reboot or what? Here is the dismax config. groups means movies. The title* fields are stemmed and stopped, the exact* fields are not. !-- groups and people -- requestHandler name=groups_people class=solr.SearchHandler lst name=defaults str name=defTypedismax/str str name=echoParamsnone/str float name=tie0.01/float str name=qf exact^6.0 exact_alt^6.0 exact_base~jw_0.7_1^8.0 exact_alias^8.0 title^3.0 title_alt^3.0 title_base^4.0 /str str name=pf exact^9.0 exact_alt^9.0 exact_base^12.0 exact_alias^12.0 title^3.0 title_alt^4.0 title_base^6.0 /str str name=bf search_popularity^100.0 /str str name=mm1/str int name=ps100/int str name=flid,type,movieid,personid,genreid/str /lst lst name=appends str name=fqtype:group OR type:person/str /lst /requestHandler wunder On 4/10/09 12:51 PM, Grant Ingersoll gsing...@apache.org wrote: On Apr 10, 2009, at 1:56 PM, Walter Underwood wrote: We have a rare, hard-to-reproduce problem with our Solr 1.3 servers, and I would appreciate any ideas. Ocassionally, a server will start returning results with really poor relevance. Single term queries work fine, but multi-term queries are scored based on the most common term (lowest IDF). I don't see anything in the logs when this happens. We have a monitor doing a search for the 100 most popular movies once per minute to catch this, so we know when it was first detected. I'm attaching two explain outputs, one for the query changeling and one for the changeling. I'm not sure what exactly you are asking, so bear with me... Are you saying that the changeling normally returns results just fine and then periodically it will go bad or are you saying you don't understand why the changeling scores differently from changeling? In looking at the explains, it is weird that in the the changeling case, the term changeling doesn't even show up as a term. Can you share your dismax configuration? That will be easier to parse than trying to make sense of the debug query parsing. -Grant -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: Index Version Number
On Fri, Apr 10, 2009 at 11:58 AM, Richard Wiseman rwise...@infosciences.com wrote: Is it possible for a Solr client to determine if the index has changed since the last time it performed a query? For example, is it possible to query the current Lucene indexVersion? Grant pointed to one way - the Luke handler. Another way is to look at the Last-Modified or ETag HTTP headers. $ curl -i http://localhost:8983/solr/select?q=solr HTTP/1.1 200 OK Last-Modified: Fri, 10 Apr 2009 17:40:54 GMT ETag: OWZlNjdkN2Q4ODAwMDAwU29scg== Content-Type: text/xml; charset=utf-8 Content-Length: 2308 Server: Jetty(6.1.3) -Yonik http://www.lucidimagination.com
Re: Its urgent! plz help in schema.xml- appending one field to another
Erik Hatcher wrote: On Apr 8, 2009, at 9:50 PM, Udaya wrote: Hi, Need your help, I would like to know how we could append or add one field value to another field in Scheme.xml My scheme is as follows (only the field part is given): Scheme.xml fields field name=topics_id type=integer indexed=true stored=true required=true / field name=topics_subject type=text indexed=true stored=true required=true/ field name=post_text type=text indexed=true stored=true multiValued=true/ field name=url type=string stored=true default=http://comp.com/portals/ForumWindow? action=1v=tp=topics_id#topics_id / field name=all_text type=text indexed=true stored=true multiValued=true/ Here for the field with name topics_id we get id from a table. I what his topics_id value to be appended into the default value attribute of the field with name url. For eg: Suppose if we get topics_id value as 512 during a search then the value of the url should be appended as http://comp.com/portals/JBossForumWindow?action=1v=tp=512#512 Is this possible, plz give me some suggestions. If you're using DIH to index your table, you could aggregate using the template transformer during indexing. If you're indexing a different way, why not let the searching client (UI) do the aggregation of an id into a URL? Erik Hi Erik thanks for your suggestion :-)... i do have appended the id into the url by writing code in solrj and java... thanks Udaya -- View this message in context: http://www.nabble.com/Its-urgent%21-plz-help-in-schema.xml--appending-one-field-to-another-tp22963108p23001380.html Sent from the Solr - User mailing list archive at Nabble.com.
Evaluating Solr
Hi All, I'm evaluating Solr to provide a single fast and effective searching solution for a range of web applications (Java, .Net PHP). So a couple questions: 1) Can we configure Solr to auto reindex/import if we point it to a datasource? I've done some searching on this and have only come up with specific mailing list questions on DataImporter - there seems to be no docs on that. If so, can we use this mechnism to point to multiple datasources for different indexes i.e. use datasource A for index A and datasource B for index B? 2) Can we create multiple indexes of the same objects i.e. index of users for customer A and another for users of customer B? If so, how would be specify context/index to use? 3) It seems that the built-in replication depends on rsync; we use a SAN for our production databases, can we easily use the SAN for replication/redundancy? Are there fiel system locking issues (as with out-of-the-box lucene) with two Solr servers using the same index shared via a SAN? 4) We've extended one of our local lucene search implementations to store the index in a database (not for speed but for redundancy load balancing) using a JDBCDirectoryProvider implementation - is it possible to use that with Solr? Thanks for any help on this! rotis 23
Re: Help with relevance failure in Solr 1.3
Restarting Solr fixes it. If I remember correctly, a sync and commit does not fix it. I have disabled snappuller this time, so I can study the broken instance. wunder On 4/11/09 5:03 AM, Grant Ingersoll gsing...@apache.org wrote: On Apr 10, 2009, at 5:50 PM, Walter Underwood wrote: Normally, both changeling and the changeling work fine. This one server is misbehaving like this for all multi-term queries. Yes, it is VERY weird that the term changeling does not show up in the explain. A server will occasionally go bad and stay in that state. In one case, two servers went bad and both gave the same wrong results. What's the solution for when they go bad? Do you have to restart Solr or reboot or what? Here is the dismax config. groups means movies. The title* fields are stemmed and stopped, the exact* fields are not. !-- groups and people -- requestHandler name=groups_people class=solr.SearchHandler lst name=defaults str name=defTypedismax/str str name=echoParamsnone/str float name=tie0.01/float str name=qf exact^6.0 exact_alt^6.0 exact_base~jw_0.7_1^8.0 exact_alias^8.0 title^3.0 title_alt^3.0 title_base^4.0 /str str name=pf exact^9.0 exact_alt^9.0 exact_base^12.0 exact_alias^12.0 title^3.0 title_alt^4.0 title_base^6.0 /str str name=bf search_popularity^100.0 /str str name=mm1/str int name=ps100/int str name=flid,type,movieid,personid,genreid/str /lst lst name=appends str name=fqtype:group OR type:person/str /lst /requestHandler wunder On 4/10/09 12:51 PM, Grant Ingersoll gsing...@apache.org wrote: On Apr 10, 2009, at 1:56 PM, Walter Underwood wrote: We have a rare, hard-to-reproduce problem with our Solr 1.3 servers, and I would appreciate any ideas. Ocassionally, a server will start returning results with really poor relevance. Single term queries work fine, but multi-term queries are scored based on the most common term (lowest IDF). I don't see anything in the logs when this happens. We have a monitor doing a search for the 100 most popular movies once per minute to catch this, so we know when it was first detected. I'm attaching two explain outputs, one for the query changeling and one for the changeling. I'm not sure what exactly you are asking, so bear with me... Are you saying that the changeling normally returns results just fine and then periodically it will go bad or are you saying you don't understand why the changeling scores differently from changeling? In looking at the explains, it is weird that in the the changeling case, the term changeling doesn't even show up as a term. Can you share your dismax configuration? That will be easier to parse than trying to make sense of the debug query parsing. -Grant -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: Evaluating Solr
On Sat, Apr 11, 2009 at 7:26 PM, Chris Masters roti...@yahoo.com wrote: I'm evaluating Solr to provide a single fast and effective searching solution for a range of web applications (Java, .Net PHP). That's good news for Solr :) So a couple questions: 1) Can we configure Solr to auto reindex/import if we point it to a datasource? I've done some searching on this and have only come up with specific mailing list questions on DataImporter - there seems to be no docs on that. If so, can we use this mechnism to point to multiple datasources for different indexes i.e. use datasource A for index A and datasource B for index B? There's a comprehensive wiki document at http://wiki.apache.org/solr/DataImportHandler 2) Can we create multiple indexes of the same objects i.e. index of users for customer A and another for users of customer B? If so, how would be specify context/index to use? Looks like you need http://wiki.apache.org/solr/MultipleIndexes 3) It seems that the built-in replication depends on rsync; we use a SAN for our production databases, can we easily use the SAN for replication/redundancy? Are there fiel system locking issues (as with out-of-the-box lucene) with two Solr servers using the same index shared via a SAN? The issues would be same as for Lucene. I've never used Solr with SAN so I do not know about if it will work or how well it will perform. 4) We've extended one of our local lucene search implementations to store the index in a database (not for speed but for redundancy load balancing) using a JDBCDirectoryProvider implementation - is it possible to use that with Solr? Again, never tried. Though, support for configurable directory provider was added recently (in trunk). See https://issues.apache.org/jira/browse/SOLR-465 for more details. -- Regards, Shalin Shekhar Mangar.
DIH XML/update
Hi, Wondering if there is a solution to this. I am using DIH to index an XML feed, I update the feed every 10 min and now I have an index of more then 10 million docs, I do DIH update/clean = false. Now more and more I update my indexing time is increasing and its coming to a point where I can see indexing time will be greater then the update time. Do you have any suggestion? I am thinking of creating a large index and then every X hours merge the small index with the large ones -- small index is the one which gets updated by the feed. any other thoughts? How are you folks doing it? Regards Antonio __ Låna pengar utan säkerhet. Jämför vilkor online hos Kelkoo. http://www.kelkoo.se/c-100390123-lan-utan-sakerhet.html?partnerId=96915014
Evaluating Solr and Ruby-Solr
We are also Evaluating Solr and Sphinx to see which one provides a suitable search solution for our ROR site. As a newbie I picked the Advanced Rails Recipes book and follow Recipe 27 before building our prototype. As I followed Recipe 27 of the Advanced Rails Recipes book. I tried to follow the steps exactly to see if I can drill down the search results using facets. Initially solr and solr-ruby appeared to work fine. The below statements worked great and I was able to query the solr index from Ruby. solr = Solr::Connection.new(’http://127.0.0.1:8983/solr’) response = solr.query(’pragmatic’, :facets = {:fields = ['pubisher_facet'], :mincount =1}) docs = response.hits But when I ran response.facet_fields(’publisher_facet’) I got the below error response.field_facets(’publisher_facet’) NoMethodError: undefined method `size’ for nil:NilClass from C:/Ruby/lib/ruby/gems/1.8/gems/solr-ruby-0.0.7/lib/solr/util.rb:17:in `paired_array_each’ from C:/Ruby/lib/ruby/gems/1.8/gems/solr-ruby-0.0.7/lib/solr/response/standard.rb:43:in `field_facets’ from (irb):13 As development box, we currently use solr 1.3, Ruby 1.8.6 on Windows Vista, solr-ruby 0.0.7. Please advise if I missed something here wnen follow the recipe? Thank you. Thanh Below are installed gems on my pc. C:\Users\thanhruby -v ruby 1.8.6 (2008-08-11 patchlevel 287) [i386-mswin32] C:\Users\thanhgem -v 1.3.1 C:\Users\thanhgem list *** LOCAL GEMS *** actionmailer (2.3.2) actionpack (2.3.2) activerecord (2.3.2) activeresource (2.3.2) activesupport (2.3.2) capistrano (2.5.5) cgi_multipart_eof_fix (2.5.0) fxri (0.3.6) fxruby (1.6.16) gem_plugin (0.2.3) highline (1.5.0) hpricot (0.6.164) log4r (1.0.5) mongrel (1.1.5) mongrel_cluster (1.0.5) net-scp (1.0.2) net-sftp (2.0.2) net-ssh (2.0.11) net-ssh-gateway (1.0.1) ptools (1.1.6) rails (2.3.2) rake (0.8.4, 0.8.1) ruby-opengl (0.60.0) solr-ruby (0.0.7) test-unit (2.0.1) win32-api (1.2.1, 1.2.0) win32-clipboard (0.4.4) win32-dir (0.3.2) win32-eventlog (0.5.0) win32-file (0.5.5) win32-file-stat (1.3.1) win32-process (0.5.9) win32-sapi (0.1.4) win32-sound (0.4.1) windows-api (0.2.4) windows-pr (0.9.3)
Re: Evaluating Solr
Thanks for the info Shalin - most useful! With regard to https://issues.apache.org/jira/browse/SOLR-465; I see it's been reopened - roughly when is v1.4 planned for release? Thanks, rotis - Original Message From: Shalin Shekhar Mangar shalinman...@gmail.com To: solr-user@lucene.apache.org Sent: Saturday, April 11, 2009 7:13:11 PM Subject: Re: Evaluating Solr On Sat, Apr 11, 2009 at 7:26 PM, Chris Masters roti...@yahoo.com wrote: I'm evaluating Solr to provide a single fast and effective searching solution for a range of web applications (Java, .Net PHP). That's good news for Solr :) So a couple questions: 1) Can we configure Solr to auto reindex/import if we point it to a datasource? I've done some searching on this and have only come up with specific mailing list questions on DataImporter - there seems to be no docs on that. If so, can we use this mechnism to point to multiple datasources for different indexes i.e. use datasource A for index A and datasource B for index B? There's a comprehensive wiki document at http://wiki.apache.org/solr/DataImportHandler 2) Can we create multiple indexes of the same objects i.e. index of users for customer A and another for users of customer B? If so, how would be specify context/index to use? Looks like you need http://wiki.apache.org/solr/MultipleIndexes 3) It seems that the built-in replication depends on rsync; we use a SAN for our production databases, can we easily use the SAN for replication/redundancy? Are there fiel system locking issues (as with out-of-the-box lucene) with two Solr servers using the same index shared via a SAN? The issues would be same as for Lucene. I've never used Solr with SAN so I do not know about if it will work or how well it will perform. 4) We've extended one of our local lucene search implementations to store the index in a database (not for speed but for redundancy load balancing) using a JDBCDirectoryProvider implementation - is it possible to use that with Solr? Again, never tried. Though, support for configurable directory provider was added recently (in trunk). See https://issues.apache.org/jira/browse/SOLR-465 for more details. -- Regards, Shalin Shekhar Mangar.
Re: Evaluating Solr and Ruby-Solr
On Apr 11, 2009, at 3:05 PM, Thanh Doan wrote: We are also Evaluating Solr and Sphinx to see which one provides a suitable search solution for our ROR site. As a newbie I picked the Advanced Rails Recipes book and follow Recipe 27 before building our prototype. As I followed Recipe 27 of the Advanced Rails Recipes book. I tried to follow the steps exactly to see if I can drill down the search results using facets. Initially solr and solr-ruby appeared to work fine. The below statements worked great and I was able to query the solr index from Ruby. solr = Solr::Connection.new(’http://127.0.0.1:8983/solr’) response = solr.query(’pragmatic’, :facets = {:fields = ['pubisher_facet'], :mincount =1}) docs = response.hits But when I ran response.facet_fields(’publisher_facet’) I got the below error response.field_facets(’publisher_facet’) NoMethodError: undefined method `size’ for nil:NilClass from It's field_facets, not facet_fields. Using Solr's example data, solr- ruby 0.0.7, and irb - Solr::Connection.new.query('ipod', :facets = {:fields = ['cat']}).field_facets('cat').each {|f| p #{f.name} (#{f.value})} electronics (3) connector (2) music (1) camera (0) card (0) copier (0) drive (0) graphics (0) hard (0) memory (0) monitor (0) multifunction (0) printer (0) scanner (0) search (0) software (0) Erik
Re: Question on StreamingUpdateSolrServer
Thanks Shalin. The problem is I don't see any error message in the catalina.out. I don't even see the request coming in - I simply get blank page on browser. If I keep trying the request goes through and I get respond from Solr, but then it become unresponsive again or sometimes throws connection reset error. I'm not sure why would it work sometimes and not the other times for the same query. As soon as I stop the Indexer process things start working fine. Any way I can debug this problem? -vivek On Fri, Apr 10, 2009 at 11:05 PM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: On Sat, Apr 11, 2009 at 3:29 AM, vivek sar vivex...@gmail.com wrote: I also noticed that the Solr app has over 6000 file handles open - lsof | grep solr | wc -l - shows 6455 I've 10 cores (using multi-core) managed by the same Solr instance. As soon as start up the Tomcat the open file count goes up to 6400. Few questions, 1) Why is Solr holding on to all the segments from all the cores - is it because of auto-warmer? You have 10 cores, so Solr opens 10 indexes, each of which contains multiple files. That is one reason. Apart from that, Tomcat will keep some file handles for incoming connections. 2) How can I reduce the open file count? Are they causing a problem? Tomcat will log messages when it cannot accept incoming connections if it runs out of available file handles. But if you experiencing issues, you can increase the file handle limit or you can set useCompoundFile=true in solrconfig.xml. 3) Is there a way to stop the auto-warmer? 4) Could this be related to Tomcat returning blank page for every request? It could be. Check the Tomcat and Solr logs. -- Regards, Shalin Shekhar Mangar.
Re: Question on StreamingUpdateSolrServer
On Sun, Apr 12, 2009 at 2:15 AM, vivek sar vivex...@gmail.com wrote: The problem is I don't see any error message in the catalina.out. I don't even see the request coming in - I simply get blank page on browser. If I keep trying the request goes through and I get respond from Solr, but then it become unresponsive again or sometimes throws connection reset error. I'm not sure why would it work sometimes and not the other times for the same query. As soon as I stop the Indexer process things start working fine. Any way I can debug this problem? I'm not sure. I've never seen this issue myself. Could you try using the bundled jetty instead of Tomcat or on a different box just to make sure this is not an environment specific issue? -- Regards, Shalin Shekhar Mangar.
Re: Evaluating Solr
On Sun, Apr 12, 2009 at 1:49 AM, Chris Masters roti...@yahoo.com wrote: With regard to https://issues.apache.org/jira/browse/SOLR-465; I see it's been reopened Yes, there are issues related to hard coded FSDirectory usage. There's patch in there but I haven't looked at it yet. - roughly when is v1.4 planned for release? That's a tough one. It is not something that I control but a few days ago I had answered mid next month for the same question :) The right answer is: when all the issues marked for 1.4 are resolved and there is consensus in the community for a release. You can see the list of unresolved issues at the following url. https://issues.apache.org/jira/secure/IssueNavigator.jspa?reset=truemode=hidesorter/order=DESCsorter/field=priorityresolution=-1pid=12310230fixfor=12313351 -- Regards, Shalin Shekhar Mangar.
Re: maxCodeLength in PhoneticFilterFactory
There's definitely a bug - I just reproduced it. Nothing obvious jumps out at me... and there's no error in the logs either (that's another bug it would seem). Could you open a JIRA issue for this? -Yonik http://www.lucidimagination.com On Fri, Apr 10, 2009 at 6:54 PM, Brian Whitman br...@echonest.com wrote: i have this version of solr running: Solr Implementation Version: 1.4-dev 747554M - bwhitman - 2009-02-24 16:37:49 and am trying to update a schema to support 8 code length metaphone instead of 4 via this (committed) issue: https://issues.apache.org/jira/browse/SOLR-813 So I change the schema to this (knowing that I have to reindex) filter class=solr.PhoneticFilterFactory encoder=DoubleMetaphone inject=true maxCodeLength=8/ But when I do queries fail with Error_initializing_DoubleMetaphoneclass_orgapachecommonscodeclanguageDoubleMetaphone__at_orgapachesolranalysisPhoneticFilterFactoryinitPhoneticFilterFactoryjava90__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava821__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava817__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava149__at_orgapachesolrschemaIndexSchemareadAnalyzerIndexSchemajava831__at_orgapachesolrschemaIndexSchemaaccess$100IndexSchemajava58__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava425__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava410__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava141__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava452__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava501__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava121
Re: maxCodeLength in PhoneticFilterFactory
OK, should hopefully be fixed in trunk. -Yonik http://www.lucidimagination.com On Sat, Apr 11, 2009 at 9:16 PM, Yonik Seeley yo...@apache.org wrote: There's definitely a bug - I just reproduced it. Nothing obvious jumps out at me... and there's no error in the logs either (that's another bug it would seem). Could you open a JIRA issue for this? -Yonik http://www.lucidimagination.com On Fri, Apr 10, 2009 at 6:54 PM, Brian Whitman br...@echonest.com wrote: i have this version of solr running: Solr Implementation Version: 1.4-dev 747554M - bwhitman - 2009-02-24 16:37:49 and am trying to update a schema to support 8 code length metaphone instead of 4 via this (committed) issue: https://issues.apache.org/jira/browse/SOLR-813 So I change the schema to this (knowing that I have to reindex) filter class=solr.PhoneticFilterFactory encoder=DoubleMetaphone inject=true maxCodeLength=8/ But when I do queries fail with Error_initializing_DoubleMetaphoneclass_orgapachecommonscodeclanguageDoubleMetaphone__at_orgapachesolranalysisPhoneticFilterFactoryinitPhoneticFilterFactoryjava90__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava821__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava817__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava149__at_orgapachesolrschemaIndexSchemareadAnalyzerIndexSchemajava831__at_orgapachesolrschemaIndexSchemaaccess$100IndexSchemajava58__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava425__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava410__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava141__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava452__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava501__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava121