Re: reindexing a solr collection of nested documents

2016-11-29 Thread adfel70
Anyone has a clue?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/reindexing-a-solr-collection-of-nested-documents-tp4307586p4307976.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: LukeRequestHandler Error getting file length for [segments_1l]

2016-11-29 Thread Shawn Heisey
On 11/29/2016 8:40 AM, halis Yılboğa wrote:
> it is not normal to get that many error actually. Main problem should be
> from your index. It seems to me your index is corrupted.
>
> 29 Kas 2016 Sal, 14:40 tarihinde, Furkan KAMACI 
> şunu yazdı:
>
>> On the other hand, my Solr instance stops frequently due to such errors:
>>
>> 2016-11-29 12:25:36.962 WARN  (qtp1528637575-14) [   x:collection1]
>> o.a.s.h.a.LukeRequestHandler Error getting file length for [segments_c]
>> java.nio.file.NoSuchFileException: data/index/segments_c

If your Solr instance is actually stopping, I would suspect the OOM
script, assuming a non-windows system.  On non-windows systems, recent
versions of Solr have a script that forcibly terminates Solr in the
event of an OutOfMemoryError.  This script has its own log, which would
be in the same place as solr.log.

I've never heard of Solr actually crashing on a normally configured
system, and I'm reasonably sure that the message you've indicated is not
something that would cause a crash.  In fact, I've never seen it cause
any real issues, just the warning message.

Thanks,
Shawn



Re: HTTP Request timeout exception with Solr+Solarium on Azure

2016-11-29 Thread Shawn Heisey
On 11/29/2016 6:40 AM, tesm...@gmail.com wrote:
> Solr server is running in a Ubuntu VM on Azure. Php pages PHPSolarium are
> hosted as webapp using the same VM as for Solr server.
>
> After deployment, I am getting the following HTTP request timeout error:
>
> Fatal error: Uncaught exception 'Solarium\Exception\HttpException' with
> message 'Solr HTTP error: HTTP request failed, Connection timed out after
> 5016 milliseconds' in

Solarium is third-party software.  The Solr project was not involved
with its development at all.  If you need help with it, you'll need to
contact their community.  This is their official source repo:

https://github.com/solariumphp/solarium

That problem looks to me like there is a 5 second timeout in Solarium
for requests.  They likely have a way to increase that timeout.

We can help with problems on the server side, but unless you've really
messed with the server-side config, chances are that this is a
client-side problem.

If you're trying to reduce the query time, a common remedy is to add
more memory in the Solr server.  Here is the current state of my
knowledge about performance issues with Solr:

https://wiki.apache.org/solr/SolrPerformanceProblems

You haven't given us the Solr version or any Solr logs, which we will
need if you are trying to fix a problem on the server side.

As for best practices ... we can't give you any information on the
client side.  For the server side, in a nutshell, a reasonable answer
requires a LOT of information, and even with that information, any guess
we make might be wrong:

https://lucidworks.com/blog/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/

My general advice:  Put as much memory as you can afford in the server.

Thanks,
Shawn



Re: Broken pipe error

2016-11-29 Thread Shawn Heisey
On 11/29/2016 4:47 AM, Srinivas Kashyap wrote:
> Can somebody guide me how to resolve this issue?
>
> Some of the parameters for Tomcat set are :
>
> maxWait="15000" maxActive="1000" maxIdle="50".

A broken pipe error usually means that the TCP connection was broken,
but you didn't include enough of the stacktrace for me to say for sure.

Normally this would mean that the client disconnected before the request
completed, and that's usually due to a timeout.  The most likely timeout
to apply is a TCP socket timeout.  Since you're not running in the Jetty
that was included with Solr, I can't say for sure that it's not a
timeout issue on the server side, because I have no way of knowing how
the server is configured.

What version of Solr?  Running in Tomcat became an unsupported config
(though still possible) with the release of 5.0, which was over a year ago.

Thanks,
Shawn



Re: Using Solr CDCR with HdfsDirectoryFactory

2016-11-29 Thread Shalin Shekhar Mangar
At present, CDCR does not work with HDFS. The update log implementation is
different and incompatible at the moment. Please open a jira issue to
support cdcr under hdfs -- patches are always welcome!

On Tue, Nov 29, 2016 at 8:36 PM, ZHOU Ran (SAFRAN IDENTITY AND SECURITY) <
ran.z...@safrangroup.com> wrote:

> Hello all,
>
>
>
> could someone help on this?
>
>
>
> Best Regards
>
> Ran
>
>
>
> *Von:* ZHOU Ran (SAFRAN IDENTITY AND SECURITY)
> *Gesendet:* Freitag, 25. November 2016 15:37
> *An:* 'solr-user@lucene.apache.org' 
> *Betreff:* Using Solr CDCR with HdfsDirectoryFactory
>
>
>
> Hello
>
>
>
>
>
> Hi All,
>
>
>
> I have followed the guide „Cross Data Center Replication (CDCR)“ and get
> my source collection replicated to the target. And then I tried to use HDFS
> as storage for both Solr clusters, but failed with the following error
> message:
>
>
>
> ERROR: Failed to create collection 'collection11' due to:
> {192.168.5.95:8983_solr=org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException:Error
> from server at http://192.168.5.95:8983/solr: Error CREATEing SolrCore
> 'collection11_shard1_replica1': Unable to create core
> [collection11_shard1_replica1] Caused by: Solr instance is not configured
> with the cdcr update log.}
>
>
>
> Actually Solr with HDFS works for me. In the configuration for CDCR, there
> is one block:
>
>
>
> 
>
>   
>
> ${solr.ulog.dir:}
>
> 
>
>   
>
> 
>
>
>
>
>
> And I know that if HdfsDirectoryFactory is used, then updateHandler will
> initialized updateLog with the class HdfsUpdateLog. Is this the problem
> that CDCR does not work with HDFS? Since the updateLog cannot be
> initialized with CdcrUpdateLog?
>
>
>
> Thanks in advance for your help!
>
>
>
> Best Regards
>
>
>
> *Ran ZHOU*
> Software Engineer
>
>
> *T* +49 (0) 234 97 87 59 <+49%20234%20978759>
>
> *E* ran.z...@safrangroup.com *NEW*
>
>
>
> L-1 Identity Solutions AG
>
> Universitätsstrasse 160 I 44801 BOCHUM - GERMANY
> *www.safran-identity-security.com*
> 
>
>
> 
> 
>
> 
>
> 
>
> 
>
>
>
>
> *Managing Board:* Dr. Martin Werner (Vors. / Chairman) I Christèle Jacqz
> I Francois Rieul* Supervisory Board: *Jean-Christophe Fondeur (Vors. /
> Chairman)
>
> *Register Court:* Amtsgericht Bochum I HRB 69 54 | *UST-ID / VAT ID:* DE
> 813124378
>
>
>
> #
> " Ce courriel et les documents qui lui sont joints peuvent contenir des
> informations confidentielles, être soumis aux règlementations relatives au
> contrôle des exportations ou ayant un caractère privé. S'ils ne vous sont
> pas destinés, nous vous signalons qu'il est strictement interdit de les
> divulguer, de les reproduire ou d'en utiliser de quelque manière que ce
> soit le contenu. Toute exportation ou réexportation non autorisée est
> interdite.Si ce message vous a été transmis par erreur, merci d'en informer
> l'expéditeur et de supprimer immédiatement de votre système informatique ce
> courriel ainsi que tous les documents qui y sont attachés."
> **
> " This e-mail and any attached documents may contain confidential or
> proprietary information and may be subject to export control laws and
> regulations. If you are not the intended recipient, you are notified that
> any dissemination, copying of this e-mail and any attachments thereto or
> use of their contents by any means whatsoever is strictly prohibited.
> Unauthorized export or re-export is prohibited. If you have received this
> e-mail in error, please advise the sender immediately and delete this
> e-mail and all attached documents from your computer system."
> #
>



-- 
Regards,
Shalin Shekhar Mangar.


Re: Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread Shalin Shekhar Mangar
On Tue, Nov 29, 2016 at 10:22 PM, Kevin Risden  wrote:
> If using CloudSolrClient or another zookeeper aware client, then a request
> gets sent to Zookeeper to determine the live nodes. If indexing,
> CloudSolrClient can find the leader and send documents directly there. The
> client then uses that information to query the correct nodes directly.

The live node information is watched and cached so it is re-fetched
only when it changes. There is no direct request to ZooKeeper on
either indexing or querying unless our cached live nodes and cluster
state information has become stale.

>
> Zookeeper is not forwarding requests to Solr. The client requests from
> Zookeeper and then the client uses that information to query Solr directly.
>
> Kevin Risden
>
> On Tue, Nov 29, 2016 at 10:49 AM, John Bickerstaff > wrote:
>
>> All,
>>
>> I've thought I understood that Solr search requests are made to the Solr
>> servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide which
>> Solr server responds to requests and requests are made directly to Solr)
>>
>> My new place tells me they're sending requests to Zookeeper - and those are
>> getting sent on to Solr by Zookeeper - -- this is news to me if it's
>> true...
>>
>> Is there any documentation of exactly the role(s) played by Zookeeper in a
>> SolrCloud setup?
>>



-- 
Regards,
Shalin Shekhar Mangar.


Re: Starting SolrCloud

2016-11-29 Thread James Muerle
Hello,

Thanks for reading this, but it has been resolved. I honestly don't know
what was happening, but restarting my shell and running the exact same
commands today instead of yesterday seems to have fixed it.

Best,
James

On Mon, Nov 28, 2016 at 8:07 PM, James Muerle  wrote:

> Hello,
>
> I am very new to Solr, and I'm excited to get it up and running on amazon
> ec2 for some prototypical testing. So, I've installed solr (and java) on
> one ec2 instance, and I've installed zookeeper on another. After starting
> the zookeeper server on the default port of 2181, I run this on the solr
> instance: "opt/solr/bin/solr start -c -z ".us-
> west-2.compute.amazonaws.com/solr"", which seems to complete successfully:
>
> Archiving 1 old GC log files to /opt/solr/server/logs/archived
> Archiving 1 console log files to /opt/solr/server/logs/archived
> Rotating solr logs, keeping a max of 9 generations
> Waiting up to 180 seconds to see Solr running on port 8983 [|]
> Started Solr server on port 8983 (pid=13038). Happy searching!
>
> But then when I run "/opt/solr/bin/solr status", I get this output:
>
> Found 1 Solr nodes:
>
> Solr process 13038 running on port 8983
>
> ERROR: Failed to get system information from http://localhost:8983/solr
> due to: org.apache.http.client.ClientProtocolException: Expected JSON
> response from server but received: 
> 
> 
> Error 500 Server Error
> 
> HTTP ERROR 500
> Problem accessing /solr/admin/info/system. Reason:
> Server ErrorCaused 
> by:org.apache.solr.common.SolrException:
> Error processing the request. CoreContainer is either not initialized or
> shutting down.
> at org.apache.solr.servlet.SolrDispatchFilter.doFilter(
> SolrDispatchFilter.java:263)
> at org.apache.solr.servlet.SolrDispatchFilter.doFilter(
> SolrDispatchFilter.java:254)
> at org.eclipse.jetty.servlet.ServletHandler$CachedChain.
> doFilter(ServletHandler.java:1668)
> at org.eclipse.jetty.servlet.ServletHandler.doHandle(
> ServletHandler.java:581)
> at org.eclipse.jetty.server.handler.ScopedHandler.handle(
> ScopedHandler.java:143)
> at org.eclipse.jetty.security.SecurityHandler.handle(
> SecurityHandler.java:548)
> at org.eclipse.jetty.server.session.SessionHandler.
> doHandle(SessionHandler.java:226)
> at org.eclipse.jetty.server.handler.ContextHandler.
> doHandle(ContextHandler.java:1160)
> at org.eclipse.jetty.servlet.ServletHandler.doScope(
> ServletHandler.java:511)
> at org.eclipse.jetty.server.session.SessionHandler.
> doScope(SessionHandler.java:185)
> at org.eclipse.jetty.server.handler.ContextHandler.
> doScope(ContextHandler.java:1092)
> at org.eclipse.jetty.server.handler.ScopedHandler.handle(
> ScopedHandler.java:141)
> at org.eclipse.jetty.server.handler.ContextHandlerCollection.
> handle(ContextHandlerCollection.java:213)
> at org.eclipse.jetty.server.handler.HandlerCollection.
> handle(HandlerCollection.java:119)
> at org.eclipse.jetty.server.handler.HandlerWrapper.handle(
> HandlerWrapper.java:134)
> at org.eclipse.jetty.server.Server.handle(Server.java:518)
> at org.eclipse.jetty.server.HttpChannel.handle(
> HttpChannel.java:308)
> at org.eclipse.jetty.server.HttpConnection.onFillable(
> HttpConnection.java:244)
> at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(
> AbstractConnection.java:273)
> at org.eclipse.jetty.io.FillInterest.fillable(
> FillInterest.java:95)
> at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(
> SelectChannelEndPoint.java:93)
> at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.
> produceAndRun(ExecuteProduceConsume.java:246)
> at org.eclipse.jetty.util.thread.strategy.
> ExecuteProduceConsume.run(ExecuteProduceConsume.java:156)
> at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(
> QueuedThreadPool.java:654)
> at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(
> QueuedThreadPool.java:572)
> at java.lang.Thread.run(Thread.java:745)
> 
>
> 
> 
>
> Typically, this indicates a problem with the Solr server; check the Solr
> server logs for more information.
>
>
> I don't quite understand what things could be causing this problem, so I'm
> really at a loss at the moment. If you need any additional information, I'd
> be glad to provide it.
>
> Thanks for reading!
> James
>


Re: Solr 6.3.0 SQL question

2016-11-29 Thread Joel Bernstein
Let's break this down:

'stmt=SELECT TextSize from main LIMIT 10' fails
This fails because CloudSolrStream does not currently support aliases. I
believe this is fixed in 6.4

'stmt=SELECT avg(TextSize) from UNCLASS' fails
This surprises me. I read through the StatsStream and don't see any reason
why an uppercase collection should fail. Under the covers the StatsStream
simply uses the CloudSolrClient to make the call. I'll need to investigate
this further.

'stmt=SELECT like_count, DocumentId from main where like_count>0' fails
Again the issue with CloudSolrStream not supporting aliases.












Joel Bernstein
http://joelsolr.blogspot.com/

On Tue, Nov 29, 2016 at 11:48 AM, Joe Obernberger <
joseph.obernber...@gmail.com> wrote:

> Just some data points.
> main is an alias for the collection UNCLASS.
>
> 'stmt=SELECT TextSize from main LIMIT 10' fails
> 'stmt=SELECT TextSize from UNCLASS LIMIT 10' succeeds
> 'stmt=SELECT avg(TextSize) from UNCLASS' fails
> 'stmt=SELECT avg(TextSize) from main' succeeds
> 'stmt=SELECT like_count, DocumentId from UNCLASS where like_count>0'
> succeeds
> 'stmt=SELECT like_count, DocumentId from main where like_count>0' fails
>
> Hope that helps.
>
>
> -Joe
>
> On 11/29/2016 9:11 AM, Joel Bernstein wrote:
>
>> I'll take a look at the StatsStream and see what the issue is.
>>
>> Joel Bernstein
>> http://joelsolr.blogspot.com/
>>
>> On Mon, Nov 28, 2016 at 8:32 PM, Damien Kamerman 
>> wrote:
>>
>> Aggregated selects only work with lower-case collection names (and no
>>> dashes). (Bug in StatsStream I think)
>>>
>>> I assume 'SOLR-9077 Streaming expressions should support collection
>>> alias'
>>> which is fixed in 6.4 is a work around.
>>>
>>> On 29 November 2016 at 08:29, Kevin Risden 
>>> wrote:
>>>
>>> Is there a longer error/stack trace in your Solr server logs? I wonder if
 the real error is being masked.

 Kevin Risden

 On Mon, Nov 28, 2016 at 3:24 PM, Joe Obernberger <
 joseph.obernber...@gmail.com> wrote:

 I'm running this query:
>
> curl --data-urlencode 'stmt=SELECT avg(TextSize) from UNCLASS'
> http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce
>
> The error that I get back is:
>
> {"result-set":{"docs":[
> {"EXCEPTION":"org.apache.solr.common.SolrException: Collection not
>
 found:

> unclass","EOF":true,"RESPONSE_TIME":2}]}}
>
> TextSize is defined as:
> 
 multiValued="false"
>>>
 indexed="true" stored="true"/>
>
> This query works fine:
> curl --data-urlencode 'stmt=SELECT TextSize from UNCLASS'
> http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce
>
> Any idea what I'm doing wrong?
> Thank you!
>
> -Joe
>
>
>
>


Re: boost solr results using users behavior

2016-11-29 Thread John Bickerstaff
This slideshow / presentation may give you some idea of the complexity
involved...  No, nothing like this in Solr itself.

At least one approach is to mine your logs for user behavior and use that
information as a starting point for either an external machine learning
piece, or for just fine-tuning pre and post processing via smart people
writing software...

http://www.slideshare.net/treygrainger/leveraging-lucenesolr-as-a-knowledge-graph-and-intent-engine

On Tue, Nov 29, 2016 at 9:55 AM,  wrote:

> Is there any way Solr Engine should “learn” based on user behaviours and
> offer intelligent suggestions
>
> How can it be implemented ?
>
>
>
>
> This e-mail and any files transmitted with it are for the sole use of the
> intended recipient(s) and may contain confidential and privileged
> information. If you are not the intended recipient(s), please reply to the
> sender and destroy all copies of the original message. Any unauthorized
> review, use, disclosure, dissemination, forwarding, printing or copying of
> this email, and/or any action taken in reliance on the contents of this
> e-mail is strictly prohibited and may be unlawful. Where permitted by
> applicable law, this e-mail and other e-mail communications sent to and
> from Cognizant e-mail addresses may be monitored.
>


Re: Highlighting, offsets -- external doc store

2016-11-29 Thread John Bickerstaff
Thank you Simon.

On Tue, Nov 29, 2016 at 11:25 AM, simon  wrote:

> You might want to take a look at
> https://issues.apache.org/jira/browse/SOLR-4722
> ( 'highlighter which generates a list of query term positions'). We used it
> a while back and doesn't appear to have been used in any Solr > 4.10)
>
> -Simon
>
> On Tue, Nov 29, 2016 at 11:43 AM, John Bickerstaff <
> j...@johnbickerstaff.com
> > wrote:
>
> > All,
> >
> > One of the questions I've been asked to answer / prove out is around the
> > question of highlighting query matches in responses.
> >
> > BTW - One assumption I'm making is that highlighting is basically a
> > function of storing offsets for terms / tokens at index time.  If that's
> > not right, I'd be grateful for pointers in the right direction.
> >
> > My underlying need is to get highlighting on search term matches for
> > returned documents.  I need to choose between doing this in Solr and
> using
> > an external document store, so I'm interested in whether Solr can provide
> > the doc store with the information necessary to identify which section(s)
> > of the doc to highlight in a query response...
> >
> > A few questions:
> >
> > 1. This page doesn't say a lot about how things work - is there somewhere
> > with more information on dealing with offsets and highlighting? On
> offsets
> > and how they're handled?
> > https://cwiki.apache.org/confluence/display/solr/Highlighting
> >
> > 2. Can I return offset information with a query response or is that
> > internal only?  If yes, can I return offset info if I have NOT stored the
> > data in Solr but indexed only?
> >
> > (Explanation: Currently my project is considering indexing only and
> storing
> > the entire text elsewhere -- using Solr to return only doc ID's for
> > searches.  If Solr could also return offsets, these could be used in
> > processing the text stored elsewhere to provide highlighting)
> >
> > 3. Do I assume correctly that in order for Solr highlighting to work
> > correctly, the text MUST also be stored in Solr (I.E. not indexed only,
> but
> > stored=true)
> >
> > Many thanks...
> >
>


Re: Highlighting, offsets -- external doc store

2016-11-29 Thread simon
You might want to take a look at
https://issues.apache.org/jira/browse/SOLR-4722
( 'highlighter which generates a list of query term positions'). We used it
a while back and doesn't appear to have been used in any Solr > 4.10)

-Simon

On Tue, Nov 29, 2016 at 11:43 AM, John Bickerstaff  wrote:

> All,
>
> One of the questions I've been asked to answer / prove out is around the
> question of highlighting query matches in responses.
>
> BTW - One assumption I'm making is that highlighting is basically a
> function of storing offsets for terms / tokens at index time.  If that's
> not right, I'd be grateful for pointers in the right direction.
>
> My underlying need is to get highlighting on search term matches for
> returned documents.  I need to choose between doing this in Solr and using
> an external document store, so I'm interested in whether Solr can provide
> the doc store with the information necessary to identify which section(s)
> of the doc to highlight in a query response...
>
> A few questions:
>
> 1. This page doesn't say a lot about how things work - is there somewhere
> with more information on dealing with offsets and highlighting? On offsets
> and how they're handled?
> https://cwiki.apache.org/confluence/display/solr/Highlighting
>
> 2. Can I return offset information with a query response or is that
> internal only?  If yes, can I return offset info if I have NOT stored the
> data in Solr but indexed only?
>
> (Explanation: Currently my project is considering indexing only and storing
> the entire text elsewhere -- using Solr to return only doc ID's for
> searches.  If Solr could also return offsets, these could be used in
> processing the text stored elsewhere to provide highlighting)
>
> 3. Do I assume correctly that in order for Solr highlighting to work
> correctly, the text MUST also be stored in Solr (I.E. not indexed only, but
> stored=true)
>
> Many thanks...
>


Re: Issue in 5.5.3 with lucene localParams with type

2016-11-29 Thread Mikhail Khludnev
Hello,

{!lucene%20type=payloadQueryParser just seems doubtful to me. The token
after ! comes as a shortcut for type=.. param. So, the behavior in case of
specifying type twice might be undetermined.
Regarding wrong output, it might help to check how it behaves as part of q,
check explain, and probably debug.

On Mon, Nov 28, 2016 at 11:15 PM, William Bell  wrote:

> In Solr 5.4.1 this used to work:
>
> fl={!lucene%20type=payloadQueryParser v='hosp_quality_spec_boost:PS628'}
>
> 24.227154,
>
> The only way I can get payloads to work is:
>
> fl={!payloadQueryParser v='hosp_quality_spec_boost:PS628'}
>
> 0.125,
>
> But the right values only come back in #2. It should be .125.
>
> Why is type not working anymore for queryParser?
>
> 
>  class="hg.payload.PayloadQParserPlugin"/>
>
>
>
> --
> Bill Bell
> billnb...@gmail.com
> cell 720-256-8076
>



-- 
Sincerely yours
Mikhail Khludnev


Re: Highlighting, offsets -- external doc store

2016-11-29 Thread John Bickerstaff
Beautiful!

Thank you all - that is exactly what I needed to be sure where I stood on
this before going into a meeting today.

On Tue, Nov 29, 2016 at 11:03 AM, Kevin Risden 
wrote:

> For #2 you might be able to get away with the following:
>
> https://cwiki.apache.org/confluence/display/solr/The+Term+Vector+Component
>
> The Term Vector component can return offsets and positions. Not sure how
> useful they would be to you, but at least is a starting point. I'm assuming
> this requires only termVecotrs and termPositions and won't require stored
> to be true.
>
> Kevin Risden
>
> On Tue, Nov 29, 2016 at 12:00 PM, Kevin Risden 
> wrote:
>
> > For #3 specifically, I've always found this page useful:
> >
> > https://cwiki.apache.org/confluence/display/solr/Field+
> > Properties+by+Use+Case
> >
> > It lists out what properties are necessary on each field based on a use
> > case.
> >
> > Kevin Risden
> >
> > On Tue, Nov 29, 2016 at 11:49 AM, Erick Erickson <
> erickerick...@gmail.com>
> > wrote:
> >
> >> (1) No that I have readily at hand. And to make it
> >> worse, there's the UnifiedHighlighter coming out soon
> >>
> >> I don't think there's a good way for (2).
> >>
> >> for (3) at least yes. The reason is simple. For analyzed text,
> >> the only thing in the index is what's made it through the
> >> analysis chains. So stopwords are missing. Stemming
> >> has been done. You could even have put a phonetic filter
> >> in there and have terms like ARDT KNTR which would
> >> be...er...not very useful to show the end user so the original
> >> text must be available.
> >>
> >>
> >>
> >>
> >> Not much help...
> >> Erick
> >>
> >> On Tue, Nov 29, 2016 at 8:43 AM, John Bickerstaff
> >>  wrote:
> >> > All,
> >> >
> >> > One of the questions I've been asked to answer / prove out is around
> the
> >> > question of highlighting query matches in responses.
> >> >
> >> > BTW - One assumption I'm making is that highlighting is basically a
> >> > function of storing offsets for terms / tokens at index time.  If
> that's
> >> > not right, I'd be grateful for pointers in the right direction.
> >> >
> >> > My underlying need is to get highlighting on search term matches for
> >> > returned documents.  I need to choose between doing this in Solr and
> >> using
> >> > an external document store, so I'm interested in whether Solr can
> >> provide
> >> > the doc store with the information necessary to identify which
> >> section(s)
> >> > of the doc to highlight in a query response...
> >> >
> >> > A few questions:
> >> >
> >> > 1. This page doesn't say a lot about how things work - is there
> >> somewhere
> >> > with more information on dealing with offsets and highlighting? On
> >> offsets
> >> > and how they're handled?
> >> > https://cwiki.apache.org/confluence/display/solr/Highlighting
> >> >
> >> > 2. Can I return offset information with a query response or is that
> >> > internal only?  If yes, can I return offset info if I have NOT stored
> >> the
> >> > data in Solr but indexed only?
> >> >
> >> > (Explanation: Currently my project is considering indexing only and
> >> storing
> >> > the entire text elsewhere -- using Solr to return only doc ID's for
> >> > searches.  If Solr could also return offsets, these could be used in
> >> > processing the text stored elsewhere to provide highlighting)
> >> >
> >> > 3. Do I assume correctly that in order for Solr highlighting to work
> >> > correctly, the text MUST also be stored in Solr (I.E. not indexed
> only,
> >> but
> >> > stored=true)
> >> >
> >> > Many thanks...
> >>
> >
> >
>


Re: Highlighting, offsets -- external doc store

2016-11-29 Thread Kevin Risden
For #2 you might be able to get away with the following:

https://cwiki.apache.org/confluence/display/solr/The+Term+Vector+Component

The Term Vector component can return offsets and positions. Not sure how
useful they would be to you, but at least is a starting point. I'm assuming
this requires only termVecotrs and termPositions and won't require stored
to be true.

Kevin Risden

On Tue, Nov 29, 2016 at 12:00 PM, Kevin Risden 
wrote:

> For #3 specifically, I've always found this page useful:
>
> https://cwiki.apache.org/confluence/display/solr/Field+
> Properties+by+Use+Case
>
> It lists out what properties are necessary on each field based on a use
> case.
>
> Kevin Risden
>
> On Tue, Nov 29, 2016 at 11:49 AM, Erick Erickson 
> wrote:
>
>> (1) No that I have readily at hand. And to make it
>> worse, there's the UnifiedHighlighter coming out soon
>>
>> I don't think there's a good way for (2).
>>
>> for (3) at least yes. The reason is simple. For analyzed text,
>> the only thing in the index is what's made it through the
>> analysis chains. So stopwords are missing. Stemming
>> has been done. You could even have put a phonetic filter
>> in there and have terms like ARDT KNTR which would
>> be...er...not very useful to show the end user so the original
>> text must be available.
>>
>>
>>
>>
>> Not much help...
>> Erick
>>
>> On Tue, Nov 29, 2016 at 8:43 AM, John Bickerstaff
>>  wrote:
>> > All,
>> >
>> > One of the questions I've been asked to answer / prove out is around the
>> > question of highlighting query matches in responses.
>> >
>> > BTW - One assumption I'm making is that highlighting is basically a
>> > function of storing offsets for terms / tokens at index time.  If that's
>> > not right, I'd be grateful for pointers in the right direction.
>> >
>> > My underlying need is to get highlighting on search term matches for
>> > returned documents.  I need to choose between doing this in Solr and
>> using
>> > an external document store, so I'm interested in whether Solr can
>> provide
>> > the doc store with the information necessary to identify which
>> section(s)
>> > of the doc to highlight in a query response...
>> >
>> > A few questions:
>> >
>> > 1. This page doesn't say a lot about how things work - is there
>> somewhere
>> > with more information on dealing with offsets and highlighting? On
>> offsets
>> > and how they're handled?
>> > https://cwiki.apache.org/confluence/display/solr/Highlighting
>> >
>> > 2. Can I return offset information with a query response or is that
>> > internal only?  If yes, can I return offset info if I have NOT stored
>> the
>> > data in Solr but indexed only?
>> >
>> > (Explanation: Currently my project is considering indexing only and
>> storing
>> > the entire text elsewhere -- using Solr to return only doc ID's for
>> > searches.  If Solr could also return offsets, these could be used in
>> > processing the text stored elsewhere to provide highlighting)
>> >
>> > 3. Do I assume correctly that in order for Solr highlighting to work
>> > correctly, the text MUST also be stored in Solr (I.E. not indexed only,
>> but
>> > stored=true)
>> >
>> > Many thanks...
>>
>
>


Re: Highlighting, offsets -- external doc store

2016-11-29 Thread Kevin Risden
For #3 specifically, I've always found this page useful:

https://cwiki.apache.org/confluence/display/solr/Field+Properties+by+Use+Case

It lists out what properties are necessary on each field based on a use
case.

Kevin Risden

On Tue, Nov 29, 2016 at 11:49 AM, Erick Erickson 
wrote:

> (1) No that I have readily at hand. And to make it
> worse, there's the UnifiedHighlighter coming out soon
>
> I don't think there's a good way for (2).
>
> for (3) at least yes. The reason is simple. For analyzed text,
> the only thing in the index is what's made it through the
> analysis chains. So stopwords are missing. Stemming
> has been done. You could even have put a phonetic filter
> in there and have terms like ARDT KNTR which would
> be...er...not very useful to show the end user so the original
> text must be available.
>
>
>
>
> Not much help...
> Erick
>
> On Tue, Nov 29, 2016 at 8:43 AM, John Bickerstaff
>  wrote:
> > All,
> >
> > One of the questions I've been asked to answer / prove out is around the
> > question of highlighting query matches in responses.
> >
> > BTW - One assumption I'm making is that highlighting is basically a
> > function of storing offsets for terms / tokens at index time.  If that's
> > not right, I'd be grateful for pointers in the right direction.
> >
> > My underlying need is to get highlighting on search term matches for
> > returned documents.  I need to choose between doing this in Solr and
> using
> > an external document store, so I'm interested in whether Solr can provide
> > the doc store with the information necessary to identify which section(s)
> > of the doc to highlight in a query response...
> >
> > A few questions:
> >
> > 1. This page doesn't say a lot about how things work - is there somewhere
> > with more information on dealing with offsets and highlighting? On
> offsets
> > and how they're handled?
> > https://cwiki.apache.org/confluence/display/solr/Highlighting
> >
> > 2. Can I return offset information with a query response or is that
> > internal only?  If yes, can I return offset info if I have NOT stored the
> > data in Solr but indexed only?
> >
> > (Explanation: Currently my project is considering indexing only and
> storing
> > the entire text elsewhere -- using Solr to return only doc ID's for
> > searches.  If Solr could also return offsets, these could be used in
> > processing the text stored elsewhere to provide highlighting)
> >
> > 3. Do I assume correctly that in order for Solr highlighting to work
> > correctly, the text MUST also be stored in Solr (I.E. not indexed only,
> but
> > stored=true)
> >
> > Many thanks...
>


Re: Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread John Bickerstaff
I think they believe this because of what you say above -- they are
misinterpreting the call for topology as a handoff where zookeeper "does
the rest"...

This info will allow me to straighten out that misunderstanding...  Thanks
all!

On Tue, Nov 29, 2016 at 10:49 AM, Walter Underwood 
wrote:

> This is easy to prove. Shut down Zookeeper, then send search requests to
> different hosts in the Solr Cloud cluster. If they work, then the requests
> are
> not going through Zookeeper.
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
>
>
> > On Nov 29, 2016, at 9:44 AM, Erick Erickson 
> wrote:
> >
> > bq: My new place tells me they're sending requests to Zookeeper - and
> those are
> > getting sent on to Solr by Zookeeper - -- this is news to me if it's
> true...
> >
> > Have you seen the code? Because if they're using SolrJ, then they do,
> indeed
> > connect to Zookeeper (i.e. CloudSolrClient takes a ZK ensemble) which
> they may
> > be misinterpreting as sending request to ZK that are forwarded to Solr.
> That is
> > _not_ what's happening. The ZK connection allows SolrJ to get the
> > topology from ZK
> > but then CloudSolrClient sends requests to Solr nodes directly based
> > on the state.json
> > fetched from ZK. And it should cache that.
> >
> > There have been JIRAs about reducing the number of times any particular
> client
> > requests state.json from ZK, but there's _no_ code on ZK that forwards
> requests.
> > ZK is just a bookkeeper, for it to "do the right thing" with a Solr
> > request it'd have to
> > know about Solr, i.e. we'd have to distribute some custom ZK code
> > which we don't.
> >
> > It would help if you provided the reason they think this. Code like
> above?
> > Network traffic? They've monitored the ZK activity and it's high? I.e.
> > is there any evidence of this?
> >
> > Best,
> > Erick
> >
> >
> >
> > On Tue, Nov 29, 2016 at 8:52 AM, Kevin Risden 
> wrote:
> >> If using CloudSolrClient or another zookeeper aware client, then a
> request
> >> gets sent to Zookeeper to determine the live nodes. If indexing,
> >> CloudSolrClient can find the leader and send documents directly there.
> The
> >> client then uses that information to query the correct nodes directly.
> >>
> >> Zookeeper is not forwarding requests to Solr. The client requests from
> >> Zookeeper and then the client uses that information to query Solr
> directly.
> >>
> >> Kevin Risden
> >>
> >> On Tue, Nov 29, 2016 at 10:49 AM, John Bickerstaff <
> j...@johnbickerstaff.com
> >>> wrote:
> >>
> >>> All,
> >>>
> >>> I've thought I understood that Solr search requests are made to the
> Solr
> >>> servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide
> which
> >>> Solr server responds to requests and requests are made directly to
> Solr)
> >>>
> >>> My new place tells me they're sending requests to Zookeeper - and
> those are
> >>> getting sent on to Solr by Zookeeper - -- this is news to me if it's
> >>> true...
> >>>
> >>> Is there any documentation of exactly the role(s) played by Zookeeper
> in a
> >>> SolrCloud setup?
> >>>
>
>


Re: Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread Walter Underwood
This is easy to prove. Shut down Zookeeper, then send search requests to
different hosts in the Solr Cloud cluster. If they work, then the requests are
not going through Zookeeper.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)


> On Nov 29, 2016, at 9:44 AM, Erick Erickson  wrote:
> 
> bq: My new place tells me they're sending requests to Zookeeper - and those 
> are
> getting sent on to Solr by Zookeeper - -- this is news to me if it's true...
> 
> Have you seen the code? Because if they're using SolrJ, then they do, indeed
> connect to Zookeeper (i.e. CloudSolrClient takes a ZK ensemble) which they may
> be misinterpreting as sending request to ZK that are forwarded to Solr. That 
> is
> _not_ what's happening. The ZK connection allows SolrJ to get the
> topology from ZK
> but then CloudSolrClient sends requests to Solr nodes directly based
> on the state.json
> fetched from ZK. And it should cache that.
> 
> There have been JIRAs about reducing the number of times any particular client
> requests state.json from ZK, but there's _no_ code on ZK that forwards 
> requests.
> ZK is just a bookkeeper, for it to "do the right thing" with a Solr
> request it'd have to
> know about Solr, i.e. we'd have to distribute some custom ZK code
> which we don't.
> 
> It would help if you provided the reason they think this. Code like above?
> Network traffic? They've monitored the ZK activity and it's high? I.e.
> is there any evidence of this?
> 
> Best,
> Erick
> 
> 
> 
> On Tue, Nov 29, 2016 at 8:52 AM, Kevin Risden  
> wrote:
>> If using CloudSolrClient or another zookeeper aware client, then a request
>> gets sent to Zookeeper to determine the live nodes. If indexing,
>> CloudSolrClient can find the leader and send documents directly there. The
>> client then uses that information to query the correct nodes directly.
>> 
>> Zookeeper is not forwarding requests to Solr. The client requests from
>> Zookeeper and then the client uses that information to query Solr directly.
>> 
>> Kevin Risden
>> 
>> On Tue, Nov 29, 2016 at 10:49 AM, John Bickerstaff >> wrote:
>> 
>>> All,
>>> 
>>> I've thought I understood that Solr search requests are made to the Solr
>>> servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide which
>>> Solr server responds to requests and requests are made directly to Solr)
>>> 
>>> My new place tells me they're sending requests to Zookeeper - and those are
>>> getting sent on to Solr by Zookeeper - -- this is news to me if it's
>>> true...
>>> 
>>> Is there any documentation of exactly the role(s) played by Zookeeper in a
>>> SolrCloud setup?
>>> 



Re: Highlighting, offsets -- external doc store

2016-11-29 Thread Erick Erickson
(1) No that I have readily at hand. And to make it
worse, there's the UnifiedHighlighter coming out soon

I don't think there's a good way for (2).

for (3) at least yes. The reason is simple. For analyzed text,
the only thing in the index is what's made it through the
analysis chains. So stopwords are missing. Stemming
has been done. You could even have put a phonetic filter
in there and have terms like ARDT KNTR which would
be...er...not very useful to show the end user so the original
text must be available.




Not much help...
Erick

On Tue, Nov 29, 2016 at 8:43 AM, John Bickerstaff
 wrote:
> All,
>
> One of the questions I've been asked to answer / prove out is around the
> question of highlighting query matches in responses.
>
> BTW - One assumption I'm making is that highlighting is basically a
> function of storing offsets for terms / tokens at index time.  If that's
> not right, I'd be grateful for pointers in the right direction.
>
> My underlying need is to get highlighting on search term matches for
> returned documents.  I need to choose between doing this in Solr and using
> an external document store, so I'm interested in whether Solr can provide
> the doc store with the information necessary to identify which section(s)
> of the doc to highlight in a query response...
>
> A few questions:
>
> 1. This page doesn't say a lot about how things work - is there somewhere
> with more information on dealing with offsets and highlighting? On offsets
> and how they're handled?
> https://cwiki.apache.org/confluence/display/solr/Highlighting
>
> 2. Can I return offset information with a query response or is that
> internal only?  If yes, can I return offset info if I have NOT stored the
> data in Solr but indexed only?
>
> (Explanation: Currently my project is considering indexing only and storing
> the entire text elsewhere -- using Solr to return only doc ID's for
> searches.  If Solr could also return offsets, these could be used in
> processing the text stored elsewhere to provide highlighting)
>
> 3. Do I assume correctly that in order for Solr highlighting to work
> correctly, the text MUST also be stored in Solr (I.E. not indexed only, but
> stored=true)
>
> Many thanks...


Re: Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread Erick Erickson
bq: My new place tells me they're sending requests to Zookeeper - and those are
getting sent on to Solr by Zookeeper - -- this is news to me if it's true...

Have you seen the code? Because if they're using SolrJ, then they do, indeed
connect to Zookeeper (i.e. CloudSolrClient takes a ZK ensemble) which they may
be misinterpreting as sending request to ZK that are forwarded to Solr. That is
_not_ what's happening. The ZK connection allows SolrJ to get the
topology from ZK
but then CloudSolrClient sends requests to Solr nodes directly based
on the state.json
fetched from ZK. And it should cache that.

There have been JIRAs about reducing the number of times any particular client
requests state.json from ZK, but there's _no_ code on ZK that forwards requests.
ZK is just a bookkeeper, for it to "do the right thing" with a Solr
request it'd have to
know about Solr, i.e. we'd have to distribute some custom ZK code
which we don't.

It would help if you provided the reason they think this. Code like above?
Network traffic? They've monitored the ZK activity and it's high? I.e.
is there any evidence of this?

Best,
Erick



On Tue, Nov 29, 2016 at 8:52 AM, Kevin Risden  wrote:
> If using CloudSolrClient or another zookeeper aware client, then a request
> gets sent to Zookeeper to determine the live nodes. If indexing,
> CloudSolrClient can find the leader and send documents directly there. The
> client then uses that information to query the correct nodes directly.
>
> Zookeeper is not forwarding requests to Solr. The client requests from
> Zookeeper and then the client uses that information to query Solr directly.
>
> Kevin Risden
>
> On Tue, Nov 29, 2016 at 10:49 AM, John Bickerstaff > wrote:
>
>> All,
>>
>> I've thought I understood that Solr search requests are made to the Solr
>> servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide which
>> Solr server responds to requests and requests are made directly to Solr)
>>
>> My new place tells me they're sending requests to Zookeeper - and those are
>> getting sent on to Solr by Zookeeper - -- this is news to me if it's
>> true...
>>
>> Is there any documentation of exactly the role(s) played by Zookeeper in a
>> SolrCloud setup?
>>


boost solr results using users behavior

2016-11-29 Thread Shanku.Vijay
Is there any way Solr Engine should “learn” based on user behaviours and offer 
intelligent suggestions

How can it be implemented ?




This e-mail and any files transmitted with it are for the sole use of the 
intended recipient(s) and may contain confidential and privileged information. 
If you are not the intended recipient(s), please reply to the sender and 
destroy all copies of the original message. Any unauthorized review, use, 
disclosure, dissemination, forwarding, printing or copying of this email, 
and/or any action taken in reliance on the contents of this e-mail is strictly 
prohibited and may be unlawful. Where permitted by applicable law, this e-mail 
and other e-mail communications sent to and from Cognizant e-mail addresses may 
be monitored.


Re: Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread Kevin Risden
If using CloudSolrClient or another zookeeper aware client, then a request
gets sent to Zookeeper to determine the live nodes. If indexing,
CloudSolrClient can find the leader and send documents directly there. The
client then uses that information to query the correct nodes directly.

Zookeeper is not forwarding requests to Solr. The client requests from
Zookeeper and then the client uses that information to query Solr directly.

Kevin Risden

On Tue, Nov 29, 2016 at 10:49 AM, John Bickerstaff  wrote:

> All,
>
> I've thought I understood that Solr search requests are made to the Solr
> servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide which
> Solr server responds to requests and requests are made directly to Solr)
>
> My new place tells me they're sending requests to Zookeeper - and those are
> getting sent on to Solr by Zookeeper - -- this is news to me if it's
> true...
>
> Is there any documentation of exactly the role(s) played by Zookeeper in a
> SolrCloud setup?
>


Documentation of Zookeeper's specific roles and functions in Solr Cloud?

2016-11-29 Thread John Bickerstaff
All,

I've thought I understood that Solr search requests are made to the Solr
servers and NOT Zookeeper directly.  (I.E. Zookeeper doesn't decide which
Solr server responds to requests and requests are made directly to Solr)

My new place tells me they're sending requests to Zookeeper - and those are
getting sent on to Solr by Zookeeper - -- this is news to me if it's true...

Is there any documentation of exactly the role(s) played by Zookeeper in a
SolrCloud setup?


Re: Solr 6.3.0 SQL question

2016-11-29 Thread Joe Obernberger

Just some data points.
main is an alias for the collection UNCLASS.

'stmt=SELECT TextSize from main LIMIT 10' fails
'stmt=SELECT TextSize from UNCLASS LIMIT 10' succeeds
'stmt=SELECT avg(TextSize) from UNCLASS' fails
'stmt=SELECT avg(TextSize) from main' succeeds
'stmt=SELECT like_count, DocumentId from UNCLASS where like_count>0' 
succeeds

'stmt=SELECT like_count, DocumentId from main where like_count>0' fails

Hope that helps.


-Joe

On 11/29/2016 9:11 AM, Joel Bernstein wrote:

I'll take a look at the StatsStream and see what the issue is.

Joel Bernstein
http://joelsolr.blogspot.com/

On Mon, Nov 28, 2016 at 8:32 PM, Damien Kamerman  wrote:


Aggregated selects only work with lower-case collection names (and no
dashes). (Bug in StatsStream I think)

I assume 'SOLR-9077 Streaming expressions should support collection alias'
which is fixed in 6.4 is a work around.

On 29 November 2016 at 08:29, Kevin Risden 
wrote:


Is there a longer error/stack trace in your Solr server logs? I wonder if
the real error is being masked.

Kevin Risden

On Mon, Nov 28, 2016 at 3:24 PM, Joe Obernberger <
joseph.obernber...@gmail.com> wrote:


I'm running this query:

curl --data-urlencode 'stmt=SELECT avg(TextSize) from UNCLASS'
http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce

The error that I get back is:

{"result-set":{"docs":[
{"EXCEPTION":"org.apache.solr.common.SolrException: Collection not

found:

unclass","EOF":true,"RESPONSE_TIME":2}]}}

TextSize is defined as:

multiValued="false"

indexed="true" stored="true"/>

This query works fine:
curl --data-urlencode 'stmt=SELECT TextSize from UNCLASS'
http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce

Any idea what I'm doing wrong?
Thank you!

-Joe






Highlighting, offsets -- external doc store

2016-11-29 Thread John Bickerstaff
All,

One of the questions I've been asked to answer / prove out is around the
question of highlighting query matches in responses.

BTW - One assumption I'm making is that highlighting is basically a
function of storing offsets for terms / tokens at index time.  If that's
not right, I'd be grateful for pointers in the right direction.

My underlying need is to get highlighting on search term matches for
returned documents.  I need to choose between doing this in Solr and using
an external document store, so I'm interested in whether Solr can provide
the doc store with the information necessary to identify which section(s)
of the doc to highlight in a query response...

A few questions:

1. This page doesn't say a lot about how things work - is there somewhere
with more information on dealing with offsets and highlighting? On offsets
and how they're handled?
https://cwiki.apache.org/confluence/display/solr/Highlighting

2. Can I return offset information with a query response or is that
internal only?  If yes, can I return offset info if I have NOT stored the
data in Solr but indexed only?

(Explanation: Currently my project is considering indexing only and storing
the entire text elsewhere -- using Solr to return only doc ID's for
searches.  If Solr could also return offsets, these could be used in
processing the text stored elsewhere to provide highlighting)

3. Do I assume correctly that in order for Solr highlighting to work
correctly, the text MUST also be stored in Solr (I.E. not indexed only, but
stored=true)

Many thanks...


Re: LukeRequestHandler Error getting file length for [segments_1l]

2016-11-29 Thread halis Yılboğa
it is not normal to get that many error actually. Main problem should be
from your index. It seems to me your index is corrupted.

29 Kas 2016 Sal, 14:40 tarihinde, Furkan KAMACI 
şunu yazdı:

> On the other hand, my Solr instance stops frequently due to such errors:
>
> 2016-11-29 12:25:36.962 WARN  (qtp1528637575-14) [   x:collection1]
> o.a.s.h.a.LukeRequestHandler Error getting file length for [segments_c]
> java.nio.file.NoSuchFileException: data/index/segments_c
> at sun.nio.fs.UnixException.translateToIOException(UnixException.java:86)
> at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102)
> at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107)
> at
>
> sun.nio.fs.UnixFileAttributeViews$Basic.readAttributes(UnixFileAttributeViews.java:55)
> at
>
> sun.nio.fs.UnixFileSystemProvider.readAttributes(UnixFileSystemProvider.java:144)
> at java.nio.file.Files.readAttributes(Files.java:1737)
> at java.nio.file.Files.size(Files.java:2332)
> at org.apache.lucene.store.FSDirectory.fileLength(FSDirectory.java:243)
> at
>
> org.apache.lucene.store.NRTCachingDirectory.fileLength(NRTCachingDirectory.java:128)
> at
>
> org.apache.solr.handler.admin.LukeRequestHandler.getFileLength(LukeRequestHandler.java:598)
> at
>
> org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo(LukeRequestHandler.java:586)
> at
>
> org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:137)
> at
>
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:153)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2213)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:654)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:303)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254)
> at
>
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1668)
> at
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581)
> at
>
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
> at
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)
> at
>
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)
> at
>
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1160)
> at
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511)
> at
>
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
> at
>
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1092)
> at
>
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
> at
>
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)
> at
>
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119)
> at
>
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
> at org.eclipse.jetty.server.Server.handle(Server.java:518)
> at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:308)
> at
> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:244)
> at
> org.eclipse.jetty.io
> .AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273)
> at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95)
> at
> org.eclipse.jetty.io
> .SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
> at
>
> org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceAndRun(ExecuteProduceConsume.java:246)
> at
>
> org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:156)
> at
>
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:654)
> at
>
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:572)
> at java.lang.Thread.run(Thread.java:745)
> 2016-11-29 12:25:36.964 INFO  (qtp1528637575-35) [   x:collection1]
> o.a.s.c.S.Request [collection1]  webapp=/solr path=/replication
> params={wt=json=details&_=1480422336954} status=0 QTime=4
> 2016-11-29 12:25:36.967 INFO  (qtp1528637575-14) [   x:collection1]
> o.a.s.c.S.Request [collection1]  webapp=/solr path=/admin/luke
> params={numTerms=0=index=json&_=1480422336954} status=0 QTime=6
> 2016-11-29 12:25:36.982 INFO  (qtp1528637575-18) [   x:collection1]
> o.a.s.c.S.Request [collection1]  webapp=/solr path=/admin/system
> params={wt=json&_=1480422336954} status=0 QTime=21
> 2016-11-29 12:25:36.983 INFO  (qtp1528637575-49) [   ] o.a.s.s.HttpSolrCall
> [admin] webapp=null path=/admin/info/system
> params={wt=json&_=1480422336924} status=0 QTime=23
> 2016-11-29 12:25:38.570 INFO  (qtp1528637575-37) [   ] o.a.s.s.HttpSolrCall
> [admin] webapp=null path=/admin/cores
> params={indexInfo=false=json&_=1480422336924} 

AW: Using Solr CDCR with HdfsDirectoryFactory

2016-11-29 Thread ZHOU Ran (SAFRAN IDENTITY AND SECURITY)
Hello all,

could someone help on this?

Best Regards
Ran

Von: ZHOU Ran (SAFRAN IDENTITY AND SECURITY)
Gesendet: Freitag, 25. November 2016 15:37
An: 'solr-user@lucene.apache.org' 
Betreff: Using Solr CDCR with HdfsDirectoryFactory

Hello


Hi All,

I have followed the guide "Cross Data Center Replication (CDCR)" and get my 
source collection replicated to the target. And then I tried to use HDFS as 
storage for both Solr clusters, but failed with the following error message:

ERROR: Failed to create collection 'collection11' due to: 
{192.168.5.95:8983_solr=org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException:Error
 from server at http://192.168.5.95:8983/solr: Error CREATEing SolrCore 
'collection11_shard1_replica1': Unable to create core 
[collection11_shard1_replica1] Caused by: Solr instance is not configured with 
the cdcr update log.}

Actually Solr with HDFS works for me. In the configuration for CDCR, there is 
one block:


  
${solr.ulog.dir:}

  



And I know that if HdfsDirectoryFactory is used, then updateHandler will 
initialized updateLog with the class HdfsUpdateLog. Is this the problem that 
CDCR does not work with HDFS? Since the updateLog cannot be initialized with 
CdcrUpdateLog?

Thanks in advance for your help!

Best Regards

Ran ZHOU
Software Engineer

T +49 (0) 234 97 87 59
E ran.z...@safrangroup.com NEW

L-1 Identity Solutions AG
Universitätsstrasse 160 I 44801 BOCHUM - GERMANY
www.safran-identity-security.com

[cid:image001.png@01D24A5A.83739D70]

[cid:image002.png@01D24A5A.83739D70][cid:image003.png@01D24A5A.83739D70]

[cid:image004.jpg@01D24A5A.83739D70]

[cid:image005.jpg@01D24A5A.83739D70]

[cid:image006.jpg@01D24A5A.83739D70]



[cid:image007.jpg@01D24A5A.83739D70]

Managing Board: Dr. Martin Werner (Vors. / Chairman) I Christèle Jacqz I 
Francois Rieul Supervisory Board: Jean-Christophe Fondeur (Vors. / Chairman)
Register Court: Amtsgericht Bochum I HRB 69 54 | UST-ID / VAT ID: DE 813124378

#
" Ce courriel et les documents qui lui sont joints peuvent contenir des 
informations confidentielles, être soumis aux règlementations relatives au 
contrôle des exportations ou ayant un caractère privé. S'ils ne vous sont pas 
destinés, nous vous signalons qu'il est strictement interdit de les divulguer, 
de les reproduire ou d'en utiliser de quelque manière que ce soit le contenu. 
Toute exportation ou réexportation non autorisée est interdite Si ce message 
vous a été transmis par erreur, merci d'en informer l'expéditeur et de 
supprimer immédiatement de votre système informatique ce courriel ainsi que 
tous les documents qui y sont attachés."
**
" This e-mail and any attached documents may contain confidential or 
proprietary information and may be subject to export control laws and 
regulations. If you are not the intended recipient, you are notified that any 
dissemination, copying of this e-mail and any attachments thereto or use of 
their contents by any means whatsoever is strictly prohibited. Unauthorized 
export or re-export is prohibited. If you have received this e-mail in error, 
please advise the sender immediately and delete this e-mail and all attached 
documents from your computer system."
#


Re: Solr 6.3.0 SQL question

2016-11-29 Thread Joe Obernberger

Created an alias called 'main' and now it works:

curl --data-urlencode 'stmt=SELECT avg(TextSize) from main' 
http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce


{"result-set":{"docs":[
{"avg(TextSize)":6024.222616504568},
{"EOF":true,"RESPONSE_TIME":1391}]}}

Thank you Damian and Joel!


-Joe

On 11/29/2016 9:11 AM, Joel Bernstein wrote:

I'll take a look at the StatsStream and see what the issue is.

Joel Bernstein
http://joelsolr.blogspot.com/

On Mon, Nov 28, 2016 at 8:32 PM, Damien Kamerman  wrote:


Aggregated selects only work with lower-case collection names (and no
dashes). (Bug in StatsStream I think)

I assume 'SOLR-9077 Streaming expressions should support collection alias'
which is fixed in 6.4 is a work around.

On 29 November 2016 at 08:29, Kevin Risden 
wrote:


Is there a longer error/stack trace in your Solr server logs? I wonder if
the real error is being masked.

Kevin Risden

On Mon, Nov 28, 2016 at 3:24 PM, Joe Obernberger <
joseph.obernber...@gmail.com> wrote:


I'm running this query:

curl --data-urlencode 'stmt=SELECT avg(TextSize) from UNCLASS'
http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce

The error that I get back is:

{"result-set":{"docs":[
{"EXCEPTION":"org.apache.solr.common.SolrException: Collection not

found:

unclass","EOF":true,"RESPONSE_TIME":2}]}}

TextSize is defined as:

multiValued="false"

indexed="true" stored="true"/>

This query works fine:
curl --data-urlencode 'stmt=SELECT TextSize from UNCLASS'
http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce

Any idea what I'm doing wrong?
Thank you!

-Joe






Re: Solr 6.3.0 SQL question

2016-11-29 Thread Joel Bernstein
I'll take a look at the StatsStream and see what the issue is.

Joel Bernstein
http://joelsolr.blogspot.com/

On Mon, Nov 28, 2016 at 8:32 PM, Damien Kamerman  wrote:

> Aggregated selects only work with lower-case collection names (and no
> dashes). (Bug in StatsStream I think)
>
> I assume 'SOLR-9077 Streaming expressions should support collection alias'
> which is fixed in 6.4 is a work around.
>
> On 29 November 2016 at 08:29, Kevin Risden 
> wrote:
>
> > Is there a longer error/stack trace in your Solr server logs? I wonder if
> > the real error is being masked.
> >
> > Kevin Risden
> >
> > On Mon, Nov 28, 2016 at 3:24 PM, Joe Obernberger <
> > joseph.obernber...@gmail.com> wrote:
> >
> > > I'm running this query:
> > >
> > > curl --data-urlencode 'stmt=SELECT avg(TextSize) from UNCLASS'
> > > http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce
> > >
> > > The error that I get back is:
> > >
> > > {"result-set":{"docs":[
> > > {"EXCEPTION":"org.apache.solr.common.SolrException: Collection not
> > found:
> > > unclass","EOF":true,"RESPONSE_TIME":2}]}}
> > >
> > > TextSize is defined as:
> > >  multiValued="false"
> > > indexed="true" stored="true"/>
> > >
> > > This query works fine:
> > > curl --data-urlencode 'stmt=SELECT TextSize from UNCLASS'
> > > http://cordelia:9100/solr/UNCLASS/sql?aggregationMode=map_reduce
> > >
> > > Any idea what I'm doing wrong?
> > > Thank you!
> > >
> > > -Joe
> > >
> > >
> >
>


Solr+Solarium deployment on Azure - Best practices

2016-11-29 Thread tesm...@gmail.com
Hi,

I am deploying a search engine on Azure. The following is my configuration:

Solr server is running on Ubuntu VM (hosted on Azure)
PHP web app is hosted on Azure using the same VM hosting Solr server.


Is there any best practices/approach guidelines?

I am getting the following exception:
Fatal error: Uncaught exception 'Solarium\Exception\HttpException' with
message 'Solr HTTP error: HTTP request failed, Connection timed out after
5000 milliseconds' in
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php:195
Stack trace: #0
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(92):
Solarium\Core\Client\Adapter\Curl->check('', Array, Resource id #3) #1
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(213):
Solarium\Core\Client\Adapter\Curl->getResponse(Resource id #3, false) #2
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(68):
Solarium\Core\Client\Adapter\Curl->getData(Object(Solarium\Core\Client\Request),
Object(Solarium\Core\Client\Endpoint)) #3
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Client.php(804):
Solarium\Core\Client\Adapter\Curl->execute(Object(Solarium\Core\Client\Request),
Object(Solarium\Core\Clie in
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php
on line 195



Regards,


HTTP Request timeout exception with Solr+Solarium on Azure

2016-11-29 Thread tesm...@gmail.com
Hi,




I am deploying Solr+PHPSolarium on Azure

Solr server is running in a Ubuntu VM on Azure. Php pages PHPSolarium are
hosted as webapp using the same VM as for Solr server.

After deployment, I am getting the following HTTP request timeout error:



Fatal error: Uncaught exception 'Solarium\Exception\HttpException' with
message 'Solr HTTP error: HTTP request failed, Connection timed out after
5016 milliseconds' in
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php:195
Stack trace: #0
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(92):
Solarium\Core\Client\Adapter\Curl->check('', Array, Resource id #3) #1
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(213):
Solarium\Core\Client\Adapter\Curl->getResponse(Resource id #3, false) #2
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php(68):
Solarium\Core\Client\Adapter\Curl->getData(Object(Solarium\Core\Client\Request),
Object(Solarium\Core\Client\Endpoint)) #3
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Client.php(804):
Solarium\Core\Client\Adapter\Curl->execute(Object(Solarium\Core\Client\Request),
Object(Solarium\Core\Clie in
D:\home\site\wwwroot\vendor\solarium\solarium\library\Solarium\Core\Client\Adapter\Curl.php
on line 195


Any help is much appreciated


Regards,


Broken pipe error

2016-11-29 Thread Srinivas Kashyap
Hello,

After starting the solr application and running full imports, running into this 
below error after a while:

null:org.apache.catalina.connector.ClientAbortException: java.io.IOException: 
Broken pipe
at 
org.apache.catalina.connector.OutputBuffer.realWriteBytes(OutputBuffer.java:393)
at org.apache.tomcat.util.buf.ByteChunk.flushBuffer(ByteChunk.java:426)
at org.apache.catalina.connector.OutputBuffer.doFlush(OutputBuffer.java:342)
at org.apache.catalina.connector.OutputBuffer.flush(OutputBuffer.java:317)
at 
org.apache.catalina.connector.CoyoteOutputStream.flush(CoyoteOutputStream.java:110)
at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:297)
at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141)
at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229)
at org.apache.solr.util.FastWriter.flush(FastWriter.java:137)
at 
org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter.java:710)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:430)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:201)
at 
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240)
at 
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207)
at 
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
at 
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
at 
org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
at org.apache.catalina


Can somebody guide me how to resolve this issue?

Some of the parameters for Tomcat set are :

maxWait="15000" maxActive="1000" maxIdle="50".

Thanks and Regards,
Srinivas

DISCLAIMER: 
E-mails and attachments from TradeStone Software, Inc. are confidential.
If you are not the intended recipient, please notify the sender immediately by
replying to the e-mail, and then delete it without making copies or using it
in any way. No representation is made that this email or any attachments are
free of viruses. Virus scanning is recommended and is the responsibility of
the recipient.

Re: LukeRequestHandler Error getting file length for [segments_1l]

2016-11-29 Thread Furkan KAMACI
On the other hand, my Solr instance stops frequently due to such errors:

2016-11-29 12:25:36.962 WARN  (qtp1528637575-14) [   x:collection1]
o.a.s.h.a.LukeRequestHandler Error getting file length for [segments_c]
java.nio.file.NoSuchFileException: data/index/segments_c
at sun.nio.fs.UnixException.translateToIOException(UnixException.java:86)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107)
at
sun.nio.fs.UnixFileAttributeViews$Basic.readAttributes(UnixFileAttributeViews.java:55)
at
sun.nio.fs.UnixFileSystemProvider.readAttributes(UnixFileSystemProvider.java:144)
at java.nio.file.Files.readAttributes(Files.java:1737)
at java.nio.file.Files.size(Files.java:2332)
at org.apache.lucene.store.FSDirectory.fileLength(FSDirectory.java:243)
at
org.apache.lucene.store.NRTCachingDirectory.fileLength(NRTCachingDirectory.java:128)
at
org.apache.solr.handler.admin.LukeRequestHandler.getFileLength(LukeRequestHandler.java:598)
at
org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo(LukeRequestHandler.java:586)
at
org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:137)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:153)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:2213)
at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:654)
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:303)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254)
at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1668)
at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)
at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)
at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1160)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511)
at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1092)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)
at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119)
at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at org.eclipse.jetty.server.Server.handle(Server.java:518)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:308)
at
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:244)
at
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95)
at
org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceAndRun(ExecuteProduceConsume.java:246)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:156)
at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:654)
at
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:572)
at java.lang.Thread.run(Thread.java:745)
2016-11-29 12:25:36.964 INFO  (qtp1528637575-35) [   x:collection1]
o.a.s.c.S.Request [collection1]  webapp=/solr path=/replication
params={wt=json=details&_=1480422336954} status=0 QTime=4
2016-11-29 12:25:36.967 INFO  (qtp1528637575-14) [   x:collection1]
o.a.s.c.S.Request [collection1]  webapp=/solr path=/admin/luke
params={numTerms=0=index=json&_=1480422336954} status=0 QTime=6
2016-11-29 12:25:36.982 INFO  (qtp1528637575-18) [   x:collection1]
o.a.s.c.S.Request [collection1]  webapp=/solr path=/admin/system
params={wt=json&_=1480422336954} status=0 QTime=21
2016-11-29 12:25:36.983 INFO  (qtp1528637575-49) [   ] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/info/system
params={wt=json&_=1480422336924} status=0 QTime=23
2016-11-29 12:25:38.570 INFO  (qtp1528637575-37) [   ] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={indexInfo=false=json&_=1480422336924} status=0 QTime=0
2016-11-29 12:25:38.582 INFO  (qtp1528637575-72) [   ] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/info/system
params={wt=json&_=1480422336924} status=0 QTime=12
2016-11-29 12:25:51.103 INFO  (qtp1528637575-73) [   x:collection1]
o.a.s.c.S.Request [collection1]  webapp=/solr path=/select
params={q=*:*=on=json&_=1480422338519} hits=0 status=0 QTime=7
2016-11-29 12:26:03.869 INFO  (Thread-0) [   ] o.e.j.s.ServerConnector

LukeRequestHandler Error getting file length for [segments_1l]

2016-11-29 Thread Furkan KAMACI
I use Solr 6.3 and get too many warning about. Is it usual:

WARN true LukeRequestHandler Error getting file length for [segments_1l]
java.nio.file.NoSuchFileException:
/home/server/solr/collection1/data/index/segments_1l
at sun.nio.fs.UnixException.translateToIOException(UnixException.java:86)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107)
at
sun.nio.fs.UnixFileAttributeViews$Basic.readAttributes(UnixFileAttributeViews.java:55)
at
sun.nio.fs.UnixFileSystemProvider.readAttributes(UnixFileSystemProvider.java:144)
at
sun.nio.fs.LinuxFileSystemProvider.readAttributes(LinuxFileSystemProvider.java:99)
at java.nio.file.Files.readAttributes(Files.java:1737)
at java.nio.file.Files.size(Files.java:2332)
at org.apache.lucene.store.FSDirectory.fileLength(FSDirectory.java:243)
at
org.apache.lucene.store.NRTCachingDirectory.fileLength(NRTCachingDirectory.java:128)
at
org.apache.solr.handler.admin.LukeRequestHandler.getFileLength(LukeRequestHandler.java:598)
at
org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo(LukeRequestHandler.java:586)
at
org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:137)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:153)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:2213)
at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:654)
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:303)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254)
at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1668)
at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)
at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)
at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1160)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511)
at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1092)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)
at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119)
at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at org.eclipse.jetty.server.Server.handle(Server.java:518)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:308)
at
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:244)
at
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95)
at
org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceAndRun(ExecuteProduceConsume.java:246)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:156)
at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:654)
at
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:572)
at java.lang.Thread.run(Thread.java:745)

Kind Regards,
Furkan KAMACI


Re: The state of Solr 5. Is it in maintenance mode only?

2016-11-29 Thread Charlie Hull

On 28/11/2016 13:11, Jaroslaw Rozanski wrote:

Hi,

Thanks for elaborate response. Missed the link to duplicate JIRA. Makes
sense.

On the 5.x front I wasn't expecting 5.6 release now that we have 6.x but
was simply surprised to see fix for 4.x and not for 5.x.

As for adoption levels, it was my subjective feel reading this list. Do
we have community survey on that subject? That would be really
interesting to see.

Hi,

We still see Solr 4.x running in the wild, though not so often these 
days. 5.x is also common especially when running under other systems 
such as Drupal or Sitecore (the developers of these systems and others 
that integrate Solr will often get something running with whatever 
version is reasonably current - often a barely modified example config - 
then not touch it again for ages. We obviously recommend 6.x for new 
installations and encourage others to upgrade to this if possible. Note 
that this upgrade process gets harder the further you are away from the 
current release.


Cheers

Charlie



Thanks,
Jaroslaw


On 28/11/16 12:59, Shawn Heisey wrote:

On 11/28/2016 4:29 AM, Jaroslaw Rozanski wrote:

Recently I have noticed that couple of Solr issues have been
resolved/added only for Solr 4.x and Solr 6.x branch. For example
https://issues.apache.org/jira/browse/SOLR-2242. Has Solr 5.x branch
been moved to maintenance mode only? The 5 wasn't around for long
before 6 came about so I appreciate its adoption might not be vast.

The 5.0 version was announced in March 2015.  The 6.0 version was
announced in April 2016.  Looks like 4.x was current for a little less
than three years (July 2012 for 4.0).  5.x had one year, which I
wouldn't call really call a short time.

Since the release of 6.0, 4.x is dead and 5.x is in maintenance mode.
Maintenance mode means that only particularly nasty bugs are fixed and
only extremely trivial features are added.  The latter is usually only
done if the lack of the feature can be considered a bug.  There is never
any guarantee that a new 5.x release will be made, but if that happens,
it will be a 5.5.x release.  The likelihood of seeing a 5.6 release is
VERY low.

SOLR-2242 is a duplicate of SOLR-6348.  It probably had 4.9 in the fixed
version field because that's what was already in it when it was resolved
as a duplicate.  It's a very old issue that's been around since the 3.x
days.  No changes were committed for SOLR-2242.  The changes for
SOLR-6348 were committed to 5.2 and 6.0.  I have updated the fix
versions in the older issue to match.  The versions should probably all
be removed, but I am not sure what our general rule is for duplicates.

Thanks,
Shawn




--
Charlie Hull
Flax - Open Source Enterprise Search

tel/fax: +44 (0)8700 118334
mobile:  +44 (0)7767 825828
web: www.flax.co.uk