Specifying shards when querying a alias.

2017-07-11 Thread philippa griggs
Hello,


Solr 5.4.1


I have two collections ‘Online’ and ‘Offline’ both collections have an implicit 
router and are sharded into weekly cores (for example an online shard would be 
Online_20170605). I have created an alias to query both collections called 
‘AllData’. I want to query the alias but specify the shards. So something like 
this:


http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=Online_20170605,Offline_20170529



However I have noticed that this doesn’t work.  If I create the alias like this 
:


http://localhost:8983/solr/admin/collections?action=CREATEALIAS=AllData=Online,Offline


(with the Online collection mentioned first)


I can use Online_20170605 as a shard but not the Offline shard, for that I 
would have to specify the location of the shard.


http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=Online_20170605,localhost:8983/solr/offline


If I delete the alias and recreate it with the Offline collection mentioned 
first:


http://localhost:8983/solr/admin/collections?action=CREATEALIAS=AllData=Offline,Online


I can use the Offline_20170529 as a shard but then have to specify the location 
of the shard for the online collection.


http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=localhost:8984/solr/online,Offline_20170529



Is this expected behaviour? Or can anyone point out what I’m doing wrong?


Many thanks

Philippa




Re: Returning results for multi-word search term

2017-07-11 Thread Shawn Heisey
On 7/10/2017 1:02 PM, Miller, William K - Norman, OK - Contractor wrote:
>
> I am trying to return results when using a multi-word term.  I am
> using “Paddle Arm” as my search term(including the quotes).  I know
> that the field that I am querying against has these words together. 
> If I run the query using Paddle* Arm* I get the following results, but
> I want to get only the last two.  I have looked at Fuzzy Searches but
> that I don’t feel will work and I have looked at the Proximity
> Searches and I get no results back with that one whether I use 0,1 or
> 10.  How can I structure my query to get the last items in the below list?
>
>  
>
> Paddle Assembly
>
> Paddle
>
> Paddle
>
> Paddle Pneumatic Piping
>
> Paddle
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle to Bucket Offset Check
>
> Paddle to Bucket Wall
>
> Paddle to Bucket Offset
>
> Paddle
>
> Paddle Assembly Troubleshooting
>
> Paddle Assembly Troubleshooting
>
> Paddle Air Pressure
>
> Paddle Assembly
>
> Paddle
>
> Paddle Stop Adjustment
>
> Paddle Stop
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle Vacuum Holes
>
> Paddle Position
>
> Paddle Detection Sensor Adjustment
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle Stop
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle Rotary Actuator
>
> Paddle Removal and Replacement
>
> Paddle Assembly
>
> Paddle Removal and Replacement
>
> Paddle Seal Removal and Replacement
>
> Paddle Location
>
> Paddle Location
>
> Paddle Removal Location
>
> Paddle/Belt Speed for Photoeye Inputs
>
> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>
> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>
>

The answer to your question is going to be less about the query
structure and more about the type of field you're using and any defined
analysis for that type.

With a schema field type that is properly configured, the query you are
trying to use (with quotes) should work, as long as it is being directed
specifically to the correct field as Erick mentioned.  Note that if you
change your schema to make this work, you will need to entirely reindex.

Nearly any of the "text" field types included in the example schemas
will do the job.  The "string" or "strings" types in the examples will
NOT work, because they do not break the text into multiple tokens
(search terms).  I think you're probably trying to search a field that
is using one of the latter types.

Can you share your schema and the name of the field that you are
querying?  Your followup message tells us the version of Solr (6.5.1),
so the most likely filename for the schema will be "managed-schema" with
no extension.  If there have been significant changes from an example in
your solrconfig.xml file, it would be a good idea to share that too.

Thanks,
Shawn



RE: Returning results for multi-word search term

2017-07-11 Thread Miller, William K - Norman, OK - Contractor
I do have my fields as strings not text, so I am going to play around with 
using the "text".  If I continue to have problems, I will post the additional 
information you are requesting.




~~~
William Kevin Miller

ECS Federal, Inc.
USPS/MTSC
(405) 573-2158


-Original Message-
From: Shawn Heisey [mailto:apa...@elyograg.org] 
Sent: Tuesday, July 11, 2017 8:34 AM
To: solr-user@lucene.apache.org
Subject: Re: Returning results for multi-word search term

On 7/10/2017 1:02 PM, Miller, William K - Norman, OK - Contractor wrote:
>
> I am trying to return results when using a multi-word term.  I am 
> using “Paddle Arm” as my search term(including the quotes).  I know 
> that the field that I am querying against has these words together.
> If I run the query using Paddle* Arm* I get the following results, but 
> I want to get only the last two.  I have looked at Fuzzy Searches but 
> that I don’t feel will work and I have looked at the Proximity 
> Searches and I get no results back with that one whether I use 0,1 or 
> 10.  How can I structure my query to get the last items in the below list?
>
>  
>
> Paddle Assembly
>
> Paddle
>
> Paddle
>
> Paddle Pneumatic Piping
>
> Paddle
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle to Bucket Offset Check
>
> Paddle to Bucket Wall
>
> Paddle to Bucket Offset
>
> Paddle
>
> Paddle Assembly Troubleshooting
>
> Paddle Assembly Troubleshooting
>
> Paddle Air Pressure
>
> Paddle Assembly
>
> Paddle
>
> Paddle Stop Adjustment
>
> Paddle Stop
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle Vacuum Holes
>
> Paddle Position
>
> Paddle Detection Sensor Adjustment
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle Stop
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle
>
> Paddle Assembly
>
> Paddle Assembly
>
> Paddle Rotary Actuator
>
> Paddle Removal and Replacement
>
> Paddle Assembly
>
> Paddle Removal and Replacement
>
> Paddle Seal Removal and Replacement
>
> Paddle Location
>
> Paddle Location
>
> Paddle Removal Location
>
> Paddle/Belt Speed for Photoeye Inputs
>
> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>
> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>
>

The answer to your question is going to be less about the query structure and 
more about the type of field you're using and any defined analysis for that 
type.

With a schema field type that is properly configured, the query you are trying 
to use (with quotes) should work, as long as it is being directed specifically 
to the correct field as Erick mentioned.  Note that if you change your schema 
to make this work, you will need to entirely reindex.

Nearly any of the "text" field types included in the example schemas will do 
the job.  The "string" or "strings" types in the examples will NOT work, 
because they do not break the text into multiple tokens (search terms).  I 
think you're probably trying to search a field that is using one of the latter 
types.

Can you share your schema and the name of the field that you are querying?  
Your followup message tells us the version of Solr (6.5.1), so the most likely 
filename for the schema will be "managed-schema" with no extension.  If there 
have been significant changes from an example in your solrconfig.xml file, it 
would be a good idea to share that too.

Thanks,
Shawn



Re: High disk write usage

2017-07-11 Thread Antonio De Miguel
Thanks Shawn!


I will try to change the values of those parameters


2017-07-10 14:57 GMT+02:00 Shawn Heisey :

> On 7/10/2017 2:57 AM, Antonio De Miguel wrote:
> > I continue deeping inside this problem...  high writing rates continues.
> >
> > Searching in logs i see this:
> >
> > 2017-07-10 08:46:18.888 INFO  (commitScheduler-11-thread-1) [c:ads
> s:shard2
> > r:core_node47 x:ads_shard2_replica3] o.a.s.u.LoggingInfoStream
> > [DWPT][commitScheduler-11-thread-1]: flushed: segment=_mb7
> ramUsed=7.531 MB
> > newFlushedSize=2.472 MB docs/MB=334.132
> > 2017-07-10 08:46:29.336 INFO  (commitScheduler-11-thread-1) [c:ads
> s:shard2
> > r:core_node47 x:ads_shard2_replica3] o.a.s.u.LoggingInfoStream
> > [DWPT][commitScheduler-11-thread-1]: flushed: segment=_mba
> ramUsed=8.079 MB
> > newFlushedSize=1.784 MB docs/MB=244.978
> >
> >
> > A flush happens each 10 seconds (my autosoftcommit time is 10 secs and
> > hardcommit 5 minutes).  ¿is the expected behaviour?
>
> If you are indexing continuously, then the auto soft commit time of 10
> seconds means that this will be happening every ten seconds.
>
> > I thought soft commits does not write into disk...
>
> If you are using the correct DirectoryFactory type, a soft commit has
> the *possibility* of not writing to disk, but the amount of memory
> reserved is fairly small.
>
> Looking into the source code for NRTCachingDirectoryFactory, I see that
> maxMergeSizeMB defaults to 4, and maxCachedMB defaults to 48.  This is a
> little bit different than what the javadoc states for
> NRTCachingDirectory (5 and 60):
>
> http://lucene.apache.org/core/6_6_0/core/org/apache/lucene/
> store/NRTCachingDirectory.html
>
> The way I read this, assuming the amount of segment data created is
> small, only the first few soft commits will be entirely handled in
> memory.  After that, older segments must be flushed to disk to make room
> for new ones.
>
> If the indexing rate is high, there's not really much difference between
> soft commits and hard commits.  This also assumes that you have left the
> directory at the default of NRTCachingDirectoryFactory.  If this has
> been changed, then there is no caching in RAM, and soft commit probably
> behaves *exactly* the same as hard commit.
>
> Thanks,
> Shawn
>
>


Re: SolrCloud - Sharing zookeeper ensemble with Kafka

2017-07-11 Thread Ganesh M
Even we use same zookeeper for HBase and SolrCloud with corresponding
folder structure.

On Tue 11 Jul, 2017 7:01 pm Joe Obernberger, 
wrote:

> Vincenzo - we do this in our environment.  Zookeeper handles, HDFS,
> HBase, Kafka, and Solr Cloud.
>
> -Joe
>
>
> On 7/11/2017 4:18 AM, Vincenzo D'Amore wrote:
> > Hi All,
> >
> > in my test environment I've two Zookeeper instances one for SolrCloud
> > (6.6.0) and another for a Kafka server (2.11-0.10.1.0).
> >
> > My task (for now) is reading from a topic queue from the Kafka instance
> and
> > then writing all the documents in a Solr Collection.
> >
> > I write here just to ask if in your experience, I can share the zookeeper
> > instance (or ensemble) between the two server (instead of have two
> separate
> > instances) and if not what are the counter-indications.
> >
> > Thanks in advance for your time and best regards,
> > Vincenzo
> >
>
>


Collections API Overseer Status

2017-07-11 Thread Webster Homer
Looking at the overseer API call as documented in the Solr Collections API
https://cwiki.apache.org/confluence/display/solr/Collections+API#CollectionsAPI-OVERSEERSTATUS:OverseerStatusandStatistics

The information returned looks like it could be useful in diagnosing
problems with Solrcloud.

It would be useful to document the metrics it returns
What constitutes an "overseer operation"?

What is "peek_wait_forever" ?

I could go on but it's difficult to interpret the results of this API
without more documentation

-- 


This message and any attachment are confidential and may be privileged or 
otherwise protected from disclosure. If you are not the intended recipient, 
you must not copy this message or attachment or disclose the contents to 
any other person. If you have received this transmission in error, please 
notify the sender immediately and delete the message and any attachment 
from your system. Merck KGaA, Darmstadt, Germany and any of its 
subsidiaries do not accept liability for any omissions or errors in this 
message which may arise as a result of E-Mail-transmission or for damages 
resulting from any unauthorized changes of the content of this message and 
any attachment thereto. Merck KGaA, Darmstadt, Germany and any of its 
subsidiaries do not guarantee that this message is free of viruses and does 
not accept liability for any damages caused by any virus transmitted 
therewith.

Click http://www.emdgroup.com/disclaimer to access the German, French, 
Spanish and Portuguese versions of this disclaimer.


Re: SolrCloud - Sharing zookeeper ensemble with Kafka

2017-07-11 Thread Joe Obernberger
Vincenzo - we do this in our environment.  Zookeeper handles, HDFS, 
HBase, Kafka, and Solr Cloud.


-Joe


On 7/11/2017 4:18 AM, Vincenzo D'Amore wrote:

Hi All,

in my test environment I've two Zookeeper instances one for SolrCloud
(6.6.0) and another for a Kafka server (2.11-0.10.1.0).

My task (for now) is reading from a topic queue from the Kafka instance and
then writing all the documents in a Solr Collection.

I write here just to ask if in your experience, I can share the zookeeper
instance (or ensemble) between the two server (instead of have two separate
instances) and if not what are the counter-indications.

Thanks in advance for your time and best regards,
Vincenzo





Re: Tlogs not being deleted/truncated

2017-07-11 Thread jmyatt
Have you found a solution to this?  I'm seeing the same behavior (running
solr v6.5.0):

- source data center replicates to target just fine
- autocommit on Source with openSearcher=false causes tlogs to rollover at a
reasonable size
- a post-import explicit commit (with the default openSearcher=true) causes
docs to be visible on Source
- autocommit on Target with openSearcher=true (and a higher maxTime /
maxDocs) causes tlogs to roll eventually and docs to become visible without
any need for a final explicit commit
- tlogs on Target clean up (get deleted) just fine, after each hard commit
- *tlogs on Source are never deleted and I can't figure out why*
 
I've tried this with the cdcr buffer both enabled and disabled, and tried it
with DataImportHandler imports as well as SolrJ posts of updates.  I can't
find any combination of settings that causes tlogs on the Source cluster to
be cleaned up.  Any ideas on how to manage this other than implement a
manual log deletion routine on the filesystem?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Tlogs-not-being-deleted-truncated-tp4341958p4345473.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Tlogs not being deleted/truncated

2017-07-11 Thread jmyatt
another interesting clue in my case (different from what WebsterHomer is
seeing): the response from /cdcr?action=QUEUES reflects what I would expect
to see in the tlog directory but it's not accurate.  By that I mean
tlogTotalSize shows 1500271 (bytes) and tlogTotalCount shows 2.  This
changes as more updates come in and autoCommit runs - sometimes
tlogTotalCount is 1 instead of 2, and the tlogTotalSize changes but stays in
that low range.  

But on the filesystem, all the tlogs are still there.  Perhaps the ignored
exception noted above is in fact a problem?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Tlogs-not-being-deleted-truncated-tp4341958p4345477.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Specifying shards when querying a alias.

2017-07-11 Thread Erick Erickson
Maybe shard.keys? See:
https://lucidworks.com/2013/06/13/solr-cloud-document-routing/

Best,
Erick

On Tue, Jul 11, 2017 at 4:24 AM, philippa griggs
 wrote:
> Hello,
>
>
> Solr 5.4.1
>
>
> I have two collections ‘Online’ and ‘Offline’ both collections have an 
> implicit router and are sharded into weekly cores (for example an online 
> shard would be Online_20170605). I have created an alias to query both 
> collections called ‘AllData’. I want to query the alias but specify the 
> shards. So something like this:
>
>
> http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=Online_20170605,Offline_20170529
>
>
>
> However I have noticed that this doesn’t work.  If I create the alias like 
> this :
>
>
> http://localhost:8983/solr/admin/collections?action=CREATEALIAS=AllData=Online,Offline
>
>
> (with the Online collection mentioned first)
>
>
> I can use Online_20170605 as a shard but not the Offline shard, for that I 
> would have to specify the location of the shard.
>
>
> http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=Online_20170605,localhost:8983/solr/offline
>
>
> If I delete the alias and recreate it with the Offline collection mentioned 
> first:
>
>
> http://localhost:8983/solr/admin/collections?action=CREATEALIAS=AllData=Offline,Online
>
>
> I can use the Offline_20170529 as a shard but then have to specify the 
> location of the shard for the online collection.
>
>
> http://localhost:8983/solr/AllData/select?q=*%3A*=Session_UTCStartTime+desc=json=true=localhost:8984/solr/online,Offline_20170529
>
>
>
> Is this expected behaviour? Or can anyone point out what I’m doing wrong?
>
>
> Many thanks
>
> Philippa
>
>


Re: Returning results for multi-word search term

2017-07-11 Thread Erick Erickson
The admin/analysis page is your friend here. Hover over the light gray
abbreviations (like "ST") and you'll see what the analysis chain
component is that does the transformation. (Standard Tokenizer in this
case). I almost always turn off the "verbose" checkbox BTW.

In general you only want string types for things where the entire
field must be considered for matching. Do _not_ fall into the habit of
indexing a string field and then searching for words with *something*
as though it were a SQL query as that's just a horribly inefficient.

Best,
Erick

On Tue, Jul 11, 2017 at 6:46 AM, Miller, William K - Norman, OK -
Contractor  wrote:
> I do have my fields as strings not text, so I am going to play around with 
> using the "text".  If I continue to have problems, I will post the additional 
> information you are requesting.
>
>
>
>
> ~~~
> William Kevin Miller
>
> ECS Federal, Inc.
> USPS/MTSC
> (405) 573-2158
>
>
> -Original Message-
> From: Shawn Heisey [mailto:apa...@elyograg.org]
> Sent: Tuesday, July 11, 2017 8:34 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Returning results for multi-word search term
>
> On 7/10/2017 1:02 PM, Miller, William K - Norman, OK - Contractor wrote:
>>
>> I am trying to return results when using a multi-word term.  I am
>> using “Paddle Arm” as my search term(including the quotes).  I know
>> that the field that I am querying against has these words together.
>> If I run the query using Paddle* Arm* I get the following results, but
>> I want to get only the last two.  I have looked at Fuzzy Searches but
>> that I don’t feel will work and I have looked at the Proximity
>> Searches and I get no results back with that one whether I use 0,1 or
>> 10.  How can I structure my query to get the last items in the below list?
>>
>>
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle
>>
>> Paddle Pneumatic Piping
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle to Bucket Offset Check
>>
>> Paddle to Bucket Wall
>>
>> Paddle to Bucket Offset
>>
>> Paddle
>>
>> Paddle Assembly Troubleshooting
>>
>> Paddle Assembly Troubleshooting
>>
>> Paddle Air Pressure
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Stop Adjustment
>>
>> Paddle Stop
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle Vacuum Holes
>>
>> Paddle Position
>>
>> Paddle Detection Sensor Adjustment
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle Stop
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle Rotary Actuator
>>
>> Paddle Removal and Replacement
>>
>> Paddle Assembly
>>
>> Paddle Removal and Replacement
>>
>> Paddle Seal Removal and Replacement
>>
>> Paddle Location
>>
>> Paddle Location
>>
>> Paddle Removal Location
>>
>> Paddle/Belt Speed for Photoeye Inputs
>>
>> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>>
>> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>>
>>
>
> The answer to your question is going to be less about the query structure and 
> more about the type of field you're using and any defined analysis for that 
> type.
>
> With a schema field type that is properly configured, the query you are 
> trying to use (with quotes) should work, as long as it is being directed 
> specifically to the correct field as Erick mentioned.  Note that if you 
> change your schema to make this work, you will need to entirely reindex.
>
> Nearly any of the "text" field types included in the example schemas will do 
> the job.  The "string" or "strings" types in the examples will NOT work, 
> because they do not break the text into multiple tokens (search terms).  I 
> think you're probably trying to search a field that is using one of the 
> latter types.
>
> Can you share your schema and the name of the field that you are querying?  
> Your followup message tells us the version of Solr (6.5.1), so the most 
> likely filename for the schema will be "managed-schema" with no extension.  
> If there have been significant changes from an example in your solrconfig.xml 
> file, it would be a good idea to share that too.
>
> Thanks,
> Shawn
>


RE: Returning results for multi-word search term

2017-07-11 Thread Miller, William K - Norman, OK - Contractor
I used the "copyField" and created a text version of the field that I wanted to 
search on and am now getting the results I was looking for.  Thanks for all 
your help.




~~~
William Kevin Miller

ECS Federal, Inc.
USPS/MTSC
(405) 573-2158


-Original Message-
From: Erick Erickson [mailto:erickerick...@gmail.com] 
Sent: Tuesday, July 11, 2017 10:54 AM
To: solr-user
Subject: Re: Returning results for multi-word search term

The admin/analysis page is your friend here. Hover over the light gray 
abbreviations (like "ST") and you'll see what the analysis chain component is 
that does the transformation. (Standard Tokenizer in this case). I almost 
always turn off the "verbose" checkbox BTW.

In general you only want string types for things where the entire field must be 
considered for matching. Do _not_ fall into the habit of indexing a string 
field and then searching for words with *something* as though it were a SQL 
query as that's just a horribly inefficient.

Best,
Erick

On Tue, Jul 11, 2017 at 6:46 AM, Miller, William K - Norman, OK - Contractor 
 wrote:
> I do have my fields as strings not text, so I am going to play around with 
> using the "text".  If I continue to have problems, I will post the additional 
> information you are requesting.
>
>
>
>
> ~~~
> William Kevin Miller
>
> ECS Federal, Inc.
> USPS/MTSC
> (405) 573-2158
>
>
> -Original Message-
> From: Shawn Heisey [mailto:apa...@elyograg.org]
> Sent: Tuesday, July 11, 2017 8:34 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Returning results for multi-word search term
>
> On 7/10/2017 1:02 PM, Miller, William K - Norman, OK - Contractor wrote:
>>
>> I am trying to return results when using a multi-word term.  I am 
>> using “Paddle Arm” as my search term(including the quotes).  I know 
>> that the field that I am querying against has these words together.
>> If I run the query using Paddle* Arm* I get the following results, 
>> but I want to get only the last two.  I have looked at Fuzzy Searches 
>> but that I don’t feel will work and I have looked at the Proximity 
>> Searches and I get no results back with that one whether I use 0,1 or 
>> 10.  How can I structure my query to get the last items in the below list?
>>
>>
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle
>>
>> Paddle Pneumatic Piping
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle to Bucket Offset Check
>>
>> Paddle to Bucket Wall
>>
>> Paddle to Bucket Offset
>>
>> Paddle
>>
>> Paddle Assembly Troubleshooting
>>
>> Paddle Assembly Troubleshooting
>>
>> Paddle Air Pressure
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Stop Adjustment
>>
>> Paddle Stop
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle Vacuum Holes
>>
>> Paddle Position
>>
>> Paddle Detection Sensor Adjustment
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle Stop
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle
>>
>> Paddle Assembly
>>
>> Paddle Assembly
>>
>> Paddle Rotary Actuator
>>
>> Paddle Removal and Replacement
>>
>> Paddle Assembly
>>
>> Paddle Removal and Replacement
>>
>> Paddle Seal Removal and Replacement
>>
>> Paddle Location
>>
>> Paddle Location
>>
>> Paddle Removal Location
>>
>> Paddle/Belt Speed for Photoeye Inputs
>>
>> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>>
>> Paddle Arm Spring, Upper Paddle Arm, and Lower Paddle Arm
>>
>>
>
> The answer to your question is going to be less about the query structure and 
> more about the type of field you're using and any defined analysis for that 
> type.
>
> With a schema field type that is properly configured, the query you are 
> trying to use (with quotes) should work, as long as it is being directed 
> specifically to the correct field as Erick mentioned.  Note that if you 
> change your schema to make this work, you will need to entirely reindex.
>
> Nearly any of the "text" field types included in the example schemas will do 
> the job.  The "string" or "strings" types in the examples will NOT work, 
> because they do not break the text into multiple tokens (search terms).  I 
> think you're probably trying to search a field that is using one of the 
> latter types.
>
> Can you share your schema and the name of the field that you are querying?  
> Your followup message tells us the version of Solr (6.5.1), so the most 
> likely filename for the schema will be "managed-schema" with no extension.  
> If there have been significant changes from an example in your solrconfig.xml 
> file, it would be a good idea to share that too.
>
> Thanks,
> Shawn
>


Re: Cross DC SolrCloud anti-patterns in presentation shalinmangar/cross-datacenter-replication-in-apache-solr-6

2017-07-11 Thread Erick Erickson
bq: if you loose DC1, then your cluster will not be able to recover
after DC1 comes back as there will be no clear majority

When ZK loses majority, no indexing takes place. So in the case where
you have 3 nodes in DC1 and 3 nodes in DC2, _neither_ of them would
allow updates if the connection was cut for any reason since updates
require 4 live ZK servers in this scenario to be available ((6/2)+1).
So when the connection was restored, there'd be nothing to reconcile
and Solr should recover just fine.

The whole ZK majority thing is about data consistency. Since querying
doesn't change the index at all there's no consistency problem here to
reconcile after the connection is restored. And since quorum was lost,
no updates are allowed.

Best,
Erick

On Mon, Jul 10, 2017 at 5:08 PM, Arcadius Ahouansou
 wrote:
> Hello Shawn.
>
> Thank you very much for the comment.
>
> On 24 June 2017 at 16:14, Shawn Heisey  wrote:
>
>> On 6/24/2017 2:14 AM, Arcadius Ahouansou wrote:
>> > Interpretation 1:
>>
>> ZooKeeper doesn't *need* an odd number of servers, but there's no
>> benefit to an even number.  If you have 5 servers, two can go down.  If
>> you have 6 servers, you can still only lose two, so you might as well
>> just run 5.  You'd have fewer possible points of failure, less power
>> usage, and less bandwidth usage.
>>
>>
> About Slide 8 and the odd/even number of nodes...
> what I meant is that on Slide 8, if you loose DC1, then your cluster will
> not be able to recover after DC1 comes back as there will be no clear
> majority
> and you will have:
> -  3 ZK nodes with up-to-date data (that is DC2+DC3) and
> -  3 ZK nodes with out-of-date data (DC1).
>
> But, if you had only 2 ZK nodes in DC1, then you could afford to loose one
> of either DC1, or DC2 or DC3 and the cluster will be able to recover and be
> OK
>
>
> Thank you very much.
>
>
> Arcadius
>
> --
> Arcadius Ahouansou
> Menelic Ltd | Applied Knowledge Is Power
> Office : +441444702101
> Mobile: +447908761999
> Web: www.menelic.com
> ---


Using HTTP and HTTPS at the same time

2017-07-11 Thread Nawab Zada Asad Iqbal
Hi,

I am reading a comment on
https://cwiki.apache.org/confluence/display/solr/Enabling+SSL which says.
Just wanted to check if this is still the same with 6.5? This used to work
in 4.5.
Shalin Shekhar Mangar


Solr does not support both HTTP and HTTPS at the same time. You can only
use one of them at a time.


Thanks

Nawab


Re: Arabic words search in solr

2017-07-11 Thread mohanmca01
Hi Steve,

thank you for your reply, it been quite long time to reply to you back.

i have tried what you suggested, and there were some improvements in terms
of searching and getting the results.

however, the team is facing some difficulty in searching using shortcut of
the indexed names which forced us to revert back the suggested changes..

below are the examples which we are facing:


-
*Example 1:*

*Indexed Text*
بنك مسقط - مركز الاتصال

*Searched*
مسقط الات

*Remarks of Example 1*
unable to get the indexed result unless I typed the two words fully (مسقط
الاتصال)


{
  "responseHeader": {
"status": 0,
"QTime": 0,
"params": {
  "indent": "true",
  "q": "businessNmBl:(مسقط الات)",
  "_": "1499758511717",
  "wt": "json"
}
  },
  "response": {
"numFound": 0,
"start": 0,
"docs": []
  }
}


-

*Example 2:*

*Indexed Text
*الطيران العماني - مركز الاتصال

*Searched*
الطير الات

*Remarks*
unable to get the indexed result unless I typed the two words fully (الطيران
الاتصال)


{
  "responseHeader": {
"status": 0,
"QTime": 2,
"params": {
  "indent": "true",
  "q": "businessNmBl:(طير الات)",
  "_": "1499758649600",
  "wt": "json"
}
  },
  "response": {
"numFound": 0,
"start": 0,
"docs": []
  }
}



Please be noted that the existing configuration (which we are facing
problems with Hamzzh (ء) and etc. )  on production is working with the above
examples. its not working only once we implement your suggested
configuration. 

Thanks in advance





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Arabic-words-search-in-solr-tp4317733p4345392.html
Sent from the Solr - User mailing list archive at Nabble.com.


SolrCloud - Sharing zookeeper ensemble with Kafka

2017-07-11 Thread Vincenzo D'Amore
Hi All,

in my test environment I've two Zookeeper instances one for SolrCloud
(6.6.0) and another for a Kafka server (2.11-0.10.1.0).

My task (for now) is reading from a topic queue from the Kafka instance and
then writing all the documents in a Solr Collection.

I write here just to ask if in your experience, I can share the zookeeper
instance (or ensemble) between the two server (instead of have two separate
instances) and if not what are the counter-indications.

Thanks in advance for your time and best regards,
Vincenzo

-- 
Vincenzo D'Amore
email: v.dam...@gmail.com
skype: free.dev
mobile: +39 349 8513251 <349%20851%203251>


Re: Tlogs not being deleted/truncated

2017-07-11 Thread Xie, Sean
Please see my previous thread. I have to disable buffer on source cluster and a 
scheduled hard commit with scheduled logscheduler to make it work.


-- Thank you
Sean

From: jmyatt >
Date: Tuesday, Jul 11, 2017, 1:56 PM
To: solr-user@lucene.apache.org 
>
Subject: [EXTERNAL] Re: Tlogs not being deleted/truncated

another interesting clue in my case (different from what WebsterHomer is
seeing): the response from /cdcr?action=QUEUES reflects what I would expect
to see in the tlog directory but it's not accurate.  By that I mean
tlogTotalSize shows 1500271 (bytes) and tlogTotalCount shows 2.  This
changes as more updates come in and autoCommit runs - sometimes
tlogTotalCount is 1 instead of 2, and the tlogTotalSize changes but stays in
that low range.

But on the filesystem, all the tlogs are still there.  Perhaps the ignored
exception noted above is in fact a problem?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Tlogs-not-being-deleted-truncated-tp4341958p4345477.html
Sent from the Solr - User mailing list archive at Nabble.com.

Confidentiality Notice::  This email, including attachments, may include 
non-public, proprietary, confidential or legally privileged information.  If 
you are not an intended recipient or an authorized agent of an intended 
recipient, you are hereby notified that any dissemination, distribution or 
copying of the information contained in or transmitted with this e-mail is 
unauthorized and strictly prohibited.  If you have received this email in 
error, please notify the sender by replying to this message and permanently 
delete this e-mail, its attachments, and any copies of it immediately.  You 
should not retain, copy or use this e-mail or any attachment for any purpose, 
nor disclose all or any part of the contents to any other person. Thank you.