Solr7.3.1 Installation

2018-07-10 Thread tapan1707
We are trying to install solr-7.3.1 into our existing system (We have also
made some changes by adding one custom query parser).

I am having some build issues and it would be really helpful if someone can
help.

While running ant test(in the process of building the solr package), it
terminates because of failed tests.
At first time (build with ant-1.9)
Tests with failures [seed: C2C0D761AEAAE8A4] (first 10 out of 23):
21:25:20[junit4]   -
org.apache.solr.client.solrj.response.TestSuggesterResponse (suite)
21:25:20[junit4]   -
org.apache.solr.client.solrj.response.TermsResponseTest (suite)
21:25:20[junit4]   - org.apache.solr.client.solrj.TestSolrJErrorHandling
(suite)
21:25:20[junit4]   - org.apache.solr.client.solrj.GetByIdTest (suite)
21:25:20[junit4]   -
org.apache.solr.client.solrj.response.TestSpellCheckResponse (suite)
21:25:20[junit4]   -
org.apache.solr.client.solrj.embedded.LargeVolumeEmbeddedTest (suite)
21:25:20[junit4]   -
org.apache.solr.client.solrj.embedded.JettyWebappTest.testAdminUI
21:25:20[junit4]   -
org.apache.solr.client.solrj.embedded.SolrExampleStreamingBinaryTest (suite)
21:25:20[junit4]   - org.apache.solr.client.solrj.SolrExampleBinaryTest
(suite)
21:25:20[junit4]   -
org.apache.solr.client.solrj.embedded.LargeVolumeBinaryJettyTest (suite)

Running the same ant test command without doing any changes (build with
ant-1.10)
Tests with failures [seed: 7E004642A6008D89]:
11:30:57[junit4]   -
org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove  

Thirds time (build with ant 1.10)
[junit4] Tests with failures [seed: EFD939D82A6EC707]:
[junit4]   - org.apache.solr.cloud.autoscaling.SystemLogListenerTest.test

Even though I'm not making any changes, build is failing with different
failed tests. Can anyone help me with this, I mean if there is any problem
with the code then shouldn't it fail with same test cases?
Also, all above-mentioned test cases work fine if I check them individually.
(using ant test -Dtests.class=)

Also, does ant version has any effects in build??

At last, at present, we are using solr-6.4.2 which has zookeeper-3.4.6
dependency but for solr-7, the zookeeper dependency has been upgraded to
3.4.10, so my question is, At what extent does this might affect our system
performance? Can we use zookeeper-3.4.6 with solr-7?
(same with the jetty version) 

Thanks in advance



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: FieldValueCache in solr 6.6

2018-07-10 Thread zhang.mingyue
hi:
   I have a question ,how to  load data to FieldValueCache in solr7 
   
   thanks



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Solr OpenNLP named entity extraction

2018-07-10 Thread Steve Rowe
Hi Jerome,

I was able to setup a configset to perform OpenNLP NER, loading the model files 
from local storage.

There is a trick though[1]: the model files must be located *in a jar* or *in a 
subdirectory* under ${solr.solr.home}/lib/ or under a directory specified via a 
solrconfig.xml  directive.

I tested with the bin/solr cloud example, and put model files under the two 
solr home directories, at example/cloud/node1/solr/lib/opennlp/ and 
example/cloud/node1/solr/lib/opennlp/.  The “opennlp/“ subdirectory is 
required, though its name can be anything else you choose.

[1] As you noted, ZkSolrResourceLoader delegates to its parent classloader when 
it can’t find resources in a configset, and the parent classloader is set up to 
load from subdirectories and jar files under ${solr.solr.home}/lib/ or under a 
directory specified via a solrconfig.xml  directive.  These directories 
themselves are not included in the set of directories from which resources are 
loaded; only their children are.

--
Steve
www.lucidworks.com

> On Jul 9, 2018, at 10:10 PM, Jerome Yang  wrote:
> 
> Hi Steve,
> 
> Put models under " ${solr.solr.home}/lib/ " is not working.
> I check the "ZkSolrResourceLoader" seems it will first try to find modes in
> config set.
> If not find, then it uses class loader to load from resources.
> 
> Regards,
> Jerome
> 
> On Tue, Jul 10, 2018 at 9:58 AM Jerome Yang  wrote:
> 
>> Thanks Steve!
>> 
>> 
>> On Tue, Jul 10, 2018 at 5:20 AM Steve Rowe  wrote:
>> 
>>> Hi Jerome,
>>> 
>>> See the ref guide[1] for a writeup of how to enable uploading files
>>> larger than 1MB into ZooKeeper.
>>> 
>>> Local storage should also work - have you tried placing OpenNLP model
>>> files in ${solr.solr.home}/lib/ ? - make sure you do the same on each node.
>>> 
>>> [1]
>>> https://lucene.apache.org/solr/guide/7_4/setting-up-an-external-zookeeper-ensemble.html#increasing-the-file-size-limit
>>> 
>>> --
>>> Steve
>>> www.lucidworks.com
>>> 
 On Jul 9, 2018, at 12:50 AM, Jerome Yang  wrote:
 
 Hi guys,
 
 In Solrcloud mode, where to put the OpenNLP models?
 Upload to zookeeper?
 As I test on solr 7.3.1, seems absolute path on local host is not
>>> working.
 And can not upload into zookeeper if the model size exceed 1M.
 
 Regards,
 Jerome
 
 On Wed, Apr 18, 2018 at 9:54 AM Steve Rowe  wrote:
 
> Hi Alexey,
> 
> First, thanks for moving the conversation to the mailing list.
>>> Discussion
> of usage problems should take place here rather than in JIRA.
> 
> I locally set up Solr 7.3 similarly to you and was able to get things
>>> to
> work.
> 
> Problems with your setup:
> 
> 1. Your update chain is missing the Log and Run update processors at
>>> the
> end (I see these are missing from the example in the javadocs for the
> OpenNLP NER update processor; I’ll fix that):
> 
>
>
> 
>  The Log update processor isn’t strictly necessary, but, from <
> 
>>> https://lucene.apache.org/solr/guide/7_3/update-request-processors.html#custom-update-request-processor-chain
>> :
> 
>  Do not forget to add RunUpdateProcessorFactory at the end of any
>  chains you define in solrconfig.xml. Otherwise update requests
>  processed by that chain will not actually affect the indexed
>>> data.
> 
> 2. Your example document is missing an “id” field.
> 
> 3. For whatever reason, the pre-trained model "en-ner-person.bin"
>>> doesn’t
> extract anything from text “This is Steve Jobs 2”.  It will extract
>>> “Steve
> Jobs” from text “This is Steve Jobs in white” e.g. though.
> 
> 4. (Not a problem necessarily) You may want to use a multi-valued
>>> “string”
> field for the “dest” field in your update chain, e.g. “people_str”
>>> (“*_str”
> in the default configset is so configured).
> 
> --
> Steve
> www.lucidworks.com
> 
>> On Apr 17, 2018, at 8:23 AM, Alexey Ponomarenko <
>>> alex1989s...@gmail.com>
> wrote:
>> 
>> Hi once more I am trying to implement named entities extraction using
> this
>> manual
>> 
> 
>>> https://lucene.apache.org/solr/7_3_0//solr-analysis-extras/org/apache/solr/update/processor/OpenNLPExtractNamedEntitiesUpdateProcessorFactory.html
>> 
>> I am modified solrconfig.xml like this:
>> 
>> 
>>  class="solr.OpenNLPExtractNamedEntitiesUpdateProcessorFactory">
>>   opennlp/en-ner-person.bin
>>   text_opennlp
>>   description_en
>>   content
>> 
>> 
>> 
>> But when I was trying to add data using:
>> 
>> *request:*
>> 
>> POST
>> 
> 
>>> http://localhost:8983/solr/numberplate/update?version=2.2=xml=multiple-extract
>> 
>> This is Steve Jobs 2
>> This is text 2> name="content">This is text for content 2
>> 
>> *response*
>> 
>> 
>> 
>>  
>>  0

Re: Sum and aggregation on nested documents field

2018-07-10 Thread Mikhail Khludnev
Hello, JB.
The rule of thumb if that you post what's you have whether it exception or
unexpected result with explain. Here I guess you need something like:
q={!parent which=object_type_s:contact score=total v=$chq}=+
object_type:order {!func}TTC_i
Make sure that + isn't mangled by url ecoding ie you might need to make it
as %2b.

On Tue, Jul 10, 2018 at 11:13 AM jeebix  wrote:

> Thanks for your reply Mikhail, I tried like this :
> "{!parent which=object_type_s:contact score=total v={!func}TTC_i}" in a
> filter query, the same in the sort parameter, I also tried to put this
> query
> "{!parent which=object_type_s:contact score=total}TTC_i" into the q
> parameter, but no results...
> I had a look to the dismax query parser, but I'm not sure it is a way to
> solve my problem.
>
> So what's wrong with my function query ? Did I explicite the child level
> correctly ?
>
> Best
>
> JB
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>


-- 
Sincerely yours
Mikhail Khludnev


Re: Get transaction count from ZooKeeper transaction logs

2018-07-10 Thread Jostein Elvaker Haande
On Wed, 11 Jul 2018 at 02:48, Shawn Heisey  wrote:

> On 7/10/2018 3:32 PM, Jostein Elvaker Haande wrote:
> > I'm trying to find an effective way to find the number of transactions
> You have detailed questions about the inner workings of ZooKeeper.  This
> is not a ZooKeeper mailing list.  It is a Solr mailing list.  The
> ZooKeeper project has its own support resources.  That project is going
> to be in a far better position to have access to the information that
> you need.
>

Thanks Shawn, I've redirect my question to this mailing list.

-- 
Yours sincerely Jostein Elvaker Haande
"A free society is a society where it is safe to be unpopular"
- Adlai Stevenson

http://tolecnal.net -- tolecnal at tolecnal dot net


Re: Get transaction count from ZooKeeper transaction logs

2018-07-10 Thread Shawn Heisey

On 7/10/2018 3:32 PM, Jostein Elvaker Haande wrote:

I'm trying to find an effective way to find the number of transactions
stored in the ZooKeeper transaction logs. The only method I've found so far
is by using the Java class 'org.apache.zookeeper.server.LogFormatter' which
outputs the following after it has formatted a log file:

   EOF reached after 203 txns.

Now I could of course make a script to process each log file through this
log formatter, and extract the count from the last line of stdout, but I'm
wondering if there's an easier method.

I've read through the ZK documentation, and tried the ZK commands (aka The
Four Letter Words) to see if any of these offer this metric, but I could
not see it.


You have detailed questions about the inner workings of ZooKeeper.  This 
is not a ZooKeeper mailing list.  It is a Solr mailing list.  The 
ZooKeeper project has its own support resources.  That project is going 
to be in a far better position to have access to the information that 
you need.


http://zookeeper.apache.org/lists.html
http://zookeeper.apache.org/irc.html

Thanks,
Shawn



Get transaction count from ZooKeeper transaction logs

2018-07-10 Thread Jostein Elvaker Haande
 Hello,

I'm trying to find an effective way to find the number of transactions
stored in the ZooKeeper transaction logs. The only method I've found so far
is by using the Java class 'org.apache.zookeeper.server.LogFormatter' which
outputs the following after it has formatted a log file:

  EOF reached after 203 txns.

Now I could of course make a script to process each log file through this
log formatter, and extract the count from the last line of stdout, but I'm
wondering if there's an easier method.

I've read through the ZK documentation, and tried the ZK commands (aka The
Four Letter Words) to see if any of these offer this metric, but I could
not see it.

So my question is - is there a simpler approach to find this count?

-- 
Yours sincerely Jostein Elvaker Haande
"A free society is a society where it is safe to be unpopular"
- Adlai Stevenson

http://tolecnal.net -- tolecnal at tolecnal dot net


Verifying autoscale events via History Api

2018-07-10 Thread Duncan, Adam
Hi all,

We are trying to confirm that a Autoscaling ‘nodeAdded’ event triggers when a 
node joins the cluster using the History Api.
We’ve added a node to the cluster but do not see any records in the Autoscale 
History endpoint.
As a side note, we had to manually create the .system collection. The history 
endpoint was erroring with '.system collection does not exist'. So we created 
that collection manually. The History Api no longer errors, but has no records
This is for solr 7.3.

Our trigger and listener were created as standard based on the solr 
documentation (below):
triggers:
{

 *   node_added_trigger:
{

*   event: "nodeAdded",
*   waitFor: 1,
*   actions:
[

   *   {
  *   name: "compute_plan",
  *   class: "solr.ComputePlanAction",
},

   *   {
  *   name: "execute_plan",
  *   class: "solr.ExecutePlanAction",
},
],

*   enabled: true,
}
},

  *   listeners:
{

 *   node_added_trigger.system:
{

*   trigger: "node_added_trigger",
*   afterAction: [ ],
*   stage:
[

   *   "STARTED",
   *   "ABORTED",
   *   "SUCCEEDED",
   *   "FAILED",
   *   "BEFORE_ACTION",
   *   "AFTER_ACTION",
   *   "IGNORED",
],

*   class: "org.apache.solr.cloud.autoscaling.SystemLogListener",
*   beforeAction: [ ],
}
},

Any thoughts on why we’re not seeing any event history ?

Thanks!
Adam


MetricRegistry instances for transient cores are not getting GC'd

2018-07-10 Thread nandakishorek
I took a heap dump when Solr heap usage kept growing to 32GB and comes down
to 15GB after GC.
Heap dump analysis shows there are MetricRegistry instances for transient
cores.i.e., cores with "isLoaded" as false in STATUS output.


 

 

I validated that this problem happens on the default configset as well (I
tested it with 5 cores and transientCache size set to 2).

Please help.

My environment is as follows 
 * Solr 7.2.1 in standalone mode. 
 * 32GB heap 
 * 150 cores with data getting continuously ingested to ~10 cores and all of 
the cores queried. 
 * transient cache size is set to 30. 

The solr.xml is as follows 
 
 

  32 
  true 
  ${configSetBaseDir:configsets} 

   
${socketTimeout:60} 
${connTimeout:6} 
   




--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Sorting and pagination in Solr json range facet

2018-07-10 Thread Anil
HI Eric,

i mean pagination is offset and limit for facet results. Basically i am
trying to sort the daily totals (from json facet field) and apply offset,
limit to the buckets.

json.facet=
{
daily_totals: {
type: range,
field: daily_window,
start : "2017-11-01T00:00:00Z",
end : "2018-03-14T00:00:00Z",
gap:"%+1DAY",
sort: daily_total,
mincount:1,
facet: {
daily_total: "sum(daily_views)"
}
}
}

please let me know if you have any questions. thanks.

Regards,
Anil

On 10 July 2018 at 20:22, Erick Erickson  wrote:

> What exactly do you mean by "pagination" here? Facets are computed over
> the entire result set. That is, if the number of documents found for the
> query
> is 1,000,000, the facets are returned counted over all 1M docs, even if
> your
> rows parameter is 10. The same numbers will be returned for facets
> regardless of the start and rows parameters.
>
> This feels like an XY problem, you're asking how to do X (paginate facets)
> to solve problem Y, but haven't stated what Y is. What's the use-case here?
>
> Best,
> Erick
>
>
>
> On Tue, Jul 10, 2018 at 5:36 AM, Anil  wrote:
> > Hi,
> >
> > Good Morning.
> >
> > I am trying solr json facet features. sort, offset, limit fields are not
> > working for Range facet.
> >
> > and could not find the support in the documentation. is there any way to
> > achieve sort and pagination for Range facet ? please help.
> >
> > Documentation of range facet says -
> >
> > Parameters:
> >
> >- field – The numeric field or date field to produce range buckets
> from
> >- mincount – Minimum document count for the bucket to be included in
> the
> >response. Defaults to 0.
> >- start – Lower bound of the ranges
> >- end – Upper bound of the ranges
> >- gap – Size of each range bucket produced
> >- hardend – A boolean, which if true means that the last bucket will
> end
> >at “end” even if it is less than “gap” wide. If false, the last
> bucket will
> >be “gap” wide, which may extend past “end”.
> >- other – This param indicates that in addition to the counts for each
> >range constraint between facet.range.start and facet.range.end, counts
> >should also be computed for…
> >   - "before" all records with field values lower then lower bound of
> >   the first range
> >   - "after" all records with field values greater then the upper
> bound
> >   of the last range
> >   - "between" all records with field values between the start and end
> >   bounds of all ranges
> >   - "none" compute none of this information
> >   - "all" shortcut for before, between, and after
> >- include – By default, the ranges used to compute range faceting
> >between facet.range.start and facet.range.end are inclusive of their
> lower
> >bounds and exclusive of the upper bounds. The “before” range is
> exclusive
> >and the “after” range is inclusive. This default, equivalent to lower
> >below, will not result in double counting at the boundaries. This
> behavior
> >can be modified by the facet.range.include param, which can be any
> >combination of the following options…
> >   - "lower" all gap based ranges include their lower bound
> >   - "upper" all gap based ranges include their upper bound
> >   - "edge" the first and last gap ranges include their edge bounds
> (ie:
> >   lower for the first one, upper for the last one) even if the
> > corresponding
> >   upper/lower option is not specified
> >   - "outer" the “before” and “after” ranges will be inclusive of
> their
> >   bounds, even if the first or last ranges already include those
> boundaries.
> >   - "all" shorthand for lower, upper, edge, outer
> >
> >
> >
> >  Thanks,
> > Anil
>


Re: Sum and aggregation on nested documents field

2018-07-10 Thread jeebix
Thanks for your reply Mikhail, I tried like this :
"{!parent which=object_type_s:contact score=total v={!func}TTC_i}" in a
filter query, the same in the sort parameter, I also tried to put this query
"{!parent which=object_type_s:contact score=total}TTC_i" into the q
parameter, but no results...
I had a look to the dismax query parser, but I'm not sure it is a way to
solve my problem.

So what's wrong with my function query ? Did I explicite the child level
correctly ?

Best

JB



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Sum and aggregation on nested documents field

2018-07-10 Thread jeebix
Thanks for your reply Mikhail, I tried like this :
"{!parent which=object_type_s:contact score=total v={!func}TTC_i}" in a
filter query, the same in the sort parameter, I also tried to put this query
"{!parent which=object_type_s:contact score=total}TTC_i" into the q
parameter, but no results...
I had a look to the dismax query parser, but I'm not sure it is a way to
solve my problem.

So what's wrong with my function query ? Did I explicite the child level
correctly ?

Best

JB



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Sorting and pagination in Solr json range facet

2018-07-10 Thread Erick Erickson
What exactly do you mean by "pagination" here? Facets are computed over
the entire result set. That is, if the number of documents found for the query
is 1,000,000, the facets are returned counted over all 1M docs, even if your
rows parameter is 10. The same numbers will be returned for facets
regardless of the start and rows parameters.

This feels like an XY problem, you're asking how to do X (paginate facets)
to solve problem Y, but haven't stated what Y is. What's the use-case here?

Best,
Erick



On Tue, Jul 10, 2018 at 5:36 AM, Anil  wrote:
> Hi,
>
> Good Morning.
>
> I am trying solr json facet features. sort, offset, limit fields are not
> working for Range facet.
>
> and could not find the support in the documentation. is there any way to
> achieve sort and pagination for Range facet ? please help.
>
> Documentation of range facet says -
>
> Parameters:
>
>- field – The numeric field or date field to produce range buckets from
>- mincount – Minimum document count for the bucket to be included in the
>response. Defaults to 0.
>- start – Lower bound of the ranges
>- end – Upper bound of the ranges
>- gap – Size of each range bucket produced
>- hardend – A boolean, which if true means that the last bucket will end
>at “end” even if it is less than “gap” wide. If false, the last bucket will
>be “gap” wide, which may extend past “end”.
>- other – This param indicates that in addition to the counts for each
>range constraint between facet.range.start and facet.range.end, counts
>should also be computed for…
>   - "before" all records with field values lower then lower bound of
>   the first range
>   - "after" all records with field values greater then the upper bound
>   of the last range
>   - "between" all records with field values between the start and end
>   bounds of all ranges
>   - "none" compute none of this information
>   - "all" shortcut for before, between, and after
>- include – By default, the ranges used to compute range faceting
>between facet.range.start and facet.range.end are inclusive of their lower
>bounds and exclusive of the upper bounds. The “before” range is exclusive
>and the “after” range is inclusive. This default, equivalent to lower
>below, will not result in double counting at the boundaries. This behavior
>can be modified by the facet.range.include param, which can be any
>combination of the following options…
>   - "lower" all gap based ranges include their lower bound
>   - "upper" all gap based ranges include their upper bound
>   - "edge" the first and last gap ranges include their edge bounds (ie:
>   lower for the first one, upper for the last one) even if the
> corresponding
>   upper/lower option is not specified
>   - "outer" the “before” and “after” ranges will be inclusive of their
>   bounds, even if the first or last ranges already include those 
> boundaries.
>   - "all" shorthand for lower, upper, edge, outer
>
>
>
>  Thanks,
> Anil


Re: Delta import not working with Oracle in Solr

2018-07-10 Thread Erick Erickson
It's actually quite easy to index from a DB to Solr via SolrJ, here's
an example.

https://lucidworks.com/2012/02/14/indexing-with-solrj/

On Tue, Jul 10, 2018 at 6:06 AM, Shawn Heisey  wrote:
> On 7/8/2018 9:44 AM, shruti suri wrote:
>>
>> I am using solr-6.1.0 version. This is the response I am getting. But
>> every
>> time I run delta import , it fetches same number of records but didn't
>> commit them.
>>
>> 
>> 0:0:42.255
>> 2
>> 10208
>> 0
>> 0
>> 2018-07-08 15:37:31
>> 2018-07-08 15:37:31
>> 2018-07-08 15:38:13
>> 2018-07-08 15:38:13
>> 1
>> 
>
>
> The DIH response won't tell you what went wrong.  It just contains numbers.
>
> You're going to need to find and examine the file named solr.log on the
> server side, look for ERROR and/or WARN log messages after the point where
> the import is started.
>
> If the logfile has been rotated, which happens by default in 6.x if it
> reaches 4MB in size, then there may be more parts to the filename beyond
> solr.log.  Exactly where this file lives is going to depend on how you
> installed and started Solr.  Default locations can vary, from a relative
> path of "server/logs" to an absolute path of "/var/solr/logs/solr.log".  If
> defaults have been overridden, then I can't tell you where the file will be
> without additional information from you.
>
> Thanks,
> Shawn
>


Re: [Solr-Zookeeper] Housekeeping Archival Logs

2018-07-10 Thread Shawn Heisey

On 7/8/2018 8:46 PM, Goh, Qing Hao wrote:

Can I kindly check what is the best practice for housekeeping archival logs for 
Solr & Zookeeper? What I meant here is the best practices recommended by Apache 
and to ensure the middleware is kept healthy?


I have no idea what options ZooKeeper has for log maintenance.  It is a 
separate project, you're on a Solr mailing list.


If you're running the ZK server that's embedded into Solr, then whatever 
ZK logs will go to the solr log, it will not be sent to its own log.  
The embedded ZK server is not recommended for production.


The current version of Solr (7.4.0) handles all rotation and deletion of 
the main solr log with log4j config.  For the GC log, this is handled by 
Java itself.  I'm not entirely sure about the console log, but I think 
that this is overwritten each time Solr starts.  Solr will typically 
silence logs sent to the console once it gets up and running.


If you want to keep logs longer than the current configuration 
specifies, then you will need to change the logging configuration, 
possibly to send the logs to a different piece of software.



The current version of Solr is 6.1.0 and Zookeeper is 3.4.6. Thanks in advance


Solr 6.1.0 handles normal log rotation and deletion with log4j config 
(1.x), but there is a log archival step at startup that is handled in 
the start script.  This involves copying some of the old logs to an 
archive directory.  I'd have to look at the script to discover the 
precise details.  This log archival in the start script was removed in 
7.4.0 with the upgrade to log4j 2.x, because log4j 2.x can do log 
rotation at startup.


Thanks,
Shawn



Re: error on health check from command line

2018-07-10 Thread Satya Marivada
Additional information: Using solr-6.3.0

Also tried, no luck:

./bin/solr healthcheck -c poi -z host1:2181,host2:2181,host3:2181
SOLR_AUTH_TYPE=“basic”
SOLR_AUTHENTICATION_OPTS="-Dbasicauth=username:password"

On Tue, Jul 10, 2018 at 9:09 AM Satya Marivada 
wrote:

> Hi,
>
> How do I supply the basic auth credentials for the below healthcheck that
> has to be done from command line?
>
> $ ./bin/solr healthcheck -c poi -z host1:2181,host2:2181,host3:2181
>
> ERROR: Solr requires authentication for https://host1:15101/solr/shard1/.
> Please supply valid credentials. HTTP code=401
>
> Thanks,
> Satya
>


error on health check from command line

2018-07-10 Thread Satya Marivada
Hi,

How do I supply the basic auth credentials for the below healthcheck that
has to be done from command line?

$ ./bin/solr healthcheck -c poi -z host1:2181,host2:2181,host3:2181

ERROR: Solr requires authentication for https://host1:15101/solr/shard1/.
Please supply valid credentials. HTTP code=401

Thanks,
Satya


Re: Delta import not working with Oracle in Solr

2018-07-10 Thread Shawn Heisey

On 7/8/2018 9:44 AM, shruti suri wrote:

I am using solr-6.1.0 version. This is the response I am getting. But every
time I run delta import , it fetches same number of records but didn't
commit them.


0:0:42.255
2
10208
0
0
2018-07-08 15:37:31
2018-07-08 15:37:31
2018-07-08 15:38:13
2018-07-08 15:38:13
1



The DIH response won't tell you what went wrong.  It just contains numbers.

You're going to need to find and examine the file named solr.log on the 
server side, look for ERROR and/or WARN log messages after the point 
where the import is started.


If the logfile has been rotated, which happens by default in 6.x if it 
reaches 4MB in size, then there may be more parts to the filename beyond 
solr.log.  Exactly where this file lives is going to depend on how you 
installed and started Solr.  Default locations can vary, from a relative 
path of "server/logs" to an absolute path of "/var/solr/logs/solr.log".  
If defaults have been overridden, then I can't tell you where the file 
will be without additional information from you.


Thanks,
Shawn



Sorting and pagination in Solr json range facet

2018-07-10 Thread Anil
Hi,

Good Morning.

I am trying solr json facet features. sort, offset, limit fields are not
working for Range facet.

and could not find the support in the documentation. is there any way to
achieve sort and pagination for Range facet ? please help.

Documentation of range facet says -

Parameters:

   - field – The numeric field or date field to produce range buckets from
   - mincount – Minimum document count for the bucket to be included in the
   response. Defaults to 0.
   - start – Lower bound of the ranges
   - end – Upper bound of the ranges
   - gap – Size of each range bucket produced
   - hardend – A boolean, which if true means that the last bucket will end
   at “end” even if it is less than “gap” wide. If false, the last bucket will
   be “gap” wide, which may extend past “end”.
   - other – This param indicates that in addition to the counts for each
   range constraint between facet.range.start and facet.range.end, counts
   should also be computed for…
  - "before" all records with field values lower then lower bound of
  the first range
  - "after" all records with field values greater then the upper bound
  of the last range
  - "between" all records with field values between the start and end
  bounds of all ranges
  - "none" compute none of this information
  - "all" shortcut for before, between, and after
   - include – By default, the ranges used to compute range faceting
   between facet.range.start and facet.range.end are inclusive of their lower
   bounds and exclusive of the upper bounds. The “before” range is exclusive
   and the “after” range is inclusive. This default, equivalent to lower
   below, will not result in double counting at the boundaries. This behavior
   can be modified by the facet.range.include param, which can be any
   combination of the following options…
  - "lower" all gap based ranges include their lower bound
  - "upper" all gap based ranges include their upper bound
  - "edge" the first and last gap ranges include their edge bounds (ie:
  lower for the first one, upper for the last one) even if the
corresponding
  upper/lower option is not specified
  - "outer" the “before” and “after” ranges will be inclusive of their
  bounds, even if the first or last ranges already include those boundaries.
  - "all" shorthand for lower, upper, edge, outer



 Thanks,
Anil


Re: truncate string field type

2018-07-10 Thread Zahra Aminolroaya
suppose I want to search the "l(i|a)*on k(i|e)*ng". there is a space between
two words. I want solr to retrieve the exact match that these two words or
their other cases are adjacent. If I want to use text field type, each one
of these words are considered as tokens, so solr may bring back other
results too; However, we have strict costumers who only need exact matches
if any result is available not more!



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: truncate string field type

2018-07-10 Thread Alexandre Rafalovitch
Are you sure Solr is the right tool for you? Regexp searches is the really
last resort approach in the domain.

I suggest that maybe you rethink your actual business case (share it here)
to benefiy from tokenization or look if other tools are better.

As it is, you are using a drill to hammer nails.

Regards,
Alex

On Tue, Jul 10, 2018, 2:44 AM Zahra Aminolroaya, 
wrote:

> Thanks Alexandre and Erick. Erick I want to use my regular expression to
> search a field and Solr text field token the document, so the regular
> expression result will not be valid. I want Solr not to token my doc,
> although I will lose some terms using solr string.
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>


Re: CDCR traffic

2018-07-10 Thread Amrit Sarkar
Hi,

In the case of CDCR, assuming both the source and target clusters are SSL
> enabled, can we say that the source clusters’ shard leaders act as clients
> to the target cluster and hence the data is encrypted while its transmitted
> between the clusters?


Yes, that is correct. SSL and Kerberized cluster will have the
payload/updates encrypted. Thank you for pointing it out.

Amrit Sarkar
Search Engineer
Lucidworks, Inc.
415-589-9269
www.lucidworks.com
Twitter http://twitter.com/lucidworks
LinkedIn: https://www.linkedin.com/in/sarkaramrit2
Medium: https://medium.com/@sarkaramrit2

On Mon, Jul 9, 2018 at 3:50 PM, Greenhorn Techie 
wrote:

> Amrit,
>
> Further to the below conversation:
>
> As I understand, Solr supports SSL encryption between nodes within a Solr
> cluster and as well communications to and from clients. In the case of
> CDCR, assuming both the source and target clusters are SSL enabled, can we
> say that the source clusters’ shard leaders act as clients to the target
> cluster and hence the data is encrypted while its transmitted between the
> clusters?
>
> Thanks
>
>
> On 25 June 2018 at 15:56:07, Amrit Sarkar (sarkaramr...@gmail.com) wrote:
>
> Hi Rajeswari,
>
> No it is not. Source forwards the update to the Target in classic manner.
>
> Amrit Sarkar
> Search Engineer
> Lucidworks, Inc.
> 415-589-9269
> www.lucidworks.com
> Twitter http://twitter.com/lucidworks
> LinkedIn: https://www.linkedin.com/in/sarkaramrit2
> Medium: https://medium.com/@sarkaramrit2
>
> On Fri, Jun 22, 2018 at 11:38 PM, Natarajan, Rajeswari <
> rajeswari.natara...@sap.com> wrote:
>
> > Hi,
> >
> > Would like to know , if the CDCR traffic is encrypted.
> >
> > Thanks
> > Ra
> >
>
>


Re: Delta import not working with Oracle in Solr

2018-07-10 Thread Rahul Singh
Agreed. DIH is not an industrial grade ETL tool.. may want to consider other 
options. May want to look into Kafka Connect as an alternative. It has 
connectors for JDBC into Kafka, and from Kafka into Solr.

--
Rahul Singh
rahul.si...@anant.us

Anant Corporation
On Jul 9, 2018, 6:14 AM -0500, Alexandre Rafalovitch , 
wrote:
> I think you are moving so fast it is hard to understand where you need help.
>
> Can you setup one clean smallest issue (maybe as test) and try our original
> suggestions.
>
> Otherwise, nobody has enough attention energy to figure out what is
> happening.
>
> And even then, this list is voluntary help, we are just trying to give you
> pointers the best we can. It is quite possible you have outgrown DIH and
> need to move up to a propper stand alone ETL tool.
>
> Regards,
> Alex
>
> On Sun, Jul 8, 2018, 11:49 PM shruti suri,  wrote:
>
> > Still not working, same issue documents are not getting pushed to index.
> >
> >
> >
> > -
> > Regards
> > Shruti
> > --
> > Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
> >


Re: truncate string field type

2018-07-10 Thread Zahra Aminolroaya
Thanks Alexandre and Erick. Erick I want to use my regular expression to
search a field and Solr text field token the document, so the regular
expression result will not be valid. I want Solr not to token my doc,
although I will lose some terms using solr string.



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html