Cross join on multivalued field

2021-02-23 Thread Luke Oak
Hi,

I am wondering whether there is planning to implement cross collections join 
query on multivalued field 

Thanks 

Sent from my iPhone

Index rich document and view

2021-02-10 Thread Luke Oak
Hi,

I have all kind of rich documents, such as excel, ppt, PDF, word, jpg ..., I 
knew Tika or ocr can convert them to text and index it. But when I open the 
document, the format is changed,  how can I keep original document format, is 
it possible in solr?

If not, can I use external field type to save original file and load it when I 
want to view the document?

Thanks 

Sent from my iPhone

Re: Solr 8.7.0 memory leak?

2021-01-28 Thread Luke Oak
Thanks Chris,  

Is the matter to use the config file ? I am using custom config instead of 
_default, my config is from solr 8.6.2 with custom solrconfig.xml

Derrick

Sent from my iPhone

> On Jan 28, 2021, at 2:48 PM, Chris Hostetter  wrote:
> 
> 
> FWIW, I just tried using 8.7.0 to run:
>bin/solr -m 200m -e cloud -noprompt
> 
> And then setup the following bash one liner to poll the heap metrics...
> 
> while : ; do date; echo "node 8989" && (curl -sS 
> http://localhost:8983/solr/admin/metrics | grep memory.heap); echo "node 
> 7574" && (curl -sS http://localhost:8983/solr/admin/metrics | grep 
> memory.heap) ; sleep 30; done
> 
> ...what i saw was about what i expected ... heap usage slowly grew on both 
> nodes as bits of garbage were generated (as expected cosidering the 
> metrics requests, let alone typical backgroup threads) until eventually it 
> garbage collected back down to low usage w/o ever encountering an OOM or 
> crash...
> 
> 
> Thu Jan 28 12:38:47 MST 2021
> node 8989
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.7613688659667969,
>  "memory.heap.used":159670624,
> node 7574
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.7713688659667969,
>  "memory.heap.used":161767776,
> Thu Jan 28 12:39:17 MST 2021
> node 8989
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.7813688659667969,
>  "memory.heap.used":163864928,
> node 7574
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.7913688659667969,
>  "memory.heap.used":165962080,
> Thu Jan 28 12:39:47 MST 2021
> node 8989
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.8063688659667969,
>  "memory.heap.used":169107808,
> node 7574
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.8113688659667969,
>  "memory.heap.used":170156384,
> Thu Jan 28 12:40:17 MST 2021
> node 8989
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.3428504943847656,
>  "memory.heap.used":71900960,
> node 7574
>  "memory.heap.committed":209715200,
>  "memory.heap.init":209715200,
>  "memory.heap.max":209715200,
>  "memory.heap.usage":0.3528504943847656,
>  "memory.heap.used":73998112,
> 
> 
> 
> 
> 
> 
> -Hoss
> http://www.lucidworks.com/


Solr 8.7.0 memory leak?

2021-01-27 Thread Luke Oak
Hi, I am using solr 8.7.0, centos 7, java 8.

I just created a few collections and no data, memory keeps growing but never go 
down, until I got OOM and solr is killed 

Any reason?

Thanks

Sent from my iPhone

Cannot start solr because oom

2021-01-23 Thread Luke Oak
Hi there,

I use default settings to start solr , I set heap to 6G, I created 10 
collections with 1node and 1 replica, however, there is not much data at all, 
just 100 documents.

My server is 32 G memory and 4 core cpu, ssd drive 300g

It was ok when i created 5 collections. It got oom killed when 10 collections 
are created. Please, no data in new collections. 

Also I found that solr doesn’t do garbage collection when the 6G is used ( from 
dashboard, jvm usage is reached 6 g)

Any idea or setting needs to be done?

Thanks

Derrick

Sent from my iPhone

Re: read/write on different node?

2021-01-21 Thread Luke Oak
Thanks Mohandoss,

It is realtime inserting and updating if a query is executed. My environment is 
three nodes, 3 shards and 2 replicas.  

I noticed there was master slave mode in the old version,  but for solr cloud, 
I don’t know whether it is doable.

Derrick

Sent from my iPhone

> On Jan 21, 2021, at 3:15 AM, Doss  wrote:
> 
> Hi,
> 
> You haven't shared information about your environment and how frequently
> you are commiting the changes, whether your user searching collection gets
> real time inserts / updates etc.,
> 
> but if you are not doing any real time analysis with the user query
> information, you can store the information in mysql table and do a batch
> import(full or delta). If you want the current approach then try increasing
> the commit interval.
> 
> Recently I have asked the same question and get to know that we can use
> shards.preference (Replica Types) to achieve our requirement
> https://lucene.472066.n3.nabble.com/Solrcloud-Reads-on-specific-nodes-tp4467568.html
> 
> - Mohandoss
> 
>> On Wed, Jan 20, 2021 at 7:22 PM Luke  wrote:
>> 
>> Hi,
>> 
>> I have one data collection on 3 shards and 2 replicas, user searches on it.
>> Also I log all user queries and save to another collection on the same solr
>> cloud, but user queries are very slow when there are a lot of logs to be
>> written to the log collection.
>> 
>> any solution for me, please advise. or does solr support separate write
>> operation on different node and read on other nodes?
>>