Cross join on multivalued field
Hi, I am wondering whether there is planning to implement cross collections join query on multivalued field Thanks Sent from my iPhone
Index rich document and view
Hi, I have all kind of rich documents, such as excel, ppt, PDF, word, jpg ..., I knew Tika or ocr can convert them to text and index it. But when I open the document, the format is changed, how can I keep original document format, is it possible in solr? If not, can I use external field type to save original file and load it when I want to view the document? Thanks Sent from my iPhone
Re: Solr 8.7.0 memory leak?
Thanks Chris, Is the matter to use the config file ? I am using custom config instead of _default, my config is from solr 8.6.2 with custom solrconfig.xml Derrick Sent from my iPhone > On Jan 28, 2021, at 2:48 PM, Chris Hostetter wrote: > > > FWIW, I just tried using 8.7.0 to run: >bin/solr -m 200m -e cloud -noprompt > > And then setup the following bash one liner to poll the heap metrics... > > while : ; do date; echo "node 8989" && (curl -sS > http://localhost:8983/solr/admin/metrics | grep memory.heap); echo "node > 7574" && (curl -sS http://localhost:8983/solr/admin/metrics | grep > memory.heap) ; sleep 30; done > > ...what i saw was about what i expected ... heap usage slowly grew on both > nodes as bits of garbage were generated (as expected cosidering the > metrics requests, let alone typical backgroup threads) until eventually it > garbage collected back down to low usage w/o ever encountering an OOM or > crash... > > > Thu Jan 28 12:38:47 MST 2021 > node 8989 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.7613688659667969, > "memory.heap.used":159670624, > node 7574 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.7713688659667969, > "memory.heap.used":161767776, > Thu Jan 28 12:39:17 MST 2021 > node 8989 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.7813688659667969, > "memory.heap.used":163864928, > node 7574 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.7913688659667969, > "memory.heap.used":165962080, > Thu Jan 28 12:39:47 MST 2021 > node 8989 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.8063688659667969, > "memory.heap.used":169107808, > node 7574 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.8113688659667969, > "memory.heap.used":170156384, > Thu Jan 28 12:40:17 MST 2021 > node 8989 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.3428504943847656, > "memory.heap.used":71900960, > node 7574 > "memory.heap.committed":209715200, > "memory.heap.init":209715200, > "memory.heap.max":209715200, > "memory.heap.usage":0.3528504943847656, > "memory.heap.used":73998112, > > > > > > > -Hoss > http://www.lucidworks.com/
Solr 8.7.0 memory leak?
Hi, I am using solr 8.7.0, centos 7, java 8. I just created a few collections and no data, memory keeps growing but never go down, until I got OOM and solr is killed Any reason? Thanks Sent from my iPhone
Cannot start solr because oom
Hi there, I use default settings to start solr , I set heap to 6G, I created 10 collections with 1node and 1 replica, however, there is not much data at all, just 100 documents. My server is 32 G memory and 4 core cpu, ssd drive 300g It was ok when i created 5 collections. It got oom killed when 10 collections are created. Please, no data in new collections. Also I found that solr doesn’t do garbage collection when the 6G is used ( from dashboard, jvm usage is reached 6 g) Any idea or setting needs to be done? Thanks Derrick Sent from my iPhone
Re: read/write on different node?
Thanks Mohandoss, It is realtime inserting and updating if a query is executed. My environment is three nodes, 3 shards and 2 replicas. I noticed there was master slave mode in the old version, but for solr cloud, I don’t know whether it is doable. Derrick Sent from my iPhone > On Jan 21, 2021, at 3:15 AM, Doss wrote: > > Hi, > > You haven't shared information about your environment and how frequently > you are commiting the changes, whether your user searching collection gets > real time inserts / updates etc., > > but if you are not doing any real time analysis with the user query > information, you can store the information in mysql table and do a batch > import(full or delta). If you want the current approach then try increasing > the commit interval. > > Recently I have asked the same question and get to know that we can use > shards.preference (Replica Types) to achieve our requirement > https://lucene.472066.n3.nabble.com/Solrcloud-Reads-on-specific-nodes-tp4467568.html > > - Mohandoss > >> On Wed, Jan 20, 2021 at 7:22 PM Luke wrote: >> >> Hi, >> >> I have one data collection on 3 shards and 2 replicas, user searches on it. >> Also I log all user queries and save to another collection on the same solr >> cloud, but user queries are very slow when there are a lot of logs to be >> written to the log collection. >> >> any solution for me, please advise. or does solr support separate write >> operation on different node and read on other nodes? >>