DIH import fails when importing multi-valued field

2019-06-26 Thread Robert Dadzie
Hi All, I'm trygin to use DIH to import about 150k documents to Solr. One of the multi-valued fields I need to import stores about 1500 unique ID per record. I tried increasing the 'ramBufferSizeMB' setting but that didn't help. I get this ArrayIndexOutOfBoundsException error and I can't make

Is it possible to documents using fields from multiple entities?

2019-06-26 Thread Robert Dadzie
I have 2 entities and I need to devise a query that can filter documents using fields from both entities, is this possible, if so can you provide a sample query for how to do so? Thanks, Robert Dadzie If you wish to view the CPA Global group email disclaimer,

Re: SolrInputDocument setField method

2019-06-26 Thread Mark Sholund
I noticed this yesterday as well. The toString() and jsonStr() (in later versions) of SolrJ both include things like toString(): {id=id=[foo123](https://www.nga.mil/careers/studentopp/Pages/default.aspx), ...} or jsonStr():

Re: Large Filter Query

2019-06-26 Thread Lucky Sharma
Thanks, David, Shawn, Jagdish Help and suggestions are really appreciated. Regards, Lucky Sharma On Thu, Jun 27, 2019 at 12:50 AM Shawn Heisey wrote: > > On 6/26/2019 12:56 PM, Lucky Sharma wrote: > > @Shawn: Sorry I forgot to mention the corpus size: the corpus size is > > around 3 million

Re: Large Filter Query

2019-06-26 Thread Shawn Heisey
On 6/26/2019 12:56 PM, Lucky Sharma wrote: @Shawn: Sorry I forgot to mention the corpus size: the corpus size is around 3 million docs, where we need to query for 1500 docs and run aggregations, sorting, search on them. Assuming the documents aren't HUGE, that sounds like something Solr

Re: Large Filter Query

2019-06-26 Thread jai dutt
Then term query parser is best way to do that. You can check below link for performance detail. http://yonik.com/solr-terms-query/ n Thu, 27 Jun, 2019, 12:31 AM Lucky Sharma, wrote: > Thanks, Jagdish > But what if we need to perform search and filtering on those 1.5k doc > ids results, also

Re: Large Filter Query

2019-06-26 Thread David Hastings
yeah there is a performance hit but that is expected. in my scenario i pass sometimes a few thousand using this method, but i pre-process my results since its a set. you will not have any issues if you are using POST with the uri length. On Wed, Jun 26, 2019 at 3:02 PM Lucky Sharma wrote: >

Re: Large Filter Query

2019-06-26 Thread Lucky Sharma
Thanks, Jagdish But what if we need to perform search and filtering on those 1.5k doc ids results, also for URI error, we can go with the POST approach, and what if the data is not sharded. Regards, Lucky Sharma On Thu, Jun 27, 2019 at 12:28 AM jai dutt wrote: > > 1. No Solr is not for id

Re: Large Filter Query

2019-06-26 Thread jai dutt
1. No Solr is not for id search. rdms a better option. 2. Yes correct it going to impact query performance. And you may got large uri error. 3 ya you can pass ids internally by writing any custom parser.or divide data into different shard. On Thu, 27 Jun, 2019, 12:01 AM Lucky Sharma,

Re: Large Filter Query

2019-06-26 Thread Lucky Sharma
@Shawn: Sorry I forgot to mention the corpus size: the corpus size is around 3 million docs, where we need to query for 1500 docs and run aggregations, sorting, search on them. @David: But will that not be a performance hit (resource incentive)? since it will have that many terms to search upon,

Re: Large Filter Query

2019-06-26 Thread David Hastings
you can use the !terms operator and send them separated by a comma: {!terms f=id}id1,id2,..id1499,id1500 and run facets normally On Wed, Jun 26, 2019 at 2:31 PM Lucky Sharma wrote: > Hi all, > > What we are doing is, we will be having a set of unique Ids of solr > document at max 1500,

Re: Large Filter Query

2019-06-26 Thread Shawn Heisey
On 6/26/2019 12:31 PM, Lucky Sharma wrote: What we are doing is, we will be having a set of unique Ids of solr document at max 1500, we need to run faceting and sorting among them. there is no direct search involved. It's a head-on search since we already know the document unique keys

Large Filter Query

2019-06-26 Thread Lucky Sharma
Hi all, What we are doing is, we will be having a set of unique Ids of solr document at max 1500, we need to run faceting and sorting among them. there is no direct search involved. It's a head-on search since we already know the document unique keys beforehand. 1. Is Solr a better use case for

Re: Invoice 6873 from Sobek Digital Hosting and Consulting, LLC 26.06.19

2019-06-26 Thread Mark Sullivan
All, THIS EMAIL IS PHISHING AND IMPERSONATED MY EMAIL ADDRESS. PLEASE IGNORE! Mark From: Mark Sullivan Sent: Wednesday, June 26, 2019 1:29:09 PM Subject: Invoice 6873 from Sobek Digital Hosting and Consulting, LLC 26.06.19 Hi, Mark used box to share

Invoice 6873 from Sobek Digital Hosting and Consulting, LLC 26.06.19

2019-06-26 Thread Mark Sullivan
Hi, Mark used box to share INV-6873 Kindly press REVIEW DOCUMENT to access the secure document Please let us know if there is any skipped invoices. Thank you Mark V. Sullivan CIO & Application Architect Sobek Digital Hosting

Solr 6.6.0 - Multiple DataSources - Performance / Delta Issues - MSSQL(Azure)

2019-06-26 Thread Joseph_Tucker
I've currently got a data configuration that uses multiple dataSources. I have a main dataSource that contains shared inventory data, and individual dataSources that contain price data that differs from database to database. (I have little to no say in how the Databases can be structured) The

Solr 6.6.0 - Indexing Multiple DataSources multiple child entities

2019-06-26 Thread Joseph_Tucker
[Using Solr 6.6.0] I've currently got a few databases that I'm indexing. To give the scenario: I have 10 different shops Each shop will have the same inventory, but different price tags on each item. (i.e. Shop 1 sells Chocolate for $1, and Shop 2 sells Chocolate for $0.95... etc) I'm connecting

Postgres Large Object Facility

2019-06-26 Thread Christopher Spooner
I am using SOLR 7.7.2 and trying to index binary data that is stored in postgresql's large object feature (OID type / lo module) and not directly in the database. Is this possible? If so are there any examples of others configuring SOLR in this way? Attached are my db-data-config and

Re: Replication issue with version 0 index in SOLR 7.5

2019-06-26 Thread Patrick Bordelon
One other question related to this. I know the change was made for a specific problem that was occurring but has this caused a similar problem as mine with anyone else? We're looking to try changing the second 'if' statement to add an extra conditional to prevent it from performing the

Re: SolrInputDocument setField method

2019-06-26 Thread Shawn Heisey
On 6/26/2019 9:52 AM, Vincenzo D'Amore wrote: I have a very basic question related to the SolrInputDocument behaviour. Looking at SolrInputDocument source code I found how the method setField works: public void setField(String name, Object value ) { SolrInputField field = new

Re: SolrInputDocument setField method

2019-06-26 Thread Samuel Kasimalla
Hi Vicenzo, May be looking at the overridden toString() would give you a clue. The second part, I don't think SolrJ holds it it twice(if you are worried about redundant usage of memory), BUT if you haven't used SolrJ so far and wanted to know if this is the format in which it pushes to Solr, I'm

Re: giving weight for SynonymFilterFactory terms

2019-06-26 Thread Ruslan Dautkhanov
Any way in Solr to give weights to synonyms? Thanks. On Sun, Jun 23, 2019 at 4:39 PM Ruslan Dautkhanov wrote: > Hello! > > Is there is a way for Solr to assign weights for terms produced > by SynonymFilterFactory? > > We'd like to give smaller weight for synonyms words/terms injected by >

SolrInputDocument setField method

2019-06-26 Thread Vincenzo D'Amore
Hi all, I have a very basic question related to the SolrInputDocument behaviour. Looking at SolrInputDocument source code I found how the method setField works: public void setField(String name, Object value ) { SolrInputField field = new SolrInputField( name ); _fields.put( name,

Migrating from JdbcDataSource to ContentStreamDataSource

2019-06-26 Thread Reinharn
Im trying to get off my jdbc data source and move to a Streaming data source. I have successfully implemented a node.js api that will push items to my solr index using the /update/json which is defined out of the box as: This process replaces the 'delta' We still have our /dataimport

Suggester not returning all possible completions for a query in a large index

2019-06-26 Thread Christian Ortner
Hello Everyone, I'm using suggesters with Solr 6.4 to get suggestions for a field with a decent number of different values across a large number of documents that is configured like this: vendorSuggester BlendedInfixLookupFactory 600 false DocumentDictionaryFactory

Problems using a suggester component in the /select handler in cloud mode

2019-06-26 Thread Alexandros Paramythis
Hi everyone, Environment: Solr 7.5.0, cloud mode (but problem should be identical in multiple versions, at least in 7.x) Summary: We have a Solr configuration that returns suggestions in the course of a normal search call (i.e., we have a 'suggest' component added to the 'last-components'