There is a more fundamental problem here: Solr/Lucene index only
implements one table. If you have data from multiple tables in a
normalized index, you have denormalize the multi-table DB schema to
make a single-table Solr/Lucene index.
Your indexing will probably be faster if you a join in SQL
HI Lance,
I think you are discussing a different issue here. We are talking about
each row from each table represents a document in index. You look to discuss
about some documents may have multi-value fields which are stored in a
separate table in RDBMS because of normalization.
On Mon, Nov
I have the same problem. I had thought we could specify multiple document
blah blah blah/documents, each of which is mapping one table in the RDBMS.
But I found it was not the case. It only picks the first documentblah blah
blah/document to do indexing.
I think Rupert's and my request are pretty
I have figured out a way to solve this problem: just specify a
single document blah blah blah /document. Under document, specify
multiple top level entity entries, each of which corresponds to one table
data.
So each top level entry will map one row in it to a document in Lucene
index. document
On Sun, Nov 8, 2009 at 8:25 AM, Bertie Shen bertie.s...@gmail.com wrote:
I have figured out a way to solve this problem: just specify a
single document blah blah blah /document. Under document, specify
multiple top level entity entries, each of which corresponds to one table
data.
So each
You can only have one document tag and the entities must be nested
within that.
From the wiki, if you issue a simple /dataimport?command=full-import
all top level entities will be processed.
Maybe I should be more clear: I have multiple tables in my DB that I
need to save to my Solr index. In
I am using the DataImportHandler with a JDBC datasource. From my
understanding of DIH, for each of my content types e.g. Blog posts,
Mesh Categories, etc I would construct a series of document/entity
sets, like
dataConfig
dataSource driver=com.mysql.jdbc.Driver url=jdbc:mysql:// /
!--
Maybe I should be more clear: I have multiple tables in my DB that I
need to save to my Solr index. In my app code I have logic to persist
each table, which maps to an application model to Solr. This is fine.
I am just trying to speed up indexing time by using DIH instead of
going through my
DIH allows only document tag. you may have multiple root entity
tags and you may invoke them by name(s). When no name is passed all
root entities are invoked one after another.
On Wed, Sep 9, 2009 at 5:12 AM, Rupert Fiascorufia...@gmail.com wrote:
Maybe I should be more clear: I have multiple