[
https://issues.apache.org/jira/browse/SOLR-284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12649986#action_12649986
]
Grant Ingersoll commented on SOLR-284:
--------------------------------------
{quote}
I think I like where this is going.
{quote}
Great! I think the nice thing is as Tika grows, we'll get many more formats
all for free. For instance, I saw someone working on a Flash extractor.
{quote}
Currently the default is ext.ignore.und.fl (IGNORE_UNDECLARED_FIELDS) == false,
which means that if Tika returns a metadata field and you haven't made an
explicit mapping from the Tika fieldname to your Solr fieldname, then Solr will
throw an exception and your document add will fail. This doesn't seem sound
very robust for a production environment, unless Tika will only ever use a
finite list of metadata field names. (That doesn't sound plausible, though I
admit I haven't looked into it.) Even in that case, I think I'd rather not have
to set up a mapping for every possible field name in order to get started with
this handler. Would true perhaps be a better default?
{quote}
I guess I was thinking that most people will probably start out with this by
sending their docs through the engine and see what happens. I think an
exception helps them see sooner what they are missing. That being said, I
don't feel particularly strong about it. It's easy enough to set it to true
in the request handler mappings. From what I see of Tika, though, the
possible values for metadata is fixed within a version. Perhaps the bigger
issue is what happens when someone updates Tika to a newer version with newer
Metadata options.
{quote}
ext.capture / CAPTURE_FIELDS: Do you have a use case in mind for this feature,
Grant? The example in the patch is of routing text from <div> tags to one Solr
field while routing text from other tags to a different Solr field. I'm kind of
curious when this would be useful, especially keeping in mind that, in general,
Tika source documents are not HTML, and so when <div> tags are generated
they're as much artifacts of Tika as reflecting anything in the underlying
document. (You could maybe ask a similar question about ext.inx.attr /
INDEX_ATTRIBUTES.)
{quote}
For capture fields, it's similar to a copy field function. Say, for example,
you want a whole document in one field, but also to be able to search within
paragraphs. Then, you could use a capture field on a <p> tag to do that.
Thus, you get the best of both worlds. The Tika output, is XHTML.
Also, since extraction is happening on the server side, I want to make sure we
have lots of options for dealing with the content. I don't know where else one
would have options to muck with the content post-extraction, but pre-indexing.
Hooking into the processor chain is too late, since then the Tika structure is
gone. That's my reasoning, anyway.
Similarly, for index attributes. When extracting from an HTML file, and it
comes across anchor tags (<a>) it will provide the attributes of the tags as
XML attributes. So, one may want to extract out the links separately from the
main content and put them into a separate field.
> Parsing Rich Document Types
> ---------------------------
>
> Key: SOLR-284
> URL: https://issues.apache.org/jira/browse/SOLR-284
> Project: Solr
> Issue Type: New Feature
> Components: update
> Reporter: Eric Pugh
> Assignee: Grant Ingersoll
> Fix For: 1.4
>
> Attachments: libs.zip, rich.patch, rich.patch, rich.patch,
> rich.patch, rich.patch, rich.patch, rich.patch, SOLR-284.patch,
> SOLR-284.patch, solr-word.pdf, source.zip, test-files.zip, test-files.zip,
> test.zip, un-hardcode-id.diff
>
>
> I have developed a RichDocumentRequestHandler based on the CSVRequestHandler
> that supports streaming a PDF, Word, Powerpoint, Excel, or PDF document into
> Solr.
> There is a wiki page with information here:
> http://wiki.apache.org/solr/UpdateRichDocuments
>
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.