[ 
https://issues.apache.org/jira/browse/TIKA-1858?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15150440#comment-15150440
 ] 

raghu commented on TIKA-1858:
-----------------------------

Thanks for your reply.

we have very large PDF,.docx,.xlsx. We are using Tika to extract content and 
dump data in Elastic Search for full-text search.If we unable to chunk these 
large file i don't think tika will process this much large file in single 
request.
may be we end up getting out of memory exception at Tika.

please suggest us better approach for this scenario.



> Unable to extract content from chunked portion of large file
> ------------------------------------------------------------
>
>                 Key: TIKA-1858
>                 URL: https://issues.apache.org/jira/browse/TIKA-1858
>             Project: Tika
>          Issue Type: Bug
>            Reporter: raghu
>
> Hi All,
> we are using Tika server(REST based api) to extract content in .NET 
> Application.
> we need to extract content from very large file(500MB). we want to split this 
> file to chunks and passing request to TIKA. we are able to get any result 
> from TIKA. 
> please help me.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to