[
https://issues.apache.org/jira/browse/FLINK-29689?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Martijn Visser closed FLINK-29689.
----------------------------------
Resolution: Invalid
This is not the Apache NiFi jira, but this is the Apache Flink Jira
> NIFI Performance issue - ConvertExcelToCSVProcessor to handle more data
> -----------------------------------------------------------------------
>
> Key: FLINK-29689
> URL: https://issues.apache.org/jira/browse/FLINK-29689
> Project: Flink
> Issue Type: Improvement
> Components: API / DataSet
> Reporter: Vrinda Palod
> Priority: Major
> Labels: CSV, excel, nifi, poi
> Attachments: image-2022-10-19-18-03-13-983.png,
> image-2022-10-19-18-04-33-092.png
>
>
> Currently , when we try to convert an excel with records more than
> 100,000,000 record type, it throws error and doesnt convert the file to csv
> format.
>
> This can be fixed by adding the below line before reading the excel in the
> code.
> IOUtils.setByteArrayMaxOverride(2147483647);
>
> The value hardcoded is the max length of int array above which we cant handle
> for now.
>
> !image-2022-10-19-18-04-33-092.png!
--
This message was sent by Atlassian Jira
(v8.20.10#820010)