[ 
https://issues.apache.org/jira/browse/NIFI-10792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17688578#comment-17688578
 ] 

David Handermann commented on NIFI-10792:
-----------------------------------------

Both of the libraries listed are licensed under Apache Software License 2.0, so 
they could be suitable for inclusion. The excel-streaming-reader library 
appears to provide a wrapper around the streaming capabilities of Apache POI, 
which sounds like the type of feature necessary. The fastexcel-reader library 
also looks like a potential candidate with a minimal set of dependencies for 
reading XML. Either of these could be worth evaluating for a new Excel Record 
Reader.

Downgrading the Apache POI library is not an option due to other bug fixes. The 
size limitation is somewhat arbitrary, but it also provides basic protection 
against the potential for attempting to process unsafe input files. The Apache 
POI library also supports some options for writing Zip elements to temporary 
files, which might be another incremental solution for this particular 
Processor.

> ConvertExcelToCSVProcessor : Failed to convert file over 10MB 
> --------------------------------------------------------------
>
>                 Key: NIFI-10792
>                 URL: https://issues.apache.org/jira/browse/NIFI-10792
>             Project: Apache NiFi
>          Issue Type: Bug
>          Components: Core UI
>    Affects Versions: 1.17.0, 1.16.3, 1.18.0
>            Reporter: mayki
>            Priority: Critical
>              Labels: Excel, csv, processor
>             Fix For: 1.15.3
>
>         Attachments: ConvertExcelToCSVProcessor_1_18_0_with_POI_OLD.PNG, 
> ConvertExcelToCSVProcessor_1_19_1.PNG
>
>
> Hello all,
> It seems all version greater 1.15.3 introduce a failure on the processor 
> *ConvertExcelToCSVProcessor* with this error :
> {code:java}
> Tried to allocate an array of length 101,695,141, but the maximum length for 
> this record type is 100,000,000. If the file is not corrupt or large, please 
> open an issue on bugzilla to request increasing the maximum allowable size 
> for this record type. As a temporary workaround, consider setting a higher 
> override value with IOUtils.setByteArrayMaxOverride() {code}
> I have tested with 2 differences instances nifi version 1.15.3 ==> Work: OK
> And since upgrade in 1.16, 1.17, 1.18 ==> same processsor *failed* with file 
> greater than 10MB.
> Could you help us to correct this bug ?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to