Skip to content

[C#] Large record batch is written with negative buffer length #23458

@asfimport

Description

@asfimport

I have a 30 gig arrow file with 100 batches.  the largest batch in the file causes get batch to fail - All other batches load fine. in 14.11 the individual batch errors.. in 15.1.1 the batch crashes R studio when it is used

14.1.1

>  rbn <- data_rbfr$get_batch(x)
Error in ipc__RecordBatchFileReader_ReadRecordBatch(self, i) : 
Invalid: negative malloc size
  

15.1.1

rbn <- data_rbfr$get_batch(x)  works!
df <- as.data.frame(rbn) - Crashes R Studio! 

 

Update

I put the data in the batch into a separate file.  The file size is over 2 gigs. 

Using 15.1.1, when I try to load this entire file via read_arrow it also fails.

ar <- arrow::read_arrow("e:\\temp\\file.arrow") 
Error in Table__from_RecordBatchFileReader(batch_reader) :
 Invalid: negative malloc size

 

Reporter: Anthony Abate / @abbotware

Original Issue Attachments:

Note: This issue was originally created as ARROW-7156. Please see the migration documentation for further details.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions