[jira] [Commented] (PARQUET-1633) Integer overflow in ParquetFileReader.ConsecutiveChunkList
[ https://issues.apache.org/jira/browse/PARQUET-1633?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17355329#comment-17355329 ] ASF GitHub Bot commented on PARQUET-1633: - eadwright commented on pull request #902: URL: https://github.com/apache/parquet-mr/pull/902#issuecomment-852434740 @gszadovszky I had a look at your changes. I feel uncomfortable relying on any behaviour at all after an OOM error. Are you sure this is the right approach? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Integer overflow in ParquetFileReader.ConsecutiveChunkList > -- > > Key: PARQUET-1633 > URL: https://issues.apache.org/jira/browse/PARQUET-1633 > Project: Parquet > Issue Type: Bug > Components: parquet-mr >Affects Versions: 1.10.1 >Reporter: Ivan Sadikov >Priority: Major > > When reading a large Parquet file (2.8GB), I encounter the following > exception: > {code:java} > Caused by: org.apache.parquet.io.ParquetDecodingException: Can not read value > at 0 in block -1 in file > dbfs:/user/hive/warehouse/demo.db/test_table/part-00014-tid-1888470069989036737-593c82a4-528b-4975-8de0-5bcbc5e9827d-10856-1-c000.snappy.parquet > at > org.apache.parquet.hadoop.InternalParquetRecordReader.nextKeyValue(InternalParquetRecordReader.java:251) > at > org.apache.parquet.hadoop.ParquetRecordReader.nextKeyValue(ParquetRecordReader.java:207) > at > org.apache.spark.sql.execution.datasources.RecordReaderIterator.hasNext(RecordReaderIterator.scala:40) > at > org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1$$anon$2.getNext(FileScanRDD.scala:228) > ... 14 more > Caused by: java.lang.IllegalArgumentException: Illegal Capacity: -212 > at java.util.ArrayList.(ArrayList.java:157) > at > org.apache.parquet.hadoop.ParquetFileReader$ConsecutiveChunkList.readAll(ParquetFileReader.java:1169){code} > > The file metadata is: > * block 1 (3 columns) > ** rowCount: 110,100 > ** totalByteSize: 348,492,072 > ** compressedSize: 165,689,649 > * block 2 (3 columns) > ** rowCount: 90,054 > ** totalByteSize: 3,243,165,541 > ** compressedSize: 2,509,579,966 > * block 3 (3 columns) > ** rowCount: 105,119 > ** totalByteSize: 350,901,693 > ** compressedSize: 144,952,177 > * block 4 (3 columns) > ** rowCount: 48,741 > ** totalByteSize: 1,275,995 > ** compressedSize: 914,205 > I don't have the code to reproduce the issue, unfortunately; however, I > looked at the code and it seems that integer {{length}} field in > ConsecutiveChunkList overflows, which results in negative capacity for array > list in {{readAll}} method: > {code:java} > int fullAllocations = length / options.getMaxAllocationSize(); > int lastAllocationSize = length % options.getMaxAllocationSize(); > > int numAllocations = fullAllocations + (lastAllocationSize > 0 ? 1 : 0); > List buffers = new ArrayList<>(numAllocations);{code} > > This is caused by cast to integer in {{readNextRowGroup}} method in > ParquetFileReader: > {code:java} > currentChunks.addChunk(new ChunkDescriptor(columnDescriptor, mc, startingPos, > (int)mc.getTotalSize())); > {code} > which overflows when total size of the column is larger than > Integer.MAX_VALUE. > I would appreciate if you could help addressing the issue. Thanks! > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [parquet-mr] eadwright commented on pull request #902: PARQUET-1633 Fix integer overflow
eadwright commented on pull request #902: URL: https://github.com/apache/parquet-mr/pull/902#issuecomment-852434740 @gszadovszky I had a look at your changes. I feel uncomfortable relying on any behaviour at all after an OOM error. Are you sure this is the right approach? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Order of encodings?
I couldn't find anything in the specification on this, but is there any constraint on the ordering of encoded pages in a column for a row group. I think in practice most implementations try to dictionary-encode first and then fallback to another encoding if the dictionary doesn't yield benefits or grows too big. But in theory is an ordering of pages described below possible? 1. Dictionary Encoded Page 2. Plain Encoded Page 3. Dictionary Encoded Page Thanks, Micah
[jira] [Created] (PARQUET-2054) TCP connection leaking when calling appendFile()
Xinli Shang created PARQUET-2054: Summary: TCP connection leaking when calling appendFile() Key: PARQUET-2054 URL: https://issues.apache.org/jira/browse/PARQUET-2054 Project: Parquet Issue Type: New Feature Components: parquet-mr Reporter: Xinli Shang When appendFile() is called, the file reader path is opened but not closed. It caused many TCP connections leaked. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (PARQUET-2051) AvroWriteSupport does not pass Configuration to AvroSchemaConverter on Creation
[ https://issues.apache.org/jira/browse/PARQUET-2051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17355130#comment-17355130 ] ASF GitHub Bot commented on PARQUET-2051: - hailuand commented on pull request #912: URL: https://github.com/apache/parquet-mr/pull/912#issuecomment-852155994 Hi @gszadovszky, can you please review this? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > AvroWriteSupport does not pass Configuration to AvroSchemaConverter on > Creation > --- > > Key: PARQUET-2051 > URL: https://issues.apache.org/jira/browse/PARQUET-2051 > Project: Parquet > Issue Type: Bug >Reporter: Andreas Hailu >Assignee: Andreas Hailu >Priority: Major > > Because of this, we're unable to fully leverage the ThreeLevelListWriter > functionality when trying to write Avro lists out using Parquet through the > AvroParquetOutputFormat. > The following record is used for testing: > Schema: > { "type": "record", "name": "NullLists", "namespace": "com.test", "fields": [ > \{ "name": "KeyID", "type": "string" }, \{ "name": "NullableList", "type": [ > "null", { "type": "array", "items": [ "null", "string" ] } ], "default": null > } ] } > Record (using basic JSON just for display purposes): > { "KeyID": "0", "NullableList": [ "foo", null, "baz" ] } > During testing, we see the following exception: > {quote}{{Caused by: java.lang.ClassCastException: repeated binary array > (STRING) is not a group}} > \{{ at org.apache.parquet.schema.Type.asGroupType(Type.java:250)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport$ThreeLevelListWriter.writeCollection(AvroWriteSupport.java:612)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport$ListWriter.writeList(AvroWriteSupport.java:397)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeValueWithoutConversion(AvroWriteSupport.java:355)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeValue(AvroWriteSupport.java:278)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeRecordFields(AvroWriteSupport.java:191)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.write(AvroWriteSupport.java:165)}} > \{{ at > org.apache.parquet.hadoop.InternalParquetRecordWriter.write(InternalParquetRecordWriter.java:128}} > {quote} > Upon review, it was found that the configuration option that was set in > AvroWriteSupport for the ThreeLevelListWriter, > parquet.avro.write-old-list-structure being set to false, was never shared > with the AvroSchemaConverter. > Once we made this change and tested locally, we observe the record with nulls > in the array being successfully written by AvroParquetOutputFormat. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [parquet-mr] hailuand commented on pull request #912: PARQUET-2051 Pass Configuration to AvroSchemaConverter as to not lose…
hailuand commented on pull request #912: URL: https://github.com/apache/parquet-mr/pull/912#issuecomment-852155994 Hi @gszadovszky, can you please review this? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (PARQUET-2051) AvroWriteSupport does not pass Configuration to AvroSchemaConverter on Creation
[ https://issues.apache.org/jira/browse/PARQUET-2051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17355097#comment-17355097 ] ASF GitHub Bot commented on PARQUET-2051: - hailuand opened a new pull request #912: URL: https://github.com/apache/parquet-mr/pull/912 Make sure you have checked _all_ steps below. ### Jira - [x] My PR addresses the following [Parquet Jira](https://issues.apache.org/jira/projects/PARQUET/issues/PARQUET-2051) issues and references them in the PR title. For example, "PARQUET-1234: My Parquet PR" - https://issues.apache.org/jira/browse/PARQUET-XXX - In case you are adding a dependency, check if the license complies with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). ### Tests - [x] My PR adds the following unit tests __OR__ does not need testing for this extremely good reason: - TestAvroWriteSupport.java ### Commits - [x] My commits all reference Jira issues in their subject lines. In addition, my commits follow the guidelines from "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)": 1. Subject is separated from body by a blank line 1. Subject is limited to 50 characters (not including Jira issue reference) 1. Subject does not end with a period 1. Subject uses the imperative mood ("add", not "adding") 1. Body wraps at 72 characters 1. Body explains "what" and "why", not "how" ### Documentation - [ ] In case of new functionality, my PR adds documentation that describes how to use it. - All the public functions and the classes in the PR contain Javadoc that explain what it does -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > AvroWriteSupport does not pass Configuration to AvroSchemaConverter on > Creation > --- > > Key: PARQUET-2051 > URL: https://issues.apache.org/jira/browse/PARQUET-2051 > Project: Parquet > Issue Type: Bug >Reporter: Andreas Hailu >Assignee: Andreas Hailu >Priority: Major > > Because of this, we're unable to fully leverage the ThreeLevelListWriter > functionality when trying to write Avro lists out using Parquet through the > AvroParquetOutputFormat. > The following record is used for testing: > Schema: > { "type": "record", "name": "NullLists", "namespace": "com.test", "fields": [ > \{ "name": "KeyID", "type": "string" }, \{ "name": "NullableList", "type": [ > "null", { "type": "array", "items": [ "null", "string" ] } ], "default": null > } ] } > Record (using basic JSON just for display purposes): > { "KeyID": "0", "NullableList": [ "foo", null, "baz" ] } > During testing, we see the following exception: > {quote}{{Caused by: java.lang.ClassCastException: repeated binary array > (STRING) is not a group}} > \{{ at org.apache.parquet.schema.Type.asGroupType(Type.java:250)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport$ThreeLevelListWriter.writeCollection(AvroWriteSupport.java:612)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport$ListWriter.writeList(AvroWriteSupport.java:397)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeValueWithoutConversion(AvroWriteSupport.java:355)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeValue(AvroWriteSupport.java:278)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.writeRecordFields(AvroWriteSupport.java:191)}} > \{{ at > org.apache.parquet.avro.AvroWriteSupport.write(AvroWriteSupport.java:165)}} > \{{ at > org.apache.parquet.hadoop.InternalParquetRecordWriter.write(InternalParquetRecordWriter.java:128}} > {quote} > Upon review, it was found that the configuration option that was set in > AvroWriteSupport for the ThreeLevelListWriter, > parquet.avro.write-old-list-structure being set to false, was never shared > with the AvroSchemaConverter. > Once we made this change and tested locally, we observe the record with nulls > in the array being successfully written by AvroParquetOutputFormat. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [parquet-mr] hailuand opened a new pull request #912: PARQUET-2051 Pass Configuration to AvroSchemaConverter as to not lose…
hailuand opened a new pull request #912: URL: https://github.com/apache/parquet-mr/pull/912 Make sure you have checked _all_ steps below. ### Jira - [x] My PR addresses the following [Parquet Jira](https://issues.apache.org/jira/projects/PARQUET/issues/PARQUET-2051) issues and references them in the PR title. For example, "PARQUET-1234: My Parquet PR" - https://issues.apache.org/jira/browse/PARQUET-XXX - In case you are adding a dependency, check if the license complies with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). ### Tests - [x] My PR adds the following unit tests __OR__ does not need testing for this extremely good reason: - TestAvroWriteSupport.java ### Commits - [x] My commits all reference Jira issues in their subject lines. In addition, my commits follow the guidelines from "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)": 1. Subject is separated from body by a blank line 1. Subject is limited to 50 characters (not including Jira issue reference) 1. Subject does not end with a period 1. Subject uses the imperative mood ("add", not "adding") 1. Body wraps at 72 characters 1. Body explains "what" and "why", not "how" ### Documentation - [ ] In case of new functionality, my PR adds documentation that describes how to use it. - All the public functions and the classes in the PR contain Javadoc that explain what it does -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org