Github user holdenk commented on a diff in the pull request:
https://github.com/apache/spark/pull/14419#discussion_r72896264
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetCompatibilityTest.scala
---
@@ -119,8 +119,19 @@ private[sql] object ParquetCompatibilityTest {
metadata: Map[String, String],
recordWriters: (RecordConsumer => Unit)*): Unit = {
val messageType = MessageTypeParser.parseMessageType(schema)
- val writeSupport = new DirectWriteSupport(messageType, metadata)
- val parquetWriter = new ParquetWriter[RecordConsumer => Unit](new
Path(path), writeSupport)
+ val testWriteSupport = new DirectWriteSupport(messageType, metadata)
--- End diff --
Sure, so the parquetWriter constructors are deprecated now and its been
replaced with a builder interface. For Avro and others there is a standard
builder - but for sort "raw" formats you need to implement your own builder.
This is equivalent to the old constructor we were using - you can see the
deprecation in https://github.com/apache/parquet-mr/pull/199/files as well as
how the builder interface ends up calling an equivalent (now protected)
constructor. Also since our WriteSupport doesn't need to change based on the
configuration we always return the same writesupport regardless of conf.
If it would be useful I can add some of this as a comment in the sourcecode.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]