[
https://issues.apache.org/jira/browse/SPARK-29248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16938050#comment-16938050
]
Ximo Guanter commented on SPARK-29248:
--------------------------------------
Thanks for the quick reply [~kabhwan]! The issue you linked to sound like a
very interesting feature, but I feel this issue is different from it. I'm not
trying to impose a specific clustering or sorting requirement to the data. The
requirement here is only providing information to the writer, not changing the
execution plan in any way. I don't think that is achievable with the
RequiresClustering interface, since the information flow is the opposite in
that case (the writers imposes a number of partitions to Spark).
> Pass in number of partitions to BuildWriter
> -------------------------------------------
>
> Key: SPARK-29248
> URL: https://issues.apache.org/jira/browse/SPARK-29248
> Project: Spark
> Issue Type: Sub-task
> Components: SQL
> Affects Versions: 3.0.0
> Reporter: Ximo Guanter
> Priority: Major
>
> When implementing a ScanBuilder, we require the implementor to provide the
> schema of the data and the number of partitions.
> However, when someone is implementing WriteBuilder we only pass them the
> schema, but not the number of partitions. This is an asymetrical developer
> experience. Passing in the number of partitions on the WriteBuilder would
> enable data sources to provision their write targets before starting to
> write. For example, it could be used to provision a Kafka topic with a
> specific number of partitions.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]