[
https://issues.apache.org/jira/browse/FLINK-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16629902#comment-16629902
]
Fabian Hueske commented on FLINK-10447:
---------------------------------------
Hi [~Suxing Lee], thanks for opening this issue. I agree that a TableSink to
write results to a file system would be a great feature. In fact, we have
thought about this as well.
However, this feature requires a lot more that what you did in your PR.
* The way that external tables (sources and sinks) are defined was improved in
recent versions. Tables are defined via properties and instantiated by
factories. Please have a look at recent documentation.
* Flink 1.6 features a new sink function (StreamingFileSink) that provides
better support for writing to S3 and blocking formats like Parquet or ORC. A
StreamingFileTableSink should use this function instead of BucketingFileSink.
* The FileTableSink should support different formats (CSV, JSON, Avro,
Parquet, ORC, ...
Unfortunately, your PR does not consider any of these points and cannot be
merged.
As stated in the contribution guidelines, it is a good idea to discuss larger
features and gather requirements and ideas before implementing them.
Thanks, Fabian
> Create Bucketing Table Sink.
> ----------------------------
>
> Key: FLINK-10447
> URL: https://issues.apache.org/jira/browse/FLINK-10447
> Project: Flink
> Issue Type: New Feature
> Components: Table API & SQL
> Reporter: Suxing Lee
> Priority: Major
> Labels: pull-request-available
> Fix For: 1.7.0
>
>
> It would be nice to integrate the table APIs with the HDFS connectors so that
> the rows in the tables can be directly pushed into HDFS.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)