[ 
https://issues.apache.org/jira/browse/FLINK-10447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16629902#comment-16629902
 ] 

Fabian Hueske commented on FLINK-10447:
---------------------------------------

Hi [~Suxing Lee], thanks for opening this issue. I agree that a TableSink to 
write results to a file system would be a great feature. In fact, we have 
thought about this as well.

However, this feature requires a lot more that what you did in your PR.
 * The way that external tables (sources and sinks) are defined was improved in 
recent versions. Tables are defined via properties and instantiated by 
factories. Please have a look at recent documentation.
 * Flink 1.6 features a new sink function (StreamingFileSink) that provides 
better support for writing to S3 and blocking formats like Parquet or ORC. A 
StreamingFileTableSink should use this function instead of BucketingFileSink.
 * The FileTableSink should support different formats (CSV, JSON, Avro, 
Parquet, ORC, ...

Unfortunately, your PR does not consider any of these points and cannot be 
merged.

As stated in the contribution guidelines, it is a good idea to discuss larger 
features and gather requirements and ideas before implementing them.

Thanks, Fabian

> Create Bucketing Table Sink.
> ----------------------------
>
>                 Key: FLINK-10447
>                 URL: https://issues.apache.org/jira/browse/FLINK-10447
>             Project: Flink
>          Issue Type: New Feature
>          Components: Table API & SQL
>            Reporter: Suxing Lee
>            Priority: Major
>              Labels: pull-request-available
>             Fix For: 1.7.0
>
>
> It would be nice to integrate the table APIs with the HDFS connectors so that 
> the rows in the tables can be directly pushed into HDFS.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to