[
https://issues.apache.org/jira/browse/FLINK-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15270677#comment-15270677
]
ASF GitHub Bot commented on FLINK-1996:
---------------------------------------
Github user yjshen commented on the pull request:
https://github.com/apache/flink/pull/1961#issuecomment-216877262
Hi @fhueske , thanks for the explanation. If I get it correctly, the
current `toSink` API is a general one to allow write table content to a large
variety of `Sinks` without blow up `flink-table` modules dependencies, the
design seems quite reasonable for me now.
BTW, if we are going to support some **native** output format, the
`register` & `reflection` seems a feasible approach, by doing this, we can not
only do
``` scala
t.toSink("csv").option("path", "/foo").option("fieldDelim", "|")
```
in Table API but also
``` sql
insert overwrite into parquet_table_a select * from table_b
```
in SQL.
> Add output methods to Table API
> -------------------------------
>
> Key: FLINK-1996
> URL: https://issues.apache.org/jira/browse/FLINK-1996
> Project: Flink
> Issue Type: Improvement
> Components: Table API
> Affects Versions: 0.9
> Reporter: Fabian Hueske
> Assignee: Fabian Hueske
>
> Tables need to be converted to DataSets (or DataStreams) to write them out.
> It would be good to have a way to emit Table results directly for example to
> print, CSV, JDBC, HBase, etc.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)