[ 
https://issues.apache.org/jira/browse/BAHIR-228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17100443#comment-17100443
 ] 

dalongliu commented on BAHIR-228:
---------------------------------

Ok, thanks, how soon can we see it?

> Flink SQL supports kudu sink
> ----------------------------
>
>                 Key: BAHIR-228
>                 URL: https://issues.apache.org/jira/browse/BAHIR-228
>             Project: Bahir
>          Issue Type: New Feature
>          Components: Flink Streaming Connectors
>            Reporter: dalongliu
>            Priority: Major
>
> currently, for Flink-1.10.0, we can use the catalog to store our stream table 
> sink for kudu, it should exist a kudu table sink so we can register it to 
> catalog, and use kudu as a table in SQL environment.
> we can use kudu table sink like this:
> {code:java}
> KuduOptions options = KuduOptions.builder() 
>             .setKuduMaster(kuduMaster) 
>             .setTableName(kuduTable) 
>             .build(); 
> KuduWriterOptions writerOptions = KuduWriterOptions.builder()          
> .setWriteMode(KuduWriterMode.UPSERT) 
>             .setFlushMode(FlushMode.AUTO_FLUSH_BACKGROUND) 
>             .build(); 
> KuduTableSink tableSink = KuduTableSink.builder() 
>             .setOptions(options) 
>             .setWriterOptions(writerOptions) 
>             .setTableSchema(schema) 
>             .build(); 
> tEnv.registerTableSink("kudu", tableSink);  
> tEnv.sqlUpdate("insert into kudu select * from source");
> {code}
> I have used kudu table sink to sync data in company's production environment, 
> the writing  speed at 5w/s in upsert mode



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to