hi, Leonard
我定义了一个ddl和一个dml,sql如下。ddl中没有定义PK,我观察到的现象是:这样在sink到es结果中,结果生成的id是index名,导致只有一条记录。
我将DDL更换为之前版本的with参数(声明使用update-mode =
‘upsert’),不使用1.11最新的with参数,观察到sink结果就正常了。不确定是不是我哪边配置的方式不太对,还是说使用方式有问题。
@[email protected] 我看了下给的源代码,貌似这个是处理upsert的情况,如果不声明pk的话,是不是会是processInsert?
CREATE TABLE ES6_SENSORDATA_SERVER_API (
event varchar,
user_id varchar,
distinct_id varchar,
_date varchar,
_event_time varchar,
recv_time varchar,
code varchar,
_current_project varchar,
api varchar,
elapsed int ,
`start` bigint,
is_err int
) WITH (
'connector' = 'elasticsearch-6',
'hosts' = '<ES_YUNTU.SERVERS>',
'index' = 'flink_sensordata_server_api',
'document-type' = 'default',
'document-id.key-delimiter' = '$',
'sink.bulk-flush.interval' = '1000',
'failure-handler' = 'fail',
'format' = 'json'
)
INSERT INTO ES6_SENSORDATA_SERVER_API
SELECT event,
user_id,
distinct_id,
ts2Date(`time`, 'yyyy-MM-dd') as _date,
ts2Date(`time`, 'yyyy-MM-dd HH:mm:ss.SSS') as _event_time,
ts2Date(recv_time, false, false) as recv_time,
properties.code as code,
properties.`project` as _current_project,
properties.api as api,
properties.elapsed as elapsed,
properties.`start` as `start`,
case when properties.code = '0' then 0 else 1 end as is_err
FROM KafkaEventTopic
where `type` in ('track') and event in ('serverApiReqEvt')
在 2020-07-13 13:44:29,"Leonard Xu" <[email protected]> 写道:
>Hello, fulin
>
>这个问题能提供段可以复现的代码吗?
>
>祝好,
>Leonard Xu
>
>
>> 在 2020年7月13日,09:50,Yangze Guo <[email protected]> 写道:
>>
>> Hi,
>>
>> 如果没有定义主键,ES connector 会把 _id设为null[1],这样ES的Java Client会将_id设为一个随机值[2].
>> 所以应该不会出现您说的这种情况。您那里的ES有没有请求日志之类的,看一下Flink发过来的请求是什么样的。
>>
>> [1]
>> https://github.com/apache/flink/blob/f0eeaec530e001ab02cb889dfe217e25913660c4/flink-connectors/flink-connector-elasticsearch-base/src/main/java/org/apache/flink/streaming/connectors/elasticsearch/table/RowElasticsearchSinkFunction.java#L102
>> [2]
>> https://github.com/elastic/elasticsearch/blob/977230a0ce89a55515dc6ef6452e9f059d9356a2/core/src/main/java/org/elasticsearch/action/index/IndexRequest.java#L509
>>
>> Best,
>> Yangze Guo
>>
>> On Sat, Jul 11, 2020 at 11:33 PM sunfulin <[email protected]> wrote:
>>>
>>> hi,
>>> 根据文档[1]的描述,1.11的es sql connector如果在ddl里没有声明primary
>>> key,将会使用append模式sink数据,并使用es本身生成的id作为document_id。但是我在测试时发现,如果我的ddl里没有定义primary
>>> key,写入时没有正确生成document_id,反而是将index作为id生成了。导致只有最新的一条记录。下面是我的ddl定义:
>>> 不确定是我配置使用的方式不对,还是确实存在bug。。
>>>
>>>
>>> CREATE TABLE ES6_SENSORDATA_OUTPUT (
>>> event varchar,
>>> user_id varchar,
>>> distinct_id varchar,
>>> _date varchar,
>>> _event_time varchar,
>>> recv_time varchar,
>>> _browser_version varchar,
>>> path_name varchar,
>>> _search varchar,
>>> event_type varchar,
>>> _current_project varchar,
>>> message varchar,
>>> stack varchar,
>>> component_stack varchar,
>>> _screen_width varchar,
>>> _screen_height varchar
>>> ) WITH (
>>> 'connector' = 'elasticsearch-6',
>>> 'hosts' = '<ES_YUNTU.SERVERS>',
>>> 'index' = 'flink_sensordata_target_event',
>>> 'document-type' = 'default',
>>> 'document-id.key-delimiter' = '$',
>>> 'sink.bulk-flush.interval' = '1000',
>>> 'failure-handler' = 'fail',
>>> 'format' = 'json'
>>> )
>>>
>>>
>>>
>>>
>>> [1]https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/table/connectors/elasticsearch.html#key-handling