Tan Kim created FLINK-32273:
-------------------------------
Summary: What is org.apache.flink.avro.generated.record_json?
Key: FLINK-32273
URL: https://issues.apache.org/jira/browse/FLINK-32273
Project: Flink
Issue Type: Bug
Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table
SQL / API
Reporter: Tan Kim
I'm trying to relay a topic from kafka to another kafka.
This is the original record in source topic.
{code:java}
"json": {
"eventName": "event-ABC",
...
} {code}
The source is json format and sink is avro format with confluent-schema
registry.
Here is my code.
{code:java}
tableEnv.executeSql("CREATE TABLE source_table (..) WITH (
'connector'='kafka',
'format'='json',
)")
tableEnv.executeSql("CREATE TABLE sink_table WITH (
'connector'='kafka',
'format'='avro-confluent',
..
)"){code}
If I run this code without 'value.avro-confluent.subject' configuration, the
record is some like this.
{code:java}
{
"json": {
"org.apache.flink.avro.generated.record_json": {
"eventName": {
"string": "event-ABC"
},
..
}
} {code}
I don't understand why flink-avro inserts
"org.apache.flink.avro.generated.record_json" between `json` and `eventName`.
Also `eventName` is not just 'event-ABC' but `string: event-ABC`.
Is this bug? or something I missed?
--
This message was sent by Atlassian Jira
(v8.20.10#820010)