chacha.tang created HUDI-3693:
---------------------------------
Summary: error will be occured when writing data into hudi using
flink job more than once in same flink session
Key: HUDI-3693
URL: https://issues.apache.org/jira/browse/HUDI-3693
Project: Apache Hudi
Issue Type: Bug
Components: flink
Reporter: chacha.tang
环境:
hudi: 0.10.1
flink:1.13.2
flink cluster setup in standalone mode, and use the below flink sql to launch
the flink job
#create the cow sink table
CREATE TABLE t1(
uuid VARCHAR(20) PRIMARY KEY NOT ENFORCED,
name VARCHAR(10),
age INT,
ts TIMESTAMP(3),
`partition` VARCHAR(20)
)
PARTITIONED BY (`partition`)
WITH (
'connector' = 'hudi',
'path' = '/user/hive/warehouse/hudi.db/t1',
'write.tasks' = '1',
'table.type' = 'COPY_ON_WRITE'
);
#write data first
INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 00:00:01','par1')
#write data in twice
INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 00:00:01','par1')
the first time the flink job is finished successfully, but it failed in twice,
and the below exception occured
org.apache.hudi.common.fs.HoodieWrapperFileSystem cannot be cast to
org.apache.hudi.common.fs.HoodieWrapperFileSystem
I search issues, and found issue#3885 mentioned same problem, But the root
casuse is different.My problem is involed by hadoop fs cache. The
HoodieWrapperFileSystem is cached in fs cache when it create by first flink job
in flink taskmanger.When I launched the second flink job to write data, it used
cached HoodieWrapperFileSystem and whose childFlinkClassload is different with
this flink job.So the exception occured.
--
This message was sent by Atlassian Jira
(v8.20.1#820001)